%0 Journal Article %J terpconnect.umd.edu %D Submitted %T Surveying the Accessibility of Mobile Touchscreen Games for Persons with Motor Impairments: A Preliminary Analysis %A Kim, Y %A Sutreja, N %A Jon Froehlich %A Findlater, L %X ABSTRACT Touchscreen devices have rapidly become one of the most pervasive video game platforms in the world and, in turn, an integral part of popular culture; however, little work exists on comprehensively examining their accessibility. In this poster paper, we ... %B terpconnect.umd.edu %G eng %U http://terpconnect.umd.edu/~leahkf/pubs/ASSETS2013-TouchscreenGames.pdf %0 Journal Article %J The Analyst %D 2020 %T Rapid, quantitative therapeutic screening for Alzheimer's enzymes enabled by optimal signal transduction with transistors %A Le, Son T. %A Morris, Michelle A. %A Cardone, Antonio %A Guros, Nicholas B. %A Klauda, Jeffery B. %A Sperling, Brent A. %A Richter, Curt A. %A Pant, Harish C. %A Balijepalli, Arvind %B The Analyst %V 145 %P 2925 - 2936 %8 Feb-04-2021 %G eng %U http://xlink.rsc.org/?DOI=C9AN01804Bhttp://pubs.rsc.org/en/content/articlepdf/2020/AN/C9AN01804Bhttp://pubs.rsc.org/en/content/articlepdf/2020/AN/C9AN01804B %N 8 %! Analyst %R 10.1039/C9AN01804B %0 Journal Article %J mBio %D 2018 %T Biofilms Comprise a Component of the Annual Cycle of Vibrio cholerae in the Bay of Bengal Estuary %A Sultana, Marzia %A Nusrin, Suraia %A Hasan, Nur A. %A Sadique, Abdus %A Ahmed, Kabir U. %A Islam, Atiqul %A Hossain, Anwar %A Longini, Ira %A Nizam, Azhar %A Huq, Anwar %A Siddique, Abul K. %A Sack, David A. %A Sack, Richard B. %A Rita R Colwell %A Alam, Munirul %E Vidaver, Anne K. %X Vibrio cholerae, an estuarine bacterium, is the causative agent of cholera, a severe diarrheal disease that demonstrates seasonal incidence in Bangladesh. In an extensive study of V. cholerae occurrence in a natural aquatic environment, water and plankton samples were collected biweekly between December 2005 and November 2006 from Mathbaria, an estuarine village of Bangladesh near the mangrove forests of the Sundarbans. Toxigenic V. cholerae exhibited two seasonal growth peaks, one in spring (March to May) and another in autumn (September to November), corresponding to the two annual seasonal outbreaks of cholera in this region. The total numbers of bacteria determined by heterotrophic plate count (HPC), representing culturable bacteria, accounted for 1% to 2.7% of the total numbers obtained using acridine orange direct counting (AODC). The highest bacterial culture counts, including toxigenic V. cholerae, were recorded in the spring. The direct fluorescent antibody (DFA) assay was used to detect V. cholerae O1 cells throughout the year, as free-living cells, within clusters, or in association with plankton. V. cholerae O1 varied significantly in morphology, appearing as distinctly rod-shaped cells in the spring months, while small coccoid cells within thick clusters of biofilm were observed during interepidemic periods of the year, notably during the winter months. Toxigenic V. cholerae O1 was culturable in natural water during the spring when the temperature rose sharply. The results of this study confirmed biofilms to be a means of persistence for bacteria and an integral component of the annual life cycle of toxigenic V. cholerae in the estuarine environment of Bangladesh. %B mBio %P e00483-18 %8 Feb-05-2018 %G eng %U https://mbio.asm.org/content/9/2/e00483-18 %N 2 %! mBio %R 10.1128/mBio.00483-18 %0 Journal Article %J Frontiers in Microbiology %D 2018 %T Characterization of the Microbiome at the World’s Largest Potable Water Reuse Facility %A Stamps, Blake W. %A Leddy, Menu B. %A Plumlee, Megan H. %A Hasan, Nur A. %A Rita R Colwell %A Spear, John R. %X Conventional water resources are not sufficient in many regions to meet the needs of growing populations. Due to cyclical weather cycles, drought, and climate change, water stress has increased worldwide including in Southern California, which serves as a model for regions that integrate reuse of wastewater for both potable and non-potable use. The Orange County Water District (OCWD) Advanced Water Purification Facility (AWPF) is a highly engineered system designed to treat and produce up to 100 million gallons per day (MGD) of purified water from a municipal wastewater source for potable reuse. Routine facility microbial water quality analysis is limited to standard indicators at this and similar facilities. Given recent advances in high throughput DNA sequencing techniques, complete microbial profiling of communities in water samples is now possible. By using 16S/18S rRNA gene sequencing, metagenomic and metatranscriptomic sequencing coupled to a highly accurate identification method along with 16S rRNA gene qPCR, we describe a detailed view of the total microbial community throughout the facility. The total bacterial load of the water at stages of the treatment train ranged from 3.02 × 106 copies in source, unchlorinated wastewater feed to 5.49 × 101 copies of 16S rRNA gene/mL after treatment (consisting of microfiltration, reverse osmosis, and ultraviolet/advanced oxidation). Microbial diversity and load decreased by several orders of magnitude after microfiltration and reverse osmosis treatment, falling to almost non-detectable levels that more closely resembled controls of molecular grade laboratory water than the biomass detected in the source water. The presence of antibiotic resistance genes and viruses was also greatly reduced. Overall, system design performance was achieved, and comprehensive microbial community analysis was found to enable a more complete characterization of the water/wastewater microbial signature. Keywords: water reuse, pathogens, water purification, metatranscriptomics, metagenomics %B Frontiers in Microbiology %8 Feb-10-2020 %G eng %U https://www.frontiersin.org/articles/10.3389/fmicb.2018.02435/full %! Front. Microbiol. %R 10.3389/fmicb.2018.02435 %0 Journal Article %J Frontiers in Microbiology %D 2018 %T Comparison of Infant Gut and Skin Microbiota, Resistome and Virulome Between Neonatal Intensive Care Unit (NICU) Environments %A Hourigan, Suchitra K. %A Subramanian, Poorani %A Hasan, Nur A. %A Ta, Allison %A Klein, Elisabeth %A Chettout, Nassim %A Huddleston, Kathi %A Deopujari, Varsha %A Levy, Shira %A Baveja, R %A Clemency, Nicole C. %A Baker, Robin L. %A Niederhuber, John E. %A Rita R Colwell %B Frontiers in Microbiology %8 Jan-06-2020 %G eng %U https://www.frontiersin.org/article/10.3389/fmicb.2018.01361/full %! Front. Microbiol. %R 10.3389/fmicb.2018.01361 %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2018 %T Deep-learning-assisted Volume Visualization %A Cheng, Hsueh-Chien %A Cardone, Antonio %A Jain, Somay %A Krokos, Eric %A Narayan, Kedar %A Subramaniam, Sriram %A Varshney, Amitabh %B IEEE Transactions on Visualization and Computer Graphics %P 1 - 1 %8 Jan-01-2018 %G eng %U http://ieeexplore.ieee.org/document/8265023/http://xplorestaging.ieee.org/ielx7/2945/4359476/08265023.pdf?arnumber=8265023 %! IEEE Trans. Visual. Comput. Graphics %R 10.1109/TVCG.2018.2796085 %0 Journal Article %J Anaerobe %D 2018 %T Distinct consequences of amoxicillin and ertapenem exposure in the porcine gut microbiome %A Connelly, Sheila %A Subramanian, Poorani %A Hasan, Nur A. %A Rita R Colwell %A Kaleko, Michael %B Anaerobe %V 53 %P 82 - 93 %8 Jan-10-2018 %G eng %U https://www.sciencedirect.com/science/article/pii/S1075996418300696?via%3Dihub %! Anaerobe %R 10.1016/j.anaerobe.2018.04.012 %0 Generic %D 2018 %T High-Throughput DNA Sequencing To Profile Microbial Water Quality Of Potable Reuse %A Menu B. Leddy %A Megan H. Plumlee %A Rose S. Kantor %A Kara L. Nelson %A Scott E. Miller %A Lauren C. Kennedy %A Blake W. Stamps %A John R. Spear %A Nur A. Hasan %A Rita R Colwell %G eng %U https://www.wateronline.com/doc/high-throughput-dna-sequencing-to-profile-microbial-water-quality-of-potable-reuse-0001 %0 Journal Article %J Biological Reviews %D 2018 %T Spores and soil from six sides: interdisciplinarity and the environmental biology of anthrax (Bacillus anthracis) %A Carlson, Colin J. %A Getz, Wayne M. %A Kausrud, Kyrre L. %A Cizauskas, Carrie A. %A Blackburn, Jason K. %A Bustos Carrillo, Fausto A. %A Rita R Colwell %A Easterday, W. Ryan %A Ganz, Holly H. %A Kamath, Pauline L. %A Økstad, Ole A. %A Turner, Wendy C. %A Kolsto, Anne-Brit %A Stenseth, Nils C. %X Environmentally transmitted diseases are comparatively poorly understood and managed, and their ecology is particularly understudied. Here we identify challenges of studying environmental transmission and persistence with a six‐sided interdisciplinary review of the biology of anthrax (Bacillus anthracis). Anthrax is a zoonotic disease capable of maintaining infectious spore banks in soil for decades (or even potentially centuries), and the mechanisms of its environmental persistence have been the topic of significant research and controversy. Where anthrax is endemic, it plays an important ecological role, shaping the dynamics of entire herbivore communities. The complex eco‐epidemiology of anthrax, and the mysterious biology of Bacillus anthracis during its environmental stage, have necessitated an interdisciplinary approach to pathogen research. Here, we illustrate different disciplinary perspectives through key advances made by researchers working in Etosha National Park, a long‐term ecological research site in Namibia that has exemplified the complexities of the enzootic process of anthrax over decades of surveillance. In Etosha, the role of scavengers and alternative routes (waterborne transmission and flies) has proved unimportant relative to the long‐term persistence of anthrax spores in soil and their infection of herbivore hosts. Carcass deposition facilitates green‐ups of vegetation to attract herbivores, potentially facilitated by the role of anthrax spores in the rhizosphere. The underlying seasonal pattern of vegetation, and herbivores' immune and behavioural responses to anthrax risk, interact to produce regular ‘anthrax seasons’ that appear to be a stable feature of the Etosha ecosystem. Through the lens of microbiologists, geneticists, immunologists, ecologists, epidemiologists, and clinicians, we discuss how anthrax dynamics are shaped at the smallest scale by population genetics and interactions within the bacterial communities up to the broadest scales of ecosystem structure. We illustrate the benefits and challenges of this interdisciplinary approach to disease ecology, and suggest ways anthrax might offer insights into the biology of other important pathogens. Bacillus anthracis, and the more recently emerged Bacillus cereus biovar anthracis, share key features with other environmentally transmitted pathogens, including several zoonoses and panzootics of special interest for global health and conservation efforts. Understanding the dynamics of anthrax, and developing interdisciplinary research programs that explore environmental persistence, is a critical step forward for understanding these emerging threats. %B Biological Reviews %P 1813 - 1831 %8 Jan-11-2018 %G eng %U https://onlinelibrary.wiley.com/doi/full/10.1111/brv.12420 %N 4 %! Biol Rev %R 10.1111/brv.12420 %0 Journal Article %J Genome Announcements %D 2018 %T Virulence-Related Genes Identified from the Genome Sequence of the Non-O1/Non-O139 Vibrio cholerae Strain VcN1, Isolated from Dhaka, Bangladesh %A Hossain, Maqsud %A Alam, Munirul %A Khaleque, Abdul %A Islam, Sohidul %A Sadique, Abdus %A Khan, Nayeim %A Halim, Zahra %A Sarker, Mrinmoy %A El-Sayed, Najib M. %A Huq, Anwar %A Ahsan, Gias Uddin %A Rita R Colwell %X We report here the first draft genome sequence of the non-O1/non-O139 Vibrio cholerae strain VcN1, isolated from Dhaka, Bangladesh. The data submitted to GenBank for this strain will contribute to advancing our understanding of this environmentally disseminated bacterium, including its virulence and its evolution as an important pathogen. %B Genome Announcements %8 Aug-03-2018 %G eng %U https://mra.asm.org/content/6/10/e01513-17 %N 10 %! Genome Announc %R 10.1128/genomeA.01513-17 %0 Journal Article %J ACS Biomaterials Science & Engineering %D 2017 %T A Bioinformatics 3D Cellular Morphotyping Strategy for Assessing Biomaterial Scaffold Niches %A Florczyk, Stephen J. %A Simon, Mylene %A Juba, Derek %A Pine, P. Scott %A Sarkar, Sumona %A Chen, Desu %A Baker, Paula J. %A Bodhak, Subhadip %A Cardone, Antonio %A Brady, Mary C. %A Bajcsy, Peter %A Simon, Carl G. %B ACS Biomaterials Science & Engineering %V 3 %P 2302 - 2313 %8 Sep-10-2017 %G eng %U http://pubs.acs.org/doi/10.1021/acsbiomaterials.7b00473http://pubs.acs.org/doi/pdf/10.1021/acsbiomaterials.7b00473 %N 10 %! ACS Biomater. Sci. Eng. %R 10.1021/acsbiomaterials.7b00473 %0 Journal Article %J Journal - American Water Works Association %D 2017 %T Characterization of Microbial Signatures From Advanced Treated Wastewater Biofilms %A Leddy, Menu B. %A Hasan, Nur A. %A Subramanian, Poorani %A Heberling, Colin %A Cotruvo, Joseph %A Rita R Colwell %X Next‐generation sequencing (NGS) and metagenomics were used to identify microbial communities in biofilms of microfiltration (MF) and reverse osmosis (RO) membranes in an advanced water purification facility (AWPF) that treats municipal wastewater to produce potable quality water. Secondary treated wastewater effluent is the source of influent to the AWPF treatment train and was also characterized by NGS. Results show low bacterial diversity in biofilms obtained from the feed‐side surfaces of MF and RO membranes. Microbial communities and antibiotic resistance genes (ARGs) in the RO biofilm were compared with those of the MF and influent, revealing lower abundance of bacterial species and ARGs in the RO biofilm than in the other samples. Opportunistic pathogens were detected in all samples; however, indicator bacteria, viruses, and bacteriophages were not detected in the RO biofilm. It is concluded that NGS has great potential for improved detection and characterization of microbial communities in biofilms that form on AWPF MF and RO membranes. %B Journal - American Water Works Association %V 109 %P E503 - E512 %8 Jan-11-2017 %G eng %U http://doi.wiley.com/10.5942/jawwa.2017.109.0116 %! jawwa %R 10.5942/jawwa.2017.109.0116 %0 Journal Article %J Frontiers in Microbiology %D 2017 %T Characterization of Two Cryptic Plasmids Isolated in Haiti from Clinical Vibrio cholerae Non-O1/Non-O139 %A Ceccarelli, Daniela %A Garriss, Genevieve %A Choi, Seon Y. %A Hasan, Nur A. %A Stepanauskas, Ramunas %A Pop, Mihai %A Huq, Anwar %A Rita R Colwell %B Frontiers in Microbiology %8 Nov-11-2018 %G eng %U http://journal.frontiersin.org/article/10.3389/fmicb.2017.02283 %! Front. Microbiol. %R 10.3389/fmicb.2017.02283 %0 Journal Article %J mBio %D 2017 %T Comparative Genomics of Escherichia coli Isolated from Skin and Soft Tissue and Other Extraintestinal Infections %A Ranjan, Amit %A Shaik, Sabiha %A Nandanwar, Nishant %A Hussain, Arif %A Tiwari, Sumeet K. %A Semmler, Torsten %A Jadhav, Savita %A Wieler, Lothar H. %A Alam, Munirul %A Rita R Colwell %A Ahmed, Niyaz %E Cossart, Pascale F. %X Escherichia coli, an intestinal Gram-negative bacterium, has been shown to be associated with a variety of diseases in addition to intestinal infections, such as urinary tract infections (UTIs), meningitis in neonates, septicemia, skin and soft tissue infections (SSTIs), and colisepticemia. Thus, for nonintestinal infections, it is categorized as extraintestinal pathogenic E. coli (ExPEC). It is also an opportunistic pathogen, causing cross infections, notably as an agent of zoonotic diseases. However, comparative genomic data providing functional and genetic coordinates for ExPEC strains associated with these different types of infections have not proven conclusive. In the study reported here, ExPEC E. coli isolated from SSTIs was characterized, including virulence and drug resistance profiles, and compared with isolates from patients suffering either pyelonephritis or septicemia. Results revealed that the majority of the isolates belonged to two pathogenic phylogroups, B2 and D. Approximately 67% of the isolates were multidrug resistant (MDR), with 85% producing extended-spectrum beta-lactamase (ESBL) and 6% producing metallo-beta-lactamase (MBL). The blaCTX-M-15 genotype was observed in at least 70% of the E. coli isolates in each category, conferring resistance to an extended range of beta-lactam antibiotics. Whole-genome sequencing and comparative genomics of the ExPEC isolates revealed that two of the four isolates from SSTIs, NA633 and NA643, belong to pandemic sequence type ST131, whereas functional characteristics of three of the ExPEC pathotypes revealed that they had equal capabilities to form biofilm and were resistant to human serum. Overall, the isolates from a variety of ExPEC infections demonstrated similar resistomes and virulomes and did not display any disease-specific functional or genetic coordinates. IMPORTANCE Infections caused by extraintestinal pathogenic E. coli (ExPEC) are of global concern as they result in significant costs to health care facilities management. The recent emergence of a multidrug-resistant pandemic clone, Escherichia coli ST131, is of primary concern as a global threat. In developing countries, such as India, skin and soft tissue infections (SSTIs) associated with E. coli are marginally addressed. In this study, we employed both genomic analysis and phenotypic assays to determine relationships, if any, among the ExPEC pathotypes. Similarity between antibiotic resistance and virulence profiles was observed, ST131 isolates from SSTIs were reported, and genomic similarities among strains isolated from different disease conditions were detected. This study provides functional molecular infection epidemiology insight into SSTI-associated E. coli compared with ExPEC pathotypes. %B mBio %8 Jun-09-2017 %G eng %U http://mbio.asm.org/lookup/doi/10.1128/mBio.01070-17 %N 4 %! mBio %R 10.1128/mBio.01070-17 %0 Journal Article %J Genome Biology %D 2017 %T Comprehensive benchmarking and ensemble approaches for metagenomic classifiers %A McIntyre, Alexa B. R. %A Ounit, Rachid %A Afshinnekoo, Ebrahim %A Prill, Robert J. %A Hénaff, Elizabeth %A Alexander, Noah %A Minot, Samuel S. %A Danko, David %A Foox, Jonathan %A Ahsanuddin, Sofia %A Tighe, Scott %A Hasan, Nur A. %A Subramanian, Poorani %A Moffat, Kelly %A Levy, Shawn %A Lonardi, Stefano %A Greenfield, Nick %A Rita R Colwell %A Rosen, Gail L. %A Mason, Christopher E. %B Genome Biology %8 Jan-12-2017 %G eng %U http://genomebiology.biomedcentral.com/articles/10.1186/s13059-017-1299-7 %N 1210 %! Genome Biol %R 10.1186/s13059-017-1299-7 %0 Conference Proceedings %B IEEE International Conference on Image Processing %D 2017 %T Deep-learning-assisted visualization for live-cell images %A Hsueh-Chien Cheng %A Cardone, Antonio %A Krokos, Eric %A Stoica, Bogdan %A Faden, Alan %A Varshney, Amitabh %K deep learning %K live-cell images %K Visualization %X Analyzing live-cell images is particularly challenging because cells move at the same time they undergo systematic changes. Visually inspecting live-cell images therefore involves simultaneously tracking individual cells and detecting relevant spatio-temporal changes. The high cognitive burden of such a complex task makes this kind of analysis inefficient and error-prone. In this paper we describe a deep-learning-assisted visualization based on automatically derived high-level features to identify target cell changes in live-cell images. Applying a novel user-mediated color assignment scheme that maps abstract features into corresponding colors, we create color-based visual annotations that facilitate visual reasoning and analysis of complex time varying live-cell imagery datasets. The visual representations can be used to study temporal changes in cells, such as the morphological changes in cell at various stages of life cycle. %B IEEE International Conference on Image Processing %I IEEE %C Beijing, China %8 09/2017 %G eng %0 Journal Article %J Journal of Biomolecular Techniques : JBT %D 2017 %T Genomic Methods and Microbiological Technologies for Profiling Novel and Extreme Environments for the Extreme Microbiome Project (XMP) %A Tighe, Scott %A Afshinnekoo, Ebrahim %A Rock, Tara M. %A McGrath, Ken %A Alexander, Noah %A McIntyre, Alexa %A Ahsanuddin, Sofia %A Bezdan, Daniela %A Green, Stefan J. %A Joye, Samantha %A Stewart Johnson, Sarah %A Baldwin, Don A. %A Bivens, Nathan %A Ajami, Nadim %A Carmical, Joseph R. %A Herriott, Ian Charold %A Rita R Colwell %A Donia, Mohamed %A Foox, Jonathan %A Greenfield, Nick %A Hunter, Tim %A Hoffman, Jessica %A Hyman, Joshua %A Jorgensen, Ellen %A Krawczyk, Diana %A Lee, Jodie %A Levy, Shawn %A Garcia-Reyero, àlia %A Settles, Matthew %A Thomas, Kelley %A ómez, Felipe %A Schriml, Lynn %A Kyrpides, Nikos %A Zaikova, Elena %A Penterman, Jon %A Mason, Christopher E. %X The Extreme Microbiome Project (XMP) is a project launched by the Association of Biomolecular Resource Facilities Metagenomics Research Group (ABRF MGRG) that focuses on whole genome shotgun sequencing of extreme and unique environments using a wide variety of biomolecular techniques. The goals are multifaceted, including development and refinement of new techniques for the following: 1) the detection and characterization of novel microbes, 2) the evaluation of nucleic acid techniques for extremophilic samples, and 3) the identification and implementation of the appropriate bioinformatics pipelines. Here, we highlight the different ongoing projects that we have been working on, as well as details on the various methods we use to characterize the microbiome and metagenome of these complex samples. In particular, we present data of a novel multienzyme extraction protocol that we developed, called Polyzyme or MetaPolyZyme. Presently, the XMP is characterizing sample sites around the world with the intent of discovering new species, genes, and gene clusters. Once a project site is complete, the resulting data will be publically available. Sites include Lake Hillier in Western Australia, the “Door to Hell” crater in Turkmenistan, deep ocean brine lakes of the Gulf of Mexico, deep ocean sediments from Greenland, permafrost tunnels in Alaska, ancient microbial biofilms from Antarctica, Blue Lagoon Iceland, Ethiopian toxic hot springs, and the acidic hypersaline ponds in Western Australia. %B Journal of Biomolecular Techniques : JBT %V 28 %P 31 - 39 %8 Jan-04-2017 %G eng %U https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5345951/ %N 1 %! J Biomol Tech %R 10.7171/jbt.17-2801-004 %0 Journal Article %J Frontiers in Microbiology %D 2017 %T Vibrio cholerae O1 with Reduced Susceptibility to Ciprofloxacin and Azithromycin Isolated from a Rural Coastal Area of Bangladesh %A Rashed, Shah M. %A Hasan, Nur A. %A Alam, Munirul %A Sadique, Abdus %A Sultana, Marzia %A Hoq, Md. Mozammel %A Sack, R. Bradley %A Rita R Colwell %A Huq, Anwar %X Cholera outbreaks occur each year in the remote coastal areas of Bangladesh and epidemiological surveillance and routine monitoring of cholera in these areas is challenging. In this study, a total of 97 Vibrio cholerae O1 isolates from Mathbaria, Bangladesh, collected during 2010 and 2014 were analyzed for phenotypic and genotypic traits, including antimicrobial susceptibility. Of the 97 isolates, 95 possessed CTX-phage mediated genes, ctxA, ace, and zot, and two lacked the cholera toxin gene, ctxA. Also both CTX+ and CTX− V. cholerae O1 isolated in this study carried rtxC, tcpAET, and hlyA. The classical cholera toxin gene, ctxB1, was detected in 87 isolates, while eight had ctxB7. Of 95 CTX+ V. cholerae O1, 90 contained rstRET and 5 had rstRCL. All isolates, except two, contained SXT related integrase intSXT. Resistance to penicillin, streptomycin, nalidixic acid, sulfamethoxazole-trimethoprim, erythromycin, and tetracycline varied between the years of study period. Most importantly, 93% of the V. cholerae O1 were multidrug resistant. Six different resistance profiles were observed, with resistance to streptomycin, nalidixic acid, tetracycline, and sulfamethoxazole-trimethoprim predominant every year. Ciprofloxacin and azithromycin MIC were 0.003–0.75 and 0.19–2.00 μg/ml, respectively, indicating reduced susceptibility to these antibiotics. Sixteen of the V. cholerae O1 isolates showed higher MIC for azithromycin (≥0.5 μg/ml) and were further examined for 10 macrolide resistance genes, erm(A), erm(B), erm(C), ere(A), ere(B), mph(A), mph(B), mph(D), mef(A), and msr(A) with none testing positive for the macrolide resistance genes. %B Frontiers in Microbiology %8 Sep-02-2018 %G eng %U http://journal.frontiersin.org/article/10.3389/fmicb.2017.00252 %! Front. Microbiol. %R 10.3389/fmicb.2017.00252 %0 Journal Article %J Water %D 2017 %T Membrane Bioreactor-Based Wastewater Treatment Plant in Saudi Arabia: Reduction of Viral Diversity, Load, and Infectious Capacity %A Jumat, Muhammad %A Hasan, Nur %A Subramanian, Poorani %A Heberling, Colin %A Rita R Colwell %A Hong, Pei-Ying %B Water %V 96046Volume 70 %P 534 %8 Jan-07-2017 %G eng %U http://www.mdpi.com/2073-4441/9/7/534 %N 7 %! Water %R 10.3390/w9070534 %0 Journal Article %J Scientific Reports %D 2017 %T The microbiomes of blowflies and houseflies as bacterial transmission reservoirs %A Junqueira, AC %A Ratan, Aakrosh %A Acerbi, Enzo %A Drautz-Moses, Daniela I. %A Premkrishnan, BNV %A Costea, PI %A Linz, Bodo %A Purbojati, Rikky W. %A Paulo, Daniel F. %A Gaultier, Nicolas E. %A Subramanian, Poorani %A Hasan, Nur A. %A Rita R Colwell %A Bork, Peer %A Azeredo-Espin, Ana Maria L. %A Bryant, Donald A. %A Schuster, Stephan C. %X Blowflies and houseflies are mechanical vectors inhabiting synanthropic environments around the world. They feed and breed in fecal and decaying organic matter, but the microbiome they harbour and transport is largely uncharacterized. We sampled 116 individual houseflies and blowflies from varying habitats on three continents and subjected them to high-coverage, whole-genome shotgun sequencing. This allowed for genomic and metagenomic analyses of the host-associated microbiome at the species level. Both fly host species segregate based on principal coordinate analysis of their microbial communities, but they also show an overlapping core microbiome. Legs and wings displayed the largest microbial diversity and were shown to be an important route for microbial dispersion. The environmental sequencing approach presented here detected a stochastic distribution of human pathogens, such as Helicobacter pylori, thereby demonstrating the potential of flies as proxies for environmental and public health surveillance. %B Scientific Reports %8 Jan-12-2017 %G eng %U http://www.nature.com/articles/s41598-017-16353-x %N 1 %! Sci Rep %R 10.1038/s41598-017-16353-x %0 Journal Article %J Journal of Applied Microbiology %D 2017 %T SYN-004 (ribaxamase), an oral beta-lactamase, mitigates antibiotic-mediated dysbiosis in a porcine gut microbiome model %A Connelly, S. %A Bristol, J.A. %A Hubert, S. %A Subramanian, P. %A Hasan, N.A. %A Rita R Colwell %A Kaleko, M. %X Aim To evaluate an antibiotic inactivation strategy to protect the gut microbiome from antibiotic‐mediated damage. Methods and Results SYN‐004 (ribaxamase) is an orally delivered beta‐lactamase intended to degrade penicillins and cephalosporins within the gastrointestinal tract to protect the microbiome. Pigs (20 kg, n = 10) were treated with ceftriaxone (CRO) (IV, 50 mg kg−1, SID) for 7 days and a cohort (n = 5) received ribaxamase (PO, 75 mg, QID) for 9 days beginning the day before antibiotic administration. Ceftriaxone serum levels were not statistically different in the antibiotic‐alone and antibiotic + ribaxamase groups, indicating ribaxamase did not alter systemic antibiotic levels. Whole‐genome metagenomic analyses of pig faecal DNA revealed that CRO caused significant changes to the gut microbiome and an increased frequency of antibiotic resistance genes. With ribaxamase, the gut microbiomes were not significantly different from pretreatment and antibiotic resistance gene frequency was not increased. Conclusion Ribaxamase mitigated CRO‐mediated gut microbiome dysbiosis and attenuated propagation of the antibiotic resistance genes in pigs. Significance and Impact of the Study Damage of the microbiome can lead to overgrowth of pathogenic organisms and antibiotic exposure can promote selection for antibiotic‐resistant micro‐organisms. Ribaxamase has the potential to become the first therapy designed to protect the gut microbiome from antibiotic‐mediated dysbiosis and reduce emergence of antibiotic resistance. %B Journal of Applied Microbiology %P 66 - 79 %8 Jan-07-2017 %G eng %U https://onlinelibrary.wiley.com/doi/10.1111/jam.13432 %N 1Suppl 11 %! J Appl Microbiol %R 10.1111/jam.13432 %0 Conference Paper %B 2016 32nd Southern Biomedical Engineering Conference (SBEC) %D 2016 %T 3D Cellular Morphotyping of Scaffold Niches %A Florczyk, Stephen J %A Simon, Mylene %A Juba, Derek %A Pine, P Scott %A Sarkar, Sumona %A Chen, Desu %A Baker, Paula J %A Bodhak, Subhadip %A Cardone, Antonio %A Brady, Mary %A others %B 2016 32nd Southern Biomedical Engineering Conference (SBEC) %I IEEE %G eng %0 Journal Article %J FEMS Microbiology Ecology %D 2016 %T Chitin promotes Mycobacterium ulcerans growth %A Sanhueza, Daniel %A Chevillon, Christine %A Rita R Colwell %A Babonneau, Jérémie %A Marion, Estelle %A Marsollier, Laurent %A égan, çois %E Sobecky, Patricia %X Mycobacterium ulcerans (MU) is the causative agent of Buruli ulcer, an emerging human infectious disease. However, both the ecology and life cycle of MU are poorly understood. The occurrence of MU has been linked to the aquatic environment, notably water bodies affected by human activities. It has been hypothesized that one or a combination of environmental factor(s) connected to human activities could favour growth of MU in aquatic systems. Here, we tested in vitro the growth effect of two ubiquitous polysaccharides and five chemical components on MU at concentration ranges shown to occur in endemic regions. Real-time PCR showed that chitin increased MU growth significantly providing a nutrient source or environmental support for the bacillus, thereby, providing a focus on the association between MU and aquatic arthropods. Aquatic environments with elevated population of arthropods provide increased chitin availability and, thereby, enhanced multiplication of MU. If calcium very slightly enhanced MU growth, iron, zinc, sulphate and phosphate did not stimulate MU growth, and at the concentration ranges of this study would limit MU population in natural ecosystems. %B FEMS Microbiology Ecology %V 928565871 %P fiw067 %8 Jun-06-2017 %G eng %U https://academic.oup.com/femsec/article-lookup/doi/10.1093/femsec/fiw067 %N 6 %! FEMS Microbiology Ecology %R 10.1093/femsec/fiw067 %0 Journal Article %J FEMS Microbiology Letters %D 2016 %T Comparison of inferred relatedness based on multilocus variable-number tandem-repeat analysis and whole genome sequencing of Vibrio cholerae O1 %A Rashid, Mahamud-ur %A Almeida, Mathieu %A Azman, Andrew S. %A Lindsay, Brianna R. %A Sack, David A. %A Rita R Colwell %A Huq, Anwar %A Morris, J. Glenn %A Alam, Munirul %A Stine, O. Colin %E Winstanley, Craig %X Vibrio cholerae causes cholera, a severe diarrheal disease. Understanding the local genetic diversity and transmission of V. cholerae will improve our ability to control cholera. Vibrio cholerae isolates clustered in genetically related groups (clonal complexes, CC) by multilocus variable tandem-repeat analysis (MLVA) were compared by whole genome sequencing (WGS). Isolates in CC1 had been isolated from two geographical locations. Isolates in a second genetically distinct group, CC2, were isolated only at one location. Using WGS, CC1 isolates from both locations revealed, on average, 43.8 nucleotide differences, while those strains comprising CC2 averaged 19.7 differences. Strains from both MLVA-CCs had an average difference of 106.6. Thus, isolates comprising CC1 were more closely related (P < 10−6) to each other than to isolates in CC2. Within a MLVA-CC, after removing all paralogs, alternative alleles were found in all possible combinations on separate chromosomes indicative of recombination within the core genome. Including recombination did not affect the distinctiveness of the MLVA-CCs when measured by WGS. We found that WGS generally reflected the same genetic relatedness of isolates as MLVA, indicating that isolates from the same MLVA-CC shared a more recent common ancestor than isolates from the same location that clustered in a distinct MLVA-CC. %B FEMS Microbiology Letters %V 36389 %P fnw116 %8 Nov-06-2017 %G eng %U https://academic.oup.com/femsle/article-lookup/doi/10.1093/femsle/fnw116 %N 12 %! FEMS Microbiology Letters %R 10.1093/femsle/fnw116 %0 Journal Article %J Journal of Computer-Aided Molecular Design %D 2016 %T Computational study of the inhibitory mechanism of the kinase CDK5 hyperactivity by peptide p5 and derivation of a pharmacophore %A Cardone, Antonio %A Brady, M. %A Sriram, R. %A Pant, H. C. %A Hassan, S. A. %B Journal of Computer-Aided Molecular Design %V 30 %P 513 - 521 %8 Jan-06-2016 %G eng %U http://link.springer.com/10.1007/s10822-016-9922-3http://link.springer.com/content/pdf/10.1007/s10822-016-9922-3http://link.springer.com/content/pdf/10.1007/s10822-016-9922-3.pdfhttp://link.springer.com/article/10.1007/s10822-016-9922-3/fulltext.html %N 6 %! J Comput Aided Mol Des %R 10.1007/s10822-016-9922-3 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2016 %T Cross-talk among flesh-eating Aeromonas hydrophila strains in mixed infection leading to necrotizing fasciitis %A Ponnusamy, Duraisamy %A Kozlova, Elena V. %A Sha, Jian %A Erova, Tatiana E. %A Azar, Sasha R. %A Fitts, Eric C. %A Kirtley, Michelle L. %A Tiner, Bethany L. %A Andersson, Jourdan A. %A Grim, Christopher J. %A Isom, Richard P. %A Hasan, Nur A. %A Rita R Colwell %A Chopra, Ashok K. %X Necrotizing fasciitis (NF) caused by flesh-eating bacteria is associated with high case fatality. In an earlier study, we reported infection of an immunocompetent individual with multiple strains of Aeromonas hydrophila (NF1–NF4), the latter three constituted a clonal group whereas NF1 was phylogenetically distinct. To understand the complex interactions of these strains in NF pathophysiology, a mouse model was used, whereby either single or mixed A. hydrophila strains were injected intramuscularly. NF2, which harbors exotoxin A (exoA) gene, was highly virulent when injected alone, but its virulence was attenuated in the presence of NF1 (exoA-minus). NF1 alone, although not lethal to animals, became highly virulent when combined with NF2, its virulence augmented by cis-exoA expression when injected alone in mice. Based on metagenomics and microbiological analyses, it was found that, in mixed infection, NF1 selectively disseminated to mouse peripheral organs, whereas the other strains (NF2, NF3, and NF4) were confined to the injection site and eventually cleared. In vitro studies showed NF2 to be more effectively phagocytized and killed by macrophages than NF1. NF1 inhibited growth of NF2 on solid media, but ExoA of NF2 augmented virulence of NF1 and the presence of NF1 facilitated clearance of NF2 from animals either by enhanced priming of host immune system or direct killing via a contact-dependent mechanism. %B Proceedings of the National Academy of Sciences %V 11312161268 %P 722 - 727 %8 Jul-01-2017 %G eng %U http://www.pnas.org/lookup/doi/10.1073/pnas.1523817113 %N 3321029 %! Proc Natl Acad Sci USA %R 10.1073/pnas.1523817113 %0 Journal Article %J BMC Microbiology %D 2016 %T Enrichment dynamics of Listeria monocytogenes and the associated microbiome from naturally contaminated ice cream linked to a listeriosis outbreak %A Ottesen, Andrea %A Ramachandran, Padmini %A Reed, Elizabeth %A White, James R. %A Hasan, Nur %A Subramanian, Poorani %A Ryan, Gina %A Jarvis, Karen %A Grim, Christopher %A Daquiqan, Ninalynn %A Hanes, Darcy %A Allard, Marc %A Rita R Colwell %A Brown, Eric %A Chen, Yi %B BMC Microbiology %8 Jan-12-2016 %G eng %U http://bmcmicrobiol.biomedcentral.com/articles/10.1186/s12866-016-0894-1 %! BMC Microbiol %R 10.1186/s12866-016-0894-1 %0 Journal Article %J National Science Review %D 2016 %T Modeling Sustainability: Population, Inequality, Consumption, and Bidirectional Coupling of the Earth and Human Systems %A Motesharrei, Safa %A Rivas, Jorge %A Kalnay, Eugenia %A Asrar, Ghassem R. %A Busalacchi, Antonio J. %A Cahalan, Robert F. %A Cane, Mark A. %A Rita R Colwell %A Feng, Kuishuang %A Franklin, Rachel S. %A Hubacek, Klaus %A Miralles-Wilhelm, Fernando %A Miyoshi, Takemasa %A Ruth, Matthias %A Sagdeev, Roald %A Shirmohammadi, Adel %A Shukla, Jagadish %A Srebric, Jelena %A Yakovenko, Victor M. %A Zeng, Ning %X Over the last two centuries, the impact of the Human System has grown dramatically, becoming strongly dominant within the Earth System in many different ways. Consumption, inequality, and population have increased extremely fast, especially since about 1950, threatening to overwhelm the many critical functions and ecosystems of the Earth System. Changes in the Earth System, in turn, have important feedback effects on the Human System, with costly and potentially serious consequences. However, current models do not incorporate these critical feedbacks. We argue that in order to understand the dynamics of either system, Earth System Models must be coupled with Human System Models through bidirectional couplings representing the positive, negative, and delayed feedbacks that exist in the real systems. In particular, key Human System variables, such as demographics, inequality, economic growth, and migration, are not coupled with the Earth System but are instead driven by exogenous estimates, such as United Nations population projections. This makes current models likely to miss important feedbacks in the real Earth–Human system, especially those that may result in unexpected or counterintuitive outcomes, and thus requiring different policy interventions from current models. The importance and imminence of sustainability challenges, the dominant role of the Human System in the Earth System, and the essential roles the Earth System plays for the Human System, all call for collaboration of natural scientists, social scientists, and engineers in multidisciplinary research and modeling to develop coupled Earth–Human system models for devising effective science-based policies and measures to benefit current and future generations. %B National Science Review %P nww081 %8 Nov-12-2016 %G eng %U https://academic.oup.com/nsr/article-lookup/doi/10.1093/nsr/nww081 %! Nat. Sci. Rev. %R 10.1093/nsr/nww081 %0 Journal Article %J mBio %D 2016 %T Phylogenetic Diversity of Vibrio cholerae Associated with Endemic Cholera in Mexico from 1991 to 2008 %A Choi, Seon Y %A Rashed, Shah M. %A Hasan, Nur A. %A Alam, Munirul %A Islam, Tarequl %A Sadique, Abdus %A Johura, Fatema-Tuz %A Eppinger, Mark %A Ravel, Jacques %A Huq, Anwar %A Cravioto, Alejandro %A Rita R Colwell %X An outbreak of cholera occurred in 1991 in Mexico, where it had not been reported for more than a century and is now endemic. Vibrio cholerae O1 prototype El Tor and classical strains coexist with altered El Tor strains (1991 to 1997). Nontoxigenic (CTX−) V. cholerae El Tor dominated toxigenic (CTX+) strains (2001 to 2003), but V. cholerae CTX+ variant El Tor was isolated during 2004 to 2008, outcompeting CTX− V. cholerae. Genomes of six Mexican V. cholerae O1 strains isolated during 1991 to 2008 were sequenced and compared with both contemporary and archived strains of V. cholerae. Three were CTX+ El Tor, two were CTX− El Tor, and the remaining strain was a CTX+ classical isolate. Whole-genome sequence analysis showed the six isolates belonged to five distinct phylogenetic clades. One CTX− isolate is ancestral to the 6th and 7th pandemic CTX+ V. cholerae isolates. The other CTX− isolate joined with CTX− non-O1/O139 isolates from Haiti and seroconverted O1 isolates from Brazil and Amazonia. One CTX+ isolate was phylogenetically placed with the sixth pandemic classical clade and the V. cholerae O395 classical reference strain. Two CTX+ El Tor isolates possessing intact Vibrio seventh pandemic island II (VSP-II) are related to hybrid El Tor isolates from Mozambique and Bangladesh. The third CTX+ El Tor isolate contained West African-South American (WASA) recombination in VSP-II and showed relatedness to isolates from Peru and Brazil. Except for one isolate, all Mexican isolates lack SXT/R391 integrative conjugative elements (ICEs) and sensitivity to selected antibiotics, with one isolate resistant to streptomycin. No isolates were related to contemporary isolates from Asia, Africa, or Haiti, indicating phylogenetic diversity. %B mBio %V 7 %8 Apr-05-2016 %G eng %U http://mbio.asm.org/lookup/doi/10.1128/mBio.02160-15 %N 2 %! mBio %R 10.1128/mBio.02160-15 %0 Journal Article %J The Milbank Quarterly %D 2016 %T Strategic Planning in Population Health and Public Health Practice: A Call to Action for Higher Education %A Phelps, Charles %A Madhavan, Guruprasad %A Rappuoli, Rino %A LEVIN, SCOTT %A Shortlife, Edward %A Rita R Colwell %X POLICY POINTS: Scarce resources, especially in population health and public health practice, underlie the importance of strategic planning. Public health agencies' current planning and priority setting efforts are often narrow, at times opaque, and focused on single metrics such as cost-effectiveness. As demonstrated by SMART Vaccines, a decision support software system developed by the Institute of Medicine and the National Academy of Engineering, new approaches to strategic planning allow the formal incorporation of multiple stakeholder views and multicriteria decision making that surpass even those sophisticated cost-effectiveness analyses widely recommended and used for public health planning. Institutions of higher education can and should respond by building on modern strategic planning tools as they teach their students how to improve population health and public health practice. CONTEXT: Strategic planning in population health and public health practice often uses single indicators of success or, when using multiple indicators, provides no mechanism for coherently combining the assessments. Cost-effectiveness analysis, the most complex strategic planning tool commonly applied in public health, uses only a single metric to evaluate programmatic choices, even though other factors often influence actual decisions. METHODS: Our work employed a multicriteria systems analysis approach--specifically, multiattribute utility theory--to assist in strategic planning and priority setting in a particular area of health care (vaccines), thereby moving beyond the traditional cost-effectiveness analysis approach. FINDINGS: (1) Multicriteria systems analysis provides more flexibility, transparency, and clarity in decision support for public health issues compared with cost-effectiveness analysis. (2) More sophisticated systems-level analyses will become increasingly important to public health as disease burdens increase and the resources to deal with them become scarcer. CONCLUSIONS: The teaching of strategic planning in public health must be expanded in order to fill a void in the profession's planning capabilities. Public health training should actively incorporate model building, promote the interactive use of software tools, and explore planning approaches that transcend restrictive assumptions of cost-effectiveness analysis. The Strategic Multi-Attribute Ranking Tool for Vaccines (SMART Vaccines), which was recently developed by the Institute of Medicine and the National Academy of Engineering to help prioritize new vaccine development, is a working example of systems analysis as a basis for decision support. %B The Milbank Quarterly %V 94906033 %P 109 - 125 %8 Jan-03-2016 %G eng %U https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4941964/ %N 143S2 %! The Milbank Quarterly %R 10.1111/1468-0009.12182 %0 Journal Article %J Frontiers in Public Health %D 2015 %T Building Infectious Disease Research Programs to Promote Security and Enhance Collaborations with Countries of the Former Soviet Union %A Bartholomew, James C. %A Pearson, Andrew D. %A Stenseth, Nils C. %A LeDuc, James W. %A Hirschberg, David L. %A Rita R Colwell %B Frontiers in Public Health %V 35361632 %8 Feb-11-2017 %G eng %U http://journal.frontiersin.org/Article/10.3389/fpubh.2015.00271/abstract %! Front. Public Health %R 10.3389/fpubh.2015.00271 %0 Journal Article %J J Comput Chem %D 2015 %T Detection and characterization of nonspecific, sparsely populated binding modes in the early stages of complexation. %A Cardone, Antonio %A Bornstein, Aaron %A Pant, Harish C %A Brady, Mary %A Sriram, Ram %A Hassan, Sergio A %X

A method is proposed to study protein-ligand binding in a system governed by specific and nonspecific interactions. Strong associations lead to narrow distributions in the proteins configuration space; weak and ultraweak associations lead instead to broader distributions, a manifestation of nonspecific, sparsely populated binding modes with multiple interfaces. The method is based on the notion that a discrete set of preferential first-encounter modes are metastable states from which stable (prerelaxation) complexes at equilibrium evolve. The method can be used to explore alternative pathways of complexation with statistical significance and can be integrated into a general algorithm to study protein interaction networks. The method is applied to a peptide-protein complex. The peptide adopts several low-population conformers and binds in a variety of modes with a broad range of affinities. The system is thus well suited to analyze general features of binding, including conformational selection, multiplicity of binding modes, and nonspecific interactions, and to illustrate how the method can be applied to study these problems systematically. The equilibrium distributions can be used to generate biasing functions for simulations of multiprotein systems from which bulk thermodynamic quantities can be calculated. © 2015 Wiley Periodicals, Inc.

%B J Comput Chem %8 2015 Mar 18 %G eng %R 10.1002/jcc.23883 %0 Journal Article %J mBio %D 2015 %T Hybrid Vibrio cholerae El Tor Lacking SXT Identified as the Cause of a Cholera Outbreak in the Philippines %A Klinzing, David C. %A Choi, Seon Young %A Hasan, Nur A. %A Matias, Ronald R. %A Tayag, Enrique %A Geronimo, Josefina %A Skowronski, Evan %A Rashed, Shah M. %A Kawashima, Kent %A Rosenzweig, C. Nicole %A Gibbons, Henry S. %A Torres, Brian C. %A Liles, Veni %A Alfon, Alicia C. %A Juan, Maria Luisa %A Natividad, Filipinas F. %A Cebula, Thomas A. %A Rita R Colwell %X Cholera continues to be a global threat, with high rates of morbidity and mortality. In 2011, a cholera outbreak occurred in Palawan, Philippines, affecting more than 500 people, and 20 individuals died. Vibrio cholerae O1 was confirmed as the etiological agent. Source attribution is critical in cholera outbreaks for proper management of the disease, as well as to control spread. In this study, three V. cholerae O1 isolates from a Philippines cholera outbreak were sequenced and their genomes analyzed to determine phylogenetic relatedness to V. cholerae O1 isolates from recent outbreaks of cholera elsewhere. The Philippines V. cholerae O1 isolates were determined to be V. cholerae O1 hybrid El Tor belonging to the seventh-pandemic clade. They clustered tightly, forming a monophyletic clade closely related to V. cholerae O1 hybrid El Tor from Asia and Africa. The isolates possess a unique multilocus variable-number tandem repeat analysis (MLVA) genotype (12-7-9-18-25 and 12-7-10-14-21) and lack SXT. In addition, they possess a novel 15-kb genomic island (GI-119) containing a predicted type I restriction-modification system. The CTXΦ-RS1 array of the Philippines isolates was similar to that of V. cholerae O1 MG116926, a hybrid El Tor strain isolated in Bangladesh in 1991. Overall, the data indicate that the Philippines V. cholerae O1 isolates are unique, differing from recent V. cholerae O1 isolates from Asia, Africa, and Haiti. Furthermore, the results of this study support the hypothesis that the Philippines isolates of V. cholerae O1 are indigenous and exist locally in the aquatic ecosystem of the Philippines. %B mBio %8 Jan-05-2015 %G eng %U http://mbio.asm.org/lookup/doi/10.1128/mBio.00047-15 %N 2 %! mBio %R 10.1128/mBio.00047-15 %0 Journal Article %J PLOS ONE %D 2015 %T Satellite Based Assessment of Hydroclimatic Conditions Related to Cholera in Zimbabwe %A Jutla, Antarpreet %A Aldaach, Haidar %A Billian, Hannah %A Akanda, Ali %A Huq, Anwar %A Rita R Colwell %E Schumann, Guy J-P. %X Introduction Cholera, an infectious diarrheal disease, has been shown to be associated with large scale hydroclimatic processes. The sudden and sporadic occurrence of epidemic cholera is linked with high mortality rates, in part, due to uncertainty in timing and location of outbreaks. Improved understanding of the relationship between pathogenic abundance and climatic processes allows prediction of disease outbreak to be an achievable goal. In this study, we show association of large scale hydroclimatic processes with the cholera epidemic in Zimbabwe reported to have begun in Chitungwiza, a city in Mashonaland East province, in August, 2008. Principal Findings Climatic factors in the region were found to be associated with triggering cholera outbreak and are shown to be related to anomalies of temperature and precipitation, validating the hypothesis that poor conditions of sanitation, coupled with elevated temperatures, and followed by heavy rainfall can initiate outbreaks of cholera. Spatial estimation by satellite of precipitation and global gridded air temperature captured sensitivities in hydroclimatic conditions that permitted identification of the location in the region where the disease outbreak began. Discussion Satellite derived hydroclimatic processes can be used to capture environmental conditions related to epidemic cholera, as occurred in Zimbabwe, thereby providing an early warning system. Since cholera cannot be eradicated because the causative agent, Vibrio cholerae, is autochthonous to the aquatic environment, prediction of conditions favorable for its growth and estimation of risks of triggering the disease in a given population can be used to alert responders, potentially decreasing infection and saving lives. %B PLOS ONE %P e0137828 %8 May-09-2017 %G eng %U https://dx.plos.org/10.1371/journal.pone.0137828 %N 9Suppl 1 %! PLoS ONE %R 10.1371/journal.pone.0137828 %0 Journal Article %J BMC bioinformatics %D 2015 %T Survey statistics of automated segmentations applied to optical imaging of mammalian cells %A Bajcsy, Peter %A Cardone, Antonio %A Chalfoun, Joe %A Halter, Michael %A Juba, Derek %A Kociolek, Marcin %A Majurski, Michael %A Peskin, Adele %A Simon, Carl %A Simon, Mylene %A others %B BMC bioinformatics %V 16 %P 1 %G eng %0 Journal Article %J PLoS ONE %D 2014 %T Microbial Community Profiling of Human Saliva Using Shotgun Metagenomic Sequencing %A Hasan, Nur A. %A Young, Brian A. %A Minard-Smith, Angela T. %A Saeed, Kelly %A Li, Huai %A Heizer, Esley M. %A McMillan, Nancy J. %A Isom, Richard %A Abdullah, Abdul Shakur %A Bornman, Daniel M. %A Faith, Seth A. %A Choi, Seon Young %A Dickens, Michael L. %A Cebula, Thomas A. %A Rita R Colwell %E Ahmed, Niyaz %X Human saliva is clinically informative of both oral and general health. Since next generation shotgun sequencing (NGS) is now widely used to identify and quantify bacteria, we investigated the bacterial flora of saliva microbiomes of two healthy volunteers and five datasets from the Human Microbiome Project, along with a control dataset containing short NGS reads from bacterial species representative of the bacterial flora of human saliva. GENIUS, a system designed to identify and quantify bacterial species using unassembled short NGS reads was used to identify the bacterial species comprising the microbiomes of the saliva samples and datasets. Results, achieved within minutes and at greater than 90% accuracy, showed more than 175 bacterial species comprised the bacterial flora of human saliva, including bacteria known to be commensal human flora but also Haemophilus influenzae, Neisseria meningitidis, Streptococcus pneumoniae, and Gamma proteobacteria. Basic Local Alignment Search Tool (BLASTn) analysis in parallel, reported ca. five times more species than those actually comprising the in silico sample. Both GENIUS and BLAST analyses of saliva samples identified major genera comprising the bacterial flora of saliva, but GENIUS provided a more precise description of species composition, identifying to strain in most cases and delivered results at least 10,000 times faster. Therefore, GENIUS offers a facile and accurate system for identification and quantification of bacterial species and/or strains in metagenomic samples. %B PLoS ONE %P e97699 %8 Aug-05-2015 %G eng %U https://dx.plos.org/10.1371/journal.pone.0097699 %N 5 %! PLoS ONE %R 10.1371/journal.pone.0097699 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2014 %T Occurrence in Mexico, 1998–2008, of Vibrio cholerae CTX + El Tor carrying an additional truncated CTX prophage %A Alam, Munirul %A Rashed, Shah M %A Mannan, Shahnewaj Bin %A Islam, Tarequl %A Lizarraga-Partida, Marcial L. %A Delgado, Gabriela %A Morales-Espinosa, Rosario %A Mendez, Jose Luis %A Navarro, Armando %A Watanabe, Haruo %A Ohnishi, Makoto %A Hasan, Nur A. %A Huq, Anwar %A Sack, R. Bradley %A Rita R Colwell %A Cravioto, Alejandro %B Proceedings of the National Academy of Sciences %V 111 %P 9917 - 9922 %8 Aug-07-2014 %G eng %U http://www.pnas.org/lookup/doi/10.1073/pnas.1323408111 %N 27 %! Proc Natl Acad Sci USA %R 10.1073/pnas.1323408111 %0 Journal Article %J mBio %D 2014 %T Phylodynamic Analysis of Clinical and Environmental Vibrio cholerae Isolates from Haiti Reveals Diversification Driven by Positive Selection %A Azarian, Taj %A Ali, Afsar %A Johnson, Judith A. %A Mohr, David %A Prosperi, Mattia %A Veras, Nazle M. %A Jubair, Mohammed %A Strickland, Samantha L. %A Rashid, Mohammad H. %A Alam, Meer T. %A Weppelmann, Thomas A. %A Katz, Lee S. %A Tarr, Cheryl L. %A Rita R Colwell %A Morris, J. Glenn %A Salemi, Marco %X Phylodynamic analysis of genome-wide single-nucleotide polymorphism (SNP) data is a powerful tool to investigate underlying evolutionary processes of bacterial epidemics. The method was applied to investigate a collection of 65 clinical and environmental isolates of Vibrio cholerae from Haiti collected between 2010 and 2012. Characterization of isolates recovered from environmental samples identified a total of four toxigenic V. cholerae O1 isolates, four non-O1/O139 isolates, and a novel nontoxigenic V. cholerae O1 isolate with the classical tcpA gene. Phylogenies of strains were inferred from genome-wide SNPs using coalescent-based demographic models within a Bayesian framework. A close phylogenetic relationship between clinical and environmental toxigenic V. cholerae O1 strains was observed. As cholera spread throughout Haiti between October 2010 and August 2012, the population size initially increased and then fluctuated over time. Selection analysis along internal branches of the phylogeny showed a steady accumulation of synonymous substitutions and a progressive increase of nonsynonymous substitutions over time, suggesting diversification likely was driven by positive selection. Short-term accumulation of nonsynonymous substitutions driven by selection may have significant implications for virulence, transmission dynamics, and even vaccine efficacy. %B mBio %8 Jul-12-2016 %G eng %U http://mbio.asm.org/lookup/doi/10.1128/mBio.01824-14 %N 6 %! mBio %R 10.1128/mBio.01824-14 %0 Journal Article %J Investigative Genetics %D 2014 %T Validation of high throughput sequencing and microbial forensics applications %A Budowle, Bruce %A Connell, Nancy D %A Bielecka-Oder, Anna %A Rita R Colwell %A Corbett, Cindi R %A Fletcher, Jacqueline %A Forsman, Mats %A Kadavy, Dana R %A Markotic, Alemka %A Morse, Stephen A %A Murch, Randall S %A Sajantila, Antti %A Schmedes, Sarah E %A Ternus, Krista L %A Turner, Stephen D %A Minot, Samuel %X High throughput sequencing (HTS) generates large amounts of high quality sequence data for microbial genomics. The value of HTS for microbial forensics is the speed at which evidence can be collected and the power to characterize microbial-related evidence to solve biocrimes and bioterrorist events. As HTS technologies continue to improve, they provide increasingly powerful sets of tools to support the entire field of microbial forensics. Accurate, credible results allow analysis and interpretation, significantly influencing the course and/or focus of an investigation, and can impact the response of the government to an attack having individual, political, economic or military consequences. Interpretation of the results of microbial forensic analyses relies on understanding the performance and limitations of HTS methods, including analytical processes, assays and data interpretation. The utility of HTS must be defined carefully within established operating conditions and tolerances. Validation is essential in the development and implementation of microbial forensics methods used for formulating investigative leads attribution. HTS strategies vary, requiring guiding principles for HTS system validation. Three initial aspects of HTS, irrespective of chemistry, instrumentation or software are: 1) sample preparation, 2) sequencing, and 3) data analysis. Criteria that should be considered for HTS validation for microbial forensics are presented here. Validation should be defined in terms of specific application and the criteria described here comprise a foundation for investigators to establish, validate and implement HTS as a tool in microbial forensics, enhancing public safety and national security. %B Investigative Genetics %V 5 %P 9 %8 Jan-01-2014 %G eng %U http://investigativegenetics.biomedcentral.com/articles/10.1186/2041-2223-5-9 %N 1 %! Invest GenetInvestigative Genetics %R 10.1186/2041-2223-5-9 %0 Journal Article %J Microbial Ecology %D 2014 %T Viewing Marine Bacteria, Their Activity and Response to Environmental Drivers from Orbit %A Grimes, D. Jay %A Ford, Tim E. %A Rita R Colwell %A Baker-Austin, Craig %A Martinez-Urtaza, Jaime %A Subramaniam, Ajit %A Capone, Douglas G. %X Satellite-based remote sensing of marine microorganisms has become a useful tool in predicting human health risks associated with these microscopic targets. Early applications were focused on harmful algal blooms, but more recently methods have been developed to interrogate the ocean for bacteria. As satellite-based sensors have become more sophisticated and our ability to interpret information derived from these sensors has advanced, we have progressed from merely making fascinating pictures from space to developing process models with predictive capability. Our understanding of the role of marine microorganisms in primary production and global elemental cycles has been vastly improved as has our ability to use the combination of remote sensing data and models to provide early warning systems for disease outbreaks. This manuscript will discuss current approaches to monitoring cyanobacteria and vibrios, their activity and response to environmental drivers, and will also suggest future directions. %B Microbial Ecology %P 489 - 500 %8 Jan-04-2014 %G eng %U http://link.springer.com/10.1007/s00248-013-0363-4 %N 38 %! Microb Ecol %R 10.1007/s00248-013-0363-4 %0 Book Section %B Advances in Cryptology - ASIACRYPT 2013 %D 2013 %T Adaptive and Concurrent Secure Computation from New Adaptive, Non-malleable Commitments %A Dana Dachman-Soled %A Malkin, Tal %A Raykova, Mariana %A Venkitasubramaniam, Muthuramakrishnan %E Sako, Kazue %E Sarkar, Palash %K Algorithm Analysis and Problem Complexity %K Applications of Mathematics %K Data Encryption %K Discrete Mathematics in Computer Science %K Management of Computing and Information Systems %K Systems and Data Security %X We present a unified approach for obtaining general secure computation that achieves adaptive-Universally Composable (UC)-security. Using our approach we essentially obtain all previous results on adaptive concurrent secure computation, both in relaxed models (e.g., quasi-polynomial time simulation), as well as trusted setup models (e.g., the CRS model, the imperfect CRS model). This provides conceptual simplicity and insight into what is required for adaptive and concurrent security, as well as yielding improvements to set-up assumptions and/or computational assumptions in known models. Additionally, we provide the first constructions of concurrent secure computation protocols that are adaptively secure in the timing model, and the non-uniform simulation model. As a corollary we also obtain the first adaptively secure multiparty computation protocol in the plain model that is secure under bounded-concurrency. Conceptually, our approach can be viewed as an adaptive analogue to the recent work of Lin, Pass and Venkitasubramaniam [STOC ‘09], who considered only non-adaptive adversaries. Their main insight was that the non-malleability requirement could be decoupled from the simulation requirement to achieve UC-security. A main conceptual contribution of this work is, quite surprisingly, that it is still the case even when considering adaptive security. A key element in our construction is a commitment scheme that satisfies a strong definition of non-malleability. Our new primitive of concurrent equivocal non-malleable commitments, intuitively, guarantees that even when a man-in-the-middle adversary observes concurrent equivocal commitments and decommitments, the binding property of the commitments continues to hold for commitments made by the adversary. This definition is stronger than previous ones, and may be of independent interest. Previous constructions that satisfy our definition have been constructed in setup models, but either require existence of stronger encryption schemes such as CCA-secure encryption or require independent “trapdoors” provided by the setup for every pair of parties to ensure non-malleability. A main technical contribution of this work is to provide a construction that eliminates these requirements and requires only a single trapdoor. %B Advances in Cryptology - ASIACRYPT 2013 %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 316 - 336 %8 2013/01/01/ %@ 978-3-642-42032-0, 978-3-642-42033-7 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-42033-7_17 %0 Journal Article %J Applied and Environmental Microbiology %D 2013 %T Distribution of Virulence Genes in Clinical and Environmental Vibrio cholerae Strains in Bangladesh %A Hasan, Nur A. %A Ceccarelli, Daniela %A Grim, Christopher J. %A Taviani, Elisa %A Choi, Jinna %A Sadique, Abdus %A Alam, Munirul %A Siddique, Abul K. %A Sack, R. Bradley %A Huq, Anwar %A Rita R Colwell %X Vibrio cholerae, an environmental organism, is a facultative human pathogen. Here, we report the virulence profiles, comprising 18 genetic markers, of 102 clinical and 692 environmental V. cholerae strains isolated in Bangladesh between March 2004 and January 2006, showing the variability of virulence determinants within the context of public health. %B Applied and Environmental Microbiology %P 5782 - 5785 %8 Mar-09-2014 %G eng %U http://aem.asm.org/lookup/doi/10.1128/AEM.01113-13 %N 18 %! Appl. Environ. Microbiol. %R 10.1128/AEM.01113-13 %0 Journal Article %J CBE—Life Sciences Education %D 2013 %T Engaging Actively with Issues in the Responsible Conduct of Science: Lessons from International Efforts Are Relevant for Undergraduate Education in the United States %A Clements, John D. %A Connell, Nancy D. %A Dirks, Clarissa %A El-Faham, Mohamed %A Hay, Alastair %A Heitman, Elizabeth %A Stith, James H. %A Bond, Enriqueta C. %A Rita R Colwell %A Anestidou, Lida %A Husbands, Jo L. %A Labov, Jay B. %X Numerous studies are demonstrating that engaging undergraduate students in original research can improve their achievement in the science, technology, engineering, and mathematics (STEM) fields and increase the likelihood that some of them will decide to pursue careers in these disciplines. Associated with this increased prominence of research in the undergraduate curriculum are greater expectations from funders, colleges, and universities that faculty mentors will help those students, along with their graduate students and postdoctoral fellows, develop an understanding and sense of personal and collective obligation for responsible conduct of science (RCS). This Feature describes an ongoing National Research Council (NRC) project and a recent report about educating faculty members in culturally diverse settings (Middle East/North Africa and Asia) to employ active-learning strategies to engage their students and colleagues deeply in issues related to RCS. The NRC report describes the first phase of this project, which took place in Aqaba and Amman, Jordan, in September 2012 and April 2013, respectively. Here we highlight the findings from that report and our subsequent experience with a similar interactive institute in Kuala Lumpur, Malaysia. Our work provides insights and perspectives for faculty members in the United States as they engage undergraduate and graduate students, as well as postdoctoral fellows, to help them better understand the intricacies of and connections among various components of RCS. Further, our experiences can provide insights for those who may wish to establish “train-the-trainer” programs at their home institutions. %B CBE—Life Sciences Education %P 596 - 603 %8 Jan-12-2013 %G eng %U https://www.lifescied.org/doi/10.1187/cbe.13-09-0184 %N 4 %! LSE %R 10.1187/cbe.13-09-0184 %0 Journal Article %J The American Journal of Tropical Medicine and Hygiene %D 2013 %T Environmental Factors Influencing Epidemic Cholera %A Huq, Anwar %A Hasan, Nur %A Akanda, Ali %A Whitcombe, Elizabeth %A Rita R Colwell %A Haley, Bradd %A Alam, Munir %A Jutla, Antarpreet %A Sack, R. Bradley %X Cholera outbreak following the earthquake of 2010 in Haiti has reaffirmed that the disease is a major public health threat. Vibrio cholerae is autochthonous to aquatic environment, hence, it cannot be eradicated but hydroclimatology-based prediction and prevention is an achievable goal. Using data from the 1800s, we describe uniqueness in seasonality and mechanism of occurrence of cholera in the epidemic regions of Asia and Latin America. Epidemic regions are located near regional rivers and are characterized by sporadic outbreaks, which are likely to be initiated during episodes of prevailing warm air temperature with low river flows, creating favorable environmental conditions for growth of cholera bacteria. Heavy rainfall, through inundation or breakdown of sanitary infrastructure, accelerates interaction between contaminated water and human activities, resulting in an epidemic. This causal mechanism is markedly different from endemic cholera where tidal intrusion of seawater carrying bacteria from estuary to inland regions, results in outbreaks. %B The American Journal of Tropical Medicine and Hygiene %V 89 %P 597 - 607 %8 Apr-09-2013 %G eng %U http://www.ajtmh.org/content/journals/10.4269/ajtmh.12-0721 %N 3 %R 10.4269/ajtmh.12-0721 %0 Journal Article %J HCIC 2013 %D 2013 %T Exploring Early Solutions for Automatically Identifying Inaccessible Sidewalks in the Physical World Using Google Street View %A Hara, K %A Le, V %A Sun, J %A Jacobs, D %A Jon Froehlich %X Abstract Poorly maintained sidewalks , missing curb ramps, and other obstacles pose considerable accessibility challenges. Although pedestrian-and bicycle-oriented maps and associated routing algorithms continue to improve, there has been a lack of work focusing ... %B HCIC 2013 %8 2013/00/01 %G eng %U http://www.cs.umd.edu/~jonf/publications/Hara_ExploringEarlySolutionsForAutomaticallyIdentifyingInaccessibleSidewalksInThePhysicalWorldUsingGoogleStreetView_HCIC2013.pdf %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2013 %T Fostering advances in interdisciplinary climate science %A Shaman, J. %A Solomon, S. %A Rita R Colwell %A Field, C. B. %B Proceedings of the National Academy of Sciences %P 3653 - 3656 %8 Feb-02-2015 %G eng %U http://www.pnas.org/cgi/doi/10.1073/pnas.1301104110 %N Supplement_114869243Supplement_1 %! Proceedings of the National Academy of Sciences %R 10.1073/pnas.1301104110 %0 Journal Article %J Aquatic Biosystems %D 2013 %T Identification of bacteria in enrichment cultures of sulfate reducers in the Cariaco Basin water column employing Denaturing Gradient Gel Electrophoresis of 16S ribosomal RNA gene fragments %A Bozo-Hurtado, Lorelei %A García-Amado, M %A Chistoserdov, Andrei %A Varela, Ramon %A Narvaez, J %A Rita R Colwell %A Suárez, Paula %X Background The Cariaco Basin is characterized by pronounced and predictable vertical layering of microbial communities dominated by reduced sulfur species at and below the redox transition zone. Marine water samples were collected in May, 2005 and 2006, at the sampling stations A (10°30′ N, 64°40′ W), B (10°40′ N, 64°45′ W) and D (10°43’N, 64°32’W) from different depths, including surface, redox interface, and anoxic zones. In order to enrich for sulfate reducing bacteria (SRB), water samples were inoculated into anaerobic media amended with lactate or acetate as carbon source. To analyze the composition of enrichment cultures, we performed DNA extraction, PCR-DGGE, and sequencing of selected bands. Results DGGE results indicate that many bacterial genera were present that are associated with the sulfur cycle, including Desulfovibrio spp., as well as heterotrophs belonging to Vibrio, Enterobacter, Shewanella, Fusobacterium, Marinifilum, Mariniliabilia, and Spirochaeta. These bacterial populations are related to sulfur coupling and carbon cycles in an environment of variable redox conditions and oxygen availability. Conclusions In our studies, we found an association of SRB-like Desulfovibrio with Vibrio species and other genera that have a previously defined relevant role in sulfur transformation and coupling of carbon and sulfur cycles in an environment where there are variable redox conditions and oxygen availability. This study provides new information about microbial species that were culturable on media for SRB at anaerobic conditions at several locations and water depths in the Cariaco Basin. %B Aquatic Biosystems %V 9 %P 17 %8 Jan-01-2013 %G eng %U http://aquaticbiosystems.biomedcentral.com/articles/10.1186/2046-9063-9-17 %N 1 %! Aquat BiosystAquatic Biosystems %R 10.1186/2046-9063-9-17 %0 Journal Article %J First AAAI Conference on Human Computation and Crowdsourcing %D 2013 %T An Initial Study of Automatic Curb Ramp Detection with Crowdsourced Verification Using Google Street View Images %A Hara, Kotaro %A Sun, Jin %A Chazan, Jonah %A Jacobs, David %A Jon Froehlich %X In our previous research, we examined whether minimally trained crowd workers could find, categorize, and assess sidewalk accessibility problems using Google Street View (GSV) images. This poster paper presents a first step towards combining automated methods ( e.g., machine vision-based curb ramp detectors) in concert with human computation to improve the overall scalability of our approach. %B First AAAI Conference on Human Computation and Crowdsourcing %8 2013/00/11 %G eng %U http://www.aaai.org/ocs/index.php/HCOMP/HCOMP13/paper/view/7507 %! First AAAI Conference on Human Computation and Crowdsourcing %0 Journal Article %J Research in Microbiology %D 2013 %T A new integrative conjugative element detected in Haitian isolates of Vibrio cholerae non-O1/non-O139 %A Ceccarelli, Daniela %A Spagnoletti, Matteo %A Hasan, Nur A. %A Lansing, Stephanie %A Huq, Anwar %A Rita R Colwell %X The presence of SXT/R391-related integrating conjugative elements (ICEs) in Vibrio cholerae O1 and non-O1/non-O139 isolated from clinical and environmental samples in Haiti in 2010 was studied. The main finding of this work was the identification of the novel ICEVchHai2 among closely related V. cholerae non-O1/non-O139 clinical strains. The mosaic structure of this element confirms the role of ICEs as efficient recombination systems whereby new genetic material can be acquired and exchanged, according V. cholerae strains new accessory functions. %B Research in Microbiology %V 164 %P 891 - 893 %8 Jan-11-2013 %G eng %U https://linkinghub.elsevier.com/retrieve/pii/S092325081300154X %N 9 %! Research in Microbiology %R 10.1016/j.resmic.2013.08.004 %0 Book Section %B Financial Cryptography and Data Security %D 2013 %T Parallel and Dynamic Searchable Symmetric Encryption %A Kamara, Seny %A Charalampos Papamanthou %E Sadeghi, Ahmad-Reza %K cloud storage %K Computer Appl. in Administrative Data Processing %K Data Encryption %K e-Commerce/e-business %K parallel search %K Searchable encryption %K Systems and Data Security %X Searchable symmetric encryption (SSE) enables a client to outsource a collection of encrypted documents in the cloud and retain the ability to perform keyword searches without revealing information about the contents of the documents and queries. Although efficient SSE constructions are known, previous solutions are highly sequential. This is mainly due to the fact that, currently, the only method for achieving sub-linear time search is the inverted index approach (Curtmola, Garay, Kamara and Ostrovsky, CCS ’06) which requires the search algorithm to access a sequence of memory locations, each of which is unpredictable and stored at the previous location in the sequence. Motivated by advances in multi-core architectures, we present a new method for constructing sub-linear SSE schemes. Our approach is highly parallelizable and dynamic. With roughly a logarithmic number of cores in place, searches for a keyword w in our scheme execute in o(r) parallel time, where r is the number of documents containing keyword w (with more cores, this bound can go down to O(logn), i.e., independent of the result size r). Such time complexity outperforms the optimal Θ(r) sequential search time—a similar bound holds for the updates. Our scheme also achieves the following important properties: (a) it enjoys a strong notion of security, namely security against adaptive chosen-keyword attacks; (b) compared to existing sub-linear dynamic SSE schemes (e.g., Kamara, Papamanthou, Roeder, CCS ’12), updates in our scheme do not leak any information, apart from information that can be inferred from previous search tokens; (c) it can be implemented efficiently in external memory (with logarithmic I/O overhead). Our technique is simple and uses a red-black tree data structure; its security is proven in the random oracle model. %B Financial Cryptography and Data Security %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 258 - 274 %8 2013/01/01/ %@ 978-3-642-39883-4, 978-3-642-39884-1 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-39884-1_22 %0 Journal Article %J Computational Science & Discovery %D 2013 %T Parallel geometric classification of stem cells by their three-dimensional morphology %A Juba,Derek %A Cardone, Antonio %A Ip, Cheuk Yiu %A Simon Jr, Carl G %A K Tison, Christopher %A Kumar, Girish %A Brady,Mary %A Varshney, Amitabh %B Computational Science & Discovery %V 6 %P 015007 %8 01/2013 %N 1 %! Comput. Sci. Disc. %R 10.1088/1749-4699/6/1/015007 %0 Conference Paper %B CCS '13 Proceedings of the 2013 ACM SIGSAC Conference on Computer & Communications Security %D 2013 %T Practical Dynamic Proofs of Retrievability %A Shi, Elaine %A Stefanov, Emil %A Charalampos Papamanthou %K dynamic proofs of retrievability %K erasure code %K por %X Proofs of Retrievability (PoR), proposed by Juels and Kaliski in 2007, enable a client to store n file blocks with a cloud server so that later the server can prove possession of all the data in a very efficient manner (i.e., with constant computation and bandwidth). Although many efficient PoR schemes for static data have been constructed, only two dynamic PoR schemes exist. The scheme by Stefanov et. al. (ACSAC 2012) uses a large of amount of client storage and has a large audit cost. The scheme by Cash (EUROCRYPT 2013) is mostly of theoretical interest, as it employs Oblivious RAM (ORAM) as a black box, leading to increased practical overhead (e.g., it requires about 300 times more bandwidth than our construction). We propose a dynamic PoR scheme with constant client storage whose bandwidth cost is comparable to a Merkle hash tree, thus being very practical. Our construction outperforms the constructions of Stefanov et. al. and Cash et. al., both in theory and in practice. Specifically, for n outsourced blocks of beta bits each, writing a block requires beta+O(lambdalog n) bandwidth and O(betalog n) server computation (lambda is the security parameter). Audits are also very efficient, requiring beta+O(lambda^2log n) bandwidth. We also show how to make our scheme publicly verifiable, providing the first dynamic PoR scheme with such a property. We finally provide a very efficient implementation of our scheme. %B CCS '13 Proceedings of the 2013 ACM SIGSAC Conference on Computer & Communications Security %S CCS '13 %I ACM %P 325 - 336 %8 2013/// %@ 978-1-4503-2477-9 %G eng %U http://doi.acm.org/10.1145/2508859.2516669 %0 Book Section %B Theory of Cryptography %D 2013 %T Signatures of Correct Computation %A Charalampos Papamanthou %A Shi, Elaine %A Tamassia, Roberto %E Sahai, Amit %K Algorithm Analysis and Problem Complexity %K Computation by Abstract Devices %K Data Encryption %K Systems and Data Security %X We introduce Signatures of Correct Computation (SCC), a new model for verifying dynamic computations in cloud settings. In the SCC model, a trusted source outsources a function f to an untrusted server, along with a public key for that function (to be used during verification). The server can then produce a succinct signature σ vouching for the correctness of the computation of f, i.e., that some result v is indeed the correct outcome of the function f evaluated on some point a. There are two crucial performance properties that we want to guarantee in an SCC construction: (1) verifying the signature should take asymptotically less time than evaluating the function f; and (2) the public key should be efficiently updated whenever the function changes. We construct SCC schemes (satisfying the above two properties) supporting expressive manipulations over multivariate polynomials, such as polynomial evaluation and differentiation. Our constructions are adaptively secure in the random oracle model and achieve optimal updates, i.e., the function’s public key can be updated in time proportional to the number of updated coefficients, without performing a linear-time computation (in the size of the polynomial). We also show that signatures of correct computation imply Publicly Verifiable Computation (PVC), a model recently introduced in several concurrent and independent works. Roughly speaking, in the SCC model, any client can verify the signature σ and be convinced of some computation result, whereas in the PVC model only the client that issued a query (or anyone who trusts this client) can verify that the server returned a valid signature (proof) for the answer to the query. Our techniques can be readily adapted to construct PVC schemes with adaptive security, efficient updates and without the random oracle model. %B Theory of Cryptography %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 222 - 242 %8 2013/01/01/ %@ 978-3-642-36593-5, 978-3-642-36594-2 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-36594-2_13 %0 Journal Article %J Journal of the Asiatic Society of Bangladesh, Science %D 2013 %T Species composition of limnetic zooplankton from the southern coastal areas (Mathbaria and Bakerganj) in Bangladesh %A Mozumder, Pronob K %A Nahar, Samsun %A Naser, M Niamul %A Alam, Munirul %A Huq, Anwar %A Sack, R Bradley %A Rita R Colwell %X We studied the coastal zooplankton community structure of six waterbodies of Mathbaria and Bakerganj from January 2008 to June 2009. In total 35 zooplankton species were identified under 26 genera under 20 families under 8 orders from Mathbaria. Among them 6 were protozoans, 24 were rotifers, 3 were copepods, 1 was cladocerans and 1 was ostracods. From Bakergonj a total of 42 zooplankton species were identified under 23 genera under 17 Families under 7 orders. Among them 3 were protozoans, 25 were rotifers, 7 were copepods, 6 were cladocerans and 1 was ostracods. Results showed that abundance of rotifera group from both area was higher in Mathbaria (64.86%) and Bakerganj (60.98%) than other groups while abundance of ostracoda of Bakerganj area was lowest. %B Journal of the Asiatic Society of Bangladesh, Science %V 38 %8 Apr-06-2014 %G eng %U http://banglajol.info/index.php/JASBS/article/view/15326 %N 1 %! J. Asiat. Soc. Bangladesh, Sci. %R 10.3329/jasbs.v38i1.15326 %0 Book Section %B Advances in Cryptology – EUROCRYPT 2013 %D 2013 %T Streaming Authenticated Data Structures %A Charalampos Papamanthou %A Shi, Elaine %A Tamassia, Roberto %A Yi, Ke %E Johansson, Thomas %E Nguyen, Phong Q. %K Algorithm Analysis and Problem Complexity %K Data Encryption %K Discrete Mathematics in Computer Science %K Systems and Data Security %X We consider the problem of streaming verifiable computation, where both a verifier and a prover observe a stream of n elements x 1,x 2,…,x n and the verifier can later delegate some computation over the stream to the prover. The prover must return the output of the computation, along with a cryptographic proof to be used for verifying the correctness of the output. Due to the nature of the streaming setting, the verifier can only keep small local state (e.g., logarithmic) which must be updatable in a streaming manner and with no interaction with the prover. Such constraints make the problem particularly challenging and rule out applying existing verifiable computation schemes. We propose streaming authenticated data structures, a model that enables efficient verification of data structure queries on a stream. Compared to previous work, we achieve an exponential improvement in the prover’s running time: While previous solutions have linear prover complexity (in the size of the stream), even for queries executing in sublinear time (e.g., set membership), we propose a scheme with O(logM logn)O(\log M\ log n) prover complexity, where n is the size of the stream and M is the size of the universe of elements. Our schemes support a series of expressive queries, such as (non-)membership, successor, range search and frequency queries, over an ordered universe and even in higher dimensions. The central idea of our construction is a new authentication tree, called generalized hash tree. We instantiate our generalized hash tree with a hash function based on lattices assumptions, showing that it enjoys suitable algebraic properties that traditional Merkle trees lack. We exploit such properties to achieve our results. %B Advances in Cryptology – EUROCRYPT 2013 %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 353 - 370 %8 2013/01/01/ %@ 978-3-642-38347-2, 978-3-642-38348-9 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-38348-9_22 %0 Journal Article %J Nucleic Acids Res %D 2013 %T TIGRFAMs and Genome Properties in 2013. %A Haft, Daniel H %A Jeremy D Selengut %A Richter, Roland A %A Harkins, Derek %A Basu, Malay K %A Beck, Erin %K Databases, Protein %K Genome, Archaeal %K Genome, Bacterial %K Genomics %K Internet %K Markov chains %K Molecular Sequence Annotation %K Proteins %K sequence alignment %X

TIGRFAMs, available online at http://www.jcvi.org/tigrfams is a database of protein family definitions. Each entry features a seed alignment of trusted representative sequences, a hidden Markov model (HMM) built from that alignment, cutoff scores that let automated annotation pipelines decide which proteins are members, and annotations for transfer onto member proteins. Most TIGRFAMs models are designated equivalog, meaning they assign a specific name to proteins conserved in function from a common ancestral sequence. Models describing more functionally heterogeneous families are designated subfamily or domain, and assign less specific but more widely applicable annotations. The Genome Properties database, available at http://www.jcvi.org/genome-properties, specifies how computed evidence, including TIGRFAMs HMM results, should be used to judge whether an enzymatic pathway, a protein complex or another type of molecular subsystem is encoded in a genome. TIGRFAMs and Genome Properties content are developed in concert because subsystems reconstruction for large numbers of genomes guides selection of seed alignment sequences and cutoff values during protein family construction. Both databases specialize heavily in bacterial and archaeal subsystems. At present, 4284 models appear in TIGRFAMs, while 628 systems are described by Genome Properties. Content derives both from subsystem discovery work and from biocuration of the scientific literature.

%B Nucleic Acids Res %V 41 %P D387-95 %8 2013 Jan %G eng %N Database issue %R 10.1093/nar/gks1234 %0 Book Section %B Theory of Cryptography %D 2013 %T Why “Fiat-Shamir for Proofs” Lacks a Proof %A Bitansky, Nir %A Dana Dachman-Soled %A Garg, Sanjam %A Jain, Abhishek %A Kalai, Yael Tauman %A López-Alt, Adriana %A Wichs, Daniel %E Sahai, Amit %K Algorithm Analysis and Problem Complexity %K Computation by Abstract Devices %K Data Encryption %K Systems and Data Security %X The Fiat-Shamir heuristic [CRYPTO ’86] is used to convert any 3-message public-coin proof or argument system into a non-interactive argument, by hashing the prover’s first message to select the verifier’s challenge. It is known that this heuristic is sound when the hash function is modeled as a random oracle. On the other hand, the surprising result of Goldwasser and Kalai [FOCS ’03] shows that there exists a computationally sound argument on which the Fiat-Shamir heuristic is never sound, when instantiated with any actual efficient hash function. This leaves us with the following interesting possibility: perhaps we can securely instantiates the Fiat-Shamir heuristic for all 3-message public-coin statistically sound proofs, even if we must fail for some computationally sound arguments. Indeed, this has been conjectured to be the case by Barak, Lindell and Vadhan [FOCS ’03], but we do not have any provably secure instantiation under any “standard assumption”. In this work, we give a broad black-box separation result showing that the security of the Fiat-Shamir heuristic for statistically sound proofs cannot be proved under virtually any standard assumption via a black-box reduction. More precisely: –If we want to have a “universal” instantiation of the Fiat-Shamir heuristic that works for all 3-message public-coin proofs, then we cannot prove its security via a black-box reduction from any assumption that has the format of a “cryptographic game”. –For many concrete proof systems, if we want to have a “specific” instantiation of the Fiat-Shamir heuristic for that proof system, then we cannot prove its security via a black box reduction from any “falsifiable assumption” that has the format of a cryptographic game with an efficient challenger. %B Theory of Cryptography %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 182 - 201 %8 2013/01/01/ %@ 978-3-642-36593-5, 978-3-642-36594-2 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-36594-2_11 %0 Journal Article %J The ISME Journal %D 2012 %T An additional step in the transmission of Yersinia pestis? %A Easterday, W Ryan %A Kausrud, Kyrre L %A Star, Bastiaan %A Heier, Lise %A Haley, Bradd J %A Ageyev, Vladimir %A Rita R Colwell %A Stenseth, Nils Chr %X Plague, caused by the bacterium Yersinia pestis, is a mammalian vector-borne disease, transmitted by fleas that serve as the vector between rodent hosts. For many pathogens, including Y. pestis, there are strong evolutionary pressures that lead to a reduction in ‘useless genes’, with only those retained that reflect function in the specific environment inhabited by the pathogen. Genetic traits critical for survival and transmission between two environments, the rodent and the flea, are conserved in epizootic/epidemic plague strains. However, there are genes that remain conserved for which no function in the flea–rodent cycle has yet been observed, indicating an additional environment may exist in the transmission cycle of plague. Here, we present evidence for highly conserved genes that suggests a role in the persistence of Y. pestis after death of its host. Furthermore, maintenance of these genes points to Y. pestis traversing a post-mortem path between, and possibly within, epizootic periods and offering insight into mechanisms that may allow Y. pestis an alternative route of transmission in the natural environment. %B The ISME Journal %P 231 - 236 %8 Jan-02-2012 %G eng %U http://www.nature.com/articles/ismej2011105 %N 2 %! ISME J %R 10.1038/ismej.2011.105 %0 Journal Article %J BMC bioinformatics %D 2012 %T AGORA: Assembly Guided by Optical Restriction Alignment %A Lin, H.C. %A Goldstein, S. %A Mendelowitz, L. %A Zhou, S. %A Wetzel, J. %A Schwartz, D.C. %A Pop, Mihai %X Genome assembly is difficult due to repeated sequences within the genome, which create ambiguities and cause the final assembly to be broken up into many separate sequences (contigs). Long range linking information, such as mate-pairs or mapping data, is necessary to help assembly software resolve repeats, thereby leading to a more complete reconstruction of genomes. Prior work has used optical maps for validating assemblies and scaffolding contigs, after an initial assembly has been produced. However, optical maps have not previously been used within the genome assembly process. Here, we use optical map information within the popular de Bruijn graph assembly paradigm to eliminate paths in the de Bruijn graph which are not consistent with the optical map and help determine the correct reconstruction of the genome.We developed a new algorithm called AGORA: Assembly Guided by Optical Restriction Alignment. AGORA is the first algorithm to use optical map information directly within the de Bruijn graph framework to help produce an accurate assembly of a genome that is consistent with the optical map information provided. Our simulations on bacterial genomes show that AGORA is effective at producing assemblies closely matching the reference sequences. Additionally, we show that noise in the optical map can have a strong impact on the final assembly quality for some complex genomes, and we also measure how various characteristics of the starting de Bruijn graph may impact the quality of the final assembly. Lastly, we show that a proper choice of restriction enzyme for the optical map may substantially improve the quality of the final assembly. Our work shows that optical maps can be used effectively to assemble genomes within the de Bruijn graph assembly framework. Our experiments also provide insights into the characteristics of the mapping data that most affect the performance of our algorithm, indicating the potential benefit of more accurate optical mapping technologies, such as nano-coding. %B BMC bioinformatics %V 13 %8 2012 %G eng %N 1 %0 Journal Article %J J Bacteriol %D 2012 %T Archaeosortases and exosortases are widely distributed systems linking membrane transit with posttranslational modification. %A Haft, Daniel H %A Payne, Samuel H %A Jeremy D Selengut %K Amino Acid Sequence %K Aminoacyltransferases %K Archaeal Proteins %K Bacterial Proteins %K Cell Membrane %K Cysteine Endopeptidases %K Gene Expression Regulation, Archaeal %K Gene Expression Regulation, Bacterial %K Gene Expression Regulation, Enzymologic %K Molecular Sequence Data %K Protein Processing, Post-Translational %X

Multiple new prokaryotic C-terminal protein-sorting signals were found that reprise the tripartite architecture shared by LPXTG and PEP-CTERM: motif, TM helix, basic cluster. Defining hidden Markov models were constructed for all. PGF-CTERM occurs in 29 archaeal species, some of which have more than 50 proteins that share the domain. PGF-CTERM proteins include the major cell surface protein in Halobacterium, a glycoprotein with a partially characterized diphytanylglyceryl phosphate linkage near its C terminus. Comparative genomics identifies a distant exosortase homolog, designated archaeosortase A (ArtA), as the likely protein-processing enzyme for PGF-CTERM. Proteomics suggests that the PGF-CTERM region is removed. Additional systems include VPXXXP-CTERM/archeaosortase B in two of the same archaea and PEF-CTERM/archaeosortase C in four others. Bacterial exosortases often fall into subfamilies that partner with very different cohorts of extracellular polymeric substance biosynthesis proteins; several species have multiple systems. Variant systems include the VPDSG-CTERM/exosortase C system unique to certain members of the phylum Verrucomicrobia, VPLPA-CTERM/exosortase D in several alpha- and deltaproteobacterial species, and a dedicated (single-target) VPEID-CTERM/exosortase E system in alphaproteobacteria. Exosortase-related families XrtF in the class Flavobacteria and XrtG in Gram-positive bacteria mark distinctive conserved gene neighborhoods. A picture emerges of an ancient and now well-differentiated superfamily of deeply membrane-embedded protein-processing enzymes. Their target proteins are destined to transit cellular membranes during their biosynthesis, during which most undergo additional posttranslational modifications such as glycosylation.

%B J Bacteriol %V 194 %P 36-48 %8 2012 Jan %G eng %N 1 %R 10.1128/JB.06026-11 %0 Journal Article %J Autonomous Robots %D 2012 %T Automated synthesis of action selection policies for unmanned vehicles operating in adverse environments %A Svec,Petr %A Gupta, Satyandra K. %K Computer science %X We address the problem of automated action selection policy synthesis for unmanned vehicles operating in adverse environments. We introduce a new evolutionary computation-based approach using which an initial version of the policy is automatically generated and then gradually refined by detecting and fixing its shortcomings. The synthesis technique consists of the automated extraction of the vehicle’s exception states and Genetic Programming (GP) for automated composition and optimization of corrective sequences of commands in the form of macro-actions to be applied locally. The focus is specifically on automated synthesis of a policy for Unmanned Surface Vehicle (USV) to efficiently block the advancement of an intruder boat toward a valuable target. This task requires the USV to utilize reactive planning complemented by short-term forward planning to generate specific maneuvers for blocking. The intruder is human-competitive and exhibits a deceptive behavior so that the USV cannot exploit regularity in its attacking behavior. We compared the performance of a hand-coded blocking policy to the performance of a policy that was automatically synthesized. Our results show that the performance of the automatically generated policy exceeds the performance of the hand-coded policy and thus demonstrates the feasibility of the proposed approach. %B Autonomous Robots %V 32 %P 149 - 164 %8 2012/// %@ 0929-5593 %G eng %U http://www.springerlink.com/content/664k84x080868141/abstract/ %N 2 %R 10.1007/s10514-011-9268-6 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2012 %T BclAF1 restriction factor is neutralized by proteasomal degradation and microRNA repression during human cytomegalovirus infection %A Lee, Song Hee %A Kalejta, Robert F. %A Kerry, Julie %A Semmes, Oliver John %A O’Connor, Christine M. %A Zia Khan %A Garcia, Benjamin A. %A Shenk, Thomas %A Murphy, Eain %K innate immunity %K intrinsic immunity %K miRNA %K Proteasome %K UL82 %X Cell proteins can restrict the replication of viruses. Here, we identify the cellular BclAF1 protein as a human cytomegalovirus restriction factor and describe two independent mechanisms the virus uses to decrease its steady-state levels. Immediately following infection, the viral pp71 and UL35 proteins, which are delivered to cells within virions, direct the proteasomal degradation of BclAF1. Although BclAF1 reaccumulates through the middle stages of infection, it is subsequently down-regulated at late times by miR-UL112-1, a virus-encoded microRNA. In the absence of BclAF1 neutralization, viral gene expression and replication are inhibited. These data identify two temporally and mechanistically distinct functions used by human cytomegalovirus to down-regulate a cellular antiviral protein. %B Proceedings of the National Academy of Sciences %V 109 %P 9575 - 9580 %8 2012/06/12/ %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/109/24/9575 %N 24 %! PNAS %0 Journal Article %J Systematic BiologySyst Biol %D 2012 %T BEAGLE: An Application Programming Interface and High-Performance Computing Library for Statistical Phylogenetics %A Ayres,Daniel L %A Darling,Aaron %A Zwickl,Derrick J %A Beerli,Peter %A Holder,Mark T %A Lewis,Paul O %A Huelsenbeck,John P %A Ronquist,Fredrik %A Swofford,David L %A Cummings, Michael P. %A Rambaut,Andrew %A Suchard,Marc A %K Bayesian phylogenetics %K gpu %K maximum likelihood %K parallel computing %X Phylogenetic inference is fundamental to our understanding of most aspects of the origin and evolution of life, and in recent years, there has been a concentration of interest in statistical approaches such as Bayesian inference and maximum likelihood estimation. Yet, for large data sets and realistic or interesting models of evolution, these approaches remain computationally demanding. High-throughput sequencing can yield data for thousands of taxa, but scaling to such problems using serial computing often necessitates the use of nonstatistical or approximate approaches. The recent emergence of graphics processing units (GPUs) provides an opportunity to leverage their excellent floating-point computational performance to accelerate statistical phylogenetic inference. A specialized library for phylogenetic calculation would allow existing software packages to make more effective use of available computer hardware, including GPUs. Adoption of a common library would also make it easier for other emerging computing architectures, such as field programmable gate arrays, to be used in the future. We present BEAGLE, an application programming interface (API) and library for high-performance statistical phylogenetic inference. The API provides a uniform interface for performing phylogenetic likelihood calculations on a variety of compute hardware platforms. The library includes a set of efficient implementations and can currently exploit hardware including GPUs using NVIDIA CUDA, central processing units (CPUs) with Streaming SIMD Extensions and related processor supplementary instruction sets, and multicore CPUs via OpenMP. To demonstrate the advantages of a common API, we have incorporated the library into several popular phylogenetic software packages. The BEAGLE library is free open source software licensed under the Lesser GPL and available from http://beagle-lib.googlecode.com. An example client program is available as public domain software. %B Systematic BiologySyst Biol %V 61 %P 170 - 173 %8 2012/01/01/ %@ 1063-5157, 1076-836X %G eng %U http://sysbio.oxfordjournals.org/content/61/1/170 %N 1 %R 10.1093/sysbio/syr100 %0 Journal Article %J Empirical Methods in Natural Language Processing %D 2012 %T Besting the Quiz Master: Crowdsourcing Incremental Classification Games %A Satinoff,Brianna %A Jordan Boyd-Graber %X Cost-sensitive classification, where the features used in machine learning tasks have a cost, has been explored as a means of balancing knowl- edge against the expense of obtaining new fea- tures. We introduce a setting where humans engage in classification with incrementally re- vealed features: the collegiate trivia circuit. By providing the community with a web-based system to practice, we collected tens of thou- sands of implicit word-by-word ratings of how useful features are for eliciting correct answers. Observing humans’ classification process, we improve the performance of a state-of-the art classifier. We also use the dataset to evaluate a system to compete in the incremental classifica- tion task through a reduction of reinforcement learning to classification. Our system learns when to answer a question, performing better than baselines and most human players. %B Empirical Methods in Natural Language Processing %8 2012/// %G eng %0 Journal Article %J PLoS Computational Biology %D 2012 %T Bioinformatics for the Human Microbiome Project %A Gevers, Dirk %A Pop, Mihai %A Schloss, Patrick D. %A Huttenhower, Curtis %B PLoS Computational Biology %V 8 %P e1002779 %8 11/2012 %N 11 %R 10.1371/journal.pcbi.1002779 %0 Journal Article %J Computer %D 2012 %T Cloud Data Protection for the Masses %A Song,D. %A Elaine Shi %A Fischer, I. %A Shankar,U. %K Cloud computing %K cloud data protection %K cloud platform architecture %K cloud users %K Maintenance %K rapid development %K security of data %K strong data protection %X Offering strong data protection to cloud users while enabling rich applications is a challenging task. Researchers explore a new cloud platform architecture called Data Protection as a Service, which dramatically reduces the per-application development effort required to offer data protection, while still allowing rapid development and maintenance. %B Computer %V 45 %P 39 - 45 %8 2012 %@ 0018-9162 %G eng %N 1 %0 Book Section %B Modeling, Learning, and Processing of Text Technological Data StructuresModeling, Learning, and Processing of Text Technological Data Structures %D 2012 %T Collecting Semantic Similarity Ratings to Connect Concepts in Assistive Communication Tools %A Nikolova,Sonya %A Jordan Boyd-Graber %A Fellbaum,Christiane %E Mehler,Alexander %E Kühnberger,Kai-Uwe %E Lobin,Henning %E Lüngen,Harald %E Storrer,Angelika %E Witt,Andreas %X To compensate for the common inability of people with lexical production impairments to access and express intended concepts, we make use of models of human semantic memory that build on the notion of semantic similarity and relatedness. Such models, constructed on evidence gained from psycholinguistic experiments, form the basis of a large lexical database, WordNet . We augment WordNet with many additional links among words and concepts that are semantically related. Making this densely connected semantic network available to people with anomic aphasia through assistive technologies should enable them to navigate among related words and concepts and retrieve the words that they intend to express. %B Modeling, Learning, and Processing of Text Technological Data StructuresModeling, Learning, and Processing of Text Technological Data Structures %S Studies in Computational Intelligence %I Springer Berlin / Heidelberg %V 370 %P 81 - 93 %8 2012/// %@ 978-3-642-22612-0 %G eng %U http://dx.doi.org/10.1007/978-3-642-22613-7_5 %0 Book %D 2012 %T Computational Analysis of Terrorist Groups: Lashkar-e-Taiba %A V.S. Subrahmanian %A Mannes, Aaron %A Sliva,Amy %A Shakarian, Jana %A Dickerson, John P. %X Computational Analysis of Terrorist Groups: Lashkar-e-Taiba provides an in-depth look at Web intelligence, and how advanced mathematics and modern computing technology can influence the insights we have on terrorist groups. This book primarily focuses on one famous terrorist group known as Lashkar-e-Taiba (or LeT), and how it operates. After 10 years of counter Al Qaeda operations, LeT is considered by many in the counter-terrorism community to be an even greater threat to the US and world peace than Al Qaeda. Computational Analysis of Terrorist Groups: Lashkar-e-Taiba is the first book that demonstrates how to use modern computational analysis techniques including methods for big data analysis. This book presents how to quantify both the environment in which LeT operate, and the actions it took over a 20-year period, and represent it as a relational database table. This table is then mined using sophisticated data mining algorithms in order to gain detailed, mathematical, computational and statistical insights into LeT and its operations. This book also provides a detailed history of Lashkar-e-Taiba based on extensive analysis conducted by using open source information and public statements. Each chapter includes a case study, as well as a slide describing the key results which are available on the authors web sites. Computational Analysis of Terrorist Groups: Lashkar-e-Taiba is designed for a professional market composed of government or military workers, researchers and computer scientists working in the web intelligence field. Advanced-level students in computer science will also find this valuable as a reference book. %I Springer Publishing Company, Incorporated %8 2012 %@ 1461447682, 9781461447689 %G eng %0 Journal Article %J Microbiology and Immunology %D 2012 %T Conversion of viable but nonculturable enteric bacteria to culturable by co‐culture with eukaryotic cells %A Senoh,Mitsutoshi %A Ghosh‐Banerjee,Jayeeta %A Ramamurthy,Thandavarayan %A Rita R Colwell %A Miyoshi,Shin-ichi %A Nair,G. Balakrish %A Takeda,Yoshifumi %K conversion to culturability %K enteric bacteria %K viable but nonculturable %X Viable but nonculturable (VBNC) Vibrio cholerae non-O1/non-O139, V. parahaemolyticus, enterohemorrhagic Escherichia coli, enterotoxigenic E. coli, enteropathogenic E. coli, Shigella flexneri, and Salmonella enterica were converted to the culturable state by co-culture with selected eukaryotic cells, e.g., HT-29, Caco-2, T84, HeLa, Intestine 407, and CHO cells. %B Microbiology and Immunology %V 56 %P 342 - 345 %8 2012/04/27/ %@ 1348-0421 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1348-0421.2012.00440.x/abstract?userIsAuthenticated=false&deniedAccessCustomisedMessage= %N 5 %R 10.1111/j.1348-0421.2012.00440.x %0 Journal Article %J PloS one %D 2012 %T Deep Sequencing of the Oral Microbiome Reveals Signatures of Periodontal Disease %A Liu,B. %A Faller, L.L. %A Klitgord, N. %A Mazumdar, V. %A Ghodsi,M. %A Sommer, D.D. %A Gibbons, T.R. %A Treangen, T.J. %A Chang, Y.C. %A Li,S. %X The oral microbiome, the complex ecosystem of microbes inhabiting the human mouth, harbors several thousands of bacterial types. The proliferation of pathogenic bacteria within the mouth gives rise to periodontitis, an inflammatory disease known to also constitute a risk factor for cardiovascular disease. While much is known about individual species associated with pathogenesis, the system-level mechanisms underlying the transition from health to disease are still poorly understood. Through the sequencing of the 16S rRNA gene and of whole community DNA we provide a glimpse at the global genetic, metabolic, and ecological changes associated with periodontitis in 15 subgingival plaque samples, four from each of two periodontitis patients, and the remaining samples from three healthy individuals. We also demonstrate the power of whole-metagenome sequencing approaches in characterizing the genomes of key players in the oral microbiome, including an unculturable TM7 organism. We reveal the disease microbiome to be enriched in virulence factors, and adapted to a parasitic lifestyle that takes advantage of the disrupted host homeostasis. Furthermore, diseased samples share a common structure that was not found in completely healthy samples, suggesting that the disease state may occupy a narrow region within the space of possible configurations of the oral microbiome. Our pilot study demonstrates the power of high-throughput sequencing as a tool for understanding the role of the oral microbiome in periodontal disease. Despite a modest level of sequencing (~2 lanes Illumina 76 bp PE) and high human DNA contamination (up to ~90%) we were able to partially reconstruct several oral microbes and to preliminarily characterize some systems-level differences between the healthy and diseased oral microbiomes. %B PloS one %V 7 %8 2012 %G eng %N 6 %0 Journal Article %J IEEE Transactions on Multimedia %D 2012 %T Design and Synthesis for Multimedia Systems Using the Targeted Dataflow Interchange Format %A Chung-Ching Shen %A Wu, Shenpei %A Sane, N. %A Wu, Hsiang-Huang %A Plishker,W. %A Bhattacharyya, Shuvra S. %K associated dataflow graph-code generation %K Computational modeling %K contextual information encapsulation %K cross-platform application design %K data flow graphs %K Data models %K Data structures %K Dataflow graphs %K design components %K design tools %K Digital signal processing %K electronic data interchange %K embedded signal processing %K Embedded software %K high-level abstract modeling %K high-performance embedded-processing architectures %K Image coding %K image registration application %K image representation %K low-level customizations %K low-level optimization %K low-level synthesis %K Multimedia communication %K multimedia systems %K multimedia systems development %K object-oriented data structures %K object-oriented methods %K parameterized schedule representation %K programming interfaces %K repetitive graph structures %K retargetable design %K Schedules %K scheduling %K software synthesis %K Streaming media %K targeted dataflow interchange format %K task-level dataflow analysis %X Development of multimedia systems that can be targeted to different platforms is challenging due to the need for rigorous integration between high-level abstract modeling, and low-level synthesis and optimization. In this paper, a new dataflow-based design tool called the targeted dataflow interchange format is introduced for retargetable design, analysis, and implementation of embedded software for multimedia systems. Our approach provides novel capabilities, based on principles of task-level dataflow analysis, for exploring and optimizing interactions across design components; object-oriented data structures for encapsulating contextual information for components; a novel model for representing parameterized schedules that are derived from repetitive graph structures; and automated code generation for programming interfaces and low-level customizations that are geared toward high-performance embedded-processing architectures. We demonstrate our design tool for cross-platform application design, parameterized schedule representation, and associated dataflow graph-code generation using a case study centered around an image registration application. %B IEEE Transactions on Multimedia %V 14 %P 630 - 640 %8 2012 %@ 1520-9210 %G eng %N 3 %0 Journal Article %J Current Protocols in Microbiology %D 2012 %T Detection, Isolation, and Identification of Vibrio cholerae from the Environment %A Huq, Anwar %A Haley, Bradd J. %A Taviani, Elisa %A Chen, Arlene %A Hasan, Nur A. %A Rita R Colwell %E Coico, Richard %E Kowalik, Tim %E Quarles, John %E Stevenson, Brian %E Taylor, Ron %X Recent molecular advances in microbiology have greatly improved the detection of bacterial pathogens in the environment. These improvements and a downward trend in the cost of molecular detection methods have contributed to increased frequency of detection of pathogenic microorganisms where traditional culture‐based detection methods have failed. Culture methods also have been greatly improved, and the confluence of the two suites of methods provides a powerful tool for detection, isolation, and characterization of pathogens. While molecular detection provides data on the presence and type of pathogens, culturing methods allow a researcher to preserve the organism of interest for “‐omics” studies, such as genomic, metabolomic, secretomic, and transcriptomic analysis, which are rapidly becoming more affordable. This has yielded a clearer understanding of the ecology and epidemiology of microorganisms that cause disease. In this unit, we present commonly accepted methods for isolation, detection, and characterization of V. cholerae, providing more extensive knowledge of the ecology and epidemiology of this organism. This unit has been fully revised and updated from the earlier version with the latest knowledge and additional information not previously included. Curr. Protoc. Microbiol. 26:6A.5.1‐6A.5.51. © 2012 by John Wiley & Sons, Inc. %B Current Protocols in Microbiology %I John Wiley & Sons, Inc. %C Hoboken, NJ, USA %V 26 %8 08/2012 %G eng %U http://doi.wiley.com/10.1002/9780471729259.mc06a05s26 %N 1 %& Unit 6A.5 %R 10.1002/9780471729259.mc06a05s26 %0 Book Section %B Protein NMR Techniques %D 2012 %T Determining Protein Dynamics from 15N Relaxation Data by Using DYNAMICS %A Fushman, David %E Shekhtman,Alexander %E Burz,David S. %E Walker,John M. %K Biomedical and Life Sciences %X Motions are essential for protein function, and knowledge of protein dynamics is a key to our understanding the mechanisms underlying protein folding and stability, ligand recognition, allostery, and catalysis. In the last two decades, NMR relaxation measurements have become a powerful tool for characterizing backbone and side chain dynamics in complex biological macromolecules such as proteins and nucleic acids. Accurate analysis of the experimental data in terms of motional parameters is an essential prerequisite for developing physical models of motions to paint an adequate picture of protein dynamics. Here, I describe in detail how to use the software package DYNAMICS that was developed for accurate characterization of the overall tumbling and local dynamics in a protein from nuclear spin-relaxation rates measured by NMR. Step-by-step instructions are provided and illustrated through an analysis of 15 N relaxation data for protein G. %B Protein NMR Techniques %S Methods in Molecular Biology %I Humana Press %V 831 %P 485 - 511 %8 2012/// %@ 978-1-61779-480-3 %G eng %U http://www.springerlink.com/content/q667p0k138058103/abstract/ %0 Conference Paper %B 1st ACM SIGSPATIAL International Workshop on Mobile Geographic Information Systems (MobiGIS 2012) %D 2012 %T Duking it out at the smartphone mobile app mapping API corral: Apple, Google, and the competition %A Samet, Hanan %A Fruin, B.C %A Nutanong,S. %X The recent introduction of the Apple iPhone 5 and the accompanying iOS6 software environment which, among other changes, replaced the use of the Google Maps API in iOS5 by Apple’s own Maps API, has led to significant changes in the user experience with apps that make use of maps and has resulted in closer scrutiny of mapping applications on mobile devices. Many of these changes in the user experience deal with the quality of the data that is being produced and presented to the user, and has led to a wide ranging discussion of data quality and the seeming lack of quality assurance policies and protocols by Apple. These are widely documented in web postings. However, equally important are significant changes in the manner in which the data is presented to the user, but, surprisingly, not much attention has been paid to this aspect of the user experience which is somewhat analogous to the concept of the “last mile” when discussing the bandwidth of communications networks and its associated costs. The changes in the presentation and in the amount of data that are presented to the user on the Apple mapping platform, with an emphasis on mobile devices with a small form factor such as smartphones, are tabulated and compared along with other mapping platforms such as the iOS apps of ESRI, MapQuest, and OpenSeaMap (using the open source map data of OpenStreetMap), as well as Bing Maps and Nokia Maps for which no iOS app exists and thus the corresponding mobile web versions are used. %B 1st ACM SIGSPATIAL International Workshop on Mobile Geographic Information Systems (MobiGIS 2012) %C Redondo Beach, California %8 11/2012 %0 Journal Article %J Applied and Environmental Microbiology %D 2012 %T Ecology of Vibrio parahaemolyticus and Vibrio vulnificus in the Coastal and Estuarine Waters of Louisiana, Maryland, Mississippi, and Washington (United States) %A Johnson, Crystal N. %A Bowers, John C. %A Griffitt, Kimberly J. %A Molina, Vanessa %A Clostio, Rachel W. %A Pei, Shaofeng %A Laws, Edward %A Paranjpye, Rohinee N. %A Strom, Mark S. %A Chen, Arlene %A Hasan, Nur A. %A Huq, Anwar %A Noriea, Nicholas F. %A Grimes, D. Jay %A Rita R Colwell %X Vibrio parahaemolyticus and Vibrio vulnificus, which are native to estuaries globally, are agents of seafood-borne or wound infections, both potentially fatal. Like all vibrios autochthonous to coastal regions, their abundance varies with changes in environmental parameters. Sea surface temperature (SST), sea surface height (SSH), and chlorophyll have been shown to be predictors of zooplankton and thus factors linked to vibrio populations. The contribution of salinity, conductivity, turbidity, and dissolved organic carbon to the incidence and distribution of Vibrio spp. has also been reported. Here, a multicoastal, 21-month study was conducted to determine relationships between environmental parameters and V. parahaemolyticus and V. vulnificus populations in water, oysters, and sediment in three coastal areas of the United States. Because ecologically unique sites were included in the study, it was possible to analyze individual parameters over wide ranges. Molecular methods were used to detect genes for thermolabile hemolysin (tlh), thermostable direct hemolysin (tdh), and tdh-related hemolysin (trh) as indicators of V. parahaemolyticus and the hemolysin gene vvhA for V. vulnificus. SST and suspended particulate matter were found to be strong predictors of total and potentially pathogenic V. parahaemolyticus and V. vulnificus. Other predictors included chlorophyll a, salinity, and dissolved organic carbon. For the ecologically unique sites included in the study, SST was confirmed as an effective predictor of annual variation in vibrio abundance, with other parameters explaining a portion of the variation not attributable to SST. %B Applied and Environmental Microbiology %P 7249 - 7257 %8 Mar-10-2013 %G eng %U http://aem.asm.org/lookup/doi/10.1128/AEM.01296-12 %N 20 %! Appl. Environ. Microbiol. %R 10.1128/AEM.01296-12 %0 Journal Article %J Image Processing, IEEE Transactions on %D 2012 %T Face Identification Using Large Feature Sets %A Schwartz, W.R. %A Guo,Huimin %A Choi,Jonghyun %A Davis, Larry S. %K (mathematics); %K approximations;trees %K challenge %K data %K data;tree-based %K discriminative %K environments;face %K feature %K FERET;FRGC %K grand %K identification %K least %K recognition %K recognition;least %K sets;face %K sets;facial %K squares %K squares;training %K structure;uncontrolled %K task;face %K technology;multichannel %K weighting;partial %X With the goal of matching unknown faces against a gallery of known people, the face identification task has been studied for several decades. There are very accurate techniques to perform face identification in controlled environments, particularly when large numbers of samples are available for each face. However, face identification under uncontrolled environments or with a lack of training data is still an unsolved problem. We employ a large and rich set of feature descriptors (with more than 70 000 descriptors) for face identification using partial least squares to perform multichannel feature weighting. Then, we extend the method to a tree-based discriminative structure to reduce the time required to evaluate probe samples. The method is evaluated on Facial Recognition Technology (FERET) and Face Recognition Grand Challenge (FRGC) data sets. Experiments show that our identification method outperforms current state-of-the-art results, particularly for identifying faces acquired across varying conditions. %B Image Processing, IEEE Transactions on %V 21 %P 2245 - 2255 %8 2012/04// %@ 1057-7149 %G eng %N 4 %R 10.1109/TIP.2011.2176951 %0 Journal Article %J SIGIR Forum %D 2012 %T The first international workshop on entity-oriented search (EOS) %A Balog,Krisztian %A de Vries,Arjen P. %A Serdyukov,Pavel %A Wen,Ji-Rong %X The First International Workshop on Entity-Oriented Search (EOS) workshop was held on July 28, 2011 in Beijing, China, in conjunction with the 34th Annual International ACM SIGIR Conference (SIGIR 2011). The objective for the workshop was to bring together academic researchers and industry practitioners working on entity-oriented search to discuss tasks and challenges, and to uncover the next frontiers for academic research on the topic. The workshop program accommodated two invited talks, eleven refereed papers divided into three technical paper sessions, and a group discussion. %B SIGIR Forum %V 45 %P 43 - 50 %8 2012/01// %@ 0163-5840 %G eng %U http://doi.acm.org/10.1145/2093346.2093353 %N 2 %R 10.1145/2093346.2093353 %0 Journal Article %J Genome Research %D 2012 %T GAGE: A Critical Evaluation of Genome Assemblies and Assembly Algorithms %A Salzberg,Steven L. %A Phillippy,Adam M %A Zimin,Aleksey %A Puiu,Daniela %A Magoc,Tanja %A Koren,Sergey %A Treangen,Todd J %A Schatz,Michael C %A Delcher,Arthur L. %A Roberts,Michael %A Marçais,Guillaume %A Pop, Mihai %A Yorke,James A. %X New sequencing technology has dramatically altered the landscape of whole-genome sequencing, allowing scientists to initiate numerous projects to decode the genomes of previously unsequenced organisms. The lowest-cost technology can generate deep coverage of most species, including mammals, in just a few days. The sequence data generated by one of these projects consist of millions or billions of short DNA sequences (reads) that range from 50 to 150 nt in length. These sequences must then be assembled de novo before most genome analyses can begin. Unfortunately, genome assembly remains a very difficult problem, made more difficult by shorter reads and unreliable long-range linking information. In this study, we evaluated several of the leading de novo assembly algorithms on four different short-read data sets, all generated by Illumina sequencers. Our results describe the relative performance of the different assemblers as well as other significant differences in assembly difficulty that appear to be inherent in the genomes themselves. Three overarching conclusions are apparent: first, that data quality, rather than the assembler itself, has a dramatic effect on the quality of an assembled genome; second, that the degree of contiguity of an assembly varies enormously among different assemblers and different genomes; and third, that the correctness of an assembly also varies widely and is not well correlated with statistics on contiguity. To enable others to replicate our results, all of our data and methods are freely available, as are all assemblers used in this study. %B Genome Research %V 22 %P 557 - 567 %8 2012 %U http://genome.cshlp.org/content/22/3/557 %N 3 %R 10.1101/gr.131383.111 %0 Journal Article %J Nucleic Acids ResearchNucl. Acids Res. %D 2012 %T Gene Prediction with Glimmer for Metagenomic Sequences Augmented by Classification and Clustering %A Kelley,David R %A Liu,Bo %A Delcher,Arthur L. %A Pop, Mihai %A Salzberg,Steven L. %X Environmental shotgun sequencing (or metagenomics) is widely used to survey the communities of microbial organisms that live in many diverse ecosystems, such as the human body. Finding the protein-coding genes within the sequences is an important step for assessing the functional capacity of a metagenome. In this work, we developed a metagenomics gene prediction system Glimmer-MG that achieves significantly greater accuracy than previous systems via novel approaches to a number of important prediction subtasks. First, we introduce the use of phylogenetic classifications of the sequences to model parameterization. We also cluster the sequences, grouping together those that likely originated from the same organism. Analogous to iterative schemes that are useful for whole genomes, we retrain our models within each cluster on the initial gene predictions before making final predictions. Finally, we model both insertion/deletion and substitution sequencing errors using a different approach than previous software, allowing Glimmer-MG to change coding frame or pass through stop codons by predicting an error. In a comparison among multiple gene finding methods, Glimmer-MG makes the most sensitive and precise predictions on simulated and real metagenomes for all read lengths and error rates tested. %B Nucleic Acids ResearchNucl. Acids Res. %V 40 %P e9-e9 - e9-e9 %8 2012/01/01/ %@ 0305-1048, 1362-4962 %G eng %U http://nar.oxfordjournals.org/content/40/1/e9 %N 1 %R 10.1093/nar/gkr1067 %0 Journal Article %J Journal of Medical Microbiology %D 2012 %T Genetic characteristics of drug-resistant Vibrio cholerae O1 causing endemic cholera in Dhaka, 2006-2011 %A Rashed, S. M. %A Mannan, S. B. %A Johura, F.-T. %A Islam, M. T. %A Sadique, A. %A Watanabe, H. %A Sack, R. B. %A Huq, A. %A Rita R Colwell %A Cravioto, A. %A Alam, M. %X Vibrio cholerae O1 biotype El Tor (ET), causing the seventh cholera pandemic, was recently replaced in Bangladesh by an altered ET possessing ctxB of the Classical (CL) biotype, which caused the first six cholera pandemics. In the present study, V. cholerae O1 strains associated with endemic cholera in Dhaka between 2006 and 2011 were analysed for major phenotypic and genetic characteristics. Of 54 representative V. cholerae isolates tested, all were phenotypically ET and showed uniform resistance to trimethoprim/sulfamethoxazole (SXT) and furazolidone (FR). Resistance to tetracycline (TE) and erythromycin (E) showed temporal fluctuation, varying from year to year, while all isolates were susceptible to gentamicin (CN) and ciprofloxacin (CIP). Year-wise data revealed erythromycin resistance to be 33.3 % in 2006 and 11 % in 2011, while tetracycline resistance accounted for 33, 78, 0, 100 and 27 % in 2006, 2007, 2008, 2009 and 2010, respectively; interestingly, all isolates tested were sensitive to TE in 2011, as observed in 2008. All V. cholerae isolates tested possessed genetic elements such as SXT, ctxAB, tcpA ET, rstR ET and rtxC; none had IntlI (Integron I). Double mismatch amplification mutation assay (DMAMA)-PCR followed by DNA sequencing and analysis of the ctxB gene revealed a point mutation at position 58 (C→A), which has resulted in an amino acid substitution from histidine (H) to asparagine (N) at position 20 (genotype 7) since 2008. Although the multi-resistant strains having tetracycline resistance showed minor genetic divergence, V. choleraestrains were clonal, as determined by a PFGE (NotI)-based dendrogram. This study shows 2008–2010 to be the time of transition from ctxB genotype 1 to genotype 7 in V. cholerae ET causing endemic cholera in Dhaka, Bangladesh. %B Journal of Medical Microbiology %P 1736 - 1745 %8 Jan-12-2012 %G eng %U http://jmm.microbiologyresearch.org/content/journal/jmm/10.1099/jmm.0.049635-0 %! Journal of Medical Microbiology %R 10.1099/jmm.0.049635-0 %0 Journal Article %J FEBS Letters %D 2012 %T Genomic analysis of ICE Vch Ban8: An atypical genetic element in Vibrio cholerae %A Taviani, Elisa %A Spagnoletti, Matteo %A Ceccarelli, Daniela %A Haley, Bradd J. %A Hasan, Nur A. %A Chen, Arlene %A Colombo, Mauro M. %A Huq, Anwar %A Rita R Colwell %X Genomic islands (GIs) and integrative conjugative elements (ICEs) are major players in bacterial evolution since they encode genes involved in adaptive functions of medical or environmental importance. Here we performed the genomic analysis of ICEVchBan8, an unusual ICE found in the genome of a clinical non‐toxigenic Vibrio cholerae O37 isolate. ICEVchBan8 shares most of its genetic structure with SXT/R391 ICEs. However, this ICE codes for a different integration/excision module is located at a different insertion site, and part of its genetic cargo shows homology to other pathogenicity islands of V. cholerae. %B FEBS Letters %P 1617 - 1621 %8 Apr-06-2012 %G eng %U http://doi.wiley.com/10.1016/j.febslet.2012.03.064 %N 11 %R 10.1016/j.febslet.2012.03.064 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2012 %T Genomic diversity of 2010 Haitian cholera outbreak strains %A Hasan, N. A. %A Choi, S. Y. %A Eppinger, M. %A Clark, P. W. %A Chen, A. %A Alam, M. %A Haley, B. J. %A Taviani, E. %A Hine, E. %A Su, Q. %A Tallon, L. J. %A Prosper, J. B. %A Furth, K. %A Hoq, M. M. %A Li, H. %A Fraser-Liggett, C. M. %A Cravioto, A. %A Huq, A. %A Ravel, J. %A Cebula, T. A. %A Rita R Colwell %B Proceedings of the National Academy of Sciences %P E2010 - E2017 %8 05/2012 %G eng %U http://www.pnas.org/cgi/doi/10.1073/pnas.1207359109 %! Proceedings of the National Academy of Sciences %R 10.1073/pnas.1207359109 %0 Journal Article %J ISME J %D 2012 %T Genomic insights to SAR86, an abundant and uncultivated marine bacterial lineage. %A Dupont, Chris L %A Rusch, Douglas B %A Yooseph, Shibu %A Lombardo, Mary-Jane %A Richter, R Alexander %A Valas, Ruben %A Novotny, Mark %A Yee-Greenbaum, Joyclyn %A Jeremy D Selengut %A Haft, Dan H %A Halpern, Aaron L %A Lasken, Roger S %A Nealson, Kenneth %A Friedman, Robert %A Venter, J Craig %K Computational Biology %K Gammaproteobacteria %K Genome, Bacterial %K Genomic Library %K metagenomics %K Oceans and Seas %K Phylogeny %K plankton %K Rhodopsin %K RNA, Ribosomal, 16S %K Seawater %X

Bacteria in the 16S rRNA clade SAR86 are among the most abundant uncultivated constituents of microbial assemblages in the surface ocean for which little genomic information is currently available. Bioinformatic techniques were used to assemble two nearly complete genomes from marine metagenomes and single-cell sequencing provided two more partial genomes. Recruitment of metagenomic data shows that these SAR86 genomes substantially increase our knowledge of non-photosynthetic bacteria in the surface ocean. Phylogenomic analyses establish SAR86 as a basal and divergent lineage of γ-proteobacteria, and the individual genomes display a temperature-dependent distribution. Modestly sized at 1.25-1.7 Mbp, the SAR86 genomes lack several pathways for amino-acid and vitamin synthesis as well as sulfate reduction, trends commonly observed in other abundant marine microbes. SAR86 appears to be an aerobic chemoheterotroph with the potential for proteorhodopsin-based ATP generation, though the apparent lack of a retinal biosynthesis pathway may require it to scavenge exogenously-derived pigments to utilize proteorhodopsin. The genomes contain an expanded capacity for the degradation of lipids and carbohydrates acquired using a wealth of tonB-dependent outer membrane receptors. Like the abundant planktonic marine bacterial clade SAR11, SAR86 exhibits metabolic streamlining, but also a distinct carbon compound specialization, possibly avoiding competition.

%B ISME J %V 6 %P 1186-99 %8 2012 Jun %G eng %N 6 %R 10.1038/ismej.2011.189 %0 Journal Article %J North American Association of Computational Linguistics %D 2012 %T Grammatical structures for word-level sentiment detection %A Jordan Boyd-Graber %A Sayeed,Asad B. %A Rusk,Bryan %A Weinberg, Amy %X Existing work in fine-grained sentiment anal- ysis focuses on sentences and phrases but ig- nores the contribution of individual words and their grammatical connections. This is because of a lack of both (1) annotated data at the word level and (2) algorithms that can leverage syn- tactic information in a principled way. We ad- dress the first need by annotating articles from the information technology business press via crowdsourcing to provide training and testing data. To address the second need, we propose a suffix-tree data structure to represent syntac- tic relationships between opinion targets and words in a sentence that are opinion-bearing. We show that a factor graph derived from this data structure acquires these relationships with a small number of word-level features. We demonstrate that our supervised model per- forms better than baselines that ignore syntac- tic features and constraints. %B North American Association of Computational Linguistics %8 2012/// %G eng %0 Journal Article %J arXiv:1205.4928 %D 2012 %T Grey-box GUI Testing: Efficient Generation of Event Sequences %A Arlt,Stephan %A Banerjee,Ishan %A Bertolini,Cristiano %A Memon, Atif M. %A Schäf,Martin %K 68N30 %K Computer Science - Software Engineering %X Graphical user interfaces (GUIs), due to their event driven nature, present a potentially unbounded space of all possible ways to interact with software. During testing it becomes necessary to effectively sample this space. In this paper we develop algorithms that sample the GUI's input space by only generating sequences that (1) are allowed by the GUI's structure, and (2) chain together only those events that have data dependencies between their event handlers. We create a new abstraction, called an event-dependency graph (EDG) of the GUI, that captures data dependencies between event handler code. We develop a mapping between EDGs and an existing black-box user-level model of the GUI's workflow, called an event-flow graph (EFG). We have implemented automated EDG construction in a tool that analyzes the bytecode of each event handler. We evaluate our "grey-box" approach using four open-source applications and compare it with the current state-of-the-art EFG approach. Our results show that using the EDG reduces the number of test cases while still achieving at least the same coverage. Furthermore, we were able to detect 2 new bugs in the subject applications. %B arXiv:1205.4928 %8 2012/05/22/ %G eng %U http://arxiv.org/abs/1205.4928 %0 Conference Paper %D 2012 %T Gripper synthesis for indirect manipulation of cells using Holographic Optical Tweezers %A Chowdhury,Sagar %A Svec,Petr %A Wang,Chenlu %A Losert,Wolfgang %A Gupta, Satyandra K. %X Optical Tweezers (OT) are used for highly accurate manipulations of biological cells. However, the direct exposure of cells to focused laser beam may negatively influence their biological functions. In order to overcome this problem, we generate multiple optical traps to grab and move a 3D ensemble of inert particles such as silica microspheres to act as a reconfigurable gripper for a manipulated cell. The relative positions of the microspheres are important in order for the gripper to be robust against external environmental forces and the exposure of high intensity laser on the cell to be minimized. In this paper, we present results of different gripper configurations, experimentally tested using our OT setup, that provide robust gripping as well as minimize laser intensity experienced by the cell. We developed a computational approach that allowed us to perform preliminary modeling and synthesis of the gripper configurations. The gripper synthesis is cast as a multi-objective optimization problem. %P 2749 - 2754 %8 2012/05// %G eng %R 10.1109/ICRA.2012.6225153 %0 Journal Article %J Infection and immunity %D 2012 %T Identification of Coli Surface Antigen 23, a Novel Adhesin of Enterotoxigenic Escherichia coli %A Del Canto, F. %A Botkin, D.J. %A Valenzuela, P. %A Popov, V. %A Ruiz-Perez, F. %A Nataro, J.P. %A Levine, M.M. %A Stine, O.C. %A Pop, Mihai %A Torres, A.G. %A others %X Enterotoxigenic Escherichia coli (ETEC) is an important cause of diarrhea, mainly in developing countries. Although there are 25 different ETEC adhesins described in strains affecting humans, between 15% and 50% of the clinical isolates from different geographical regions are negative for these adhesins, suggesting that additional unidentified adhesion determinants might be present. Here, we report the discovery of Coli Surface Antigen 23 (CS23), a novel adhesin expressed by an ETEC serogroup O4 strain (ETEC 1766a), which was negative for the previously known ETEC adhesins, albeit it has the ability to adhere to Caco-2 cells. CS23 is encoded by an 8.8-kb locus which contains 9 open reading frames (ORFs), 7 of them sharing significant identity with genes required for assembly of K88-related fimbriae. This gene locus, named aal (adhesion-associated locus), is required for the adhesion ability of ETEC 1766a and was able to confer this adhesive phenotype to a nonadherent E. coli HB101 strain. The CS23 major structural subunit, AalE, shares limited identity with known pilin proteins, and it is more closely related to the CS13 pilin protein CshE, carried by human ETEC strains. Our data indicate that CS23 is a new member of the diverse adhesin repertoire used by ETEC strains. %B Infection and immunity %V 80 %P 2791 - 2801 %8 2012 %G eng %N 8 %0 Journal Article %J Pervasive and Mobile Computing %D 2012 %T Individuals among commuters: Building personalised transport information services from fare collection systems %A Lathia,N. %A Smith,C. %A Jon Froehlich %A Capra,L. %B Pervasive and Mobile Computing %8 2012 %G eng %0 Journal Article %J Queue - Micoprocessors %D 2012 %T Interactive Dynamics for Visual Analysis %A Heer,Jeffrey %A Shneiderman, Ben %X A taxonomy of tools that support the fluent and flexible use of visualizations %B Queue - Micoprocessors %V 10 %P 30:30–30:55 - 30:30–30:55 %8 2012/02// %@ 1542-7730 %G eng %U http://doi.acm.org/10.1145/2133416.2146416 %N 2 %R 10.1145/2133416.2146416 %0 Journal Article %J Nucleic Acids Res %D 2012 %T InterPro in 2011: new developments in the family and domain prediction database. %A Hunter, Sarah %A Jones, Philip %A Mitchell, Alex %A Apweiler, Rolf %A Attwood, Teresa K %A Bateman, Alex %A Bernard, Thomas %A Binns, David %A Bork, Peer %A Burge, Sarah %A de Castro, Edouard %A Coggill, Penny %A Corbett, Matthew %A Das, Ujjwal %A Daugherty, Louise %A Duquenne, Lauranne %A Finn, Robert D %A Fraser, Matthew %A Gough, Julian %A Haft, Daniel %A Hulo, Nicolas %A Kahn, Daniel %A Kelly, Elizabeth %A Letunic, Ivica %A Lonsdale, David %A Lopez, Rodrigo %A Madera, Martin %A Maslen, John %A McAnulla, Craig %A McDowall, Jennifer %A McMenamin, Conor %A Mi, Huaiyu %A Mutowo-Muellenet, Prudence %A Mulder, Nicola %A Natale, Darren %A Orengo, Christine %A Pesseat, Sebastien %A Punta, Marco %A Quinn, Antony F %A Rivoire, Catherine %A Sangrador-Vegas, Amaia %A Jeremy D Selengut %A Sigrist, Christian J A %A Scheremetjew, Maxim %A Tate, John %A Thimmajanarthanan, Manjulapramila %A Thomas, Paul D %A Wu, Cathy H %A Yeats, Corin %A Yong, Siew-Yit %K Databases, Protein %K Protein Structure, Tertiary %K Proteins %K Sequence Analysis, Protein %K software %K Terminology as Topic %K User-Computer Interface %X

InterPro (http://www.ebi.ac.uk/interpro/) is a database that integrates diverse information about protein families, domains and functional sites, and makes it freely available to the public via Web-based interfaces and services. Central to the database are diagnostic models, known as signatures, against which protein sequences can be searched to determine their potential function. InterPro has utility in the large-scale analysis of whole genomes and meta-genomes, as well as in characterizing individual protein sequences. Herein we give an overview of new developments in the database and its associated software since 2009, including updates to database content, curation processes and Web and programmatic interfaces.

%B Nucleic Acids Res %V 40 %P D306-12 %8 2012 Jan %G eng %N Database issue %R 10.1093/nar/gkr948 %0 Journal Article %J arXiv:1207.1813 [cs] %D 2012 %T Introspective Pushdown Analysis of Higher-Order Programs %A Earl, Christopher %A Sergey, Ilya %A Might, Matthew %A David Van Horn %K Computer Science - Programming Languages %K D.3.4 %K F.3.2 %X In the static analysis of functional programs, pushdown flow analysis and abstract garbage collection skirt just inside the boundaries of soundness and decidability. Alone, each method reduces analysis times and boosts precision by orders of magnitude. This work illuminates and conquers the theoretical challenges that stand in the way of combining the power of these techniques. The challenge in marrying these techniques is not subtle: computing the reachable control states of a pushdown system relies on limiting access during transition to the top of the stack; abstract garbage collection, on the other hand, needs full access to the entire stack to compute a root set, just as concrete collection does. \emph{Introspective} pushdown systems resolve this conflict. Introspective pushdown systems provide enough access to the stack to allow abstract garbage collection, but they remain restricted enough to compute control-state reachability, thereby enabling the sound and precise product of pushdown analysis and abstract garbage collection. Experiments reveal synergistic interplay between the techniques, and the fusion demonstrates "better-than-both-worlds" precision. %B arXiv:1207.1813 [cs] %8 2012/07/07/ %G eng %U http://arxiv.org/abs/1207.1813 %0 Journal Article %J Environmental Microbiology Reports %D 2012 %T Vibrio cholerae in a historically cholera-free country %A Haley, Bradd J. %A Chen, Arlene %A Grim, Christopher J. %A Clark, Philip %A Diaz, Celia M %A Taviani, Elisa %A Hasan, Nur A. %A Sancomb, Elizabeth %A Elnemr, Wessam M %A Islam, Muhammad A. %A Huq, Anwar %A Rita R Colwell %A Benediktsdóttir, Eva %X We report the autochthonous existence of Vibrio cholerae in coastal waters of Iceland, a geothermally active country where cholera is absent and has never been reported. Seawater, mussel and macroalgae samples were collected close to, and distant from, sites where geothermal activity causes a significant increase in water temperature during low tides. Vibrio cholerae was detected only at geothermal‐influenced sites during low‐tides. None of the V. cholerae isolates encoded cholera toxin (ctxAB) and all were non‐O1/non‐O139 serogroups. However, all isolates encoded other virulence factors that are associated with cholera as well as extra‐intestinal V. cholerae infections. The virulence factors were functional at temperatures of coastal waters of Iceland, suggesting an ecological role. It is noteworthy that V. cholerae was isolated from samples collected at sites distant from anthropogenic influence, supporting the conclusion that V. cholerae is autochthonous to the aquatic environment of Iceland. %B Environmental Microbiology Reports %P 381 - 389 %8 Jan-08-2012 %G eng %U http://doi.wiley.com/10.1111/j.1758-2229.2012.00332.x %N 4 %R 10.1111/j.1758-2229.2012.00332.x %0 Conference Paper %B Eight International Conference on Language Resources and Evaluation (LREC'12) %D 2012 %T Linguistic Resources for Handwriting Recognition and Translation Evaluation %A Song, Zhiyi %A Ismael, Safa %A Grimes, Stephen %A David Doermann %A Strassel,Stephanie %B Eight International Conference on Language Resources and Evaluation (LREC'12) %0 Journal Article %J Journal of Signal Processing Systems %D 2012 %T Mapping Parameterized Cyclo-static Dataflow Graphs onto Configurable Hardware %A Kee, Hojin %A Chung-Ching Shen %A Bhattacharyya, Shuvra S. %A Wong, Ian %A Yong Rao %A Kornerup, Jacob %K 4G communication systems %K Circuits and Systems %K Computer Imaging, Vision, Pattern Recognition and Graphics %K Dataflow modeling %K Electrical Engineering %K FPGA implementation %K Image Processing and Computer Vision %K Parameterized dataflow %K pattern recognition %K scheduling %K Signal, Image and Speech Processing %X In recent years, parameterized dataflow has evolved as a useful framework for modeling synchronous and cyclo-static graphs in which arbitrary parameters can be changed dynamically. Parameterized dataflow has proven to have significant expressive power for managing dynamics of DSP applications in important ways. However, efficient hardware synthesis techniques for parameterized dataflow representations are lacking. This paper addresses this void; specifically, the paper investigates efficient field programmable gate array (FPGA)-based implementation of parameterized cyclo-static dataflow (PCSDF) graphs. We develop a scheduling technique for throughput-constrained minimization of dataflow buffering requirements when mapping PCSDF representations of DSP applications onto FPGAs. The proposed scheduling technique is integrated with an existing formal schedule model, called the generalized schedule tree, to reduce schedule cost. To demonstrate our new, hardware-oriented PCSDF scheduling technique, we have designed a real-time base station emulator prototype based on a subset of long-term evolution (LTE), which is a key cellular standard. %B Journal of Signal Processing Systems %V 66 %P 285 - 301 %8 2012 %@ 1939-8018, 1939-8115 %G eng %U http://link.springer.com/article/10.1007/s11265-011-0599-5 %N 3 %! J Sign Process Syst %0 Journal Article %J Mobile Computing, IEEE Transactions on %D 2012 %T Mobile Data Offloading through Opportunistic Communications and Social Participation %A Han,Bo %A Hui,Pan %A Kumar,V. S.A %A Marathe,M. V %A Shao,Jianhua %A Srinivasan, Aravind %K 3G mobile communication %K 3G networks %K Bluetooth interface %K cellular networks %K cellular radio %K content transfer %K device/service discovery %K graph theory %K greedy algorithm %K heuristic algorithm %K information delivery %K information dissemination %K mobile data offloading %K mobile data traffic %K mobile phones %K mobile social networks %K MoSoNets %K Nokia N900 smartphones %K Opp-off %K opportunistic communications %K random algorithm %K real-world mobility traces %K smart phones %K social networking (online) %K social participation %K target-set selection problem %K telecommunication network routing %K Telecommunication traffic %K trace-driven simulation study %X 3G networks are currently overloaded, due to the increasing popularity of various applications for smartphones. Offloading mobile data traffic through opportunistic communications is a promising solution to partially solve this problem, because there is almost no monetary cost for it. We propose to exploit opportunistic communications to facilitate information dissemination in the emerging Mobile Social Networks (MoSoNets) and thus reduce the amount of mobile data traffic. As a case study, we investigate the target-set selection problem for information delivery. In particular, we study how to select the target set with only k users, such that we can minimize the mobile data traffic over cellular networks. We propose three algorithms, called Greedy, Heuristic, and Random, for this problem and evaluate their performance through an extensive trace-driven simulation study. Our simulation results verify the efficiency of these algorithms for both synthetic and real-world mobility traces. For example, the Heuristic algorithm can offload mobile data traffic by up to 73.66 percent for a real-world mobility trace. Moreover, to investigate the feasibility of opportunistic communications for mobile phones, we implement a proof-of-concept prototype, called Opp-off, on Nokia N900 smartphones, which utilizes their Bluetooth interface for device/service discovery and content transfer. %B Mobile Computing, IEEE Transactions on %V 11 %P 821 - 834 %8 2012/05// %@ 1536-1233 %G eng %N 5 %R 10.1109/TMC.2011.101 %0 Conference Paper %B 2012 Fourth International Conference on Communication Systems and Networks (COMSNETS) %D 2012 %T Networking lessons: From computers to water %A Narayanan,I. %A Sarangan,V. %A Vasan, A. %A Srinivasan, Aravind %A Sivasubramaniam,A. %K Biological cells %K computer networking %K Computer networks %K COMPUTERS %K energy footprint %K Genetic algorithms %K green nonIT domains %K infrastructure enhancement problem %K Internet %K network theory (graphs) %K networking lessons %K planning %K pricing %K water distribution networks %K water supply %K water utility network %X As an instance of using IT to green non-IT domains, we consider the question whether lessons from computer networking can be applied in water distribution networks to improve their energy footprint and/or efficiency. Our contributions in this work are: (i) we identify several areas where principles from computer networking can be used to better water distribution; (ii) we focus on a specific infrastructure enhancement problem caused by increasing demands on a water utility network and present solutions (similar to those used in computer networks) that optimize both operational expenditure and total cost of ownership. We validate our solutions through simulations and compare their efficacy against techniques that are traditionally used in enhancing water networks. Our results show that lessons from computer networks can help in enhancing water networks. %B 2012 Fourth International Conference on Communication Systems and Networks (COMSNETS) %I IEEE %P 1 - 6 %8 2012/01/03/7 %@ 978-1-4673-0296-8 %G eng %R 10.1109/COMSNETS.2012.6151373 %0 Journal Article %J University Journal of Zoology, Rajshahi University %D 2012 %T Occurrence of protozoans & their limnological relationships in some ponds of Mathbaria, Bangladesh %A Mozumder,P. K. %A Banu,M. A. %A Naser,M. N. %A Ali,M. S. %A Alam,M. %A Sack,R. B. %A Rita R Colwell %A Huq,A. %B University Journal of Zoology, Rajshahi University %V 29 %P 69 - 71 %8 2012/// %@ 1023-6104 %G eng %U http://journals.sfu.ca/bd/index.php/UJZRU %N 1 %0 Conference Paper %B Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium %D 2012 %T Optimizing epidemic protection for socially essential workers %A Barrett,Chris %A Beckman,Richard %A Bisset,Keith %A Chen,Jiangzhuo %A DuBois,Thomas %A Eubank,Stephen %A Kumar,V. S. Anil %A Lewis,Bryan %A Marathe,Madhav V. %A Srinivasan, Aravind %A Stretz,Paula E. %K epidemiology %K OPTIMIZATION %K public health informatics %X Public-health policy makers have many tools to mitigate an epidemic's effects. Most related research focuses on the direct effects on those infected (in terms of health, life, or productivity). Interventions including treatment, prophylaxis, quarantine, and social distancing are well studied in this context. These interventions do not address indirect effects due to the loss of critical services and infrastructures when too many of those responsible for their day-to-day operations fall ill. We examine, both analytically and through simulation, the protection of such essential subpopulations by sequestering them, effectively isolating them into groups during an epidemic. We develop a framework for studying the benefits of sequestering and heuristics for when to sequester. We also prove a key property of sequestering placement which helps partition the subpopulations optimally. Thus we provide a first step toward determining how to allocate resources between the direct protection of a population, and protection of those responsible for critical services. %B Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium %S IHI '12 %I ACM %C New York, NY, USA %P 31 - 40 %8 2012/// %@ 978-1-4503-0781-9 %G eng %U http://doi.acm.org/10.1145/2110363.2110371 %R 10.1145/2110363.2110371 %0 Journal Article %J Arxiv preprint arXiv:1202.5150 %D 2012 %T Path O-RAM: An Extremely Simple Oblivious RAM Protocol %A Stefanov, Emil %A Elaine Shi %K Computer Science - Cryptography and Security %X We present Path O-RAM, an extremely simple Oblivious RAM protocol. %B Arxiv preprint arXiv:1202.5150 %8 2012 %G eng %U http://arxiv.org/abs/1202.5150 %0 Journal Article %J arXiv:1208.6189 [cs] %D 2012 %T Preserving Link Privacy in Social Network Based Systems %A Mittal, Prateek %A Charalampos Papamanthou %A Song, Dawn %K Computer Science - Cryptography and Security %K Computer Science - Social and Information Networks %X A growing body of research leverages social network based trust relationships to improve the functionality of the system. However, these systems expose users' trust relationships, which is considered sensitive information in today's society, to an adversary. In this work, we make the following contributions. First, we propose an algorithm that perturbs the structure of a social graph in order to provide link privacy, at the cost of slight reduction in the utility of the social graph. Second we define general metrics for characterizing the utility and privacy of perturbed graphs. Third, we evaluate the utility and privacy of our proposed algorithm using real world social graphs. Finally, we demonstrate the applicability of our perturbation algorithm on a broad range of secure systems, including Sybil defenses and secure routing. %B arXiv:1208.6189 [cs] %8 2012/08/30/ %G eng %U http://arxiv.org/abs/1208.6189 %0 Journal Article %J Radio Science %D 2012 %T Prototyping scalable digital signal processing systems for radio astronomy using dataflow models %A Sane, N. %A Ford, J. %A Harris, A. I. %A Bhattacharyya, Shuvra S. %K dataflow models %K digital downconverter %K Digital signal processing %K model-based design %K radio astronomy %K rapid prototyping %X There is a growing trend toward using high-level tools for design and implementation of radio astronomy digital signal processing (DSP) systems. Such tools, for example, those from the Collaboration for Astronomy Signal Processing and Electronics Research (CASPER), are usually platform-specific, and lack high-level, platform-independent, portable, scalable application specifications. This limits the designer's ability to experiment with designs at a high-level of abstraction and early in the development cycle. We address some of these issues using a model-based design approach employing dataflow models. We demonstrate this approach by applying it to the design of a tunable digital downconverter (TDD) used for narrow-bandwidth spectroscopy. Our design is targeted toward an FPGA platform, called theInterconnect Break-out Board (IBOB), that is available from the CASPER. We use the term TDD to refer to a digital downconverter for which the decimation factor and center frequency can be reconfigured without the need for regenerating the hardware code. Such a design is currently not available in the CASPER DSP library. The work presented in this paper focuses on two aspects. First, we introduce and demonstrate a dataflow-based design approach using thedataflow interchange format (DIF) tool for high-level application specification, and we integrate this approach with the CASPER tool flow. Secondly, we explore the trade-off between the flexibility of TDD designs and the low hardware cost of fixed-configuration digital downconverter (FDD) designs that use the available CASPER DSP library. We further explore this trade-off in the context of a two-stage downconversion scheme employing a combination of TDD or FDD designs. %B Radio Science %V 47 %P n/a - n/a %8 2012 %@ 1944-799X %G eng %U http://onlinelibrary.wiley.com/doi/10.1029/2011RS004924/abstract %N 3 %0 Journal Article %J Molecular Systems Biology %D 2012 %T Quantitative measurement of allele‐specific protein expression in a diploid yeast hybrid by LC‐MS %A Zia Khan %A Bloom, Joshua S. %A Amini, Sasan %A Singh, Mona %A Perlman, David H. %A Caudy, Amy A. %A Kruglyak, Leonid %K allele specific %K divergence %K mass spectrometry %K protein expression %K proteomics %X Understanding the genetic basis of gene regulatory variation is a key goal of evolutionary and medical genetics. Regulatory variation can act in an allele‐specific manner (cis‐acting) or it can affect both alleles of a gene (trans‐acting). Differential allele‐specific expression (ASE), in which the expression of one allele differs from another in a diploid, implies the presence of cis‐acting regulatory variation. While microarrays and high‐throughput sequencing have enabled genome‐wide measurements of transcriptional ASE, methods for measurement of protein ASE (pASE) have lagged far behind. We describe a flexible, accurate, and scalable strategy for measurement of pASE by liquid chromatography‐coupled mass spectrometry (LC‐MS). We apply this approach to a hybrid between the yeast species Saccharomyces cerevisiae and Saccharomyces bayanus. Our results provide the first analysis of the relative contribution of cis‐acting and trans‐acting regulatory differences to protein expression divergence between yeast species.Synopsis A novel strategy for the quantitative measurement of allele‐specific protein expression is used to infer the contributions of cis‐ and trans‐acting factors influencing the divergence of protein levels between yeast species. Rigorous experimental controls and analyses confirm the accuracy of the new strategy for the quantitative measurement of allele‐specific protein expression by high‐throughput mass spectrometry.Analysis of allele‐specific protein expression in an interspecies yeast hybrid and protein expression differences between species reveals that both cis‐effects and trans‐effects contribute to protein expression divergence between two yeast species, Saccharomyces cerevisiae and Saccharomyces bayanus. %B Molecular Systems Biology %V 8 %8 2012/01/01/ %G eng %U http://msb.embopress.org/content/8/1/602 %N 1 %0 Journal Article %J Interacting with Computers %D 2012 %T Querying event sequences by exact match or similarity search: Design and empirical evaluation %A Wongsuphasawat,Krist %A Plaisant, Catherine %A Taieb-Maimon,Meirav %A Shneiderman, Ben %K Event sequence %K Similan %K similarity measure %K Similarity Search %K temporal categorical data %K Temporal query interface %X Specifying event sequence queries is challenging even for skilled computer professionals familiar with SQL. Most graphical user interfaces for database search use an exact match approach, which is often effective, but near misses may also be of interest. We describe a new similarity search interface, in which users specify a query by simply placing events on a blank timeline and retrieve a similarity-ranked list of results. Behind this user interface is a new similarity measure for event sequences which the users can customize by four decision criteria, enabling them to adjust the impact of missing, extra, or swapped events or the impact of time shifts. We describe a use case with Electronic Health Records based on our ongoing collaboration with hospital physicians. A controlled experiment with 18 participants compared exact match and similarity search interfaces. We report on the advantages and disadvantages of each interface and suggest a hybrid interface combining the best of both. %B Interacting with Computers %V 24 %P 55 - 68 %8 2012/03// %@ 0953-5438 %G eng %U http://www.sciencedirect.com/science/article/pii/S0953543812000124 %N 2 %R 10.1016/j.intcom.2012.01.003 %0 Journal Article %J Environmental Microbiology Reports %D 2012 %T Role of GbpA protein, an important virulence-related colonization factor, for Vibrio cholerae's survival in the aquatic environment %A Stauder, Monica %A Huq, Anwar %A Pezzati, Elisabetta %A Grim, Christopher J. %A Ramoino, Paola %A Pane, Luigi %A Rita R Colwell %A Pruzzo, Carla %A Vezzulli, Luigi %X Vibrio cholerae N‐acetyl glucosamine‐binding protein A (GbpA) is a chitin binding protein and a virulence factor involved in the colonization of human intestine. We investigated the distribution and genetic variations of gbpA in 488 V. cholerae strains of environmental and clinical origin, belonging to different serogroups and biotypes. We found that the gene is consistently present and highly conserved including an environmental V. cholerae‐related strain of ancestral origin. The gene was also consistently expressed in a number of representative V. cholerae strains cultured in laboratory aquatic microcosms under conditions simulating those found in temperate marine environments. Functional analysis carried out on V. cholerae O1 El Tor N16961 showed that GbpA is not involved in adhesion to inorganic surfaces but promotes interaction with environmental biotic substrates (plankton and bivalve hepatopancreas cells) representing known marine reservoir or host for the bacterium. It is suggested that the ability of GbpA to colonize human intestinal cells most probably originated from its primary function in the aquatic environment. %B Environmental Microbiology Reports %P 439 - 445 %8 Jan-08-2012 %G eng %U http://doi.wiley.com/10.1111/j.1758-2229.2012.00356.x %N 4 %R 10.1111/j.1758-2229.2012.00356.x %0 Journal Article %J Frontiers in Microbiology %D 2012 %T Role of Shrimp Chitin in the Ecology of Toxigenic Vibrio cholerae and Cholera Transmission %A Nahar,Shamsun %A Sultana,Marzia %A Naser,M. Niamul %A Nair,Gopinath B. %A Watanabe,Haruo %A Ohnishi,Makoto %A Yamamoto,Shouji %A Endtz,Hubert %A Cravioto,Alejandro %A Sack,R. Bradley %A Hasan,Nur A. %A Sadique,Abdus %A Huq,Anwar %A Rita R Colwell %A Alam,Munirul %X Seasonal plankton blooms correlate with occurrence of cholera in Bangladesh, although the mechanism of how dormant Vibrio cholerae, enduring interepidemic period in biofilms and plankton, initiates seasonal cholera is not fully understood. In this study, laboratory microcosms prepared with estuarine Mathbaria water (MW) samples supported active growth of toxigenic V. cholerae O1 up to 7 weeks as opposed to 6 months when microcosms were supplemented with dehydrated shrimp chitin chips (CC) as the single source of nutrient. Bacterial counting and detection of wbe and ctxA genes were done employing culture, direct fluorescent antibody (DFA) assay, and multiplex-polymerase chain reaction methods. In MW microcosm, the aqueous phase became clear as the non-culturable cells settled, whereas the aqueous phase of the MW–CC microcosm became turbid from bacterial growth stimulated by chitin. Bacterial chitin degradation and biofilm formation proceeded from an initial steady state to a gradually declining bacterial culturable count. V. cholerae within the microenvironments of chitin and chitin-associated biofilms remained metabolically active even in a high acidic environment without losing either viability or virulence. It is concluded that the abundance of chitin that occurs during blooms plays an important role in the aquatic life cycle of V. cholerae and, ultimately, in the seasonal transmission of cholera. %B Frontiers in Microbiology %V 2:260 %8 01/2012 %@ 1664-302X %G eng %R 10.3389/fmicb.2011.00260 %0 Book Section %B Advances in Cryptology – CRYPTO 2012 %D 2012 %T Securing Circuits against Constant-Rate Tampering %A Dana Dachman-Soled %A Kalai, Yael Tauman %E Safavi-Naini, Reihaneh %E Canetti, Ran %K circuit compiler %K Computer Communication Networks %K computers and society %K Data Encryption %K Discrete Mathematics in Computer Science %K Management of Computing and Information Systems %K PCP of proximity %K side-channel attacks %K Systems and Data Security %K tampering %X We present a compiler that converts any circuit into one that remains secure even if a constant fraction of its wires are tampered with. Following the seminal work of Ishai et. al. (Eurocrypt 2006), we consider adversaries who may choose an arbitrary set of wires to corrupt, and may set each such wire to 0 or to 1, or may toggle with the wire. We prove that such adversaries, who continuously tamper with the circuit, can learn at most logarithmically many bits of secret information (in addition to black-box access to the circuit). Our results are information theoretic. %B Advances in Cryptology – CRYPTO 2012 %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 533 - 551 %8 2012/01/01/ %@ 978-3-642-32008-8, 978-3-642-32009-5 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-32009-5_31 %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 2012 %T A Temporal Pattern Search Algorithm for Personal History Event Visualization %A Wang,T. D %A Deshpande, Amol %A Shneiderman, Ben %K automaton-based approach %K binary search %K bit-parallel approach %K data visualisation %K electronic health records %K event array %K finite automata %K interactive visualization program %K Lifelines2 visualization tool %K medical information systems %K NFA approach %K nondeterministic finite automaton %K O(m2n lg(n)) problem %K pattern matching %K personal history event visualization %K Shift-And approach %K temporal pattern search algorithm %K time stamp %X We present Temporal Pattern Search (TPS), a novel algorithm for searching for temporal patterns of events in historical personal histories. The traditional method of searching for such patterns uses an automaton-based approach over a single array of events, sorted by time stamps. Instead, TPS operates on a set of arrays, where each array contains all events of the same type, sorted by time stamps. TPS searches for a particular item in the pattern using a binary search over the appropriate arrays. Although binary search is considerably more expensive per item, it allows TPS to skip many unnecessary events in personal histories. We show that TPS's running time is bounded by O(m2n lg(n)), where m is the length of (number of events) a search pattern, and n is the number of events in a record (history). Although the asymptotic running time of TPS is inferior to that of a nondeterministic finite automaton (NFA) approach (O(mn)), TPS performs better than NFA under our experimental conditions. We also show TPS is very competitive with Shift-And, a bit-parallel approach, with real data. Since the experimental conditions we describe here subsume the conditions under which analysts would typically use TPS (i.e., within an interactive visualization program), we argue that TPS is an appropriate design choice for us. %B Knowledge and Data Engineering, IEEE Transactions on %V 24 %P 799 - 812 %8 2012/05// %@ 1041-4347 %G eng %N 5 %R 10.1109/TKDE.2010.257 %0 Conference Paper %B Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium %D 2012 %T Towards event sequence representation, reasoning and visualization for EHR data %A Tao,Cui %A Wongsuphasawat,Krist %A Clark,Kimberly %A Plaisant, Catherine %A Shneiderman, Ben %A Chute,Christopher G. %K ehr %K semantic web %K temporal relation reasoning %K time trend visualization %X Efficient analysis of event sequences and the ability to answer time-related, clinically important questions can accelerate clinical research in several areas such as causality assessments, decision support systems, and retrospective studies. The Clinical Narrative Temporal Reasoning Ontology (CNTRO)-based system is designed for semantically representing, annotating, and inferring temporal relations and constraints for clinical events in Electronic Health Records (EHR) represented in both structured and unstructured ways. The LifeFlow system is designed to support an interactive exploration of event sequences using visualization techniques. The combination of the two systems will provide a comprehensive environment for users to visualize inferred temporal relationships from EHR data. This paper discusses our preliminary efforts on connecting the two systems and the benefits we envision from such an environment. %B Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium %S IHI '12 %I ACM %C New York, NY, USA %P 801 - 806 %8 2012/// %@ 978-1-4503-0781-9 %G eng %U http://doi.acm.org/10.1145/2110363.2110461 %R 10.1145/2110363.2110461 %0 Conference Paper %B World Congress on Social Simulation (WCSS), Taipei, Taiwan %D 2012 %T When Does Simulated Data Match Real Data? Comparing Model Calibration Functions using Genetic Algorithms %A Stonedahl,Forrest %A Rand, William %X Agent-based models can be manipulated to replicate real- world datasets, but choosing the best set of parameters to achieve this result can be difficult. To validate a model, the real-world dataset is often divided into a training and test set. The training set is used to calibrate the parameters and the test set is used to determine if the calibrated model represents the real-world data. The difference between the real-world data and the simulated data is determined using an error measure. When using an evolutionary computation technique to choose the parameters, this error measure becomes the fitness function, and choosing the appropriate measure becomes even more crucial for a successful calibration process. We survey the effect of five different error measures in the context of a toy problem and a real world problem (simulating on-line news consumption). We use each error measure in turn to calibrate on the training dataset, and then examine the results of all five error measures on both the training and testing datasets. For the toy problem, one measure was the Pareto-dominant choice for calibration, but no error measure dominated all the others for the real-world problem. %B World Congress on Social Simulation (WCSS), Taipei, Taiwan %0 Journal Article %J PLoS Negl Trop Dis %D 2012 %T Whole genome analysis of Leptospira licerasiae provides insight into leptospiral evolution and pathogenicity. %A Ricaldi, Jessica N %A Fouts, Derrick E %A Jeremy D Selengut %A Harkins, Derek M %A Patra, Kailash P %A Moreno, Angelo %A Lehmann, Jason S %A Purushe, Janaki %A Sanka, Ravi %A Torres, Michael %A Webster, Nicholas J %A Vinetz, Joseph M %A Matthias, Michael A %K DNA, Bacterial %K Evolution, Molecular %K Gene Transfer, Horizontal %K Genome, Bacterial %K Genomic islands %K HUMANS %K Leptospira %K Molecular Sequence Data %K Multigene Family %K Prophages %K Sequence Analysis, DNA %K Virulence factors %X

The whole genome analysis of two strains of the first intermediately pathogenic leptospiral species to be sequenced (Leptospira licerasiae strains VAR010 and MMD0835) provides insight into their pathogenic potential and deepens our understanding of leptospiral evolution. Comparative analysis of eight leptospiral genomes shows the existence of a core leptospiral genome comprising 1547 genes and 452 conserved genes restricted to infectious species (including L. licerasiae) that are likely to be pathogenicity-related. Comparisons of the functional content of the genomes suggests that L. licerasiae retains several proteins related to nitrogen, amino acid and carbohydrate metabolism which might help to explain why these Leptospira grow well in artificial media compared with pathogenic species. L. licerasiae strains VAR010(T) and MMD0835 possess two prophage elements. While one element is circular and shares homology with LE1 of L. biflexa, the second is cryptic and homologous to a previously identified but unnamed region in L. interrogans serovars Copenhageni and Lai. We also report a unique O-antigen locus in L. licerasiae comprised of a 6-gene cluster that is unexpectedly short compared with L. interrogans in which analogous regions may include >90 such genes. Sequence homology searches suggest that these genes were acquired by lateral gene transfer (LGT). Furthermore, seven putative genomic islands ranging in size from 5 to 36 kb are present also suggestive of antecedent LGT. How Leptospira become naturally competent remains to be determined, but considering the phylogenetic origins of the genes comprising the O-antigen cluster and other putative laterally transferred genes, L. licerasiae must be able to exchange genetic material with non-invasive environmental bacteria. The data presented here demonstrate that L. licerasiae is genetically more closely related to pathogenic than to saprophytic Leptospira and provide insight into the genomic bases for its infectiousness and its unique antigenic characteristics.

%B PLoS Negl Trop Dis %V 6 %P e1853 %8 2012 %G eng %N 10 %R 10.1371/journal.pntd.0001853 %0 Conference Paper %B Technical Communications of the 27th International Conference on Logic Programming (ICLP'11) %D 2011 %T Abduction in Annotated Probabilistic Temporal Logic %A Molinaro,Cristian %A Sliva,Amy %A V.S. Subrahmanian %X Annotated Probabilistic Temporal (APT) logic programs are a form of logic programs that allow users to state (or systems to automatically learn)rules of the form ``formula G becomes true K time units after formula F became true with L to U% probability.'' In this paper, we develop a theory of abduction for APT logic programs. Specifically, given an APT logic program Pi, a set of formulas H that can be ``added'' to Pi, and a goal G, is there a subset S of H such that Pi \cup S is consistent and entails the goal G? In this paper, we study the complexity of the Basic APT Abduction Problem (BAAP). We then leverage a geometric characterization of BAAP to suggest a set of pruning strategies when solving BAAP and use these intuitions to develop a sound and complete algorithm. %B Technical Communications of the 27th International Conference on Logic Programming (ICLP'11) %S Leibniz International Proceedings in Informatics (LIPIcs) %I Schloss Dagstuhl--Leibniz-Zentrum fuer Informatik %V 11 %P 240 - 250 %8 2011/// %@ 978-3-939897-31-6 %G eng %R http://dx.doi.org/10.4230/LIPIcs.ICLP.2011.240 %0 Journal Article %J Genome Biology %D 2011 %T Accurate proteome-wide protein quantification from high-resolution 15N mass spectra %A Zia Khan %A Amini, Sasan %A Bloom, Joshua S. %A Ruse, Cristian %A Caudy, Amy A. %A Kruglyak, Leonid %A Singh, Mona %A Perlman, David H. %A Tavazoie, Saeed %X In quantitative mass spectrometry-based proteomics, the metabolic incorporation of a single source of 15N-labeled nitrogen has many advantages over using stable isotope-labeled amino acids. However, the lack of a robust computational framework for analyzing the resulting spectra has impeded wide use of this approach. We have addressed this challenge by introducing a new computational methodology for analyzing 15N spectra in which quantification is integrated with identification. Application of this method to an Escherichia coli growth transition reveals significant improvement in quantification accuracy over previous methods.PMID: 22182234 %B Genome Biology %V 12 %8 2011/12/19/ %@ 1465-6906 %G eng %U http://genomebiology.com/2011/12/12/R122/abstract %N 12 %0 Conference Paper %B Twenty-Second International Joint Conference on Artificial Intelligence %D 2011 %T Active Surveying: A Probabilistic Approach for Identifying Key Opinion Leaders %A Sharara,H. %A Getoor, Lise %A Norton,M. %X Opinion leaders play an important role in influenc- ing people’s beliefs, actions and behaviors. Al- though a number of methods have been proposed for identifying influentials using secondary sources of information, the use of primary sources, such as surveys, is still favored in many domains. In this work we present a new surveying method which combines secondary data with partial knowl- edge from primary sources to guide the information gathering process. We apply our proposed active surveying method to the problem of identifying key opinion leaders in the medical field, and show how we are able to accurately identify the opinion lead- ers while minimizing the amount of primary data required, which results in significant cost reduction in data acquisition without sacrificing its integrity. %B Twenty-Second International Joint Conference on Artificial Intelligence %8 2011/// %G eng %0 Conference Paper %B Proceedings of the 20th international conference companion on World wide web %D 2011 %T Adapting a map query interface for a gesturing touch screen interface %A Samet, Hanan %A Teitler,Benjamin E. %A Adelfio,Marco D. %A Lieberman,Michael D. %K map query interface %K newsstand %K touch screen gesturing interface %X NewsStand is an example application of a general framework that we are developing to enable searching for information using a map query interface, where the information results from monitoring the output of over 8,000 RSS news sources and is available for retrieval within minutes of publication. The user interface of NewsStand was recently adapted so that NewsStand can execute on mobile and tablet devices with a gesturing touch screen interface such as the iPhone, iPod Touch, and iPad. This action led to a discovery of some shortcomings of current mapping APIs as well as devising some interesting new widgets. These issues are discussed, and the realization can be seen by a demo at http://newsstand.umiacs.umd.edu on any of the above Apple devices as well as other devices that support gestures such as an Android phone. %B Proceedings of the 20th international conference companion on World wide web %S WWW '11 %I ACM %C New York, NY, USA %P 257 - 260 %8 2011/// %@ 978-1-4503-0637-9 %G eng %U http://doi.acm.org/10.1145/1963192.1963303 %R 10.1145/1963192.1963303 %0 Journal Article %J ACM Trans. Comput. Logic %D 2011 %T Annotated probabilistic temporal logic %A Shakarian,Paulo %A Parker,Austin %A Simari,Gerardo %A V.S. Subrahmanian %K frequency functions %K imprecise probabilities %K Probabilistic and temporal reasoning %K threads %X The semantics of most logics of time and probability is given via a probability distribution over threads, where a thread is a structure specifying what will be true at different points in time (in the future). When assessing the probabilities of statements such as “Event a will occur within 5 units of time of event b,” there are many different semantics possible, even when assessing the truth of this statement within a single thread. We introduce the syntax of annotated probabilistic temporal (APT) logic programs and axiomatically introduce the key notion of a frequency function (for the first time) to capture different types of intrathread reasoning, and then provide a semantics for intrathread and interthread reasoning in APT logic programs parameterized by such frequency functions. We develop a comprehensive set of complexity results for consistency checking and entailment in APT logic programs, together with sound and complete algorithms to check consistency and entailment. The basic algorithms use linear programming, but we then show how to substantially and correctly reduce the sizes of these linear programs to yield better computational properties. We describe a real world application we are developing using APT logic programs. %B ACM Trans. Comput. Logic %V 12 %P 14:1–14:44 - 14:1–14:44 %8 2011/01// %@ 1529-3785 %G eng %U http://doi.acm.org/10.1145/1877714.1877720 %N 2 %R 10.1145/1877714.1877720 %0 Conference Paper %B 2011 AAAI Fall Symposium Series %D 2011 %T Ant Colony Optimization in a Changing Environment %A Seymour,John Jefferson %A Tuzo,Joseph %A desJardins, Marie %X Ant colony optimization (ACO) algorithms are computational problem-solving methods that are inspired by the complex behaviors of ant colonies; specifically, the ways in which ants interact with each other and their environment to optimize the overall performance of the ant colony. Our eventual goal is to develop and experiment with ACO methods that can more effectively adapt to dynamically changing environments and problems. We describe biological ant systems and the dynamics of their environments and behaviors. We then introduce a family of dynamic ACO algorithms that can handle dynamic modifications of their inputs. We report empirical results, showing that dynamic ACO algorithms can effectively adapt to time-varying environments. %B 2011 AAAI Fall Symposium Series %8 2011/03/11/ %G eng %U http://www.aaai.org/ocs/index.php/FSS/FSS11/paper/viewPaper/4223 %0 Conference Paper %B 2011 Proceedings IEEE INFOCOM %D 2011 %T Approximation algorithms for throughput maximization in wireless networks with delay constraints %A Guanhong Pei %A Anil Kumar,V. S %A Parthasarathy,S. %A Srinivasan, Aravind %K Approximation algorithms %K Approximation methods %K approximation theory %K Delay %K delay constraints %K delays %K general interference model %K Interference %K multihop wireless networks %K optimisation %K Optimized production technology %K radio networks %K radiofrequency interference %K target delay bound %K Throughput %K throughput maximization %K Wireless networks %X We study the problem of throughput maximization in multi-hop wireless networks with end-to-end delay constraints for each session. This problem has received much attention starting with the work of Grossglauser and Tse (2002), and it has been shown that there is a significant tradeoff between the end-to-end delays and the total achievable rate. We develop algorithms to compute such tradeoffs with provable performance guarantees for arbitrary instances, with general interference models. Given a target delay-bound Δ(c) for each session c, our algorithm gives a stable flow vector with a total throughput within a factor of O (logΔm/log log Δm) of the maximum, so that the per-session (end-to-end) delay is O ((logΔm/log log Δm Δ(c))2), where Δm = maxc{Δ(c)}; note that these bounds depend only on the delays, and not on the network size, and this is the first such result, to our knowledge. %B 2011 Proceedings IEEE INFOCOM %I IEEE %P 1116 - 1124 %8 2011/04/10/15 %@ 978-1-4244-9919-9 %G eng %R 10.1109/INFCOM.2011.5934887 %0 Journal Article %J SIGCOMM Comput. Commun. Rev. %D 2011 %T Architecting for innovation %A Koponen,Teemu %A Shenker,Scott %A Balakrishnan,Hari %A Feamster, Nick %A Ganichev,Igor %A Ghodsi,Ali %A Godfrey,P. Brighten %A McKeown,Nick %A Parulkar,Guru %A Raghavan,Barath %A Rexford,Jennifer %A Arianfar,Somaya %A Kuptsov,Dmitriy %K diversity %K Evolution %K innovation %K internet architecture %X We argue that the biggest problem with the current Internet architecture is not a particular functional deficiency, but its inability to accommodate innovation. To address this problem we propose a minimal architectural "framework" in which comprehensive architectures can reside. The proposed Framework for Internet Innovation (FII) --- which is derived from the simple observation that network interfaces should be extensible and abstract --- allows for a diversity of architectures to coexist, communicate, and evolve. We demonstrate FII's ability to accommodate diversity and evolution with a detailed examination of how information flows through the architecture and with a skeleton implementation of the relevant interfaces. %B SIGCOMM Comput. Commun. Rev. %V 41 %P 24 - 36 %8 2011/// %@ 0146-4833 %G eng %U http://doi.acm.org/10.1145/2002250.2002256 %N 3 %R 10.1145/2002250.2002256 %0 Journal Article %J Journal of the ACM-Association for ComputingMachinery %D 2011 %T Article 28 (28 pages)-New Constructive Aspects of the Lovász Local Lemma %A Haeupler,B. %A Saha,B. %A Srinivasan, Aravind %B Journal of the ACM-Association for ComputingMachinery %V 58 %8 2011/// %G eng %N 6 %0 Book Section %B New Horizons in Evolutionary Robotics %D 2011 %T Automated Planning Logic Synthesis for Autonomous Unmanned Vehicles in Competitive Environments with Deceptive Adversaries %A Svec,Petr %A Gupta, Satyandra K. %E Doncieux,Stéphane %E Bredèche,Nicolas %E Mouret,Jean-Baptiste %K engineering %X We developed a new approach for automated synthesis of a planning logic for autonomous unmanned vehicles. This new approach can be viewed as an automated iterative process during which an initial version of a logic is synthesized and then gradually improved by detecting and fixing its shortcomings. This is achieved by combining data mining for extraction of vehicle’s states of failure and Genetic Programming (GP) technique for synthesis of corresponding navigation code. We verified the feasibility of the approach using unmanned surface vehicles (USVs) simulation. Our focus was specifically on the generation of a planning logic used for blocking the advancement of an intruder boat towards a valuable target. Developing autonomy logic for this behavior is challenging as the intruder’s attacking logic is human-competitive with deceptive behavior so the USV is required to learn specific maneuvers for specific situations to do successful blocking. We compared the performance of the generated blocking logic to the performance of logic that was manually implemented. Our results show that the new approach was able to synthesize a blocking logic with performance closely approaching the performance of the logic coded by hand. %B New Horizons in Evolutionary Robotics %S Studies in Computational Intelligence %I Springer Berlin / Heidelberg %V 341 %P 171 - 193 %8 2011/// %@ 978-3-642-18271-6 %G eng %U http://www.springerlink.com/content/f454477212518671/abstract/ %0 Conference Paper %B Advanced Video and Signal-Based Surveillance (AVSS), 2011 8th IEEE International Conference on %D 2011 %T AVSS 2011 demo session: A large-scale benchmark dataset for event recognition in surveillance video %A Oh,Sangmin %A Hoogs,Anthony %A Perera,Amitha %A Cuntoor,Naresh %A Chen,Chia-Chih %A Lee,Jong Taek %A Mukherjee,Saurajit %A Aggarwal, JK %A Lee,Hyungtae %A Davis, Larry S. %A Swears,Eran %A Wang,Xiaoyang %A Ji,Qiang %A Reddy,Kishore %A Shah,Mubarak %A Vondrick,Carl %A Pirsiavash,Hamed %A Ramanan,Deva %A Yuen,Jenny %A Torralba,Antonio %A Song,Bi %A Fong,Anesco %A Roy-Chowdhury,Amit %A Desai,Mita %X We introduce to the surveillance community the VIRAT Video Dataset[1], which is a new large-scale surveillance video dataset designed to assess the performance of event recognition algorithms in realistic scenes1. %B Advanced Video and Signal-Based Surveillance (AVSS), 2011 8th IEEE International Conference on %P 527 - 528 %8 2011/09/30/2 %G eng %R 10.1109/AVSS.2011.6027400 %0 Journal Article %J Proceedings of the National Academy of SciencesPNAS %D 2011 %T Bacillus Anthracis Comparative Genome Analysis in Support of the Amerithrax Investigation %A Rasko,David A %A Worsham,Patricia L %A Abshire,Terry G %A Stanley,Scott T %A Bannan,Jason D %A Wilson,Mark R %A Langham,Richard J %A Decker,R. Scott %A Jiang,Lingxia %A Read,Timothy D. %A Phillippy,Adam M %A Salzberg,Steven L. %A Pop, Mihai %A Van Ert,Matthew N %A Kenefic,Leo J %A Keim,Paul S %A Fraser-Liggett,Claire M %A Ravel,Jacques %X Before the anthrax letter attacks of 2001, the developing field of microbial forensics relied on microbial genotyping schemes based on a small portion of a genome sequence. Amerithrax, the investigation into the anthrax letter attacks, applied high-resolution whole-genome sequencing and comparative genomics to identify key genetic features of the letters’ Bacillus anthracis Ames strain. During systematic microbiological analysis of the spore material from the letters, we identified a number of morphological variants based on phenotypic characteristics and the ability to sporulate. The genomes of these morphological variants were sequenced and compared with that of the B. anthracis Ames ancestor, the progenitor of all B. anthracis Ames strains. Through comparative genomics, we identified four distinct loci with verifiable genetic mutations. Three of the four mutations could be directly linked to sporulation pathways in B. anthracis and more specifically to the regulation of the phosphorylation state of Spo0F, a key regulatory protein in the initiation of the sporulation cascade, thus linking phenotype to genotype. None of these variant genotypes were identified in single-colony environmental B. anthracis Ames isolates associated with the investigation. These genotypes were identified only in B. anthracis morphotypes isolated from the letters, indicating that the variants were not prevalent in the environment, not even the environments associated with the investigation. This study demonstrates the forensic value of systematic microbiological analysis combined with whole-genome sequencing and comparative genomics. %B Proceedings of the National Academy of SciencesPNAS %V 108 %P 5027 - 5032 %8 2011/03/22/ %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/108/12/5027 %N 12 %R 10.1073/pnas.1016657108 %0 Conference Paper %B 2011 IEEE International Conference on Computer Vision (ICCV) %D 2011 %T Blurring-invariant Riemannian metrics for comparing signals and images %A Zhengwu Zhang %A Klassen, E. %A Srivastava, A. %A Turaga,P. %A Chellapa, Rama %K blurring-invariant Riemannian metrics %K Estimation %K Fourier transforms %K Gaussian blur function %K Gaussian processes %K image representation %K log-Fourier representation %K measurement %K Orbits %K Polynomials %K signal representation %K Space vehicles %K vectors %X We propose a novel Riemannian framework for comparing signals and images in a manner that is invariant to their levels of blur. This framework uses a log-Fourier representation of signals/images in which the set of all possible Gaussian blurs of a signal, i.e. its orbits under semigroup action of Gaussian blur functions, is a straight line. Using a set of Riemannian metrics under which the group actions are by isometries, the orbits are compared via distances between orbits. We demonstrate this framework using a number of experimental results involving 1D signals and 2D images. %B 2011 IEEE International Conference on Computer Vision (ICCV) %I IEEE %P 1770 - 1775 %8 2011/11/06/13 %@ 978-1-4577-1101-5 %G eng %R 10.1109/ICCV.2011.6126442 %0 Journal Article %J SIGCOMM-Computer Communication Review %D 2011 %T Broadband internet performance: A view from the gateway %A Sundaresan,S. %A de Donato,W. %A Feamster, Nick %A Teixeira,R. %A Crawford,S. %A Pescapè,A. %X We present the first study of network access link performance mea-sured directly from home gateway devices. Policymakers, ISPs, and users are increasingly interested in studying the performance of Internet access links. Because of many confounding factors in a home network or on end hosts, however, thoroughly understanding access network performance requires deploying measurement in- frastructure in users’ homes as gateway devices. In conjunction with the Federal Communication Commission’s study of broad- band Internet access in the United States, we study the throughput and latency of network access links using longitudinal measure- ments from nearly 4,000 gateway devices across 8 ISPs from a de- ployment of over 4,200 devices. We study the performance users achieve and how various factors ranging from the user’s choice of modem to the ISP’s traffic shaping policies can affect performance. Our study yields many important findings about the characteristics of existing access networks. Our findings also provide insights into the ways that access network performance should be measured and presented to users, which can help inform ongoing broader efforts to benchmark the performance of access networks. %B SIGCOMM-Computer Communication Review %V 41 %P 134 - 134 %8 2011/// %G eng %N 4 %0 Conference Paper %B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on %D 2011 %T Bypassing synthesis: PLS for face recognition with pose, low-resolution and sketch %A Sharma,A. %A Jacobs, David W. %K approximations;pose %K COMPARISON %K estimation; %K extraction;image %K framework;face %K intensities;subspace %K least %K PLS;bypassing %K recognition;feature %K recognition;partial %K resolution;least %K selection;multimodal %K squares %K squares;pixel %K synthesis;face %X This paper presents a novel way to perform multi-modal face recognition. We use Partial Least Squares (PLS) to linearly map images in different modalities to a common linear subspace in which they are highly correlated. PLS has been previously used effectively for feature selection in face recognition. We show both theoretically and experimentally that PLS can be used effectively across modalities. We also formulate a generic intermediate subspace comparison framework for multi-modal recognition. Surprisingly, we achieve high performance using only pixel intensities as features. We experimentally demonstrate the highest published recognition rates on the pose variations in the PIE data set, and also show that PLS can be used to compare sketches to photos, and to compare images taken at different resolutions. %B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on %P 593 - 600 %8 2011/06// %G eng %R 10.1109/CVPR.2011.5995350 %0 Book Section %B Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques %D 2011 %T A Canonical Form for Testing Boolean Function Properties %A Dana Dachman-Soled %A Servedio, Rocco A. %E Goldberg, Leslie Ann %E Jansen, Klaus %E Ravi, R. %E Rolim, José D. P. %K Algorithm Analysis and Problem Complexity %K Boolean functions %K Computation by Abstract Devices %K Computer Communication Networks %K Computer Graphics %K Data structures %K Discrete Mathematics in Computer Science %K property testing %X In a well-known result Goldreich and Trevisan (2003) showed that every testable graph property has a “canonical” tester in which a set of vertices is selected at random and the edges queried are the complete graph over the selected vertices. We define a similar-in-spirit canonical form for Boolean function testing algorithms, and show that under some mild conditions property testers for Boolean functions can be transformed into this canonical form. Our first main result shows, roughly speaking, that every “nice” family of Boolean functions that has low noise sensitivity and is testable by an “independent tester,” has a canonical testing algorithm. Our second main result is similar but holds instead for families of Boolean functions that are closed under ID-negative minors. Taken together, these two results cover almost all of the constant-query Boolean function testing algorithms that we know of in the literature, and show that all of these testing algorithms can be automatically converted into a canonical form. %B Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 460 - 471 %8 2011/01/01/ %@ 978-3-642-22934-3, 978-3-642-22935-0 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-22935-0_39 %0 Journal Article %J Wireless Networks %D 2011 %T Capacity of wireless networks under SINR interference constraints %A Chafekar,Deepti %A Anil Kumar,V. %A Marathe,Madhav %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %X A fundamental problem in wireless networks is to estimate their throughput capacity—given a set of wireless nodes and a set of connections, what is the maximum rate at which data can be sent on these connections. Most of the research in this direction has focused either on random distributions of points, or has assumed simple graph-based models for wireless interference. In this paper, we study the capacity estimation problem using a realistic Signal to Interference Plus Noise Ratio (SINR) model for interference, on arbitrary wireless networks without any assumptions on node distributions. The problem becomes much more challenging for this setting, because of the non-locality of the SINR model. Recent work by Moscibroda et al. (IEEE INFOCOM 2006, ACM MobiHoc 2006) has shown that the throughput achieved by using SINR models can differ significantly from that obtained by using graph-based models. In this work, we develop polynomial time algorithms to provably approximate the throughput capacity of wireless network under the SINR model. %B Wireless Networks %V 17 %P 1605 - 1624 %8 2011/// %@ 1022-0038 %G eng %U http://dx.doi.org/10.1007/s11276-011-0367-2 %N 7 %0 Conference Paper %B Proceedings of the 20th international conference on World wide web %D 2011 %T A case for query by image and text content: searching computer help using screenshots and keywords %A Tom Yeh %A White,Brandyn %A San Pedro,Jose %A Katz,Boriz %A Davis, Larry S. %K content-based image retrieval %K multimodal search %K online help %X The multimedia information retrieval community has dedicated extensive research effort to the problem of content-based image retrieval (CBIR). However, these systems find their main limitation in the difficulty of creating pictorial queries. As a result, few systems offer the option of querying by visual examples, and rely on automatic concept detection and tagging techniques to provide support for searching visual content using textual queries. This paper proposes and studies a practical multimodal web search scenario, where CBIR fits intuitively to improve the retrieval of rich information queries. Many online articles contain useful know-how knowledge about computer applications. These articles tend to be richly illustrated by screenshots. We present a system to search for such software know-how articles that leverages the visual correspondences between screenshots. Users can naturally create pictorial queries simply by taking a screenshot of the application to retrieve a list of articles containing a matching screenshot. We build a prototype comprising 150k articles that are classified into walkthrough, book, gallery, and general categories, and provide a comprehensive evaluation of this system, focusing on technical (accuracy of CBIR techniques) and usability (perceived system usefulness) aspects. We also consider the study of added value features of such a visual-supported search, including the ability to perform cross-lingual queries. We find that the system is able to retrieve matching screenshots for a wide variety of programs, across language boundaries, and provide subjectively more useful results than keyword-based web and image search engines. %B Proceedings of the 20th international conference on World wide web %S WWW '11 %I ACM %C New York, NY, USA %P 775 - 784 %8 2011/// %@ 978-1-4503-0632-4 %G eng %U http://doi.acm.org/10.1145/1963405.1963513 %R 10.1145/1963405.1963513 %0 Conference Paper %B Proceedings of the 20th international conference on World wide web %D 2011 %T A case for query by image and text content: searching computer help using screenshots and keywords %A Tom Yeh %A White,Brandyn %A San Pedro,Jose %A Katz,Boriz %A Davis, Larry S. %K content-based image retrieval %K multimodal search %K online help %X The multimedia information retrieval community has dedicated extensive research effort to the problem of content-based image retrieval (CBIR). However, these systems find their main limitation in the difficulty of creating pictorial queries. As a result, few systems offer the option of querying by visual examples, and rely on automatic concept detection and tagging techniques to provide support for searching visual content using textual queries.This paper proposes and studies a practical multimodal web search scenario, where CBIR fits intuitively to improve the retrieval of rich information queries. Many online articles contain useful know-how knowledge about computer applications. These articles tend to be richly illustrated by screenshots. We present a system to search for such software know-how articles that leverages the visual correspondences between screenshots. Users can naturally create pictorial queries simply by taking a screenshot of the application to retrieve a list of articles containing a matching screenshot. We build a prototype comprising 150k articles that are classified into walkthrough, book, gallery, and general categories, and provide a comprehensive evaluation of this system, focusing on technical (accuracy of CBIR techniques) and usability (perceived system usefulness) aspects. We also consider the study of added value features of such a visual-supported search, including the ability to perform cross-lingual queries. We find that the system is able to retrieve matching screenshots for a wide variety of programs, across language boundaries, and provide subjectively more useful results than keyword-based web and image search engines. %B Proceedings of the 20th international conference on World wide web %S WWW '11 %I ACM %C New York, NY, USA %P 775 - 784 %8 2011/// %@ 978-1-4503-0632-4 %G eng %U http://doi.acm.org/10.1145/1963405.1963513 %R 10.1145/1963405.1963513 %0 Journal Article %J Journal of mechanisms and robotics %D 2011 %T A Case Study in Optimization of Gait and Physical Parameters for a Snake-Inspired Robot Based on a Rectilinear Gait %A Hopkins,James K. %A Spranklin,Brent W. %A Gupta, Satyandra K. %X Physical parameters of the constituent modules and gait parameters affect the overall performance of snake-inspired robots. Hence, a system-level optimization model needs to concurrently optimize the module parameters and the gait. Incorporating a physics-based model of rectilinear gaits in the system-level optimization model is a computationally challenging problem. This paper presents a case study to illustrate how metamodels of the precomputed optimal rectilinear gaits can be utilized to reduce the complexity of the system-level optimization model. An example is presented to illustrate the importance of concurrently optimizing the module parameters and the gait to obtain the optimal performance for a given mission. %B Journal of mechanisms and robotics %V 3 %8 2011/// %@ 1942-4302 %G eng %U http://cat.inist.fr/?aModele=afficheN&cpsidt=23978760 %N 1 %0 Journal Article %J Cytometry. Part A : the journal of the International Society for Analytical Cytology %D 2011 %T Cell cycle dependent TN-C promoter activity determined by live cell imaging. %A Halter, Michael %A Sisan, Daniel R %A Chalfoun, Joe %A Stottrup, Benjamin L %A Cardone, Antonio %A Dima,Alden A. %A Tona, Alessandro %A Plant,Anne L. %A Elliott, John T %K Animals %K cell cycle %K Gene Expression Regulation %K Green Fluorescent Proteins %K Image Processing, Computer-Assisted %K Mice %K Microscopy, Fluorescence %K Microscopy, Phase-Contrast %K NIH 3T3 Cells %K Promoter Regions, Genetic %K Tenascin %X The extracellular matrix protein tenascin-C plays a critical role in development, wound healing, and cancer progression, but how it is controlled and how it exerts its physiological responses remain unclear. By quantifying the behavior of live cells with phase contrast and fluorescence microscopy, the dynamic regulation of TN-C promoter activity is examined. We employ an NIH 3T3 cell line stably transfected with the TN-C promoter ligated to the gene sequence for destabilized green fluorescent protein (GFP). Fully automated image analysis routines, validated by comparison with data derived from manual segmentation and tracking of single cells, are used to quantify changes in the cellular GFP in hundreds of individual cells throughout their cell cycle during live cell imaging experiments lasting 62 h. We find that individual cells vary substantially in their expression patterns over the cell cycle, but that on average TN-C promoter activity increases during the last 40% of the cell cycle. We also find that the increase in promoter activity is proportional to the activity earlier in the cell cycle. This work illustrates the application of live cell microscopy and automated image analysis of a promoter-driven GFP reporter cell line to identify subtle gene regulatory mechanisms that are difficult to uncover using population averaged measurements. %B Cytometry. Part A : the journal of the International Society for Analytical Cytology %V 79 %P 192-202 %8 2011 Mar %G eng %N 3 %1 http://www.ncbi.nlm.nih.gov/pubmed/22045641?dopt=Abstract %R 10.1002/cyto.a.21028 %0 Conference Paper %D 2011 %T Characterizing Attackers and Attacks: An Empirical Study %A Salles-Loustau,G. %A Berthier,R. %A Collange,E. %A Sobesto,B. %A Michel Cukier %K attack sessions %K attacker characterization %K attacker skill measurement %K honey net infrastructure %K honey pot configurations %K IP address %K keystroke profile analysis %K opportunity target %K rogue software exploitation %K security of data %K SSH-based authentication proxy %X This paper describes an empirical research study to characterize attackers and attacks against targets of opportunity. A honey net infrastructure was built and deployed over 167 days that leveraged three different honey pot configurations and a SSH-based authentication proxy to attract and follow attackers over several weeks. A total of 211 attack sessions were recorded and evidence was collected at each stage of the attack sequence: from discovery to intrusion and exploitation of rogue software. This study makes two important contributions: 1) we introduce a new approach to measure attacker skills, and 2) we leverage keystroke profile analysis to differentiate attackers beyond their IP address of origin. %P 174 - 183 %8 2011/12// %G eng %R 10.1109/PRDC.2011.29 %0 Journal Article %J Interactions %D 2011 %T Claiming success, charting the future: micro-HCI and macro-HCI %A Shneiderman, Ben %B Interactions %V 18 %P 10 - 11 %8 2011/// %G eng %N 5 %0 Journal Article %J ACM SIGCOMM Computer Communication Review %D 2011 %T Communicating with caps: managing usage caps in home networks %A Kim,Hyojoon %A Sundaresan,Srikanth %A Marshini Chetty %A Feamster, Nick %A Edwards,W. Keith %K home network %K openflow %K usage cap %X As Internet service providers increasingly implement and impose "usage caps", consumers need better ways to help them understand and control how devices in the home use up the available network resources or available capacity. Towards this goal, we will demonstrate a system that allows users to monitor and manage their usage caps. The system uses the BISMark firmware running on network gateways to collect usage statistics and report them to a logically centralized controller, which displays usage information. The controller allows users to specify policies about how different people, devices, and applications should consume the usage cap; it implements and enforces these policies via a secure OpenFlow control channel to each gateway device. The demonstration will show various use cases, such as limiting the usage of a particular application, visualizing usage statistics, and allowing users within a single household to "trade" caps with one another. %B ACM SIGCOMM Computer Communication Review %V 41 %P 470 - 471 %8 2011/08// %@ 0146-4833 %G eng %U http://doi.acm.org/10.1145/2043164.2018526 %N 4 %R 10.1145/2043164.2018526 %0 Journal Article %J SIGCOMM Comput. Commun. Rev. %D 2011 %T Communicating with Caps: Managing Usage Caps in Home Networks %A Kim, Hyojoon %A Sundaresan, Srikanth %A Marshini Chetty %A Feamster, Nick %A Edwards, W. Keith %K home network %K openflow %K usage cap %X As Internet service providers increasingly implement and impose "usage caps", consumers need better ways to help them understand and control how devices in the home use up the available network resources or available capacity. Towards this goal, we will demonstrate a system that allows users to monitor and manage their usage caps. The system uses the BISMark firmware running on network gateways to collect usage statistics and report them to a logically centralized controller, which displays usage information. The controller allows users to specify policies about how different people, devices, and applications should consume the usage cap; it implements and enforces these policies via a secure OpenFlow control channel to each gateway device. The demonstration will show various use cases, such as limiting the usage of a particular application, visualizing usage statistics, and allowing users within a single household to "trade" caps with one another. %B SIGCOMM Comput. Commun. Rev. %V 41 %P 470 - 471 %8 2011/08// %@ 0146-4833 %G eng %U http://doi.acm.org/10.1145/2043164.2018526 %N 4 %0 Journal Article %J Development %D 2011 %T A computational statistics approach for estimating the spatial range of morphogen gradients %A Kanodia, Jitendra S. %A Kim, Yoosik %A Tomer, Raju %A Zia Khan %A Chung, Kwanghun %A Storey, John D. %A Lu, Hang %A Keller, Philipp J. %A Shvartsman, Stanislav Y. %K Computational Biology %K Confidence interval %K Dorsal gradient %K Drosophila embryo %K Morphogen gradient %K Statistics %X A crucial issue in studies of morphogen gradients relates to their range: the distance over which they can act as direct regulators of cell signaling, gene expression and cell differentiation. To address this, we present a straightforward statistical framework that can be used in multiple developmental systems. We illustrate the developed approach by providing a point estimate and confidence interval for the spatial range of the graded distribution of nuclear Dorsal, a transcription factor that controls the dorsoventral pattern of the Drosophila embryo. %B Development %V 138 %P 4867 - 4874 %8 2011/11/15/ %@ 0950-1991, 1477-9129 %G eng %U http://dev.biologists.org/content/138/22/4867 %N 22 %! Development %0 Journal Article %J Journal of Discrete Algorithms %D 2011 %T A cost-aggregating integer linear program for motif finding %A Kingsford, Carl %A Zaslavsky,Elena %A Singh,Mona %K Computational Biology %K Integer linear programming %K Sequence motif finding %X In the motif finding problem one seeks a set of mutually similar substrings within a collection of biological sequences. This is an important and widely-studied problem, as such shared motifs in DNA often correspond to regulatory elements. We study a combinatorial framework where the goal is to find substrings of a given length such that the sum of their pairwise distances is minimized. We describe a novel integer linear program for the problem, which uses the fact that distances between substrings come from a limited set of possibilities allowing for aggregate consideration of sequence position pairs with the same distances. We show how to tighten its linear programming relaxation by adding an exponential set of constraints and give an efficient separation algorithm that can find violated constraints, thereby showing that the tightened linear program can still be solved in polynomial time. We apply our approach to find optimal solutions for the motif finding problem and show that it is effective in practice in uncovering known transcription factor binding sites. %B Journal of Discrete Algorithms %V 9 %P 326 - 334 %8 2011/12// %@ 1570-8667 %G eng %U http://www.sciencedirect.com/science/article/pii/S157086671100044X %N 4 %R 10.1016/j.jda.2011.04.001 %0 Report %D 2011 %T Dataflow-based Design and Implementation of Image Processing Applications %A Chung-Ching Shen %A Plishker,William %A Bhattacharyya, Shuvra S. %K Technical Report %X Dataflow is a well known computational model and is widely used forexpressing the functionality of digital signal processing (DSP) applications, such as audio and video data stream processing, digital communications, and image processing. These applications usually require real-time processing capabilities and have critical performance constraints. Dataflow provides a formal mechanism for describing specifications of DSP applications, imposes minimal data-dependency constraints in specifications, and is effective in exposing and exploiting task or data level parallelism for achieving high performance implementations. To demonstrate dataflow-based design methods in a manner that is concrete and easily adapted to different platforms and back-end design tools, we present in this report a number of case studies based on the lightweight dataflow (LWDF) programming methodology. LWDF is designed as a "minimalistic" approach for integrating coarse grain dataflow programming structures into arbitrary simulation- or platform-oriented languages, such as C, C++, CUDA, MATLAB, SystemC, Verilog, and VHDL. In particular, LWDF requires minimal dependence on specialized tools or libraries. This feature --- together with the rigorous adherence to dataflow principles throughout the LWDF design framework --- allows designers to integrate and experiment with dataflow modeling approaches relatively quickly and flexibly into existing design methodologies and processes. %B Technical Reports from UMIACS %8 2011 %@ UMIACS-TR-2011-11 %G eng %U http://drum.lib.umd.edu//handle/1903/11403 %R Technical Report %0 Report %D 2011 %T Dataflow-Based Implementation of Layered Sensing Applications %A Bhattacharyya, Shuvra S. %A Chung-Ching Shen %A Plishker,William %A Sane, Nimish %A Wu, Hsiang-Huang %A Gu, Ruirui %K *COMPUTER AIDED DESIGN %K *DATA FUSION %K *DATAFLOW %K *LAYERS %K *SOFTWARE TOOLS %K COMPUTER PROGRAMMING AND SOFTWARE %K DETECTION %K High performance computing %K LAYERED SENSING %K OPTIMIZATION %K Signal processing %K synthesis %K T2KA %X This report describes a new dataflow-based technology and associated design tools for high-productivity design, analysis, and optimization of layered sensing software for signal processing systems. Our approach provides novel capabilities, based on the principles of task-level dataflow analysis, for exploring and optimizing interactions across application behavior; operational context; high performance embedded processing platforms, and implementation constraints. Particularly, we introduce and deliver novel software tools, called the targeted dataflow interchange format (TDIF) and Dataflow Interchange Format Markup Language (DIFML), for design and implementation of layered sensing and signal processing systems. The TDIF-CUDA (Compute Unified Device Architecture) environment is a graphics processing unit targeted software synthesis tool that provides a unique integration of dynamic dataflow modeling; retargetable actor construction; software synthesis; and instrumentation-based schedule evaluation and tuning. The DIFML package is a software package for the DIFML format, which is an Extensible Markup Language (XML)-based format for exchanging information between DIF and other tools. %8 2011 %G eng %0 Conference Paper %B 2011 Proceedings IEEE INFOCOM %D 2011 %T Decentralized, accurate, and low-cost network bandwidth prediction %A Sukhyun Song %A Keleher,P. %A Bhattacharjee, Bobby %A Sussman, Alan %K accuracy %K approximate tree metric space %K Bandwidth %K bandwidth allocation %K bandwidth measurement %K decentralized low cost system %K distributed tree %K end-to-end prediction %K Extraterrestrial measurements %K Internet %K low-cost network bandwidth prediction %K Measurement uncertainty %K pairwise bandwidth %K Peer to peer computing %K Prediction algorithms %K trees (mathematics) %X The distributed nature of modern computing makes end-to-end prediction of network bandwidth increasingly important. Our work is inspired by prior work that treats the Internet and bandwidth as an approximate tree metric space. This paper presents a decentralized, accurate, and low cost system that predicts pairwise bandwidth between hosts. We describe an algorithm to construct a distributed tree that embeds bandwidth measurements. The correctness of the algorithm is provable when driven by precise measurements. We then describe three novel heuristics that achieve high accuracy for predicting bandwidth even with imprecise input data. Simulation experiments with a real-world dataset confirm that our approach shows high accuracy with low cost. %B 2011 Proceedings IEEE INFOCOM %I IEEE %P 6 - 10 %8 2011/04/10/15 %@ 978-1-4244-9919-9 %G eng %R 10.1109/INFCOM.2011.5935251 %0 Conference Paper %B 2011 IEEE 36th Conference on Local Computer Networks (LCN) %D 2011 %T Design methods for Wireless Sensor Network Building Energy Monitoring Systems %A Cho, Inkeun %A Chung-Ching Shen %A Potbhare, S. %A Bhattacharyya, Shuvra S. %A Goldsman,N. %K Analytical models %K application-level interfacing behavior %K building energy monitoring system %K Buildings %K dataflow technique %K embedded sensor node %K energy analysis method %K Energy consumption %K energy management systems %K Energy resolution %K IEEE 802.15.4 MAC functionality %K Monitoring %K OPTIMIZATION %K wireless sensor network %K Wireless sensor networks %K WSNBEMS %K Zigbee %X In this paper, we present a new energy analysis method for evaluating energy consumption of embedded sensor nodes at the application level and the network level. Then we apply the proposed energy analysis method to develop new energy management schemes in order to maximize lifetime for Wireless Sensor Network Building Energy Monitoring Systems (WSNBEMS). At the application level, we develop a new design approach that uses dataflow techniques to model the application-level interfacing behavior between the processor and sensors on an embedded sensor node. At the network level, we analyze the energy consumption of the IEEE 802.15.4 MAC functionality. Based on our techniques for modeling and energy analysis, we have implemented an optimized WSNBEMS for a real building, and validated our energy analysis techniques through measurements on this implementation. The performance of our implementation is also evaluated in terms of monitoring accuracy and energy consumption savings. We have demonstrated that by applying the proposed scheme, system lifetime can be improved significantly without affecting monitoring accuracy. %B 2011 IEEE 36th Conference on Local Computer Networks (LCN) %P 974 - 981 %8 2011 %G eng %0 Conference Paper %B 2011 IEEE International Conference on Multimedia and Expo (ICME) %D 2011 %T A design tool for efficient mapping of multimedia applications onto heterogeneous platforms %A Chung-Ching Shen %A Wu, Hsiang-Huang %A Sane, N. %A Plishker,W. %A Bhattacharyya, Shuvra S. %K Dataflow graphs %K design tools %K embedded signal processing %K software synthesis %X Development of multimedia systems on heterogeneous platforms is a challenging task with existing design tools due to a lack of rigorous integration between high level abstract modeling, and low level synthesis and analysis. In this paper, we present a new dataflow-based design tool, called the targeted dataflow interchange format (TDIF), for design, analysis, and implementation of embedded software for multimedia systems. Our approach provides novel capabilities, based on the principles of task-level dataflow analysis, for exploring and optimizing interactions across application behavior; operational context; heterogeneous platforms, including high performance embedded processing architectures; and implementation constraints. %B 2011 IEEE International Conference on Multimedia and Expo (ICME) %P 1 - 6 %8 2011 %G eng %0 Journal Article %J Software Engineering, IEEE Transactions on %D 2011 %T Developing a Single Model and Test Prioritization Strategies for Event-Driven Software %A Bryce,R.C. %A Sampath,S. %A Memon, Atif M. %K EDS %K event-driven software %K graphical user interface %K Graphical user interfaces %K GUI testing %K Internet %K program testing %K service-oriented architecture %K test prioritization strategy %K Web application testing %X Event-Driven Software (EDS) can change state based on incoming events; common examples are GUI and Web applications. These EDSs pose a challenge to testing because there are a large number of possible event sequences that users can invoke through a user interface. While valuable contributions have been made for testing these two subclasses of EDS, such efforts have been disjoint. This work provides the first single model that is generic enough to study GUI and Web applications together. In this paper, we use the model to define generic prioritization criteria that are applicable to both GUI and Web applications. Our ultimate goal is to evolve the model and use it to develop a unified theory of how all EDS should be tested. An empirical study reveals that the GUI and Web-based applications, when recast using the new model, show similar behavior. For example, a criterion that gives priority to all pairs of event interactions did well for GUI and Web applications; another criterion that gives priority to the smallest number of parameter value settings did poorly for both. These results reinforce our belief that these two subclasses of applications should be modeled and studied together. %B Software Engineering, IEEE Transactions on %V 37 %P 48 - 64 %8 2011/// %@ 0098-5589 %G eng %N 1 %R 10.1109/TSE.2010.12 %0 Conference Paper %B Fifth International AAAI Conference on Weblogs and Social Media %D 2011 %T Differential adaptive diffusion: Understanding diversity and learning whom to trust in viral marketing %A Sharara,H. %A Rand, William %A Getoor, Lise %X Viral marketing mechanisms use the existing social network between customers to spread information about products and encourage product adoption. Existing viral marketing models focus on the dynamics of the diffusion process, however they typically: (a) only consider a single product campaign and (b) fail to model the evolution of the social network, as the trust between individuals changes over time, during the course of multiple campaigns. In this work, we propose an adaptive vi- ral marketing model which captures: (1) multiple different product campaigns, (2) the diversity in customer preferences among different product categories, and (3) changing con- fidence in peers’ recommendations over time. By applying our model to a real-world network extracted from the Digg social news website, we provide insights into the effects of network dynamics on the different products’ adoption. Our experiments show that our proposed model outperforms ear- lier non-adaptive diffusion models in predicting future prod- uct adoptions. We also show how this model can be used to explore new viral marketing strategies that are more success- ful than classic strategies which ignore the dynamic nature of social networks. %B Fifth International AAAI Conference on Weblogs and Social Media %8 2011/// %G eng %0 Book Section %B Distributed Video Sensor NetworksDistributed Video Sensor Networks %D 2011 %T Distributed Sensing and Processing for Multi-Camera Networks %A Sankaranarayanan,Aswin C. %A Chellapa, Rama %A Baraniuk,Richard G. %E Bhanu,Bir %E Ravishankar,Chinya V. %E Roy-Chowdhury,Amit K. %E Aghajan,Hamid %E Terzopoulos,Demetri %X Sensor networks with large numbers of cameras are becoming increasingly prevalent in a wide range of applications, including video conferencing, motion capture, surveillance, and clinical diagnostics. In this chapter, we identify some of the fundamental challenges in designing such systems: robust statistical inference, computationally efficiency, and opportunistic and parsimonious sensing. We show that the geometric constraints induced by the imaging process are extremely useful for identifying and designing optimal estimators for object detection and tracking tasks. We also derive pipelined and parallelized implementations of popular tools used for statistical inference in non-linear systems, of which multi-camera systems are examples. Finally, we highlight the use of the emerging theory of compressive sensing in reducing the amount of data sensed and communicated by a camera network. %B Distributed Video Sensor NetworksDistributed Video Sensor Networks %I Springer London %P 85 - 101 %8 2011/// %@ 978-0-85729-127-1 %G eng %U http://dx.doi.org/10.1007/978-0-85729-127-1_6 %0 Conference Paper %D 2011 %T Do You Know Where Your Data Are? Secure Data Capsules for Deployable Data Protection %A Maniatis, P. %A Akhawe, D. %A Fall,K. %A Elaine Shi %A McCamant, S. %A Song,D. %8 2011 %G eng %U http://static.usenix.org/event/hotos11/tech/final_files/ManiatisAkhawe.pdf %0 Report %D 2011 %T The DSPCAD Integrative Command Line Environment: Introduction to DICE Version 1.1 %A Bhattacharyya, Shuvra S. %A Plishker,William %A Chung-Ching Shen %A Sane, Nimish %A Zaki, George %K *SOFTWARE ENGINEERING %K COMPUTER PROGRAMMING %K COMPUTER PROGRAMMING AND SOFTWARE %K COMPUTER PROGRAMS %K DICE(COMPUTER PROGRAM) %K programming languages %K Project management %X DICE (the DSPCAD Integrative Command Line Environment) is a package of utilities that facilitates efficient management of software projects. Key areas of emphasis in DICE are cross-platform operation, support for projects that integrate heterogeneous programming languages, and support for applying and integrating different kinds of design and testing methodologies. The package is being developed at the University of Maryland to facilitate the research and teaching of methods for implementation, testing, evolution, and revision of engineering software. The package is also being developed as a foundation for developing experimental research software for techniques and tools in the area of computer-aided design (CAD) of digital signal processing (DSP) systems. The package is intended for cross-platform operation, and is currently being developed and used actively on the Linux, Mac OS, Solaris, and Windows (equipped with Cygwin) platforms. This report provides an introduction to DICE, and provides background on some of the key features in DICE Version 1.1. This report also gives a brief introduction to dicelang, which is a plug-in package for DICE that provides additional utilities, libraries, and tools for managing software projects in specific programming languages. %8 2011 %G eng %0 Report %D 2011 %T The DSPCAD Lightweight Dataflow Environment: Introduction to LIDE Version 0.1 %A Chung-Ching Shen %A Wang, Lai-Huei %A Cho, Inkeun %A Kim, Scott %A Won, Stephen %A Plishker,William %A Bhattacharyya, Shuvra S. %K Technical Report %X LIDE (the DSPCAD Lightweight Dataflow Environment) is a flexible,lightweight design environment that allows designers to experiment with dataflow-based approaches for design and implementation of digital signal processing (DSP) systems. LIDE contains libraries of dataflow graph elements (primitive actors, hierarchical actors, and edges) and utilities that assist designers in modeling, simulating, and implementing DSP systems using formal dataflow techniques. The libraries of dataflow graph elements (mainly actors) contained in LIDE provide useful building blocks that can be used to construct signal processing applications, and that can be used as examples that designers can adapt to create their own, customized LIDE actors. Furthermore, by using LIDE along with the DSPCAD Integrative Command Line Environment (DICE), designers can efficiently create and execute unit tests for user-designed actors. This report provides an introduction to LIDE. The report includes details on the process for setting up the LIDE environment, and covers methods for using pre-designed libraries of graph elements, as well as creating user-designed libraries and associated utilities using the C language. The report also gives an introduction to the C language plug-in for dicelang. This plug-in, called dicelang-C, provides features for efficient C-based project development and maintenance that are useful to apply when working with LIDE. %B Technical Reports from UMIACS %8 2011 %@ UMIACS-TR-2011-17 %G eng %U http://drum.lib.umd.edu//handle/1903/12147 %R Technical Report %0 Conference Paper %B Computer Security Foundations Symposium (CSF), 2011 IEEE 24th %D 2011 %T Dynamic Enforcement of Knowledge-Based Security Policies %A Mardziel,P. %A Magill,S. %A Hicks, Michael W. %A Srivatsa,M. %K abstract interpretation %K belief networks %K belief tracking %K Data models %K dynamic enforcement %K Facebook %K information flow %K knowledge based systems %K knowledge-based security %K knowledge-based security policy %K privacy %K probabilistic computation %K probabilistic logic %K probabilistic polyhedral domain %K probabilistic polyhedron %K probability %K query analysis %K Security %K security of data %K semantics %K Waste materials %X This paper explores the idea of knowledge-based security policies, which are used to decide whether to answer queries over secret data based on an estimation of the querier's (possibly increased) knowledge given the results. Limiting knowledge is the goal of existing information release policies that employ mechanisms such as noising, anonymization, and redaction. Knowledge-based policies are more general: they increase flexibility by not fixing the means to restrict information flow. We enforce a knowledge-based policy by explicitly tracking a model of a querier's belief about secret data, represented as a probability distribution, and denying any query that could increase knowledge above a given threshold. We implement query analysis and belief tracking via abstract interpretation using a novel probabilistic polyhedral domain, whose design permits trading off precision with performance while ensuring estimates of a querier's knowledge are sound. Experiments with our implementation show that several useful queries can be handled efficiently, and performance scales far better than would more standard implementations of probabilistic computation based on sampling. %B Computer Security Foundations Symposium (CSF), 2011 IEEE 24th %I IEEE %P 114 - 128 %8 2011/06/27/29 %@ 978-1-61284-644-6 %G eng %R 10.1109/CSF.2011.15 %0 Journal Article %J AI Magazine %D 2011 %T EAAI-10: The First Symposium on Educational Advances in Artificial Intelligence %A desJardins, Marie %A Sahami,Mehran %A Wagstaff,Kiri %K education, AI, symposium %B AI Magazine %V 32 %P 91 - 91 %8 2011/04/18/ %@ 0738-4602 %G eng %U http://www.aaai.org/ojs/index.php/aimagazine/article/view/2323 %N 1 %R 10.1609/aimag.v32i1.2323 %0 Conference Paper %B Proceedings of the 42nd ACM technical symposium on Computer science education %D 2011 %T Educational advances in artificial intelligence %A Sahami,Mehran %A desJardins, Marie %A Dodds,Zachary %A Neller,Todd %K artificial intelligence education %K model AI assignments %X In 2010 a new annual symposium on Educational Advances in Artificial Intelligence (EAAI) was launched as part of the AAAI annual meeting. The event was held in cooperation with ACM SIGCSE and has many similar goals related to broadening and disseminating work in computer science education. EAAI has a particular focus, however, as the event is specific to educational work in Artificial Intelligence and collocated with a major research conference (AAAI) to promote more interaction between researchers and educators in that domain. This panel seeks to introduce participants to EAAI as a way of fostering more interaction between educational communities in computing. Specifically, the panel will discuss the goals of EAAI, provide an overview of the kinds of work presented at the symposium, and identify potential synergies between that EAAI and SIGCSE as a way of better linking the two communities going forward. %B Proceedings of the 42nd ACM technical symposium on Computer science education %S SIGCSE '11 %I ACM %C New York, NY, USA %P 81 - 82 %8 2011/// %@ 978-1-4503-0500-6 %G eng %U http://doi.acm.org/10.1145/1953163.1953189 %R 10.1145/1953163.1953189 %0 Conference Paper %D 2011 %T Efficient and secure threshold-based event validation for vanets %A Hsiao, H. C. %A Studer, A. %A Dubey, R. %A Elaine Shi %A Perrig, A. %X Determining whether the number of vehicles reporting anevent is above a threshold is an important mechanism for VANETs, because many applications rely on a threshold number of notifications to reach agreement among vehicles, to determine the validity of an event, or to prevent the abuse of emergency alarms. We present the first efficient and se- cure threshold-based event validation protocol for VANETs. Quite counter-intuitively, we found that the z-smallest ap- proach offers the best tradeoff between security and effi- ciency since other approaches perform better for probabilis- tic counting. Analysis and simulation shows that our pro- tocol provides > 99% accuracy despite the presence of at- tackers, collection and distribution of alerts in less than 1 second, and negligible impact on network performance. %8 2011 %G eng %U http://www.eecs.berkeley.edu/~elaines/docs/hsiao_wisec2011.pdf %0 Journal Article %J Information Systems Security %D 2011 %T Efficient Secure Computation with Garbled Circuits %A Huang,Y. %A Shen,C. %A Evans,D. %A Katz, Jonathan %A Shelat,A. %X Secure two-party computation enables applications in which participants compute the output of a function that depends on their private inputs, without revealing those inputs or relying on any trusted third party. In this paper, we show the potential of building privacy-preserving applications using garbled circuits, a generic technique that until recently was believed to be too inefficient to scale to realistic problems. We present a Java-based framework that uses pipelining and circuit-level optimizations to build efficient and scalable privacy-preserving applications. Although the standard garbled circuit protocol assumes a very week, honest-but-curious adversary, techniques are available for converting such protocols to resist stronger adversaries, including fully malicious adversaries. We summarize approaches to producing malicious-resistant secure computations that reduce the costs of transforming a protocol to be secure against stronger adversaries. In addition, we summarize results on ensuring fairness, the property that either both parties receive the result or neither party does. Several open problems remain, but as theory and pragmatism advance, secure computation is approaching the point where it offers practical solutions for a wide variety of important problems. %B Information Systems Security %P 28 - 48 %8 2011/// %G eng %R 10.1007/978-3-642-25560-1_2 %0 Journal Article %J Interactions %D 2011 %T Electronic medical records: usability challenges and opportunities %A Hochheiser,H. %A Shneiderman, Ben %B Interactions %V 18 %P 48 - 49 %8 2011/// %G eng %N 6 %0 Journal Article %J RNA Biology %D 2011 %T Epigenomic and RNA structural correlates of polyadenylation %A Khaladkar,Mugdha %A Smyda,Mark %A Hannenhalli, Sridhar %X Polyadenylation (poly(A)) of mRNA plays a critical role in regulating gene expression. Identifying the sequence, structural, and epigenomic determinants of poly(A) site usage is an important long term goal. Several cis elements that mediate poly(A) regulation have been identified. Highly used poly(A) sites are also known to have a greater nucleosome occupancy in the immediate downstream. However, a detailed exploration of additional epigenomic and mRNA structural correlates of poly(A) site usage has not been reported. Importantly, functional interaction between sequence, structure, and the epigenome in determining the poly(A) site usage is not known. We show that highly used poly(A) sites are positively associated with an mRNA structure that is energetically more favorable and one that better exposes a critical polyadenylation cis element. In exploring potential interplay between RNA and chromatin structure, we found that a stronger nucleosome occupancy downstream of poly(A) site strongly correlated with (1) a more favorable mRNA structure, and (2) a greater accumulation of RNA Polymerase II (PolII) at the poly(A) site. Further analysis suggested a causal relationship pointing from PolII accumulation to a stable RNA structure. Additionally, we found that distinct patterns of histone modifications characterize poly(A) sites and these epigenetic patterns alone can distinguish true poly(A) sites with ~76% accuracy and also discriminate between high and low usage poly(A) sites with ~74% accuracy. Our results suggest a causative link between chromatin structure and mRNA structure whereby a compacted chromatin downstream of the poly(A) site slows down the elongating transcript, thus facilitating the folding of nascent mRNA in a favorable structure at poly(A) site during transcription. Additionally we report hitherto unknown epigenomic correlates for poly(A) site usage. %B RNA Biology %V 8 %P 529 - 537 %8 2011/05/01/ %@ 1547-6286 %G eng %U http://www.landesbioscience.com/journals/rnabiology/article/15194/ %N 3 %R 10.4161/rna.8.3.15194 %0 Journal Article %J EMERGING TOPICS IN COMPUTER VISION AND ITS APPLICATIONS %D 2011 %T ON THE ESTIMATION OF 3D HUMAN BODY MODELS AND POSE FROM MULTIPLE CAMERAS %A Sundaresan, A. %A Chellapa, Rama %X We present a completely automatic algorithm for initializing and tracking the articulated motion of humans using image sequences obtained from multiple cameras. We discuss the challenges in solving this problem and compare our work to some of the state of the art techniques today. We use a detailed articulated human body model composed of sixteen rigid segments that allows both translation and rotation at joints. Voxel data of the subject obtained from the images is segmented into the different articulated chains using Laplacian Eigenmaps. The segmented chains are registered in a subset of the frames using a single-frame registration technique and subsequently used to initialize the pose in the sequence. A temporal registration method is then used to identify the partially segmented or unregistered articulated chains in the remaining frames in the sequence. The tracker uses motion cues such as pixel displacement as well as 2D and 3D shape cues such as silhouettes, motion residues and skeleton curves. The use of complementary cues in the tracking algorithm alleviates the twin problems of drift and convergence to incorrect solutions. The use of multiple cameras also allows us to deal with the problems due to self-occlusion and kinematic singularity. We present tracking results on sequences with different kinds of motion to illustrate the effectiveness of our approach. %B EMERGING TOPICS IN COMPUTER VISION AND ITS APPLICATIONS %P 3 - 25 %8 2011/09//undefin %G eng %0 Journal Article %J IEEE Transactions on Software Engineering %D 2011 %T Evaluating Dynamic Software Update Safety Using Systematic Testing %A Hayden,C. %A Smith,E. %A Hardisty,E. %A Hicks, Michael W. %A Foster, Jeffrey S. %K and serviceability %K Availability %K Computer Systems Organization %K Delay %K Manuals %K Performance of Systems %K reliability %K Safety %K servers %K software %K software engineering %K Test execution %K Testing and Debugging %X Dynamic software updating (DSU) systems patch programs on the fly without incurring downtime. To avoid failures due to the updating process itself, many DSU systems employ timing restrictions. However, timing restrictions are theoretically imperfect, and their practical effectiveness is an open question. This paper presents the first significant empirical evaluation of three popular timing restrictions: activeness safety (AS), which prevents updates to active functions; con-freeness safety (CFS), which only allows modifications to active functions when doing so is provably type-safe; and manual identification of the event-handling loops during which an update may occur. We evaluated these timing restrictions using a series of DSU patches to three programs: OpenSSH, vsftpd, and ngIRCd. We systematically applied updates at each distinct update point reached during execution of a suite of system tests for these programs to determine which updates pass and which fail. We found that all three timing restrictions prevented most failures, but only manual identification allowed none. Further, although CFS and AS allowed many more update points, manual identification still supported updates with minimal delay. Finally, we found that manual identification required the least developer effort. Overall, we conclude that manual identification is most effective. %B IEEE Transactions on Software Engineering %V PP %P 1 - 1 %@ 0098-5589 %G eng %N 99 %R 10.1109/TSE.2011.101 %0 Conference Paper %B 2011 IEEE Symposium on Visual Languages and Human-Centric Computing (VL/HCC) %D 2011 %T Evaluating visual and statistical exploration of scientific literature networks %A Gove,R. %A Dunne,C. %A Shneiderman, Ben %A Klavans,J. %A Dorr, Bonnie J %K abstracting %K academic literature %K action science explorer %K automatic clustering %K citation analysis %K citation network visualization %K Communities %K Context %K custom exploration goal %K Data visualization %K Databases %K Document filtering %K document handling %K document ranking %K easy-to-understand metrics %K empirical evaluation %K Google %K Graphical user interfaces %K Information filtering %K Information Visualization %K Libraries %K literature exploration %K network statistics %K paper filtering %K paper ranking %K scientific literature network %K statistical exploration %K summarization technique %K user-defined tasks %K visual exploration %K Visualization %X Action Science Explorer (ASE) is a tool designed to support users in rapidly generating readily consumable summaries of academic literature. It uses citation network visualization, ranking and filtering papers by network statistics, and automatic clustering and summarization techniques. We describe how early formative evaluations of ASE led to a mature system evaluation, consisting of an in-depth empirical evaluation with four domain experts. The evaluation tasks were of two types: predefined tasks to test system performance in common scenarios, and user-defined tasks to test the system's usefulness for custom exploration goals. The primary contribution of this paper is a validation of the ASE design and recommendations to provide: easy-to-understand metrics for ranking and filtering documents, user control over which document sets to explore, and overviews of the document set in coordinated views along with details-on-demand of specific papers. We contribute a taxonomy of features for literature search and exploration tools and describe exploration goals identified by our participants. %B 2011 IEEE Symposium on Visual Languages and Human-Centric Computing (VL/HCC) %I IEEE %P 217 - 224 %8 2011/09/18/22 %@ 978-1-4577-1246-3 %G eng %R 10.1109/VLHCC.2011.6070403 %0 Conference Paper %B 2011 44th Hawaii International Conference on System Sciences (HICSS) %D 2011 %T EventGraphs: Charting Collections of Conference Connections %A Hansen,D. %A Smith,M. A %A Shneiderman, Ben %K conference connections %K Conferences %K Data visualization %K EventGraphs %K hashtag %K measurement %K Media %K message identification %K multimedia computing %K NodeXL %K Real time systems %K social media network diagrams %K social networking (online) %K Twitter %X EventGraphs are social media network diagrams of conversations related to events, such as conferences. Many conferences now communicate a common "hashtag" or keyword to identify messages related to the event. EventGraphs help make sense of the collections of connections that form when people follow, reply or mention one another and a keyword. This paper defines EventGraphs, characterizes different types, and shows how the social media network analysis add-in NodeXL supports their creation and analysis. The structural patterns to look for in EventGraphs are highlighted and design ideas for their improvement are discussed. %B 2011 44th Hawaii International Conference on System Sciences (HICSS) %I IEEE %P 1 - 10 %8 2011/01/04/7 %@ 978-1-4244-9618-1 %G eng %R 10.1109/HICSS.2011.196 %0 Journal Article %J IFIP Lecture Notes in Computer Science (LNCS) %D 2011 %T Evolution in Action: Using Active Networking to Evolve Network Support for Mobility %A Hicks, Michael W. %A Nettles,S. %A Seong-Kyu,S. %A Shannon,S. %X key early objective of Active Networking (AN) was to support on-the-fly network evolution. Although AN has been used relatively extensively to build application-customized protocols and even whole networking systems, demonstrations of evolution have been limited. %B IFIP Lecture Notes in Computer Science (LNCS) %V 2546 %P 146 - 161 %8 2011/// %G eng %N 2546 %0 Conference Paper %B Image Processing (ICIP), 2011 18th IEEE International Conference on %D 2011 %T Exploring compression effects for improved source camera identification using strongly compressed video %A Chuang,Wei-Hong %A Su,Hui %A M. Wu %K camera %K cameras;video %K coding; %K compression;computer %K compression;video %K estimation;photoresponse %K forensics;data %K identification;video %K non-uniformity;reliability;source %K PRNU %X This paper presents a study of the video compression effect on source camera identification based on the Photo-Response Non-Uniformity (PRNU). Specifically, the reliability of different types of frames in a compressed video is first investigated, which shows quantitatively that I-frames are more reliable than P-frames for PRNU estimation. Motivated by this observation, a new mechanism for estimating the reference PRNU and two mechanisms for estimating the test-video PRNU are proposed to achieve higher accuracy with fewer frames used. Experiments are performed to validate the effectiveness of the proposed mechanisms. %B Image Processing (ICIP), 2011 18th IEEE International Conference on %P 1953 - 1956 %8 2011/09// %G eng %R 10.1109/ICIP.2011.6115855 %0 Journal Article %J Journal of Medical Systems %D 2011 %T Extracting Insights from Electronic Health Records: Case Studies, a Visual Analytics Process Model, and Design Recommendations %A Wang,Taowei %A Wongsuphasawat,Krist %A Plaisant, Catherine %A Shneiderman, Ben %X Current electronic health record (EHR) systems facilitate the storage, retrieval, persistence, and sharing of patient data. However, the way physicians interact with EHRs has not changed much. More specifically, support for temporal analysis of a large number of EHRs has been lacking. A number of information visualization techniques have been proposed to alleviate this problem. Unfortunately, due to their limited application to a single case study, the results are often difficult to generalize across medical scenarios. We present the usage data of Lifelines2 (Wang et al. 2008 ), our information visualization system, and user comments, both collected over eight different medical case studies. We generalize our experience into a visual analytics process model for multiple EHRs. Based on our analysis, we make seven design recommendations to information visualization tools to explore EHR systems. %B Journal of Medical Systems %V 35 %P 1135 - 1152 %8 2011/// %@ 0148-5598 %G eng %U http://dx.doi.org/10.1007/s10916-011-9718-x %N 5 %0 Book Section %B Visual Analysis of HumansVisual Analysis of Humans %D 2011 %T Face Detection %A Gopalan,Raghuraman %A Schwartz,William R. %A Chellapa, Rama %A Srivastava,Ankur %E Moeslund,Thomas B. %E Hilton,Adrian %E Krüger,Volker %E Sigal,Leonid %X Face detection in still images and videos has been extensively studied over the last two decades. Attributed to the recent proliferation of cameras in consumer applications, research in face detection has gradually transformed into more unconstrained settings, with the goal of achieving performance close to humans. This presents two main challenges: (i) in addition to modeling the facial characteristics, understanding the information portrayed by the surrounding scene is important in resolving visual ambiguities, and (ii) the computational time needed for decision making should be compatible for real-time applications, since detection is primarily a front-end process on which additional knowledge extraction is built upon. This chapter begins with a review of recent work in modeling face-specific information, including appearance-based methods used by sliding window classifiers, concepts from learning and local interest-point descriptors, and then focuses on representing the contextual information shared by faces with the surrounding scene. To provide better understanding of working concepts, we discuss a method for learning the semantic context shared by the face with other human body parts that facilitates reasoning under occlusion, and then present an image representation which efficiently encodes contour information to enable fast detection of faces. We conclude the chapter by discussing some existing challenges. %B Visual Analysis of HumansVisual Analysis of Humans %I Springer London %P 71 - 90 %8 2011/// %@ 978-0-85729-997-0 %G eng %U http://dx.doi.org/10.1007/978-0-85729-997-0_5 %0 Journal Article %J International Journal for Numerical Methods in Fluids %D 2011 %T Fast solvers for models of ICEO microfluidic flows %A Shuttleworth, R. R %A Elman, Howard %A Long, K. R %A Templeton, J. A %B International Journal for Numerical Methods in Fluids %V 65 %P 383 - 404 %8 2011/// %G eng %N 4 %0 Conference Paper %B Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %D 2011 %T From slacktivism to activism: participatory culture in the age of social media %A Rotman,Dana %A Vieweg,Sarah %A Yardi,Sarita %A Chi,Ed %A Preece,Jenny %A Shneiderman, Ben %A Pirolli,Peter %A Glaisyer,Tom %K activism %K change %K design %K participation %K slacktivism %K social media %X Social networking sites (e.g. Facebook), microblogging services (e.g. Twitter), and content-sharing sites (e.g. YouTube and Flickr) have introduced the opportunity for wide-scale, online social participation. Visibility of national and international priorities such as public health, political unrest, disaster relief, and climate change has increased, yet we know little about the benefits - and possible costs - of engaging in social activism via social media. These powerful social issues introduce a need for scientific research into technology mediated social participation. What are the actual, tangible benefits of "greening" Twitter profile pictures in support of the Iranian elections? Does cartooning a Facebook profile picture really raise awareness of child abuse? Are there unintended negative effects through low-risk, low-cost technology-mediated participation? And, is there a difference - in both outcome and engagement level - between different types of online social activism? This SIG will investigate technology mediated social participation through a critical lens, discussing both the potential positive and negative outcomes of such participation. Approaches to designing for increased participation, evaluating effects of participation, and next steps in scientific research directions will be discussed. %B Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %S CHI EA '11 %I ACM %C New York, NY, USA %P 819 - 822 %8 2011/// %@ 978-1-4503-0268-5 %G eng %U http://doi.acm.org/10.1145/1979742.1979543 %R 10.1145/1979742.1979543 %0 Journal Article %J Circulation: Cardiovascular GeneticsCirc Cardiovasc Genet %D 2011 %T Gene Coexpression Network Topology of Cardiac Development, Hypertrophy, and FailureClinical Perspective %A Dewey,Frederick E %A Perez,Marco V %A Wheeler,Matthew T %A Watt,Clifton %A Spin,Joshua %A Langfelder,Peter %A Horvath,Steve %A Hannenhalli, Sridhar %A Cappola,Thomas P. %A Ashley,Euan A %K fetal %K Gene expression %K heart failure %K hypertrophy %K myocardium %X Background— Network analysis techniques allow a more accurate reflection of underlying systems biology to be realized than traditional unidimensional molecular biology approaches. Using gene coexpression network analysis, we define the gene expression network topology of cardiac hypertrophy and failure and the extent of recapitulation of fetal gene expression programs in failing and hypertrophied adult myocardium.Methods and Results— We assembled all myocardial transcript data in the Gene Expression Omnibus (n=1617). Because hierarchical analysis revealed species had primacy over disease clustering, we focused this analysis on the most complete (murine) dataset (n=478). Using gene coexpression network analysis, we derived functional modules, regulatory mediators, and higher-order topological relationships between genes and identified 50 gene coexpression modules in developing myocardium that were not present in normal adult tissue. We found that known gene expression markers of myocardial adaptation were members of upregulated modules but not hub genes. We identified ZIC2 as a novel transcription factor associated with coexpression modules common to developing and failing myocardium. Of 50 fetal gene coexpression modules, 3 (6%) were reproduced in hypertrophied myocardium and 7 (14%) were reproduced in failing myocardium. One fetal module was common to both failing and hypertrophied myocardium. Conclusions— Network modeling allows systems analysis of cardiovascular development and disease. Although we did not find evidence for a global coordinated program of fetal gene expression in adult myocardial adaptation, our analysis revealed specific gene expression modules active during both development and disease and specific candidates for their regulation. %B Circulation: Cardiovascular GeneticsCirc Cardiovasc Genet %V 4 %P 26 - 35 %8 2011/02/01/ %@ 1942-325X, 1942-3268 %G eng %U http://circgenetics.ahajournals.org/content/4/1/26 %N 1 %R 10.1161/CIRCGENETICS.110.941757 %0 Journal Article %J Molecular Biology and EvolutionMol Biol Evol %D 2011 %T Genome-Wide Survey of Natural Selection on Functional, Structural, and Network Properties of Polymorphic Sites in Saccharomyces Paradoxus %A Vishnoi,Anchal %A Sethupathy,Praveen %A Simola,Daniel %A Plotkin,Joshua B. %A Hannenhalli, Sridhar %K derived allele frequency %K Evolution %K natural selection %K yeast %X Background. To characterize the genetic basis of phenotypic evolution, numerous studies have identified individual genes that have likely evolved under natural selection. However, phenotypic changes may represent the cumulative effect of similar evolutionary forces acting on functionally related groups of genes. Phylogenetic analyses of divergent yeast species have identified functional groups of genes that have evolved at significantly different rates, suggestive of differential selection on the functional properties. However, due to environmental heterogeneity over long evolutionary timescales, selection operating within a single lineage may be dramatically different, and it is not detectable via interspecific comparisons alone. Moreover, interspecific studies typically quantify selection on protein-coding regions using the Dn/Ds ratio, which cannot be extended easily to study selection on noncoding regions or synonymous sites. The population genetic-based analysis of selection operating within a single lineage ameliorates these limitations. Findings. We investigated selection on several properties associated with genes, promoters, or polymorphic sites, by analyzing the derived allele frequency spectrum of single nucleotide polymorphisms (SNPs) in 28 strains of Saccharomyces paradoxus. We found evidence for significant differential selection between many functionally relevant categories of SNPs, underscoring the utility of function-centric approaches for discovering signatures of natural selection. When comparable, our findings are largely consistent with previous studies based on interspecific comparisons, with one notable exception: our study finds that mutations from an ancient amino acid to a relatively new amino acid are selectively disfavored, whereas interspecific comparisons have found selection against ancient amino acids. Several of our findings have not been addressed through prior interspecific studies: we find that synonymous mutations from preferred to unpreferred codons are selected against and that synonymous SNPs in the linker regions of proteins are relatively less constrained than those within protein domains. Conclusions. We present the first global survey of selection acting on various functional properties in S. paradoxus. We found that selection pressures previously detected over long evolutionary timescales have also shaped the evolution of S. paradoxus. Importantly, we also make novel discoveries untenable via conventional interspecific analyses. %B Molecular Biology and EvolutionMol Biol Evol %V 28 %P 2615 - 2627 %8 2011/09/01/ %@ 0737-4038, 1537-1719 %G eng %U http://mbe.oxfordjournals.org/content/28/9/2615 %N 9 %R 10.1093/molbev/msr085 %0 Conference Paper %B Visual Analytics Science and Technology (VAST), 2011 IEEE Conference on %D 2011 %T G-PARE: A visual analytic tool for comparative analysis of uncertain graphs %A Sharara,H. %A Sopan,A. %A Namata,G. %A Getoor, Lise %A Singh,L. %B Visual Analytics Science and Technology (VAST), 2011 IEEE Conference on %P 61 - 70 %8 2011/// %G eng %0 Magazine Article %D 2011 %T Graph Analytics-Lessons Learned and Challenges Ahead %A Wong,Pak Chung %A Chen,Chaomei %A Gorg,Carsten %A Shneiderman, Ben %A Stasko,John %A Thomas,Jim %K citation analysis %K citespace %K Computer Graphics %K document analysis %K graphics and multimedia %K greengrid %K jigsaw system %K modeling %K power grid analysis %K semantic substrates %K simulation %K social networks %K text analysis %K Visualization %X Graph analytics is one of the most influential and important R&D topics in the visual analytics community. Researchers with diverse backgrounds from information visualization, human-computer interaction, computer graphics, graph drawing, and data mining have pursued graph analytics from scientific, technical, and social approaches. These studies have addressed both distinct and common challenges. Past successes and mistakes can provide valuable lessons for revising the research agenda. In this article, six researchers from four academic and research institutes identify graph analytics' fundamental challenges and present both insightful lessons learned from their experience and good practices in graph analytics research. The goal is to critically assess those lessons and shed light on how they can stimulate research and draw attention to grand challenges for graph analytics. The article also establishes principles that could lead to measurable standards and criteria for research. %B IEEE Computer Graphics and Applications %V 31 %P 18 - 29 %8 2011/// %@ 0272-1716 %G eng %N 5 %0 Conference Paper %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %D 2011 %T Group-in-a-Box Layout for Multi-faceted Analysis of Communities %A Rodrigues,E.M. %A Milic-Frayling,N. %A Smith,M. %A Shneiderman, Ben %A Hansen,D. %K Algorithm design and analysis %K category based social graph partitions %K clustered graphs %K clustering %K Clustering algorithms %K Communities %K data visualisation %K force-directed %K gender %K geographic location %K graph layout algorithms %K graph theory %K group-in-a-box %K group-in-a-box layout %K Image edge detection %K Layout %K meta-layout %K multifaceted community analysis %K network subgraph visualization %K network visualization %K pattern clustering %K profession %K semantic substrates %K Social network services %K social networking (online) %K social networks %K treemap space filling technique %K Visualization %X Communities in social networks emerge from interactions among individuals and can be analyzed through a combination of clustering and graph layout algorithms. These approaches result in 2D or 3D visualizations of clustered graphs, with groups of vertices representing individuals that form a community. However, in many instances the vertices have attributes that divide individuals into distinct categories such as gender, profession, geographic location, and similar. It is often important to investigate what categories of individuals comprise each community and vice-versa, how the community structures associate the individuals from the same category. Currently, there are no effective methods for analyzing both the community structure and the category-based partitions of social graphs. We propose Group-In-a-Box (GIB), a meta-layout for clustered graphs that enables multi-faceted analysis of networks. It uses the tree map space filling technique to display each graph cluster or category group within its own box, sized according to the number of vertices therein. GIB optimizes visualization of the network sub-graphs, providing a semantic substrate for category-based and cluster-based partitions of social graphs. We illustrate the application of GIB to multi-faceted analysis of real social networks and discuss desirable properties of GIB using synthetic datasets. %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %I IEEE %P 354 - 361 %8 2011/10/09/11 %@ 978-1-4577-1931-8 %G eng %R 10.1109/PASSAT/SocialCom.2011.139 %0 Journal Article %J Briefings in Bioinformatics %D 2011 %T Hawkeye and AMOS: Visualizing and Assessing the Quality of Genome Assemblies %A Schatz,Michael C %A Phillippy,Adam M %A Sommer,Daniel D %A Delcher,Arthur L. %A Puiu,Daniela %A Narzisi,Giuseppe %A Salzberg,Steven L. %A Pop, Mihai %K assembly forensics %K DNA Sequencing %K genome assembly %K visual analytics %X Since its launch in 2004, the open-source AMOS project has released several innovative DNA sequence analysis applications including: Hawkeye, a visual analytics tool for inspecting the structure of genome assemblies; the Assembly Forensics and FRCurve pipelines for systematically evaluating the quality of a genome assembly; and AMOScmp, the first comparative genome assembler. These applications have been used to assemble and analyze dozens of genomes ranging in complexity from simple microbial species through mammalian genomes. Recent efforts have been focused on enhancing support for new data characteristics brought on by second- and now third-generation sequencing. This review describes the major components of AMOS in light of these challenges, with an emphasis on methods for assessing assembly quality and the visual analytics capabilities of Hawkeye. These interactive graphical aspects are essential for navigating and understanding the complexities of a genome assembly, from the overall genome structure down to individual bases. Hawkeye and AMOS are available open source at http://amos.sourceforge.net. %B Briefings in Bioinformatics %8 2011/12/23/ %@ 1467-5463, 1477-4054 %G eng %U http://bib.oxfordjournals.org/content/early/2011/12/23/bib.bbr074 %R 10.1093/bib/bbr074 %0 Conference Paper %B PART 2 ———– Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %D 2011 %T HCI for peace: from idealism to concrete steps %A Hourcade,Juan Pablo %A Bullock-Rest,Natasha E. %A Friedman,Batya %A Nelson,Mark %A Shneiderman, Ben %A Zaphiris,Panayiotis %K cyprus %K peace %K persuasive technology %K post-conflict reconciliation %K social media %K value sensitive design %K war %X This panel will contribute diverse perspectives on the use of computer technology to promote peace and prevent armed conflict. These perspectives include: the use of social media to promote democracy and citizen participation, the role of computers in helping people communicate across division lines in zones of conflict, how persuasive technology can promote peace, and how interaction design can play a role in post-conflict reconciliation. %B PART 2 ———– Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %S CHI EA '11 %I ACM %C New York, NY, USA %P 613 - 616 %8 2011/// %G eng %U http://doi.acm.org/10.1145/1979482.1979493 %R 10.1145/1979482.1979493 %0 Conference Paper %B ACM SIGCOMM Workshop on Home Networks '11 %D 2011 %T Helping Users Shop for ISPs with Internet Nutrition Labels %A Sundaresan, Srikanth %A Feamster, Nick %A Teixeira, Renata %A Tang, Anthony %A Edwards, W. Keith %A Grinter, Rebecca E. %A Marshini Chetty %A de Donato, Walter %K access networks %K benchmarking %K bismark %K broadband networks %K gateway measurements %X When purchasing home broadband access from Internet service providers (ISPs), users must decide which service plans are most appropriate for their needs. Today, ISPs advertise their available service plans using only generic upload and download speeds. Unfortunately, these metrics do not always accurately reflect the varying performance that home users will experience for a wide range of applications. In this paper, we propose that each ISP service plan carry a "nutrition label" that conveys more comprehensive information about network metrics along many dimensions, including various aspects of throughput, latency, loss rate, and jitter. We first justify why these metrics should form the basis of a network nutrition label. Then, we demonstrate that current plans that are superficially similar with respect to advertised download rates may have different performance according to the label metrics. We close with a discussion of the challenges involved in presenting a nutrition label to users in a way that is both accurate and easy to understand. %B ACM SIGCOMM Workshop on Home Networks '11 %S HomeNets '11 %I ACM %P 13 - 18 %8 2011/// %@ 978-1-4503-0798-7 %G eng %U http://doi.acm.org/10.1145/2018567.2018571 %0 Book Section %B Transactions on High-Performance Embedded Architectures and Compilers IV %D 2011 %T Heterogeneous Design in Functional DIF %A Plishker,William %A Sane, Nimish %A Kiemb, Mary %A Bhattacharyya, Shuvra S. %E Stenström, Per %K Arithmetic and Logic Structures %K Computer Communication Networks %K Dataflow %K heterogeneous %K Input/Output and Data Communications %K Logic Design %K Processor Architectures %K Programming Languages, Compilers, Interpreters %K Signal processing %X Dataflow formalisms have provided designers of digital signal processing (DSP) systems with analysis and optimizations for many years. As system complexity increases, designers are relying on more types of dataflow models to describe applications while retaining these implementation benefits. The semantic range of DSP-oriented dataflow models has expanded to cover heterogeneous models and dynamic applications, but efficient design, simulation, and scheduling of such applications has not. To facilitate implementing heterogeneous applications, we utilize a new dataflow model of computation and show how actors designed in other dataflow models are directly supported by this framework, allowing system designers to immediately compose and simulate actors from different models. Using examples, we show how this approach can be applied to quickly describe and functionally simulate a heterogeneous dataflow-based application such that a designer may analyze and tune trade-offs among different models and schedules for simulation time, memory consumption, and schedule size. %B Transactions on High-Performance Embedded Architectures and Compilers IV %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 391 - 408 %8 2011 %@ 978-3-642-24567-1, 978-3-642-24568-8 %G eng %U http://link.springer.com/chapter/10.1007/978-3-642-24568-8_20 %0 Book Section %B Visual Analysis of HumansVisual Analysis of Humans %D 2011 %T Image and Video-Based Biometrics %A Patel, Vishal M. %A Pillai,Jaishanker K. %A Chellapa, Rama %E Moeslund,Thomas B. %E Hilton,Adrian %E Krüger,Volker %E Sigal,Leonid %X Biometrics deals with the problem of identifying individuals based on physiological or behavioral characteristics. Since many physical characteristics, such as face, iris, etc., and behavioral characteristics, such as voice, expression, etc., are unique to an individual, biometric analysis offers a reliable and natural solution to the problem of identity verification. In this chapter, we discuss image and video-based biometrics involving face, iris and gait. In particular, we discuss several recent approaches to physiological biometrics based on Sparse Representations and Compressed Sensing. Some of the most compelling challenges and issues that confront research in biometrics are also addressed. %B Visual Analysis of HumansVisual Analysis of Humans %I Springer London %P 437 - 454 %8 2011/// %@ 978-0-85729-997-0 %G eng %U http://dx.doi.org/10.1007/978-0-85729-997-0_22 %0 Conference Paper %B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on %D 2011 %T Image ranking and retrieval based on multi-attribute queries %A Siddiquie,B. %A Feris,R.S. %A Davis, Larry S. %K datasets;image %K datasets;PASCAL %K faces %K FaceTracer %K in %K methods;labeled %K queries;image %K ranking;image %K retrieval %K retrieval; %K the %K VOC %K wild;multiattribute %X We propose a novel approach for ranking and retrieval of images based on multi-attribute queries. Existing image retrieval methods train separate classifiers for each word and heuristically combine their outputs for retrieving multiword queries. Moreover, these approaches also ignore the interdependencies among the query terms. In contrast, we propose a principled approach for multi-attribute retrieval which explicitly models the correlations that are present between the attributes. Given a multi-attribute query, we also utilize other attributes in the vocabulary which are not present in the query, for ranking/retrieval. Furthermore, we integrate ranking and retrieval within the same formulation, by posing them as structured prediction problems. Extensive experimental evaluation on the Labeled Faces in the Wild(LFW), FaceTracer and PASCAL VOC datasets show that our approach significantly outperforms several state-of-the-art ranking and retrieval methods. %B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on %P 801 - 808 %8 2011/06// %G eng %R 10.1109/CVPR.2011.5995329 %0 Book Section %B Information Security %D 2011 %T Implicit Authentication through Learning User Behavior %A Elaine Shi %A Niu, Yuan %A Jakobsson, Markus %A Chow, Richard %E Burmester, Mike %E Tsudik, Gene %E Magliveras, Spyros %E Ilic, Ivana %K Computer science %X Users are increasingly dependent on mobile devices. However, current authentication methods like password entry are significantly more frustrating and difficult to perform on these devices, leading users to create and reuse shorter passwords and pins, or no authentication at all. We present implicit authentication - authenticating users based on behavior patterns. We describe our model for performing implicit authentication and assess our techniques using more than two weeks of collected data from over 50 subjects. %B Information Security %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6531 %P 99 - 113 %8 2011 %@ 978-3-642-18177-1 %G eng %U http://www.springerlink.com/content/m57u551u3133475m/abstract/ %0 Journal Article %J Marine Biotechnology %D 2011 %T The Importance of Chitin in the Marine Environment %A Souza,C. P. %A Almeida,B. C. %A Rita R Colwell %A Rivera,I. N. G. %X Chitin is the most abundant renewable polymer in the oceans and is an important source of carbon and nitrogen for marine organisms. The process of chitin degradation is a key step in the cycling of nutrients in the oceans and chitinolytic bacteria play a significant role in this process. These bacteria are autochthonous to both marine and freshwater ecosystems and produce chitinases that degrade chitin, an insoluble polysaccharide, to a biologically useful form. In this brief review, a description of the structure of chitin and diversity of chitinolytic bacteria in the oceans is provided, in the context of the significance of chitin degradation for marine life. %B Marine Biotechnology %P 1 - 8 %8 2011/// %G eng %R 10.1007/s10126-011-9388-1 %0 Journal Article %J Theory of Cryptography %D 2011 %T Impossibility of blind signatures from one-way permutations %A Katz, Jonathan %A Schröder,D. %A Yerukhimovich,A. %X A seminal result in cryptography is that signature schemes can be constructed (in a black-box fashion) from any one-way function. The minimal assumptions needed to construct blind signature schemes, however, have remained unclear. Here, we rule out black-box constructions of blind signature schemes from one-way functions. In fact, we rule out constructions even from a random permutation oracle, and our results hold even for blind signature schemes for 1-bit messages that achieve security only against honest-but-curious behavior. %B Theory of Cryptography %P 615 - 629 %8 2011/// %G eng %R 10.1007/978-3-642-19571-6_37 %0 Journal Article %J Annals of Emergency Medicine %D 2011 %T Improved Identification and Visualization of Emergency Department Patient Visits %A Hettinger,AZ %A Rackoff,A. %A Wongsuphasawat,K. %A Cheng,H. %A Fairbanks,RJ %A Plaisant, Catherine %A Smith,M. S %B Annals of Emergency Medicine %V 58 %P S309 - S309 %8 2011/// %G eng %N 4 %0 Journal Article %J 2011 10th International Conference on Machine Learning and Applications Workshops %D 2011 %T Improving Classifier Performance by Autonomously Collecting Background Knowledge from the Web %A Minton,S.N. %A Michelson,M. %A See,K. %A Macskassy,S. %A Gazen,B.C. %A Getoor, Lise %X Many websites allow users to tag data items to makethem easier to find. In this paper we consider the problem of classifying tagged data according to user-specified interests. We present an approach for aggregating background knowledge from the Web to improve the performance of a classier. In previous work, researchers have developed technology for extracting knowledge, in the form of relational tables, from semi- structured websites. In this paper we integrate this extraction technology with generic machine learning algorithms, showing that knowledge extracted from the Web can significantly benefit the learning process. Specifically, the knowledge can lead to better generalizations, reduce the number of samples required for supervised learning, and eliminate the need to retrain the system when the environment changes. We validate the approach with an application that classifies tagged Fickr data. %B 2011 10th International Conference on Machine Learning and Applications Workshops %P 1 - 6 %8 2011/// %G eng %0 Journal Article %J Proceedings of the VLDB Endowment %D 2011 %T An incremental Hausdorff distance calculation algorithm %A Nutanong,Sarana %A Jacox,Edwin H. %A Samet, Hanan %X The Hausdorff distance is commonly used as a similarity measure between two point sets. Using this measure, a set X is considered similar to Y iff every point in X is close to at least one point in Y. Formally, the Hausdorff distance HausDist(X, Y) can be computed as the Max-Min distance from X to Y, i.e., find the maximum of the distance from an element in X to its nearest neighbor (NN) in Y. Although this is similar to the closest pair and farthest pair problems, computing the Hausdorff distance is a more challenging problem since its Max-Min nature involves both maximization and minimization rather than just one or the other. A traditional approach to computing HausDist(X, Y) performs a linear scan over X and utilizes an index to help compute the NN in Y for each x in X. We present a pair of basic solutions that avoid scanning X by applying the concept of aggregate NN search to searching for the element in X that yields the Hausdorff distance. In addition, we propose a novel method which incrementally explores the indexes of the two sets X and Y simultaneously. As an example application of our techniques, we use the Hausdorff distance as a measure of similarity between two trajectories (represented as point sets). We also use this example application to compare the performance of our proposed method with the traditional approach and the basic solutions. Experimental results show that our proposed method outperforms all competitors by one order of magnitude in terms of the tree traversal cost and total response time. %B Proceedings of the VLDB Endowment %V 4 %P 506 - 517 %8 2011/05// %@ 2150-8097 %G eng %U http://dl.acm.org/citation.cfm?id=2002974.2002978 %N 8 %0 Journal Article %J Information VisualizationInformation Visualization %D 2011 %T Information Visualization: State of the Field and New Research Directions %A Kerren,Andreas %A Plaisant, Catherine %A Stasko,John T %B Information VisualizationInformation Visualization %V 10 %P 269 - 270 %8 2011/10/01/ %@ 1473-8716, 1473-8724 %G eng %U http://ivi.sagepub.com/content/10/4/269 %N 4 %R 10.1177/1473871611418138 %0 Journal Article %J Microbial ecology %D 2011 %T Interaction of Vibrio cholerae non-O1/non-O139 with Copepods, Cladocerans and Competing Bacteria in the Large Alkaline Lake Neusiedler See, Austria %A Kirschner,A. K. T. %A Schauer,S. %A Steinberger,B. %A Wilhartitz,I. %A Grim,C. J. %A Huq,A. %A Rita R Colwell %A Herzig,A. %A Sommer,R. %X Vibrio cholerae is a human pathogen and natural inhabitant of aquatic environments. Serogroups O1/O139 have been associated with epidemic cholera, while non-O1/non-O139 serogroups usually cause human disease other than classical cholera. V. cholerae non-O1/non-O139 from the Neusiedler See, a large Central European lake, have caused ear and wound infections, including one case of fatal septicaemia. Recent investigations demonstrated rapid planktonic growth of V. cholerae non-O1/non-O139 and correlation with zooplankton biomass. The aim of this study was to elucidate the interaction of autochthonous V. cholerae with two dominant crustacean zooplankton species in the lake and investigate the influence of the natural bacterial community on this interaction. An existing data set was evaluated for statistical relationships between zooplankton species and V. cholerae and co-culture experiments were performed in the laboratory. A new fluorescence in situ hybridisation protocol was applied for quantification of V. cholerae non-O1/non-O139 cells, which significantly reduced analysis time. The experiments clearly demonstrated a significant relationship of autochthonous V. cholerae non-O1/non-O139 with cladocerans by promoting growth of V. cholerae non-O1/non-O139 in the water and on the surfaces of the cladocerans. In contrast, copepods had a negative effect on the growth of V. cholerae non-O1/non-O139 via competing bacteria from their surfaces. Thus, beside other known factors, biofilm formation by V. cholerae on crustacean zooplankton appears to be zooplankton taxon specific and may be controlled by the natural bacterial community. %B Microbial ecology %V 61 %P 496 - 506 %8 2011/// %G eng %N 3 %R 10.1007/s00248-010-9764-9 %0 Journal Article %D 2011 %T Interactive topic modeling %A Hu, Yuening %A Jordan Boyd-Graber %A Satinoff, Brianna %X Abstract Topic models have been used extensively as a tool for corpus exploration, and a cottage industry has developed to tweak topic models to better encode human intuitions or to better model data. However, creating such extensions requires expertise in machine ... %I Association for Computational Linguistics %P 248 - 257 %8 2011/00/19 %G eng %U http://dl.acm.org/citation.cfm?id=2002472.2002505 %0 Conference Paper %B Proceedings of the ACM SIGMETRICS joint international conference on Measurement and modeling of computer systems %D 2011 %T IP geolocation in metropolitan areas %A Singh,Satinder Pal %A Baden,Randolph %A Lee,Choon %A Bhattacharjee, Bobby %A La,Richard %A Shayman,Mark %K geolocation %K pattern recognition %K perturbation %X Current IP geoloation techniques can geolocate an IP address to a region approximately 700 square miles, roughly the size of a metropolitan area. We model geolocation as a pattern-recognition problem, and introduce techniques that geolocate addresses to within 5 miles inside a metropolitan area. We propose two complementary algorithms: The first algorithm, Pattern Based Geolocation (PBG), models the distribution of latencies to the target and compares it to those of the reference landmarks to resolve an address to within 5 miles in a metropolitan area. The second approach, Perturbation Augmented PBG (PAPBG), provides higher resolution by sending extra traffic in the network. While sending an aggregate of 600 Kbps extra traffic to 20 nodes for approximately 2 minutes, PAPBG geolocates addresses to within 3 miles. %B Proceedings of the ACM SIGMETRICS joint international conference on Measurement and modeling of computer systems %S SIGMETRICS '11 %I ACM %C New York, NY, USA %P 155 - 156 %8 2011/// %@ 978-1-4503-0814-4 %G eng %U http://doi.acm.org/10.1145/1993744.1993803 %R 10.1145/1993744.1993803 %0 Conference Paper %B Twelfth Annual Conference of the International Speech Communication Association %D 2011 %T Kernel partial least squares for speaker recognition %A Srinivasan,B.V. %A Garcia-Romero,D. %A Zotkin,Dmitry N %A Duraiswami, Ramani %X I-vectors are a concise representation of speaker characteristics. Recent advances in speaker recognition have utilized their ability to capture speaker and channel variability to develop efficient recognition engines. Inter-speaker relationships in the i-vector space are non-linear. Accomplishing effective speaker recognition requires a good modeling of these non-linearities and can be cast as a machine learning problem. In this paper, we propose a kernel partial least squares (kernel PLS, or KPLS) framework for modeling speakers in the i-vectors space. The resulting recognition system is tested across several conditions of the NIST SRE 2010 extended core data set and compared against state-of-the-art systems: Joint Factor Analysis (JFA), Probabilistic Linear Discriminant Analysis (PLDA), and Cosine Distance Scoring (CDS) classifiers. Improvements are shown. %B Twelfth Annual Conference of the International Speech Communication Association %8 2011/// %G eng %0 Journal Article %J Technical Reports of the Computer Science Department %D 2011 %T Kernelized Renyi distance for subset selection and similarity scoring %A Srinivasan,Balaji Vasan %A Duraiswami, Ramani %K Technical Report %X Renyi entropy refers to a generalized class of entropies that have beenused in several applications. In this work, we derive a non-parametric distance between distributions based on the quadratic Renyi entropy. The distributions are estimated via Parzen density estimates. The quadratic complexity of the distance evaluation is mitigated with GPU-based parallelization. This results in an efficiently evaluated non-parametric entropic distance - the kernelized Renyi distance or the KRD. We adapt the KRD into a similarity measure and show its application to speaker recognition. We further extend KRD to measure dissimilarities between distributions and illustrate its applications to statistical subset selection and dictionary learning for object recognition and pose estimation. %B Technical Reports of the Computer Science Department %8 2011/10/12/ %G eng %U http://drum.lib.umd.edu/handle/1903/12132 %0 Conference Paper %B Workshops at the Twenty-Fifth AAAI Conference on Artificial Intelligence %D 2011 %T Language Models for Semantic Extraction and Filtering in Video Action Recognition %A Tzoukermann,Evelyne %A Neumann, Jan %A Kosecka,Jana %A Fermüller, Cornelia %A Perera,Ian %A Ferraro,Frank %A Sapp,Ben %A Chaudhry,Rizwan %A Singh,Gautam %X The paper addresses the following issues: (a) how to represent semantic information from natural language so that a vision model can utilize it? (b) how to extract the salient textual information relevant to vision? For a given domain, we present a new model of semantic extraction that takes into account word relatedness as well as word disambiguation in order to apply to a vision model. We automatically process the text transcripts and perform syntactic analysis to extract dependency relations. We then perform semantic extraction on the output to filter semantic entities related to actions. The resulting data are used to populate a matrix of co-occurrences utilized by the vision processing modules. Results show that explicitly modeling the co-occurrence of actions and tools significantly improved performance. %B Workshops at the Twenty-Fifth AAAI Conference on Artificial Intelligence %8 2011/08/24/ %G eng %U https://www.aaai.org/ocs/index.php/WS/AAAIW11/paper/viewPaper/3919 %0 Conference Paper %B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on %D 2011 %T A large-scale benchmark dataset for event recognition in surveillance video %A Oh,Sangmin %A Hoogs, A. %A Perera,A. %A Cuntoor, N. %A Chen,Chia-Chih %A Lee,Jong Taek %A Mukherjee,S. %A Aggarwal, JK %A Lee,Hyungtae %A Davis, Larry S. %A Swears,E. %A Wang,Xioyang %A Ji,Qiang %A Reddy,K. %A Shah,M. %A Vondrick,C. %A Pirsiavash,H. %A Ramanan,D. %A Yuen,J. %A Torralba,A. %A Song,Bi %A Fong,A. %A Roy-Chowdhury, A. %A Desai,M. %K algorithm;evaluation %K CVER %K databases; %K databases;video %K dataset;moving %K event %K metrics;large-scale %K object %K recognition %K recognition;diverse %K recognition;video %K scenes;surveillance %K surveillance;visual %K tasks;computer %K tracks;outdoor %K video %K video;computer %K vision;continuous %K vision;image %K visual %X We introduce a new large-scale video dataset designed to assess the performance of diverse visual event recognition algorithms with a focus on continuous visual event recognition (CVER) in outdoor areas with wide coverage. Previous datasets for action recognition are unrealistic for real-world surveillance because they consist of short clips showing one action by one individual [15, 8]. Datasets have been developed for movies [11] and sports [12], but, these actions and scene conditions do not apply effectively to surveillance videos. Our dataset consists of many outdoor scenes with actions occurring naturally by non-actors in continuously captured videos of the real world. The dataset includes large numbers of instances for 23 event types distributed throughout 29 hours of video. This data is accompanied by detailed annotations which include both moving object tracks and event examples, which will provide solid basis for large-scale evaluation. Additionally, we propose different types of evaluation modes for visual recognition tasks and evaluation metrics along with our preliminary experimental results. We believe that this dataset will stimulate diverse aspects of computer vision research and help us to advance the CVER tasks in the years ahead. %B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on %P 3153 - 3160 %8 2011/06// %G eng %R 10.1109/CVPR.2011.5995586 %0 Conference Paper %B Proceedings of the 2011 annual conference on Human factors in computing systems %D 2011 %T LifeFlow: visualizing an overview of event sequences %A Wongsuphasawat,Krist %A Guerra Gómez,John Alexis %A Plaisant, Catherine %A Wang,Taowei David %A Taieb-Maimon,Meirav %A Shneiderman, Ben %K Information Visualization %K overview visualization %K temporal categorical data %K timestamped event sequences %X Event sequence analysis is an important task in many domains: medical researchers may study the patterns of transfers within the hospital for quality control; transportation experts may study accident response logs to identify best practices. In many cases they deal with thousands of records. While previous research has focused on searching and browsing, overview tasks are often overlooked. We introduce a novel interactive visual overview of event sequences called \emph{LifeFlow}. LifeFlow is scalable, can summarize all possible sequences, and represents the temporal spacing of the events within sequences. Two case studies with healthcare and transportation domain experts are presented to illustrate the usefulness of LifeFlow. A user study with ten participants confirmed that after 15 minutes of training novice users were able to rapidly answer questions about the prevalence and temporal characteristics of sequences, find anomalies, and gain significant insight from the data. %B Proceedings of the 2011 annual conference on Human factors in computing systems %S CHI '11 %I ACM %C New York, NY, USA %P 1747 - 1756 %8 2011/// %@ 978-1-4503-0228-9 %G eng %U http://doi.acm.org/10.1145/1978942.1979196 %R 10.1145/1978942.1979196 %0 Conference Paper %B Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %D 2011 %T LifeFlow: visualizing an overview of event sequences (video preview) %A Wongsuphasawat,Krist %A Guerra Gómez,John Alexis %A Plaisant, Catherine %A Wang,Taowei %A Taieb-Maimon,Meirav %A Shneiderman, Ben %K emergency room %K healthcare %K Information Visualization %K overview visualization %K temporal categorical data %K timestamped event sequences %X Event sequence analysis is an important task in many domains: medical researchers may study the patterns of transfers within the hospital for quality control; transportation experts may study accident response logs to identify best practices. In many cases they deal with thousands of records. While previous research has focused on searching and browsing, overview tasks are often overlooked. We introduce a novel interactive visual overview of event sequences called LifeFlow. LifeFlow is scalable, can summarize all possible sequences, and represents the temporal spacing of the events within sequences. In this video, we show an example of patient transfer data and briefly demonstrate how to analyze them with LifeFlow. Please see [11] or visit http:www.cs.umd.eduhcillifeflow for more detail. %B Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %S CHI EA '11 %I ACM %C New York, NY, USA %P 507 - 510 %8 2011/// %@ 978-1-4503-0268-5 %G eng %U http://doi.acm.org/10.1145/1979742.1979557 %R 10.1145/1979742.1979557 %0 Conference Paper %B Proceedings of the 16th ACM SIGPLAN international conference on Functional programming %D 2011 %T Lightweight monadic programming in ML %A Swamy,Nikhil %A Guts,Nataliya %A Leijen,Daan %A Hicks, Michael W. %K coercion %K coherence %K monad %K rewriting %K type %X Many useful programming constructions can be expressed as monads. Examples include probabilistic modeling, functional reactive programming, parsing, and information flow tracking, not to mention effectful functionality like state and I/O. In this paper, we present a type-based rewriting algorithm to make programming with arbitrary monads as easy as using ML's built-in support for state and I/O. Developers write programs using monadic values of type m τ as if they were of type τ, and our algorithm inserts the necessary binds, units, and monad-to-monad morphisms so that the program type checks. Our algorithm, based on Jones' qualified types, produces principal types. But principal types are sometimes problematic: the program's semantics could depend on the choice of instantiation when more than one instantiation is valid. In such situations we are able to simplify the types to remove any ambiguity but without adversely affecting typability; thus we can accept strictly more programs. Moreover, we have proved that this simplification is efficient (linear in the number of constraints) and coherent: while our algorithm induces a particular rewriting, all related rewritings will have the same semantics. We have implemented our approach for a core functional language and applied it successfully to simple examples from the domains listed above, which are used as illustrations throughout the paper. %B Proceedings of the 16th ACM SIGPLAN international conference on Functional programming %S ICFP '11 %I ACM %C New York, NY, USA %P 15 - 27 %8 2011/// %@ 978-1-4503-0865-6 %G eng %U http://doi.acm.org/10.1145/2034773.2034778 %R 10.1145/2034773.2034778 %0 Journal Article %J Theory of Cryptography %D 2011 %T Limits on the power of zero-knowledge proofs in cryptographic constructions %A Brakerski,Z. %A Katz, Jonathan %A Segev,G. %A Yerukhimovich,A. %X For over 20 years, black-box impossibility results have been used to argue the infeasibility of constructing certain cryptographic primitives (e.g., key agreement) from others (e.g., one-way functions). A widely recognized limitation of such impossibility results, however, is that they say nothing about the usefulness of (known) nonblack-box techniques. This is unsatisfying, as we would at least like to rule out constructions using the set of techniques we have at our disposal.With this motivation in mind, we suggest a new framework for black-box constructions that encompasses constructions with a nonblack-box flavor: specifically, those that rely on zero-knowledge proofs relative to some oracle. We show that our framework is powerful enough to capture the Naor-Yung/Sahai paradigm for building a (shielding) CCA-secure public-key encryption scheme from a CPA-secure one, something ruled out by prior black-box separation results. On the other hand, we show that several black-box impossibility results still hold even in a setting that allows for zero-knowledge proofs. %B Theory of Cryptography %P 559 - 578 %8 2011/// %G eng %R 10.1007/978-3-642-19571-6_34 %0 Journal Article %J IEEE Automatic Speech Recognition and Understanding Workshop %D 2011 %T Linear versus Mel Frequency Cepstral Coefficients for Speaker Recognition %A Zhou,X. %A Garcia-Romero,D. %A Duraiswami, Ramani %A Espy-Wilson,C. %A Shamma,S. %X Mel-frequency cepstral coefficients (MFCC) havebeen dominantly used in speaker recognition as well as in speech recognition. However, based on theories in speech production, some speaker characteristics associated with the structure of the vocal tract, particularly the vocal tract length, are reflected more in the high frequency range of speech. This insight suggests that a linear scale in frequency may provide some advantages in speaker recognition over the mel scale. Based on two state-of-the- art speaker recognition back-end systems (one Joint Factor Analysis system and one Probabilistic Linear Discriminant Analysis system), this study compares the performances between MFCC and LFCC (Linear frequency cepstral coefficients) in the NIST SRE (Speaker Recognition Evaluation) 2010 extended-core task. Our results in SRE10 show that, while they are complementary to each other, LFCC consistently outperforms MFCC, mainly due to its better performance in the female trials. This can be explained by the relatively shorter vocal tract in females and the resulting higher formant frequencies in speech. LFCC benefits more in female speech by better capturing the spectral characteristics in the high frequency region. In addition, our results show some advantage of LFCC over MFCC in reverberant speech. LFCC is as robust as MFCC in the babble noise, but not in the white noise. It is concluded that LFCC should be more widely used, at least for the female trials, by the mainstream of the speaker recognition community. %B IEEE Automatic Speech Recognition and Understanding Workshop %8 2011/// %G eng %0 Conference Paper %D 2011 %T Link prediction by de-anonymization: How We Won the Kaggle Social Network Challenge %A Narayanan, A. %A Elaine Shi %A Rubinstein, B.I.P. %K deanonymization %K Flickr social photo sharing Website %K graph theory %K IJCNN 2011 social network challenge %K Kaggle social network challenge %K learning (artificial intelligence) %K machine learning %K realworld link prediction %K Simulated annealing %K simulated annealing-based weighted graph matching algorithm %K social networking (online) %X This paper describes the winning entry to the IJCNN 2011 Social Network Challenge run by Kaggle.com. The goal of the contest was to promote research on real-world link prediction, and the dataset was a graph obtained by crawling the popular Flickr social photo sharing website, with user identities scrubbed. By de-anonymizing much of the competition test set using our own Flickr crawl, we were able to effectively game the competition. Our attack represents a new application of de-anonymization to gaming machine learning contests, suggesting changes in how future competitions should be run. We introduce a new simulated annealing-based weighted graph matching algorithm for the seeding step of de-anonymization. We also show how to combine de-anonymization with link prediction-the latter is required to achieve good performance on the portion of the test set not de-anonymized-for example by training the predictor on the de-anonymized portion of the test set, and combining probabilistic predictions from de-anonymization and link prediction. %P 1825 - 1834 %8 2011 %G eng %0 Journal Article %J Proceedings of the National Academy of SciencesPNAS %D 2011 %T Local Balancing Influences Global Structure in Social Networks %A Srinivasan, Aravind %X Although social networks can often be viewed as graphs in their most basic form, many networks of interest include explicit sentiments (positive or negative views) of the nodes (constituent entities, also known as vertices) toward each other. This is particularly the case in geopolitics, settings where networks of firms compete in the creation of standards, and situations where polarization is frequent, such as in national elections, etc. The classical theory of structural balance developed by Heider (1) suggests how nodes may modify their relationships locally to maintain a type of balance within triads of nodes; this theory has been enriched in terms of explicit temporal dynamics by Kulakowski et al. (2). The work of Marvel et al. (3) in PNAS shows a rigorous analysis of these dynamics and brings out interesting properties and applications thereof. %B Proceedings of the National Academy of SciencesPNAS %V 108 %P 1751 - 1752 %8 2011/02/01/ %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/108/5/1751 %N 5 %R 10.1073/pnas.1018901108 %0 Journal Article %J Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications %D 2011 %T Local Response Context Applied to Pedestrian Detection %A Schwartz,W. %A Davis, Larry S. %A Pedrini,H. %X Appearing as an important task in computer vision, pedestrian detection has been widely investigated in the recent years. To design a robust detector, we propose a feature descriptor called Local Response Context (LRC). This descriptor captures discriminative information regarding the surrounding of the person’s location by sampling the response map obtained by a generic sliding window detector. A partial least squares regression model using LRC descriptors is learned and employed as a second classification stage (after the execution of the generic detector to obtain the response map). Experiments based on the ETHZ pedestrian dataset show that the proposed approach improves significantly the results achieved by the generic detector alone and is comparable to the state-of-the-art methods. %B Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications %P 181 - 188 %8 2011/// %G eng %0 Book Section %B Pervasive Computing %D 2011 %T A Longitudinal Study of Pressure Sensing to Infer Real-World Water Usage Events in the Home %A Jon Froehlich %A Larson,Eric %A Saba,Elliot %A Campbell,Tim %A Atlas,Les %A Fogarty,James %A Patel,Shwetak %E Lyons,Kent %E Hightower,Jeffrey %E Huang,Elaine %X We present the first longitudinal study of pressure sensing to infer real-world water usage events in the home (e.g., dishwasher, upstairs bathroom sink, downstairs toilet). In order to study the pressure-based approach out in the wild , we deployed a ground truth sensor network for five weeks in three homes and two apartments that directly monitored valve-level water usage by fixtures and appliances . We use this data to, first, demonstrate the practical challenges in constructing water usage activity inference algorithms and, second, to inform the design of a new probabilistic-based classification approach. Inspired by algorithms in speech recognition, our novel Bayesian approach incorporates template matching, a language model, grammar, and prior probabilities. We show that with a single pressure sensor, our probabilistic algorithm can classify real-world water usage at the fixture level with 90% accuracy and at the fixturecategory level with 96% accuracy. With two pressure sensors, these accuracies increase to 94% and 98%. Finally, we show how our new approach can be trained with fewer examples than a strict template-matching approach alone. %B Pervasive Computing %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6696 %P 50 - 69 %8 2011 %@ 978-3-642-21725-8 %G eng %U http://dx.doi.org/10.1007/978-3-642-21726-5_4 %0 Book Section %B Pervasive ComputingPervasive Computing %D 2011 %T A Longitudinal Study of Pressure Sensing to Infer Real-World Water Usage Events in the Home %A Jon Froehlich %A Larson,Eric %A Saba,Elliot %A Campbell,Tim %A Atlas,Les %A Fogarty,James %A Patel,Shwetak %E Lyons,Kent %E Hightower,Jeffrey %E Huang,Elaine %X We present the first longitudinal study of pressure sensing to infer real-world water usage events in the home (e.g., dishwasher, upstairs bathroom sink, downstairs toilet). In order to study the pressure-based approach out in the wild , we deployed a ground truth sensor network for five weeks in three homes and two apartments that directly monitored valve-level water usage by fixtures and appliances . We use this data to, first, demonstrate the practical challenges in constructing water usage activity inference algorithms and, second, to inform the design of a new probabilistic-based classification approach. Inspired by algorithms in speech recognition, our novel Bayesian approach incorporates template matching, a language model, grammar, and prior probabilities. We show that with a single pressure sensor, our probabilistic algorithm can classify real-world water usage at the fixture level with 90% accuracy and at the fixturecategory level with 96% accuracy. With two pressure sensors, these accuracies increase to 94% and 98%. Finally, we show how our new approach can be trained with fewer examples than a strict template-matching approach alone. %B Pervasive ComputingPervasive Computing %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6696 %P 50 - 69 %8 2011/// %@ 978-3-642-21725-8 %G eng %U http://dx.doi.org/10.1007/978-3-642-21726-5_4 %0 Journal Article %J Theoretical Computer Science %D 2011 %T Maximum bipartite flow in networks with adaptive channel width %A Azar,Yossi %A Mądry,Aleksander %A Moscibroda,Thomas %A Panigrahi,Debmalya %A Srinivasan, Aravind %K Adaptive channel width %K graph algorithm %K Linear program rounding %K Maximum flow %K Wireless networks %X Traditionally, network optimization problems assume that each link in the network has a fixed capacity. Recent research in wireless networking has shown that it is possible to design networks where the capacity of the links can be changed adaptively to suit the needs of specific applications. In particular, one gets a choice of having a few high capacity outgoing links or many low capacity ones at any node of the network. This motivates us to have a re-look at classical network optimization problems and design algorithms to solve them in this new framework. In particular, we consider the problem of maximum bipartite flow, which has been studied extensively in the fixed-capacity network model. One of the motivations for studying this problem arises from the need to maximize the throughput of an infrastructure wireless network comprising base-stations (one set of vertices in the bipartition) and clients (the other set of vertices in the bipartition). We show that this problem has a significantly different combinatorial structure in this new network model from the fixed-capacity one. While there are several polynomial time algorithms for the maximum bipartite flow problem in traditional networks, we show that the problem is NP-hard in the new model. In fact, our proof extends to showing that the problem is APX-hard. We complement our lower bound by giving two algorithms for solving the problem approximately. The first algorithm is deterministic and achieves an approximation factor of O ( log N ) , where N is the number of nodes in the network, while the second algorithm is randomized and achieves an approximation factor of e e − 1 . %B Theoretical Computer Science %V 412 %P 2577 - 2587 %8 2011/05/27/ %@ 0304-3975 %G eng %U http://www.sciencedirect.com/science/article/pii/S0304397510005852 %N 24 %R 10.1016/j.tcs.2010.10.023 %0 Journal Article %J AMIA Annual Symposium ProceedingsAMIA Annu Symp Proc %D 2011 %T Medication Reconciliation: Work Domain Ontology, Prototype Development, and a Predictive Model %A Markowitz,Eliz %A Bernstam,Elmer V. %A Herskovic,Jorge %A Zhang,Jiajie %A Shneiderman, Ben %A Plaisant, Catherine %A Johnson,Todd R. %X Medication errors can result from administration inaccuracies at any point of care and are a major cause for concern. To develop a successful Medication Reconciliation (MR) tool, we believe it necessary to build a Work Domain Ontology (WDO) for the MR process. A WDO defines the explicit, abstract, implementation-independent description of the task by separating the task from work context, application technology, and cognitive architecture. We developed a prototype based upon the WDO and designed to adhere to standard principles of interface design. The prototype was compared to Legacy Health System’s and Pre-Admission Medication List Builder MR tools via a Keystroke-Level Model analysis for three MR tasks. The analysis found the prototype requires the fewest mental operations, completes tasks in the fewest steps, and completes tasks in the least amount of time. Accordingly, we believe that developing a MR tool, based upon the WDO and user interface guidelines, improves user efficiency and reduces cognitive load. %B AMIA Annual Symposium ProceedingsAMIA Annu Symp Proc %V 2011 %P 878 - 887 %8 2011/// %@ 1942-597X %G eng %0 Journal Article %J Bangladesh Journal of Microbiology %D 2011 %T Metagenomic 16S rDNA Targeted PCR-DGGE in Determining Bacterial Diversity in Aquatic Ecosystem %A Hasan,Nur A. %A Chowdhury,W Bari %A Rahim,Niaz %A Sultana,Marzia %A Shabnam,S Antara %A Mai,Volker %A Ali,Afsar %A Morris,Glen J %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %A Endtz,Hubert Ph %A Cravioto,Alejandro %A Alam,Munirul %X Bacterial numbers in surface water samples, collected randomly from six different water bodies, were estimated by acridine orange direct counting (AODC) and conventional culture-based heterotrophic plate counting (HPC). Bacterial genomic DNA was prepared from water samples by employing methods used for stool samples, including the population dynamics, were determined by primer extension of the 16S rDNA (V6/V8 region) using polymerase chain reaction (PCR), followed by denaturing gradient gel electrophoresis (DGGE), a metagenomic tool that is capable of separating unrelated DNAs based on the differences in their sequences and GC contents. The bacterial numbers in water samples ranged from 103 – 106 CFU/ mL for HPC and 104 – 107 cells/ mL for AODC, showing that a great majority of bacteria prevail as uncultivable which do not respond to culture methods that are used widely for tracking bacterial pathogens. The acridine orange-stained bacteria varied in sizes and shapes, and appeared either as planktonic (solitary) cells or as clusters of biofilms, showing the presence of diverse community under the epifluorescence microscope. The DGGE of the ca. 457 bp amplicons, as confirmed by agarose gel electrophoresis, produced bands that ranged in intensities and numbers from 18 to 31, with each band possibly indicating the presence of one or more closely related bacterial species. The enrichment of pathogenic bacteria in the aquatic ecosystem is known to precede the seasonal diarrhoeal outbreaks; therefore, bacterial community dynamics determined by Metagenomic 16S PCR-DGGE during pre-epidemic enrichment appears promising in predicting the upcoming diarrheal outbreaks. %B Bangladesh Journal of Microbiology %V 27 %8 2011/12/13/ %@ 1011-9981 %G eng %U http://www.banglajol.info/bd/index.php/BJM/article/view/9171 %N 2 %R 10.3329/bjm.v27i2.9171 %0 Patent %D 2011 %T Method and system for markerless motion capture using multiple cameras %A Sundaresan,Aravind %A Chellapa, Rama %E University of Maryland %X Completely automated end-to-end method and system for markerless motion capture performs segmentation of articulating objects in Laplacian Eigenspace and is applicable to handling of the poses of some complexity. 3D voxel representation of acquired images are mapped to a higher dimensional space (k), where k depends on the number of articulated chains of the subject body, so as to extract the 1-D representations of the articulating chains. A bottom-up approach is suggested in order to build a parametric (spline-based) representation of a general articulated body in the high dimensional space followed by a top-down probabilistic approach that registers the segments to an average human body model. The parameters of the model are further optimized using the segmented and registered voxels. %V 11/937,834 %8 2011/09/20/ %G eng %U http://www.google.com/patents?id=-THwAQAAEBAJ %N 8023726 %0 Conference Paper %B 2011 IEEE International Symposium on Parallel and Distributed Processing Workshops and Phd Forum (IPDPSW) %D 2011 %T A Model-Based Schedule Representation for Heterogeneous Mapping of Dataflow Graphs %A Wu, Hsiang-Huang %A Chung-Ching Shen %A Sane, N. %A Plishker,W. %A Bhattacharyya, Shuvra S. %K Computational modeling %K data flow graphs %K dataflow schedule graph %K dataflow semantics %K dataflow-based application specifications %K Dynamic scheduling %K heterogeneous mapping %K heterogeneous signal processing system design %K model-based design methodologies %K model-based schedule representation %K Processor scheduling %K Program processors %K Schedules %K semantics %K Signal processing %K synchronization %X Dataflow-based application specifications are widely used in model-based design methodologies for signal processing systems. In this paper, we develop a new model called the dataflow schedule graph (DSG) for representing a broad class of dataflow graph schedules. The DSG provides a graphical representation of schedules based on dataflow semantics. In conventional approaches, applications are represented using dataflow graphs, whereas schedules for the graphs are represented using specialized notations, such as various kinds of sequences or looping constructs. In contrast, the DSG approach employs dataflow graphs for representing both application models and schedules that are derived from them. Our DSG approach provides a precise, formal framework for unambiguously representing, analyzing, manipulating, and interchanging schedules. We develop detailed formulations of the DSG representation, and present examples and experimental results that demonstrate the utility of DSGs in the context of heterogeneous signal processing system design. %B 2011 IEEE International Symposium on Parallel and Distributed Processing Workshops and Phd Forum (IPDPSW) %P 70 - 81 %8 2011 %G eng %0 Conference Paper %B 2011 8th IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSS) %D 2011 %T Modeling and optimization of dynamic signal processing in resource-aware sensor networks %A Bhattacharyya, Shuvra S. %A Plishker,W. %A Sane, N. %A Chung-Ching Shen %A Wu, Hsiang-Huang %K Adaptation models %K Aerodynamics %K Computational modeling %K data flow graphs %K dataflow interchange format %K Dynamic scheduling %K dynamic signal processing %K power consumption overhead %K Program processors %K resource-aware sensor networks %K run-time adaptation %K Schedules %K sensor node processing %K Signal processing %K Wireless sensor networks %X Sensor node processing in resource-aware sensor networks is often critically dependent on dynamic signal processing functionality - i.e., signal processing functionality in which computational structure must be dynamically assessed and adapted based on time-varying environmental conditions, operating constraints or application requirements. In dynamic signal processing systems, it is important to provide flexibility for run-time adaptation of application behavior and execution characteristics, but in the domain of resource-aware sensor networks, such flexibility cannot come with significant costs in terms of power consumption overhead or reduced predictability. In this paper, we review a variety of complementary models of computation that are being developed as part of the dataflow interchange format (DIF) project to facilitate efficient and reliable implementation of dynamic signal processing systems. We demonstrate these methods in the context of resource-aware sensor networks. %B 2011 8th IEEE International Conference on Advanced Video and Signal-Based Surveillance (AVSS) %P 449 - 454 %8 2011 %G eng %0 Conference Paper %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %D 2011 %T Motivation for Participation in Online Neighborhood Watch Communities: An Empirical Study Involving Invitation Letters %A Violi,N. %A Shneiderman, Ben %A Hanson,A. %A Rey,P. J %K Art %K Communities %K community safety %K Electronic mail %K Interviews %K invitation email %K invitations %K motivation %K neighborhood watch %K Online communities %K online neighborhood watch communities %K online survey %K participation %K Safety %K Security %K social media %K Social network services %K social networking (online) %K social networking Website %X This paper presents a three-part experiment designed to investigate the motivations of users of a community safety and neighborhood watch social networking website. The experiment centers around an intervention into the invitation system that current users employ to invite nonmembers to join the site, and involves several versions of an invitation email which differ by expressing one of four possible motivations for using such a site. The research presented investigates how potential users' choice of whether or not to join the site is affected by the use case presented by the invitation. It also includes an investigation of the motivations of current users of the site, as reported in an online survey. The experiment yielded no significant difference in responses to the emails. Overall, invitations that included a specific motivation slightly outperformed those which did not, but not to a statistically significant degree. We conclude that although users have specific motivations for using the site, as reported in the survey, attempting to increase response rates to invitation emails by suggesting use cases of the site is surprisingly unlikely to be successful. %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %I IEEE %P 760 - 765 %8 2011/10/09/11 %@ 978-1-4577-1931-8 %G eng %R 10.1109/PASSAT/SocialCom.2011.108 %0 Conference Paper %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %D 2011 %T NetVisia: Heat Map & Matrix Visualization of Dynamic Social Network Statistics & Content %A Gove,R. %A Gramsky,N. %A Kirby,R. %A Sefer,E. %A Sopan,A. %A Dunne,C. %A Shneiderman, Ben %A Taieb-Maimon,M. %K business intelligence concept %K business intelligence entity %K competitive intelligence %K data visualisation %K dynamic networks %K dynamic social network %K heat map %K Heating %K Image color analysis %K Information Visualization %K Layout %K matrix visualization %K measurement %K NetVisia system %K network evolution %K network visualization %K node-link diagrams %K outlier node %K social network content %K Social network services %K social network statistics %K social networking (online) %K social networks %K static network visualization %K time period %K topological similarity %K Training %K usability %K user evaluation %K User interfaces %X Visualizations of static networks in the form of node-link diagrams have evolved rapidly, though researchers are still grappling with how best to show evolution of nodes over time in these diagrams. This paper introduces NetVisia, a social network visualization system designed to support users in exploring temporal evolution in networks by using heat maps to display node attribute changes over time. NetVisia's novel contributions to network visualizations are to (1) cluster nodes in the heat map by similar metric values instead of by topological similarity, and (2) align nodes in the heat map by events. We compare NetVisia to existing systems and describe a formative user evaluation of a NetVisia prototype with four participants that emphasized the need for tool tips and coordinated views. Despite the presence of some usability issues, in 30-40 minutes the user evaluation participants discovered new insights about the data set which had not been discovered using other systems. We discuss implemented improvements to NetVisia, and analyze a co-occurrence network of 228 business intelligence concepts and entities. This analysis confirms the utility of a clustered heat map to discover outlier nodes and time periods. %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %I IEEE %P 19 - 26 %8 2011/10/09/11 %@ 978-1-4577-1931-8 %G eng %R 10.1109/PASSAT/SocialCom.2011.216 %0 Journal Article %J arXiv:1110.3563 %D 2011 %T Network Clustering Approximation Algorithm Using One Pass Black Box Sampling %A DuBois,Thomas %A Golbeck,Jennifer %A Srinivasan, Aravind %K Computer Science - Social and Information Networks %K Physics - Physics and Society %X Finding a good clustering of vertices in a network, where vertices in the same cluster are more tightly connected than those in different clusters, is a useful, important, and well-studied task. Many clustering algorithms scale well, however they are not designed to operate upon internet-scale networks with billions of nodes or more. We study one of the fastest and most memory efficient algorithms possible - clustering based on the connected components in a random edge-induced subgraph. When defining the cost of a clustering to be its distance from such a random clustering, we show that this surprisingly simple algorithm gives a solution that is within an expected factor of two or three of optimal with either of two natural distance functions. In fact, this approximation guarantee works for any problem where there is a probability distribution on clusterings. We then examine the behavior of this algorithm in the context of social network trust inference. %B arXiv:1110.3563 %8 2011/10/16/ %G eng %U http://arxiv.org/abs/1110.3563 %0 Journal Article %J Current Protocols in Bioinformatics %D 2011 %T Next Generation Sequence Assembly with AMOS %A Treangen, T.J. %A Sommer, D.D. %A Angly, F.E. %A Koren, S. %A Pop, Mihai %B Current Protocols in Bioinformatics %V 11 %P 1 - 11 %8 2011 %G eng %0 Journal Article %J Computational Linguistics %D 2011 %T Nouveau-ROUGE: A Novelty Metric for Update Summarization %A Conroy,John M. %A Schlesinger,Judith D. %A O'Leary, Dianne P. %X An update summary should provide a fluent summarization of new information on a time-evolving topic, assuming that the reader has already reviewed older documents or summaries. In 2007 and 2008, an annual summarization evaluation included an update summarization task. Several participating systems produced update summaries indistinguishable from human-generated summaries when measured using ROUGE. However, no machine system performed near human-level performance in manual evaluations such as pyramid and overall responsiveness scoring. %B Computational Linguistics %V 37 %P 1 - 8 %8 2011/// %@ 0891-2017 %G eng %U http://dx.doi.org/10.1162/coli_a_00033 %N 1 %R 10.1162/coli_a_00033 %0 Conference Paper %B Image Processing (ICIP), 2011 18th IEEE International Conference on %D 2011 %T A novel feature descriptor based on the shearlet transform %A Schwartz, W.R. %A da Silva,R.D. %A Davis, Larry S. %A Pedrini,H. %K analysis;multiscale %K analysis;object %K classification;face %K classification;image %K classification;intensity %K coefficients;image %K descriptor;feature %K detection;object %K distribution %K edge %K EXTRACTION %K extraction;image %K gradient %K gradients;histograms %K identification;feature %K methods;histograms %K of %K orientations;face %K oriented %K recognition;feature %K recognition;shearlet %K recognition;transforms; %K shearlet %K singularities;texture %K texture;object %K transform;signal %X Problems such as image classification, object detection and recognition rely on low-level feature descriptors to represent visual information. Several feature extraction methods have been proposed, including the Histograms of Oriented Gradients (HOG), which captures edge information by analyzing the distribution of intensity gradients and their directions. In addition to directions, the analysis of edge at different scales provides valuable information. Shearlet transforms provide a general framework for analyzing and representing data with anisotropic information at multiple scales. As a consequence, signal singularities, such as edges, can be precisely detected and located in images. Based on the idea of employing histograms to estimate the distribution of edge orientations and on the accurate multi-scale analysis provided by shearlet transforms, we propose a feature descriptor called Histograms of Shearlet Coefficients (HSC). Experimental results comparing HOG with HSC show that HSC provides significantly better results for the problems of texture classification and face identification. %B Image Processing (ICIP), 2011 18th IEEE International Conference on %P 1033 - 1036 %8 2011/09// %G eng %R 10.1109/ICIP.2011.6115600 %0 Journal Article %J SIAM Journal on Matrix Analysis and Applications %D 2011 %T On the Numerical Analysis of Oblique Projectors %A Stewart, G.W. %K $B$-orthogonality %K oblique projector %K Perturbation theory %K updating algorithms %K XQRY form %X An oblique projector is an idempotent matrix whose null space is oblique to its range, in contrast to an orthogonal projector, whose null space is orthogonal to its range. Oblique projectors arise naturally in many applications and have a substantial literature. Missing from that literature, however, are systematic expositions of their numerical properties, including their perturbation theory, their various representations, their behavior in the presence of rounding error, the computation of complementary projections, and updating algorithms. This article is intended to make a start at filling this gap. The first part of the article is devoted to the first four of the above topics, with particular attention given to complementation. In the second part, stable algorithms are derived for updating an XQRY representation of projectors, which was introduced in the first part. %B SIAM Journal on Matrix Analysis and Applications %V 32 %P 309 - 348 %8 2011/// %G eng %U http://link.aip.org/link/?SML/32/309/1 %N 1 %R 10.1137/100792093 %0 Book Section %B Advances in Cryptology – ASIACRYPT 2011 %D 2011 %T Oblivious RAM with O((logN)^3) Worst-Case Cost %A Elaine Shi %A Chan, T. %A Stefanov, Emil %A Li, Mingfei %E Lee, Dong %E Wang, Xiaoyun %K Computer science %X Oblivious RAM is a useful primitive that allows a client to hide its data access patterns from an untrusted server in storage outsourcing applications. Until recently, most prior works on Oblivious RAM aim to optimize its amortized cost, while suffering from linear or even higher worst-case cost. Such poor worst-case behavior renders these schemes impractical in realistic settings, since a data access request can occasionally be blocked waiting for an unreasonably large number of operations to complete. This paper proposes novel Oblivious RAM constructions that achieves poly-logarithmic worst-case cost, while consuming constant client-side storage. To achieve the desired worst-case asymptotic performance, we propose a novel technique in which we organize the O-RAM storage into a binary tree over data buckets, while moving data blocks obliviously along tree edges. %B Advances in Cryptology – ASIACRYPT 2011 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 7073 %P 197 - 214 %8 2011 %@ 978-3-642-25384-3 %G eng %U http://www.springerlink.com/content/2214q8013376rj37/abstract/ %0 Journal Article %J EcoHealth %D 2011 %T Occurrence of Vibrio cholerae in Municipal and Natural Waters and Incidence of Cholera in Azerbaijan %A Gurbanov, Shair %A Akhmadov, Rashid %A Shamkhalova, Gulnara %A Akhmadova, Sevinj %A Haley, Bradd J. %A Rita R Colwell %A Huq, Anwar %X Cholera, a waterborne disease caused by Vibrio cholerae, is an autochthonous member of the aquatic environment and predominantly reported from developing countries. Technical reports and proceedings were reviewed to determine the relationship between occurrence of V. cholerae in natural waters, including sources of municipal water, and cases of cholera in Azerbaijan. Water samples collected from different environmental sources from 1970 to 1998 were tested for V. cholerae and 0.73% (864/117,893) were positive. The results showed that in April of each year, when the air temperature rose by approximately 5°C, V. cholerae could be isolated. With each increase in air temperature, 6–8 weeks after, impact on cases of cholera was recorded. The incidence of cholera peaked when the air temperature reached >25°C during the month of September. It is concluded that a distinct seasonality in cholera incidence exists in Azerbaijan, with increased occurrence during warmer months. %B EcoHealth %P 468 - 477 %8 Jan-12-2011 %G eng %U http://link.springer.com/10.1007/s10393-012-0756-8 %N 4 %! EcoHealth %R 10.1007/s10393-012-0756-8 %0 Conference Paper %B Privacy, security, risk and trust (passat), 2011 ieee third international conference on and 2011 ieee third international conference on social computing (socialcom) %D 2011 %T Odd Leaf Out: Improving Visual Recognition with Games %A Hansen,D. L %A Jacobs, David W. %A Lewis,D. %A Biswas,A. %A Preece,J. %A Rotman,D. %A Stevens,E. %K algorithm;educational %K classification;object %K computational %K computing;botany;computer %K datasets;misclassification %K errors;scientific %K feedback;labeled %K game;human %K games;computer %K games;image %K image %K leaf %K Odd %K Out;complex %K recognition; %K recognition;biology %K tags;visual %K tasks;computer %K tasks;textual %K VISION %X A growing number of projects are solving complex computational and scientific tasks by soliciting human feedback through games. Many games with a purpose focus on generating textual tags for images. In contrast, we introduce a new game, Odd Leaf Out, which provides players with an enjoyable and educational game that serves the purpose of identifying misclassification errors in a large database of labeled leaf images. The game uses a novel mechanism to solicit useful information from players' incorrect answers. A study of 165 players showed that game data can be used to identify mislabeled leaves much more quickly than would have been possible using a computer vision algorithm alone. Domain novices and experts were equally good at identifying mislabeled images, although domain experts enjoyed the game more. We discuss the successes and challenges of this new game, which can be applied to other domains with labeled image datasets. %B Privacy, security, risk and trust (passat), 2011 ieee third international conference on and 2011 ieee third international conference on social computing (socialcom) %P 87 - 94 %8 2011/10// %G eng %R 10.1109/PASSAT/SocialCom.2011.225 %0 Journal Article %J Snowbird Learning Workshop %D 2011 %T Partial least squares based speaker recognition system %A Srinivasan,B.V. %A Zotkin,Dmitry N %A Duraiswami, Ramani %B Snowbird Learning Workshop %8 2011/// %G eng %0 Conference Paper %B Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on %D 2011 %T A partial least squares framework for speaker recognition %A Srinivasan,B.V. %A Zotkin,Dmitry N %A Duraiswami, Ramani %K approximations;speaker %K attribute %K background %K GMM;Gaussian %K least %K mixture %K model;Gaussian %K model;NIST %K modeling %K processes;least %K recognition; %K recognition;speaker %K separability;latent %K squares %K squares;partial-least-squares;speaker %K SRE;interclass %K technique;multiple %K utterances;nuisance %K variability;partial %K variable %K verification;universal %X Modern approaches to speaker recognition (verification) operate in a space of "supervectors" created via concatenation of the mean vectors of a Gaussian mixture model (GMM) adapted from a universal background model (UBM). In this space, a number of approaches to model inter-class separability and nuisance attribute variability have been proposed. We develop a method for modeling the variability associated with each class (speaker) by using partial-least-squares - a latent variable modeling technique, which isolates the most informative subspace for each speaker. The method is tested on NIST SRE 2008 data and provides promising results. The method is shown to be noise-robust and to be able to efficiently learn the subspace corresponding to a speaker on training data consisting of multiple utterances. %B Acoustics, Speech and Signal Processing (ICASSP), 2011 IEEE International Conference on %P 5276 - 5279 %8 2011/05// %G eng %R 10.1109/ICASSP.2011.5947548 %0 Conference Paper %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %D 2011 %T Predicting Trust and Distrust in Social Networks %A DuBois,T. %A Golbeck,J. %A Srinivasan, Aravind %K distrust prediction %K Electronic publishing %K Encyclopedias %K graph theory %K inference algorithm %K Inference algorithms %K inference mechanisms %K Internet %K negative trust %K online social networks %K positive trust %K Prediction algorithms %K probability %K random graphs %K security of data %K social media %K social networking (online) %K spring-embedding algorithm %K Training %K trust inference %K trust probabilistic interpretation %K user behavior %K user satisfaction %K user-generated content %K user-generated interactions %X As user-generated content and interactions have overtaken the web as the default mode of use, questions of whom and what to trust have become increasingly important. Fortunately, online social networks and social media have made it easy for users to indicate whom they trust and whom they do not. However, this does not solve the problem since each user is only likely to know a tiny fraction of other users, we must have methods for inferring trust - and distrust - between users who do not know one another. In this paper, we present a new method for computing both trust and distrust (i.e., positive and negative trust). We do this by combining an inference algorithm that relies on a probabilistic interpretation of trust based on random graphs with a modified spring-embedding algorithm. Our algorithm correctly classifies hidden trust edges as positive or negative with high accuracy. These results are useful in a wide range of social web applications where trust is important to user behavior and satisfaction. %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %I IEEE %P 418 - 424 %8 2011/10/09/11 %@ 978-1-4577-1931-8 %G eng %R 10.1109/PASSAT/SocialCom.2011.56 %0 Conference Paper %D 2011 %T Privacy settings from contextual attributes: A case study using Google Buzz %A Mashima, D. %A Elaine Shi %A Chow, R. %A Sarkar, P. %A Li,C. %A Song,D. %K contextual attributes %K data privacy %K Google Buzz %K privacy settings %K social networking (online) %K social networks %K Statistics %X Social networks provide users with privacy settings to control what information is shared with connections and other users. In this paper, we analyze factors influencing changes in privacy-related settings in the Google Buzz social network. Specifically, we show statistics on contextual data related to privacy settings that are derived from crawled datasets and analyze the characteristics of users who changed their privacy settings. We also investigate potential neighboring effects among such users. %P 257 - 262 %8 2011 %G eng %0 Conference Paper %D 2011 %T Privacy-preserving aggregation of time-series data %A Elaine Shi %A Chan, T. %A Rieffel, E. %A Chow, R. %A Song,D. %X We consider how an untrusted data aggregator canlearn desired statistics over multiple participants’ data, without compromising each individual’s privacy. We propose a construction that allows a group of partici- pants to periodically upload encrypted values to a data aggregator, such that the aggregator is able to compute the sum of all participants’ values in every time period, but is unable to learn anything else. We achieve strong privacy guarantees using two main techniques. First, we show how to utilize applied cryptographic techniques to allow the aggregator to decrypt the sum from multiple ciphertexts encrypted under different user keys. Second, we describe a distributed data randomization procedure that guarantees the differential privacy of the outcome statistic, even when a subset of participants might be compromised. %V 17 %8 2011 %G eng %U http://www.eecs.berkeley.edu/~elaines/docs/ndss2011.pdf %0 Journal Article %J ACM Transactions on Information and System Security (TISSEC) %D 2011 %T Private and Continual Release of Statistics %A Chan, T.-H. Hubert %A Elaine Shi %A Song, Dawn %K continual mechanism %K Differential privacy %K streaming algorithm %X We ask the question: how can Web sites and data aggregators continually release updated statistics, and meanwhile preserve each individual user’s privacy? Suppose we are given a stream of 0’s and 1’s. We propose a differentially private continual counter that outputs at every time step the approximate number of 1’s seen thus far. Our counter construction has error that is only poly-log in the number of time steps. We can extend the basic counter construction to allow Web sites to continually give top-k and hot items suggestions while preserving users’ privacy. %B ACM Transactions on Information and System Security (TISSEC) %V 14 %P 26:1 - 26:24 %8 2011 %@ 1094-9224 %G eng %U http://doi.acm.org/10.1145/2043621.2043626 %N 3 %0 Journal Article %J BMC Bioinformatics %D 2011 %T ProPhylo: partial phylogenetic profiling to guide protein family construction and assignment of biological process. %A Basu, Malay K %A Jeremy D Selengut %A Haft, Daniel H %K algorithms %K Archaea %K Archaeal Proteins %K DNA %K Methane %K Phylogeny %K software %X

BACKGROUND: Phylogenetic profiling is a technique of scoring co-occurrence between a protein family and some other trait, usually another protein family, across a set of taxonomic groups. In spite of several refinements in recent years, the technique still invites significant improvement. To be its most effective, a phylogenetic profiling algorithm must be able to examine co-occurrences among protein families whose boundaries are uncertain within large homologous protein superfamilies.

RESULTS: Partial Phylogenetic Profiling (PPP) is an iterative algorithm that scores a given taxonomic profile against the taxonomic distribution of families for all proteins in a genome. The method works through optimizing the boundary of each protein family, rather than by relying on prebuilt protein families or fixed sequence similarity thresholds. Double Partial Phylogenetic Profiling (DPPP) is a related procedure that begins with a single sequence and searches for optimal granularities for its surrounding protein family in order to generate the best query profiles for PPP. We present ProPhylo, a high-performance software package for phylogenetic profiling studies through creating individually optimized protein family boundaries. ProPhylo provides precomputed databases for immediate use and tools for manipulating the taxonomic profiles used as queries.

CONCLUSION: ProPhylo results show universal markers of methanogenesis, a new DNA phosphorothioation-dependent restriction enzyme, and efficacy in guiding protein family construction. The software and the associated databases are freely available under the open source Perl Artistic License from ftp://ftp.jcvi.org/pub/data/ppp/.

%B BMC Bioinformatics %V 12 %P 434 %8 2011 %G eng %R 10.1186/1471-2105-12-434 %0 Journal Article %J University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2011 %D 2011 %T Rapid understanding of scientific paper collections: integrating statistics, text analysis, and visualization %A Dunne,C. %A Shneiderman, Ben %A Gove,R. %A Klavans,J. %A Dorr, Bonnie J %X Keeping up with rapidly growing research fields, especially when there aremultiple interdisciplinary sources, requires substantial effort for researchers, program managers, or venture capital investors. Current theories and tools are directed at finding a paper or website, not gaining an understanding of the key papers, authors, controversies, and hypotheses. This report presents an effort to integrate statistics, text analytics, and visualization in a multiple coordinated window environment that supports exploration. Our prototype system, Action Science Explorer (ASE), provides an environment for demon- strating principles of coordination and conducting iterative usability tests of them with interested and knowledgeable users. We developed an under- standing of the value of reference management, statistics, citation context extraction, natural language summarization for single and multiple docu- ments, filters to interactively select key papers, and network visualization to see citation patterns and identify clusters. The three-phase usability study guided our revisions to ASE and led us to improve the testing methods. %B University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2011 %8 2011/// %G eng %0 Journal Article %J Communications of the ACM %D 2011 %T Realizing the value of social media requires innovative computing research %A Shneiderman, Ben %A Preece,Jennifer %A Pirolli,Peter %X How social media are expanding traditional research and development topics for computer and information scientists. %B Communications of the ACM %V 54 %P 34 - 37 %8 2011/09// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/1995376.1995389 %N 9 %R 10.1145/1995376.1995389 %0 Conference Paper %D 2011 %T Real-Time Planning for Covering an Initially-Unknown Spatial Environment %A Shivashankar,V. %A Jain, R. %A Kuter,U. %A Nau, Dana S. %X We consider the problem of planning, on the fly, a path whereby a robotic vehicle will cover every point in an ini- tially unknown spatial environment. We describe four strate- gies (Iterated WaveFront, Greedy-Scan, Delayed Greedy- Scan and Closest-First Scan) for generating cost-effective coverage plans in real time for unknown environments. We give theorems showing the correctness of our planning strate- gies. Our experiments demonstrate that some of these strate- gies work significantly better than others, and that the best ones work very well; e.g., in environments having an average of 64,000 locations for the robot to cover, the best strategy re- turned plans with less than 6% redundant coverage, and took only an average of 0.1 milliseconds per action. %8 2011/// %G eng %U http://www.aaai.org/ocs/index.php/FLAIRS/FLAIRS11/paper/download/2566/2992 %0 Journal Article %J AMIA Annual Symposium ProceedingsAMIA Annu Symp Proc %D 2011 %T Reducing Missed Laboratory Results: Defining Temporal Responsibility, Generating User Interfaces for Test Process Tracking, and Retrospective Analyses to Identify Problems %A Tarkan,Sureyya %A Plaisant, Catherine %A Shneiderman, Ben %A Hettinger,A. Zachary %X Researchers have conducted numerous case studies reporting the details on how laboratory test results of patients were missed by the ordering medical providers. Given the importance of timely test results in an outpatient setting, there is limited discussion of electronic versions of test result management tools to help clinicians and medical staff with this complex process. This paper presents three ideas to reduce missed results with a system that facilitates tracking laboratory tests from order to completion as well as during follow-up: (1) define a workflow management model that clarifies responsible agents and associated time frame, (2) generate a user interface for tracking that could eventually be integrated into current electronic health record (EHR) systems, (3) help identify common problems in past orders through retrospective analyses. %B AMIA Annual Symposium ProceedingsAMIA Annu Symp Proc %V 2011 %P 1382 - 1391 %8 2011/// %@ 1942-597X %G eng %0 Conference Paper %B PART 2 ———– Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %D 2011 %T Re-engineering health care with information technology: the role of computer-human interaction %A Butler,Keith %A Payne,Thomas %A Shneiderman, Ben %A Brennan,Patricia %A Zhang,Jiajie %K clinical workflow %K electronic medical records %K health information technology visualization %K healthcare informatics %K participatory design %K usability standards & evaluation %X There is critical, nation-wide need to improve health care and its cost. Health information technology has great promise that is yet to be realized. In this panel four noted experts will discuss key issues that should drive health IT, and the challenges for the CHI community to play a leading role. %B PART 2 ———– Proceedings of the 2011 annual conference extended abstracts on Human factors in computing systems %S CHI EA '11 %I ACM %C New York, NY, USA %P 451 - 454 %8 2011/// %G eng %U http://doi.acm.org/10.1145/1979482.1979490 %R 10.1145/1979482.1979490 %0 Journal Article %J Applied and Environmental Microbiology %D 2011 %T Role of Zooplankton Diversity in Vibrio Cholerae Population Dynamics and in the Incidence of Cholera in the Bangladesh Sundarbans %A De Magny,Guillaume Constantin %A Mozumder,Pronob K. %A Grim,Christopher J. %A Hasan,Nur A. %A Naser,M. Niamul %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X Vibrio cholerae, a bacterium autochthonous to the aquatic environment, is the causative agent of cholera, a severe watery, life-threatening diarrheal disease occurring predominantly in developing countries. V. cholerae, including both serogroups O1 and O139, is found in association with crustacean zooplankton, mainly copepods, and notably in ponds, rivers, and estuarine systems globally. The incidence of cholera and occurrence of pathogenic V. cholerae strains with zooplankton were studied in two areas of Bangladesh: Bakerganj and Mathbaria. Chitinous zooplankton communities of several bodies of water were analyzed in order to understand the interaction of the zooplankton population composition with the population dynamics of pathogenic V. cholerae and incidence of cholera. Two dominant zooplankton groups were found to be consistently associated with detection of V. cholerae and/or occurrence of cholera cases, namely, rotifers and cladocerans, in addition to copepods. Local differences indicate there are subtle ecological factors that can influence interactions between V. cholerae, its plankton hosts, and the incidence of cholera. %B Applied and Environmental Microbiology %V 77 %P 6125 - 6132 %8 09/2011 %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/77/17/6125 %N 17 %R 10.1128/AEM.01472-10 %0 Journal Article %J Computer Vision and Image Understanding %D 2011 %T Silhouette-based gesture and action recognition via modeling trajectories on Riemannian shape manifolds %A Abdelkader,Mohamed F. %A Abd-Almageed, Wael %A Srivastava,Anuj %A Chellapa, Rama %K Action recognition %K Gesture recognition %K Riemannian manifolds %K Shape space %K Silhouette-based approaches %X This paper addresses the problem of recognizing human gestures from videos using models that are built from the Riemannian geometry of shape spaces. We represent a human gesture as a temporal sequence of human poses, each characterized by a contour of the associated human silhouette. The shape of a contour is viewed as a point on the shape space of closed curves and, hence, each gesture is characterized and modeled as a trajectory on this shape space. We propose two approaches for modeling these trajectories. In the first template-based approach, we use dynamic time warping (DTW) to align the different trajectories using elastic geodesic distances on the shape space. The gesture templates are then calculated by averaging the aligned trajectories. In the second approach, we use a graphical model approach similar to an exemplar-based hidden Markov model, where we cluster the gesture shapes on the shape space, and build non-parametric statistical models to capture the variations within each cluster. We model each gesture as a Markov model of transitions between these clusters. To evaluate the proposed approaches, an extensive set of experiments was performed using two different data sets representing gesture and action recognition applications. The proposed approaches not only are successfully able to represent the shape and dynamics of the different classes for recognition, but are also robust against some errors resulting from segmentation and background subtraction. %B Computer Vision and Image Understanding %V 115 %P 439 - 455 %8 2011/03// %@ 1077-3142 %G eng %U http://www.sciencedirect.com/science/article/pii/S1077314210002377 %N 3 %R 16/j.cviu.2010.10.006 %0 Conference Paper %B Proceedings of the 8th ACM conference on Creativity and cognition %D 2011 %T Social discovery framework: building capacity and seeking solutions %A Shneiderman, Ben %K Collaboration %K collective intelligence %K community %K creativity support tools %K crowdsourcing %K search %K social discovery %K User interfaces %X While journalists often portray discovery as the thrilling insight of a brilliant individual, many discoveries require years of work by competing and collaborating teams. Often large amounts of foundational work are necessary and dialogs among participants help clarify goals. The Social Discovery Framework suggests that (1) there are important processes in building capacity and then seeking solutions and (2) those that initiate requests are often as important as those who seek solutions. The implications of the Social Discovery Framework are that improved social tools to build capacity, initiate requests, and support dialog would accelerate the discovery process as much as the more visible tools for individuals seeking solutions. %B Proceedings of the 8th ACM conference on Creativity and cognition %S C&C '11 %I ACM %C New York, NY, USA %P 307 - 308 %8 2011/// %@ 978-1-4503-0820-5 %G eng %U http://doi.acm.org/10.1145/2069618.2069669 %R 10.1145/2069618.2069669 %0 Journal Article %J Information Services and Use %D 2011 %T Social discovery in an information abundant world: Designing to create capacity and seek solutions %A Shneiderman, Ben %X The first generation of search tools provided remarkable capabilities to many researchers for finding specific information, navigating to desired websites or tracking down needed documents. A second generation of tools is helping researchers conduct exploratory search in situations where the goals are less clear, where complete coverage is necessary, and where proof of absence is required (patents, legal precedents, etc.). However, the third generation of tools turns search into social discovery, in which colleagues, informal groups, companies, non-profit organizations, professional societies, and international federations are motivated to find, filter, organize, annotate and summarize voluminous information resources. The Reader-to-Leader framework describes the usability and sociability design features needed to motivate readers, contributors, collaborators and leaders. The Social Discovery framework, proposed in this paper, suggests that effective design enables people to engage in dialogs over weeks and months to create capacity (in the manner described by the Reader-to-Leader framework), and become more effective solution seekers. Much work remains to be done to validate these social discovery frameworks and refine them to fit diverse contexts. %B Information Services and Use %V 31 %P 3 - 13 %8 2011/01/01/ %G eng %U http://dx.doi.org/10.3233/ISU-2011-0628 %N 1 %R 10.3233/ISU-2011-0628 %0 Journal Article %J IEEE Transactions on Circuits and Systems for Video Technology %D 2011 %T Special Issue on Video Analysis on Resource-Limited Systems %A Chellapa, Rama %A Cavallaro, A. %A Wu,Y. %A Shan, C. %A Fu, Y. %A Pulli, K. %K computational complexity %K Image Enhancement %K Special issues and sections %K Video compression %X The 17 papers in this special issue focus on resource-limited systems. %B IEEE Transactions on Circuits and Systems for Video Technology %V 21 %P 1349 - 1352 %8 2011/10// %@ 1051-8215 %G eng %N 10 %R 10.1109/TCSVT.2011.2165795 %0 Book Section %B Social Computing, Behavioral-Cultural Modeling and PredictionSocial Computing, Behavioral-Cultural Modeling and Prediction %D 2011 %T Speeding Up Network Layout and Centrality Measures for Social Computing Goals %A Sharma,Puneet %A Khurana,Udayan %A Shneiderman, Ben %A Scharrenbroich,Max %A Locke,John %E Salerno,John %E Yang,Shanchieh %E Nau, Dana S. %E Chai,Sun-Ki %X This paper presents strategies for speeding up calculation of graph metrics and layout by exploiting the parallel architecture of modern day Graphics Processing Units (GPU), specifically Compute Unified Device Architecture ( CUDA ) by Nvidia . Graph centrality metrics like Eigenvector , Betweenness , Page Rank and layout algorithms like Fruchterman − Rheingold are essential components of Social Network Analysis ( SNA ). With the growth in adoption of SNA in different domains and increasing availability of huge networked datasets for analysis, social network analysts require faster tools that are also scalable. Our results, using NodeXL, show up to 802 times speedup for a Fruchterman-Rheingold graph layout and up to 17,972 times speedup for Eigenvector centrality metric calculations on a 240 core CUDA-capable GPU. %B Social Computing, Behavioral-Cultural Modeling and PredictionSocial Computing, Behavioral-Cultural Modeling and Prediction %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6589 %P 244 - 251 %8 2011/// %@ 978-3-642-19655-3 %G eng %U http://dx.doi.org/10.1007/978-3-642-19656-0_35 %0 Conference Paper %B 2011 IEEE 27th International Conference on Data Engineering Workshops (ICDEW) %D 2011 %T State transfer for clear and efficient runtime updates %A Hayden,C.M. %A Smith,E. K %A Hicks, Michael W. %A Foster, Jeffrey S. %K C++ language %K C++ program %K Cognition %K complex tool support %K dynamic software updating %K efficient runtime update %K in-place DSU approach %K Libraries %K Operating systems %K program modification %K Program processors %K program transfer %K program version %K Runtime %K servers %K software libraries %K Software maintenance %K state transfer update %K Steady-state %K updating library %K VSFTPD %X Dynamic software updating (DSU), the practice of updating software while it executes, is a lively area of research. The DSU approach most prominent in both commercial and research systems is in-place updating, in which patches containing program modifications are loaded into a running process. However, in-place updating suffers from several problems: it requires complex tool support, it may adversely affect the performance of normal execution, it requires challenging reasoning to understand the behavior of an updated program, and it requires extra effort to modify program state to be compatible with an update. This paper presents preliminary work investigating the potential for state transfer updating to address these problems. State transfer updates work by launching a new process running the updated program version and transferring program state from the running process to the updated version. In this paper, we describe the use and implementation of Ekiden, a new state transfer updating library for C/C++ programs. Ekiden seeks to redress the difficulties of in-place updating, and we report on our experience updating VSFTPD using Ekiden. This initial experience suggests that state transfer provides the availability benefits of in-place DSU approaches while addressing many of their shortcomings. %B 2011 IEEE 27th International Conference on Data Engineering Workshops (ICDEW) %I IEEE %P 179 - 184 %8 2011/04/11/16 %@ 978-1-4244-9195-7 %G eng %R 10.1109/ICDEW.2011.5767632 %0 Journal Article %J IEEE Transactions on Pattern Analysis and Machine Intelligence %D 2011 %T Statistical Computations on Grassmann and Stiefel Manifolds for Image and Video-Based Recognition %A Turaga,P. %A Veeraraghavan,A. %A Srivastava, A. %A Chellapa, Rama %K activity based video clustering %K activity recognition %K computational geometry %K Computational modeling %K Data models %K face recognition %K feature representation %K finite dimensional linear subspaces %K geometric properties %K Geometry %K Grassmann Manifolds %K Grassmann. %K HUMANS %K Image and video models %K image recognition %K linear dynamic models %K linear subspace structure %K Manifolds %K maximum likelihood classification %K maximum likelihood estimation %K Object recognition %K Riemannian geometry %K Riemannian metrics %K SHAPE %K statistical computations %K statistical models %K Stiefel %K Stiefel Manifolds %K unsupervised clustering %K video based face recognition %K video based recognition %K video signal processing %X In this paper, we examine image and video-based recognition applications where the underlying models have a special structure-the linear subspace structure. We discuss how commonly used parametric models for videos and image sets can be described using the unified framework of Grassmann and Stiefel manifolds. We first show that the parameters of linear dynamic models are finite-dimensional linear subspaces of appropriate dimensions. Unordered image sets as samples from a finite-dimensional linear subspace naturally fall under this framework. We show that an inference over subspaces can be naturally cast as an inference problem on the Grassmann manifold. To perform recognition using subspace-based models, we need tools from the Riemannian geometry of the Grassmann manifold. This involves a study of the geometric properties of the space, appropriate definitions of Riemannian metrics, and definition of geodesics. Further, we derive statistical modeling of inter and intraclass variations that respect the geometry of the space. We apply techniques such as intrinsic and extrinsic statistics to enable maximum-likelihood classification. We also provide algorithms for unsupervised clustering derived from the geometry of the manifold. Finally, we demonstrate the improved performance of these methods in a wide variety of vision applications such as activity recognition, video-based face recognition, object recognition from image sets, and activity-based video clustering. %B IEEE Transactions on Pattern Analysis and Machine Intelligence %V 33 %P 2273 - 2286 %8 2011/11// %@ 0162-8828 %G eng %N 11 %R 10.1109/TPAMI.2011.52 %0 Journal Article %J Electronic Journal of Biotechnology %D 2011 %T Suppression subtractive hybridization PCR isolation of cDNAs from a Caribbean soft coral %A Lopez, J.V. %A Ledger, A. %A Santiago-Vázquez, L.Z. %A Pop, Mihai %A Sommer, D.D. %A Ranzer, L.K. %A Feldman, R.A. %A Russell, G.K. %B Electronic Journal of Biotechnology %V 14 %P 8 - 9 %8 2011 %G eng %N 1 %0 Journal Article %J Communications Magazine, IEEE %D 2011 %T A survey of virtual LAN usage in campus networks %A Yu,Minlan %A Rexford,J. %A Sun,Xin %A Rao,Sanjay %A Feamster, Nick %K academic department %K campus network %K educational computing %K Educational institutions %K Ethernet scalability %K Local area networks %K network policy support %K university campus %K virtual LAN usage %K VLAN %X VLANs are widely used in today's enterprise networks to improve Ethernet scalability and support network policies. However, manuals and textbooks offer very little information about how VLANs are actually used in practice. Through discussions with network administrators and analysis of configuration data, we describe how three university campuses and one academic department use VLANs to achieve a variety of goals. We argue that VLANs are ill-suited to some of these goals (e.g., VLANs are often used to realize access control policies, but constrain the types of policies that can be expressed). Furthermore, the use of VLANs leads to significant complexity in the configuration of network devices. %B Communications Magazine, IEEE %V 49 %P 98 - 103 %8 2011/07// %@ 0163-6804 %G eng %N 7 %R 10.1109/MCOM.2011.5936161 %0 Conference Paper %B 2011 International Joint Conference on Biometrics (IJCB) %D 2011 %T Synthesis-based recognition of low resolution faces %A Shekhar, S. %A Patel, Vishal M. %A Chellapa, Rama %K Dictionaries %K Face %K face images %K face recognition %K face recognition literature %K face recognition systems %K illumination variations %K image resolution %K low resolution faces %K Organizations %K PROBES %K support vector machines %K synthesis based recognition %X Recognition of low resolution face images is a challenging problem in many practical face recognition systems. Methods have been proposed in the face recognition literature for the problem when the probe is of low resolution, and a high resolution gallery is available for recognition. These methods modify the probe image such that the resultant image provides better discrimination. We formulate the problem differently by leveraging the information available in the high resolution gallery image and propose a generative approach for classifying the probe image. An important feature of our algorithm is that it can handle resolution changes along with illumination variations. The effective- ness of the proposed method is demonstrated using standard datasets and a challenging outdoor face dataset. It is shown that our method is efficient and can perform significantly better than many competitive low resolution face recognition algorithms. %B 2011 International Joint Conference on Biometrics (IJCB) %I IEEE %P 1 - 6 %8 2011/10/11/13 %@ 978-1-4577-1358-3 %G eng %R 10.1109/IJCB.2011.6117545 %0 Patent %D 2011 %T SYSTEM AND METHOD FOR DATA MANAGEMENT IN LARGE DATA NETWORKS %A BROECHELER,M. %A V.S. Subrahmanian %A Pugliese, A. %X A system and method for storing an input data network, in the form of graph is provided. The system includes a master node and a plurality of slave nodes. The master node is operable to receive the data network in the form of a graph, the graph including a plurality of vertices connected by edges; calculate a probability of co-retrieval for each of the plurality of vertices; and assign each of the plurality of vertices to one of the plurality of compute nodes based on the calculated probability of co-retrieval. Another method and system are provided for converting a dataset into a graph based index and storing the index on disk. Respective systems and methods of querying such data networks are also provided. %V WO/2011/032077 %8 2011/03// %G eng %N WO/2011/032077 %0 Journal Article %J University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2011-09 %D 2011 %T A task taxonomy of network evolution analysis %A Ahn,JW %A Plaisant, Catherine %A Shneiderman, Ben %X Visualization is a useful tool for understanding the nature of networks. The recent growth of social media requires morepowerful visualization techniques beyond static network diagrams. One of the most important challenges is the visualization of temporal network evolution. In order to provide strong temporal visualization methods, we need to understand what tasks users accomplish. This study provides a taxonomy of the temporal network visualization tasks. We identify (1) the entities, (2) the properties to be visualized, and (3) the hierarchy of temporal features, which were extracted by surveying existing temporal network visualization systems. By building and examining the task taxonomy, we report which tasks have been covered so far and suggest additions for designing the future visualizations. We also present example visualizations constructed using the task taxonomy for a social networking site in order to validate the quality of the taxonomy. %B University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2011-09 %8 2011/// %G eng %0 Conference Paper %D 2011 %T Teaching cross-platform design and testing methods for embedded systems using DICE %A Bhattacharyya, Shuvra S. %A Plishker,William %A Gupta, Ayush %A Chung-Ching Shen %X DICE (the DSPCAD Integrative Command Line Environment) is a package of utilities that facilitates efficient management of software projects. Key areas of emphasis in DICE are cross-platform operation, support for projects that integrate heterogeneous programming languages, and support for applying and integrating different kinds of design and testing methodologies. The package is being developed at the University of Maryland to facilitate the research and teaching of methods for implementation, testing, evolution, and revision of engineering software. The platform- and language-independent focus of DICE makes it an effective vehicle for teaching high-productivity, high-reliability methods for design and implementation of embedded systems for a variety of courses. In this paper, we provide an overview of features of DICE --- particularly as they relate to testing driven design practices --- that are useful in embedded systems education, and discuss examples and experiences of applying the tool in courses at the University of Maryland aimed at diverse groups of students --- undergraduate programming concepts for engineers, graduate VLSI architectures (aimed at research-oriented students), and graduate FPGA system design (aimed at professional Master's students). %S WESE '11 %I ACM %P 38 - 45 %8 2011 %@ 978-1-4503-1046-8 %G eng %U http://doi.acm.org/10.1145/2077370.2077376 %0 Journal Article %J Human-Computer Interaction. Design and Development Approaches %D 2011 %T Technology-mediated social participation: the next 25 years of HCI challenges %A Shneiderman, Ben %X The dramatic success of social media such as Facebook, Twitter, YouTube, blogs, and traditional discussion groups empowers individuals to become active in local and global communities. Some enthusiasts believe that with modest redesign, these technologies can be harnessed to support national priorities such as healthcare/wellness, disaster response, community safety, energy sustainability, etc. However, accomplishing these ambitious goals will require long-term research to develop validated scientific theories and reliable, secure, and scalable technology strategies. The enduring questions of how to motivate participation, increase social trust, and promote collaboration remain grand challenges even as the technology rapidly evolves. This talk invites researchers across multiple disciplines to participate in redefining our discipline of Human-Computer Interaction (HCI) along more social lines to answer vital research questions while creating inspirational prototypes, conducting innovative evaluations, and developing robust technologies. By placing greater emphasis on social media, the HCI community could constructively influence these historic changes. %B Human-Computer Interaction. Design and Development Approaches %P 3 - 14 %8 2011/// %G eng %0 Book Section %B Social Computing, Behavioral-Cultural Modeling and PredictionSocial Computing, Behavioral-Cultural Modeling and Prediction %D 2011 %T Temporal Visualization of Social Network Dynamics: Prototypes for Nation of Neighbors %A Ahn,Jae-wook %A Taieb-Maimon,Meirav %A Sopan,Awalin %A Plaisant, Catherine %A Shneiderman, Ben %E Salerno,John %E Yang,Shanchieh %E Nau, Dana S. %E Chai,Sun-Ki %X Information visualization is a powerful tool for analyzing the dynamic nature of social communities. Using Nation of Neighbors community network as a testbed, we propose five principles of implementing temporal visualizations for social networks and present two research prototypes: NodeXL and TempoVis. Three different states are defined in order to visualize the temporal changes of social networks. We designed the prototypes to show the benefits of the proposed ideas by letting users interactively explore temporal changes of social networks. %B Social Computing, Behavioral-Cultural Modeling and PredictionSocial Computing, Behavioral-Cultural Modeling and Prediction %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6589 %P 309 - 316 %8 2011/// %@ 978-3-642-19655-3 %G eng %U http://dx.doi.org/10.1007/978-3-642-19656-0_43 %0 Journal Article %J Journal of Signal Processing Systems %D 2011 %T Topological Patterns for Scalable Representation and Analysis of Dataflow Graphs %A Sane, Nimish %A Kee, Hojin %A Seetharaman, Gunasekaran %A Bhattacharyya, Shuvra S. %K Circuits and Systems %K Computer Imaging, Vision, Pattern Recognition and Graphics %K Dataflow graphs %K Electrical Engineering %K High-level languages %K Image Processing and Computer Vision %K model-based design %K pattern recognition %K Signal processing systems %K Signal, Image and Speech Processing %K Topological patterns %X Tools for designing signal processing systems with their semantic foundation in dataflow modeling often use high-level graphical user interfaces (GUIs) or text based languages that allow specifying applications as directed graphs. Such graphical representations serve as an initial reference point for further analysis and optimizations that lead to platform-specific implementations. For large-scale applications, the underlying graphs often consist of smaller substructures that repeat multiple times. To enable more concise representation and direct analysis of such substructures in the context of high level DSP specification languages and design tools, we develop the modeling concept of topological patterns, and propose ways for supporting this concept in a high-level language. We augment the dataflow interchange format (DIF) language—a language for specifying DSP-oriented dataflow graphs—with constructs for supporting topological patterns, and we show how topological patterns can be effective in various aspects of embedded signal processing design flows using specific application examples. %B Journal of Signal Processing Systems %V 65 %P 229 - 244 %8 2011 %@ 1939-8018, 1939-8115 %G eng %U http://link.springer.com/article/10.1007/s11265-011-0610-1 %N 2 %! J Sign Process Syst %0 Conference Paper %B BADGERS '11 Proceedings of the First Workshop on Building Analysis Datasets and Gathering Experience Returns for Security %D 2011 %T Toward a Standard Benchmark for Computer Security Research: The Worldwide Intelligence Network Environment (WINE) %A Tudor Dumitras %A Shou, Darren %X Unlike benchmarks that focus on performance or reliability evaluations, a benchmark for computer security must necessarily include sensitive code and data. Because these artifacts could damage systems or reveal personally identifiable information about the users affected by cyber attacks, publicly disseminating such a benchmark raises several scientific, ethical and legal challenges. We propose the Worldwide Intelligence Network Environment (WINE), a security-benchmarking approach based on rigorous experimental methods. WINE includes representative field data, collected worldwide from 240,000 sensors, for new empirical studies, and it will enable the validation of research on all the phases in the lifecycle of security threats. We tackle the key challenges for security benchmarking by designing a platform for repeatable experimentation on the WINE data sets and by collecting the metadata required for understanding the results. In this paper, we review the unique characteristics of the WINE data, we discuss why rigorous benchmarking will provide fresh insights on the security arms race and we propose a research agenda for this area. %B BADGERS '11 Proceedings of the First Workshop on Building Analysis Datasets and Gathering Experience Returns for Security %S BADGERS '11 %I ACM %P 89 - 96 %8 2011/// %@ 978-1-4503-0768-0 %G eng %U http://doi.acm.org/10.1145/1978672.1978683 %0 Journal Article %J Proceedings 67th Annual Forum of the American Helicopter Society %D 2011 %T Toward improved aeromechanics simulations using recent advancements in scientific computing %A Hu,Q. %A Syal,M. %A Gumerov, Nail A. %A Duraiswami, Ramani %A Leishman,J.G. %B Proceedings 67th Annual Forum of the American Helicopter Society %P 3 - 5 %8 2011/// %G eng %0 Journal Article %J Proceedings of the 14th international ACM Sigsoft symposium on Component based software engineering, CBSE %D 2011 %T Towards incremental component compatibility testing %A Yoon,I. %A Sussman, Alan %A Memon, Atif M. %A Porter, Adam %X Software components are increasingly assembled from othercomponents. Each component may further depend on oth- ers, and each may have multiple active versions. The to- tal number of configurations—combinations of components and their versions—deployed by end users can be very large. Component developers, therefore, spend considerable time and effort doing compatibility testing – determining whether their components can be built correctly for all deployed con- figurations. In previous work we developed Rachet to sup- port large-scale compatibility testing of components. In this paper, we describe and evaluate methods to enable Rachet to perform incremental compatibility testing. We de- scribe algorithms to compute differences in component com- patibilities between current and previous component builds, a formal test adequacy criterion based on covering the differ- ences, and cache-aware configuration sampling and testing methods that attempt to reuse effort from previous testing sessions. We evaluate our approach using the 5-year evo- lution history of a scientific middleware component. Our results show significant performance improvements over Ra- chet’s previous retest-all approach, making the process of compatibility testing practical for evolving components. %B Proceedings of the 14th international ACM Sigsoft symposium on Component based software engineering, CBSE %V 11 %P 119 - 128 %8 2011/// %G eng %0 Journal Article %J arXiv:1106.3652 %D 2011 %T Towards Practical Oblivious RAM %A Stefanov, Emil %A Elaine Shi %A Song, Dawn %K Computer Science - Cryptography and Security %X We take an important step forward in making Oblivious RAM (O-RAM) practical. We propose an O-RAM construction achieving an amortized overhead of 20X-35X (for an O-RAM roughly 1 terabyte in size), about 63 times faster than the best existing scheme. On the theoretic front, we propose a fundamentally novel technique for constructing Oblivious RAMs: specifically, we partition a bigger O-RAM into smaller O-RAMs, and employ a background eviction technique to obliviously evict blocks from the client-side cache into a randomly assigned server-side partition. This novel technique is the key to achieving the gains in practical performance. %B arXiv:1106.3652 %8 2011 %G eng %U http://arxiv.org/abs/1106.3652 %0 Journal Article %J Interactions %D 2011 %T Tragic errors: usability and electronic health records %A Shneiderman, Ben %B Interactions %V 18 %P 60 - 63 %8 2011/11// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/2029976.2029992 %N 6 %R 10.1145/2029976.2029992 %0 Conference Paper %D 2011 %T Trajectory planning with look-ahead for Unmanned Sea Surface Vehicles to handle environmental disturbances %A Svec,P. %A Schwartz,M. %A Thakur,A. %A Gupta,S.K. %P 1154 - 1159 %8 2011/// %G eng %U http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=6048670 %0 Journal Article %J Government Information Quarterly %D 2011 %T TreeCovery: Coordinated dual treemap visualization for exploring the Recovery Act %A Rios-Berrios,Miguel %A Sharma,Puneet %A Lee,Tak Yeon %A Schwartz,Rachel %A Shneiderman, Ben %K Recovery Act %K user interface %K Visual analytic %X The American Recovery and Reinvestment Act dedicated $787 billion to stimulate the U.S. economy and mandated the release of the data describing the exact distribution of that money. The dataset is a large and complex one; one of its distinguishing features is its bi-hierarchical structure, arising from the distribution of money through agencies to specific projects and the natural aggregation of awards based on location. To offer a comprehensive overview of the data, a visualization must incorporate both these hierarchies. We present TreeCovery, a tool that accomplishes this through the use of two coordinated treemaps. The tool includes a number of innovative features, including coordinated zooming and filtering and a proportional highlighting technique across the two trees. TreeCovery was designed to facilitate data exploration, and initial user studies suggest that it will be helpful in insight generation. RATB (Recovery Accountability and Transparency Board) has tested TreeCovery and is considering including the concept in their visual analytics. %B Government Information Quarterly %8 2011/// %@ 0740-624X %G eng %U http://www.sciencedirect.com/science/article/pii/S0740624X11001055 %R 10.1016/j.giq.2011.07.004 %0 Conference Paper %B 2011 IEEE Conference on Visual Analytics Science and Technology (VAST) %D 2011 %T TreeVersity: Comparing tree structures by topology and node's attributes differences %A Gomez,J.A.G. %A Buck-Coleman,A. %A Plaisant, Catherine %A Shneiderman, Ben %K Computer science %K data classification %K Data visualization %K Educational institutions %K hierarchy %K Image color analysis %K LifeFlow %K node attributes differences %K Pattern classification %K structural changes %K Topology %K topology attributes differences %K traffic agencies %K tree structures comparison %K trees (mathematics) %K TreeVersity %K Vegetation %K Visualization %X It is common to classify data in hierarchies, they provide a comprehensible way of understanding big amounts of data. From budgets to organizational charts or even the stock market, trees are everywhere and people find them easy to use. However when analysts need to compare two versions of the same tree structure, or two related taxonomies, the task is not so easy. Much work has been done on this topic, but almost all of it has been restricted to either compare the trees by topology, or by the node attribute values. With this project we are proposing TreeVersity, a framework for comparing tree structures, both by structural changes and by differences in the node attributes. This paper is based on our previous work on comparing traffic agencies using LifeFlow [1, 2] and on a first prototype of TreeVersity. %B 2011 IEEE Conference on Visual Analytics Science and Technology (VAST) %I IEEE %P 275 - 276 %8 2011/10/23/28 %@ 978-1-4673-0015-5 %G eng %R 10.1109/VAST.2011.6102471 %0 Conference Paper %B Visual Analytics Science and Technology (VAST), 2011 IEEE Conference on %D 2011 %T TreeVersity: Comparing tree structures by topology and node's attributes differences %A Gomez,J.A.G. %A Buck-Coleman,A. %A Plaisant, Catherine %A Shneiderman, Ben %K (mathematics); %K agencies;tree %K attributes %K changes;topology %K classification;hierarchy;node %K classification;trees %K comparison;pattern %K differences;structural %K differences;traffic %K LifeFlow;TreeVersity;data %K STRUCTURES %X It is common to classify data in hierarchies, they provide a comprehensible way of understanding big amounts of data. From budgets to organizational charts or even the stock market, trees are everywhere and people find them easy to use. However when analysts need to compare two versions of the same tree structure, or two related taxonomies, the task is not so easy. Much work has been done on this topic, but almost all of it has been restricted to either compare the trees by topology, or by the node attribute values. With this project we are proposing TreeVersity, a framework for comparing tree structures, both by structural changes and by differences in the node attributes. This paper is based on our previous work on comparing traffic agencies using LifeFlow [1, 2] and on a first prototype of TreeVersity. %B Visual Analytics Science and Technology (VAST), 2011 IEEE Conference on %P 275 - 276 %8 2011/10// %G eng %R 10.1109/VAST.2011.6102471 %0 Journal Article %J IEEE Computer Graphics and Applications %D 2011 %T Tutorial-Graph Analytics — Lessons Learned and Challenges Ahead %A Wong,P.C. %A Chen,C. %A Görg,C. %A Shneiderman, Ben %A Stasko,J. %A Thomas,J. %B IEEE Computer Graphics and Applications %V 31 %P 18 - 18 %8 2011/// %G eng %N 5 %0 Journal Article %J Social Network Analysis and Mining %D 2011 %T Understanding actor loyalty to event-based groups in affiliation networks %A Sharara,H. %A Singh,L. %A Getoor, Lise %A Mann,J. %X In this paper, we introduce a method for analyzing the temporal dynamics of affiliation networks. We define affiliation groups which describe temporally related subsets of actors and describe an approach for exploring changing memberships in these affiliation groups over time. To model the dynamic behavior in these networks, we consider the concept of loyalty and introduce a measure that captures an actor’s loyalty to an affiliation group as the degree of ‘commitment’ an actor shows to the group over time. We evaluate our measure using three real world affiliation networks: a publication network, a senate bill cosponsorship network, and a dolphin network. The results show the utility of our measure for analyzing the dynamic behavior of actors and quantifying their loyalty to different time-varying affiliation groups. %B Social Network Analysis and Mining %V 1 %P 115 - 126 %8 2011/// %G eng %N 2 %R 10.1007/s13278-010-0005-5 %0 Report %D 2011 %T Understanding Scientific Literature Networks: An Evaluation of Action Science Explorer %A Gove,R. %A Dunne,C. %A Shneiderman, Ben %A Klavans,J. %A Dorr, Bonnie J %X Action Science Explorer (ASE) is a tool designed to supportusers in rapidly generating readily consumable summaries of academic literature. The authors describe ASE and report on how early formative evaluations led to a mature system evaluation, consisting of an in-depth empirical evaluation with 4 domain expert participants. The user study tasks were of two types: predefined tasks to test system perfor- mance in common scenarios, and user-defined tasks to test the system’s usefulness for custom exploration goals. This paper concludes by describing ASE’s attribute ranking ca- pability which is a novel contribution for exploring scientific literature networks. It makes design recommendations to: give the users control over which documents to explore, easy- to-understand metrics for ranking documents, and overviews of the document set in coordinated views along with details- on-demand of specific papers. %I University of Maryland at College Park %8 2011/// %G eng %0 Journal Article %J The VLDB Journal %D 2011 %T A unified approach to ranking in probabilistic databases %A Li,Jian %A Saha,Barna %A Deshpande, Amol %K Approximation techniques %K Graphical models %K Learning to rank %K Probabilistic databases %K Ranking %X Ranking is a fundamental operation in data analysis and decision support and plays an even more crucial role if the dataset being explored exhibits uncertainty. This has led to much work in understanding how to rank the tuples in a probabilistic dataset in recent years. In this article, we present a unified approach to ranking and top-k query processing in probabilistic databases by viewing it as a multi-criterion optimization problem and by deriving a set of features that capture the key properties of a probabilistic dataset that dictate the ranked result. We contend that a single, specific ranking function may not suffice for probabilistic databases, and we instead propose two parameterized ranking functions, called PRF ¿ and PRF e, that generalize or can approximate many of the previously proposed ranking functions. We present novel generating functions-based algorithms for efficiently ranking large datasets according to these ranking functions, even if the datasets exhibit complex correlations modeled using probabilistic and/xor trees or Markov networks. We further propose that the parameters of the ranking function be learned from user preferences, and we develop an approach to learn those parameters. Finally, we present a comprehensive experimental study that illustrates the effectiveness of our parameterized ranking functions, especially PRF e, at approximating other ranking functions and the scalability of our proposed algorithms for exact or approximate ranking. %B The VLDB Journal %V 20 %P 249 - 275 %8 2011/04// %@ 1066-8888 %G eng %U http://dx.doi.org/10.1007/s00778-011-0220-3 %N 2 %R 10.1007/s00778-011-0220-3 %0 Journal Article %J American Journal of Preventive Medicine %D 2011 %T Usability and Accessibility in Consumer Health Informatics: Current Trends and Future Challenges %A Goldberg,Larry %A Lide,Bettijoyce %A Lowry,Svetlana %A Massett,Holly A. %A O'Connell,Trisha %A Preece,Jennifer %A Quesenbery,Whitney %A Shneiderman, Ben %X It is a truism that, for innovative eHealth systems to have true value and impact, they must first and foremost be usable and accessible by clinicians, consumers, and other stakeholders. In this paper, current trends and future challenges in the usability and accessibility of consumer health informatics will be described. Consumer expectations of their healthcare providers and healthcare records in this new era of consumer-directed care will be explored, and innovative visualizations, assistive technologies, and other ways that healthcare information is currently being provided and/or shared will be described. Challenges for ensuring the usability of current and future systems will also be discussed. An innovative model for conducting systematic, timely, user-centered research on consumer-facing websites at the National Cancer Institute (NCI) and the ongoing efforts at the National Institute of Standards and Technology (NIST) to promote health information technology (HIT) usability standards and evaluation criteria will also be presented. %B American Journal of Preventive Medicine %V 40 %P S187-S197 - S187-S197 %8 2011/05// %@ 0749-3797 %G eng %U http://www.sciencedirect.com/science/article/pii/S0749379711000869 %N 5, Supplement 2 %R 10.1016/j.amepre.2011.01.009 %0 Conference Paper %B 2011 AAAI Fall Symposium Series %D 2011 %T Using Agent-Based Simulation to Determine an Optimal Lane-Changing Strategy on a Multi-Lane Highway %A Tuzo,J. %A Seymour,J. %A desJardins, Marie %A others %B 2011 AAAI Fall Symposium Series %8 2011/// %G eng %0 Report %D 2011 %T Using the DSPCAD Integrative Command-Line Environment: User's Guide for DICE Version 1.1 %A Bhattacharyya, Shuvra S. %A Chung-Ching Shen %A Plishker,William %A Sane, Nimish %A Zaki, George %K Technical Report %X This document provides instructions on setting up, starting up, andbuilding DICE and its key companion packages, dicemin and dicelang. This installation process is based on a general set of conventions, which we refer to as the DICE organizational conventions, for software packages. The DICE organizational conventions are specified in this report. These conventions are applied in DICE, dicemin, and dicelang, and also to other software packages that are developed in the Maryland DSPCAD Research Group. %B Technical Reports from UMIACS %8 2011 %@ UMIACS-TR-2011-13 %G eng %U http://drum.lib.umd.edu//handle/1903/11804 %R Technical Report %0 Conference Paper %B 2011 18th IEEE International Conference on Image Processing (ICIP) %D 2011 %T Variable remapping of images from very different sources %A Wei Zhang %A Yanlin Guo %A Meth, R. %A Sokoloff, H. %A Pope, A. %A Strat, T. %A Chellapa, Rama %K automatic object identification %K Buildings %K CAMERAS %K Conferences %K constrained motion estimation %K coordinates system %K Estimation %K G-RANSAC framework %K image context enlargement %K Image Enhancement %K image registration %K image sequence registration %K Image sequences %K Motion estimation %K Robustness %K temporal integration %K variable image remapping %X We present a system which registers image sequences acquired by very different sources, so that multiple views could be transformed to the same coordinates system. This enables the functionality of automatic object identification and confirmation across views and platforms. The capability of the system comes from three ingredients: 1) image context enlargement through temporal integration; 2) robust motion estimation using the G-RANSAC framework with a relaxed correspondence criteria; 3) constrained motion estimation within the G-RANSAC framework. The proposed system has worked successfully on thousands of frames from multiple collections with significant variations in scale and resolution. %B 2011 18th IEEE International Conference on Image Processing (ICIP) %I IEEE %P 1501 - 1504 %8 2011/09/11/14 %@ 978-1-4577-1304-0 %G eng %R 10.1109/ICIP.2011.6115729 %0 Conference Paper %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %D 2011 %T Visual Analysis of Temporal Trends in Social Networks Using Edge Color Coding and Metric Timelines %A Khurana,U. %A Nguyen,Viet-An %A Hsueh-Chien Cheng %A Ahn,Jae-wook %A Chen,Xi %A Shneiderman, Ben %K Color %K computer scientists %K data analysts %K data visualisation %K Data visualization %K dynamic social network %K dynamic timeslider %K edge color coding %K excel sheet %K Image coding %K image colour analysis %K Layout %K measurement %K metric timelines %K Microsoft excel %K multiple graph metrics %K Net EvViz %K network components %K network layout %K network visualization tool %K NodeXL template %K social networking (online) %K temporal trends %K Twitter %K Visualization %X We present Net EvViz, a visualization tool for analysis and exploration of a dynamic social network. There are plenty of visual social network analysis tools but few provide features for visualization of dynamically changing networks featuring the addition or deletion of nodes or edges. Our tool extends the code base of the Node XL template for Microsoft Excel, a popular network visualization tool. The key features of this work are (1) The ability of the user to specify and edit temporal annotations to the network components in an Excel sheet, (2) See the dynamics of the network with multiple graph metrics plotted over the time span of the graph, called the Timeline, and (3) Temporal exploration of the network layout using an edge coloring scheme and a dynamic Time slider. The objectives of the new features presented in this paper are to let the data analysts, computer scientists and others to observe the dynamics or evolution in a network interactively. We presented Net EvViz to five users of Node XL and received positive responses. %B Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom) %I IEEE %P 549 - 554 %8 2011/10/09/11 %@ 978-1-4577-1931-8 %G eng %R 10.1109/PASSAT/SocialCom.2011.212 %0 Journal Article %J IEEE/ACM Transactions on Computational Biology and Bioinformatics %D 2011 %T Visual Exploration across Biomedical Databases %A Lieberman,M.D. %A Taheri, S. %A Guo,Huimin %A Mirrashed,F. %A Yahav,I. %A Aris,A. %A Shneiderman, Ben %K Bioinformatics %K Biomedical computing %K biomedical databases %K cross-database exploration %K Data exploration and discovery %K Data visualization %K database management systems %K Databases, Factual %K DNA %K graph theory %K Information Storage and Retrieval %K information visualization. %K Keyword search %K medical computing %K natural language processing %K Proteins %K semantic networks %K semantics %K sequences %K text mining %K User-Computer Interface %K user-defined semantics %K visual databases %X Though biomedical research often draws on knowledge from a wide variety of fields, few visualization methods for biomedical data incorporate meaningful cross-database exploration. A new approach is offered for visualizing and exploring a query-based subset of multiple heterogeneous biomedical databases. Databases are modeled as an entity-relation graph containing nodes (database records) and links (relationships between records). Users specify a keyword search string to retrieve an initial set of nodes, and then explore intra- and interdatabase links. Results are visualized with user-defined semantic substrates to take advantage of the rich set of attributes usually present in biomedical data. Comments from domain experts indicate that this visualization method is potentially advantageous for biomedical knowledge exploration. %B IEEE/ACM Transactions on Computational Biology and Bioinformatics %V 8 %P 536 - 550 %8 2011/04//March %@ 1545-5963 %G eng %N 2 %R 10.1109/TCBB.2010.1 %0 Conference Paper %B Proceedings of the 13th annual conference companion on Genetic and evolutionary computation %D 2011 %T When does simulated data match real data? %A Stonedahl,Forrest %A Anderson,David %A Rand, William %K Agent-based modeling %K business %K Calibration %K Genetic algorithms %K information search %K network analysis %X Agent-based models can replicate real-world patterns, but finding parameters that achieve the best match can be difficult. To validate a model, a real-world dataset is often divided into a training set (to calibrate the parameters) and a test set (to validate the calibrated model). The difference between the training and test data and the simulated data is determined using an error measure. In the context of evolutionary computation techniques, the error measure also serves as a fitness function, and thus affects evolutionary search dynamics. We survey the effect of five different error measures on both a toy problem and a real world problem of matching a model to empirical online news consumption behavior. We use each error measure separately for calibration on the training dataset, and then examine the results of all five error measures on both the training and testing datasets. We show that certain error measures sometimes serve as better fitness functions than others, and in fact using one error measure may result in better calibration (on a different measure) than using the different measure directly. For the toy problem, the Pearson's correlation measure dominated all other measures, but no single error measure was Pareto dominant for the real world problem. %B Proceedings of the 13th annual conference companion on Genetic and evolutionary computation %S GECCO '11 %I ACM %C New York, NY, USA %P 231 - 232 %8 2011/// %@ 978-1-4503-0690-4 %G eng %U http://doi.acm.org/10.1145/2001858.2001988 %R 10.1145/2001858.2001988 %0 Conference Paper %B SIGCHI '11 %D 2011 %T Why is My Internet Slow?: Making Network Speeds Visible %A Marshini Chetty %A Haslem, David %A Baird, Andrew %A Ofoha, Ugochi %A Sumner, Bethany %A Grinter, Rebecca %K broadband speed %K broadband tools %K home networks %X With widespread broadband adoption, more households report experiencing sub-optimal speeds. Not only are slow speeds frustrating, they may indicate consumers are not receiving the services they are paying for from their internet service providers. Yet, determining the speed and source of slow-downs is difficult because few tools exist for broadband management. We report on results of a field trial with 10 households using a visual network probe designed to address these problems. We describe the results of the study and provide design implications for future tools. More importantly, we argue that tools like this can educate and empower consumers by making broadband speeds and sources of slow-downs more visible. %B SIGCHI '11 %S CHI '11 %I ACM %P 1889 - 1898 %8 2011/// %@ 978-1-4503-0228-9 %G eng %U http://doi.acm.org/10.1145/1978942.1979217 %0 Journal Article %J AI Magazine %D 2010 %T AI Theory and Practice: A Discussion on Hard Challenges and Opportunities Ahead %A Horvitz,Eric %A Getoor, Lise %A Guestrin,Carlos %A Hendler,James %A Konstan,Joseph %A Subramanian,Devika %A Wellman,Michael %A Kautz,Henry %X AI Theory and Practice: A Discussion on Hard Challenges and Opportunities Ahead %B AI Magazine %V 31 %P 103 - 114 %8 2010/07/28/ %@ 0738-4602 %G eng %U http://www.aaai.org/ojs/index.php/aimagazine/article/viewArticle/2293 %N 3 %R 10.1609/aimag.v31i3.2293 %0 Book %D 2010 %T Analyzing Social Media Networks with NodeXL %A Hansen,Derek %A Shneiderman, Ben %A Smith,Marc A. %K Business & Economics / Marketing / Research %K Computers / Computer Science %K Computers / Data Processing %K Computers / Database Management / Data Mining %K Computers / General %K Computers / Information Theory %K Computers / Interactive & Multimedia %K Computers / Social Aspects / General %K Computers / Social Aspects / Human-Computer Interaction %K Computers / User Interfaces %K data mining %K Data mining - Computer programs %K Data mining/ Computer programs %K Information Visualization %K Information visualization - Computer programs %K Information visualization/ Computer programs %K NodeXL %K online social networks %K Social Science / General %K Social Science / Reference %K Social Science / Sociology / General %K Technology & Engineering / General %K Technology & Engineering / Social Aspects %K Webometrics %K Webometrics - Computer programs %K Webometrics/ Computer programs %X "Analyzing Social Media Networks with NodeXL provides a much needed resource for the social media research community, as it describes network theory, provides compelling examples using data sources like Twitter and Flickr, and highlights how to use a free sophisticated tool for analysis. This is the perfect book for anyone trying to analyze the behavior of online social networks and beyond." ---Adam Perer, Research Scientist, IBM Research "This book provides a basic introduction to social network analysis, followed by practical instruction and examples on gathering data from online sources, importing into Excel, and then analyzing the data through Excel. The book will be important for promoting research in the area for those in information science, sociology, cultural studies, virtual community, and e-commerce."---Caroline Haythornthwaite, PhD, Professor, Graduate School of Library and Information Science, University of Illinois at Urbana-Champaign Businesses, entrepreneurs, individuals, and government agencies alike are looking to social network analysis (SNA) tools for insight into trends, connections, and fluctuations in social media. Microsoft's NodeXL is a free, open-source SNA plug-in for use with Excel. It provides instant graphical representation of relationships of complex networked data. But it goes further than other SNA tools - NodeXL was developed by a multidisciplinary team of experts that bring together information studies, computer science, sociology, human-computer interaction, and over 20 years of visual analytic theory and information visualization into a simple tool anyone can use. This makes NodeXL of interest not only to end-users but also to researchers and students studying visual and network analytics and their application in the real world. %I Morgan Kaufmann %8 2010/08/27/ %@ 9780123822291 %G eng %0 Book %D 2010 %T Analyzing Social Media Networks with NodeXL: Insights from a Connected World %A Hansen,Derek %A Shneiderman, Ben %A Smith,Marc A. %K Computers / Computer Science %K Computers / General %K Computers / Social Aspects / General %K Computers / Social Aspects / Human-Computer Interaction %K Computers / User Interfaces %X Businesses, entrepreneurs, individuals, and government agencies alike are looking to social network analysis (SNA) tools for insight into trends, connections, and fluctuations in social media. Microsoft’s NodeXL is a free, open-source SNA plug-in for use with Excel. It provides instant graphical representation of relationships of complex networked data. But it goes further than other SNA tools -- NodeXL was developed by a multidisciplinary team of experts that bring together information studies, computer science, sociology, human-computer interaction, and over 20 years of visual analytic theory and information visualization into a simple tool anyone can use. This makes NodeXL of interest not only to end-users but also to researchers and students studying visual and network analytics and their application in the real world. In Analyzing Social Media Networks with NodeXL, members of the NodeXL development team up to provide readers with a thorough and practical guide for using the tool while also explaining the development behind each feature. Blending the theoretical with the practical, this book applies specific SNA instructions directly to NodeXL, but the theory behind the implementation can be applied to any SNA. To learn more about Analyzing Social Media Networks and NodeXL, visit the companion site at www.mkp.com/nodexl*Walks you through NodeXL, while explaining the theory and development behind each step, providing takeaways that can apply to any SNA *Demonstrates how visual analytics research can be applied to SNA tools for the mass market *Includes case studies from researchers who use NodeXL on popular networks like email, Facebook, Twitter, and wikis %I Morgan Kaufmann %P 302 %8 2010 %@ 9780123822307 %G eng %0 Journal Article %J ACM Trans. Comput. Logic %D 2010 %T Annotated RDF %A Udrea,Octavian %A Recupero,Diego Reforgiato %A V.S. Subrahmanian %K annotated RDF %K Query processing %K view maintenance %X Real-world use of RDF requires the ability to transparently represent and explain metadata associated with RDF triples. For example, when RDF triples are extracted automatically by information extraction programs, there is a need to represent where the triples came from, what their temporal validity is, and how certain we are that the triple is correct. Today, there is no theoretically clean and practically scalable mechanism that spans these different needs - reification is the only solution propose to date, and its implementations have been ugly. In this paper, we present Annotated RDF (or aRDF for short) in which RDF triples are annotated by members of a partially ordered set (with bottom element) that can be selected in any way desired by the user. We present a formal declarative semantics (model theory) for annotated RDF and develop algorithms to check consistency of aRDF theories and to answer queries to aRDF theories. We show that annotated RDF supports users who need to think about the uncertainty, temporal aspects, and provenance of the RDF triples in an RDF database. We develop a prototype aRDF implementation and show that our algorithms work efficiently even on real world data sets containing over 10 million triples. %B ACM Trans. Comput. Logic %V 11 %P 10:1–10:41 - 10:1–10:41 %8 2010/01// %@ 1529-3785 %G eng %U http://doi.acm.org/10.1145/1656242.1656245 %N 2 %R 10.1145/1656242.1656245 %0 Journal Article %J Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on %D 2010 %T Applications of a Simple Characterization of Human Gait in Surveillance %A Yang Ran %A Qinfen Zheng %A Chellapa, Rama %A Strat, T.M. %K algorithms %K Artificial intelligence %K Automated;Photography;Reproducibility of Results;Sensitivity and Specificity;Video Recording; %K Biometry %K Computer-Assisted;Pattern Recognition %K double helical signatures %K Gait %K gait analysis %K human gait %K human motion kinematics %K HUMANS %K Image Enhancement %K Image Interpretation %K Image motion analysis %K iterative local curve embedding algorithm %K Object detection %K simple spatiotemporal characterization %K video sequence %K Video surveillance %X Applications of a simple spatiotemporal characterization of human gait in the surveillance domain are presented. The approach is based on decomposing a video sequence into x-t slices, which generate periodic patterns referred to as double helical signatures (DHSs). The features of DHS are given as follows: 1) they naturally encode the appearance and kinematics of human motion and reveal geometric symmetries and 2) they are effective and efficient for recovering gait parameters and detecting simple events. We present an iterative local curve embedding algorithm to extract the DHS from video sequences. Two applications are then considered. First, the DHS is used for simultaneous segmentation and labeling of body parts in cluttered scenes. Experimental results showed that the algorithm is robust to size, viewing angles, camera motion, and severe occlusion. Then, the DHS is used to classify load-carrying conditions. By examining various symmetries in DHS, activities such as carrying, holding, and walking with objects that are attached to legs are detected. Our approach possesses several advantages: a compact representation that can be computed in real time is used; furthermore, it does not depend on silhouettes or landmark tracking, which are sensitive to errors in background subtraction stage. %B Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on %V 40 %P 1009 - 1020 %8 2010/08// %@ 1083-4419 %G eng %N 4 %R 10.1109/TSMCB.2010.2044173 %0 Journal Article %J SIAM Journal on Computing %D 2010 %T Approximation Algorithms for Non-Uniform Buy-at-Bulk Network Design %A Chekuri,C. %A Hajiaghayi, Mohammad T. %A Kortsarz,G. %A Salavatipour,M. R %B SIAM Journal on Computing %V 39 %P 1772 - 1798 %8 2010/// %G eng %N 5 %0 Conference Paper %B Proceedings of the 2010 ACM-IEEE International Symposium on Empirical Software Engineering and Measurement %D 2010 %T Are developers complying with the process: an XP study %A Zazworka, Nico %A Stapel,Kai %A Knauss,Eric %A Shull, Forrest %A Basili, Victor R. %A Schneider,Kurt %K process conformance %K process improvement %K XP programming %X Adapting new software processes and practices in organizational and academic environments requires training the developers and validating the applicability of the newly introduced activities. Investigating process conformance during training and understanding if programmers are able and willing to follow the specific steps are crucial to evaluating whether the process improves various software product quality factors. In this paper we present a process model independent approach to detect process non-conformance. Our approach is based on non-intrusively collected data captured by a version control system and provides the project manager with timely updates. Further, we provide evidence of the applicability of our approach by investigating process conformance in a five day training class on eXtreme Programming (XP) practices at the Leibniz Universität Hannover. Our results show that the approach enabled researchers to formulate minimal intrusive methods to check for conformance and that for the majority of the investigated XP practices violations could be detected. %B Proceedings of the 2010 ACM-IEEE International Symposium on Empirical Software Engineering and Measurement %S ESEM '10 %I ACM %C New York, NY, USA %P 14:1–14:10 - 14:1–14:10 %8 2010/// %@ 978-1-4503-0039-1 %G eng %U http://doi.acm.org/10.1145/1852786.1852805 %R 10.1145/1852786.1852805 %0 Journal Article %J BMC Bioinformatics %D 2010 %T Assembly complexity of prokaryotic genomes using short reads %A Kingsford, Carl %A Schatz,Michael C %A Pop, Mihai %X De Bruijn graphs are a theoretical framework underlying several modern genome assembly programs, especially those that deal with very short reads. We describe an application of de Bruijn graphs to analyze the global repeat structure of prokaryotic genomes. %B BMC Bioinformatics %V 11 %P 21 - 21 %8 2010/01/12/ %@ 1471-2105 %G eng %U http://www.biomedcentral.com/1471-2105/11/21 %N 1 %R 10.1186/1471-2105-11-21 %0 Conference Paper %B Proceedings of the 19th ACM international conference on Information and knowledge management %D 2010 %T Assessor error in stratified evaluation %A Webber,W. %A Oard, Douglas %A Scholer,F. %A Hedin,B. %B Proceedings of the 19th ACM international conference on Information and knowledge management %P 539 - 548 %8 2010/// %G eng %0 Conference Paper %D 2010 %T Authentication in the clouds: a framework and its application to mobile users %A Chow, Richard %A Jakobsson, Markus %A Masuoka, Ryusuke %A Molina,Jesus %A Niu, Yuan %A Elaine Shi %A Song,Zhexuan %K Authentication %K Cloud computing %K Mobile computing %X Cloud computing is a natural fit for mobile security. Typical handsets have input constraints and practical computational and power limitations, which must be respected by mobile security technologies in order to be effective. We describe how cloud computing can address these issues. Our approach is based on a flexible framework for supporting authentication decisions we call TrustCube (to manage the authentication infrastructure) and on a behavioral authentication approach referred to as implicit authentication (to translate user behavior into authentication scores). The combination results in a new authentication paradigm for users of mobile technologies, one where an appropriate balance between usability and trust can be managed through flexible policies and dynamic tuning. %S CCSW '10 %I ACM %P 1 - 6 %8 2010 %@ 978-1-4503-0089-6 %G eng %U http://doi.acm.org/10.1145/1866835.1866837 %0 Report %D 2010 %T Autonomous Traffic Engineering using Self-Configuring Link Weights %A Sundaresan,S. %A Lumezanu,C. %A Feamster, Nick %A François,P. %X Network operators use traffic engineering to control the flow of traffic across their networks. Existing TE methods establish static topologies offline, either by setting link weights or by configuring paths a priori. These methods require manual configuration and may not be robust in the face of failures. Some methods also require knowledge about traffic demands and may not be able to handle traffic fluctuations. Even when changes in demand are expected, operators must manually tune network configurations to prepare for them. Because adjusting configurations is difficult to get right, we start from an extreme design point, asking instead whether it is possible to perform traffic engineering online without having to perform any a priori configuration. Our traffic engineering technique, SculpTE, adapts to changing traffic demands by automatically configuring link weights in a stable manner. SculpTE balances load across the network by continually adjusting link weights to expose lightly-loaded paths. We evaluate SculpTE using a simple analytical model and simulations on realistic ISP network topologies. Our results show that SculpTE achieves excellent load balancing, responsiveness, and stability compared to state-of-the-art TE schemes, without requiring network operators to perform any offline configuration. %I School of Computer Science, Georgia Tech %V GT-CS-10-16 %8 2010/// %G eng %U http://hdl.handle.net/1853/35011 %0 Journal Article %J SIGCOMM Comput. Commun. Rev. %D 2010 %T Autonomous traffic engineering with self-configuring topologies %A Sundaresan,Srikanth %A Lumezanu,Cristian %A Feamster, Nick %A Francois,Pierre %K multi-path routing %K online %K sculpte %K self-configuring %K traffic engineering %X Network operators use traffic engineering (TE) to control the flow of traffic across their networks. Existing TE methods require manual configuration of link weights or tunnels, which is difficult to get right, or prior knowledge of traffic demands and hence may not be robust to link failures or traffic fluctuations. We present a self-configuring TE scheme, SculpTE, which automatically adapts the network-layer topology to changing traffic demands. SculpTE is responsive, stable, and achieves excellent load balancing. %B SIGCOMM Comput. Commun. Rev. %V 41 %P – - – %8 2010/08// %@ 0146-4833 %G eng %U http://dl.acm.org/citation.cfm?id=2043164.1851239 %N 4 %0 Journal Article %J Distributed Computing %D 2010 %T Brief announcement: decentralized network bandwidth prediction %A Song,S. %A Keleher,P. %A Bhattacharjee, Bobby %A Sussman, Alan %X Distributed applications can often benefit from knowledge of the bandwidth between hosts, without performing measurements between all host pairs. For example, if a peer-to-peer (P2P) computational grid system predicts pairwise bandwidth between all nodes in the system, that information could increase overall system performance by finding high-bandwidth nodes to store large scientific input or output datasets. Another possible beneficiary is a P2P online game, which can provide users a seamless gaming experience by selecting a coordinator node that has high-bandwidth connections to the players in a game region. %B Distributed Computing %P 198 - 200 %8 2010/// %G eng %0 Conference Paper %D 2010 %T Building an IT Taxonomy with Co-occurrence Analysis, Hierarchical Clustering, and Multidimensional Scaling %A Tsui,C. %A Wang,P. %A Fleischmann,K.R. %A Sayeed,A. B. %A Weinberg, Amy %X Different information technologies (ITs) are related in complex ways. How can the relationships among a large number of ITs be described and analyzed in a representative, dynamic, and scalable way? In this study, we employed co-occurrence analysis to explore the relationships among 50 information technologies discussed in six magazines over ten years (1998-2007). Using hierarchical clustering and multidimensional scaling, we have found that the similarities of the technologies can be depicted in hierarchies and two-dimensional plots, and that similar technologies can be classified into meaningful categories. The results imply reasonable validity of our approach for understanding technology relationships and building an IT taxonomy. The methodology that we offer not only helps IT practitioners and researchers make sense of numerous technologies in the iField but also bridges two related but thus far largely separate research streams in iSchools - information management and IT management. %8 2010/// %G eng %U https://www.ideals.illinois.edu/handle/2142/14918 %0 Conference Paper %B Proceedings of the ACL 2010 System Demonstrations %D 2010 %T cdec: A decoder, alignment, and learning framework for finite-state and context-free translation models %A Dyer,C. %A Weese,J. %A Setiawan,H. %A Lopez,A. %A Ture,F. %A Eidelman,V. %A Ganitkevitch,J. %A Blunsom,P. %A Resnik, Philip %B Proceedings of the ACL 2010 System Demonstrations %P 7 - 12 %8 2010/// %G eng %0 Conference Paper %B Proceedings of the 5th ACM workshop on Challenged networks %D 2010 %T Cellular traffic offloading through opportunistic communications: a case study %A Han,Bo %A Hui,Pan %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Guanhong Pei %A Srinivasan, Aravind %K cellular traffic offloading %K mobile social networks %K opportunistic communications %K target-set selection %X Due to the increasing popularity of various applications for smartphones, 3G networks are currently overloaded by mobile data traffic. Offloading cellular traffic through opportunistic communications is a promising solution to partially solve this problem, because there is no monetary cost for it. As a case study, we investigate the target-set selection problem for information delivery in the emerging Mobile Social Networks (MoSoNets). We propose to exploit opportunistic communications to facilitate the information dissemination and thus reduce the amount of cellular traffic. In particular, we study how to select the target set with only k users, such that we can minimize the cellular data traffic. In this scenario, initially the content service providers deliver information over cellular networks to only users in the target set. Then through opportunistic communications, target-users will further propagate the information among all the subscribed users. Finally, service providers will send the information to users who fail to receive it before the delivery deadline (i.e., delay-tolerance threshold). We propose three algorithms, called Greedy, Heuristic, and Random, for this problem and evaluate their performance through an extensive trace-driven simulation study. The simulation results verify the efficiency of these algorithms for both synthetic and real-world mobility traces. For example, the Heuristic algorithm can offload cellular traffic by up to 73.66% for a real-world mobility trace. %B Proceedings of the 5th ACM workshop on Challenged networks %S CHANTS '10 %I ACM %C New York, NY, USA %P 31 - 38 %8 2010/// %@ 978-1-4503-0139-8 %G eng %U http://doi.acm.org/10.1145/1859934.1859943 %R 10.1145/1859934.1859943 %0 Conference Paper %B Proceedings of the 19th ACM international conference on Information and knowledge management %D 2010 %T Challenges in personalized authority flow based ranking of social media %A Sayyadi,H. %A Edmonds,J. %A Hristidis,V. %A Raschid, Louiqa %B Proceedings of the 19th ACM international conference on Information and knowledge management %P 1409 - 1412 %8 2010/// %G eng %0 Journal Article %J Nature %D 2010 %T Communication: Mapping science %A Shneiderman, Ben %K Communication %X Ben Shneiderman enjoys a tome full of tools for discovery. %B Nature %V 468 %P 1037 - 1037 %8 2010/12/22/ %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v468/n7327/full/4681037a.html %N 7327 %R 10.1038/4681037a %0 Conference Paper %B Proceedings of the FSE/SDP workshop on Future of software engineering research %D 2010 %T Community-based, collaborative testing and analysis %A Memon, Atif M. %A Porter, Adam %A Sussman, Alan %K component-based software development communities %K testing and analysis %X This article proposes a research agenda aimed at enabling optimized testing and analysis processes and tools to support component-based software development communities. We hypothesize that de facto communities---sets of projects that provide, maintain and integrate many shared infrastructure components---are commonplace. Currently, community members, often unknown to each other, tend to work in isolation, duplicating work, failing to learn from each other's effort, and missing opportunities to efficiently improve the common infrastructure. We further hypothesize that as software integration continues to become the predominant mode of software development, there will be increasing value in tools and techniques that empower these communities to coordinate and optimize their development efforts, and to generate and broadly share information. Such tools and techniques will greatly improve the robustness, quality and usability of the common infrastructure which, in turn, will greatly reduce the time and effort needed to produce and use the end systems that are the true goal of the entire community. %B Proceedings of the FSE/SDP workshop on Future of software engineering research %S FoSER '10 %I ACM %C New York, NY, USA %P 239 - 244 %8 2010/// %@ 978-1-4503-0427-6 %G eng %U http://doi.acm.org/10.1145/1882362.1882412 %R 10.1145/1882362.1882412 %0 Journal Article %J BMC Microbiology %D 2010 %T Comparative genomic analysis reveals evidence of two novel Vibrio species closely related to V. cholerae %A Bradd,H. %A Christopher,G. %A Nur,H. %A Seon-Young,C. %A Jongsik,C. %A Thomas,B. %A David,B. %A Jean,C. %A Chris,D. J. %A Cliff,H. %A Rita R Colwell %X In recent years genome sequencing has been used to characterize new bacterial species, a method of analysis available as a result of improved methodology and reduced cost. Included in a constantly expanding list of Vibrio species are several that have been reclassified as novel members of the Vibrionaceae. The description of two putative new Vibrio species, Vibrio sp. RC341 and Vibrio sp. RC586 for which we propose the names V. metecus and V. parilis, respectively, previously characterized as non-toxigenic environmental variants of V. cholerae is presented in this study. Results Based on results of whole-genome average nucleotide identity (ANI), average amino acid identity (AAI), rpoB similarity, MLSA, and phylogenetic analysis, the new species are concluded to be phylogenetically closely related to V. cholerae and V. mimicus. Vibrio sp. RC341 and Vibrio sp. RC586 demonstrate features characteristic of V. cholerae and V. mimicus, respectively, on differential and selective media, but their genomes show a 12 to 15% divergence (88 to 85% ANI and 92 to 91% AAI) compared to the sequences of V. cholerae and V. mimicus genomes (ANI <95% and AAI <96% indicative of separate species). Vibrio sp. RC341 and Vibrio sp. RC586 share 2104 ORFs (59%) and 2058 ORFs (56%) with the published core genome of V. cholerae and 2956 (82%) and 3048 ORFs (84%) with V. mimicus MB-451, respectively. The novel species share 2926 ORFs with each other (81% Vibrio sp. RC341 and 81% Vibrio sp. RC586). Virulence-associated factors and genomic islands of V. cholerae and V. mimicus, including VSP-I and II, were found in these environmental Vibrio spp. Conclusions Results of this analysis demonstrate these two environmental vibrios, previously characterized as variant V. cholerae strains, are new species which have evolved from ancestral lineages of the V. cholerae and V. mimicus clade. The presence of conserved integration loci for genomic islands as well as evidence of horizontal gene transfer between these two new species, V. cholerae, and V. mimicus suggests genomic islands and virulence factors are transferred between these species. %B BMC Microbiology %V 10 %8 2010/// %G eng %0 Book Section %B Computer Vision – ECCV 2010Computer Vision – ECCV 2010 %D 2010 %T Compressive Acquisition of Dynamic Scenes %A Sankaranarayanan,Aswin %A Turaga,Pavan %A Baraniuk,Richard %A Chellapa, Rama %E Daniilidis,Kostas %E Maragos,Petros %E Paragios,Nikos %X Compressive sensing (CS) is a new approach for the acquisition and recovery of sparse signals and images that enables sampling rates significantly below the classical Nyquist rate. Despite significant progress in the theory and methods of CS, little headway has been made in compressive video acquisition and recovery. Video CS is complicated by the ephemeral nature of dynamic events, which makes direct extensions of standard CS imaging architectures and signal models infeasible. In this paper, we develop a new framework for video CS for dynamic textured scenes that models the evolution of the scene as a linear dynamical system (LDS). This reduces the video recovery problem to first estimating the model parameters of the LDS from compressive measurements, from which the image frames are then reconstructed. We exploit the low-dimensional dynamic parameters (the state sequence) and high-dimensional static parameters (the observation matrix) of the LDS to devise a novel compressive measurement strategy that measures only the dynamic part of the scene at each instant and accumulates measurements over time to estimate the static parameters. This enables us to considerably lower the compressive measurement rate considerably. We validate our approach with a range of experiments including classification experiments that highlight the effectiveness of the proposed approach. %B Computer Vision – ECCV 2010Computer Vision – ECCV 2010 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6311 %P 129 - 142 %8 2010/// %@ 978-3-642-15548-2 %G eng %U http://dx.doi.org/10.1007/978-3-642-15549-9_10 %0 Journal Article %J SIGACT News %D 2010 %T Concentration of measure for the analysis of randomized algorithms by Devdatt P. Dubhashi and Alessandro Panconesi Cambridge University Press, 2009 %A Srinivasan, Aravind %B SIGACT News %V 41 %P 28 - 30 %8 2010/03// %@ 0163-5700 %G eng %U http://doi.acm.org/10.1145/1753171.1753179 %N 1 %R 10.1145/1753171.1753179 %0 Journal Article %J Behaviour & Information Technology %D 2010 %T Connecting generations: developing co-design methods for older adults and children %A Xie,B. %A Druin, Allison %A Fails,J. %A Massey,S. %A Golub,E. %A Franckel,S. %A Schneider,K. %B Behaviour & Information Technology %V 99999 %P 1 - 11 %8 2010/// %G eng %N 1 %0 Journal Article %J Microbiology and Immunology %D 2010 %T Conversion of viable but nonculturable Vibrio cholerae to the culturable state by co‐culture with eukaryotic cells %A Senoh,Mitsutoshi %A Ghosh‐Banerjee,Jayeeta %A Ramamurthy,Thandavarayan %A Hamabata,Takashi %A Kurakawa,Takashi %A Takeda,Makoto %A Rita R Colwell %A Nair,G. Balakrish %A Takeda,Yoshifumi %K conversion to culturability %K co‐culture %K eukaryotic cell %K viable but nonculturable (VBNC) Vibrio cholerae %X VBNC Vibrio cholerae O139 VC-280 obtained by incubation in 1% solution of artificial sea water IO at 4°C for 74 days converted to the culturable state when co-cultured with CHO cells. Other eukaryotic cell lines, including HT-29, Caco-2, T84, HeLa, and Intestine 407, also supported conversion of VBNC cells to the culturable state. Conversion of VBNC V. cholerae O1 N16961 and V. cholerae O139 VC-280/pG13 to the culturable state, under the same conditions, was also confirmed. When VBNC V. cholerae O139 VC-280 was incubated in 1% IO at 4°C for up to 91 days, the number of cells converted by co-culture with CHO cells declined with each additional day of incubation and after 91 days conversion was not observed. %B Microbiology and Immunology %V 54 %P 502 - 507 %8 2010/06/14/ %@ 1348-0421 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1348-0421.2010.00245.x/full %N 9 %R 10.1111/j.1348-0421.2010.00245.x %0 Journal Article %J Nucleic Acids ResearchNucl. Acids Res. %D 2010 %T Correlated Changes Between Regulatory Cis Elements and Condition-Specific Expression in Paralogous Gene Families %A Singh,Larry N. %A Hannenhalli, Sridhar %X Gene duplication is integral to evolution, providing novel opportunities for organisms to diversify in function. One fundamental pathway of functional diversification among initially redundant gene copies, or paralogs, is via alterations in their expression patterns. Although the mechanisms underlying expression divergence are not completely understood, transcription factor binding sites and nucleosome occupancy are known to play a significant role in the process. Previous attempts to detect genomic variations mediating expression divergence in orthologs have had limited success for two primary reasons. First, it is inherently challenging to compare expressions among orthologs due to variable trans-acting effects and second, previous studies have quantified expression divergence in terms of an overall similarity of expression profiles across multiple samples, thereby obscuring condition-specific expression changes. Moreover, the inherently inter-correlated expressions among homologs present statistical challenges, not adequately addressed in many previous studies. Using rigorous statistical tests, here we characterize the relationship between cis element divergence and condition-specific expression divergence among paralogous genes in Saccharomyces cerevisiae. In particular, among all combinations of gene family and TFs analyzed, we found a significant correlation between TF binding and the condition-specific expression patterns in over 20% of the cases. In addition, incorporating nucleosome occupancy reveals several additional correlations. For instance, our results suggest that GAL4 binding plays a major role in the expression divergence of the genes in the sugar transporter family. Our work presents a novel means of investigating the cis regulatory changes potentially mediating expression divergence in paralogous gene families under specific conditions. %B Nucleic Acids ResearchNucl. Acids Res. %V 38 %P 738 - 749 %8 2010/01/01/ %@ 0305-1048, 1362-4962 %G eng %U http://nar.oxfordjournals.org/content/38/3/738 %N 3 %R 10.1093/nar/gkp989 %0 Conference Paper %B Advances in Social Networks Analysis and Mining (ASONAM), 2010 International Conference on %D 2010 %T COSI: Cloud Oriented Subgraph Identification in Massive Social Networks %A Bröcheler,M. %A Pugliese, A. %A V.S. Subrahmanian %K (online); %K answering;social %K complexity;data %K computing;edge %K data;graph %K estimation;query %K handling;graph %K matching;Internet;computational %K matching;probability;query %K network;subgraph %K networking %K NP-complete;cloud %K pattern;probability %K processing;social %K theory;pattern %K weight;graph %X Subgraph matching is a key operation on graph data. Social network (SN) providers may want to find all subgraphs within their social network that match certain query graph patterns. Unfortunately, subgraph matching is NP-complete, making its application to massive SNs a major challenge. Past work has shown how to implement subgraph matching on a single processor when the graph has 10-25M edges. In this paper, we show how to use cloud computing in conjunction with such existing single processor methods to efficiently match complex subgraphs on graphs as large as 778M edges. A cloud consists of one master compute node and k slave compute nodes. We first develop a probabilistic method to estimate probabilities that a vertex will be retrieved by a random query and that a pair of vertices will be successively retrieved by a random query. We use these probability estimates to define edge weights in an SN and to compute minimal edge cuts to partition the graph amongst k slave nodes. We develop algorithms for both master and slave nodes that try to minimize communication overhead. The resulting COSI system can answer complex queries over real-world SN data containing over 778M edges very efficiently. %B Advances in Social Networks Analysis and Mining (ASONAM), 2010 International Conference on %P 248 - 255 %8 2010/08// %G eng %R 10.1109/ASONAM.2010.80 %0 Conference Paper %D 2010 %T Crowdsourcing the evaluation of a domain-adapted named entity recognition system %A Sayeed,Asad B. %A Meyer,Timothy J. %A Nguyen,Hieu C. %A Buzek,Olivia %A Weinberg, Amy %X Named entity recognition systems sometimes have difficulty when applied to data from domains that do not closely match the training data. We first use a simple rule-based technique for domain adaptation. Data for robust validation of the technique is then generated, and we use crowdsourcing techniques to show that this strategy produces reliable results even on data not seen by the rule designers. We show that it is possible to extract large improvements on the target data rapidly at low cost using these techniques. %S HLT '10 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 345 - 348 %8 2010/// %@ 1-932432-65-5 %G eng %U http://dl.acm.org/citation.cfm?id=1857999.1858050 %0 Journal Article %J Computer %D 2010 %T Cyberinfrastructure for Social Action on National Priorities %A Pirolli,Peter %A Preece,Jenny %A Shneiderman, Ben %K Collaborative tools %K Peer to peer computing %K Public policy %K Research initiatives %K Social network services %K Special issues and sections %K Technology-mediated social participation %X Extensive research is needed to build upon currently used media and tools to foster wider participation, address national priorities, and deal with potential dangers associated with technology-mediated social participation. %B Computer %V 43 %P 20 - 21 %8 2010/11// %@ 0018-9162 %G eng %N 11 %R 10.1109/MC.2010.315 %0 Patent %D 2010 %T Data hiding in compiled program binaries for supplementing computer functionality %A M. Wu %A Swaminathan,Ashwin %A Mao,Yinian %E University of Maryland %X Bit reductions in program instructions are achieved by determining the set of bit patterns in bit locations of the instructions. If only a subset of bit patterns is present in the instructions, they may be represented by an index value having a smaller number of bits. The index value may be substituted into the instruction and later used in decoding by referencing a corresponding bit pattern in a lookup table. The bit-reduction in the instruction makes way for supplemental data bits which may then be embedded. %V 11/445,374 %8 2010/11/23/ %G eng %U http://www.google.com/patents?id=tNHbAAAAEBAJ %N 7840789 %0 Conference Paper %B 2010 IEEE International Symposium on Parallel & Distributed Processing, Workshops and Phd Forum (IPDPSW) %D 2010 %T Decentralized dynamic scheduling across heterogeneous multi-core desktop grids %A Jaehwan Lee %A Keleher,P. %A Sussman, Alan %K backfill jobs %K bounded waiting time %K Computer science %K decentralized dynamic scheduling %K desktop grid resource management %K Dynamic scheduling %K Educational institutions %K Environmental management %K grid computing %K heterogeneous multicore desktop grid %K job assignment %K job migration %K load balancing %K Load management %K multicore computing environment %K Peer to peer computing %K Processor scheduling %K residual resources %K resource allocation %K Resource management %K scheduling %K Scheduling algorithm %K Throughput %X The recent advent of multi-core computing environments increases both the heterogeneity and complexity of managing desktop grid resources, making efficient load balancing challenging even for a centralized manager. Even with good initial job assignments, dynamic scheduling is still needed to adapt to dynamic environments, as well as for applications whose running times are not known a priori. In this paper, we propose new decentralized scheduling schemes that backfill jobs locally and dynamically migrate waiting jobs across nodes to leverage residual resources, while guaranteeing bounded waiting times for all jobs. The methods attempt to maximize total throughput while balancing load across available grid resources. Experimental results via simulation show that our scheduling scheme has performance competitive with an online centralized scheduler. %B 2010 IEEE International Symposium on Parallel & Distributed Processing, Workshops and Phd Forum (IPDPSW) %I IEEE %P 1 - 9 %8 2010/04/19/23 %@ 978-1-4244-6533-0 %G eng %R 10.1109/IPDPSW.2010.5470877 %0 Conference Paper %B SIGCHI '10 %D 2010 %T Deliberate Interactions: Characterizing Technology Use in Nairobi, Kenya %A Wyche, Susan P. %A Smyth, Thomas N. %A Marshini Chetty %A Aoki, Paul M. %A Grinter, Rebecca E. %K everyday technology %K hci4d %K Kenya %K urban computing %X We present results from a qualitative study examining how professionals living and working in Nairobi, Kenya regularly use ICT in their everyday lives. There are two contributions of this work for the HCI community. First, we provide empirical evidence demonstrating constraints our participants encountered when using technology in an infrastructure-poor setting. These constraints are limited bandwidth, high costs, differing perceptions of responsiveness, and threats to physical and virtual security. Second, we use our findings to critically evaluate the "access, anytime and anywhere" construct shaping the design of future technologies. We present an alternative vision called deliberate interactions--a planned and purposeful interaction style that involves offline preparation and discuss ways ICT can support this online usage behavior. %B SIGCHI '10 %S CHI '10 %I ACM %P 2593 - 2602 %8 2010/// %@ 978-1-60558-929-9 %G eng %U http://doi.acm.org/10.1145/1753326.1753719 %0 Journal Article %J Research in Computational Molecular Biology %D 2010 %T Dense Subgraphs with Restrictions and Applications to Gene Annotation Graphs %A Saha,B. %A Hoch,A. %A Khuller, Samir %A Raschid, Louiqa %A Zhang,X. N %X In this paper, we focus on finding complex annotation patterns representing novel and interesting hypotheses from gene annotation data. We define a generalization of the densest subgraph problem by adding an additional distance restriction (defined by a separate metric) to the nodes of the subgraph. We show that while this generalization makes the problem NP-hard for arbitrary metrics, when the metric comes from the distance metric of a tree, or an interval graph, the problem can be solved optimally in polynomial time. We also show that the densest subgraph problem with a specified subset of vertices that have to be included in the solution can be solved optimally in polynomial time. In addition, we consider other extensions when not just one solution needs to be found, but we wish to list all subgraphs of almost maximum density as well. We apply this method to a dataset of genes and their annotations obtained from The Arabidopsis Information Resource (TAIR). A user evaluation confirms that the patterns found in the distance restricted densest subgraph for a dataset of photomorphogenesis genes are indeed validated in the literature; a control dataset validates that these are not random patterns. Interestingly, the complex annotation patterns potentially lead to new and as yet unknown hypotheses. We perform experiments to determine the properties of the dense subgraphs, as we vary parameters, including the number of genes and the distance. %B Research in Computational Molecular Biology %P 456 - 472 %8 2010/// %G eng %R 10.1007/978-3-642-12683-3_30 %0 Journal Article %J Computer Vision and Image Understanding %D 2010 %T Design and implementation of embedded computer vision systems based on particle filters %A Sankalita Saha %A Bambha, Neal K. %A Bhattacharyya, Shuvra S. %K design space exploration %K Particle filters %K Reconfigurable platforms %X Particle filtering methods are gradually attaining significant importance in a variety of embedded computer vision applications. For example, in smart camera systems, object tracking is a very important application and particle filter based tracking algorithms have shown promising results with robust tracking performance. However, most particle filters involve vast amount of computational complexity, thereby intensifying the challenges faced in their real-time, embedded implementation. Many of these applications share common characteristics, and the same system design can be reused by identifying and varying key system parameters and varying them appropriately. In this paper, we present a System-on-Chip (SoC) architecture involving both hardware and software components for a class of particle filters. The framework uses parameterization to enable fast and efficient reuse of the architecture with minimal re-design effort for a wide range of particle filtering applications as well as implementation platforms. Using this framework, we explore different design options for implementing three different particle filtering applications on field-programmable gate arrays (FPGAs). The first two applications involve particle filters with one-dimensional state transition models, and are used to demonstrate the key features of the framework. The main focus of this paper is on design methodology for hardware/software implementation of multi-dimensional particle filter application and we explore this in the third application which is a 3D facial pose tracking system for videos. In this multi-dimensional particle filtering application, we extend our proposed architecture with models for hardware/software co-design so that limited hardware resources can be utilized most effectively. Our experiments demonstrate that the framework is easy and intuitive to use, while providing for efficient design and implementation. We present different memory management schemes along with results on trade-offs between area (FPGA resource requirement) and execution speed. %B Computer Vision and Image Understanding %V 114 %P 1203 - 1214 %8 2010 %@ 1077-3142 %G eng %U http://www.sciencedirect.com/science/article/pii/S1077314210000950 %N 11 %! Computer Vision and Image Understanding %0 Conference Paper %B 2010 Conference Record of the Forty Fourth Asilomar Conference on Signals, Systems and Computers (ASILOMAR) %D 2010 %T Design and implementation of real-time signal processing applications on heterogeneous multiprocessor arrays %A Wu, Hsiang-Huang %A Chung-Ching Shen %A Bhattacharyya, Shuvra S. %A Compton, K. %A Schulte, M. %A Wolf, M. %A Zhang, Tong %K application specific integrated circuits %K application-specific integrated circuits %K computational elements %K Computer architecture %K decoding %K Field programmable gate arrays %K field programmable X arrays %K FPGA %K FPXA %K Integrated circuit modeling %K Logic Design %K microprocessor chips %K multicore processors %K multiprocessor arrays %K real-time signal processing %K reconfigurable architectures %K reconfigurable processors %K Routing %K Signal processing %K Signal processing systems %K systolic arrays %K Viterbi algorithm %X Processing structures based on arrays of computational elements form an important class of architectures, which includes field programmable gate arrays (FPGAs), systolic arrays, and various forms of multicore processors. A wide variety of design methods and tools have been targeted to regular processing arrays involving homogeneous processing elements. In this paper, we introduce the concept of field programmable X arrays (FPXAs) as an abstract model for design and implementation of heterogeneous multiprocessor arrays for signal processing systems. FPXAs are abstract structures that can be targeted for implementation on application-specific integrated circuits, FPGAs, or other kinds of reconfigurable processors. FPXAs can also be mapped onto multicore processors for flexible emulation. We discuss the use of dataflow models as an integrated application representation and intermediate representation for efficient specification and mapping of signal processing systems on FPXAs. We demonstrate our proposed models and techniques with a case study involving the embedding of an application-specific FPXA system on an off-the-shelf FPGA device. %B 2010 Conference Record of the Forty Fourth Asilomar Conference on Signals, Systems and Computers (ASILOMAR) %P 2121 - 2125 %8 2010 %G eng %0 Conference Paper %B Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems %D 2010 %T Determining the spatial reader scopes of news sources using local lexicons %A Quercini,Gianluca %A Samet, Hanan %A Sankaranarayanan,Jagan %A Lieberman,Michael D. %K geotagging %K local lexicon %K news search engine %K spatial reader scope %K spatial relatedness %X Information sources on the Internet (e.g., Web versions of newspapers) usually have an implicit spatial reader scope, which is the geographical location for which the content has been primarily produced. Knowledge of the spatial reader scope facilitates the construction of a news search engine that provides readers a set of news sources relevant to the location in which they are interested. In particular, it plays an important role in disambiguating toponyms (e.g., textual specifications of geographical locations) in news articles, as the process of selecting an interpretation for the toponym often reduces to one of selecting an interpretation that seems natural in the context of the spatial reader scope. The key to determining the spatial reader scope of news sources is the notion of local lexicon, which for a location s is a set of concepts such as, but not limited to, names of people, landmarks, and historical events, that are spatially related to s. Techniques to automatically generate the local lexicon of a location by using the link structure of Wikipedia are described and evaluated. A key contribution is the improvement of existing methods used in the semantic relatedness domain to extract concepts spatially related to a given location from the Wikipedia. Results of experiments are presented that indicate that the knowledge of the spatial reader scope significantly improves the disambiguation of textually specified locations in news articles and that using local lexicons is an effective method to determine the spatial reader scopes of news sources. %B Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems %S GIS '10 %I ACM %C New York, NY, USA %P 43 - 52 %8 2010/// %@ 978-1-4503-0428-3 %G eng %U http://doi.acm.org/10.1145/1869790.1869800 %R 10.1145/1869790.1869800 %0 Report %D 2010 %T Distributed Implementation of Coordinated, Network-Wide Policies and Protocols with FlowFlex %A Mundada,Y. %A Sherwood,R. %A Feamster, Nick %X The increasing programmability of network devices gives protocol designers and network operators considerably more flexibility in defining custom protocols and traffic processing functions. Today, network operators and protocol designers have the option of either operating at flow-level granularity, which offers coordinated control; or packet-level granularity, which offers flexibility, but not coordinated control. Today’s network programming paradigms force operators to choose between the fine-grained control and expressiveness of packet processing and the coordination of flow processing, which makes it difficult to quickly realize a distributed implementation of a global, network-wide policy. Designers must also choose between the flexibility of hardware-based solutions and the fast development cycles offered by software. This paper proposes a system called FlowFlex that offers network designers the best of both worlds: with FlowFlex, operators can quickly design, implement, and deploy network systems and protocols that offer fast, distributed, implementations that require coordinated control and fine-grained operations on packets. We present the design and implementation of the FlowFlex framework and show how it can improve both expressiveness and efficiency for three real-world networking applications. %I Georgia Institute of Technology %V GT-CS-10-07 %8 2010/// %G eng %U http://hdl.handle.net/1853/35002 %0 Conference Paper %B Proceedings of the 9th international conference on Peer-to-peer systems %D 2010 %T Don't love thy nearest neighbor %A Lumezanu,Cristian %A Levin,Dave %A Han,Bo %A Spring, Neil %A Bhattacharjee, Bobby %B Proceedings of the 9th international conference on Peer-to-peer systems %S IPTPS'10 %I USENIX Association %C Berkeley, CA, USA %P 5 - 5 %8 2010/// %G eng %U http://dl.acm.org/citation.cfm?id=1863145.1863150 %0 Conference Paper %B Proceedings of the 10th annual conference on Internet measurement %D 2010 %T The effect of packet loss on redundancy elimination in cellular wireless networks %A Lumezanu,Cristian %A Guo,Katherine %A Spring, Neil %A Bhattacharjee, Bobby %K cellular networks %K loss %K redundancy elimination %X Network-level redundancy elimination (RE) algorithms reduce traffic volume on bandwidth-constrained network paths by avoiding the transmission of repeated byte sequences. Previous work shows that RE can suppress the transmission of 20-50% bytes when deployed at ISP access links or between routers. In this paper, we focus on the challenges of deploying RE in cellular networks. The potential benefifit is substantial, since cellular networks have a growing subscriber base and network links, including wired backhaul, are often oversubscribed. Using three large traces captured at two North American and one European wireless network providers, we show that RE can reduce the bandwidth consumption of the majority of mobile users by at least 10%. However, cellular links have much higher packet loss rates than their wired counterparts, which makes applying RE much more difficult. Our experiments also show that the loss of only a few packets can disrupt RE and eliminate the bandwidth savings. We propose informed marking, a lightweight scheme that detects lost packets and prevents RE algorithms from using them for future encodings. We implement RE with informed marking and deploy it in a real-world cellular network. Our results show that with informed marking, more than 60% of the bandwidth savings of RE are preserved, even when packet loss rates are high. %B Proceedings of the 10th annual conference on Internet measurement %S IMC '10 %I ACM %C New York, NY, USA %P 294 - 300 %8 2010/// %@ 978-1-4503-0483-2 %G eng %U http://doi.acm.org/10.1145/1879141.1879179 %R 10.1145/1879141.1879179 %0 Journal Article %J Proceedings of the 20 th Conference on Probability and Statistics in the Atmospheric Sciences %D 2010 %T Efficient kriging for real-time spatio-temporal interpolation %A Srinivasan,B.V. %A Duraiswami, Ramani %A Murtugudde,R. %X Atmospheric data is often recorded at scattered stationlocations. While the data is generally available over a long period of time it cannot be used directly for extract- ing coherent patterns and mechanistic correlations. The only recourse is to spatially and temporally interpolate the data both to organize the station recording to a reg- ular grid and to query the data for predictions at a par- ticular location or time of interest. Spatio-temporal in- terpolation approaches require the evaluation of weights at each point of interest. A widely used interpolation ap- proach is kriging. However, kriging has a computational cost that scales as the cube of the number of data points N, resulting in cubic time complexity for each point of interest, which leads to a time complexity of O(N4) for interpolation at O(N) points. In this work, we formulate the kriging problem, to first reduce the computational cost to O(N3). We use an iterative solver (Saad, 2003), and further accelerate the solver using fast summation algo- rithms like GPUML (Srinivasan and Duraiswami, 2009) or FIGTREE (Morariu et al., 2008). We illustrate the speedup on synthetic data and compare the performance with other standard kriging approaches to demonstrate substantial improvement in the performance of our ap- proach. We then apply the developed approach on ocean color data from the Chesapeake Bay and present some quantitative analysis of the kriged results. %B Proceedings of the 20 th Conference on Probability and Statistics in the Atmospheric Sciences %P 228 - 235 %8 2010/// %G eng %0 Conference Paper %B Scalable Uncertainty Management: 4th International Conference %D 2010 %T Efficient Policy-Based Inconsistency Management in Relational Knowledge Bases %A Martinez,M. V %A Parisi,F. %A Pugliese, A. %A Simari,G. I %A V.S. Subrahmanian %X Real-world databases are frequently inconsistent. Even though the users who work with a body of data are far more familiar not only with that data, but also their own job and the risks they are willing to take and the inferences they are willing to make from inconsistent data, most DBMSs force them to use the policy embedded in the DBMS. Inconsistency management policies (IMPs) were introduced so that users can apply policies that they deem are appropriate for data they know and understand better than anyone else. In this paper, we develop an efficient “cluster table” method to implement IMPs and show that using cluster tables instead of a standard DBMS index is far more efficient when less than about 3% of a table is involved in an inconsistency (which is hopefully the case in most real world DBs), while standard DBMS indexes perform better when the amount of inconsistency in a database is over 3%. %B Scalable Uncertainty Management: 4th International Conference %V 6379 %P 264 - 264 %8 2010/// %G eng %R 10.1007/978-3-642-15951-0_26 %0 Conference Paper %B Proceedings of the Twenty-First Annual ACM-SIAM Symposium on Discrete Algorithms %D 2010 %T Energy efficient scheduling via partial shutdown %A Khuller, Samir %A Li,Jian %A Saha,Barna %X Motivated by issues of saving energy in data centers we define a collection of new problems referred to as "machine activation" problems. The central framework we introduce considers a collection of m machines (unrelated or related) with each machine i having an activation cost of ai. There is also a collection of n jobs that need to be performed, and pi, j is the processing time of job j on machine i. Standard scheduling models assume that the set of machines is fixed and all machines are available. However, in our setting, we assume that there is an activation cost budget of A - we would like to select a subset S of the machines to activate with total cost a(S) ≤ A and find a schedule for the n jobs on the machines in S minimizing the makespan (or any other metric). We consider both the unrelated machines setting, as well as the setting of scheduling uniformly related parallel machines, where machine i has activation cost ai and speed si, and the processing time of job j on machine i is pi, j = pjsi, where pj is the processing requirement of job j. For the general unrelated machine activation problem, our main results are that if there is a schedule with makespan T and activation cost A then we can obtain a schedule with makespan (2 + ε)T and activation cost 2(1 + 1/ε)(ln n/OPT + 1)A, for any ε > 0. We also consider assignment costs for jobs as in the generalized assignment problem, and using our framework, provide algorithms that minimize the machine activation and the assignment cost simultaneously. In addition, we present a greedy algorithm which only works for the basic version and yields a makespan of 2T and an activation cost A(1 + ln n). For the uniformly related parallel machine sceduling problem, we develop a polynomial time approximation scheme that outputs a schedule with the property that the activation cost of the subset of machines is at most A and the makespan is at most (1 + ε)T for any ε > 0. For the special case of m identical speed machines, the machine activation problem is trivial, since the cheapest subset of k machines is always the best choice if the optimal solution activates k machines. In addition, we consider the case whem some jobs can be dropped (and are treated as outliers). %B Proceedings of the Twenty-First Annual ACM-SIAM Symposium on Discrete Algorithms %S SODA '10 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 1360 - 1372 %8 2010/// %@ 978-0-898716-98-6 %G eng %U http://dl.acm.org/citation.cfm?id=1873601.1873711 %0 Journal Article %J ACM Trans. Sen. Netw. %D 2010 %T Energy-driven distribution of signal processing applications across wireless sensor networks %A Chung-Ching Shen %A Plishker, William L. %A Ko,Dong-Ik %A Bhattacharyya, Shuvra S. %A Goldsman,Neil %K DSP %K Energy efficiency %K network lifetime %K Speech recognition %K Wireless sensor networks %X Wireless sensor network (WSN) applications have been studied extensively in recent years. Such applications involve resource-limited embedded sensor nodes that have small size and low power requirements. Based on the need for extended network lifetimes in WSNs in terms of energy use, the energy efficiency of computation and communication operations in the sensor nodes becomes critical. Digital Signal Processing (DSP) applications typically require intensive data processing operations and as a result are difficult to implement directly in resource-limited WSNs. In this article, we present a novel design methodology for modeling and implementing computationally intensive DSP applications applied to wireless sensor networks. This methodology explores efficient modeling techniques for DSP applications, including data sensing and processing; derives formulations of Energy-Driven Partitioning (EDP) for distributing such applications across wireless sensor networks; and develops efficient heuristic algorithms for finding partitioning results that maximize the network lifetime. To address such an energy-driven partitioning problem, this article provides a new way of aggregating data and reducing communication traffic among nodes based on application analysis. By considering low data token delivery points and the distribution of computation in the application, our approach finds energy-efficient trade-offs between data communication and computation. %B ACM Trans. Sen. Netw. %V 6 %P 24:1 - 24:32 %8 2010 %@ 1550-4859 %G eng %U http://doi.acm.org/10.1145/1754414.1754420 %N 3 %0 Journal Article %J Journal of Computational Biology %D 2010 %T Evaluation of the Interaction of Cyclin-Dependent Kinase 5 with Activator p25 and with p25-Derived Inhibitor CIP %A Cardone, Antonio %A Albers,R.W. %A Sriram,R.D. %A Pant,H.C. %B Journal of Computational Biology %V 17 %P 707 - 721 %8 2010/// %G eng %N 5 %0 Book Section %B Genetics and Molecular Biology of LepidopteraGenetics and Molecular Biology of Lepidoptera %D 2010 %T Evolutionary framework for Lepidoptera model systems %A Roe,A. %A Weller,S. %A Baixeras,J. %A Brown,J. W %A Cummings, Michael P. %A Davis,DR %A Horak,M %A Kawahara,A. Y %A Mitter,C %A Parr,C.S. %A Regier,J. C %A Rubinoff,D %A Simonsen,TJ %A Wahlberg,N %A Zwick,A. %E Goldsmith,M %E Marec,F %X “Model systems” are specific organisms upon which detailed studies have been conducted examining a fundamental biological question. If the studies are robust, their results can be extrapolated among an array of organisms that possess features in common with the subject organism. The true power of model systems lies in the ability to extrapolate these details across larger groups of organisms. In order to generalize these results, comparative studies are essential and require that model systems be placed into their evolutionary or phylogenetic context. This chapter examines model systems in the insect order Lepidoptera from the perspective of several different superfamilies. Historically, many species of Lepidoptera have been essential in the development of invaluable model systems in the fields of development biology, genetics, molecular biology, physiology, co-evolution, population dynamics, and ecology. %B Genetics and Molecular Biology of LepidopteraGenetics and Molecular Biology of Lepidoptera %I Taylor & Francis %C Boca Raton %P 1 - 24 %8 2010/// %G eng %0 Conference Paper %B Proceedings of the 12th annual conference on Genetic and evolutionary computation %D 2010 %T Evolving viral marketing strategies %A Stonedahl,Forrest %A Rand, William %A Wilensky,Uri %K Agent-based modeling %K business %K diffusion %K Genetic algorithms %K simulation %K social networks %K viral marketing %X One method of viral marketing involves seeding certain consumers within a population to encourage faster adoption of the product throughout the entire population. However, determining how many and which consumers within a particular social network should be seeded to maximize adoption is challenging. We define a strategy space for consumer seeding by weighting a combination of network characteristics such as average path length, clustering coefficient, and degree. We measure strategy effectiveness by simulating adoption on a Bass-like agent-based model, with five different social network structures: four classic theoretical models (random, lattice, small-world, and preferential attachment) and one empirical (extracted from Twitter friendship data). To discover good seeding strategies, we have developed a new tool, called BehaviorSearch, which uses genetic algorithms to search through the parameter-space of agent-based models. This evolutionary search also provides insight into the interaction between strategies and network structure. Our results show that one simple strategy (ranking by node degree) is near-optimal for the four theoretical networks, but that a more nuanced strategy performs significantly better on the empirical Twitter-based network. We also find a correlation between the optimal seeding budget for a network, and the inequality of the degree distribution. %B Proceedings of the 12th annual conference on Genetic and evolutionary computation %S GECCO '10 %I ACM %C New York, NY, USA %P 1195 - 1202 %8 2010/// %@ 978-1-4503-0072-8 %G eng %U http://doi.acm.org/10.1145/1830483.1830701 %R 10.1145/1830483.1830701 %0 Journal Article %J University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2010-01 %D 2010 %T Exploring distributions: design and evaluation %A Sopan,A. %A Freire,M. %A Taieb-Maimon,M. %A Golbeck,J. %A Shneiderman, Ben %X Visual overviews of tables of numerical and categorical data have been proposed for tables with a single value per cell.In this paper we address the problem of exploring tables including columns consisting of distributions, e.g. the distributions of movie ratings or trust ratings in recommender systems, age distributions in demographic data, usage distributions in logs of telephone calls etc. We propose a novel way of displaying and interacting with distribution data, and present the results of a usability study that demonstrates the benefits of the interface in providing an overview of the data and facilitating the discovery of interesting clusters, patterns, outliers and relationships between columns. %B University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2010-01 %8 2010/// %G eng %0 Conference Paper %B Hawaii International Conference on System Sciences %D 2010 %T Exploring the Relationships among ICTs: A Scalable Computational Approach Using KL Divergence and Hierarchical Clustering %A Tsui,Chia-jung %A Wang,Ping %A Fleischmann,Kenneth R. %A Oard, Douglas %A Sayeed,Asad B. %X Different information and communication technologies (ICTs) are related in complex ways and, accordingly, their diffusion trajectories are related, too. How can the relationships among multiple ICTs be described and analyzed in a scalable way? In this study, we offer a scalable methodology, based on computational analysis of discourse, to examine the relationships among ICTs. Specifically, we employed Kullback-Leibler (KL) divergence to compare the semantic similarity of forty-seven ICTs discussed in the trade magazine InformationWeek over a decade. Using hierarchical clustering, we have found that the similarity of the technologies can be mapped in a hierarchy and similar technologies demonstrated similar discourses. The results establish the validity of our approach and demonstrate its scalability and richness. This analytical approach not only enables diffusion researchers to undertake multi-innovation, multi-source, and multi-period studies, but also helps practitioners effectively adopt and efficiently use new ICTs in their organizations. %B Hawaii International Conference on System Sciences %I IEEE Computer Society %C Los Alamitos, CA, USA %P 1 - 10 %8 2010/// %@ 978-0-7695-3869-3 %G eng %R http://doi.ieeecomputersociety.org/10.1109/HICSS.2010.203 %0 Conference Paper %D 2010 %T Expresses-an-opinion-about: using corpus statistics in an information extraction approach to opinion mining %A Sayeed,Asad B. %A Nguyen,Hieu C. %A Meyer,Timothy J. %A Weinberg, Amy %X We present a technique for identifying the sources and targets of opinions without actually identifying the opinions themselves. We are able to use an information extraction approach that treats opinion mining as relation mining; we identify instances of a binary "expresses-an-opinion-about" relation. We find that we can classify source-target pairs as belonging to the relation at a performance level significantly higher than two relevant baselines. This technique is particularly suited to emerging approaches in corpus-based social science which focus on aggregating interactions between sources to determine their effects on socio-economically significant targets. Our application is the analysis of information technology (IT) innovations. This is an example of a more general problem where opinion is expressed using either sub- or supersets of expressive words found in newswire. We present an annotation scheme and an SVM-based technique that uses the local context as well as the corpus-wide frequency of a source-target pair as data to determine membership in "expresses-an-opinion-about". While the presence of conventional subjectivity keywords appears significant in the success of this technique, we are able to find the most domain-relevant keywords without sacrificing recall. %S COLING '10 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 1095 - 1103 %8 2010/// %G eng %U http://dl.acm.org/citation.cfm?id=1944566.1944692 %0 Journal Article %J Computer %D 2010 %T Face Recognition by Computers and Humans %A Chellapa, Rama %A Sinha, P. %A Phillips,P.J. %K access %K Computer %K control;aging;computer %K control;face %K entertainment;database %K interaction;national %K interaction;unconstrained %K reality;access %K recognition;homeland %K recognition;human %K retrieval;face %K security; %K security;human %K sequences;virtual %K video %X This article talks about how the study of how humans perceive faces can be used to help design practical systems for face recognition. Besides applications related to identification and verification-such as access control, law enforcement, ID and licensing, and surveillance-face recognition has also proven useful in applications such as human-computer interaction, virtual reality, database retrieval, multimedia, and computer entertainment. Continuing research into face recognition will provide scientists and engineers with many vital projects, in areas such as homeland security, human-computer interaction, and numerous consumer applications. The areas we are considering pursuing are recognition from unconstrained video sequences, incorporating familiarity into algorithms, modeling effects of aging, and developing biologically plausible models for human face recognition ability. %B Computer %V 43 %P 46 - 55 %8 2010/02// %@ 0018-9162 %G eng %N 2 %R 10.1109/MC.2010.37 %0 Journal Article %J Information Forensics and Security, IEEE Transactions on %D 2010 %T Face Verification Across Age Progression Using Discriminative Methods %A Ling,Haibin %A Soatto,S. %A Ramanathan,N. %A Jacobs, David W. %K algorithms;support %K databases; %K dataset;age %K FGnet %K hair;gradient %K information;image %K information;recognition %K machine;face %K machines;visual %K methods;eyewear;face %K methods;support %K orientation %K orientation;gradient %K progression;commercial %K pyramid;hierarchical %K quality;magnitude %K recognition;gradient %K systems;discriminative %K vector %K verification;facial %X Face verification in the presence of age progression is an important problem that has not been widely addressed. In this paper, we study the problem by designing and evaluating discriminative approaches. These directly tackle verification tasks without explicit age modeling, which is a hard problem by itself. First, we find that the gradient orientation, after discarding magnitude information, provides a simple but effective representation for this problem. This representation is further improved when hierarchical information is used, which results in the use of the gradient orientation pyramid (GOP). When combined with a support vector machine GOP demonstrates excellent performance in all our experiments, in comparison with seven different approaches including two commercial systems. Our experiments are conducted on the FGnet dataset and two large passport datasets, one of them being the largest ever reported for recognition tasks. Second, taking advantage of these datasets, we empirically study how age gaps and related issues (including image quality, spectacles, and facial hair) affect recognition algorithms. We found surprisingly that the added difficulty of verification produced by age gaps becomes saturated after the gap is larger than four years, for gaps of up to ten years. In addition, we find that image quality and eyewear present more of a challenge than facial hair. %B Information Forensics and Security, IEEE Transactions on %V 5 %P 82 - 91 %8 2010/03// %@ 1556-6013 %G eng %N 1 %R 10.1109/TIFS.2009.2038751 %0 Journal Article %J Copper Mountain Conference on Iterative Methods %D 2010 %T Fast matrix-vector product based fgmres for kernel machines %A Srinivasan,B.V. %A Duraiswami, Ramani %A Gumerov, Nail A. %X Kernel based approaches for machine learning have gained huge interest in the past decades because of their robustness. In somealgorithms, the primary problem is the solution of a linear system involving the kernel matrix. Iterative Krylov approaches are often used to solve these efficiently [2, 3]. Fast matrix-vector products can be used to accelerate each Krylov iteration to further optimize the performance. In order to reduce the number of iterations of the Krylov approach, a preconditioner becomes necessary in many cases. Several researchers have proposed flexible preconditioning methods where the preconditioner changes with each iteration, and this class of preconditioners are shown to have good performance [6, 12]. In this paper, we use a Tikhonov regularized kernel matrix as a preconditioner for flexible GMRES [12] to solve kernel matrix based systems of equations. We use a truncated conjugate gradient (CG) method to solve the preconditioner system and further accelerate each CG iteration using fast matrix-vector products. The convergence of the proposed preconditioned GMRES is shown on synthetic data. The performance is further validated on problems in Gaussian process regression and radial basis function interpolation. Improvements are seen in each case. %B Copper Mountain Conference on Iterative Methods %V 2 %8 2010/// %G eng %0 Book Section %B Integer Programming and Combinatorial OptimizationInteger Programming and Combinatorial Optimization %D 2010 %T Fault-Tolerant Facility Location: A Randomized Dependent LP-Rounding Algorithm %A Byrka,Jaroslaw %A Srinivasan, Aravind %A Swamy,Chaitanya %E Eisenbrand,Friedrich %E Shepherd,F. %B Integer Programming and Combinatorial OptimizationInteger Programming and Combinatorial Optimization %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6080 %P 244 - 257 %8 2010/// %@ 978-3-642-13035-9 %G eng %U http://dx.doi.org/10.1007/978-3-642-13036-6_19 %0 Journal Article %J Interactions %D 2010 %T From bowling alone to tweeting together: technology-mediated social participation %A Hochheiser,H. %A Shneiderman, Ben %B Interactions %V 17 %P 64 - 67 %8 2010/// %G eng %N 2 %0 Conference Paper %B ACM SIGPLAN Notices %D 2010 %T From program verification to program synthesis %A Srivastava,S. %A Gulwani,S. %A Foster, Jeffrey S. %B ACM SIGPLAN Notices %V 45 %P 313 - 326 %8 2010/// %G eng %0 Book Section %B Pervasive Computing %D 2010 %T GasSense: Appliance-Level, Single-Point Sensing of Gas Activity in the Home %A Cohn,Gabe %A Gupta,Sidhant %A Jon Froehlich %A Larson,Eric %A Patel,Shwetak %E Floréen,Patrik %E Krüger,Antonio %E Spasojevic,Mirjana %X This paper presents GasSense, a low-cost, single-point sensing solution for automatically identifying gas use down to its source (e.g., water heater, furnace, fireplace). This work adds a complementary sensing solution to the growing body of work in infrastructure-mediated sensing. GasSense analyzes the acoustic response of a home’s government mandated gas regulator, which provides the unique capability of sensing both the individual appliance at which gas is currently being consumed as well as an estimate of the amount of gas flow. Our approach provides a number of appealing features including the ability to be easily and safely installed without the need of a professional. We deployed our solution in nine different homes and initial results show that GasSense has an average accuracy of 95.2% in identifying individual appliance usage. %B Pervasive Computing %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6030 %P 265 - 282 %8 2010 %@ 978-3-642-12653-6 %G eng %U http://dx.doi.org/10.1007/978-3-642-12654-3_16 %0 Journal Article %J Genome biology %D 2010 %T Genomic characterization of the Yersinia genus %A Chen, P.E. %A Cook, C. %A Stewart, A.C. %A Nagarajan,N. %A Sommer, D.D. %A Pop, Mihai %A Thomason, B. %A Thomason, M.P.K. %A Lentz, S. %A Nolan, N. %A others %B Genome biology %V 11 %8 2010 %G eng %N 1 %0 Conference Paper %B Proceedings of the 6th Workshop on Geographic Information Retrieval %D 2010 %T Geotagging: using proximity, sibling, and prominence clues to understand comma groups %A Lieberman,Michael D. %A Samet, Hanan %A Sankaranayananan,Jagan %K comma groups %K geotagging %K toponyms %X Geotagging is the process of recognizing textual references to geographic locations, known as toponyms, and resolving these references by assigning each lat/long values. Typical geotagging algorithms use a variety of heuristic evidence to select the correct interpretation for each toponym. A study is presented of one such heuristic which aids in recognizing and resolving lists of toponyms, referred to as comma groups. Comma groups of toponyms are recognized and resolved by inferring the common threads that bind them together, based on the toponyms' shared geographic attributes. Three such common threads are proposed and studied --- population-based prominence, distance-based proximity, and sibling relationships in a geographic hierarchy --- and examples of each are noted. In addition, measurements are made of these comma groups' usage and variety in a large dataset of news articles, indicating that the proposed heuristics, and in particular the proximity and sibling heuristics, are useful for resolving comma group toponyms. %B Proceedings of the 6th Workshop on Geographic Information Retrieval %S GIR '10 %I ACM %C New York, NY, USA %P 6:1–6:8 - 6:1–6:8 %8 2010/// %@ 978-1-60558-826-1 %G eng %U http://doi.acm.org/10.1145/1722080.1722088 %R 10.1145/1722080.1722088 %0 Conference Paper %B Data Engineering (ICDE), 2010 IEEE 26th International Conference on %D 2010 %T Geotagging with local lexicons to build indexes for textually-specified spatial data %A Lieberman,M.D. %A Samet, Hanan %A Sankaranarayanan,J. %K data;Internet;data %K databases; %K databases;spatial %K document %K indexes;spatial %K information %K Internet;document-independent %K knowledge;feature-based %K lexicon %K lexicons;location-based %K locations;geotagging;inference %K method;internal %K methods;geographic %K mining;geographic %K model;external %K model;textually-specified %K queries;generic %K queries;spatial %K spatial %K structure;local %K systems;visual %X The successful execution of location-based and feature-based queries on spatial databases requires the construction of spatial indexes on the spatial attributes. This is not simple when the data is unstructured as is the case when the data is a collection of documents such as news articles, which is the domain of discourse, where the spatial attribute consists of text that can be (but is not required to be) interpreted as the names of locations. In other words, spatial data is specified using text (known as a toponym) instead of geometry, which means that there is some ambiguity involved. The process of identifying and disambiguating references to geographic locations is known as geotagging and involves using a combination of internal document structure and external knowledge, including a document-independent model of the audience's vocabulary of geographic locations, termed its spatial lexicon. In contrast to previous work, a new spatial lexicon model is presented that distinguishes between a global lexicon of locations known to all audiences, and an audience-specific local lexicon. Generic methods for inferring audiences' local lexicons are described. Evaluations of this inference method and the overall geotagging procedure indicate that establishing local lexicons cannot be overlooked, especially given the increasing prevalence of highly local data sources on the Internet, and will enable the construction of more accurate spatial indexes. %B Data Engineering (ICDE), 2010 IEEE 26th International Conference on %P 201 - 212 %8 2010/03// %G eng %R 10.1109/ICDE.2010.5447903 %0 Journal Article %J Proceedings of HCI %D 2010 %T Getting serious about social media: strategies for increasing civic participation %A Shneiderman, Ben %B Proceedings of HCI %8 2010/// %G eng %0 Journal Article %J Workshop on High Performance Analytics-Algorithms, Implementations, and Applications %D 2010 %T GPUML: Graphical processors for speeding up kernel machines %A Srinivasan,B.V. %A Hu,Q. %A Duraiswami, Ramani %X Algorithms based on kernel methods play a central rolein statistical machine learning. At their core are a num- ber of linear algebra operations on matrices of kernel functions which take as arguments the training and test- ing data. These range from the simple matrix-vector product, to more complex matrix decompositions, and iterative formulations of these. Often the algorithms scale quadratically or cubically, both in memory and op- erational complexity, and as data sizes increase, kernel methods scale poorly. We use parallelized approaches on a multi-core graphical processor (GPU) to partially address this lack of scalability. GPUs are used to scale three different classes of problems, a simple kernel- matrix-vector product, iterative solution of linear sys- tems of kernel function and QR and Cholesky decom- position of kernel matrices. Application of these accel- erated approaches in scaling several kernel based learn- ing approaches are shown, and in each case substantial speedups are obtained. The core software is released as an open source package, GPUML. %B Workshop on High Performance Analytics-Algorithms, Implementations, and Applications %8 2010/// %G eng %0 Conference Paper %B Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1 %D 2010 %T A graph-theoretic approach to protect static and moving targets from adversaries %A Dickerson,J. P. %A Simari,G. I %A V.S. Subrahmanian %A Kraus,Sarit %K adversarial reasoning %K agent systems %K game theory %X The static asset protection problem (SAP) in a road network is that of allocating resources to protect vertices, given any possible behavior by an adversary determined to attack those assets. The dynamic asset protection (DAP) problem is a version of SAP where the asset is following a fixed and widely known route (e.g., a parade route) and needs to be protected. We formalize what it means for a given allocation of resources to be "optimal" for protecting a desired set of assets, and show that randomly allocating resources to a single edge cut in the road network solves this problem. Unlike SAP, we show that DAP is not only an NP-complete problem, but that approximating DAP is also NP-hard. We provide the GreedyDAP heuristic algorithm to solve DAP and show experimentally that it works well in practice, using road network data for real cities. %B Proceedings of the 9th International Conference on Autonomous Agents and Multiagent Systems: volume 1 - Volume 1 %S AAMAS '10 %I International Foundation for Autonomous Agents and Multiagent Systems %C Richland, SC %P 299 - 306 %8 2010/// %@ 978-0-9826571-1-9 %G eng %U http://dl.acm.org/citation.cfm?id=1838206.1838248 %0 Journal Article %J NIPS 2010 Workshop on Networks Across Disciplines: Theory and Applications, Whistler BC, Canada %D 2010 %T Higher-order graphical models for classification in social and affiliation networks %A Zheleva,E. %A Getoor, Lise %A Sarawagi,S. %X In this work we explore the application of higher-order Markov Random Fields(MRF) to classification in social and affiliation networks. We consider both friend- ship links and group membership for inferring hidden attributes in a collective inference framework. We explore different ways of using the social groups as ei- ther node features or to construct the graphical model structure. The bottleneck in applying higher-order MRFs to a domain with many overlapping large cliques is the complexity of inference which is exponential in the size of the largest clique. To circumvent the slow inference problem, we borrow recent advancements in the computer vision community to achieve fast approximate inference results. We provide preliminary results using a dataset from facebook which suggest that our higher-order MRF models are capturing the structural dependencies in the net- works and they yield higher accuracy than linear classifiers. %B NIPS 2010 Workshop on Networks Across Disciplines: Theory and Applications, Whistler BC, Canada %8 2010/// %G eng %0 Journal Article %J Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %D 2010 %T Identification of Pathogenic Vibrio Species by Multilocus PCR-Electrospray Ionization Mass Spectrometry and Its Application to Aquatic Environments of the Former Soviet Republic of Georgia %A Whitehouse,Chris A. %A Baldwin,Carson %A Sampath,Rangarajan %A Blyn,Lawrence B. %A Melton,Rachael %A Li,Feng %A Hall,Thomas A. %A Harpin,Vanessa %A Matthews,Heather %A Tediashvili,Marina %A Jaiani,Ekaterina %A Kokashvili,Tamar %A Janelidze,Nino %A Grim,Christopher %A Rita R Colwell %A Huq,Anwar %X The Ibis T5000 is a novel diagnostic platform that couples PCR and mass spectrometry. In this study, we developed an assay that can identify all known pathogenic Vibrio species and field-tested it using natural water samples from both freshwater lakes and the Georgian coastal zone of the Black Sea. Of the 278 total water samples screened, 9 different Vibrio species were detected, 114 (41%) samples were positive for V. cholerae, and 5 (0.8%) samples were positive for the cholera toxin A gene (ctxA). All ctxA-positive samples were from two freshwater lakes, and no ctxA-positive samples from any of the Black Sea sites were detected. %B Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %V 76 %P 1996 - 2001 %8 2010/03/15/ %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/76/6/1996 %N 6 %R 10.1128/AEM.01919-09 %0 Conference Paper %B Pattern Recognition (ICPR), 2010 20th International Conference on %D 2010 %T Images in News %A Sankaranarayanan,J. %A Samet, Hanan %K algorithm;feature %K articles;news %K clustering %K clustering;publishing; %K extraction;hypermedia %K extraction;image %K feeds;caption %K HTML %K images;news %K LANGUAGE %K languages;image %K markup %K page;RSS %K processing;news %K processing;pattern %K repository;meta-information;natural %K retrieval;natural %K stand;online %K text;feature %K vector;image %X A system, called News Stand, is introduced that automatically extracts images from news articles. The system takes RSS feeds of news article and applies an online clustering algorithm so that articles belonging to the same news topic can be associated with the same cluster. Using the feature vector associated with the cluster, the images from news articles that form the cluster are extracted. First, the caption text associated with each of the images embedded in the news article is determined. This is done by analyzing the structure of the news article's HTML page. If the caption and feature vector of the cluster are found to contain keywords in common, then the image is added to an image repository. Additional meta-information are now associated with each image such as caption, cluster features, names of people in the news article, etc. A very large repository containing more than 983k images from 12 million news articles was built using this approach. This repository also contained more than 86.8 million keywords associated with the images. The key contribution of this work is that it combines clustering and natural language processing tasks to automatically create a large corpus of news images with good quality tags or meta-information so that interesting vision tasks can be performed on it. %B Pattern Recognition (ICPR), 2010 20th International Conference on %P 3240 - 3243 %8 2010/08// %G eng %R 10.1109/ICPR.2010.792 %0 Journal Article %J SIGKDD explorations %D 2010 %T On the importance of sharing negative results %A Giraud-Carrier,C. %A Dunham,M.H. %A Atreya,A. %A Elkan,C. %A Perlich,C. %A Swirszcz,G. %A Shi,X. %A Philip,S.Y. %A Fürnkranz,J. %A Sima,J.F. %B SIGKDD explorations %V 12 %P 3 - 4 %8 2010/// %G eng %N 2 %0 Journal Article %J International Journal of Computer Mathematics %D 2010 %T On the initialization methods of an exterior point algorithm for the assignment problem %A Charalampos Papamanthou %A Paparrizos, K. %A Samaras, N. %A Sifaleras, A. %X In this paper, we present a theoretical investigation and an extensive computational study of exterior point simplex algorithm (EPSA) initialization methods for the assignment problem (AP). We describe the exterior point algorithm using three different initialization methods. Effective implementations are explored for each initialization method. Then we perform an experimental evaluation on a large set of benchmark problems from the TSPLib 95 and OR Library collections. The results obtained demonstrate the advantages of the three initialization methods. Finally, we give a theoretical justification of the initialization methods efficiency. We explain theoretically the computational ranking for these methods. %B International Journal of Computer Mathematics %V 87 %P 1831 - 1846 %8 2010/// %@ 0020-7160 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/00207160802524739 %N 8 %0 Report %D 2010 %T Interactive information visualization for exploring and querying electronic health records: A systematic review %A Rind,A. %A Wang,T. D %A Aigner,W. %A Miksh,S. %A Wongsuphasawat,K. %A Plaisant, Catherine %A Shneiderman, Ben %X To overcome the complexity and scale of making medical decisions based on electronic health records (EHRs) a variety of visual methods have been proposed. This paper surveys twelve state-of-the-art information visualization systems described in the scientific literature and compares them based on a set of carefully selected criteria. It aims to systematically examine the systems’ strengths and weaknesses to inform future information visualization designs.We select twelve state-of-the-art information visualization systems from information visualization, medical information systems and human-computer interaction conferences and journals. We compare the systems using the following criteria: (1) data types covered, (2) multivariate analysis support, (3) number of patients records used (one or many), and (4) user intents addressed. The review describes the twelve systems in detail and evaluates the systems using the aforementioned criteria. We discuss how the systems differ in their features and highlight how these differences are related to their design and affect the user intent model. Examples of findings include: (1) most systems handle numerical or categorical data but not both, (2) most systems are specifically designed for looking at a single patient or multiple patients but not both, (3) most systems utilize horizontal time lines to represent time, (4) only systems that handle multiple patient records have good support for Filter, and (5) some specific user intents (e.g. the Encode and Connect intents) are rarely addressed. Based on our review results, we believe that effective information visualization can facilitate analysis of patient records, and we encourage the information visualization community to study the application of their systems and conduct more in depth evaluations. We identify potential future research topics in interactive support for data abstraction and medical tasks that involve looking at a single or multiple records. Finally, we propose to create a repository for data and tasks so benchmarks can be established for both academic and commercial patient record visualization systems. %I Human-Computer Interaction Lab, University of Maryland %8 2010/// %G eng %0 Journal Article %J Natural Language Engineering %D 2010 %T Interlingual Annotation of Parallel Text Corpora: A New Framework for Annotation and Evaluation %A Dorr, Bonnie J %A Passonneau,Rebecca J. %A Farwell,David %A Green,Rebecca %A Habash,Nizar %A Helmreich,Stephen %A Hovy,Eduard %A Levin,Lori %A Miller,Keith J. %A Mitamura,Teruko %A Rambow,Owen %A Siddharthan,Advaith %X This paper focuses on an important step in the creation of a system of meaning representation and the development of semantically annotated parallel corpora, for use in applications such as machine translation, question answering, text summarization, and information retrieval. The work described below constitutes the first effort of any kind to annotate multiple translations of foreign-language texts with interlingual content. Three levels of representation are introduced: deep syntactic dependencies (IL0), intermediate semantic representations (IL1), and a normalized representation that unifies conversives, nonliteral language, and paraphrase (IL2). The resulting annotated, multilingually induced, parallel corpora will be useful as an empirical basis for a wide range of research, including the development and evaluation of interlingual NLP systems and paraphrase-extraction systems as well as a host of other research and development efforts in theoretical and applied linguistics, foreign language pedagogy, translation studies, and other related disciplines. %B Natural Language Engineering %V 16 %P 197 - 243 %8 2010/// %G eng %N 03 %R 10.1017/S1351324910000070 %0 Journal Article %J Human-Computer Interaction Lab 27th Annual Symposium, University of Maryland, College Park, MD %D 2010 %T iOpener Workbench: Tools for rapid understanding of scientific literature %A Dunne,C. %A Shneiderman, Ben %A Dorr, Bonnie J %A Klavans,J. %B Human-Computer Interaction Lab 27th Annual Symposium, University of Maryland, College Park, MD %8 2010/// %G eng %0 Book Section %B Integer Programming and Combinatorial OptimizationInteger Programming and Combinatorial Optimization %D 2010 %T On k-Column Sparse Packing Programs %A Bansal,Nikhil %A Korula,Nitish %A Nagarajan,Viswanath %A Srinivasan, Aravind %E Eisenbrand,Friedrich %E Shepherd,F. %X We consider the class of packing integer programs (PIPs) that are column sparse, where there is a specified upper bound k on the number of constraints that each variable appears in. We give an improved (ek + o(k))-approximation algorithm for k-column sparse PIPs. Our algorithm is based on a linear programming relaxation, and involves randomized rounding combined with alteration. We also show that the integrality gap of our LP relaxation is at least 2k − 1; it is known that even special cases of k-column sparse PIPs are (klogk)-hard to approximate.We generalize our result to the case of maximizing monotone submodular functions over k-column sparse packing constraints, and obtain an e2ke−1+o(k) -approximation algorithm. In obtaining this result, we prove a new property of submodular functions that generalizes the fractionally subadditive property, which might be of independent interest. %B Integer Programming and Combinatorial OptimizationInteger Programming and Combinatorial Optimization %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6080 %P 369 - 382 %8 2010/// %@ 978-3-642-13035-9 %G eng %U http://dx.doi.org/10.1007/978-3-642-13036-6_28 %0 Journal Article %J Advances in Neural Information Processing Systems %D 2010 %T Large-scale matrix factorization with missing data under additional constraints %A Mitra, K. %A Sheorey,S. %A Chellapa, Rama %X Matrix factorization in the presence of missing data is at the core of many com-puter vision problems such as structure from motion (SfM), non-rigid SfM and photometric stereo. We formulate the problem of matrix factorization with miss- ing data as a low-rank semidefinite program (LRSDP) with the advantage that: 1) an efficient quasi-Newton implementation of the LRSDP enables us to solve large-scale factorization problems, and 2) additional constraints such as ortho- normality, required in orthographic SfM, can be directly incorporated in the new formulation. Our empirical evaluations suggest that, under the conditions of ma- trix completion theory, the proposed algorithm finds the optimal solution, and also requires fewer observations compared to the current state-of-the-art algorithms. We further demonstrate the effectiveness of the proposed algorithm in solving the affine SfM problem, non-rigid SfM and photometric stereo problems. %B Advances in Neural Information Processing Systems %V 23 %P 1642 - 1650 %8 2010/// %G eng %0 Journal Article %J IEEE Transactions on Intelligent Transportation Systems %D 2010 %T A Learning Approach Towards Detection and Tracking of Lane Markings %A Gopalan,R. %A Hong, T. %A Shneier, M. %A Chellapa, Rama %K Boosting %K Context %K Context modeling %K Feature extraction %K lane marking detection %K outlier robustness %K Roads %K tracking and learning %K Training %K Vehicles %X Road scene analysis is a challenging problem that has applications in autonomous navigation of vehicles. An integral component of this system is the robust detection and tracking of lane markings. It is a hard problem primarily due to large appearance variations in lane markings caused by factors such as occlusion (traffic on the road), shadows (from objects like trees), and changing lighting conditions of the scene (transition from day to night). In this paper, we address these issues through a learning-based approach using visual inputs from a camera mounted in front of a vehicle. We propose the following: 1) a pixel-hierarchy feature descriptor to model the contextual information shared by lane markings with the surrounding road region; 2) a robust boosting algorithm to select relevant contextual features for detecting lane markings; and 3) particle filters to track the lane markings, without knowledge of vehicle speed, by assuming the lane markings to be static through the video sequence and then learning the possible road scene variations from the statistics of tracked model parameters. We investigate the effectiveness of our algorithm on challenging daylight and night-time road video sequences. %B IEEE Transactions on Intelligent Transportation Systems %V PP %P 1 - 12 %8 2010/02/17/ %@ 1524-9050 %G eng %N 99 %R 10.1109/TITS.2012.2184756 %0 Conference Paper %B Wireless Innovation Conference and Product Exposition, Washington DC, USA %D 2010 %T A Lightweight Dataflow Approach for Design and Implementation of SDR Systems %A Chung-Ching Shen %A Plishker,William %A Wu, Hsiang-Huang %A Bhattacharyya, Shuvra S. %X Model-based design methods based on dataflow modelsof computation are attractive for design and implementation of wireless communication systems because of their intuitive correspondence to communication system block diagrams, and the formal structure that is exposed through formal dataflow representations (e.g., see [2]). In this paper, we introduce a novel lightweight dataflow (LWDF) programming model for model-based design and implementation of wireless communication and software-defined radio systems. The approach is suitable for improving the productivity of the design process; the agility with which designs can be retargeted across different platforms; and the quality of derived implementations. By “lightweight”, we meant that the programming model is designed to be minimally intrusive on existing design processes, and require minimal dependence on specialized tools or libraries. This allows designers to integrate and experiment with dataflow modeling approaches relatively quickly and flexibly into existing design methodologies and processes. %B Wireless Innovation Conference and Product Exposition, Washington DC, USA %8 2010 %G eng %U http://www.researchgate.net/publication/228788399_A_lightweight_dataflow_approach_for_design_and_implementation_of_SDR_systems/file/d912f511472fa25833.pdf %0 Journal Article %J Computer %D 2010 %T Linking Software Development and Business Strategy Through Measurement %A Basili, Victor R. %A Lindvall,M. %A Regardie,M. %A Seaman,C. %A Heidrich,J. %A Munch,J. %A Rombach,D. %A Trendowicz,A. %K approach;business %K aspects;software %K development;organisational %K engineering; %K GQM+Strategies %K program;software %K strategy;enterprise %K support;measurement %K wide %X The GQM+Strategies approach extends the goal/question/metric paradigm for measuring the success or failure of goals and strategies, adding enterprise-wide support for determining action on the basis of measurement results. An organization can thus integrate its measurement program across all levels. %B Computer %V 43 %P 57 - 65 %8 2010/04// %@ 0018-9162 %G eng %N 4 %R 10.1109/MC.2010.108 %0 Journal Article %J Journal of Signal Processing Systems %D 2010 %T A Low-overhead Scheduling Methodology for Fine-grained Acceleration of Signal Processing Systems %A Boutellier, Jani %A Bhattacharyya, Shuvra S. %A Silvén, Olli %X Fine-grained accelerators have the potential to deliver significant benefits in various platforms for embedded signal processing. Due to the moderate complexity of their targeted operations, these accelerators must be managed with minimal run-time overhead. In this paper, we present a methodology for applying flow-shop scheduling techniques to make effective, low-overhead use of fine-grained DSP accelerators. We formulate the underlying scheduling approach in terms of general flow-shop scheduling concepts, and demonstrate our methodology concretely by applying it to MPEG-4 video decoding. We present quantitative experiments on a soft processor that runs on a field-programmable gate array, and provide insight on trends and trade-offs among different flow-shop scheduling approaches when applied to run-time management of fine-grained acceleration. %B Journal of Signal Processing Systems %V 60 %P 333 - 343 %8 2010 %G eng %U http://link.springer.com/article/10.1007/s11265-009-0366-z %N 3 %0 Journal Article %J arXiv:1007.3611 %D 2010 %T LP-rounding algorithms for facility-location problems %A Byrka,Jaroslaw %A Ghodsi,Mohammadreza %A Srinivasan, Aravind %K Computer Science - Data Structures and Algorithms %X We study LP-rounding approximation algorithms for metric uncapacitated facility-location problems. We first give a new analysis for the algorithm of Chudak and Shmoys, which differs from the analysis of Byrka and Aardal in that now we do not need any bound based on the solution to the dual LP program. Besides obtaining the optimal bifactor approximation as do Byrka and Aardal, we can now also show that the algorithm with scaling parameter equaling 1.58 is, in fact, an 1.58-approximation algorithm. More importantly, we suggest an approach based on additional randomization and analyses such as ours, which could achieve or approach the conjectured optimal 1.46...--approximation for this basic problem. Next, using essentially the same techniques, we obtain improved approximation algorithms in the 2-stage stochastic variant of the problem, where we must open a subset of facilities having only stochastic information about the future demand from the clients. For this problem we obtain a 2.2975-approximation algorithm in the standard setting, and a 2.4957-approximation in the more restricted, per-scenario setting. We then study robust fault-tolerant facility location, introduced by Chechik and Peleg: solutions here are designed to provide low connection cost in case of failure of up to $k$ facilities. Chechik and Peleg gave a 6.5-approximation algorithm for $k=1$ and a ($7.5k + 1.5$)-approximation algorithm for general $k$. We improve this to an LP-rounding $(k+5+4/k)$-approximation algorithm. We also observe that in case of oblivious failures the expected approximation ratio can be reduced to $k + 1.5$, and that the integrality gap of the natural LP-relaxation of the problem is at least $k + 1$. %B arXiv:1007.3611 %8 2010/07/21/ %G eng %U http://arxiv.org/abs/1007.3611 %0 Conference Paper %B Proceedings of the 28th international conference on Human factors in computing systems %D 2010 %T ManyNets: an interface for multiple network analysis and visualization %A Freire,Manuel %A Plaisant, Catherine %A Shneiderman, Ben %A Golbeck,Jen %K exploratory analysis %K graphical user interface %K Information Visualization %K interaction %K network analysis %K table interface %X Traditional network analysis tools support analysts in studying a single network. ManyNets offers these analysts a powerful new approach that enables them to work on multiple networks simultaneously. Several thousand networks can be presented as rows in a tabular visualization, and then inspected, sorted and filtered according to their attributes. The networks to be displayed can be obtained by subdivision of larger networks. Examples of meaningful subdivisions used by analysts include ego networks, community extraction, and time-based slices. Cell visualizations and interactive column overviews allow analysts to assess the distribution of attributes within particular sets of networks. Details, such as traditional node-link diagrams, are available on demand. We describe a case study analyzing a social network geared towards film recommendations by means of decomposition. A small usability study provides feedback on the use of the interface on a set of tasks issued from the case study. %B Proceedings of the 28th international conference on Human factors in computing systems %S CHI '10 %I ACM %C New York, NY, USA %P 213 - 222 %8 2010/// %@ 978-1-60558-929-9 %G eng %U http://doi.acm.org/10.1145/1753326.1753358 %R 10.1145/1753326.1753358 %0 Conference Paper %B Proceedings of the 7th USENIX conference on Networked systems design and implementation %D 2010 %T Maranello: practical partial packet recovery for 802.11 %A Han,Bo %A Schulman,Aaron %A Gringoli,Francesco %A Spring, Neil %A Bhattacharjee, Bobby %A Nava,Lorenzo %A Ji,Lusheng %A Lee,Seungjoon %A Miller,Robert %X Partial packet recovery protocols attempt to repair corrupted packets instead of retransmitting them in their entirety. Recent approaches have used physical layer confidence estimates or additional error detection codes embedded in each transmission to identify corrupt bits, or have applied forward error correction to repair without such explicit knowledge. In contrast to these approaches, our goal is a practical design that simultaneously: (a) requires no extra bits in correct packets, (b) reduces recovery latency, except in rare instances, (c) remains compatible with existing 802.11 devices by obeying timing and backoff standards, and (d) can be incrementally deployed on widely available access points and wireless cards. In this paper, we design, implement, and evaluate Maranello, a novel partial packet recovery mechanism for 802.11. In Maranello, the receiver computes checksums over blocks in corrupt packets and bundles these checksums into a negative acknowledgment sent when the sender expects to receive an acknowledgment. The sender then retransmits only those blocks for which the checksum is incorrect, and repeats this partial retransmission until it receives an acknowledgment. Successful transmissions are not burdened by additional bits and the receiver needs not infer which bits were corrupted. We implemented Maranello using OpenFWWF (open source firmware for Broadcom wireless cards) and deployed it in a small testbed. We compare Maranello to alternative recovery protocols using a trace-driven simulation and to 802.11 using a live implementation under various channel conditions. To our knowledge, Maranello is the first partial packet recovery design to be implemented in commonly available firmware. %B Proceedings of the 7th USENIX conference on Networked systems design and implementation %S NSDI'10 %I USENIX Association %C Berkeley, CA, USA %P 14 - 14 %8 2010/// %G eng %U http://dl.acm.org/citation.cfm?id=1855711.1855725 %0 Conference Paper %B 2010 AAAI Fall Symposium Series %D 2010 %T The Metacognitive Loop: An Architecture for Building Robust Intelligent Systems %A Shahri,Hamid Haidarian %A Dinalankara,Wikum %A Fults,Scott %A Wilson,Shomir %A Perlis, Don %A Schmill,Matt %A Oates,Tim %A Josyula,Darsana %A Anderson,Michael %K commonsense %K ontologies %K robust intelligent systems %X The Metacognitive Loop: An Architecture for Building Robust Intelligent Systems %B 2010 AAAI Fall Symposium Series %8 2010/03/11/ %G eng %U http://www.aaai.org/ocs/index.php/FSS/FSS10/paper/view/2161 %0 Journal Article %J Algorithms for Molecular Biology %D 2010 %T Mimosa: Mixture model of co-expression to detect modulators of regulatory interaction %A Hansen,Matthew %A Everett,Logan %A Singh,Larry %A Hannenhalli, Sridhar %X Functionally related genes tend to be correlated in their expression patterns across multiple conditions and/or tissue-types. Thus co-expression networks are often used to investigate functional groups of genes. In particular, when one of the genes is a transcription factor (TF), the co-expression-based interaction is interpreted, with caution, as a direct regulatory interaction. However, any particular TF, and more importantly, any particular regulatory interaction, is likely to be active only in a subset of experimental conditions. Moreover, the subset of expression samples where the regulatory interaction holds may be marked by presence or absence of a modifier gene, such as an enzyme that post-translationally modifies the TF. Such subtlety of regulatory interactions is overlooked when one computes an overall expression correlation. %B Algorithms for Molecular Biology %V 5 %P 4 - 4 %8 2010/01/04/ %@ 1748-7188 %G eng %U http://www.almob.org/content/5/1/4 %N 1 %R 10.1186/1748-7188-5-4 %0 Journal Article %J SIGMOBILE Mob. Comput. Commun. Rev. %D 2010 %T Mobile data offloading in metropolitan area networks %A Han,Bo %A Hui,Pan %A Srinivasan, Aravind %X Cellular networks, especially 3G networks, are currently overloaded with mobile data traffic. It is thus imperative to develop novel communication architectures and protocols for this problem. In this paper, we propose to exploit the delay-tolerant nature of non-realtime applications to offload mobile data traffic through opportunistic communications and WiFi networks. We argue that it is necessary for cellular network operators, WiFi service providers, and end-users to cooperate to make the offloading more effective. We also present several preliminary experiment results about the work we have been conducting, which show that it is promising to offload mobile data traffic in metropolitan area networks. %B SIGMOBILE Mob. Comput. Commun. Rev. %V 14 %P 28 - 30 %8 2010/11// %@ 1559-1662 %G eng %U http://doi.acm.org/10.1145/1942268.1942279 %N 4 %R 10.1145/1942268.1942279 %0 Book Section %B Exponential Data Fitting and Its ApplicationsExponential Data Fitting and Its Applications %D 2010 %T Modelling Type 1a Supernova Light Curves %A Rust,Bert W. %A O'Leary, Dianne P. %A Mullen,Katharine M. %E Pereyra,V. %E Scherer,G. %B Exponential Data Fitting and Its ApplicationsExponential Data Fitting and Its Applications %I Bentham Science Publishers Ltd. %P 169 - 186 %8 2010/// %G eng %U http://www.bentham.org/ebooks/9781608050482/http://www.bentham.org/ebooks/9781608050482/ %0 Conference Paper %B Proceedings of the 2010 Roadmap for Digital Preservation Interoperability Framework Workshop %D 2010 %T Monitoring distributed collections using the Audit Control Environment (ACE) %A Smorul,Michael %A Song,Sangchul %A JaJa, Joseph F. %K digital %K preservation %X The Audit Control Environment (ACE) is a system which provides a scalable, auditable platform that actively monitors collections to ensure their integrity over the lifetime of an archive. It accomplishes this by using a small integrity token issued for each monitored item. This token is part of a larger externally auditable cryptographic system. We will describe how this system has been implemented for a set of applications designed to run in an archive or library environment. ACE has been used for almost two years by the Chronopolis Preservation Environment to monitor the integrity of collections replicated between the three independent archive partners. During this time, ACE has been expanded to better support the requirements of this distributed archive. We will describe how ACE has been used and expanded to support the Chronopolis preservation requirements. We conclude by discussing several future requirements for integrity monitoring that have been identified by users of ACE. These include securely monitoring remote data, monitoring offline data, and scaling monitoring activities in a way that does not impact the normal operational activity of an archive. %B Proceedings of the 2010 Roadmap for Digital Preservation Interoperability Framework Workshop %S US-DPIF '10 %I ACM %C New York, NY, USA %P 13:1–13:5 - 13:1–13:5 %8 2010/// %@ 978-1-4503-0109-1 %G eng %U http://doi.acm.org/10.1145/2039274.2039287 %R 10.1145/2039274.2039287 %0 Conference Paper %B 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) %D 2010 %T Moving vistas: Exploiting motion for describing scenes %A Shroff, N. %A Turaga,P. %A Chellapa, Rama %K Application software %K Automation %K Chaos %K chaotic system %K Computational modeling %K Computer vision %K dynamic scene categorization %K Educational institutions %K HUMANS %K image recognition %K in the wild dynamic scene %K Layout %K motion attribute %K natural scenes %K Physics %K probability %K scene recognition %K Snow %K video data %X Scene recognition in an unconstrained setting is an open and challenging problem with wide applications. In this paper, we study the role of scene dynamics for improved representation of scenes. We subsequently propose dynamic attributes which can be augmented with spatial attributes of a scene for semantically meaningful categorization of dynamic scenes. We further explore accurate and generalizable computational models for characterizing the dynamics of unconstrained scenes. The large intra-class variation due to unconstrained settings and the complex underlying physics present challenging problems in modeling scene dynamics. Motivated by these factors, we propose using the theory of chaotic systems to capture dynamics. Due to the lack of a suitable dataset, we compiled a dataset of `in-the-wild' dynamic scenes. Experimental results show that the proposed framework leads to the best classification rate among other well-known dynamic modeling techniques. We also show how these dynamic features provide a means to describe dynamic scenes with motion-attributes, which then leads to meaningful organization of the video data. %B 2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) %I IEEE %P 1911 - 1918 %8 2010/06/13/18 %@ 978-1-4244-6984-0 %G eng %R 10.1109/CVPR.2010.5539864 %0 Book Section %B Algorithms and theory of computation handbookAlgorithms and theory of computation handbook %D 2010 %T Multidimensional data structures for spatial applications %A Samet, Hanan %E Atallah,Mikhail J. %E Blanton,Marina %X An overview is presented of a number of representations of multidimensional data that arise in spatial applications. Multidimensional spatial data consists of points as well as objects that have extent such as line segments, rectangles, regions, and volumes. The points may have locational as well as nonlocational attributes. The focus is on spatial data which is a subset of multidimensional data consisting of points with locational attributes and objects with extent. The emphasis is on hierarchical representations based on the "divide-and-conquer" problem-solving paradigm. They are of interest because they enable focusing computational resources on the interesting subsets of data. Thus, there is no need to expend work where the payoff is small. These representations are of use in operations such as range searching and finding nearest neighbors. %B Algorithms and theory of computation handbookAlgorithms and theory of computation handbook %I Chapman & Hall/CRC %P 6 - 6 %8 2010/// %@ 978-1-58488-822-2 %G eng %U http://dl.acm.org/citation.cfm?id=1882757.1882763 %0 Conference Paper %B 2010 51st Annual IEEE Symposium on Foundations of Computer Science (FOCS) %D 2010 %T New Constructive Aspects of the Lovasz Local Lemma %A Haeupler,B. %A Saha,B. %A Srinivasan, Aravind %K acyclic edge coloring %K Algorithm design and analysis %K Approximation algorithms %K Approximation methods %K computational complexity %K Computer science %K constant factor approximation algorithm %K graph colouring %K Linearity %K Lovasz Local Lemma %K MAX k-SAT %K Monte Carlo Algorithm %K Monte Carlo methods %K Moser-Tardos algorithm %K nonrepetitive graph coloring %K output distribution %K polynomial sized core subset %K Polynomials %K Probabilistc Method %K probabilistic analysis %K probabilistic logic %K probability %K Ramsey type graph %K Sampling methods %K Santa Claus Problem %X The Lov'{a}sz Local Lemma (LLL) is a powerful tool that gives sufficient conditions for avoiding all of a given set of ``bad'' events, with positive probability. A series of results have provided algorithms to efficiently construct structures whose existence is non-constructively guaranteed by the LLL, culminating in the recent breakthrough of Moser & Tardos. We show that the output distribution of the Moser-Tardos algorithm well-approximates the emph{conditional LLL-distribution} – the distribution obtained by conditioning on all bad events being avoided. We show how a known bound on the probabilities of events in this distribution can be used for further probabilistic analysis and give new constructive and non-constructive results. We also show that when an LLL application provides a small amount of slack, the number of resamplings of the Moser-Tardos algorithm is nearly linear in the number of underlying independent variables (not events!), and can thus be used to give efficient constructions in cases where the underlying proof applies the LLL to super-polynomially many events. Even in cases where finding a bad event that holds is computationally hard, we show that applying the algorithm to avoid a polynomial-sized ``core'' subset of bad events leads to a desired outcome with high probability. We demonstrate this idea on several applications. We give the first constant-factor approximation algorithm for the Santa Claus problem by making an LLL-based proof of Feige constructive. We provide Monte Carlo algorithms for acyclic edge coloring, non-repetitive graph colorings, and Ramsey-type graphs. In all these applications the algorithm falls directly out of the non-constructive LLL-based proof. Our algorithms are very simple, often provide better bounds than previous algorithms, and are in several cases the first efficient algorithms known. As a second type of application we consider settings beyond the critical dependency threshold of the LLL: - - avoiding all bad events is impossible in these cases. As the first (even non-constructive) result of this kind, we show that by sampling from the LLL-distribution of a selected smaller core, we can avoid a fraction of bad events that is higher than the expectation. MAX $k$-SAT is an example of this. %B 2010 51st Annual IEEE Symposium on Foundations of Computer Science (FOCS) %I IEEE %P 397 - 406 %8 2010/10/23/26 %@ 978-1-4244-8525-3 %G eng %R 10.1109/FOCS.2010.45 %0 Journal Article %J Information Visualization %D 2010 %T New forms of Human-Computer Interaction for Visualizing Information %A Reiterer,H. %A Kerren,A. %A Plaisant, Catherine %A Stasko,J.T. %X The Graphical User Interface (GUI) – although developed in research laboratories in the late 1970s – is still the dominant interaction paradigm in Information Visualization. We propose a new interaction paradigm called Blended Interaction. It combines ideas of Embodied Cognition, Multimodal Interaction, Reality-Based Interaction & Ubiquitous Computing. This is intended to stress that a single increase in the reality aspect of the interaction cannot go far enough. The particular challenge – and from the user's standpoint, the key advantage – lies in a meaningful marriage between the tested real-world options and the digital world. As a minimum this marriage must exist on the levels of the interaction, communication, of the way we solve problems with conventional tools (workflows), and of the design of the space or the architecture of buildings and places. The digital world often offers entirely new possibilities and takes the form of interactive devices of various shapes but also of intelligent everyday objects (e.g. the 'Internet of things'). In our view, interaction concepts can indeed offer a new quality of interaction, but only when the design of the interaction includes all these domains at the same time and with equal weighting. We test the suitability of our ideas of Blended Interaction concepts by using specific application examples that are being worked on as part of current research projects. Our experiences show that this new interaction paradigm has also great potential for interacting with visualization. For example, we have developed multi-touch scatter plots & facet maps for tangible user interfaces supporting the searching & browsing in Digital Libraries. We have embedded different visualizations into a Zoomable Object-oriented Information Landscape (ZOIL), which supports our vision of using visualizations on different displays of different size at the same time. We have developed specific kind of search tokens that supports collaborative search activities. For example, we try to address the following research questions: * How can future interactive InfoVis tools look like, especially in the light of the idea Blended Interaction? * How can future interactive InfoVis tools benefit from Multi-Displays & Multimodal environments used by Multiple Users? * What are the specific design requirements for multi-touch visualizations? * How can we support the collaborative use visualization tools? %B Information Visualization %8 2010/// %G eng %0 Conference Paper %D 2010 %T Nfsight: netflow-based network awareness tool %A Berthier,R. %A Michel Cukier %A Hiltunen,M. %A Kormann,D. %A Vesonder,G. %A Sheleheda,D. %X Network awareness is highly critical for network and se-curity administrators. It enables informed planning and management of network resources, as well as detection and a comprehensive understanding of malicious activ- ity. It requires a set of tools to efficiently collect, process, and represent network data. While many such tools al- ready exist, there is no flexible and practical solution for visualizing network activity at various granularities, and quickly gaining insights about the status of network as- sets. To address this issue, we developed Nfsight, a Net- Flow processing and visualization application designed to offer a comprehensive network awareness solution. Nfsight constructs bidirectional flows out of the unidi- rectional NetFlow flows and leverages these bidirectional flows to provide client/server identification and intrusion detection capabilities. We present in this paper the in- ternal architecture of Nfsight, the evaluation of the ser- vice, and intrusion detection algorithms. We illustrate the contributions of Nfsight through several case studies conducted by security administrators on a large univer- sity network. %8 2010/// %G eng %U http://www.usenix.org/event/lisa10/tech/full_papers/Berthier.pdf %0 Conference Paper %D 2010 %T One experience collecting sensitive mobile data %A Niu, Y. %A Elaine Shi %A Chow, R. %A Golle, P. %A Jakobsson, M. %X We report on our efforts to collect behavioral data basedon activities recorded by phones. We recruited Android de- vice owners and offered entry into a raffle for participants. Our application was distributed from the Android Market, and its placement there unexpectedly helped us find par- ticipants from casual users browsing for free applications. We collected data from 267 total participants who gave us varying amounts of data. %8 2010 %G eng %U http://www2.parc.com/csl/members/eshi/docs/users.pdf %0 Journal Article %J IEEE Transactions on Pattern Analysis and Machine Intelligence %D 2010 %T Online Empirical Evaluation of Tracking Algorithms %A Wu,Hao %A Sankaranarayanan,A. C %A Chellapa, Rama %K Back %K Biomedical imaging %K Computer vision %K Filtering %K formal model validation techniques %K formal verification %K ground truth %K Kanade Lucas Tomasi feature tracker %K Karhunen-Loeve transforms %K lighting %K Markov processes %K mean shift tracker %K model validation. %K online empirical evaluation %K particle filtering (numerical methods) %K Particle filters %K Particle tracking %K performance evaluation %K receiver operating characteristic curves %K Robustness %K SNR %K Statistics %K Surveillance %K time reversed Markov chain %K tracking %K tracking algorithms %K visual tracking %X Evaluation of tracking algorithms in the absence of ground truth is a challenging problem. There exist a variety of approaches for this problem, ranging from formal model validation techniques to heuristics that look for mismatches between track properties and the observed data. However, few of these methods scale up to the task of visual tracking, where the models are usually nonlinear and complex and typically lie in a high-dimensional space. Further, scenarios that cause track failures and/or poor tracking performance are also quite diverse for the visual tracking problem. In this paper, we propose an online performance evaluation strategy for tracking systems based on particle filters using a time-reversed Markov chain. The key intuition of our proposed methodology relies on the time-reversible nature of physical motion exhibited by most objects, which in turn should be possessed by a good tracker. In the presence of tracking failures due to occlusion, low SNR, or modeling errors, this reversible nature of the tracker is violated. We use this property for detection of track failures. To evaluate the performance of the tracker at time instant t, we use the posterior of the tracking algorithm to initialize a time-reversed Markov chain. We compute the posterior density of track parameters at the starting time t = 0 by filtering back in time to the initial time instant. The distance between the posterior density of the time-reversed chain (at t = 0) and the prior density used to initialize the tracking algorithm forms the decision statistic for evaluation. It is observed that when the data are generated by the underlying models, the decision statistic takes a low value. We provide a thorough experimental analysis of the evaluation methodology. Specifically, we demonstrate the effectiveness of our approach for tackling common challenges such as occlusion, pose, and illumination changes and provide the Receiver Operating Characteristic (ROC) curves. Finally, we also s how the applicability of the core ideas of the paper to other tracking algorithms such as the Kanade-Lucas-Tomasi (KLT) feature tracker and the mean-shift tracker. %B IEEE Transactions on Pattern Analysis and Machine Intelligence %V 32 %P 1443 - 1458 %8 2010/08// %@ 0162-8828 %G eng %N 8 %R 10.1109/TPAMI.2009.135 %0 Conference Paper %B Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems %D 2010 %T Ontuition: intuitive data exploration via ontology navigation %A Adelfio,Marco D. %A Lieberman,Michael D. %A Samet, Hanan %A Firozvi,Kashif A. %K MAPPING %K ontology %K ontuition %K spatio-textual %X Ontuition, a system for mapping ontologies, is presented. Transforming data to a usable format for Ontuition involves recognizing and resolving data values corresponding to concepts in multiple ontological domains. In particular, for datasets with a geographic component an attempt is made to identify and extract enough spatio-textual data that specific lat/long values to dataset entries can be assigned. Next, a gazetteer is used to transform the textually-specified locations into lat/long values that can be displayed on a map. Non-spatial ontological concepts are also discovered. This methodology is applied to the National Library of Medicine's very popular clinical trials website (http://clinicaltrials.gov/) whose users are generally interested in locating trials near where they live. The trials are specified using XML files. The location data is extracted and coupled with a disease ontology to enable general queries on the data with the result being of use to a very large group of people. The goal is to do this automatically for such ontology datasets with a locational component. %B Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems %S GIS '10 %I ACM %C New York, NY, USA %P 540 - 541 %8 2010/// %@ 978-1-4503-0428-3 %G eng %U http://doi.acm.org/10.1145/1869790.1869887 %R 10.1145/1869790.1869887 %0 Journal Article %J Pattern Analysis and Machine Intelligence, IEEE Transactions on %D 2010 %T PADS: A Probabilistic Activity Detection Framework for Video Data %A Albanese, M. %A Chellapa, Rama %A Cuntoor, N. %A Moscato, V. %A Picariello, A. %A V.S. Subrahmanian %A Udrea,O. %K Automated;Programming Languages;Video Recording; %K Computer-Assisted;Models %K PADL;PADS;image processing algorithms;offPad algorithm;onPad algorithm;probabilistic activity description language;probabilistic activity detection framework;video data;video sequence;image sequences;object detection;probability;video surveillance;Algorit %K Statistical;Movement;Pattern Recognition %X There is now a growing need to identify various kinds of activities that occur in videos. In this paper, we first present a logical language called Probabilistic Activity Description Language (PADL) in which users can specify activities of interest. We then develop a probabilistic framework which assigns to any subvideo of a given video sequence a probability that the subvideo contains the given activity, and we finally develop two fast algorithms to detect activities within this framework. OffPad finds all minimal segments of a video that contain a given activity with a probability exceeding a given threshold. In contrast, the OnPad algorithm examines a video during playout (rather than afterwards as OffPad does) and computes the probability that a given activity is occurring (even if the activity is only partially complete). Our prototype Probabilistic Activity Detection System (PADS) implements the framework and the two algorithms, building on top of existing image processing algorithms. We have conducted detailed experiments and compared our approach to four different approaches presented in the literature. We show that-for complex activity definitions-our approach outperforms all the other approaches. %B Pattern Analysis and Machine Intelligence, IEEE Transactions on %V 32 %P 2246 - 2261 %8 2010/12// %@ 0162-8828 %G eng %N 12 %R 10.1109/TPAMI.2010.33 %0 Journal Article %J Journal of cryptology %D 2010 %T Parallel and concurrent security of the HB and HB+ protocols %A Katz, Jonathan %A Shin,J. S %A Smith,A. %X Hopper and Blum (Asiacrypt 2001) and Juels and Weis (Crypto 2005) recently proposed two shared-key authentication protocols—HB and HB+, respectively—whose extremely low computational cost makes them attractive for low-cost devices such as radio-frequency identification (RFID) tags. The security of these protocols is based on the conjectured hardness of the “learning parity with noise” (LPN) problem, which is equivalent to the problem of decoding random binary linear codes. The HB protocol is proven secure against a passive (eavesdropping) adversary, while the HB+ protocol is proven secure against active attacks. %B Journal of cryptology %V 23 %P 402 - 421 %8 2010/// %G eng %N 3 %R 10.1007/s00145-010-9061-2 %0 Journal Article %J Technical Reports of the Computer Science Department %D 2010 %T Partial least squares on graphical processor for efficient pattern recognition %A Srinivasan,Balaji Vasan %A Schwartz,William Robson %A Duraiswami, Ramani %A Davis, Larry S. %K Technical Report %X Partial least squares (PLS) methods have recently been used for many patternrecognition problems in computer vision. Here, PLS is primarily used as a supervised dimensionality reduction tool to obtain effective feature combinations for better learning. However, application of PLS to large datasets is hindered by its higher computational cost. We propose an approach to accelerate the classical PLS algorithm on graphical processors to obtain the same performance at a reduced cost. Although, PLS modeling is practically an offline training process, accelerating it helps large scale modeling. The proposed acceleration is shown to perform well and it yields upto ~30X speedup, It is applied on standard datasets in human detection and face recognition. %B Technical Reports of the Computer Science Department %8 2010/10/18/ %G eng %U http://drum.lib.umd.edu/handle/1903/10975 %0 Conference Paper %B International Conference on Pattern Recognition %D 2010 %T Performance Evaluation Tools for Zone Segmentation and Classification (PETS) %A Seo,W. %A Agrawal,Mudit %A David Doermann %X This paper overviews a set of Performance Evaluation ToolS (PETS) for zone segmentation and classification. The tools allow researchers and developers to evaluate, optimize and compare their algorithms by providing a variety of quantitative performance metrics. The evaluation of segmentation quality is based on the pixel-based overlaps between two sets of regions proposed by Randriamasy and Vincent. PETS extends the approach by providing a set of metrics for overlap analysis, RLE and polygonal representation of regions and introduces type-matching to evaluate zone classification. The software is available for research use. %B International Conference on Pattern Recognition %P 503 - 506 %8 2010/// %G eng %0 Journal Article %J PLoS ONEPLoS ONE %D 2010 %T Perturbing the Ubiquitin Pathway Reveals How Mitosis Is Hijacked to Denucleate and Regulate Cell Proliferation and Differentiation In Vivo %A Caceres,Andrea %A Shang,Fu %A Wawrousek,Eric %A Liu,Qing %A Avidan,Orna %A Cvekl,Ales %A Yang,Ying %A Haririnia,Aydin %A Storaska,Andrew %A Fushman, David %A Kuszak,Jer %A Dudek,Edward %A Smith,Donald %A Taylor,Allen %X BackgroundThe eye lens presents a unique opportunity to explore roles for specific molecules in cell proliferation, differentiation and development because cells remain in place throughout life and, like red blood cells and keratinocytes, they go through the most extreme differentiation, including removal of nuclei and cessation of protein synthesis. Ubiquitination controls many critical cellular processes, most of which require specific lysines on ubiquitin (Ub). Of the 7 lysines (K) least is known about effects of modification of K6. Methodology and Principal Findings We replaced K6 with tryptophan (W) because K6 is the most readily modified K and W is the most structurally similar residue to biotin. The backbone of K6W-Ub is indistinguishable from that of Wt-Ub. K6W-Ub is effectively conjugated and deconjugated but the conjugates are not degraded via the ubiquitin proteasome pathways (UPP). Expression of K6W-ubiquitin in the lens and lens cells results in accumulation of intracellular aggregates and also slows cell proliferation and the differentiation program, including expression of lens specific proteins, differentiation of epithelial cells into fibers, achieving proper fiber cell morphology, and removal of nuclei. The latter is critical for transparency, but the mechanism by which cell nuclei are removed has remained an age old enigma. This was also solved by expressing K6W-Ub. p27kip, a UPP substrate accumulates in lenses which express K6W-Ub. This precludes phosphorylation of nuclear lamin by the mitotic kinase, a prerequisite for disassembly of the nuclear membrane. Thus the nucleus remains intact and DNAseIIβ neither gains entry to the nucleus nor degrades the DNA. These results could not be obtained using chemical proteasome inhibitors that cannot be directed to specific tissues. Conclusions and Significance K6W-Ub provides a novel, genetic means to study functions of the UPP because it can be targeted to specific cells and tissues. A fully functional UPP is required to execute most stages of lens differentiation, specifically removal of cell nuclei. In the absence of a functional UPP, small aggregate prone, cataractous lenses are formed. %B PLoS ONEPLoS ONE %V 5 %P e13331 - e13331 %8 2010/10/20/ %G eng %U http://dx.doi.org/10.1371/journal.pone.0013331 %N 10 %R 10.1371/journal.pone.0013331 %0 Patent %D 2010 %T Plasmonic Systems and Devices Utilizing Surface Plasmon Polariton %A Smolyaninov,Igor I. %A Vishkin, Uzi %A Davis,Christopher C. %X Plasmonic systems and devices that utilize surface plasmon polaritons (or “plasmons”) for inter-chip and/or intra-chip communications are provided. A plasmonic system includes a microchip that has an integrated circuit module and a plasmonic device configured to interface with the integrated circuit module. The plasmonic device includes a first electrode, a second electrode positioned at a non-contact distance from the first electrode, and a tunneling-junction configured to create a plasmon when a potential difference is created between the first electrode and the second electrode. %V 12/697,595 %8 2010/05/27/ %G eng %U http://www.google.com/patents?id=2VnRAAAAEBAJ %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 2010 %T Query Processing Using Distance Oracles for Spatial Networks %A Sankaranarayanan,Jagan %A Samet, Hanan %X The popularity of location-based services and the need to do real-time processing on them has led to an interest in performing queries on transportation networks, such as finding shortest paths and finding nearest neighbors. The challenge here is that the efficient execution of spatial operations usually involves the computation of distance along a spatial network instead of "as the crow flies, #x0201D; which is not simple. Techniques are described that enable the determination of the network distance between any pair of points (i.e., vertices) with as little as O(n) space rather than having to store the n^2 distances between all pairs. This is done by being willing to expend a bit more time to achieve this goal such as O(log n) instead of O(1), as well as by accepting an error varepsilon in the accuracy of the distance that is provided. The strategy that is adopted reduces the space requirements and is based on the ability to identify groups of source and destination vertices for which the distance is approximately the same within some varepsilon. The reductions are achieved by introducing a construct termed a distance oracle that yields an estimate of the network distance (termed the varepsilon-approximate distance) between any two vertices in the spatial network. The distance oracle is obtained by showing how to adapt the well-separated pair technique from computational geometry to spatial networks. Initially, an varepsilon-approximate distance oracle of size O(nover varepsilon^2 ) is used that is capable of retrieving the approximate network distance in O(log n) time using a B-tree. The retrieval time can be theoretically reduced further to O(1) time by proposing another varepsilon-approximate distance oracle of size O(n log nover varepsilon^2 ) that uses a hash table. Experimental results indicate that the proposed technique is scalable and can be applied to sufficiently large road networks. For example, a 10-percent-approximate oracle (vareps- - ilon = 0.1) on a large network yielded an average error of 0.9 percent with 90 percent of the answers having an error of 2 percent or less and an average retrieval time of 68 mu rm seconds. The fact that the network distance can be approximated by one value is used to show how a number of spatial queries can be formulated using appropriate SQL constructs and a few built-in primitives. The result is that these operations can be executed on almost any modern database with no modifications, while taking advantage of the existing query optimizers and query processing strategies. %B Knowledge and Data Engineering, IEEE Transactions on %V 22 %P 1158 - 1175 %8 2010/08// %@ 1041-4347 %G eng %N 8 %R 10.1109/TKDE.2010.75 %0 Journal Article %J Technical Reports of the Computer Science Department %D 2010 %T Random sampling for estimating the performance of fast summations %A Srinivasan,Balaji Vasan %A Duraiswami, Ramani %K Technical Report %X Summation of functions of N source points evaluated at M target pointsoccurs commonly in many applications. To scale these approaches for large datasets, many fast algorithms have been proposed. In this technical report, we propose a Chernoff bound based efficient approach to test the performance of a fast summation algorithms providing a probabilistic accuracy. We further validate and use our approach in separate comparisons. %B Technical Reports of the Computer Science Department %8 2010/10/18/ %G eng %U http://drum.lib.umd.edu/handle/1903/10976 %0 Conference Paper %B 2010 21st IEEE International Symposium on Rapid System Prototyping (RSP) %D 2010 %T Rapid prototyping for digital signal processing systems using Parameterized Synchronous Dataflow graphs %A Wu, Hsiang-Huang %A Kee, Hojin %A Sane, N. %A Plishker,W. %A Bhattacharyya, Shuvra S. %K abstract scheduling %K Computational modeling %K Computer architecture %K data flow graphs %K dataflow based design %K dataflow interchange format %K design flow %K design language %K Digital signal processing %K digital signal processing systems %K dynamic parameter reconfiguration %K Dynamic scheduling %K efficient hardware mapping %K efficient quasistatic scheduling %K Embedded software %K embedded systems %K Field programmable gate arrays %K flexible dynamic reconfiguration %K FPGA based systems %K FPGA implementations %K functional simulation %K Hardware %K parameterized synchronous dataflow graphs %K rapid prototyping %K Schedules %K scheduling %K semantics %K simulation tool %K software package %K systematic design methodology %X Parameterized Synchronous Dataflow (PSDF) has been used previously for abstract scheduling and as a model for architecting embedded software and FPGA implementations. PSDF has been shown to be attractive for these purposes due to its support for flexible dynamic reconfiguration, and efficient quasi-static scheduling. To apply PSDF techniques more deeply into the design flow, support for comprehensive functional simulation and efficient hardware mapping is important. By building on the DIF (Dataflow Interchange Format), which is a design language and associated software package for developing and experimenting with dataflow-based design techniques for signal processing systems, we have developed a tool for functional simulation of PSDF specifications. This simulation tool allows designers to model applications in PSDF and simulate their functionality, including use of the dynamic parameter reconfiguration capabilities offered by PSDF. Based on this simulation tool, we also present a systematic design methodology for applying PSDF to the design and implementation of digital signal processing systems, with emphasis on FPGA-based systems for signal processing. We demonstrate capabilities for rapid and accurate prototyping offered by our proposed design methodology, along with its novel support for PSDF-based FPGA system implementation. %B 2010 21st IEEE International Symposium on Rapid System Prototyping (RSP) %P 1 - 7 %8 2010 %G eng %0 Journal Article %J Proc. VLDB Endow. %D 2010 %T Read-once functions and query evaluation in probabilistic databases %A Sen,Prithviraj %A Deshpande, Amol %A Getoor, Lise %X Probabilistic databases hold promise of being a viable means for large-scale uncertainty management, increasingly needed in a number of real world applications domains. However, query evaluation in probabilistic databases remains a computational challenge. Prior work on efficient exact query evaluation in probabilistic databases has largely concentrated on query-centric formulations (e.g., safe plans, hierarchical queries), in that, they only consider characteristics of the query and not the data in the database. It is easy to construct examples where a supposedly hard query run on an appropriate database gives rise to a tractable query evaluation problem. In this paper, we develop efficient query evaluation techniques that leverage characteristics of both the query and the data in the database. We focus on tuple-independent databases where the query evaluation problem is equivalent to computing marginal probabilities of Boolean formulas associated with the result tuples. This latter task is easy if the Boolean formulas can be factorized into a form that has every variable appearing at most once (called read-once). However, a naive approach that directly uses previously developed Boolean formula factorization algorithms is inefficient, because those algorithms require the input formulas to be in the disjunctive normal form (DNF). We instead develop novel, more efficient factorization algorithms that directly construct the read-once expression for a result tuple Boolean formula (if one exists), for a large subclass of queries (specifically, conjunctive queries without self-joins). We empirically demonstrate that (1) our proposed techniques are orders of magnitude faster than generic inference algorithms for queries where the result Boolean formulas can be factorized into read-once expressions, and (2) for the special case of hierarchical queries, they rival the efficiency of prior techniques specifically designed to handle such queries. %B Proc. VLDB Endow. %V 3 %P 1068 - 1079 %8 2010/09// %@ 2150-8097 %G eng %U http://dl.acm.org/citation.cfm?id=1920841.1920975 %N 1-2 %0 Journal Article %J AI Magazine %D 2010 %T REPORTS AI Theory and Practice: A Discussion on Hard Challenges and Opportunities Ahead %A Horvitz,E. %A Getoor, Lise %A Guestrin,C. %A Hendler,J. %A Kautz,H. %A Konstan,J. %A Subramanian,D. %A Wellman,M. %B AI Magazine %V 31 %P 87 - 87 %8 2010/// %G eng %N 3 %0 Conference Paper %B PLDI '10 Proceedings of the 2010 ACM SIGPLAN Conference on Programming Language Design and Implementation %D 2010 %T Resolving and Exploiting the k-CFA Paradox: Illuminating Functional vs. Object-oriented Program Analysis %A Might, Matthew %A Smaragdakis, Yannis %A David Van Horn %K control-flow analysis %K functional %K k-cfa %K m-cfa %K object-oriented %K pointer analysis %K static analysis %X Low-level program analysis is a fundamental problem, taking the shape of "flow analysis" in functional languages and "points-to" analysis in imperative and object-oriented languages. Despite the similarities, the vocabulary and results in the two communities remain largely distinct, with limited cross-understanding. One of the few links is Shivers's k-CFA work, which has advanced the concept of "context-sensitive analysis" and is widely known in both communities. Recent results indicate that the relationship between the functional and object-oriented incarnations of k-CFA is not as well understood as thought. Van Horn and Mairson proved k-CFA for k ≥ 1 to be EXPTIME-complete; hence, no polynomial-time algorithm can exist. Yet, there are several polynomial-time formulations of context-sensitive points-to analyses in object-oriented languages. Thus, it seems that functional k-CFA may actually be a profoundly different analysis from object-oriented k-CFA. We resolve this paradox by showing that the exact same specification of k-CFA is polynomial-time for object-oriented languages yet exponential-time for functional ones: objects and closures are subtly different, in a way that interacts crucially with context-sensitivity and complexity. This illumination leads to an immediate payoff: by projecting the object-oriented treatment of objects onto closures, we derive a polynomial-time hierarchy of context-sensitive CFAs for functional programs. %B PLDI '10 Proceedings of the 2010 ACM SIGPLAN Conference on Programming Language Design and Implementation %S PLDI '10 %I ACM %P 305 - 315 %8 2010/// %@ 978-1-4503-0019-3 %G eng %U http://doi.acm.org/10.1145/1806596.1806631 %0 Journal Article %J Biochemistry %D 2010 %T Reversible Post-Translational Carboxylation Modulates the Enzymatic Activity of N-Acetyl-l-ornithine Transcarbamylase %A Li,Yongdong %A Yu,Xiaolin %A Ho,Jeremy %A Fushman, David %A Allewell,Norma M. %A Tuchman,Mendel %A Shi,Dashuang %X N-Acetyl-l-ornithine transcarbamylase (AOTCase), rather than ornithine transcarbamylase (OTCase), is the essential carbamylase enzyme in the arginine biosynthesis of several plant and human pathogens. The specificity of this unique enzyme provides a potential target for controlling the spread of these pathogens. Recently, several crystal structures of AOTCase from Xanthomonas campestris (xc) have been determined. In these structures, an unexplained electron density at the tip of the Lys302 side chain was observed. Using 13C NMR spectroscopy, we show herein that Lys302 is post-translationally carboxylated. The structure of wild-type AOTCase in a complex with the bisubstrate analogue Nδ-(phosphonoacetyl)-Nα-acetyl-l-ornithine (PALAO) indicates that the carboxyl group on Lys302 forms a strong hydrogen bonding network with surrounding active site residues, Lys252, Ser253, His293, and Glu92 from the adjacent subunit either directly or via a water molecule. Furthermore, the carboxyl group is involved in binding N-acetyl-l-ornithine via a water molecule. Activity assays with the wild-type enzyme and several mutants demonstrate that the post-translational modification of lysine 302 has an important role in catalysis.N-Acetyl-l-ornithine transcarbamylase (AOTCase), rather than ornithine transcarbamylase (OTCase), is the essential carbamylase enzyme in the arginine biosynthesis of several plant and human pathogens. The specificity of this unique enzyme provides a potential target for controlling the spread of these pathogens. Recently, several crystal structures of AOTCase from Xanthomonas campestris (xc) have been determined. In these structures, an unexplained electron density at the tip of the Lys302 side chain was observed. Using 13C NMR spectroscopy, we show herein that Lys302 is post-translationally carboxylated. The structure of wild-type AOTCase in a complex with the bisubstrate analogue Nδ-(phosphonoacetyl)-Nα-acetyl-l-ornithine (PALAO) indicates that the carboxyl group on Lys302 forms a strong hydrogen bonding network with surrounding active site residues, Lys252, Ser253, His293, and Glu92 from the adjacent subunit either directly or via a water molecule. Furthermore, the carboxyl group is involved in binding N-acetyl-l-ornithine via a water molecule. Activity assays with the wild-type enzyme and several mutants demonstrate that the post-translational modification of lysine 302 has an important role in catalysis. %B Biochemistry %V 49 %P 6887 - 6895 %8 2010/// %@ 0006-2960 %G eng %U http://dx.doi.org/10.1021/bi1007386 %N 32 %R 10.1021/bi1007386 %0 Journal Article %J Data Engineering %D 2010 %T Roads Belong in Databases %A Samet, Hanan %B Data Engineering %P 4 - 4 %8 2010/// %G eng %0 Journal Article %J Computer Vision–ECCV 2010 %D 2010 %T A robust and scalable approach to face identification %A Schwartz,W. %A Guo,H. %A Davis, Larry S. %X The problem of face identification has received significant attention over the years. For a given probe face, the goal of face identification is to match this unknown face against a gallery of known people. Due to the availability of large amounts of data acquired in a variety of conditions, techniques that are both robust to uncontrolled acquisition conditions and scalable to large gallery sizes, which may need to be incrementally built, are challenges. In this work we tackle both problems. Initially, we propose a novel approach to robust face identification based on Partial Least Squares (PLS) to perform multi-channel feature weighting. Then, we extend the method to a tree-based discriminative structure aiming at reducing the time required to evaluate novel probe samples. The method is evaluated through experiments on FERET and FRGC datasets. In most of the comparisons our method outperforms state-of-art face identification techniques. Furthermore, our method presents scalability to large datasets. %B Computer Vision–ECCV 2010 %P 476 - 489 %8 2010/// %G eng %0 Journal Article %J IEEE Transactions on Image Processing %D 2010 %T Robust Height Estimation of Moving Objects From Uncalibrated Videos %A Jie Shao %A Zhou,S. K %A Chellapa, Rama %K algorithms %K Biometry %K Calibration %K EM algorithm %K geometric properties %K Geometry %K Image Enhancement %K Image Interpretation, Computer-Assisted %K Imaging, Three-Dimensional %K least median of squares %K least squares approximations %K MOTION %K motion information %K multiframe measurements %K Pattern Recognition, Automated %K Reproducibility of results %K Robbins-Monro stochastic approximation %K robust height estimation %K Sensitivity and Specificity %K Signal Processing, Computer-Assisted %K stochastic approximation %K Subtraction Technique %K tracking data %K uncalibrated stationary camera %K uncalibrated videos %K uncertainty analysis %K vanishing point %K video metrology %K Video Recording %K video signal processing %X This paper presents an approach for video metrology. From videos acquired by an uncalibrated stationary camera, we first recover the vanishing line and the vertical point of the scene based upon tracking moving objects that primarily lie on a ground plane. Using geometric properties of moving objects, a probabilistic model is constructed for simultaneously grouping trajectories and estimating vanishing points. Then we apply a single view mensuration algorithm to each of the frames to obtain height measurements. We finally fuse the multiframe measurements using the least median of squares (LMedS) as a robust cost function and the Robbins-Monro stochastic approximation (RMSA) technique. This method enables less human supervision, more flexibility and improved robustness. From the uncertainty analysis, we conclude that the method with auto-calibration is robust in practice. Results are shown based upon realistic tracking data from a variety of scenes. %B IEEE Transactions on Image Processing %V 19 %P 2221 - 2232 %8 2010/08// %@ 1057-7149 %G eng %N 8 %R 10.1109/TIP.2010.2046368 %0 Conference Paper %B 2010 IEEE Workshop on Signal Processing Systems (SIPS) %D 2010 %T Scalable representation of dataflow graph structures using topological patterns %A Sane, N. %A Kee, Hojin %A Seetharaman, G. %A Bhattacharyya, Shuvra S. %K arrays %K data flow graphs %K Dataflow graphs %K DIF language %K Digital signal processing %K directed graphs %K DSP specification languages %K embedded signal processing design flows %K embedded systems %K Field programmable gate arrays %K graphical user interface %K Graphical user interfaces %K High-level languages %K modelbased design %K optimisation %K optimizations %K scalable dataflow graph structures representation %K semantics %K Signal processing %K Signal processing systems %K Specification languages %K text based languages %K Topological patterns %K Topology %K Transform coding %X Tools for designing signal processing systems with their semantic foundation in dataflow modeling often use high-level graphical user interface (GUI) or text based languages that allow specifying applications as directed graphs. Such graphical representations serve as an initial reference point for further analysis and optimizations that lead to platform-specific implementations. For large-scale applications, the underlying graphs often consist of smaller substructures that repeat multiple times. To enable more concise representation and direct analysis of such substructures in the context of high level DSP specification languages and design tools, we develop the modeling concept of topological patterns, and propose ways for supporting this concept in a high-level language. We augment the DIF language - a language for specifying DSP-oriented dataflow graphs - with constructs for supporting topological patterns, and we show how topological patterns can be effective in various aspects of embedded signal processing design flows using specific application examples. %B 2010 IEEE Workshop on Signal Processing Systems (SIPS) %P 13 - 18 %8 2010 %G eng %0 Book Section %B Video Search and MiningVideo Search and Mining %D 2010 %T Semantic Video Content Analysis %A Albanese,Massimiliano %A Turaga,Pavan %A Chellapa, Rama %A Pugliese,Andrea %A Subrahmanian,V. %E Schonfeld,Dan %E Shan,Caifeng %E Tao,Dacheng %E Wang,Liang %X In recent years, there has been significant interest in the area of automatically recognizing activities occurring in a camera’s field of view and detecting abnormalities. The practical applications of such a system could include airport tarmac monitoring, or monitoring of activities in secure installations, to name a few. The difficulty of the problem is compounded by several factors: detection of primitive actions in spite of changes in illumination, occlusions and noise; complexmultiagent interaction;mapping of higher-level activities to lower-level primitive actions; variations in which the same semantic activity can be performed. In this chapter, we develop a theory of semantic activity analysis that addresses each of these issues in an integrated manner. Specifically, we discuss ontological representations of knowledge of a domain, integration of domain knowledge and statistical models for achieving semantic mappings, definition of logical languages to describe activities, and design of frameworks which integrate all the above aspects in a coherent way, thus laying the foundations of effective Semantic Video Content Analysis systems. %B Video Search and MiningVideo Search and Mining %S Studies in Computational Intelligence %I Springer Berlin / Heidelberg %V 287 %P 147 - 176 %8 2010/// %@ 978-3-642-12899-8 %G eng %U http://dx.doi.org/10.1007/978-3-642-12900-1_6 %0 Book Section %B Financial Cryptography and Data Security %D 2010 %T Signatures of Reputation %A Bethencourt, John %A Elaine Shi %A Song, Dawn %E Sion, Radu %K Computer science %B Financial Cryptography and Data Security %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6052 %P 400 - 407 %8 2010 %@ 978-3-642-14576-6 %G eng %U http://www.springerlink.com/content/8484428n77226512/abstract/ %0 Conference Paper %B 2010 IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP) %D 2010 %T Simulating dynamic communication systems using the core functional dataflow model %A Sane, N. %A Chia-Jui Hsu %A Pino,J. L %A Bhattacharyya, Shuvra S. %K adaptive modulation %K Analytical models %K Application software %K Computational modeling %K core functional dataflow model %K Dataflow %K dataflow modeling semantics %K design tools %K Digital signal processing %K dynamic communication systems %K functional specification %K Hardware %K modeling and simulation %K Power system modeling %K Predictive models %K Processor scheduling %K Production %K Signal processing %K software tools %K wireless communication %X The latest communication technologies invariably consist of modules with dynamic behavior. There exists a number of design tools for communication system design with their foundation in dataflow modeling semantics. These tools must not only support the functional specification of dynamic communication modules and subsystems but also provide accurate estimation of resource requirements for efficient simulation and implementation. We explore this trade-off - between flexible specification of dynamic behavior and accurate estimation of resource requirements - using a representative application employing an adaptive modulation scheme. We propose an approach for precise modeling of such applications based on a recently-introduced form of dynamic dataflow called core functional dataflow. From our proposed modeling approach, we show how parameterized looped schedules can be generated and analyzed to simulate applications with low run-time overhead as well as guaranteed bounded memory execution. We demonstrate our approach using the Advanced Design System from Agilent Technologies, Inc., which is a commercial tool for design and simulation of communication systems. %B 2010 IEEE International Conference on Acoustics Speech and Signal Processing (ICASSP) %P 1538 - 1541 %8 2010 %G eng %0 Journal Article %J BMC Bioinformatics %D 2010 %T Sites Inferred by Metabolic Background Assertion Labeling (SIMBAL): adapting the Partial Phylogenetic Profiling algorithm to scan sequences for signatures that predict protein function. %A Jeremy D Selengut %A Rusch, Douglas B %A Haft, Daniel H %K algorithms %K Amino Acid Sequence %K Gene Expression Profiling %K Molecular Sequence Data %K Phylogeny %K Proteins %K Sequence Analysis, Protein %K Structure-Activity Relationship %X

BACKGROUND: Comparative genomics methods such as phylogenetic profiling can mine powerful inferences from inherently noisy biological data sets. We introduce Sites Inferred by Metabolic Background Assertion Labeling (SIMBAL), a method that applies the Partial Phylogenetic Profiling (PPP) approach locally within a protein sequence to discover short sequence signatures associated with functional sites. The approach is based on the basic scoring mechanism employed by PPP, namely the use of binomial distribution statistics to optimize sequence similarity cutoffs during searches of partitioned training sets.

RESULTS: Here we illustrate and validate the ability of the SIMBAL method to find functionally relevant short sequence signatures by application to two well-characterized protein families. In the first example, we partitioned a family of ABC permeases using a metabolic background property (urea utilization). Thus, the TRUE set for this family comprised members whose genome of origin encoded a urea utilization system. By moving a sliding window across the sequence of a permease, and searching each subsequence in turn against the full set of partitioned proteins, the method found which local sequence signatures best correlated with the urea utilization trait. Mapping of SIMBAL "hot spots" onto crystal structures of homologous permeases reveals that the significant sites are gating determinants on the cytosolic face rather than, say, docking sites for the substrate-binding protein on the extracellular face. In the second example, we partitioned a protein methyltransferase family using gene proximity as a criterion. In this case, the TRUE set comprised those methyltransferases encoded near the gene for the substrate RF-1. SIMBAL identifies sequence regions that map onto the substrate-binding interface while ignoring regions involved in the methyltransferase reaction mechanism in general. Neither method for training set construction requires any prior experimental characterization.

CONCLUSIONS: SIMBAL shows that, in functionally divergent protein families, selected short sequences often significantly outperform their full-length parent sequence for making functional predictions by sequence similarity, suggesting avenues for improved functional classifiers. When combined with structural data, SIMBAL affords the ability to localize and model functional sites.

%B BMC Bioinformatics %V 11 %P 52 %8 2010 %G eng %R 10.1186/1471-2105-11-52 %0 Conference Paper %B ACM SIGGRAPH ASIA 2010 Courses %D 2010 %T Sorting in space: multidimensional, spatial, and metric data structures for computer graphics applications %A Samet, Hanan %X The representation of spatial data is an important issue in game programming, computer graphics, visualization, solid modeling, and related areas including computer vision and geographic information systems (GIS). Many representations are currently used. Recently, there has been much interest in hierarchical representations such as quadtrees, octrees, and pyramids which are based on image hierarchies, as well methods that use bounding boxes which are based on object hierarchies. The key advantage of these representations is that they provide a way to index into space. In fact, they are little more than multidimensional sorts. They are compact and depending on the nature of the spatial data they save space as well as time and also facilitate operations such as search. This course provides a brief overview of hierarchical spatial data structures and related algorithms that make use of them. We describe hierarchical representations of points, lines, collections of small rectangles, regions, surfaces, and volumes. For region data, we point out the dimension-reduction property of the region quadtree and octree, as how to navigate between nodes in the same tree, thereby leading to the popularity of these representations in ray tracing applications. We also demonstrate how to use these representations for both raster and vector data. In the case of nonregion data, we show how these data structures can be used to compute nearest objects in an incremental fashion so that the number of objects need not be known in advance. We also review a number of different tessellations and show why hierarchical decomposition into squares instead of triangles or hexagons is preferred. In addition a demonstration of the SAND spatial browser based on the SAND spatial database system and of the VASCO JAVA applet illustrating these methods (found at http://www.cs.umd.edu/~hjs/quadtree/index.html) will be presented. %B ACM SIGGRAPH ASIA 2010 Courses %S SA '10 %I ACM %C New York, NY, USA %P 3:1–3:52 - 3:1–3:52 %8 2010/// %@ 978-1-4503-0527-3 %G eng %U http://doi.acm.org/10.1145/1900520.1900523 %R 10.1145/1900520.1900523 %0 Journal Article %J Image Processing, IEEE Transactions on %D 2010 %T Special Section on Distributed Camera Networks: Sensing, Processing, Communication, and Implementation %A Chellapa, Rama %A Heinzelman, W. %A Konrad, J. %A Schonfeld, D. %A Wolf, M. %X The eight papers in this special section span across theoretical and practical considerations of various aspects of distributed camera networks, including adaptive sensing, distributed processing, efficient communications, and versatile implementations. %B Image Processing, IEEE Transactions on %V 19 %P 2513 - 2515 %8 2010/10// %@ 1057-7149 %G eng %N 10 %R 10.1109/TIP.2010.2063650 %0 Journal Article %J IEEE Transactions on Pattern Analysis and Machine Intelligence %D 2010 %T SPECIAL SECTION ON SHAPE ANALYSIS AND ITS APPLICATIONS IN IMAGE UNDERSTANDING %A Srivastava, A. %A Damon,J.N. %A Dryden,I.L. %A Jermyn,I.H. %A Das,S. %A Vaswani, N. %A Huckemann,S. %A Hotz,T. %A Munk,A. %A Lin,Z. %A others %B IEEE Transactions on Pattern Analysis and Machine Intelligence %V 32 %P 0162 - 8828 %8 2010/// %G eng %N 4 %0 Journal Article %J Magnetics, IEEE Transactions on %D 2010 %T On Stability of Magnetization Dynamics in Nanoparticles %A Mayergoyz, Issak D %A Serpico,C. %A Bertotti,G. %K (electron);magnetic %K DYNAMICS %K ellipsoidal %K field;magnetization %K inequality;exchange %K INTERACTIONS %K magnetization %K nanoparticles;spatially %K nonuniform %K oscillations;spin-wave %K particles;magnetisation;nanoparticles;spin %K perturbations;exchange %K perturbations;spatially %K Poincare %K stability;small %K uniform %K waves; %X It is rigorously demonstrated that spatially uniform magnetization oscillations in sufficiently small ellipsoidal nanoparticles can be unconditionally stable with respect to spatially nonuniform perturbations. The proof reveals the dominant stabilizing effect of exchange field which is mathematically manifested through the Poincare inequality. %B Magnetics, IEEE Transactions on %V 46 %P 1718 - 1721 %8 2010/06// %@ 0018-9464 %G eng %N 6 %R 10.1109/TMAG.2009.2039119 %0 Conference Paper %B 2010 IEEE Symposium on Visual Analytics Science and Technology (VAST) %D 2010 %T The state of visual analytics: Views on what visual analytics is and where it is going %A May,R. %A Hanrahan,P. %A Keim,D. A %A Shneiderman, Ben %A Card,S. %K analytical reasoning %K Data analysis %K data mining %K data visualisation %K European VisMaster program %K interactive visual interfaces %K knowledge discovery %K United States %K visual analytics %K VisWeek community %X In the 2005 publication "Illuminating the Path" visual analytics was defined as "the science of analytical reasoning facilitated by interactive visual interfaces." A lot of work has been done in visual analytics over the intervening five years. While visual analytics started in the United States with a focus on security, it is now a worldwide research agenda with a broad range of application domains. This is evidenced by efforts like the European VisMaster program and the upcoming Visual Analytics and Knowledge Discovery (VAKD) workshop, just to name two.There are still questions concerning where and how visual analytics fits in the large body of research and applications represented by the VisWeek community. This panel will present distinct viewpoints on what visual analytics is and its role in understanding complex information in a complex world. The goal of this panel is to engender discussion from the audience on the emergence and continued advancement of visual analytics and its role relative to fields of related research. Four distinguished panelists will provide their perspective on visual analytics focusing on what it is, what it should be, and thoughts about a development path between these two states. The purpose of the presentations is not to give a critical review of the literature but rather to give a review on the field and to provide a contextual perspective based on the panelists' years of experience and accumulated knowledge. %B 2010 IEEE Symposium on Visual Analytics Science and Technology (VAST) %I IEEE %P 257 - 259 %8 2010/10/25/26 %@ 978-1-4244-9488-0 %G eng %R 10.1109/VAST.2010.5649078 %0 Book Section %B Video Search and MiningVideo Search and Mining %D 2010 %T Statistical Analysis on Manifolds and Its Applications to Video Analysis %A Turaga,Pavan %A Veeraraghavan,Ashok %A Srivastava,Anuj %A Chellapa, Rama %E Schonfeld,Dan %E Shan,Caifeng %E Tao,Dacheng %E Wang,Liang %X The analysis and interpretation of video data is an important component of modern vision applications such as biometrics, surveillance, motionsynthesis and web-based user interfaces. A common requirement among these very different applications is the ability to learn statistical models of appearance and motion from a collection of videos, and then use them for recognizing actions or persons in a new video. These applications in video analysis require statistical inference methods to be devised on non-Euclidean spaces or more formally on manifolds. This chapter outlines a broad survey of applications in video analysis that involve manifolds. We develop the required mathematical tools needed to perform statistical inference on manifolds and show their effectiveness in real video-understanding applications. %B Video Search and MiningVideo Search and Mining %S Studies in Computational Intelligence %I Springer Berlin / Heidelberg %V 287 %P 115 - 144 %8 2010/// %@ 978-3-642-12899-8 %G eng %U http://dx.doi.org/10.1007/978-3-642-12900-1_5 %0 Book %D 2010 %T Statistical Methods and Models for Video-Based Tracking, Modeling, and Recognition %A Chellapa, Rama %A Sankaranarayanan,Aswin C. %A Veeraraghavan,Ashok %K Computers / Computer Vision & Pattern Recognition %K Computers / Image Processing %K Technology & Engineering / Electronics / General %K Technology & Engineering / General %X Computer vision systems attempt to understand a scene and its components from mostly visual information. The geometry exhibited by the real world, the influence of material properties on scattering of incident light, and the process of imaging introduce constraints and properties that are key to solving some of these tasks. In the presence of noisy observations and other uncertainties, the algorithms make use of statistical methods for robust inference. Statistical Methods and Models for Video-based Tracking, Modeling, and Recognition highlights the role of geometric constraints in statistical estimation methods, and how the interplay of geometry and statistics leads to the choice and design of algorithms. In particular, it illustrates the role of imaging, illumination, and motion constraints in classical vision problems such as tracking, structure from motion, metrology, activity analysis and recognition, and appropriate statistical methods used in each of these problems. %I Now Publishers Inc %8 2010/// %@ 9781601983145 %G eng %0 Conference Paper %D 2010 %T Strategy generation in multi-agent imperfect-information pursuit games %A Raboin,Eric %A Nau, Dana S. %A Kuter,Ugur %A Gupta, Satyandra K. %A Svec,Petr %K game tree search %K multi-agent planning %K visibility-based pursuit-evasion games %X We describe a formalism and algorithms for game-tree search in partially-observable Euclidean space, and implementation and tests in a scenario where a multi-agent team, called tracking agents, pursues a target agent that wants to evade the tracking agents. Our contributions include--- • A formalism that combines geometric elements (agents' locations and trajectories and observable regions, and obstacles that restrict mobility and observability) with game-theoretic elements (information sets, utility functions, and strategies). • A recursive formula for information-set minimax values based on our formalism, and a implementation of the formula in a game-tree search algorithm. • A heuristic evaluation function for use at the leaf nodes of the game-tree search. It works by doing a quick lookahead search of its own, in a relaxed version of the problem. • Experimental results in 500 randomly generated trials. With the strategies generated by our heuristic, the tracking agents were more than twice as likely to know the target agent's location at the end of the game than with the strategies generated by heuristics that compute estimates of the target's possible locations. %S AAMAS '10 %I International Foundation for Autonomous Agents and Multiagent Systems %C Richland, SC %P 947 - 954 %8 2010/// %@ 978-0-9826571-1-9 %G eng %U http://dl.acm.org/citation.cfm?id=1838206.1838333 %0 Journal Article %J Journal of molecular biology %D 2010 %T Structural and dynamic determinants of ligand binding and regulation of cyclin-dependent kinase 5 by pathological activator p25 and inhibitory peptide CIP. %A Cardone, Antonio %A Hassan, S A %A Albers,R.W. %A Sriram,R.D. %A Pant,H.C. %K Crystallography, X-Ray %K Cyclin-Dependent Kinase 5 %K Cyclin-Dependent Kinase Inhibitor Proteins %K HUMANS %K Ligands %K molecular dynamics simulation %K Nerve Tissue Proteins %K Principal component analysis %K Protein Binding %K Protein Conformation %X The crystal structure of the cdk5/p25 complex has provided information on possible molecular mechanisms of the ligand binding, specificity, and regulation of the kinase. Comparative molecular dynamics simulations are reported here for physiological conditions. This study provides new insight on the mechanisms that modulate such processes, which may be exploited to control pathological activation by p25. The structural changes observed in the kinase are stabilized by a network of interactions involving highly conserved residues within the cyclin-dependent kinase (cdk) family. Collective motions of the proteins (cdk5, p25, and CIP) and their complexes are identified by principal component analysis, revealing two conformational states of the activation loop upon p25 complexation, which are absent in the uncomplexed kinase and not apparent from the crystal. Simulations of the uncomplexed inhibitor CIP show structural rearrangements and increased flexibility of the interfacial loop containing the critical residue E240, which becomes fully hydrated and available for interactions with one of several positively charged residues in the kinase. These changes provide a rationale for the observed high affinity and enhanced inhibitory action of CIP when compared to either p25 or the physiological activators of cdk5. %B Journal of molecular biology %V 401 %P 478-92 %8 2010 Aug 20 %G eng %N 3 %1 http://www.ncbi.nlm.nih.gov/pubmed/20599546?dopt=Abstract %R 10.1016/j.jmb.2010.06.040 %0 Journal Article %J Link Mining: Models, Algorithms, and Applications %D 2010 %T A survey of link mining tasks for analyzing noisy and incomplete networks %A Namata,G.M. %A Sharara,H. %A Getoor, Lise %X Many data sets of interest today are best described as networks or graphs of interlinked entities. Examples include Web and text collections, social networks and social media sites, information, transaction and communication networks, and all manner of scientific networks, including biological networks. Unfortunately, often the data collection and extraction process for gathering these network data sets is imprecise, noisy, and/or incomplete. In this chapter, we review a collection of link mining algorithms that are well suited to analyzing and making inferences about networks, especially in the case where the data is noisy or missing. %B Link Mining: Models, Algorithms, and Applications %P 107 - 133 %8 2010/// %G eng %R 10.1007/978-1-4419-6515-8_4 %0 Patent %D 2010 %T System and method for analysis of an opinion expressed in documents with regard to a particular topic %A V.S. Subrahmanian %A Picariello,Antonio %A Dorr, Bonnie J %A Reforgiato,Diego Recupero %A Cesarano,Carmine %A Sagoff,Amelia %X System and method for analysis of an opinion expressed in documents on a particular topic computes opinion strength on a continuous numeric scale, or qualitatively. A variety of opinion scoring techniques are plugged in to score opinion expressing words and sentences in documents. These scores are aggregated to measure the opinion intensity of documents. Multilingual opinion analysis is supported by capability to concurrently identify and visualize the opinion intensity expressed in documents in multiple languages. A multi-dimensional representation of the measured opinion intensity is generated which is agreeable with multi-lingual domain. %V 11/808,278 %8 2010//01/28 %G eng %U http://www.google.com/patents?id=j9fLAAAAEBAJ %0 Patent %D 2010 %T System and Method for Confidentiality-Preserving Rank-Ordered Search %A Swaminathan,Ashwin %A Mao,Yinian %A Su,Guan-Ming %A Gou,Hongmei %A Varna,Avinash L. %A He,Shan %A M. Wu %A Oard, Douglas %X A confidentiality preserving system and method for performing a rank-ordered search and retrieval of contents of a data collection. The system includes at least one computer system including a search and retrieval algorithm using term frequency and/or similar features for rank-ordering selective contents of the data collection, and enabling secure retrieval of the selective contents based on the rank-order. The search and retrieval algorithm includes a baseline algorithm, a partially server oriented algorithm, and/or a fully server oriented algorithm. The partially and/or fully server oriented algorithms use homomorphic and/or order preserving encryption for enabling search capability from a user other than an owner of the contents of the data collection. The confidentiality preserving method includes using term frequency for rank-ordering selective contents of the data collection, and retrieving the selective contents based on the rank-order. %V 12/608,724 %8 2010/06/10/ %G eng %U http://www.google.com/patents?id=kInVAAAAEBAJ %0 Conference Paper %D 2010 %T A Systematic Methodology for Accurate Design-Stage Estimation of Energy Consumption for Injection Molded Parts %A Weissman,A. %A Ananthanarayanan,A. %A Gupta,S.K. %A Sriram,R.D. %X Today's ubiquitous use of plastics in product designand manufacturing presents significant environmental and human health challenges. Injection molding, one of the most commonly used processes for making plastic products, consumes a significant amount of energy. A methodology for accurately estimating the energy consumed to injection-mold a part would enable environmentally conscious decision making during the product design. Unfortunately, only limited information is available at the design stage. Therefore, accurately estimating energy consumption before the part has gone into production can be challenging. In this paper, we describe a methodology for energy estimation that works with the limited amount of data available during the design stage, namely the CAD model of the part, the material name, and the production requirements. This methodology uses this data to estimate the parameters of the runner system and an appropriately sized molding machine. It then uses these estimates to compute the machine setup time and the cycle time required for the injection molding operation. This is done by appropriately abstracting information available from the mold flow simulation tools and analytical models that are traditionally used during the manufacturing stage. These times are then multiplied by the power consumed by the appropriately sized machine during each stage of the molding cycle to compute the estimated energy consumption per part. %8 2010/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.165.4139&rep=rep1&type=pdf %0 Journal Article %J Proc. VLDB Endow. %D 2010 %T Techniques for similarity searching in multimedia databases %A Samet, Hanan %X Techniques for similarity searching in multimedia databases are reviewed. This includes a discussion of the curse of dimensionality, as well as multidimensional indexing, distance-based indexing, and the actual search process which is realized by nearest neighbor finding. %B Proc. VLDB Endow. %V 3 %P 1649 - 1650 %8 2010/09// %@ 2150-8097 %G eng %U http://dl.acm.org/citation.cfm?id=1920841.1921064 %N 1-2 %0 Book Section %B Active Media TechnologyActive Media Technology %D 2010 %T Technology-Mediated Social Participation: Deep Science and Extreme Technology %A Shneiderman, Ben %E An,Aijun %E Lingras,Pawan %E Petty,Sheila %E Huang,Runhe %X The dramatic success of social media such as Facebook, Twitter, YouTube, Flickr, blogs, and traditional discussion groups empowers individuals to become active in local and global communities. With modest redesign, these technologies can be harnessed to support national priorities such as healthcare/wellness, disaster response, community safety, energy sustainability, etc. This talk describes a research agenda for these topics that develops deep science questions and extreme technology challenges. %B Active Media TechnologyActive Media Technology %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6335 %P 1 - 4 %8 2010/// %@ 978-3-642-15469-0 %G eng %U http://dx.doi.org/10.1007/978-3-642-15470-6_1 %0 Conference Paper %B Proceedings of the 28th international conference on Human factors in computing systems %D 2010 %T Toque: designing a cooking-based programming language for and with children %A Tarkan,S. %A Sazawal,V. %A Druin, Allison %A Golub,E. %A Bonsignore,E. M %A Walsh,G. %A Atrash,Z. %B Proceedings of the 28th international conference on Human factors in computing systems %P 2417 - 2426 %8 2010/// %G eng %0 Journal Article %J J Bacteriol %D 2010 %T Unexpected abundance of coenzyme F(420)-dependent enzymes in Mycobacterium tuberculosis and other actinobacteria. %A Jeremy D Selengut %A Haft, Daniel H %K Actinobacteria %K Amino Acid Sequence %K Binding Sites %K Coenzymes %K Flavonoids %K Gene Expression Profiling %K Gene Expression Regulation, Bacterial %K Genome, Bacterial %K molecular biology %K Molecular Sequence Data %K Molecular Structure %K Mycobacterium tuberculosis %K Phylogeny %K Protein Conformation %K Riboflavin %X

Regimens targeting Mycobacterium tuberculosis, the causative agent of tuberculosis (TB), require long courses of treatment and a combination of three or more drugs. An increase in drug-resistant strains of M. tuberculosis demonstrates the need for additional TB-specific drugs. A notable feature of M. tuberculosis is coenzyme F(420), which is distributed sporadically and sparsely among prokaryotes. This distribution allows for comparative genomics-based investigations. Phylogenetic profiling (comparison of differential gene content) based on F(420) biosynthesis nominated many actinobacterial proteins as candidate F(420)-dependent enzymes. Three such families dominated the results: the luciferase-like monooxygenase (LLM), pyridoxamine 5'-phosphate oxidase (PPOX), and deazaflavin-dependent nitroreductase (DDN) families. The DDN family was determined to be limited to F(420)-producing species. The LLM and PPOX families were observed in F(420)-producing species as well as species lacking F(420) but were particularly numerous in many actinobacterial species, including M. tuberculosis. Partitioning the LLM and PPOX families based on an organism's ability to make F(420) allowed the application of the SIMBAL (sites inferred by metabolic background assertion labeling) profiling method to identify F(420)-correlated subsequences. These regions were found to correspond to flavonoid cofactor binding sites. Significantly, these results showed that M. tuberculosis carries at least 28 separate F(420)-dependent enzymes, most of unknown function, and a paucity of flavin mononucleotide (FMN)-dependent proteins in these families. While prevalent in mycobacteria, markers of F(420) biosynthesis appeared to be absent from the normal human gut flora. These findings suggest that M. tuberculosis relies heavily on coenzyme F(420) for its redox reactions. This dependence and the cofactor's rarity may make F(420)-related proteins promising drug targets.

%B J Bacteriol %V 192 %P 5788-98 %8 2010 Nov %G eng %N 21 %R 10.1128/JB.00425-10 %0 Journal Article %J ICLP %D 2010 %T Using Generalized Annotated Programs to Solve Social Network Optimization Problems %A Shakarian,P. %A V.S. Subrahmanian %A Sapino,M. L %A Hermenegildo,M. %A Schaub,T. %X Reasoning about social networks (labeled, directed, weighted graphs) is be-coming increasingly important and there are now models of how certain phenomena (e.g. adoption of products/services by consumers, spread of a given disease) “diffuse” through the network. Some of these diffusion models can be expressed via generalized annotated programs (GAPs). In this paper, we consider the following problem: suppose we have a given goal to achieve (e.g. maximize the expected number of adoptees of a product or minimize the spread of a disease) and suppose we have limited resources to use in trying to achieve the goal (e.g. give out a few free plans, provide medication to key people in the SN) - how should these resources be used so that we optimize a given objective function related to the goal? We define a class of social network optimization problems (SNOPs) that supports this type of reasoning. We formalize and study the complexity of SNOPs and show how they can be used in conjunction with existing economic and disease diffusion models. %B ICLP %8 2010/// %G eng %0 Conference Paper %B Software Engineering, 2010 ACM/IEEE 32nd International Conference on %D 2010 %T Using symbolic evaluation to understand behavior in configurable software systems %A Reisner,E. %A Song,C. %A Ma,K.K. %A Foster, Jeffrey S. %A Porter, Adam %B Software Engineering, 2010 ACM/IEEE 32nd International Conference on %V 1 %P 445 - 454 %8 2010/// %G eng %0 Journal Article %J IEEE Signal Processing Magazine %D 2010 %T Utilizing Hierarchical Multiprocessing for Medical Image Registration %A Plishker,W. %A Dandekar,O. %A Bhattacharyya, Shuvra S. %A Shekhar,R. %K Acceleration %K application parallelism %K Biomedical imaging %K domain-specific taxonomy %K GPU acceleration %K gradient descent approach %K Graphics processing unit %K hierarchical multiprocessing %K image registration %K Magnetic resonance imaging %K Medical diagnostic imaging %K medical image processing %K medical image registration %K multicore platform set %K Multicore processing %K PARALLEL PROCESSING %K parallel programming %K Robustness %K Signal processing algorithms %K Ultrasonic imaging %X This work discusses an approach to utilize hierarchical multiprocessing in the context of medical image registration. By first organizing application parallelism into a domain-specific taxonomy, an algorithm is structured to target a set of multicore platforms.The approach on a cluster of graphics processing units (GPUs) requiring the use of two parallel programming environments to achieve fast execution times is demonstrated.There is negligible loss in accuracy for rigid registration when employing GPU acceleration, but it does adversely effect our nonrigid registration implementation due to our usage of a gradient descent approach. %B IEEE Signal Processing Magazine %V 27 %P 61 - 68 %8 2010 %@ 1053-5888 %G eng %N 2 %0 Conference Paper %B Visual Analytics Science and Technology (VAST), 2010 IEEE Symposium on %D 2010 %T VAST 2010 Challenge: Arms dealings and pandemics %A Grinstein,G. %A Konecni,S. %A Scholtz,J. %A Whiting,M. %A Plaisant, Catherine %K 2010 %K administrative %K admission;intelligence %K analysis;interactive %K analysis;medical %K challenge;arms %K challenge;hospital %K data %K data;grand %K dealings;bioinformatics;dangerous %K mutations;death %K processing;weapons; %K records;genetic %K reports;bioinformatics;data %K VAST %K viral %K visualizations;minichallenge;pandemics;text %X The 5th VAST Challenge consisted of three mini-challenges that involved both intelligence analysis and bioinformatics. Teams could solve one, two or all three mini-challenges and assess the overall situation to enter the Grand Challenge. Mini-challenge one involved text reports about people and events giving information about arms dealers, situations in various countries and linkages between different countries. Mini-challenge two involved hospital admission and death records from various countries providing information about the spread of a world wide pandemic. Mini-challenge three involved genetic data to be used to identify the origin of the pandemic and the most dangerous viral mutations. The Grand Challenge was to determine how these various mini-challenges were connected. As always the goal was to analyze the data and provide novel interactive visualizations useful in the analytic process. We received 58 submissions in total and gave 15 awards. %B Visual Analytics Science and Technology (VAST), 2010 IEEE Symposium on %P 263 - 264 %8 2010/10// %G eng %R 10.1109/VAST.2010.5649054 %0 Journal Article %J IEEE Transactions on Multimedia %D 2010 %T Video Précis: Highlighting Diverse Aspects of Videos %A Shroff, N. %A Turaga,P. %A Chellapa, Rama %K $K$-means %K CAMERAS %K combinatorial mathematics %K combinatorial optimization %K Cost function %K data compression %K Exemplar selection %K Image segmentation %K Internet %K Iron %K Length measurement %K multimedia systems %K Ncut %K optimisation %K Optimization methods %K original video %K Permission %K shot segmentation %K Surveillance %K user specified summary length %K video précis %K Video sharing %K video signal processing %K Video summarization %X Summarizing long unconstrained videos is gaining importance in surveillance, web-based video browsing, and video-archival applications. Summarizing a video requires one to identify key aspects that contain the essence of the video. In this paper, we propose an approach that optimizes two criteria that a video summary should embody. The first criterion, “coverage,” requires that the summary be able to represent the original video well. The second criterion, “diversity,” requires that the elements of the summary be as distinct from each other as possible. Given a user-specified summary length, we propose a cost function to measure the quality of a summary. The problem of generating a précis is then reduced to a combinatorial optimization problem of minimizing the proposed cost function. We propose an efficient method to solve the optimization problem. We demonstrate through experiments (on KTH data, unconstrained skating video, a surveillance video, and a YouTube home video) that optimizing the proposed criterion results in meaningful video summaries over a wide range of scenarios. Summaries thus generated are then evaluated using both quantitative measures and user studies. %B IEEE Transactions on Multimedia %V 12 %P 853 - 868 %8 2010/12// %@ 1520-9210 %G eng %N 8 %R 10.1109/TMM.2010.2058795 %0 Conference Paper %B Proceedings of the 1st ACM International Health Informatics Symposium %D 2010 %T Visual information seeking in multiple electronic health records: design recommendations and a process model %A Wang,Taowei David %A Wongsuphasawat,Krist %A Plaisant, Catherine %A Shneiderman, Ben %K design requriements %K electronic health records %K human-computer interaction (hci) %K Information Visualization %X Current electronic health record (EHR) systems facilitate the storage, retrieval, persistence, and sharing of patient data. However, the way physicians interact with EHRs has not changed much. More specifically, support for temporal analysis of a large number of EHRs has been lacking. A number of information visualization techniques have been proposed to alleviate this problem. Unfortunately, due to their limited application to a single case study, the results are often difficult to generalize across medical scenarios. We present the usage data of Lifelines2,[22] our information visualization system, and user comments, both collected over eight different medical case studies. We generalize our experience into an information-seeking process model for multiple EHRs. Based on our analysis, we make recommendations to future information visualization designers for EHRs on design requirements and future research directions. %B Proceedings of the 1st ACM International Health Informatics Symposium %S IHI '10 %I ACM %C New York, NY, USA %P 46 - 55 %8 2010/// %@ 978-1-4503-0030-8 %G eng %U http://doi.acm.org/10.1145/1882992.1883001 %R 10.1145/1882992.1883001 %0 Book Section %B Active Media TechnologyActive Media Technology %D 2010 %T Visualizing Threaded Conversation Networks: Mining Message Boards and Email Lists for Actionable Insights %A Hansen,Derek %A Shneiderman, Ben %A Smith,Marc %E An,Aijun %E Lingras,Pawan %E Petty,Sheila %E Huang,Runhe %X Analyzing complex online relationships is a difficult job, but new information visualization tools are enabling a wider range of users to make actionable insights from the growing volume of online data. This paper describes the challenges and methods for conducting analyses of threaded conversations such as found in enterprise message boards, email lists, and forums. After defining threaded conversation, we characterize the types of networks that can be extracted from them. We then provide 3 mini case studies to illustrate how actionable insights for community managers can be gained by applying the network analysis metrics and visualizations available in the free, open source NodeXL tool, which is a powerful, yet easy-to-use tool embedded in Excel 2007/2010. %B Active Media TechnologyActive Media Technology %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 6335 %P 47 - 62 %8 2010/// %@ 978-3-642-15469-0 %G eng %U http://dx.doi.org/10.1007/978-3-642-15470-6_7 %0 Journal Article %J Digital Mammography %D 2010 %T A Web Database for Computer-Aided Detection and Diagnosis of Medical Images %A Tahmoush,D. %A Samet, Hanan %B Digital Mammography %P 265 - 272 %8 2010/// %G eng %0 Report %D 2010 %T Which Factors Affect Access Network Performance? %A Sundaresan,S. %A Feamster, Nick %A Dicioccio,L. %A Teixeira,R. %X This paper presents an analysis of the performance of residential access networks using over four months of round-trip, download, and upload measurements from more than 7,000 users across four ADSL and cable providers in France. Previous studies have characterized residential access network performance, but this paper presents the first study of how access network performance relates to other factors, such as choice of access provider, service-level agreement, and geographic location. We first explore the extent to which user performance matches the capacity advertised by an access provider, and whether the ability to achieve this capacity depends on the user’s access network. We then analyze the extent to which various factors influence the performance that users experience. Finally, we explore how different groups of users experience simultaneous performance anomalies and analyze the common characteristics of users that share fate (e.g., whether users that experience simultaneous performance degradation share the same provider, city). Our analysis informs both users and designers of networked services who wish to improve the reliability and performance of access networks through multihoming and may also assist operators with troubleshooting network issues by narrowing down likely causes. %I Georgia Institute of Technology %V GT-CS-10-04 %8 2010/// %G eng %U http://hdl.handle.net/1853/37336 %0 Conference Paper %B SIGCHI '10 %D 2010 %T Who's Hogging the Bandwidth: The Consequences of Revealing the Invisible in the Home %A Marshini Chetty %A Banks, Richard %A Harper, Richard %A Regan, Tim %A Sellen, Abigail %A Gkantsidis, Christos %A Karagiannis, Thomas %A Key, Peter %K bandwidth monitoring %K home broadband %K home networks %X As more technologies enter the home, householders are burdened with the task of digital housekeeping-managing and sharing digital resources like bandwidth. In response to this, we created and evaluated a domestic tool for bandwidth management called Home Watcher. Our field trial showed that when resource contention amongst different household members is made visible, people's understanding of bandwidth changes and household politics are revealed. In this paper, we describe the consequences of showing real time resource usage in a home, and how this varies depending on the social make up of the household. %B SIGCHI '10 %S CHI '10 %I ACM %P 659 - 668 %8 2010/// %@ 978-1-60558-929-9 %G eng %U http://doi.acm.org/10.1145/1753326.1753423 %0 Conference Paper %B Proceedings of the ACM international conference companion on Object oriented programming systems languages and applications companion %D 2010 %T Workshop on experimental evaluation of software and systems in computer science (Evaluate 2010) %A Blackburn,Steven M. %A Diwan,Amer %A Hauswirth,Matthias %A Memon, Atif M. %A Sweeney,Peter F. %K Evaluation %K METHODOLOGY %X We call ourselves 'computer scientists', but are we scientists? If we are scientists, then we must practice the scientific method. This includes a solid experimental evaluation. In our experience, our experimental methodology is ad hoc at best, and nonexistent at worst. This workshop brings together experts from different areas of computer science to discuss, explore, and attempt to identify the principles of sound experimental evaluation %B Proceedings of the ACM international conference companion on Object oriented programming systems languages and applications companion %S SPLASH '10 %I ACM %C New York, NY, USA %P 291 - 292 %8 2010/// %@ 978-1-4503-0240-1 %G eng %U http://doi.acm.org/10.1145/1869542.1869618 %R 10.1145/1869542.1869618 %0 Journal Article %J PLoS ONEPLoS ONE %D 2009 %T 2009 Swine-Origin Influenza A (H1N1) Resembles Previous Influenza Isolates %A Kingsford, Carl %A Nagarajan,Niranjan %A Salzberg,Steven L. %X In April 2009, novel swine-origin influenza viruses (S-OIV) were identified in patients from Mexico and the United States. The viruses were genetically characterized as a novel influenza A (H1N1) strain originating in swine, and within a very short time the S-OIV strain spread across the globe via human-to-human contact.We conducted a comprehensive computational search of all available sequences of the surface proteins of H1N1 swine influenza isolates and found that a similar strain to S-OIV appeared in Thailand in 2000. The earlier isolates caused infections in pigs but only one sequenced human case, A/Thailand/271/2005 (H1N1). Differences between the Thai cases and S-OIV may help shed light on the ability of the current outbreak strain to spread rapidly among humans. %B PLoS ONEPLoS ONE %V 4 %P e6402 - e6402 %8 2009/07/28/ %G eng %U http://dx.doi.org/10.1371/journal.pone.0006402 %N 7 %R 10.1371/journal.pone.0006402 %0 Journal Article %J Information VisualizationInformation Visualization %D 2009 %T Advancing User-Centered Evaluation of Visual Analytic Environments Through Contests %A Costello,Loura %A Grinstein,Georges %A Plaisant, Catherine %A Scholtz,Jean %K metrics %K synthetic data %K user-centered evaluation %K visual analytics %X In this paper, the authors describe the Visual Analytics Science and Technology (VAST) Symposium contests run in 2006 and 2007 and the VAST 2008 and 2009 challenges. These contests were designed to provide researchers with a better understanding of the tasks and data that face potential end users. Access to these end users is limited because of time constraints and the classified nature of the tasks and data. In that respect, the contests serve as an intermediary, with the metrics and feedback serving as measures of utility to the end users. The authors summarize the lessons learned and the future directions for VAST Challenges. %B Information VisualizationInformation Visualization %V 8 %P 230 - 238 %8 2009/09/21/ %@ 1473-8716, 1473-8724 %G eng %U http://ivi.sagepub.com/content/8/3/230 %N 3 %R 10.1057/ivs.2009.16 %0 Conference Paper %B Data Engineering, 2009. ICDE '09. IEEE 25th International Conference on %D 2009 %T Aggregate Query Answering under Uncertain Schema Mappings %A Gal,A. %A Martinez,M. V %A Simari,G. I %A V.S. Subrahmanian %K aggregate %K algorithm;probabilistic %K answering;by-table %K complexity;data %K complexity;distributed %K database;polynomial %K databases; %K databases;probability;query %K integration;distribution %K mapping;computational %K mapping;query %K processing;range %K processing;statistical %K query %K schema %K semantics;by-tuple %K semantics;computational %K semantics;expected %K semantics;multiple %K semantics;uncertain %K TIME %K value %X Recent interest in managing uncertainty in data integration has led to the introduction of probabilistic schema mappings and the use of probabilistic methods to answer queries across multiple databases using two semantics: by-table and by-tuple. In this paper, we develop three possible semantics for aggregate queries: the range, distribution, and expected value semantics, and show that these three semantics combine with the by-table and by-tuple semantics in six ways. We present algorithms to process COUNT, AVG, SUM, MIN, and MAX queries under all six semantics and develop results on the complexity of processing such queries under all six semantics. We show that computing COUNT is in PTIME for all six semantics and computing SUM is in PTIME for all but the by-tuple/distribution semantics. Finally, we show that AVG, MIN, and MAX are PTIME computable for all by-table semantics and for the by-tuple/range semantics.We developed a prototype implementation and experimented with both real-world traces and simulated data. We show that, as expected, naive processing of aggregates does not scale beyond small databases with a small number of mappings. The results also show that the polynomial time algorithms are scalable up to several million tuples as well as with a large number of mappings. %B Data Engineering, 2009. ICDE '09. IEEE 25th International Conference on %P 940 - 951 %8 2009/04/29/2 %G eng %R 10.1109/ICDE.2009.55 %0 Journal Article %J Environmental Microbiology Reports %D 2009 %T Analysis of clonally related environmental Vibrio cholerae O1 El Tor isolated before 1992 from Varanasi, India reveals origin of SXT‐ICEs belonging to O139 and O1 serogroups %A Mohapatra,Saswat S. %A Mantri,Chinmay K. %A Mohapatra,Harapriya %A Rita R Colwell %A Singh,Durg V. %X In this study, we report the presence of SXT in environmental Vibrio cholerae O1 El Tor strains isolated before 1992 from Varanasi, India. All isolates, except one, were resistant to Tm, and/or Sul, Sm, Fr, Na and Am. None contained plasmids. PCR and DNA sequencing revealed the presence of SXT containing dfrA1 and/or sulII, strAB in six isolates and dfr18, sulII and strAB in five isolates. Three clinical V. cholerae O1 isolated during 1992 contained the antibiotic resistance gene cassette aadA1 in the class 1 integron. Conjugation experiments, followed by PCR analysis of transconjugants, provided evidence of the transferable nature of SXT and associated antibiotic resistance genes, and its integration into the prfC site. Results of phylogenetic analysis of the intSXT gene of clonally similar V. cholerae showed a clear difference between dfr18+ and dfrA1+V. cholerae O1 isolates. This is the first report of the occurrence of SXT harbouring sulII, strAB, dfr18 and/or dfrA1 genes in environmental V. cholerae O1 isolated prior to 1992 from Varanasi, India, and suggests emergence of SXT+ antibiotic-resistant V. cholerae O139 and O1 from an environmental V. cholerae progenitor by acquisition of SXT and antibiotic-resistant gene clusters. %B Environmental Microbiology Reports %V 2 %P 50 - 57 %8 2009/07/21/ %@ 1758-2229 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1758-2229.2009.00051.x/abstract?userIsAuthenticated=false&deniedAccessCustomisedMessage= %N 1 %R 10.1111/j.1758-2229.2009.00051.x %0 Journal Article %J Magnetics, IEEE Transactions on %D 2009 %T Analytical Description of Quasi-Random Magnetization Relaxation to Equilibrium %A Serpico,C. %A d'Aquino,M. %A Bertotti,G. %A Mayergoyz, Issak D %K approach;quasirandom %K crossing;uniformly %K dynamical %K dynamics;averaging %K Hamiltonian %K magnetization %K magnetized %K particle;magnetic %K particles;magnetic %K perturbed %K relaxation;magnetisation;probability; %K relaxation;separatrix %K system;Landau-Lifshitz %K technique;damping;probabilistic %X The Landau-Lifshitz (LL) dynamics of a uniformly magnetized particle is considered. The LL equation is written in the form of a Hamiltonian perturbed dynamical system. By using suitable averaging technique, the equation for the slow dynamics of the energy is derived. The averaging technique breaks up in the case of separatrix crossing. It is shown that, in the limit of small damping, the separatrix crossing can be described by using a probabilistic approach. %B Magnetics, IEEE Transactions on %V 45 %P 5224 - 5227 %8 2009/11// %@ 0018-9464 %G eng %N 11 %R 10.1109/TMAG.2009.2031067 %0 Conference Paper %B Proceedings of the fourth international conference on Communities and technologies %D 2009 %T Analyzing (social media) networks with NodeXL %A Smith,Marc A. %A Shneiderman, Ben %A Milic-Frayling,Natasa %A Mendes Rodrigues,Eduarda %A Barash,Vladimir %A Dunne,Cody %A Capone,Tony %A Perer,Adam %A Gleave,Eric %K excel %K network analysis %K social media %K social network %K spreadsheet %K Visualization %X We present NodeXL, an extendible toolkit for network overview, discovery and exploration implemented as an add-in to the Microsoft Excel 2007 spreadsheet software. We demonstrate NodeXL data analysis and visualization features with a social media data sample drawn from an enterprise intranet social network. A sequence of NodeXL operations from data import to computation of network statistics and refinement of network visualization through sorting, filtering, and clustering functions is described. These operations reveal sociologically relevant differences in the patterns of interconnection among employee participants in the social media space. The tool and method can be broadly applied. %B Proceedings of the fourth international conference on Communities and technologies %S C&T '09 %I ACM %C New York, NY, USA %P 255 - 264 %8 2009/// %@ 978-1-60558-713-4 %G eng %U http://doi.acm.org/10.1145/1556460.1556497 %R 10.1145/1556460.1556497 %0 Conference Paper %B Proceedings of the ACL-IJCNLP 2009 Conference Short Papers$}$ %D 2009 %T Arabic Cross-Document Coreference Resolution %A Sayeed,A. %A Elsayed,T. %A Garera,N. %A Alexander,D. %A Xu,T. %A Oard, Douglas %A Yarowsky,D. %A Piatko,C. %B Proceedings of the ACL-IJCNLP 2009 Conference Short Papers$}$ %P 357 - 360 %8 2009/// %G eng %0 Journal Article %J International Journal of Embedded Systems %D 2009 %T An architectural level design methodology for smart camera applications %A Saha,S. %A Kianzad,V. %A Schlessman,J. %A Aggarwal,G. %A Bhattacharyya, Shuvra S. %A Chellapa, Rama %A Wolf,W. %X Today's embedded computing applications are characterised by increased functionality, and hence increased design complexity and processing requirements. The resulting design spaces are vast and designers are typically able to evaluate only small subsets of solutions due to lack of efficient design tools. In this paper, we propose an architectural level design methodology that provides a means for a comprehensive design space exploration for smart camera applications and enable designers to select higher quality solutions and provides substantial savings on the overall cost of the system. We present efficient, accurate and intuitive models for performance estimation and validate them with experiments. %B International Journal of Embedded Systems %V 4 %P 83 - 97 %8 2009/// %G eng %N 1 %0 Journal Article %J Nature Structural & Molecular Biology %D 2009 %T Avid interactions underlie the Lys63-linked polyubiquitin binding specificities observed for UBA domains %A Sims,Joshua J. %A Haririnia,Aydin %A Dickinson,Bryan C. %A Fushman, David %A Cohen,Robert E. %K apoptosis %K basic cellular processes %K Biochemistry %K biophysics %K cell biology %K cell cycle %K cell surface proteins %K cell-cell interactions %K checkpoints %K chromatin %K chromatin remodeling %K chromatin structure %K content %K DNA recombination %K DNA repair %K DNA replication %K Gene expression %K Genetics %K intracellular signaling %K journal %K macromolecules %K mechanism %K membrane processes %K molecular %K molecular basis of disease %K molecular biology %K molecular interactions %K multi-component complexes %K nature publishing group %K nature structural molecular biology %K nucleic acids %K protein degradation %K protein folding %K protein processing %K Proteins %K regulation of transcription %K regulation of translation %K RNA %K RNA processing %K RNAi %K signal transduction %K single molecule studies %K structure and function of proteins %K transcription %K translation %X Ubiquitin (denoted Ub) receptor proteins as a group must contain a diverse set of binding specificities to distinguish the many forms of polyubiquitin (polyUb) signals. Previous studies suggested that the large class of ubiquitin-associated (UBA) domains contains members with intrinsic specificity for Lys63-linked polyUb or Lys48-linked polyUb, thus explaining how UBA-containing proteins can mediate diverse signaling events. Here we show that previously observed Lys63-polyUb selectivity in UBA domains is the result of an artifact in which the dimeric fusion partner, glutathione S-transferase (GST), positions two UBAs for higher affinity, avid interactions with Lys63-polyUb, but not with Lys48-polyUb. Freed from GST, these UBAs are either nonselective or prefer Lys48-polyUb. Accordingly, NMR experiments reveal no Lys63-polyUb–specific binding epitopes for these UBAs. We reexamine previous conclusions based on GST-UBAs and present an alternative model for how UBAs achieve a diverse range of linkage specificities. %B Nature Structural & Molecular Biology %V 16 %P 883 - 889 %8 2009/// %@ 1545-9993 %G eng %U http://www.nature.com/nsmb/journal/v16/n8/abs/nsmb.1637.html %N 8 %R 10.1038/nsmb.1637 %0 Conference Paper %B Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence %D 2009 %T Bisimulation-based approximate lifted inference %A Sen,Prithviraj %A Deshpande, Amol %A Getoor, Lise %X There has been a great deal of recent interest in methods for performing lifted inference; however, most of this work assumes that the first-order model is given as input to the system. Here, we describe lifted inference algorithms that determine symmetries and automatically lift the probabilistic model to speedup inference. In particular, we describe approximate lifted inference techniques that allow the user to trade off inference accuracy for computational efficiency by using a handful of tunable parameters, while keeping the error bounded. Our algorithms are closely related to the graph-theoretic concept of bisimulation. We report experiments on both synthetic and real data to show that in the presence of symmetries, run-times for inference can be improved significantly, with approximate lifted inference providing orders of magnitude speedup over ground inference. %B Proceedings of the Twenty-Fifth Conference on Uncertainty in Artificial Intelligence %S UAI '09 %I AUAI Press %C Arlington, Virginia, United States %P 496 - 505 %8 2009/// %@ 978-0-9749039-5-8 %G eng %U http://dl.acm.org/citation.cfm?id=1795114.1795172 %0 Conference Paper %B Proceedings of the twenty-first annual symposium on Parallelism in algorithms and architectures %D 2009 %T Brief announcement: performance potential of an easy-to-program PRAM-on-chip prototype versus state-of-the-art processor %A Caragea,G.C. %A Saybasili,A. B %A Wen,X. %A Vishkin, Uzi %B Proceedings of the twenty-first annual symposium on Parallelism in algorithms and architectures %P 163 - 165 %8 2009/// %G eng %0 Journal Article %J Mathematical Methods in Counterterrorism %D 2009 %T CAPE: Automatically Predicting Changes in Group Behavior %A Sliva,A. %A V.S. Subrahmanian %A Martinez,V. %A Simari,G. %X There is now intense interest in the problem of forecasting what a group will do in the future. Past work [1, 2, 3] has built complex models of a group’s behavior and used this to predict what the group might do in the future. However, almost all past work assumes that the group will not change its past behavior. Whether the group is a group of investors, or a political party, or a terror group, there is much interest in when and how the group will change its behavior. In this paper, we develop an architecture and algorithms called CAPE to forecast the conditions under which a group will change its behavior. We have tested CAPE on social science data about the behaviors of seven terrorist groups and show that CAPE is highly accurate in its predictions—at least in this limited setting. %B Mathematical Methods in Counterterrorism %P 253 - 269 %8 2009/// %G eng %R 10.1007/978-3-211-09442-6_15 %0 Journal Article %J ScienceScience %D 2009 %T Civic Collaboration %A Shneiderman, Ben %X Drawing on her work developing the Peer-to-Patent network, Noveck describes how Web 2.0 technologies could be used to promote participatory democracy and more efficient government. %B ScienceScience %V 325 %P 540 - 540 %8 2009/07/31/ %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/325/5940/540.1 %N 5940 %R 10.1126/science.1178326 %0 Book Section %B TAC 2009 Workshop ProceedingsTAC 2009 Workshop Proceedings %D 2009 %T CLASSY 2009: Summarization and Metrics %A Conroy,John M. %A Schlesinger,Judith D. %A O'Leary, Dianne P. %B TAC 2009 Workshop ProceedingsTAC 2009 Workshop Proceedings %I NIST %8 2009/11/16/17 %G eng %U http://www.nist.gov/tac/publications/2009/participant.papers/CLASSY.proceedings.pdfhttp://www.nist.gov/tac/publications/2009/participant.papers/CLASSY.proceedings.pdf %0 Conference Paper %B Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining %D 2009 %T Co-evolution of social and affiliation networks %A Zheleva,Elena %A Sharara,Hossam %A Getoor, Lise %K affiliation network %K Evolution %K graph generator %K groups %K social network %X In our work, we address the problem of modeling social network generation which explains both link and group formation. Recent studies on social network evolution propose generative models which capture the statistical properties of real-world networks related only to node-to-node link formation. We propose a novel model which captures the co-evolution of social and affiliation networks. We provide surprising insights into group formation based on observations in several real-world networks, showing that users often join groups for reasons other than their friends. Our experiments show that the model is able to capture both the newly observed and previously studied network properties. This work is the first to propose a generative model which captures the statistical properties of these complex networks. The proposed model facilitates controlled experiments which study the effect of actors' behavior on the evolution of affiliation networks, and it allows the generation of realistic synthetic datasets. %B Proceedings of the 15th ACM SIGKDD international conference on Knowledge discovery and data mining %S KDD '09 %I ACM %C New York, NY, USA %P 1007 - 1016 %8 2009/// %@ 978-1-60558-495-9 %G eng %U http://doi.acm.org/10.1145/1557019.1557128 %R 10.1145/1557019.1557128 %0 Book Section %B Text Mining: Classification, Clustering, and ApplicationsText Mining: Classification, Clustering, and Applications %D 2009 %T Collective classification for text classification %A Namata,G. %A Sen,P. %A Bilgic,M. %A Getoor, Lise %B Text Mining: Classification, Clustering, and ApplicationsText Mining: Classification, Clustering, and Applications %P 51 - 69 %8 2009/// %G eng %0 Journal Article %J Advances in Cryptology-CRYPTO 2009 %D 2009 %T Collusion-free multiparty computation in the mediated model %A Alwen,J. %A Katz, Jonathan %A Lindell,Y. %A Persiano,G. %A Shelat,A. %A Visconti,I. %X Collusion-free protocols prevent subliminal communication (i.e., covert channels) between parties running the protocol. In the standard communication model, if one-way functions exist, then protocols satisfying any reasonable degree of privacy cannot be collusion-free. To circumvent this impossibility, Alwen, shelat and Visconti (CRYPTO 2008) recently suggested the mediated model where all communication passes through a mediator. The goal is to design protocols where collusion-freeness is guaranteed as long as the mediator is honest, while standard security guarantees hold if the mediator is dishonest. In this model, they gave constructions of collusion-free protocols for commitments and zero-knowledge proofs in the two-party setting.We strengthen the definition of Alwen et al., and resolve the main open questions in this area by showing a collusion-free protocol (in the mediated model) for computing any multi-party functionality. %B Advances in Cryptology-CRYPTO 2009 %P 524 - 540 %8 2009/// %G eng %R 10.1007/978-3-642-03356-8_31 %0 Conference Paper %B Applications of Computer Vision (WACV), 2009 Workshop on %D 2009 %T Combining multiple kernels for efficient image classification %A Siddiquie,B. %A Vitaladevuni,S.N. %A Davis, Larry S. %K (artificial %K AdaBoost;base %K channels;multiple %K classification;kernel %K classification;learning %K decision %K feature %K function;discriminative %K intelligence);support %K Kernel %K kernel;image %K kernels;composite %K learning;support %K machine;image %K machines; %K similarity;multiple %K vector %X We investigate the problem of combining multiple feature channels for the purpose of efficient image classification. Discriminative kernel based methods, such as SVMs, have been shown to be quite effective for image classification. To use these methods with several feature channels, one needs to combine base kernels computed from them. Multiple kernel learning is an effective method for combining the base kernels. However, the cost of computing the kernel similarities of a test image with each of the support vectors for all feature channels is extremely high. We propose an alternate method, where training data instances are selected, using AdaBoost, for each of the base kernels. A composite decision function, which can be evaluated by computing kernel similarities with respect to only these chosen instances, is learnt. This method significantly reduces the number of kernel computations required during testing. Experimental results on the benchmark UCI datasets, as well as on a challenging painting dataset, are included to demonstrate the effectiveness of our method. %B Applications of Computer Vision (WACV), 2009 Workshop on %P 1 - 8 %8 2009/12// %G eng %R 10.1109/WACV.2009.5403040 %0 Journal Article %J Signal Processing Magazine, IEEE %D 2009 %T Component forensics %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K ACQUISITION %K component %K data; %K forensics;digital %K image %K of %K processing;image %K sensor %K sensors;security %K technique;visual %K technology;image %X Visual sensor technologies have experienced tremendous growth in recent decades, and digital devices are becoming ubiquitous. Digital images taken by various imaging devices have been used in a growing number of applications, from military and reconnaissance to medical diagnosis and consumer photography. Consequently, a series of new forensic issues arise amidst such rapid advancement and widespread adoption of imaging technologies. For example, one can readily ask what kinds of hardware and software components as well as their parameters have been employed inside these devices? Given a digital image, which imaging sensor or which brand of sensor was used to acquire the image? How was the image acquired? Was it captured using a digital camera, cell phone camera, image scanner, or was it created artificially using an imageediting software? Has the image undergone any manipulation after capture? Is it authentic, or has it been tampered in any way? Does it contain any hidden information or steganographic data? Many of these forensic questions are related to tracing the origin of the digital image to its creation process. Evidence obtained from such analysis would provide useful forensic information to law enforcement, security, and intelligence agencies. Knowledge of image acquisition techniques can also help answer further forensic questions regarding the nature of additional processing that the image might have undergone after capture. %B Signal Processing Magazine, IEEE %V 26 %P 38 - 48 %8 2009/03// %@ 1053-5888 %G eng %N 2 %R 10.1109/MSP.2008.931076 %0 Journal Article %J Theory of Cryptography %D 2009 %T Composability and on-line deniability of authentication %A Dodis,Y. %A Katz, Jonathan %A Smith,A. %A Walfish,S. %X Protocols for deniable authentication achieve seemingly paradoxical guarantees: upon completion of the protocol the receiver is convinced that the sender authenticated the message, but neither party can convince anyone else that the other party took part in the protocol. We introduce and study on-line deniability, where deniability should hold even when one of the parties colludes with a third party during execution of the protocol. This turns out to generalize several realistic scenarios that are outside the scope of previous models.We show that a protocol achieves our definition of on-line deniability if and only if it realizes the message authentication functionality in the generalized universal composability framework; any protocol satisfying our definition thus automatically inherits strong composability guarantees. Unfortunately, we show that our definition is impossible to realize in the PKI model if adaptive corruptions are allowed (even if secure erasure is assumed). On the other hand, we show feasibility with respect to static corruptions (giving the first separation in terms of feasibility between the static and adaptive setting), and show how to realize a relaxation termed deniability with incriminating abort under adaptive corruptions. %B Theory of Cryptography %P 146 - 162 %8 2009/// %G eng %R 10.1007/978-3-642-00457-5_10 %0 Journal Article %J Engineering Analysis with Boundary Elements %D 2009 %T Computation of singular and hypersingular boundary integrals by Green identity and application to boundary value problems %A Seydou,F. %A Duraiswami, Ramani %A Seppänen,T. %A Gumerov, Nail A. %K Bounday integral method %K Green identity %K Hypersingular integrals %K Nyström method %K Singular integrals %X The problem of computing singular and hypersingular integrals involved in a large class of boundary value problems is considered. The method is based on Green's theorem for calculating the diagonal elements of the resulting discretized matrix using the Nyström discretization method. The method is successfully applied to classical boundary value problems. Convergence of the method is also discussed. %B Engineering Analysis with Boundary Elements %V 33 %P 1124 - 1131 %8 2009/08// %@ 0955-7997 %G eng %U http://www.sciencedirect.com/science/article/pii/S0955799709000447 %N 8–9 %R 10.1016/j.enganabound.2009.02.004 %0 Journal Article %J Multimedia Tools and Applications %D 2009 %T Computational linguistics for metadata building (CLiMB): using text mining for the automatic identification, categorization, and disambiguation of subject terms for image metadata %A Klavans,J. L %A Sheffield,C. %A Abels,E. %A Jimmy Lin %A Passonneau,R. %A Sidhu,T. %A Soergel,D. %X In this paper, we present a system using computational linguistic techniques to extract metadata for image access. We discuss the implementation, functionality and evaluation of an image catalogers’ toolkit, developed in the Computational Linguistics for Metadata Building (CLiMB) research project. We have tested components of the system, including phrase finding for the art and architecture domain, functional semantic labeling using machine learning, and disambiguation of terms in domain-specific text vis a vis a rich thesaurus of subject terms, geographic and artist names. We present specific results on disambiguation techniques and on the nature of the ambiguity problem given the thesaurus, resources, and domain-specific text resource, with a comparison of domain-general resources and text. Our primary user group for evaluation has been the cataloger expert with specific expertise in the fields of painting, sculpture, and vernacular and landscape architecture. %B Multimedia Tools and Applications %V 42 %P 115 - 138 %8 2009/// %G eng %N 1 %R 10.1007/s11042-008-0253-9 %0 Conference Paper %D 2009 %T Controlling data in the cloud: outsourcing computation without outsourcing control %A Chow, Richard %A Golle, Philippe %A Jakobsson, Markus %A Elaine Shi %A Staddon, Jessica %A Masuoka, Ryusuke %A Molina,Jesus %K Cloud computing %K privacy %K Security %X Cloud computing is clearly one of today's most enticing technology areas due, at least in part, to its cost-efficiency and flexibility. However, despite the surge in activity and interest, there are significant, persistent concerns about cloud computing that are impeding momentum and will eventually compromise the vision of cloud computing as a new IT procurement model. In this paper, we characterize the problems and their impact on adoption. In addition, and equally importantly, we describe how the combination of existing research thrusts has the potential to alleviate many of the concerns impeding adoption. In particular, we argue that with continued research advances in trusted computing and computation-supporting encryption, life in the cloud can be advantageous from a business intelligence standpoint over the isolated alternative that is more common today. %S CCSW '09 %I ACM %P 85 - 90 %8 2009 %@ 978-1-60558-784-4 %G eng %U http://doi.acm.org/10.1145/1655008.1655020 %0 Conference Paper %B INFOCOM 2009, IEEE %D 2009 %T CPM: Adaptive Video-on-Demand with Cooperative Peer Assists and Multicast %A Gopalakrishnan,V. %A Bhattacharjee, Bobby %A Ramakrishnan,K.K. %A Jana,R. %A Srivastava,D. %K assists;multicast;peer-to-peer %K communication;peer-to-peer %K computing;video %K CPM;cooperative %K demand; %K on %K parameters;video-on-demand;multicast %K peer %K schemes;synthetic %X We present CPM, a unified approach that exploits server multicast, assisted by peer downloads, to provide efficient video-on-demand (VoD) in a service provider environment. We describe our architecture and show how CPM is designed to dynamically adapt to a wide range of situations including highly different peer-upload bandwidths, content popularity, user request arrival patterns, video library size, and subscriber population. We demonstrate the effectiveness of CPM using simulations (based on an actual implementation codebase) across the range of situations described above and show that CPM does significantly better than traditional unicast, different forms of multicast, as well as peer-to-peer schemes. Along with synthetic parameters, we augment our experiments using data from a deployed VoD service to evaluate the performance of CPM. %B INFOCOM 2009, IEEE %P 91 - 99 %8 2009/04// %G eng %R 10.1109/INFCOM.2009.5061910 %0 Conference Paper %B Proceedings of the 27th international conference extended abstracts on Human factors in computing systems %D 2009 %T Creativity challenges and opportunities in social computing %A Fischer,Gerhard %A Jennings,Pamela %A Maher,Mary Lou %A Resnick,Mitchel %A Shneiderman, Ben %K creativity %K social computing %X There is a convergence in recent theories of creativity that go beyond characteristics and cognitive processes of individuals to recognize the importance of the social construction of creativity. In parallel, there has been a rise in social computing supporting the collaborative construction of knowledge. The panel will discuss the challenges and opportunities from the confluence of these two developments by bringing together the contrasting and controversial perspective of the individual panel members. It will synthesize from different perspectives an analytic framework to understand these new developments, and how to promote rigorous research methods and how to identify the unique challenges in developing evaluation and assessment methods for creativity research. %B Proceedings of the 27th international conference extended abstracts on Human factors in computing systems %S CHI EA '09 %I ACM %C New York, NY, USA %P 3283 - 3286 %8 2009/// %@ 978-1-60558-247-4 %G eng %U http://doi.acm.org/10.1145/1520340.1520470 %R 10.1145/1520340.1520470 %0 Book Section %B Engineering the User InterfaceEngineering the User Interface %D 2009 %T Creativity Support Tools: A Grand Challenge for HCI Researchers %A Shneiderman, Ben %X Human-computer interaction researchers can play a key role in designing, implementing, and evaluating a new generation of creativity support tools. Understanding creative processes, especially when mediated by user interfaces will remain a continuing challenge, especially in dealing with novice and expert users, across a variety of disciplines. The expected outcomes include (1) refined theories of technology-supported creative processes, (2) active discussion of user-oriented empirical research methods, (3) new software architectures, database management strategies, and networking technologies, and (4) improved user interfaces to support discover and innovation, especially in collaborative environments. %B Engineering the User InterfaceEngineering the User Interface %I Springer London %P 1 - 9 %8 2009/// %@ 978-1-84800-136-7 %G eng %U http://dx.doi.org/10.1007/978-1-84800-136-7_1 %0 Conference Paper %B Proceedings of the 35th SIGMOD international conference on Management of data %D 2009 %T Cross-tier, label-based security enforcement for web applications %A Corcoran,Brian J. %A Swamy,Nikhil %A Hicks, Michael W. %K compilers %K database programming %K security enforcement %K type systems %K web applications %X This paper presents SELinks, a programming language focused on building secure multi-tier web applications. SELinks provides a uniform programming model, in the style of LINQ and Ruby on Rails, with language syntax for accessing objects residing either in the database or at the server. Object-level security policies are expressed as fully-customizable, first-class labels which may themselves be subject to security policies. Access to labeled data is mediated via trusted, user-provided policy enforcement functions. SELinks has two novel features that ensure security policies are enforced correctly and efficiently. First, SELinks implements a type system called Fable that allows a protected object's type to refer to its protecting label. The type system can check that labeled data is never accessed directly by the program without first consulting the appropriate policy enforcement function. Second, SELinks compiles policy enforcement code to database-resident user-defined functions that can be called directly during query processing. Database-side checking avoids transferring data to the server needlessly, while still allowing policies to be expressed in a customizable and portable manner. Our experience with two sizable web applications, a modelhealth-care database and a secure wiki with fine-grained security policies, indicates that cross-tier policy enforcement in SELinks is flexible, relatively easy to use, and, when compared to a single-tier approach, improves throughput by nearly an order of magnitude. SELinks is freely available. %B Proceedings of the 35th SIGMOD international conference on Management of data %S SIGMOD '09 %I ACM %C New York, NY, USA %P 269 - 282 %8 2009/// %@ 978-1-60558-551-2 %G eng %U http://doi.acm.org/10.1145/1559845.1559875 %R 10.1145/1559845.1559875 %0 Journal Article %J Genome Biology %D 2009 %T CTCF binding site classes exhibit distinct evolutionary, genomic, epigenomic and transcriptomic features %A Essien,Kobby %A Vigneau,Sebastien %A Apreleva,Sofia %A Singh,Larry N. %A Bartolomei,Marisa S. %A Hannenhalli, Sridhar %X CTCF (CCCTC-binding factor) is an evolutionarily conserved zinc finger protein involved in diverse functions ranging from negative regulation of MYC, to chromatin insulation of the beta-globin gene cluster, to imprinting of the Igf2 locus. The 11 zinc fingers of CTCF are known to differentially contribute to the CTCF-DNA interaction at different binding sites. It is possible that the differences in CTCF-DNA conformation at different binding sites underlie CTCF's functional diversity. If so, the CTCF binding sites may belong to distinct classes, each compatible with a specific functional role. %B Genome Biology %V 10 %P R131 - R131 %8 2009/11/18/ %@ 1465-6906 %G eng %U http://genomebiology.com/2009/10/11/R131 %N 11 %R 10.1186/gb-2009-10-11-r131 %0 Journal Article %J Web and Wireless Geographical Information Systems %D 2009 %T Database and Representation Issues in Geographic Information Systems (GIS) %A Samet, Hanan %X A review is provided of some database and representation issues involved in the implementation of geographic information systems (GIS). %B Web and Wireless Geographical Information Systems %P 1 - 6 %8 2009/// %G eng %R 10.1007/978-3-642-10601-9_1 %0 Conference Paper %B Proceedings of the 18th conference on USENIX security symposium %D 2009 %T Detecting spammers with SNARE: spatio-temporal network-level automatic reputation engine %A Hao,Shuang %A Syed,Nadeem Ahmed %A Feamster, Nick %A Gray,Alexander G. %A Krasser,Sven %X Users and network administrators need ways to filter email messages based primarily on the reputation of the sender. Unfortunately, conventional mechanisms for sender reputation--notably, IP blacklists--are cumbersome to maintain and evadable. This paper investigates ways to infer the reputation of an email sender based solely on network-level features, without looking at the contents of a message. First, we study first-order properties of network-level features that may help distinguish spammers from legitimate senders. We examine features that can be ascertained without ever looking at a packet's contents, such as the distance in IP space to other email senders or the geographic distance between sender and receiver. We derive features that are lightweight, since they do not require seeing a large amount of email from a single IP address and can be gleaned without looking at an email's contents--many such features are apparent from even a single packet. Second, we incorporate these features into a classification algorithm and evaluate the classifier's ability to automatically classify email senders as spammers or legitimate senders. We build an automated reputation engine, SNARE, based on these features using labeled data from a deployed commercial spam-filtering system. We demonstrate that SNARE can achieve comparable accuracy to existing static IP blacklists: about a 70%detection rate for less than a 0.3%false positive rate. Third, we show how SNARE can be integrated into existing blacklists, essentially as a first-pass filter. %B Proceedings of the 18th conference on USENIX security symposium %S SSYM'09 %I USENIX Association %C Berkeley, CA, USA %P 101 - 118 %8 2009/// %G eng %U http://dl.acm.org/citation.cfm?id=1855768.1855775 %0 Journal Article %J Research in Microbiology %D 2009 %T Determination of relationships among non-toxigenic Vibrio cholerae O1 biotype El Tor strains from housekeeping gene sequences and ribotype patterns %A Mohapatra,Saswat S. %A Ramachandran,Dhanya %A Mantri,Chinmay K. %A Rita R Colwell %A Singh,Durg V. %K Housekeeping genes %K Ribotyping %K sequencing %K Vibrio cholerae %X Sequencing of three housekeeping genes, mdh, dnaE and recA, and ribotyping for seven non-toxigenic Vibrio cholerae O1 strains isolated from different geographic sources indicate a phylogenetic relationship among the strains. Results of MLST and ribotyping indicate a clear difference between three toxigenic strains (N16961, O395, and 569B) and three non-toxigenic strains from India (GS1, GS2, and GW87) and one Guam strain (X392), the latter of which were similar in both MLST and ribotyping, while two other non-toxigenic strains from the USA and India (2740-80 and OR69) appeared to be more closely related to toxigenic strains than to non-toxigenic strains, although this was not supported by ribotyping. These results provide clues to the emergence of toxigenic strains from a non-toxigenic progenitor by acquisition of virulence gene clusters. Results of split decomposition analysis suggest that widespread recombination occurs among the three housekeeping genes and that recombination plays an important role in the emergence of toxigenic strains of V. cholerae O1. %B Research in Microbiology %V 160 %P 57 - 62 %8 2009/01// %@ 0923-2508 %G eng %U http://www.sciencedirect.com/science/article/pii/S0923250808001794 %N 1 %R 10.1016/j.resmic.2008.10.008 %0 Conference Paper %B Proceedings of the 5th symposium on Dynamic languages %D 2009 %T Directing JavaScript with arrows %A Khoo,Yit Phang %A Hicks, Michael W. %A Foster, Jeffrey S. %A Sazawal,Vibha %K arrowlets %K arrows %K events %K javascript %K web programming %X JavaScript programmers make extensive use of event-driven programming to help build responsive web applications. However, standard approaches to sequencing events are messy, and often lead to code that is difficult to understand and maintain. We have found that arrows, a generalization of monads, are an elegant solution to this problem. Arrows allow us to easily write asynchronous programs in small, modular units of code, and flexibly compose them in many different ways, while nicely abstracting the details of asynchronous program composition. In this paper, we present Arrowlets, a new JavaScript library that offers arrows to the everyday JavaScript programmer. We show how to use Arrowlets to construct a variety of state machines, including state machines that branch and loop. We also demonstrate how Arrowlets separate computation from composition with examples such as a drag-and-drop handler and a bubblesort animation. %B Proceedings of the 5th symposium on Dynamic languages %S DLS '09 %I ACM %C New York, NY, USA %P 49 - 58 %8 2009/// %@ 978-1-60558-769-1 %G eng %U http://doi.acm.org/10.1145/1640134.1640143 %R 10.1145/1640134.1640143 %0 Conference Paper %B Data Engineering, 2009. ICDE '09. IEEE 25th International Conference on %D 2009 %T Distance Oracles for Spatial Networks %A Sankaranarayanan,J. %A Samet, Hanan %K B-tree;distance %K data %K database %K databases;tree %K distance;hash %K languages;query %K neighbor %K network %K network;transportation %K networks;spatial %K networks;well-separated %K oracles;epsiv-approximate %K pair %K processing;region %K processing;relational %K search;location-based %K search;relational %K services;real-time %K structures; %K system;road %K table;k-nearest %K technique;programming %X The popularity of location-based services and the need to do real-time processing on them has led to an interest in performing queries on transportation networks, such as finding shortest paths and finding nearest neighbors. The challenge is that these operations involve the computation of distance along a spatial network rather than "as the crow flies." In many applications an estimate of the distance is sufficient, which can be achieved by use of an oracle. An approximate distance oracle is proposed for spatial networks that exploits the coherence between the spatial position of vertices and the network distance between them. Using this observation, a distance oracle is introduced that is able to obtain the epsiv-approximate network distance between two vertices of the spatial network. The network distance between every pair of vertices in the spatial network is efficiently represented by adapting the well-separated pair technique to spatial networks. Initially, use is made of an epsilon-approximate distance oracle of size O(n/epsivd) that is capable of retrieving the approximate network distance in O(log n) time using a B-tree. The retrieval time can be theoretically reduced further to O(1) time by proposing another epsiv-approximate distance oracle of size O(n log n/epsivd) that uses a hash table. Experimental results indicate that the proposed technique is scalable and can be applied to sufficiently large road networks. A 10%-approximate oracle (epsiv = 0.1) on a large network yielded an average error of 0.9% with 90% of the answers making an error of 2% or less and an average retrieval timeof 68 mu seconds. Finally, a strategy for the integration of the distance oracle into any relational database system as well as using it to perform a variety of spatial queries such as region search, k-nearest neighbor search, and spatial joins on spatial networks is discussed. %B Data Engineering, 2009. ICDE '09. IEEE 25th International Conference on %P 652 - 663 %8 2009/04/29/2 %G eng %R 10.1109/ICDE.2009.53 %0 Journal Article %J Proceedings of the ICWSM %D 2009 %T Distinguishing knowledge vs social capital in social media with roles and context %A Barash,V. %A Smith,M. %A Getoor, Lise %A Welser,H. T %X Social media communities (e.g. Wikipedia, Flickr, Live Q&A) give rise to distinct types of content, foremost among which are relational content (discussion, chat) and factual content (answering questions, problem-solving). Both users and researchers are increasingly interested in developing strategies that can rapidly distinguish these types of content. While many text-based and structural strategies are possible, we extend two bodies of research that show how social context, and the social roles of answerers can predict content type. We test our framework on a dataset of manually labeled contributions to Microsoft's Live Q&A and find that it reliably extracts factual and relational messages from the data. %B Proceedings of the ICWSM %V 9 %8 2009/// %G eng %0 Conference Paper %B IEEE INFOCOM 2009 %D 2009 %T Distributed Strategies for Channel Allocation and Scheduling in Software-Defined Radio Networks %A Han,Bo %A Kumar,V. S.A %A Marathe,M. V %A Parthasarathy,S. %A Srinivasan, Aravind %K access hash function %K Channel allocation %K channel assignment algorithm %K channel capacity %K collision avoidance %K Computer science %K cryptography %K distributed algorithm %K distributed algorithms %K Educational institutions %K inductive-scheduling technique %K Interference %K interference set %K packet scheduling algorithm %K Peer to peer computing %K Radio network %K radio networks %K radiofrequency interference %K random oracle methodology %K scheduling %K Scheduling algorithm %K simultaneous channel allocation %K software radio %K software-defined radio wireless network capacity %K telecommunication congestion control %K telecommunication security %K Throughput %K wireless channels %K Wireless networks %X Equipping wireless nodes with multiple radios can significantly increase the capacity of wireless networks, by making these radios simultaneously transmit over multiple non-overlapping channels. However, due to the limited number of radios and available orthogonal channels, designing efficient channel assignment and scheduling algorithms in such networks is a major challenge. In this paper, we present provably-good distributed algorithms for simultaneous channel allocation of individual links and packet-scheduling, in software-defined radio (SDR) wireless networks. Our distributed algorithms are very simple to implement, and do not require any coordination even among neighboring nodes. A novel access hash function or random oracle methodology is one of the key drivers of our results. With this access hash function, each radio can know the transmitters' decisions for links in its interference set for each time slot without introducing any extra communication overhead between them. Further, by utilizing the inductive-scheduling technique, each radio can also backoff appropriately to avoid collisions. Extensive simulations demonstrate that our bounds are valid in practice. %B IEEE INFOCOM 2009 %I IEEE %P 1521 - 1529 %8 2009/04/19/25 %@ 978-1-4244-3512-8 %G eng %R 10.1109/INFCOM.2009.5062069 %0 Journal Article %J University of Maryland Tech Report: HCIL-2009-17 %D 2009 %T Do you know the way to SNA?: A process model for analyzing and visualizing social media data %A Hansen,D. L %A Rotman,D. %A Bonsignore,E. %A Milic-Frayling,N. %A Rodrigues,E.M. %A Smith,M. %A Shneiderman, Ben %X Traces of activity left by social media users can shed lighton individual behavior, social relationships, and community efficacy. Tools and processes to analyze social traces are essential for enabling practitioners to study and nurture meaningful and sustainable social interaction. Yet such tools and processes remain in their infancy. We conducted a study of 15 graduate students who were learning to apply Social Network Analysis (SNA) to data from online communities. Based on close observations of their emergent practices, we derived the Network Analysis and Visualization (NAV) process model and identified stages where intervention from peers, experts, and an SNA tool were most useful. We show how the NAV model informs the design of SNA tools and services, education practices, and support for social media practitioners. %B University of Maryland Tech Report: HCIL-2009-17 %8 2009/// %G eng %0 Conference Paper %B Proceedings of the 2009 ACM SIGPLAN conference on Programming language design and implementation %D 2009 %T Dynamic software updates: a VM-centric approach %A Subramanian,Suriya %A Hicks, Michael W. %A McKinley,Kathryn S. %K dynamic software updating %K garbage collection %K virtual machine technology %X Software evolves to fix bugs and add features. Stopping and restarting programs to apply changes is inconvenient and often costly. Dynamic software updating (DSU) addresses this problem by updating programs while they execute, but existing DSU systems for managed languages do not support many updates that occur in practice and are inefficient. This paper presents the design and implementation of Jvolve, a DSU-enhanced Java VM. Updated programs may add, delete, and replace fields and methods anywhere within the class hierarchy. Jvolve implements these updates by adding to and coordinating VM classloading, just-in-time compilation, scheduling, return barriers, on-stack replacement, and garbage collection. Jvolve, is safe: its use of bytecode verification and VM thread synchronization ensures that an update will always produce type-correct executions. Jvolve is flexible: it can support 20 of 22 updates to three open-source programs--Jetty web server, JavaEmailServer, and CrossFTP server--based on actual releases occurring over 1 to 2 years. Jvolve is efficient: performance experiments show that incurs no overhead during steady-state execution. These results demonstrate that this work is a significant step towards practical support for dynamic updates in virtual machines for managed languages. %B Proceedings of the 2009 ACM SIGPLAN conference on Programming language design and implementation %S PLDI '09 %I ACM %C New York, NY, USA %P 1 - 12 %8 2009/// %@ 978-1-60558-392-1 %G eng %U http://doi.acm.org/10.1145/1542476.1542478 %R 10.1145/1542476.1542478 %0 Journal Article %J Proceedings of the ACM Conference on Programming Language Design and Implementation (PLDI) %D 2009 %T Dynamic software updates for Java: A VM-centric approach %A Subramanian,S. %A Hicks, Michael W. %A McKinley,K.S. %X Software evolves to fix bugs and add features, but stopping andrestarting existing programs to take advantage of these changes can be inconvenient and costly. Dynamic software updating (DSU) addresses these problems by updating programs while they run. The challenge is to develop DSU infrastructure that is flexible, safe, and efficient—DSU should enable updates that are likely to occur in practice, and updated programs should be as reliable and as efficient as those started from scratch. This paper presents the design and implementation of a JVM we call JVOLVE that is enhanced with DSU support. The pa- per’s key insight is that flexible, safe, and efficient DSU can be supported by naturally extending existing VM services. By piggybacking on classloading and garbage collection, JVOLVE can flexibly support additions and replacements of fields and methods anywhere within the class hierarchy, and in a manner that may alter class signatures. By utilizing bytecode verifi- cation and thread synchronization support, JVOLVE can ensure that an applied update will never violate type-safety. Finally, by employing JIT compilation, all DSU-related overhead be- fore or after an update can be effectively eliminated. Using JVOLVE, we successfully applied dynamic continuous updates corresponding to 20 of the 22 releases that occurred over nearly two years’ time, one update per release, for three open-source programs, Jetty web server, JavaEmailServer, and CrossFTP server. Our results indicate that the VM is well-suited to sup- port practical DSU services. %B Proceedings of the ACM Conference on Programming Language Design and Implementation (PLDI) %8 2009/// %G eng %0 Conference Paper %B Social Network Analysis and Mining, 2009. ASONAM '09. International Conference on Advances in %D 2009 %T The Dynamics of Actor Loyalty to Groups in Affiliation Networks %A Sharara,H. %A Singh,L. %A Getoor, Lise %A Mann,J. %K actor %K bill %K co-sponsorship %K dynamics;social %K groups;affiliation %K loyalty;affiliation %K network;senate %K network;temporal %K networks;dolphin %K sciences; %X In this paper, we introduce a method for analyzing the temporal dynamics of affiliation networks. We define affiliation groups which describe temporally related subsets of actors and describe an approach for exploring changing memberships in these affiliation groups over time. To model the dynamic behavior in these networks, we consider the concept of loyalty and introduce a measure that captures an actorpsilas loyalty to an affiliation group as the degree of dasiacommitmentpsila an actor shows to the group over time. We evaluate our measure using two real world affiliation networks: a senate bill co-sponsorship network and a dolphin network. The results show how the behavior of actors in different affiliation groups change dynamically over time, reinforcing the utility of our measure for understanding the loyalty of actors to time-varying affiliation groups. %B Social Network Analysis and Mining, 2009. ASONAM '09. International Conference on Advances in %P 101 - 106 %8 2009/07// %G eng %R 10.1109/ASONAM.2009.27 %0 Journal Article %J KSII Trans. Internet and Information Systems %D 2009 %T Efficient peer-to-Peer lookup in multi-hop wireless networks %A Shin,M. %A Arbaugh, William A. %B KSII Trans. Internet and Information Systems %V 3 %P 5 - 25 %8 2009/// %G eng %N 1 %0 Conference Paper %B Computer Vision, 2009 IEEE 12th International Conference on %D 2009 %T Efficient subset selection via the kernelized Rényi distance %A Srinivasan,B.V. %A Duraiswami, Ramani %X With improved sensors, the amount of data available in many vision problems has increased dramatically and allows the use of sophisticated learning algorithms to perform inference on the data. However, since these algorithms scale with data size, pruning the data is sometimes necessary. The pruning procedure must be statistically valid and a representative subset of the data must be selected without introducing selection bias. Information theoretic measures have been used for sampling the data, retaining its original information content. We propose an efficient Rényi entropy based subset selection algorithm. The algorithm is first validated and then applied to two sample applications where machine learning and data pruning are used. In the first application, Gaussian process regression is used to learn object pose. Here it is shown that the algorithm combined with the subset selection is significantly more efficient. In the second application, our subset selection approach is used to replace vector quantization in a standard object recognition algorithm, and improvements are shown. %B Computer Vision, 2009 IEEE 12th International Conference on %P 1081 - 1088 %8 2009/// %G eng %0 Journal Article %J SPIE/IS&T Media Forensics and Security %D 2009 %T Enabling search over encrypted multimedia databases %A Lu,W. %A Swaminathan,A. %A Varna,A.L. %A Wu,M. %X Performing information retrieval tasks while preserving data confidentiality is a desirable capability when adatabase is stored on a server maintained by a third-party service provider. This paper addresses the problem of enabling content-based retrieval over encrypted multimedia databases. Search indexes, along with multimedia documents, are first encrypted by the content owner and then stored onto the server. Through jointly apply- ing cryptographic techniques, such as order preserving encryption and randomized hash functions, with image processing and information retrieval techniques, secure indexing schemes are designed to provide both privacy protection and rank-ordered search capability. Retrieval results on an encrypted color image database and se- curity analysis of the secure indexing schemes under different attack models show that data confidentiality can be preserved while retaining very good retrieval performance. This work has promising applications in secure multimedia management. %B SPIE/IS&T Media Forensics and Security %P 7254 - 18 %8 2009/// %G eng %0 Journal Article %J Jisuanji Gongcheng/ Computer Engineering %D 2009 %T Fast and Accurate Method of Uniform-colored Video Text Extraction %A Shen,R.D. %A Li,S.F. %A Zhuolin Jiang %X For most video text is rich in edge,uniform-colored and horizontally ranged,the candidate text areas in video image are determined by using a fast deriche edge based algorithm,and the accurate binary text images are extracted from the areas by using a color based algorithm.The experimental results show that the method is effective in text extraction in complex-background video frame,and has higher processing speed and better extraction result in comparison with color alone based method. %B Jisuanji Gongcheng/ Computer Engineering %V 35 %8 2009/// %G eng %N 9 %0 Conference Paper %B IEEE INFOCOM 2009 %D 2009 %T Fighting Spam with the NeighborhoodWatch DHT %A Bender,A. %A Sherwood,R. %A Monner,D. %A Goergen,N. %A Spring, Neil %A Bhattacharjee, Bobby %K Communications Society %K computer crime %K cryptography %K Databases %K IP addresses %K IP networks %K on-line trusted authority %K Peer to peer computing %K peer-to-peer computing %K peer-to-peer distributed hash table %K Postal services %K Relays %K Resilience %K Routing %K Security %K table size routing %K Unsolicited electronic mail %X In this paper, we present DHTBL, an anti-spam blacklist built upon a novel secure distributed hash table (DHT). We show how DHTBL can be used to replace existing DNS-based blacklists (DNSBLs) of IP addresses of mail relays that forward spam. Implementing a blacklist on a DHT improves resilience to DoS attacks and secures message delivery, when compared to DNSBLs. However, due to the sensitive nature of the blacklist, storing the data in a peer-to-peer DHT would invite attackers to infiltrate the system. Typical DHTs can withstand fail-stop failures, but malicious nodes may provide incorrect routing information, refuse to return published items, or simply ignore certain queries. The neighborhoodwatch DHT is resilient to malicious nodes and maintains the O(logiV) bounds on routing table size and expected lookup time. NeighborhoodWatch depends on two assumptions in order to make these guarantees: (1) the existence of an on-line trusted authority that periodically contacts and issues signed certificates to each node, and (2) for every sequence of k + 1 consecutive nodes in the ID space, at least one is alive and non-malicious. We show how NeighborhoodWatch maintains many of its security properties even when the second assumption is violated. Honest nodes in NeighborhoodWatch can detect malicious behavior and expel the responsible nodes from the DHT. %B IEEE INFOCOM 2009 %I IEEE %P 1755 - 1763 %8 2009/04/19/25 %@ 978-1-4244-3512-8 %G eng %R 10.1109/INFCOM.2009.5062095 %0 Conference Paper %B IEEE Symposium on Visual Analytics Science and Technology, 2009. VAST 2009 %D 2009 %T Finding comparable temporal categorical records: A similarity measure with an interactive visualization %A Wongsuphasawat,K. %A Shneiderman, Ben %K data visualisation %K Educational institutions %K Feedback %K Information retrieval %K interactive search tool %K interactive systems %K interactive visualization tool %K large databases %K M&M Measure %K Match & Mismatch measure %K Medical services %K numerical time series %K parameters customization %K Particle measurements %K Similan %K similarity measure %K Similarity Search %K temporal categorical databases %K Temporal Categorical Records %K temporal databases %K Testing %K Time measurement %K time series %K transportation %K usability %K very large databases %K visual databases %K Visualization %X An increasing number of temporal categorical databases are being collected: Electronic Health Records in healthcare organizations, traffic incident logs in transportation systems, or student records in universities. Finding similar records within these large databases requires effective similarity measures that capture the searcher's intent. Many similarity measures exist for numerical time series, but temporal categorical records are different. We propose a temporal categorical similarity measure, the M&M (Match & Mismatch) measure, which is based on the concept of aligning records by sentinel events, then matching events between the target and the compared records. The M&M measure combines the time differences between pairs of events and the number of mismatches. To accom-modate customization of parameters in the M&M measure and results interpretation, we implemented Similan, an interactive search and visualization tool for temporal categorical records. A usability study with 8 participants demonstrated that Similan was easy to learn and enabled them to find similar records, but users had difficulty understanding the M&M measure. The usability study feedback, led to an improved version with a continuous timeline, which was tested in a pilot study with 5 participants. %B IEEE Symposium on Visual Analytics Science and Technology, 2009. VAST 2009 %I IEEE %P 27 - 34 %8 2009/10/12/13 %@ 978-1-4244-5283-5 %G eng %R 10.1109/VAST.2009.5332595 %0 Journal Article %J Information Forensics and Security, IEEE Transactions on %D 2009 %T Fingerprinting Compressed Multimedia Signals %A Varna,A.L. %A He,Shan %A Swaminathan,A. %A M. Wu %K coding; %K coding;copy %K collusion;digital %K communication;video %K compression;digital %K compression;video %K dither;collusion %K domain %K fingerprinting;anti-collusion %K fingerprinting;multimedia %K Gaussian %K management;multimedia %K protection;data %K resistance;compressed %K rights %X Digital fingerprinting is a technique to deter unauthorized redistribution of multimedia content by embedding a unique identifying signal in each legally distributed copy. The embedded fingerprint can later be extracted and used to trace the originator of an unauthorized copy. A group of users may collude and attempt to create a version of the content that cannot be traced back to any of them. As multimedia data is commonly stored in compressed form, this paper addresses the problem of fingerprinting compressed signals. Analysis is carried out to show that due to the quantized nature of the host signal and the embedded fingerprint, directly extending traditional fingerprinting techniques for uncompressed signals to the compressed case leads to low collusion resistance. To overcome this problem and improve the collusion resistance, a new technique for fingerprinting compressed signals called Anti-Collusion Dither (ACD) is proposed, whereby a random dither signal is added to the compressed host before embedding so as to make the effective host signal appear more continuous. The proposed technique is shown to reduce the accuracy with which attackers can estimate the host signal, and from an information theoretic perspective, the proposed ACD technique increases the maximum number of users that can be supported by the fingerprinting system under a given attack. Both analytical and experimental studies confirm that the proposed technique increases the probability of identifying a guilty user and can approximately quadruple the collusion resistance compared to conventional Gaussian fingerprinting. %B Information Forensics and Security, IEEE Transactions on %V 4 %P 330 - 345 %8 2009/09// %@ 1556-6013 %G eng %N 3 %R 10.1109/TIFS.2009.2025860 %0 Conference Paper %B International Conference on Computational Science and Engineering, 2009. CSE '09 %D 2009 %T First Steps to Netviz Nirvana: Evaluating Social Network Analysis with NodeXL %A Bonsignore,E. M %A Dunne,C. %A Rotman,D. %A Smith,M. %A Capone,T. %A Hansen,D. L %A Shneiderman, Ben %K Computer science %K computer science education %K data visualisation %K Data visualization %K Educational institutions %K graph drawing %K graph layout algorithm %K Information services %K Information Visualization %K Internet %K Libraries %K Microsoft Excel open-source template %K MILC %K multi-dimensional in-depth long-term case studies %K Netviz Nirvana %K NodeXL %K Open source software %K Programming profession %K SNA %K social network analysis %K Social network services %K social networking (online) %K spreadsheet programs %K structural relationship %K teaching %K visual analytics %K visualization tool %K Web sites %X Social Network Analysis (SNA) has evolved as a popular, standard method for modeling meaningful, often hidden structural relationships in communities. Existing SNA tools often involve extensive pre-processing or intensive programming skills that can challenge practitioners and students alike. NodeXL, an open-source template for Microsoft Excel, integrates a library of common network metrics and graph layout algorithms within the familiar spreadsheet format, offering a potentially low-barrier-to-entry framework for teaching and learning SNA. We present the preliminary findings of 2 user studies of 21 graduate students who engaged in SNA using NodeXL. The majority of students, while information professionals, had little technical background or experience with SNA techniques. Six of the participants had more technical backgrounds and were chosen specifically for their experience with graph drawing and information visualization. Our primary objectives were (1) to evaluate NodeXL as an SNA tool for a broad base of users and (2) to explore methods for teaching SNA. Our complementary dual case-study format demonstrates the usability of NodeXL for a diverse set of users, and significantly, the power of a tightly integrated metrics/visualization tool to spark insight and facilitate sense-making for students of SNA. %B International Conference on Computational Science and Engineering, 2009. CSE '09 %I IEEE %V 4 %P 332 - 339 %8 2009/08/29/31 %@ 978-1-4244-5334-4 %G eng %R 10.1109/CSE.2009.120 %0 Conference Paper %B Proceedings of the Fourth Workshop on Statistical Machine Translation %D 2009 %T Fluency, adequacy, or HTER?: exploring different human judgments with a tunable MT metric %A Snover,Matthew %A Madnani,Nitin %A Dorr, Bonnie J %A Schwartz,Richard %X Automatic Machine Translation (MT) evaluation metrics have traditionally been evaluated by the correlation of the scores they assign to MT output with human judgments of translation performance. Different types of human judgments, such as Fluency, Adequacy, and HTER, measure varying aspects of MT performance that can be captured by automatic MT metrics. We explore these differences through the use of a new tunable MT metric: TER-Plus, which extends the Translation Edit Rate evaluation metric with tunable parameters and the incorporation of morphology, synonymy and paraphrases. TER-Plus was shown to be one of the top metrics in NIST's Metrics MATR 2008 Challenge, having the highest average rank in terms of Pearson and Spearman correlation. Optimizing TER-Plus to different types of human judgments yields significantly improved correlations and meaningful changes in the weight of different types of edits, demonstrating significant differences between the types of human judgments. %B Proceedings of the Fourth Workshop on Statistical Machine Translation %S StatMT '09 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 259 - 268 %8 2009/// %G eng %U http://dl.acm.org/citation.cfm?id=1626431.1626480 %0 Conference Paper %D 2009 %T Formal Representation of Product Design Specifications for Validating Product Design %A Weissman,Alexander %A Gupta, Satyandra K. %A Fiorentini,Xenia %A Sudarsan,Rachuri %A Sriram,Ram %X As collaborative efforts in electro-mechanical design have scaled to large, distributed groups working for years on a single product, an increasingly large gulf has developed between the original stated goals of the project and the final design solution. It has thus become necessary to validate the final design solution against a set of requirements to ensure that these goals have, in fact, been met. This process has become tedious for complex products with hundreds of design aspects and requirements. By formalizing the representation of requirements and the design solution, tools can be developed to a large extent automatically perform this validation. In this paper, we propose a formal approach for relating product requirements to the design solution. First, we present a formal model for representing product requirements. Then, we introduce the Core Product Model (CPM) and the Open Assembly Model (OAM) for representing the design solution. Finally, we link these models formally and provide an example with an actual consumer device. %I ASME %P 1411 - 1422 %8 2009/// %@ 978-0-7918-4899-9 %G eng %U http://link.aip.org/link/ASMECP/v2009/i48999/p1411/s1&Agg=doi %R 10.1115/DETC2009-87307 %0 Journal Article %J Proceedings of the 9th SIAM International Conference on Data Mining %D 2009 %T FutureRank: Ranking scientific articles by predicting their future PageRank %A Sayyadi,H. %A Getoor, Lise %X The dynamic nature of citation networks makes the taskof ranking scientific articles hard. Citation networks are continually evolving because articles obtain new citations every day. For ranking scientific articles, we can define the popularity or prestige of a paper based on the number of past citations at the user query time; however, we argue that what is most useful is the expected future references. We define a new measure, FutureRank, which is the expected future PageRank score based on citations that will be obtained in the future. In addition to making use of the citation network, FutureRank uses the authorship network and the publication time of the article in order to predict future citations. Our experiments compare FutureRank with existing approaches, and show that FutureRank is accurate and useful for finding and ranking publications. %B Proceedings of the 9th SIAM International Conference on Data Mining %P 533 - 544 %8 2009/// %G eng %0 Journal Article %J DiabetesDiabetes %D 2009 %T Gene Profiling of Human Adipose Tissue During Evoked Inflammation In Vivo %A Shah,Rachana %A Lu,Yun %A Hinkle,Christine C %A McGillicuddy,Fiona C %A Kim,Roy %A Hannenhalli, Sridhar %A Cappola,Thomas P. %A Heffron,Sean %A Wang,XingMei %A Mehta,Nehal N %A Putt,Mary %A Reilly,Muredach P %X OBJECTIVE Adipose inflammation plays a central role in obesity-related metabolic and cardiovascular complications. However, few human adipose-secreted proteins are known to mediate these processes. We hypothesized that microarray mRNA profiling of human adipose during evoked inflammation could identify novel adipocytokines.RESEARCH DESIGN AND METHODS Healthy human volunteers (n = 14) were treated with intravenous endotoxin (3 ng/kg lipopolysaccharide [LPS]) and underwent subcutaneous adipose biopsies before and after LPS. On Affymetrix U133Plus 2.0 arrays, adipose mRNAs modulated >1.5-fold (with P < 0.00001) were selected. SignalP 3.0 and SecretomeP 2.0 identified genes predicted to encode secreted proteins. Of these, 86 candidates were chosen for validation in adipose from an independent human endotoxemia protocol (N = 7, with 0.6 ng/kg LPS) and for exploration of cellular origin in primary human adipocytes and macrophages in vitro. RESULTS Microarray identified 776 adipose genes modulated by LPS; 298 were predicted to be secreted. Of detectable prioritized genes, 82 of 85 (96% [95% CI 90–99]) were upregulated (fold changes >1.0) during the lower-dose (LPS 0.6 ng/kg) validation study and 51 of 85 (59% [49–70]) were induced greater than 1.5-fold. Treatment of primary adipocytes with LPS and macrophage polarization to M1 proinflammatory phenotype increased expression by 1.5-fold for 58 and 73% of detectable genes, respectively. CONCLUSIONS We demonstrate that evoked inflammation of human adipose in vivo modulated expression of multiple genes likely secreted by adipocytes and monocytes. These included established adipocytokines and chemokines implicated in recruitment and activation of lymphocytes, adhesion molecules, antioxidants, and several novel genes with unknown function. Such candidates may represent biomarkers and therapeutic targets for obesity-related complications. %B DiabetesDiabetes %V 58 %P 2211 - 2219 %8 2009/10/01/ %@ 0012-1797, 1939-327X %G eng %U http://diabetes.diabetesjournals.org/content/58/10/2211 %N 10 %R 10.2337/db09-0256 %0 Journal Article %J Nature %D 2009 %T Genome assortment, not serogroup, defines Vibrio cholerae pandemic strains %A Brettin,Thomas S[Los Alamos National Laboratory %A Bruce,David C[Los Alamos National Laboratory %A Challacombe,Jean F[Los Alamos National Laboratory %A Detter,John C[Los Alamos National Laboratory %A Han,Cliff S[Los Alamos National Laboratory %A Munik,A. C[Los Alamos National Laboratory %A Chertkov,Olga[Los Alamos National Laboratory %A Meincke,Linda[Los Alamos National Laboratory %A Saunders,Elizabeth[Los Alamos National Laboratory %A Choi,Seon Y[SEOUL NATL UNIV %A Haley,Bradd J[U MARYLAND %A Taviani,Elisa[U MARYLAND %A Jeon,Yoon-Seong[INTL VACCINE INST SEOUL %A Kim,Dong Wook[INTL VACCINE INST SEOUL %A Lee,Jae-Hak[SEOUL NATL UNIV %A Walters,Ronald A[PNNL %A Hug,Anwar[NATL INST CHOLERIC ENTERIC DIS %A Rita R Colwell %K 59; CHOLERA; GENES; GENETICS; GENOTYPE; ISLANDS; ORIGIN; PHENOTYPE; PUBLIC HEALTH; RECOMBINATION; STRAINS; TOXINS %X Vibrio cholerae, the causative agent of cholera, is a bacterium autochthonous to the aquatic environment, and a serious public health threat. V. cholerae serogroup O1 is responsible for the previous two cholera pandemics, in which classical and El Tor biotypes were dominant in the 6th and the current 7th pandemics, respectively. Cholera researchers continually face newly emerging and re-emerging pathogenic clones carrying combinations of new serogroups as well as of phenotypic and genotypic properties. These genotype and phenotype changes have hampered control of the disease. Here we compare the complete genome sequences of 23 strains of V. cholerae isolated from a variety of sources and geographical locations over the past 98 years in an effort to elucidate the evolutionary mechanisms governing genetic diversity and genesis of new pathogenic clones. The genome-based phylogeny revealed 12 distinct V. cholerae phyletic lineages, of which one, designated the V. cholerae core genome (CG), comprises both O1 classical and EI Tor biotypes. All 7th pandemic clones share nearly identical gene content, i.e., the same genome backbone. The transition from 6th to 7th pandemic strains is defined here as a 'shift' between pathogenic clones belonging to the same O1 serogroup, but from significantly different phyletic lineages within the CG clade. In contrast, transition among clones during the present 7th pandemic period can be characterized as a 'drift' between clones, differentiated mainly by varying composition of laterally transferred genomic islands, resulting in emergence of variants, exemplified by V.cholerae serogroup O139 and V.cholerae O1 El Tor hybrid clones that produce cholera toxin of classical biotype. Based on the comprehensive comparative genomics presented in this study it is concluded that V. cholerae undergoes extensive genetic recombination via lateral gene transfer, and, therefore, genome assortment, not serogroup, should be used to define pathogenic V. cholerae clones. %B Nature %8 2009/// %G eng %U http://www.osti.gov/energycitations/servlets/purl/962365-icnke9/ %0 Journal Article %J Managing and Mining Uncertain Data %D 2009 %T Graphical models for uncertain data %A Deshpande, Amol %A Getoor, Lise %A Sen,P. %B Managing and Mining Uncertain Data %P 77 - 77 %8 2009/// %G eng %0 Journal Article %J Information Systems %D 2009 %T GrouPeer: Dynamic clustering of P2P databases %A Kantere,Verena %A Tsoumakos,Dimitrios %A Sellis,Timos %A Roussopoulos, Nick %K Peer-to-Peer databases %K Query reformulation in P2P databases %K Semantics in P2P query answering %K Structured data in unstructured P2P overlays %X Sharing structured data in a P2P network is a challenging problem, especially in the absence of a mediated schema. The standard practice of answering a consecutively rewritten query along the propagation path often results in significant loss of information. On the opposite, the use of mediated schemas requires human interaction and global agreement, both during creation and maintenance. In this paper we present GrouPeer, an adaptive, automated approach to both issues in the context of unstructured P2P database overlays. By allowing peers to individually choose which rewritten version of a query to answer and evaluate the received answers, information-rich sources left hidden otherwise are discovered. Gradually, the overlay is restructured as semantically similar peers are clustered together. Experimental results show that our technique produces very accurate answers and builds clusters that are very close to the optimal ones by contacting a very small number of nodes in the overlay. %B Information Systems %V 34 %P 62 - 86 %8 2009/03// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306437908000252 %N 1 %R 16/j.is.2008.04.002 %0 Journal Article %J The fourth paradigm: data-intensive scientific discovery %D 2009 %T HEALTH AND WELLBEING %A Gillam,M. %A Feied,C. %A MOODY,E. %A Shneiderman, Ben %A Smith,M. %A DICKASON,J. %B The fourth paradigm: data-intensive scientific discovery %P 57 - 57 %8 2009/// %G eng %0 Conference Paper %B Computer Vision, 2009 IEEE 12th International Conference on %D 2009 %T Human detection using partial least squares analysis %A Schwartz,William Robson %A Kembhavi,Aniruddha %A Harwood,David %A Davis, Larry S. %X Significant research has been devoted to detecting people in images and videos. In this paper we describe a human detection method that augments widely used edge-based features with texture and color information, providing us with a much richer descriptor set. This augmentation results in an extremely high-dimensional feature space (more than 170,000 dimensions). In such high-dimensional spaces, classical machine learning algorithms such as SVMs are nearly intractable with respect to training. Furthermore, the number of training samples is much smaller than the dimensionality of the feature space, by at least an order of magnitude. Finally, the extraction of features from a densely sampled grid structure leads to a high degree of multicollinearity. To circumvent these data characteristics, we employ Partial Least Squares (PLS) analysis, an efficient dimensionality reduction technique, one which preserves significant discriminative information, to project the data onto a much lower dimensional subspace (20 dimensions, reduced from the original 170,000). Our human detection system, employing PLS analysis over the enriched descriptor set, is shown to outperform state-of-the-art techniques on three varied datasets including the popular INRIA pedestrian dataset, the low-resolution gray-scale DaimlerChrysler pedestrian dataset, and the ETHZ pedestrian dataset consisting of full-length videos of crowded scenes. %B Computer Vision, 2009 IEEE 12th International Conference on %P 24 - 31 %8 2009/10/29/2 %G eng %R 10.1109/ICCV.2009.5459205 %0 Conference Paper %B Proc. of workshop on Hot Topics in Networks (HotNets-VIII) %D 2009 %T Identifying close friends on the internet %A Baden,R. %A Spring, Neil %A Bhattacharjee, Bobby %X Online Social Networks (OSNs) encourage users to createan online presence that reflects their offline identity. OSNs create the illusion that these online accounts correspond to the correct offline person, but in reality the OSN lacks the re- sources to detect impersonation. We propose that OSN users identify each other based on interaction and experience. We believe that impersonation can be thwarted by users who possess exclusive shared knowledge, secret informa- tion shared only between a pair of OSN friends. We describe existing protocols that use shared secrets to exchange public keys without revealing those secrets to attackers. We present results from a user study on Facebook to show that users do share exclusive knowledge with their Facebook friends and attackers are rarely able to guess that knowledge. Finally, we show that friend identification can be extended using a web of trust built on the OSN friend graph. %B Proc. of workshop on Hot Topics in Networks (HotNets-VIII) %8 2009/// %G eng %0 Journal Article %J Proceedings of DigCCurr2009 Digital Curation: Practice, Promise and Prospects %D 2009 %T An Implementation of the Audit Control Environment (ACE) to Support the Long Term Integrity of Digital Archives %A Smorul,M. %A Song,S. %A JaJa, Joseph F. %X In this paper, we describe the implementation of the AuditControl Environment (ACE)[1] system that provides a scalable, auditable platform for ensuring the integrity of digital archival holdings. The core of ACE is a small integrity token issued for each monitored item, which is part of a larger, externally auditable cryptographic system. Two components that describe this system, an Audit Manager and Integrity Management Service, have been developed and released. The Audit Manager component is designed to be installed locally at the archive, while the Integrity Management Service is a centralized, publically available service. ACE allows for the monitoring of collections on a variety of disk and grid based storage systems. Each collection in ACE is subject to monitoring based on a customizable policy. The released ACE Version 1.0 has been tested extensively on a wide variety of collections in both centralized and distributed environments. %B Proceedings of DigCCurr2009 Digital Curation: Practice, Promise and Prospects %P 164 - 164 %8 2009/// %G eng %0 Conference Paper %D 2009 %T Implicit authentication for mobile devices %A Jakobsson, Markus %A Elaine Shi %A Golle, Philippe %A Chow, Richard %X We introduce the notion of implicit authentication - the ability to authenticate mobile users based on actions they would carry out anyway. We develop a model for how to perform implicit authentication, and describe experiments aimed at assessing the benefits of our techniques. Our preliminary findings support that this is a meaningful approach, whether used to increase usability or increase security. %S HotSec'09 %I USENIX Association %P 9 - 9 %8 2009 %G eng %U http://dl.acm.org/citation.cfm?id=1855628.1855637 %0 Journal Article %J University of Maryland, HCIL Tech Report HCIL-2009-13 %D 2009 %T Improving graph drawing readability by incorporating readability metrics: A software tool for network analysts %A Dunne,C. %A Shneiderman, Ben %X Designing graph drawings that effectively communicate the under-lying network is challenging as for every network there are many potential unintelligible or even misleading drawings. Automated graph layout algorithms have helped, but frequently generate in- effective drawings. In order to build awareness of effective graph drawing strategies, we detail readability metrics on a [0,1] contin- uous scale for node occlusion, edge crossing, edge crossing angle, and edge tunneling and summarize many more. Additionally, we define new node & edge readability metrics to provide more lo- calized identification of where improvement is needed. These are implemented in SocialAction, a tool for social network analysis, in order to direct users towards poor areas of the drawing and provide real-time readability metric feedback as users manipulate it. These contributions are aimed at heightening the awareness of network analysts that the images they share or publish could be of higher quality, so that readers could extract relevant information. %B University of Maryland, HCIL Tech Report HCIL-2009-13 %8 2009/// %G eng %0 Journal Article %J Recommender Systems & the Social Web %D 2009 %T Improving recommendation accuracy by clustering social networks with trust %A DuBois,T. %A Golbeck,J. %A Kleint,J. %A Srinivasan, Aravind %X Social trust relationships between users in social networksspeak to the similarity in opinions between the users, both in general and in important nuanced ways. They have been used in the past to make recommendations on the web. New trust metrics allow us to easily cluster users based on trust. In this paper, we investigate the use of trust clusters as a new way of improving recommendations. Previous work on the use of clusters has shown the technique to be relatively un- successful, but those clusters were based on similarity rather than trust. Our results show that when trust clusters are integrated into memory-based collaborative filtering algo- rithms, they lead to statistically significant improvements in accuracy. In this paper we discuss our methods, experi- ments, results, and potential future applications of the tech- nique. %B Recommender Systems & the Social Web %P 1 - 8 %8 2009/// %G eng %0 Conference Paper %B Computer Vision, 2009 IEEE 12th International Conference on %D 2009 %T Incremental Multiple Kernel Learning for object recognition %A Kembhavi,Aniruddha %A Siddiquie,Behjat %A Miezianko,Roland %A McCloskey,Scott %A Davis, Larry S. %X A good training dataset, representative of the test images expected in a given application, is critical for ensuring good performance of a visual categorization system. Obtaining task specific datasets of visual categories is, however, far more tedious than obtaining a generic dataset of the same classes. We propose an Incremental Multiple Kernel Learning (IMKL) approach to object recognition that initializes on a generic training database and then tunes itself to the classification task at hand. Our system simultaneously updates the training dataset as well as the weights used to combine multiple information sources. We demonstrate our system on a vehicle classification problem in a video stream overlooking a traffic intersection. Our system updates itself with images of vehicles in poses more commonly observed in the scene, as well as with image patches of the background, leading to an increase in performance. A considerable change in the kernel combination weights is observed as the system gathers scene specific training data over time. The system is also seen to adapt itself to the illumination change in the scene as day transitions to night. %B Computer Vision, 2009 IEEE 12th International Conference on %P 638 - 645 %8 2009/10/29/2 %G eng %R 10.1109/ICCV.2009.5459179 %0 Journal Article %J Proceedings of the VLDB Endowment %D 2009 %T Index interactions in physical design tuning: modeling, analysis, and applications %A Schnaitter,Karl %A Polyzotis,Neoklis %A Getoor, Lise %X One of the key tasks of a database administrator is to optimize the set of materialized indices with respect to the current workload. To aid administrators in this challenging task, commercial DBMSs provide advisors that recommend a set of indices based on a sample workload. It is left for the administrator to decide which of the recommended indices to materialize and when. This decision requires some knowledge of how the indices benefit the workload, which may be difficult to understand if there are any dependencies or interactions among indices. Unfortunately, advisors do not provide this crucial information as part of the recommendation. Motivated by this shortcoming, we propose a framework and associated tools that can help an administrator understand the interactions within the recommended set of indices. We formalize the notion of index interactions and develop a novel algorithm to identify the interaction relationships that exist within a set of indices. We present experimental results with a prototype implementation over IBM DB2 that demonstrate the efficiency of our approach. We also describe two new database tuning tools that utilize information about index interactions. The first tool visualizes interactions based on a partitioning of the index-set into non-interacting subsets, and the second tool computes a schedule that materializes the indices over several maintenance windows with maximal overall benefit. In both cases, we provide strong analytical results showing that index interactions can enable enhanced functionality. %B Proceedings of the VLDB Endowment %V 2 %P 1234 - 1245 %8 2009/08// %@ 2150-8097 %G eng %U http://dl.acm.org/citation.cfm?id=1687627.1687766 %N 1 %0 Book Section %B Human-Computer InteractionHuman-Computer Interaction %D 2009 %T Information Visualization %A Card,Stuart %E Sears,Andrew %E Jacko,Julie %B Human-Computer InteractionHuman-Computer Interaction %I CRC Press %V 20093960 %P 181 - 215 %8 2009/03/02/ %@ 978-1-4200-8885-4, 978-1-4200-8886-1 %G eng %U http://www.crcnetbase.com/doi/abs/10.1201/9781420088861.ch10 %0 Journal Article %J ACM Trans. Comput.-Hum. Interact. %D 2009 %T The Ins and Outs of Home Networking: The Case for Useful and Usable Domestic Networking %A Grinter, Rebecca E. %A Edwards, W. Keith %A Marshini Chetty %A Poole, Erika S. %A Sung, Ja-Young %A Yang, Jeonghwa %A Crabtree, Andy %A Tolmie, Peter %A Rodden, Tom %A Greenhalgh, Chris %A Benford, Steve %K home networking %K Human computer interaction %X Householders are increasingly adopting home networking as a solution to the demands created by the presence of multiple computers, devices, and the desire to access the Internet. However, current network solutions are derived from the world of work (and initially the military) and provide poor support for the needs of the home. We present the key findings to emerge from empirical studies of home networks in the UK and US. The studies reveal two key kinds of work that effective home networking relies upon: one, the technical work of setting up and maintaining the home network, and the other, the collaborative and socially organized work of the home which the network is embedded in and supports. The two are thoroughly intertwined and rely upon one another for their realization, yet neither is adequately supported by current networking technologies and applications. Explication of the “work to make the home network work” opens up the design space for the continued integration of the home network in domestic life and elaboration of future support. Key issues for development include the development of networking facilities that do not require advanced networking knowledge, that are flexible and support the local social order of the home and the evolution of its routines, and which ultimately make the home network visible and accountable to household members. %B ACM Trans. Comput.-Hum. Interact. %V 16 %P 8:1 - 8:28 %8 2009/06// %@ 1073-0516 %G eng %U http://doi.acm.org/10.1145/1534903.1534905 %N 2 %0 Journal Article %J IEEE Computer Graphics and Applications %D 2009 %T Integrating Statistics and Visualization for Exploratory Power: From Long-Term Case Studies to Design Guidelines %A Perer,A. %A Shneiderman, Ben %K case studies %K Control systems %K Data analysis %K data mining %K data visualisation %K Data visualization %K data-mining %K design guidelines %K Employment %K exploration %K Filters %K Guidelines %K Information Visualization %K insights %K laboratory-based controlled experiments %K Performance analysis %K social network analysis %K Social network services %K social networking (online) %K social networks %K SocialAction %K statistical analysis %K Statistics %K visual analytics %K visual-analytics systems %K Visualization %X Evaluating visual-analytics systems is challenging because laboratory-based controlled experiments might not effectively represent analytical tasks. One such system, Social Action, integrates statistics and visualization in an interactive exploratory tool for social network analysis. This article describes results from long-term case studies with domain experts and extends established design goals for information visualization. %B IEEE Computer Graphics and Applications %V 29 %P 39 - 51 %8 2009/06//May %@ 0272-1716 %G eng %N 3 %R 10.1109/MCG.2009.44 %0 Book Section %B Multilingual FrameNets in Computational LexicographyMultilingual FrameNets in Computational Lexicography %D 2009 %T Interlingual annotation of multilingual text corpora and FrameNet %A Farwell,David %A Dorr, Bonnie J %A Habash,Nizar %A Helmreich,Stephen %A Hovy,Eduard %A Green,Rebecca %A Levin,Lori %A Miller,Keith %A Mitamura,Teruko %A Rambow,Owen %A Reeder,Flo %A Siddharthan,Advaith %E Bisang,Walter %E Hock,Hans Henrich %E Winter,Werner %E Boas,Hans C. %B Multilingual FrameNets in Computational LexicographyMultilingual FrameNets in Computational Lexicography %I Mouton de Gruyter %C Berlin, New York %V 200 %P 287 - 318 %8 2009/07/14/ %@ 978-3-11-021296-9, 978-3-11-021297-6 %G eng %U http://www.degruyter.com/view/books/9783110212976/9783110212976.4.287/9783110212976.4.287.xml %0 Journal Article %J Information Forensics and Security, IEEE Transactions on %D 2009 %T Intrinsic Sensor Noise Features for Forensic Analysis on Scanners and Scanned Images %A Gou,Hongmei %A Swaminathan,A. %A M. Wu %K analysis;image %K camera-taken %K cameras;digital %K denoising;intrinsic %K denoising;wavelet %K features;natural %K forensics;scanned %K graphics;digital %K identifier;wavelet %K images;digital %K images;scanner %K NOISE %K photographs;computer-generated %K prediction;nonintrusive %K reproduction;forensic %K scanner %K scenes;neighborhood %K sensor %K transforms; %X A large portion of digital images available today are acquired using digital cameras or scanners. While cameras provide digital reproduction of natural scenes, scanners are often used to capture hard-copy art in a more controlled environment. In this paper, new techniques for nonintrusive scanner forensics that utilize intrinsic sensor noise features are proposed to verify the source and integrity of digital scanned images. Scanning noise is analyzed from several aspects using only scanned image samples, including through image denoising, wavelet analysis, and neighborhood prediction, and then obtain statistical features from each characterization. Based on the proposed statistical features of scanning noise, a robust scanner identifier is constructed to determine the model/brand of the scanner used to capture a scanned image. Utilizing these noise features, we extend the scope of acquisition forensics to differentiating scanned images from camera-taken photographs and computer-generated graphics. The proposed noise features also enable tampering forensics to detect postprocessing operations on scanned images. Experimental results are presented to demonstrate the effectiveness of employing the proposed noise features for performing various forensic analysis on scanners and scanned images. %B Information Forensics and Security, IEEE Transactions on %V 4 %P 476 - 491 %8 2009/09// %@ 1556-6013 %G eng %N 3 %R 10.1109/TIFS.2009.2026458 %0 Book Section %B Wiley Handbook of Science and Technology for Homeland SecurityWiley Handbook of Science and Technology for Homeland Security %D 2009 %T Knowledge Extraction from Surveillance Sensors %A Chellapa, Rama %A Veeraraghavan,Ashok %A Sankaranarayanan,Aswin C. %K multi‐camera tracking %K multi‐modal fusion %K recognition %K sensor networks %K Surveillance %B Wiley Handbook of Science and Technology for Homeland SecurityWiley Handbook of Science and Technology for Homeland Security %I John Wiley & Sons, Inc. %8 2009/// %@ 9780470087923 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/9780470087923.hhs510/abstract;jsessionid=D0D752EF5D04327BE003BDFBD2F96134.d01t03 %0 Conference Paper %B Computer Graphics and Image Processing (SIBGRAPI), 2009 XXII Brazilian Symposium on %D 2009 %T Learning Discriminative Appearance-Based Models Using Partial Least Squares %A Schwartz, W.R. %A Davis, Larry S. %K (artificial %K analysis;learning %K appearance %K approximations;object %K based %K colour %K descriptors;learning %K discriminative %K intelligence);least %K learning %K least %K models;machine %K person %K recognition; %K recognition;feature %K squares %K squares;image %K techniques;partial %X Appearance information is essential for applications such as tracking and people recognition. One of the main problems of using appearance-based discriminative models is the ambiguities among classes when the number of persons being considered increases. To reduce the amount of ambiguity, we propose the use of a rich set of feature descriptors based on color, textures and edges. Another issue regarding appearance modeling is the limited number of training samples available for each appearance. The discriminative models are created using a powerful statistical tool called partial least squares (PLS), responsible for weighting the features according to their discriminative power for each different appearance. The experimental results, based on appearance-based person recognition, demonstrate that the use of an enriched feature set analyzed by PLS reduces the ambiguity among different appearances and provides higher recognition rates when compared to other machine learning techniques. %B Computer Graphics and Image Processing (SIBGRAPI), 2009 XXII Brazilian Symposium on %P 322 - 329 %8 2009/10// %G eng %R 10.1109/SIBGRAPI.2009.42 %0 Journal Article %J Autonomous Agents and Multi-Agent Systems %D 2009 %T Learning to trust in the competence and commitment of agents %A Smith,Michael %A desJardins, Marie %X For agents to collaborate in open multi-agent systems, each agent must trust in the other agents’ ability to complete tasks and willingness to cooperate. Agents need to decide between cooperative and opportunistic behavior based on their assessment of another agents’ trustworthiness. In particular, an agent can have two beliefs about a potential partner that tend to indicate trustworthiness: that the partner is competent and that the partner expects to engage in future interactions . This paper explores an approach that models competence as an agent’s probability of successfully performing an action, and models belief in future interactions as a discount factor. We evaluate the underlying decision framework’s performance given accurate knowledge of the model’s parameters in an evolutionary game setting. We then introduce a game-theoretic framework in which an agent can learn a model of another agent online, using the Harsanyi transformation. The learning agents evaluate a set of competing hypotheses about another agent during the simulated play of an indefinitely repeated game. The Harsanyi strategy is shown to demonstrate robust and successful online play against a variety of static, classic, and learning strategies in a variable-payoff Iterated Prisoner’s Dilemma setting. %B Autonomous Agents and Multi-Agent Systems %V 18 %P 36 - 82 %8 2009/// %@ 1387-2532 %G eng %U http://dx.doi.org/10.1007/s10458-008-9055-8 %N 1 %0 Journal Article %J 2009 SIAM International Conference on Data Mining (SDM09) %D 2009 %T On maximum coverage in the streaming model & application to multi-topic blog-watch %A Saha,B. %A Getoor, Lise %X We generalize the graph streaming model to hypergraphs.In this streaming model, hyperedges are arriving online and any computation has to be done on-the-fly using a small amount of space. Each hyperedge can be viewed as a set of elements (nodes), so we refer to our proposed model as the “set-streaming” model of computation. We consider the problem of “maximum coverage”, in which k sets have to be selected that maximize the total weight of the covered elements. In the set-streaming model of computation, we show that our algorithm for maximum- coverage achieves an approximation factor of 1 . When multiple passes are allowed, we also provide a Θ(log n) approximation algorithm for the set-cover. We next consider a multi-topic blog-watch application, an extension of blog- alert like applications for handling simultaneous multiple- topic requests. We show how the problems of maximum- coverage and set-cover in the set-streaming model can be utilized to give efficient online solutions to this problem. We verify the effectiveness of our methods both on synthetic and real weblog data. %B 2009 SIAM International Conference on Data Mining (SDM09) %8 2009/// %G eng %0 Journal Article %J BMC Genomics %D 2009 %T Measuring differential gene expression by short read sequencing: quantitative comparison to 2-channel gene expression microarrays %A Bloom, Joshua S. %A Zia Khan %A Kruglyak, Leonid %A Singh, Mona %A Caudy, Amy A. %K Animal Genetics and Genomics %K Life Sciences, general %K Microarrays %K Microbial Genetics and Genomics %K Plant Genetics & Genomics %K proteomics %X Background High-throughput cDNA synthesis and sequencing of poly(A)-enriched RNA is rapidly emerging as a technology competing to replace microarrays as a quantitative platform for measuring gene expression. Results Consequently, we compared full length cDNA sequencing to 2-channel gene expression microarrays in the context of measuring differential gene expression. Because of its comparable cost to a gene expression microarray, our study focused on the data obtainable from a single lane of an Illumina 1 G sequencer. We compared sequencing data to a highly replicated microarray experiment profiling two divergent strains of S. cerevisiae. Conclusion Using a large number of quantitative PCR (qPCR) assays, more than previous studies, we found that neither technology is decisively better at measuring differential gene expression. Further, we report sequencing results from a diploid hybrid of two strains of S. cerevisiae that indicate full length cDNA sequencing can discover heterozygosity and measure quantitative allele-specific expression simultaneously. %B BMC Genomics %V 10 %P 1 - 10 %8 2009/12/01/ %@ 1471-2164 %G eng %U http://link.springer.com/article/10.1186/1471-2164-10-221 %N 1 %! BMC Genomics %0 Patent %D 2009 %T Method and system for optimal data diagnosis %A V.S. Subrahmanian %A Ernst,Jason %E University of Maryland %X In the method and system for automated data diagnosis, a relational data base is formed of attributes characterizing an entity. The framework computes optimal diagnosis for a diagnostic inference problem which separates the desirable attributes from the undesirable attributes. The user of the system is permitted to specify certain parameters based on which the system computes a set of optimal or near optimal association rules between the attributes of a specific process, product, or other entity. The system and method of the present invention considers the simplicity of conditions in addition to support and confidence when ordering them, contributes the notion of tight conditions and semi-equivalence to remove redundant rules, uses the concept of the top fringes which allows near optimal conditions to be found, in addition to optimal conditions. Numeric as well as non-numeric attributes can be analyzed. %V 10/774,516 %8 2009/01/06/ %G eng %U http://www.google.com/patents?id=fHnVAAAAEBAJ %N 7474987 %0 Conference Paper %B Proceedings of the 9th international conference on Algorithms in bioinformatics %D 2009 %T Mimosa: mixture model of co-expression to detect modulators of regulatory interaction %A Hansen,Matthew %A Everett,Logan %A Singh,Larry %A Hannenhalli, Sridhar %X Functionally related genes tend to be correlated in their expression patterns across multiple conditions and/or tissue-types. Thus co-expression networks are often used to investigate functional groups of genes. In particular, when one of the genes is a transcription factor (TF), the co-expression-based interaction is interpreted, with caution, as a direct regulatory interaction. However, any particular TF, and more importantly, any particular regulatory interaction, is likely to be active only in a subset of experimental conditions. Moreover, the subset of expression samples where the regulatory interaction holds may be marked by presence or absence of a modifier gene, such as an enzyme that post-translationally modifies the TF. Such subtlety of regulatory interactions is overlooked when one computes an overall expression correlation. Here we present a novel mixture modeling approach where a TF-Gene pair is presumed to be significantly correlated (with unknown coefficient) in a (unknown) subset of expression samples. The parameters of the model are estimated using a Maximum Likelihood approach. The estimated mixture of expression samples is then mined to identify genes potentially modulating the TF-Gene interaction. We have validated our approach using synthetic data and on three biological cases in cow and in yeast. While limited in some ways, as discussed, the work represents a novel approach to mine expression data and detect potential modulators of regulatory interactions. %B Proceedings of the 9th international conference on Algorithms in bioinformatics %S WABI'09 %I Springer-Verlag %C Berlin, Heidelberg %P 133 - 144 %8 2009/// %@ 3-642-04240-6, 978-3-642-04240-9 %G eng %U http://dl.acm.org/citation.cfm?id=1812906.1812918 %0 Journal Article %J Computer %D 2009 %T Mining Online Opinions %A V.S. Subrahmanian %X Emerging opinion-mining technology makes it possible to aggregate and financially leverage the opinions of vast numbers of people. %B Computer %V 42 %P 88 - 90 %8 2009/// %G eng %N 7 %R 10.1109/MC.2009.229 %0 Journal Article %J EURASIP Journal on Image and Video Processing %D 2009 %T Modeling and visualization of human activities for multicamera networks %A Sankaranarayanan,A. C %A Patro,R. %A Turaga,P. %A Varshney, Amitabh %A Chellapa, Rama %B EURASIP Journal on Image and Video Processing %V 2009 %8 2009/// %G eng %0 Journal Article %J Image Processing, IEEE Transactions on %D 2009 %T Multicamera Tracking of Articulated Human Motion Using Shape and Motion Cues %A Sundaresan, A. %A Chellapa, Rama %K 2D shape cues %K 3D shape cues %K algorithms %K Anatomic;Models %K articulated human motion %K automatic algorithm %K Biological;Movement;Posture;Skeleton;Video Recording; %K Computer-Assisted;Models %K Eigenvalues and eigenfunctions %K human pose estimation %K HUMANS %K Image motion analysis %K IMAGE PROCESSING %K image registration %K Image segmentation %K Image sequences %K kinematic singularity %K Laplacian eigenmaps %K multicamera tracking algorithm %K pixel displacement %K pose estimation %K single-frame registration technique %K temporal registration method %K tracking %X We present a completely automatic algorithm for initializing and tracking the articulated motion of humans using image sequences obtained from multiple cameras. A detailed articulated human body model composed of sixteen rigid segments that allows both translation and rotation at joints is used. Voxel data of the subject obtained from the images is segmented into the different articulated chains using Laplacian eigenmaps. The segmented chains are registered in a subset of the frames using a single-frame registration technique and subsequently used to initialize the pose in the sequence. A temporal registration method is proposed to identify the partially segmented or unregistered articulated chains in the remaining frames in the sequence. The proposed tracker uses motion cues such as pixel displacement as well as 2-D and 3-D shape cues such as silhouettes, motion residue, and skeleton curves. The tracking algorithm consists of a predictor that uses motion cues and a corrector that uses shape cues. The use of complementary cues in the tracking alleviates the twin problems of drift and convergence to local minima. The use of multiple cameras also allows us to deal with the problems due to self-occlusion and kinematic singularity. We present tracking results on sequences with different kinds of motion to illustrate the effectiveness of our approach. The pose of the subject is correctly tracked for the duration of the sequence as can be verified by inspection. %B Image Processing, IEEE Transactions on %V 18 %P 2114 - 2126 %8 2009/09// %@ 1057-7149 %G eng %N 9 %R 10.1109/TIP.2009.2022290 %0 Journal Article %J Statistical Bioinformatics %D 2009 %T Multidimensional Analysis and Visualization on Large Biomedical Data %A Seo,J. %A Shneiderman, Ben %B Statistical Bioinformatics %P 157 - 184 %8 2009/// %G eng %0 Journal Article %J ScienceScience %D 2009 %T A National Initiative for Social Participation %A Shneiderman, Ben %B ScienceScience %V 323 %P 1426 - 1427 %8 2009/03/13/ %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/323/5920/1426 %N 5920 %R 10.1126/science.323.5920.1426 %0 Conference Paper %B Proceedings of the 18th international conference on World wide web %D 2009 %T Network-aware forward caching %A Erman,Jeffrey %A Gerber,Alexandre %A Hajiaghayi, Mohammad T. %A Pei,Dan %A Spatscheck,Oliver %K caching %K Web %X This paper proposes and evaluates a Network Aware Forward Caching approach for determining the optimal deployment strategy of forward caches to a network. A key advantage of this approach is that we can reduce the network costs associated with forward caching to maximize the benefit obtained from their deployment. We show in our simulation that a 37% increase to net benefits could be achieved over the standard method of full cache deployment to cache all POPs traffic. In addition, we show that this maximal point occurs when only 68% of the total traffic is cached. Another contribution of this paper is the analysis we use to motivate and evaluate this problem. We characterize the Internet traffic of 100K subscribers of a US residential broadband provider. We use both layer 4 and layer 7 analysis to investigate the traffic volumes of the flows as well as study the general characteristics of the applications used. We show that HTTP is a dominant protocol and account for 68% of the total downstream traffic and that 34% of that traffic is multimedia. In addition, we show that multimedia content using HTTP exhibits a 83% annualized growth rate and other HTTP traffic has a 53% growth rate versus the 26% over all annual growth rate of broadband traffic. This shows that HTTP traffic will become ever more dominent and increase the potential caching opportunities. Furthermore, we characterize the core backbone traffic of this broadband provider to measure the distance travelled by content and traffic. We find that CDN traffic is much more efficient than P2P content and that there is large skew in the Air Miles between POP in a typical network. Our findings show that there are many opportunties in broadband provider networks to optimize how traffic is delivered and cached. %B Proceedings of the 18th international conference on World wide web %S WWW '09 %I ACM %C New York, NY, USA %P 291 - 300 %8 2009/// %@ 978-1-60558-487-4 %G eng %U http://doi.acm.org/10.1145/1526709.1526749 %R 10.1145/1526709.1526749 %0 Journal Article %J Bangladesh Journal of Plant Taxonomy %D 2009 %T New records of phytoplankton for Bangladesh. 9. Some rare and a new species %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasmin,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X Ten taxa belonging to Chlorophyceae, Cyanophyceae, Bacillariophyceae and Euglenophyceae, and one with an uncertain taxonomic position have been described in this paper. Of these, 10 taxa have been found to be globally rare and new records for Bangladesh, whereas Strombomonas islamii Khondker sp. nov. has been described as new to science. %B Bangladesh Journal of Plant Taxonomy %V 16 %8 2009/06/25/ %@ 1028-2092 %G eng %U http://www.banglajol.info/bd/index.php/BJPT/article/viewArticle/2734 %N 1 %R 10.3329/bjpt.v16i1.2734 %0 Journal Article %J Magnetics, IEEE Transactions on %D 2009 %T Nonlinear Resonant and Chaotic Dynamics in Microwave Assisted Magnetization Switching %A d'Aquino,M. %A Serpico,C. %A Bertotti,G. %A Mayergoyz, Issak D %A Bonin,R. %K amplitude;chaotic %K components;magnetic %K dynamical %K dynamics;coercive %K dynamics;numerical %K field %K field;coercivity;linearly-polarized %K force;magnetic %K magnetization %K nanoparticle;microwave-assisted %K particles;magnetic %K radiofrequency %K resonant %K RF %K simulations;perturbation %K switching;nanoparticles;nonlinear %K switching;nonlinear %K systems;perturbation %K technique;chaos;coercive %K theory; %X The switching process of a uniformly magnetized magnetic nanoparticle is considered. The particle is subject to applied fields having both dc and linearly-polarized radio-frequency (RF) components. The possibility of using the RF power to obtain a reduced coercivity of the particle is related to the onset of chaotic magnetization dynamics for moderately low values of the RF field amplitude. Perturbation technique for the evaluation of the reduced coercive field is developed and applied to the microwave assisted switching of the particle. Numerical simulations confirm the predictions of the theory. %B Magnetics, IEEE Transactions on %V 45 %P 3950 - 3953 %8 2009/10// %@ 0018-9464 %G eng %N 10 %R 10.1109/TMAG.2009.2023242 %0 Journal Article %J IEEETransactions on Pattern Analysis and Machine Intelligence %D 2009 %T Off-Line Loop Investigation for Handwriting Analysis %A Steinherz,T. %A David Doermann %A Rivlin,E. %A Intrator,N. %B IEEETransactions on Pattern Analysis and Machine Intelligence %V 31 %P 193 - 209 %8 2009/02// %G eng %N 2 %0 Journal Article %J Symposium on Click Modular Router %D 2009 %T An OpenFlow switch element for Click %A Mundada,Y. %A Sherwood,R. %A Feamster, Nick %B Symposium on Click Modular Router %8 2009/// %G eng %0 Conference Paper %B Proceedings of the 2009 Workshop on Graph-based Methods for Natural Language Processing %D 2009 %T Opinion graphs for polarity and discourse classification %A Somasundaran,Swapna %A Namata,Galileo %A Getoor, Lise %A Wiebe,Janyce %X This work shows how to construct discourse-level opinion graphs to perform a joint interpretation of opinions and discourse relations. Specifically, our opinion graphs enable us to factor in discourse information for polarity classification, and polarity information for discourse-link classification. This inter-dependent framework can be used to augment and improve the performance of local polarity and discourse-link classifiers. %B Proceedings of the 2009 Workshop on Graph-based Methods for Natural Language Processing %S TextGraphs-4 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 66 - 74 %8 2009/// %@ 978-1-932432-54-1 %G eng %U http://dl.acm.org/citation.cfm?id=1708124.1708138 %0 Journal Article %J The VLDB Journal %D 2009 %T P r DB: managing and exploiting rich correlations in probabilistic databases %A Sen,P. %A Deshpande, Amol %A Getoor, Lise %X Due to numerous applications producing noisy data, e.g., sensor data, experimental data, data from uncurated sources, information extraction, etc., there has been a surge of interest in the development of probabilistic databases. Most probabilistic database models proposed to date, however, fail to meet the challenges of real-world applications on two counts: (1) they often restrict the kinds of uncertainty that the user can represent; and (2) the query processing algorithms often cannot scale up to the needs of the application. In this work, we define a probabilistic database model, PrDB, that uses graphical models, a state-of-the-art probabilistic modeling technique developed within the statistics and machine learning community, to model uncertain data. We show how this results in a rich, complex yet compact probabilistic database model, which can capture the commonly occurring uncertainty models (tuple uncertainty, attribute uncertainty), more complex models (correlated tuples and attributes) and allows compact representation (shared and schema-level correlations). In addition, we show how query evaluation in PrDB translates into inference in an appropriately augmented graphical model. This allows us to easily use any of a myriad of exact and approximate inference algorithms developed within the graphical modeling community. While probabilistic inference provides a generic approach to solving queries, we show how the use of shared correlations, together with a novel inference algorithm that we developed based on bisimulation, can speed query processing significantly. We present a comprehensive experimental evaluation of the proposed techniques and show that even with a few shared correlations, significant speedups are possible. %B The VLDB Journal %V 18 %P 1065 - 1090 %8 2009/// %G eng %N 5 %R 10.1007/s00778-009-0153-2 %0 Journal Article %J Proc. VLDB Endow. %D 2009 %T Path oracles for spatial networks %A Sankaranarayanan,Jagan %A Samet, Hanan %A Alborzi,Houman %X The advent of location-based services has led to an increased demand for performing operations on spatial networks in real time. The challenge lies in being able to cast operations on spatial networks in terms of relational operators so that they can be performed in the context of a database. A linear-sized construct termed a path oracle is introduced that compactly encodes the n2 shortest paths between every pair of vertices in a spatial network having n vertices thereby reducing each of the paths to a single tuple in a relational database and enables finding shortest paths by repeated application of a single SQL SELECT operator. The construction of the path oracle is based on the observed coherence between the spatial positions of both source and destination vertices and the shortest paths between them which facilitates the aggregation of source and destination vertices into groups that share common vertices or edges on the shortest paths between them. With the aid of the Well-Separated Pair (WSP) technique, which has been applied to spatial networks using the network distance measure, a path oracle is proposed that takes O(sdn) space, where s is empirically estimated to be around 12 for road networks, but that can retrieve an intermediate link in a shortest path in O(logn) time using a B-tree. An additional construct termed the path-distance oracle of size O(n · max(sd, 1/εd)) (empirically (n · max(122, 2.5/ε2))) is proposed that can retrieve an intermediate vertex as well as an ε-approximation of the network distances in O(logn) time using a B-tree. Experimental results indicate that the proposed oracles are linear in n which means that they are scalable and can enable complicated query processing scenarios on massive spatial network datasets. %B Proc. VLDB Endow. %V 2 %P 1210 - 1221 %8 2009/08// %@ 2150-8097 %G eng %U http://dl.acm.org/citation.cfm?id=1687627.1687763 %N 1 %0 Conference Paper %B Proceedings of the ACM SIGCOMM 2009 conference on Data communication %D 2009 %T Persona: an online social network with user-defined privacy %A Baden,Randy %A Bender,Adam %A Spring, Neil %A Bhattacharjee, Bobby %A Starin,Daniel %K ABE %K Facebook %K OSN %K persona %K privacy %K social networks %X Online social networks (OSNs) are immensely popular, with some claiming over 200 million users. Users share private content, such as personal information or photographs, using OSN applications. Users must trust the OSN service to protect personal information even as the OSN provider benefits from examining and sharing that information. We present Persona, an OSN where users dictate who may access their information. Persona hides user data with attribute-based encryption (ABE), allowing users to apply fine-grained policies over who may view their data. Persona provides an effective means of creating applications in which users, not the OSN, define policy over access to private data. We demonstrate new cryptographic mechanisms that enhance the general applicability of ABE. We show how Persona provides the functionality of existing online social networks with additional privacy benefits. We describe an implementation of Persona that replicates Facebook applications and show that Persona provides acceptable performance when browsing privacy-enhanced web pages, even on mobile devices. %B Proceedings of the ACM SIGCOMM 2009 conference on Data communication %S SIGCOMM '09 %I ACM %C New York, NY, USA %P 135 - 146 %8 2009/// %@ 978-1-60558-594-9 %G eng %U http://doi.acm.org/10.1145/1592568.1592585 %R 10.1145/1592568.1592585 %0 Journal Article %J Computer %D 2009 %T Persona : An Online Social Network with User-Defined Privacy Categories and Subject Descriptors %A Starin,Daniel %A Baden,Randy %A Bender,Adam %A Spring, Neil %A Bhattacharjee, Bobby %X Online social networks (OSNs) are immensely popular, with some claiming over 200 million users. Users share private content, such as personal information or photographs, using OSN applications. Users must trust the OSN service to protect personal information even as the OSN provider benefits from examining and sharing that information. We present Persona, an OSN where users dictate who may access their information. Persona hides user data with attribute-based encryption (ABE), allowing users to apply fine-grained policies over who may view their data. Persona provides an effective means of creating applications in which users, not the OSN, define policy over access to private data. We demonstrate new cryptographic mechanisms that enhance the general applicability of ABE. We show how Persona provides the functionality of existing online social networks with additional privacy benefits. We describe an implementation of Persona that replicates Facebook applications and show that Persona provides acceptable performance when browsing privacy-enhanced web pages, even on mobile devices. %B Computer %V 39 %P 135 - 146 %8 2009/// %G eng %N 4 %R 10.1145/1594977.1592585 %0 Book Section %B Programming Multi-Agent Systems %D 2009 %T Planning for Interactions among Autonomous Agents %A Au,Tsz-Chiu %A Kuter,Ugur %A Nau, Dana S. %E Hindriks,Koen %E Pokahr,Alexander %E Sardina,Sebastian %K Computer science %X AI planning research has traditionally focused on offline pl- anning for static single-agent environments. In environments where an agent needs to plan its interactions with other autonomous agents, planning is much more complicated, because the actions of the other agents can induce a combinatorial explosion in the number of contingencies that the planner will need to consider. This paper discusses several ways to alleviate the combinatorial explosion, and illustrates their use in several different kinds of multi-agent planning domains. %B Programming Multi-Agent Systems %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5442 %P 1 - 23 %8 2009/// %@ 978-3-642-03277-6 %G eng %U http://www.springerlink.com/content/j258015ux2p38383/abstract/ %0 Journal Article %J SIAM Journal on Optimization %D 2009 %T A Polynomial-Time Interior-Point Method for Conic Optimization, with Inexact Barrier Evaluations %A Schurr,Simon P. %A O'Leary, Dianne P. %A Tits,Andre' L. %B SIAM Journal on Optimization %V 20 %P 548 - 571 %8 2009/// %G eng %R DOI: 10.1137/080722825 %0 Journal Article %J Bioinformatics %D 2009 %T A practical algorithm for finding maximal exact matches in large sequence datasets using sparse suffix arrays %A Zia Khan %A Bloom, Joshua S. %A Kruglyak, Leonid %A Singh, Mona %X Motivation: High-throughput sequencing technologies place ever increasing demands on existing algorithms for sequence analysis. Algorithms for computing maximal exact matches (MEMs) between sequences appear in two contexts where high-throughput sequencing will vastly increase the volume of sequence data: (i) seeding alignments of high-throughput reads for genome assembly and (ii) designating anchor points for genome–genome comparisons.Results: We introduce a new algorithm for finding MEMs. The algorithm leverages a sparse suffix array (SA), a text index that stores every K-th position of the text. In contrast to a full text index that stores every position of the text, a sparse SA occupies much less memory. Even though we use a sparse index, the output of our algorithm is the same as a full text index algorithm as long as the space between the indexed suffixes is not greater than a minimum length of a MEM. By relying on partial matches and additional text scanning between indexed positions, the algorithm trades memory for extra computation. The reduced memory usage makes it possible to determine MEMs between significantly longer sequences. Availability: Source code for the algorithm is available under a BSD open source license at http://compbio.cs.princeton.edu/mems. The implementation can serve as a drop-in replacement for the MEMs algorithm in MUMmer 3. Contact: zkhan@cs.princeton.edu;mona@cs.princeton.edu Supplementary information: Supplementary data are available at Bioinformatics online. %B Bioinformatics %V 25 %P 1609 - 1616 %8 2009/07/01/ %@ 1367-4803, 1460-2059 %G eng %U http://bioinformatics.oxfordjournals.org/content/25/13/1609 %N 13 %! Bioinformatics %0 Journal Article %J The VLDB Journal %D 2009 %T PrDB: managing and exploiting rich correlations in probabilistic databases %A Sen,Prithviraj %A Deshpande, Amol %A Getoor, Lise %X Due to numerous applications producing noisy data, e.g., sensor data, experimental data, data from uncurated sources, information extraction, etc., there has been a surge of interest in the development of probabilistic databases. Most probabilistic database models proposed to date, however, fail to meet the challenges of real-world applications on two counts: (1) they often restrict the kinds of uncertainty that the user can represent; and (2) the query processing algorithms often cannot scale up to the needs of the application. In this work, we define a probabilistic database model, P r DB, that uses graphical models, a state-of-the-art probabilistic modeling technique developed within the statistics and machine learning community, to model uncertain data. We show how this results in a rich, complex yet compact probabilistic database model, which can capture the commonly occurring uncertainty models (tuple uncertainty, attribute uncertainty), more complex models (correlated tuples and attributes) and allows compact representation (shared and schema-level correlations). In addition, we show how query evaluation in P r DB translates into inference in an appropriately augmented graphical model. This allows us to easily use any of a myriad of exact and approximate inference algorithms developed within the graphical modeling community. While probabilistic inference provides a generic approach to solving queries, we show how the use of shared correlations, together with a novel inference algorithm that we developed based on bisimulation, can speed query processing significantly. We present a comprehensive experimental evaluation of the proposed techniques and show that even with a few shared correlations, significant speedups are possible. %B The VLDB Journal %V 18 %P 1065 - 1090 %8 2009/// %@ 1066-8888 %G eng %U http://dx.doi.org/10.1007/s00778-009-0153-2 %N 5 %0 Book Section %B Theory of Cryptography %D 2009 %T Predicate Privacy in Encryption Systems %A Shen, Emily %A Elaine Shi %A Waters,Brent %E Reingold, Omer %K Computer science %X Predicate encryption is a new encryption paradigm which gives a master secret key owner fine-grained control over access to encrypted data. The master secret key owner can generate secret key tokens corresponding to predicates. An encryption of data x can be evaluated using a secret token corresponding to a predicate f ; the user learns whether the data satisfies the predicate, i.e., whether f ( x ) = 1. Prior work on public-key predicate encryption has focused on the notion of data or plaintext privacy, the property that ciphertexts reveal no information about the encrypted data to an attacker other than what is inherently revealed by the tokens the attacker possesses. In this paper, we consider a new notion called predicate privacy , the property that tokens reveal no information about the encoded query predicate. Predicate privacy is inherently impossible to achieve in the public-key setting and has therefore received little attention in prior work. In this work, we consider predicate encryption in the symmetric-key setting and present a symmetric-key predicate encryption scheme which supports inner product queries. We prove that our scheme achieves both plaintext privacy and predicate privacy. %B Theory of Cryptography %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5444 %P 457 - 473 %8 2009 %@ 978-3-642-00456-8 %G eng %U http://www.springerlink.com/content/1717x5445k4718rp/abstract/ %0 Conference Paper %B Software Maintenance, 2009. ICSM 2009. IEEE International Conference on %D 2009 %T Prioritizing component compatibility tests via user preferences %A Yoon,Il-Chul %A Sussman, Alan %A Memon, Atif M. %A Porter, Adam %K compatibility testing prioritization %K component configurations %K computer clusters %K Middleware %K Middleware systems %K object-oriented programming %K program testing %K software engineering %K Software systems %K third-party components %K user preferences %X Many software systems rely on third-party components during their build process. Because the components are constantly evolving, quality assurance demands that developers perform compatibility testing to ensure that their software systems build correctly over all deployable combinations of component versions, also called configurations. However, large software systems can have many configurations, and compatibility testing is often time and resource constrained. We present a prioritization mechanism that enhances compatibility testing by examining the ldquomost importantrdquo configurations first, while distributing the work over a cluster of computers. We evaluate our new approach on two large scientific middleware systems and examine tradeoffs between the new prioritization approach and a previously developed lowest-cost-configuration-first approach. %B Software Maintenance, 2009. ICSM 2009. IEEE International Conference on %P 29 - 38 %8 2009/09// %G eng %R 10.1109/ICSM.2009.5306357 %0 Conference Paper %B Proceedings of the seventh ACM conference on Creativity and cognition %D 2009 %T Promoting social creativity: a component of a national initiative for social participation %A Shneiderman, Ben %A Churchill,Elizabeth %A Fischer,Gerhard %A Goldberg,Ken %K research agenda %K social creativity %K social participation %X This panel will discuss group processes that promote social creativity in science, engineering, arts, and humanities. We will offer positive and negative examples of social creativity projects, while suggesting research directions for dramatically increased social participation. The goal is to develop strategies that would expand resources and opportunities for research and education in social creativity. This requires our community to develop a unified position, then reach out to national science funding agencies, while building the case for the importance of this topic beyond our own community. How can social creativity, collaborative discovery, distributed innovation, and collective intelligence be framed as an international priority to cope with the problems of the 21st century and how can we identify a clear set of research challenges? The theme of technology-mediated social participation is outlined in the white paper for a National Initiative for Social Participation (http://iparticipate.wikispaces.com). The white paper suggests that successful research challenges should have three key elements: (1) compelling national need (healthcare, national security, community safety, education, innovation, cultural heritage, energy sustainability, environmental protection, etc.), (2) scientific foundation based on established theories and well-defined research questions (privacy, reciprocity, trust, motivation, recognition, etc.), and (3) computer science research challenges (security, privacy protection, scalability, visualization, end-user development, distributed data handling for massive user-generated content, network analysis of community evolution, cross network comparison, etc.). We seek recommendations for ways to increase the resources and attention for this field. We hope to inspire: universities to change course content, add courses, and offer new degree programs industry to help researchers on social creativity government to support these ideas and try them out in government applications scientists and artists to open themselves to more social/collaborative approaches %B Proceedings of the seventh ACM conference on Creativity and cognition %S C&C '09 %I ACM %C New York, NY, USA %P 7 - 8 %8 2009/// %@ 978-1-60558-865-0 %G eng %U http://doi.acm.org/10.1145/1640233.1640237 %R 10.1145/1640233.1640237 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2009 %T Protein quantification across hundreds of experimental conditions %A Zia Khan %A Bloom, Joshua S. %A Garcia, Benjamin A. %A Singh, Mona %A Kruglyak, Leonid %K kd-tree %K orthogonal range query %K quantitative proteomics %K space partitioning data structures %K tandem mass spectrometry %X Quantitative studies of protein abundance rarely span more than a small number of experimental conditions and replicates. In contrast, quantitative studies of transcript abundance often span hundreds of experimental conditions and replicates. This situation exists, in part, because extracting quantitative data from large proteomics datasets is significantly more difficult than reading quantitative data from a gene expression microarray. To address this problem, we introduce two algorithmic advances in the processing of quantitative proteomics data. First, we use space-partitioning data structures to handle the large size of these datasets. Second, we introduce techniques that combine graph-theoretic algorithms with space-partitioning data structures to collect relative protein abundance data across hundreds of experimental conditions and replicates. We validate these algorithmic techniques by analyzing several datasets and computing both internal and external measures of quantification accuracy. We demonstrate the scalability of these techniques by applying them to a large dataset that comprises a total of 472 experimental conditions and replicates. %B Proceedings of the National Academy of Sciences %V 106 %P 15544 - 15548 %8 2009/09/15/ %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/106/37/15544 %N 37 %! PNAS %0 Report %D 2009 %T Querying timestamped event sequences by exact search or similarity-based search: design and empirical evaluation %A Wongsuphasawat,K. %A Plaisant, Catherine %A Shneiderman, Ben %I Human-Computer Interaction Lab, University of Maryland %V HCIL-2009-20 %8 2009/// %G eng %0 Conference Paper %B Proceedings of the 10th ACM conference on Electronic commerce %D 2009 %T On random sampling auctions for digital goods %A Alaei,Saeed %A Malekian,Azarakhsh %A Srinivasan, Aravind %K auction %K mechanism design %K random sampling %X In the context of auctions for digital goods, an interesting Random Sampling Optimal Price auction (RSOP) has been proposed by Goldberg, Hartline and Wright; this leads to a truthful mechanism. Since random sampling is a popular approach for auctions that aims to maximize the seller's revenue, this method has been analyzed further by Feige, Flaxman, Hartline and Kleinberg, who have shown that it is 15-competitive in the worst case -- which is substantially better than the previously proved bounds but still far from the conjectured competitive ratio of 4. In this paper, we prove that RSOP is indeed 4-competitive for a large class of instances in which the number λ of bidders receiving the item at the optimal uniform price, is at least 6. We also show that it is 4.68 competitive for the small class of remaining instances thus leaving a negligible gap between the lower and upper bound. Furthermore, we develop a robust version of RSOP -- one in which the seller's revenue is, with high probability, not much below its mean -- when the above parameter λ grows large. We employ a mix of probabilistic techniques and dynamic programming to compute these bounds. %B Proceedings of the 10th ACM conference on Electronic commerce %S EC '09 %I ACM %C New York, NY, USA %P 187 - 196 %8 2009/// %@ 978-1-60558-458-4 %G eng %U http://doi.acm.org/10.1145/1566374.1566402 %R 10.1145/1566374.1566402 %0 Journal Article %J Image Processing, IEEE Transactions on %D 2009 %T Rate-Invariant Recognition of Humans and Their Activities %A Veeraraghavan,A. %A Srivastava, A. %A Roy-Chowdhury, A.K. %A Chellapa, Rama %K Automated;Video Recording; %K Bayesian algorithm;Monte Carlo sampling;action recognition;activity-specific time warping transformations;computationally efficient inference algorithms;distance measures;execution rate function;function space;gait-based person identification;geodesics;no %K Statistical;Monte Carlo Method;Movement;Pattern Recognition %X Pattern recognition in video is a challenging task because of the multitude of spatio-temporal variations that occur in different videos capturing the exact same event. While traditional pattern-theoretic approaches account for the spatial changes that occur due to lighting and pose, very little has been done to address the effect of temporal rate changes in the executions of an event. In this paper, we provide a systematic model-based approach to learn the nature of such temporal variations (time warps) while simultaneously allowing for the spatial variations in the descriptors. We illustrate our approach for the problem of action recognition and provide experimental justification for the importance of accounting for rate variations in action recognition. The model is composed of a nominal activity trajectory and a function space capturing the probability distribution of activity-specific time warping transformations. We use the square-root parameterization of time warps to derive geodesics, distance measures, and probability distributions on the space of time warping functions. We then design a Bayesian algorithm which treats the execution rate function as a nuisance variable and integrates it out using Monte Carlo sampling, to generate estimates of class posteriors. This approach allows us to learn the space of time warps for each activity while simultaneously capturing other intra- and interclass variations. Next, we discuss a special case of this approach which assumes a uniform distribution on the space of time warping functions and show how computationally efficient inference algorithms may be derived for this special case. We discuss the relative advantages and disadvantages of both approaches and show their efficacy using experiments on gait-based person identification and activity recognition. %B Image Processing, IEEE Transactions on %V 18 %P 1326 - 1339 %8 2009/06// %@ 1057-7149 %G eng %N 6 %R 10.1109/TIP.2009.2017143 %0 Journal Article %J AIS Transactions on Human-Computer Interaction %D 2009 %T The reader-to-leader framework: Motivating technology-mediated social participation %A Preece,J. %A Shneiderman, Ben %X Billions of people participate in online social activities. Most users participate as readers of discussion boards, searchers of blogposts, or viewers of photos. A fraction of users become contributors of user-generated content by writing consumer product reviews, uploading travel photos, or expressing political opinions. Some users move beyond such individual efforts to become collaborators, forming tightly connected groups with lively discussions whose outcome might be a Wikipedia article or a carefully edited YouTube video. A small fraction of users becomes leaders, who participate in governance by setting and upholding policies, repairing vandalized materials, or mentoring novices. We analyze these activities and offer the Reader-to-Leader Framework with the goal of helping researchers, designers, and managers understand what motivates technology-mediated social participation. This will enable them to improve interface design and social support for their companies, government agencies, and non-governmental organizations. These improvements could reduce the number of failed projects, while accelerating the application of social media for national priorities such as healthcare, energy sustainability, emergency response, economic development, education, and more. %B AIS Transactions on Human-Computer Interaction %V 1 %P 13 - 32 %8 2009/// %G eng %N 1 %0 Journal Article %J Journal of cryptology %D 2009 %T Reducing complexity assumptions for statistically-hiding commitment %A Haitner,I. %A Horvitz,O. %A Katz, Jonathan %A Koo,C. Y %A Morselli,R. %A Shaltiel,R. %X We revisit the following question: what are the minimal assumptions needed to construct statistically-hiding commitment schemes? Naor et al. show how to construct such schemes based on any one-way permutation. We improve upon this by showing a construction based on any approximable preimage-size one-way function. These are one-way functions for which it is possible to efficiently approximate the number of pre-images of a given output. A special case is the class of regular one-way functions where all points in the image of the function have the same (known) number of pre-images.We also prove two additional results related to statistically-hiding commitment. First, we prove a (folklore) parallel composition theorem showing, roughly speaking, that the statistical hiding property of any such commitment scheme is amplified exponentially when multiple independent parallel executions of the scheme are carried out. Second, we show a compiler which transforms any commitment scheme which is statistically hiding against an honest-but-curious receiver into one which is statistically hiding even against a malicious receiver. %B Journal of cryptology %V 22 %P 283 - 310 %8 2009/// %G eng %N 3 %R 10.1007/s00145-007-9012-8 %0 Journal Article %J Journal of Computational Biology %D 2009 %T Revealing biological modules via graph summarization %A Navlakha,S. %A Schatz,M. C %A Kingsford, Carl %B Journal of Computational Biology %V 16 %P 253 - 264 %8 2009/// %G eng %N 2 %0 Journal Article %J Proceedings of the 17th Italian Symposium on Advanced Database Systems (SEBD’09) %D 2009 %T Review of Spatial Databases and Geographic Information Systems %A Samet, Hanan %B Proceedings of the 17th Italian Symposium on Advanced Database Systems (SEBD’09) %V 23 %P 26 - 26 %8 2009/06// %G eng %0 Conference Paper %B IEEE/WIC/ACM International Joint Conferences on Web Intelligence and Intelligent Agent Technologies, 2009. WI-IAT '09 %D 2009 %T Rigorous Probabilistic Trust-Inference with Applications to Clustering %A DuBois,Thomas %A Golbeck,Jennifer %A Srinivasan, Aravind %K Clustering algorithms %K Conferences %K Educational institutions %K Extraterrestrial measurements %K Inference algorithms %K Intelligent agent %K random graphs %K Social network services %K trust inferrence %K Visualization %K Voting %K Web sites %X The World Wide Web has transformed into an environment where users both produce and consume information. In order to judge the validity of information, it is important to know how trustworthy its creator is. Since no individual can have direct knowledge of more than a small fraction of information authors, methods for inferring trust are needed. We propose a new trust inference scheme based on the idea that a trust network can be viewed as a random graph, and a chain of trust as a path in that graph. In addition to having an intuitive interpretation, our algorithm has several advantages, noteworthy among which is the creation of an inferred trust-metric space where the shorter the distance between two people, the higher their trust. Metric spaces have rigorous algorithms for clustering, visualization, and related problems, any of which is directly applicable to our results. %B IEEE/WIC/ACM International Joint Conferences on Web Intelligence and Intelligent Agent Technologies, 2009. WI-IAT '09 %I IEEE %V 1 %P 655 - 658 %8 2009/09/15/18 %@ 978-0-7695-3801-3 %G eng %R 10.1109/WI-IAT.2009.109 %0 Journal Article %J Proceedings of the Third International Conference on Advances in Biometrics %D 2009 %T Robust human detection under occlusion by integrating face and person detectors %A Schwartz, W.R. %A Gopalan,R. %A Chellapa, Rama %A Davis, Larry S. %X Human detection under occlusion is a challenging problemin computer vision. We address this problem through a framework which integrates face detection and person detection. We first investigate how the response of a face detector is correlated with the response of a person detector. From these observations, we formulate hypotheses that capture the intuitive feedback between the responses of face and person detectors and use it to verify if the individual detectors’ outputs are true or false. We illustrate the performance of our integration framework on challeng- ing images that have considerable amount of occlusion, and demonstrate its advantages over individual face and person detectors. %B Proceedings of the Third International Conference on Advances in Biometrics %P 970 - 979 %8 2009/// %G eng %0 Journal Article %J GPU Technology Conference %D 2009 %T Scaling kernel machine learning algorithm via the use of GPUs %A Srinivasan,B.V. %A Duraiswami, Ramani %B GPU Technology Conference %8 2009/// %G eng %0 Journal Article %J Proceedings of the International Conference on Computer Vision %D 2009 %T Scene it or not? incremental multiple kernel learning for object detection %A Kembhavi,A. %A Siddiquie,B. %A Miezianko,R. %A McCloskey,S. %A Davis, Larry S. %A Schwartz, W.R. %A Harwood,D. %A Gupta,A. %A Farrell,R. %A Luo,Y. %A others %B Proceedings of the International Conference on Computer Vision %8 2009/// %G eng %0 Journal Article %J Algorithmica %D 2009 %T Scheduling on Unrelated Machines under Tree-Like Precedence Constraints %A Anil Kumar,V. %A Marathe,Madhav %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %X We present polylogarithmic approximations for the R | prec | C max and R | prec |∑ j w j C j problems, when the precedence constraints are “treelike”—i.e., when the undirected graph underlying the precedences is a forest. These are the first non-trivial generalizations of the job shop scheduling problem to scheduling with precedence constraints that are not just chains. These are also the first non-trivial results for the weighted completion time objective on unrelated machines with precedence constraints of any kind . We obtain improved bounds for the weighted completion time and flow time for the case of chains with restricted assignment—this generalizes the job shop problem to these objective functions. We use the same lower bound of “congestion + dilation”, as in other job shop scheduling approaches (e.g. Shmoys, Stein and Wein, SIAM J. Comput. 23, 617–632, 1994 ). The first step in our algorithm for the R | prec | C max problem with treelike precedences involves using the algorithm of Lenstra, Shmoys and Tardos to obtain a processor assignment with the congestion + dilation value within a constant factor of the optimal. We then show how to generalize the random-delays technique of Leighton, Maggs and Rao to the case of trees. For the special case of chains, we show a dependent rounding technique which leads to a bicriteria approximation algorithm for minimizing the flow time, a notoriously hard objective function. %B Algorithmica %V 55 %P 205 - 226 %8 2009/// %@ 0178-4617 %G eng %U http://dx.doi.org/10.1007/s00453-007-9004-y %N 1 %0 Conference Paper %B Proceedings of the 2009 3rd International Symposium on Empirical Software Engineering and Measurement %D 2009 %T Scope error detection and handling concerning software estimation models %A Sarcia,Salvatore Alessandro %A Basili, Victor R. %A Cantone,Giovanni %X Over the last 25+ years, the software community has been searching for the best models for estimating variables of interest (e.g., cost, defects, and fault proneness). However, little research has been done to improve the reliability of the estimates. Over the last decades, scope error and error analysis have been substantially ignored by the community. This work attempts to fill this gap in the research and enhance a common understanding within the community. Results provided in this study can eventually be used to support human judgment-based techniques and be an addition to the portfolio. The novelty of this work is that, we provide a way of detecting and handling the scope error arising from estimation models. The answer whether or not scope error will occur is a pre-condition to safe use of an estimation model. We also provide a handy procedure for dealing with outliers as to whether or not to include them in the training set for building a new version of the estimation model. The majority of the work is empirically based, applying computational intelligence techniques to some COCOMO model variations with respect to a publicly available cost estimation data set in the PROMISE repository. %B Proceedings of the 2009 3rd International Symposium on Empirical Software Engineering and Measurement %S ESEM '09 %I IEEE Computer Society %C Washington, DC, USA %P 123 - 132 %8 2009/// %@ 978-1-4244-4842-5 %G eng %U http://dx.doi.org/10.1109/ESEM.2009.5316020 %R 10.1109/ESEM.2009.5316020 %0 Conference Paper %B Proceedings of IS&T Archiving 2009 %D 2009 %T Search and Access Strategies for Web Archives %A Song,Sangchul %A JaJa, Joseph F. %B Proceedings of IS&T Archiving 2009 %8 2009/// %G eng %0 Journal Article %J Genome Biology %D 2009 %T Searching for SNPs with cloud computing %A Langmead,Ben %A Schatz,Michael C %A Jimmy Lin %A Pop, Mihai %A Salzberg,Steven L. %X As DNA sequencing outpaces improvements in computer speed, there is a critical need to accelerate tasks like alignment and SNP calling. Crossbow is a cloud-computing software tool that combines the aligner Bowtie and the SNP caller SOAPsnp. Executing in parallel using Hadoop, Crossbow analyzes data comprising 38-fold coverage of the human genome in three hours using a 320-CPU cluster rented from a cloud computing service for about $85. Crossbow is available from http://bowtie-bio.sourceforge.net/crossbow/. %B Genome Biology %V 10 %P R134 - R134 %8 2009/11/20/ %@ 1465-6906 %G eng %U http://genomebiology.com/2009/10/11/R134 %N 11 %R 10.1186/gb-2009-10-11-r134 %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on %D 2009 %T Secure image retrieval through feature protection %A Lu,Wenjun %A Varna,A.L. %A Swaminathan,A. %A M. Wu %K bit-plane %K confidentiality;data %K database;image %K databases; %K encoding;signal %K extraction;image %K feature %K processes;visual %K processing;content-based %K projection;randomized %K protection;image %K randomization;cryptographic %K retrieval;cryptography;feature %K retrieval;random %K storage;encrypted %K technique;data %K unary %X This paper addresses the problem of image retrieval from an encrypted database, where data confidentiality is preserved both in the storage and retrieval process. The paper focuses on image feature protection techniques which enable similarity comparison among protected features. By utilizing both signal processing and cryptographic techniques, three schemes are investigated and compared, including bit-plane randomization, random projection, and randomized unary encoding. Experimental results show that secure image retrieval can achieve comparable retrieval performance to conventional image retrieval techniques without revealing information about image content. This work enriches the area of secure information retrieval and can find applications in secure online services for images and videos. %B Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on %P 1533 - 1536 %8 2009/04// %G eng %R 10.1109/ICASSP.2009.4959888 %0 Journal Article %J SIAM Journal on Matrix Analysis and Applications %D 2009 %T On the Semidefinite B-Arnoldi Method %A Stewart, G.W. %K Arnoldi's method %K B-inner product %K null-space component %K semidefinite B %X The B-Arnoldi method is a variant of the ordinary Arnoldi method in which orthogonalization is done with respect to the inner product generated by a positive definite matrix $B$. It arises in connection with the generalized eigenvalue problem $Ax = \lambda Bx$. When $B$ is semidefinite, the algorithm can proceed formally, with “orthogonalization” taking place in the semi-inner product generated by $B$. However, it has been observed that components of the Arnoldi vectors lying in the null space of $B$ can grow rapidly. In this paper we examine the source and consequences of this growth. %B SIAM Journal on Matrix Analysis and Applications %V 31 %P 1458 - 1468 %8 2009/// %G eng %U http://link.aip.org/link/?SML/31/1458/1 %N 3 %R 10.1137/090759252 %0 Journal Article %J Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %D 2009 %T Serogroup, Virulence, and Genetic Traits of Vibrio Parahaemolyticus in the Estuarine Ecosystem of Bangladesh %A Alam,Munirul %A Chowdhury,Wasimul B. %A Bhuiyan,N. A. %A Islam,Atiqul %A Hasan,Nur A. %A Nair,G. Balakrish %A Watanabe,H. %A Siddique,A. K. %A Huq,Anwar %A Sack,R. Bradley %A Akhter,M. Z. %A Grim,Christopher J. %A Kam,K.-M. %A Luey,C. K. Y. %A Endtz,Hubert P. %A Cravioto,Alejandro %A Rita R Colwell %X Forty-two strains of Vibrio parahaemolyticus were isolated from Bay of Bengal estuaries and, with two clinical strains, analyzed for virulence, phenotypic, and molecular traits. Serological analysis indicated O8, O3, O1, and K21 to be the major O and K serogroups, respectively, and O8:K21, O1:KUT, and O3:KUT to be predominant. The K antigen(s) was untypeable, and pandemic serogroup O3:K6 was not detected. The presence of genes toxR and tlh were confirmed by PCR in all but two strains, which also lacked toxR. A total of 18 (41%) strains possessed the virulence gene encoding thermostable direct hemolysin (TDH), and one had the TDH-related hemolysin (trh) gene, but not tdh. Ten (23%) strains exhibited Kanagawa phenomenon that surrogates virulence, of which six, including the two clinical strains, possessed tdh. Of the 18 tdh-positive strains, 17 (94%), including the two clinical strains, had the seromarker O8:K21, one was O9:KUT, and the single trh-positive strain was O1:KUT. None had the group-specific or ORF8 pandemic marker gene. DNA fingerprinting employing pulsed-field gel electrophoresis (PFGE) of SfiI-digested DNA and cluster analysis showed divergence among the strains. Dendrograms constructed using PFGE (SfiI) images from a soft database, including those of pandemic and nonpandemic strains of diverse geographic origin, however, showed that local strains formed a cluster, i.e., “clonal cluster,” as did pandemic strains of diverse origin. The demonstrated prevalence of tdh-positive and diarrheagenic serogroup O8:K21 strains in coastal villages of Bangladesh indicates a significant human health risk for inhabitants. %B Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %V 75 %P 6268 - 6274 %8 2009/10/01/ %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/75/19/6268 %N 19 %R 10.1128/AEM.00266-09 %0 Thesis %D 2009 %T Shape Identification in Temporal Data Sets %A Gregory,M.B. %A Shneiderman, Ben %X Shapes are a concise way to describe temporal variable behaviors.Some commonly used shapes are spikes, sinks, rises, and drops. A spike describes a set of variable values that rapidly increase, then immediately rapidly decrease. The variable may be the value of a stock or a person’s blood sugar levels. Shapes are abstract. Details such as the height of spike or its rate increase, are lost in the ab- straction. These hidden details make it difficult to define shapes and compare one to another. For example, what attributes of a spike determine its “spikiness”? The ability to define and com- pare shapes is important because it allows shapes to be identified and ranked, according to an attribute of interest. Work has been done in the area of shape identification through pattern matching and other data mining techniques, but ideas combining the identifi- cation and comparison of shapes have received less attention. This paper fills the gap by presenting a set of shapes and the attributes by which they can identified, compared, and ranked. Neither the set of shapes, nor their attributes presented in this paper are exhaustive, but it provides an example of how a shape’s attributes can be used for identification and comparison. The intention of this paper is not to replace any particular mathematical method of identifying a par- ticular behavior, but to provide a toolset for knowledge discovery and an intuitive method of data mining for novices. Spikes, sinks, rises, drops, lines, plateaus, valleys, and gaps are the shapes pre- sented in this paper. Several attributes for each shape are defined. These attributes will be the basis for constructing definitions that allow the shapes to be identified and ranked. The second contri- bution is an information visualization tool, TimeSearcher: Shape Search Edition (SSE), which allows users to explore data sets using the identification and ranking ideas in this paper. %I Master’s thesis, University of Maryland %8 2009/// %G eng %0 Journal Article %J GeoSpatial Visual Analytics %D 2009 %T Sorting Spatial Data by Spatial Occupancy %A Samet, Hanan %X The increasing popularity of web-based mapping services such as Microsoft Virtual Earth and Google Maps/Earth has led to a dramatic increase in awareness of the importance of location as a component of data for the purposes of further processing as a means of enhancing the value of the nonspatial data and of visualization. Both of these purposes inevitably involve searching. The efficiency of searching is dependent on the extent to which the underlying data is sorted. The sorting is encapsulated by the data structure known as an index that is used to represent the spatial data thereby making it more accessible. The traditional role of the indexes is to sort the data, which means that they order the data. However, since generally no ordering exists in dimensions greater than 1 without a transformation of the data to one dimension, the role of the sort process is one of differentiating between the data and what is usually done is to sort the spatial objects with respect to the space that they occupy. The resulting ordering should be implicit rather than explicit so that the data need not be resorted (i.e., the index need not be rebuilt) when the queries change. The indexes are said to order the space and the characteristics of such indexes are explored further. %B GeoSpatial Visual Analytics %P 31 - 43 %8 2009/// %G eng %R 10.1007/978-90-481-2899-0_3 %0 Conference Paper %B Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems %D 2009 %T Spatio-textual spreadsheets: geotagging via spatial coherence %A Lieberman,Michael D. %A Samet, Hanan %A Sankaranarayanan,Jagan %A Sperling,Jon %K geotagging %K spatial coherence %K spatio-textual %K spreadsheets %X The spatio-textual spreadsheet is a conventional spreadsheet where spatial attribute values are specified textually. Techniques are presented to automatically find the textually-specified spatial attributes that are present in spreadsheets. Once the spatial attributes have been identified, an accurate translation of the values of the spatial attributes to their actual geographic locations is needed (known as geotagging). The key observation is that spreadsheets with spatial data exhibit spatial coherence --- that is, cells with spatial data that are nearby in the spreadsheet contain data that share spatial characteristics in the real world. These techniques also allow richer search engine results by returning actual tuples from spreadsheets instead of simply links to the spreadsheets. Moreover, when the search key is a particular location, results in proximity to the query can be provided rather than just exact matches. %B Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems %S GIS '09 %I ACM %C New York, NY, USA %P 524 - 527 %8 2009/// %@ 978-1-60558-649-6 %G eng %U http://doi.acm.org/10.1145/1653771.1653860 %R 10.1145/1653771.1653860 %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 2009 %T SPOT Databases: Efficient Consistency Checking and Optimistic Selection in Probabilistic Spatial Databases %A Parker,A. %A Infantes,G. %A Grant,J. %A V.S. Subrahmanian %K Cartesian %K checking;database %K database;data %K databases; %K databases;visual %K indexing;inference %K indexing;linear %K integrity;database %K mechanisms;linear %K probabilistic %K problem;spatial %K problems;temporal %K processing;search %K program;optimistic %K programming;probability;query %K query;probabilistic %K reasoning;search %K selection %K selection;consistency %K space;cautious %K temporal %X Spatial probabilistic temporal (SPOT) databases are a paradigm for reasoning with probabilistic statements about where a vehicle may be now or in the future. They express statements of the form "Object O is in spatial region R at some time t with some probability in the interval [L,U]." Past work on SPOT databases has developed selection operators based on selecting SPOT atoms that are entailed by the SPOT database-we call this "cautious" selection. In this paper, we study several problems. First, we note that the runtime of consistency checking and cautious selection algorithms in past work is influenced greatly by the granularity of the underlying Cartesian space. In this paper, we first introduce the notion of "optimistic" selection, where we are interested in returning all SPOT atoms in a database that are consistent with respect to a query, rather than having an entailment relationship. We then develop an approach to scaling SPOT databases that has three main contributions: 1) We develop methods to eliminate variables from the linear programs used in past work, thus greatly reducing the size of the linear programs used-the resulting advances apply to consistency checking, optimistic selection, and cautious selection. 2) We develop a host of theorems to show how we can prune the search space when we are interested in optimistic selection. 3) We use the above contributions to build an efficient index to execute optimistic selection queries over SPOT databases. Our approach is superior to past work in two major respects: First, it makes fewer assumptions than all past works on this topic except that in. Second, our experiments, which are based on real-world data about ship movements, show that our algorithms are much more efficient than those in. %B Knowledge and Data Engineering, IEEE Transactions on %V 21 %P 92 - 107 %8 2009/01// %@ 1041-4347 %G eng %N 1 %R 10.1109/TKDE.2008.93 %0 Journal Article %J Molecular Cell %D 2009 %T Structure of the S5a:K48-Linked Diubiquitin Complex and Its Interactions with Rpn13 %A Zhang,Naixia %A Wang,Qinghua %A Ehlinger,Aaron %A Randles,Leah %A Lary,Jeffrey W. %A Kang,Yang %A Haririnia,Aydin %A Storaska,Andrew J. %A Cole,James L. %A Fushman, David %A Walters,Kylie J. %K Proteins %X SummaryDegradation by the proteasome typically requires substrate ubiquitination. Two ubiquitin receptors exist in the proteasome, S5a/Rpn10 and Rpn13. Whereas Rpn13 has only one ubiquitin-binding surface, S5a binds ubiquitin with two independent ubiquitin-interacting motifs (UIMs). Here, we use nuclear magnetic resonance (NMR) and analytical ultracentrifugation to define at atomic level resolution how S5a binds K48-linked diubiquitin, in which K48 of one ubiquitin subunit (the “proximal” one) is covalently bonded to G76 of the other (the “distal” subunit). We demonstrate that S5a's UIMs bind the two subunits simultaneously with a preference for UIM2 binding to the proximal subunit while UIM1 binds to the distal one. In addition, NMR experiments reveal that Rpn13 and S5a bind K48-linked diubiquitin simultaneously with subunit specificity, and a model structure of S5a and Rpn13 bound to K48-linked polyubiquitin is provided. Altogether, our data demonstrate that S5a is highly adaptive and cooperative toward binding ubiquitin chains. %B Molecular Cell %V 35 %P 280 - 290 %8 2009/08/14/ %@ 1097-2765 %G eng %U http://www.sciencedirect.com/science/article/pii/S1097276509004018 %N 3 %R 10.1016/j.molcel.2009.06.010 %0 Conference Paper %B Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 1 - Volume 1 %D 2009 %T Supervised and unsupervised methods in employing discourse relations for improving opinion polarity classification %A Somasundaran,Swapna %A Namata,Galileo %A Wiebe,Janyce %A Getoor, Lise %X This work investigates design choices in modeling a discourse scheme for improving opinion polarity classification. For this, two diverse global inference paradigms are used: a supervised collective classification framework and an unsupervised optimization framework. Both approaches perform substantially better than baseline approaches, establishing the efficacy of the methods and the underlying discourse scheme. We also present quantitative and qualitative analyses showing how the improvements are achieved. %B Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing: Volume 1 - Volume 1 %S EMNLP '09 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 170 - 179 %8 2009/// %@ 978-1-932432-59-6 %G eng %U http://dl.acm.org/citation.cfm?id=1699510.1699533 %0 Journal Article %J Bioinspiration & Biomimetics %D 2009 %T A survey of snake-inspired robot designs %A Hopkins,James K. %A Spranklin,Brent W. %A Gupta, Satyandra K. %X Body undulation used by snakes and the physical architecture of a snake body may offer significant benefits over typical legged or wheeled locomotion designs in certain types of scenarios. A large number of research groups have developed snake-inspired robots to exploit these benefits. The purpose of this review is to report different types of snake-inspired robot designs and categorize them based on their main characteristics. For each category, we discuss their relative advantages and disadvantages. This review will assist in familiarizing a newcomer to the field with the existing designs and their distinguishing features. We hope that by studying existing robots, future designers will be able to create new designs by adopting features from successful robots. The review also summarizes the design challenges associated with the further advancement of the field and deploying snake-inspired robots in practice. %B Bioinspiration & Biomimetics %V 4 %P 021001 - 021001 %8 2009/06/01/ %@ 1748-3182, 1748-3190 %G eng %U http://iopscience.iop.org/1748-3182/4/2/021001 %N 2 %R 10.1088/1748-3182/4/2/021001 %0 Conference Paper %B Proceedings of the 6th USENIX symposium on Networked systems design and implementation %D 2009 %T Symbiotic relationships in internet routing overlays %A Lumezanu,Cristian %A Baden,Randy %A Levin,Dave %A Spring, Neil %A Bhattacharjee, Bobby %X We propose to construct routing overlay networks using the following principle: that overlay edges should be based on mutual advantage between pairs of hosts. Upon this principle, we design, implement, and evaluate Peer-Wise, a latency-reducing overlay network. To show the feasibility of PeerWise, we must show first that mutual advantage exists in the Internet: perhaps contrary to expectation, that there are not only "haves" and "have nots" of low-latency connectivity. Second, we must provide a scalable means of finding promising edges and overlay routes; we seek embedding error in network coordinates to expose both shorter-than-default "detour" routes and longer-than-expected default routes. We evaluate the cost of limiting PeerWise to mutually advantageous links, then build the intelligent components that put PeerWise into practice. We design and evaluate "virtual" network coordinates for destinations not participating in the overlay, neighbor selection algorithms to find promising relays, and relay selection algorithms to choose the neighbor to traverse for a good detour. Finally, we show that PeerWise is practical through a wide-area deployment and evaluation. %B Proceedings of the 6th USENIX symposium on Networked systems design and implementation %S NSDI'09 %I USENIX Association %C Berkeley, CA, USA %P 467 - 480 %8 2009/// %G eng %U http://dl.acm.org/citation.cfm?id=1558977.1559009 %0 Patent %D 2009 %T System and method for fast illumination-invariant background subtraction ... %A Lim,Ser-Nam %A Mittal,Anurag %A Davis, Larry S. %E Siemens Corporate Research, Inc. %X A method for eliminating errors in foreground object detection in digitized images comprises providing a reference camera and a secondary camera, vertically aligning each said camera with a baseline that is approximately perpendicular to a ground plane, wherein said reference camera is placed lower than said secondary camera, selecting a foreground pixel in a reference view of a first point in a foreground object, finding a conjugate pixel of the foreground pixel in a secondary view, using the foreground and conjugate pixels to determine an image base pixel of a base point in the reference view, wherein said base point is a point on the ground plane below the first point, and using the foreground and image base pixels to find a location where the ground plane is first visible. %V 11/282,513 %8 2009/03/31/ %G eng %U http://www.google.com/patents?id=8yK_AAAAEBAJ %N 7512250 %0 Conference Paper %B In Proceedings of the ACM International Workshop on Timing Issues in the Specification and Synthesis of Digital Systems, Austin, Texas, February 2009. %D 2009 %T System-level Clustering and Timing Analysis for GALS-based Dataflow Architectures %A Shen,C. %A Bhattacharyya, Shuvra S. %X In this paper, we propose an approach based ondataflow techniques for modeling application-specific, globally asynchronous, locally synchronous (GALS) architectures for digital signal processing (DSP) applications, and analyzing the performance of such architectures. Dataflow-based techniques are attractive for DSP applications because they allow applica- tion behavior to be represented formally, analyzed at a high level of abstraction, and synthesized to software/hardware implementations through an optimized, automated process. In our proposed methodology, we employ dataflow-based compu- tational models to expose relevant structure in the targeted applications, and facilitate the manual or automatic derivation of efficient implementations. We demonstrate the utility of our modeling and analysis techniques by applying them as core parts of a novel clustering algorithm that is geared towards optimizing the throughput of GALS-based DSP architectures. %B In Proceedings of the ACM International Workshop on Timing Issues in the Specification and Synthesis of Digital Systems, Austin, Texas, February 2009. %8 2009/// %G eng %0 Conference Paper %B Game Theory for Networks, 2009. GameNets '09. International Conference on %D 2009 %T Systems-compatible incentives %A Levin,D. %A Spring, Neil %A Bhattacharjee, Bobby %K computing; %K incentives;game %K Internet;distributed %K system;distributed %K systems;game %K systems;systems-compatible %K theory;incentive-compatible %K theory;peer-to-peer %X Selfish participants in a distributed system attempt to gain from the system without regard to how their actions may affect others. To maintain desirable system-wide properties in the presence of selfish users, designers are increasingly turning to the powerful mechanisms offered by economics and game theory. Combining the two fields of economics and systems design introduces new challenges of achieving incentive-compatibility in systems we can deploy in today's Internet. In this paper, we explore the interactions between systems and the mechanisms that give users incentives to cooperate. Using findings from recent work on incentive-compatible systems, we discuss several economic mechanisms and assumptions: money, punishment, and altruism. We seek to understand when these mechanisms violate system properties. Among the potential pitfalls we present is a phenomenon we call the price of altruism: altruistic peers can impose a loss of social good in some systems. We also discuss systems-compatible mechanisms that have been used in real, distributed systems, and attempt to extract the underlying design principles that have led to their success. %B Game Theory for Networks, 2009. GameNets '09. International Conference on %P 100 - 106 %8 2009/05// %G eng %R 10.1109/GAMENETS.2009.5137390 %0 Conference Paper %D 2009 %T TACKing Together Efficient Authentication, Revocation, and Privacy in VANETs %A Studer, A. %A Elaine Shi %A Bai, Fan %A Perrig, A. %K ad hoc networks %K data privacy %K eavesdropper %K IEEE 1609.2 standard %K Message authentication %K mobile radio %K public key cryptography %K public key infrastructure %K remove malevolent vehicle %K telecommunication security %K temporary anonymous certified key %K valid vehicle identification %K VANET key management %K VANET security %K vehicle-to-vehicle communication %K Vehicles %K vehicular ad hoc network %X Vehicular ad hoc networks (VANETs) require a mechanism to help authenticate messages, identify valid vehicles, and remove malevolent vehicles. A public key infrastructure (PKI) can provide this functionality using certificates and fixed public keys. However, fixed keys allow an eavesdropper to associate a key with a vehicle and a location, violating drivers' privacy. In this work we propose a VANET key management scheme based on temporary anonymous certified keys (TACKs). Our scheme efficiently prevents eavesdroppers from linking a vehicle's different keys and provides timely revocation of misbehaving participants while maintaining the same or less overhead for vehicle-to-vehicle communication as the current IEEE 1609.2 standard for VANET security. %P 1 - 9 %8 2009 %G eng %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on %D 2009 %T Tampering identification using Empirical Frequency Response %A Chuang,Wei-Hong %A Swaminathan,A. %A M. Wu %K aproach;JPEG %K coding;multimedia %K compression;frequency %K compression;LSI %K computing;security %K data; %K editing %K EFR %K forensics;tampering-type %K Frequency %K identification;data %K image %K invariant;multimedia %K of %K response;image %K response;linear %K shift %K software;empirical %K system;digital %X With the widespread popularity of digital images and the presence of easy-to-use image editing software, content integrity can no longer be taken for granted, and there is a strong need for techniques that not only detect the presence of tampering but also identify its type. This paper focusses on tampering-type identification and introduces a new approach based on the empirical frequency response (EFR) to address this problem. We show that several types of tampering operations, both linear shift invariant (LSI) and non-LSI, can be characterized consistently and distinctly by their EFRs. We then extend the approach to estimate the EFR for scenarios where only the final image is available. Theoretical reasoning supported by experimental results verify the effectiveness of this method for identifying the type of a tampering operation. %B Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on %P 1517 - 1520 %8 2009/04// %G eng %R 10.1109/ICASSP.2009.4959884 %0 Journal Article %J JSCUT (Natural Science Edition) %D 2009 %T Target Tracking Based on Nonparametric Clustering and Multi-Scale Images %A Shao-fa,J.Z.L. %A Hong-li,J.X.Z. %X In order to track a target in space and scale in a complex background,a target tracking algorithm based on the nonparametric clustering and multi-scale images is presented.In this algorithm,first,a modified nonparame-tric color-clustering method is employed to adaptively partition the color space of a tracked object,and the Gaussian function is used to model the spatial information of each bin of the color histogram.Next,the Bhattacharyya coefficient is adopted to derive a function describing the similarity between the target model and the target candidate.Then,a coarse-to-fine approach of multi-scale images is employed to implement the spatial location of the tracked object.Finally,the derived automatic bandwidth selection method of kernel function is applied to obtain the scale of the tracked object.Experimental results show that the proposed algorithm outperforms the classical mean shift tracker. %B JSCUT (Natural Science Edition) %V 37 %P 34 - 41 %8 2009/// %G eng %N 1 %0 Journal Article %J International Journal on Digital Libraries %D 2009 %T Techniques to audit and certify the long-term integrity of digital archives %A Song,Sangchul %A JaJa, Joseph F. %K Computer science %X A fundamental requirement for a digital archive is to set up mechanisms that will ensure the authenticity of its holdings in the long term. In this article, we develop a new methodology to address the long-term integrity of digital archives using rigorous cryptographic techniques. Our approach involves the generation of a small-size integrity token for each object, some cryptographic summary information, and a framework that enables cost-effective regular and periodic auditing of the archive’s holdings depending on the policy set by the archive. Our scheme is very general, architecture and platform independent, and can detect with high probability any alteration to an object, including malicious alterations introduced by the archive or by an external intruder. The scheme can be shown to be mathematically correct as long as a small amount of cryptographic information, in the order of 100 KB/year, can be kept intact. Using this approach, a prototype system called ACE (Auditing Control Environment) has been built and tested in an operational large scale archiving environment. %B International Journal on Digital Libraries %V 10 %P 123 - 131 %8 2009/// %@ 1432-5012 %G eng %U http://www.springerlink.com/content/y52815g805h96334/abstract/ %N 2 %R 10.1007/s00799-009-0056-2 %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2009 %T Temporal Summaries: Supporting Temporal Categorical Searching, Aggregation and Comparison %A Wang,T. D %A Plaisant, Catherine %A Shneiderman, Ben %A Spring, Neil %A Roseman,D. %A Marchand,G. %A Mukherjee,V. %A Smith,M. %K Aggregates %K Collaborative work %K Computational Biology %K Computer Graphics %K Data analysis %K data visualisation %K Data visualization %K Databases, Factual %K Displays %K Event detection %K Filters %K Heparin %K History %K Human computer interaction %K Human-computer interaction %K HUMANS %K Information Visualization %K Interaction design %K interactive visualization technique %K Medical Records Systems, Computerized %K Pattern Recognition, Automated %K Performance analysis %K Springs %K temporal categorical data visualization %K temporal categorical searching %K temporal ordering %K temporal summaries %K Thrombocytopenia %K Time factors %X When analyzing thousands of event histories, analysts often want to see the events as an aggregate to detect insights and generate new hypotheses about the data. An analysis tool must emphasize both the prevalence and the temporal ordering of these events. Additionally, the analysis tool must also support flexible comparisons to allow analysts to gather visual evidence. In a previous work, we introduced align, rank, and filter (ARF) to accentuate temporal ordering. In this paper, we present temporal summaries, an interactive visualization technique that highlights the prevalence of event occurrences. Temporal summaries dynamically aggregate events in multiple granularities (year, month, week, day, hour, etc.) for the purpose of spotting trends over time and comparing several groups of records. They provide affordances for analysts to perform temporal range filters. We demonstrate the applicability of this approach in two extensive case studies with analysts who applied temporal summaries to search, filter, and look for patterns in electronic health records and academic records. %B IEEE Transactions on Visualization and Computer Graphics %V 15 %P 1049 - 1056 %8 2009/12//Nov %@ 1077-2626 %G eng %N 6 %R 10.1109/TVCG.2009.187 %0 Journal Article %J Visualization and Computer Graphics, IEEE Transactions on %D 2009 %T Temporal Summaries: Supporting Temporal Categorical Searching, Aggregation and Comparison %A Wang,T. D %A Plaisant, Catherine %A Shneiderman, Ben %A Spring, Neil %A Roseman,D. %A Marchand,G. %A Mukherjee,V. %A Smith,M. %K Automated;Thrombocytopenia;Time Factors; %K Computerized;Pattern Recognition %K Factual;Heparin;Humans;Medical Records Systems %K interactive visualization technique;temporal categorical searching;temporal ordering;temporal summaries;data visualisation;human computer interaction;Computational Biology;Computer Graphics;Databases %X When analyzing thousands of event histories, analysts often want to see the events as an aggregate to detect insights and generate new hypotheses about the data. An analysis tool must emphasize both the prevalence and the temporal ordering of these events. Additionally, the analysis tool must also support flexible comparisons to allow analysts to gather visual evidence. In a previous work, we introduced align, rank, and filter (ARF) to accentuate temporal ordering. In this paper, we present temporal summaries, an interactive visualization technique that highlights the prevalence of event occurrences. Temporal summaries dynamically aggregate events in multiple granularities (year, month, week, day, hour, etc.) for the purpose of spotting trends over time and comparing several groups of records. They provide affordances for analysts to perform temporal range filters. We demonstrate the applicability of this approach in two extensive case studies with analysts who applied temporal summaries to search, filter, and look for patterns in electronic health records and academic records. %B Visualization and Computer Graphics, IEEE Transactions on %V 15 %P 1049 - 1056 %8 2009/12//nov %@ 1077-2626 %G eng %N 6 %R 10.1109/TVCG.2009.187 %0 Journal Article %J Machine Translation %D 2009 %T TER-Plus: paraphrase, semantic, and alignment enhancements to Translation Edit Rate %A Snover,Matthew %A Madnani,Nitin %A Dorr, Bonnie J %A Schwartz,Richard %X This paper describes a new evaluation metric, TER-Plus (TERp) for automatic evaluation of machine translation (MT). TERp is an extension of Translation Edit Rate (TER). It builds on the success of TER as an evaluation metric and alignment tool and addresses several of its weaknesses through the use of paraphrases, stemming, synonyms, as well as edit costs that can be automatically optimized to correlate better with various types of human judgments. We present a correlation study comparing TERp to BLEU, METEOR and TER, and illustrate that TERp can better evaluate translation adequacy. %B Machine Translation %V 23 %P 117 - 127 %8 2009/// %@ 0922-6567 %G eng %U http://dx.doi.org/10.1007/s10590-009-9062-9 %N 2 %0 Journal Article %J Encyclopedia of Information Science and Technology %D 2009 %T Testing graphical user interfaces %A Strecker,J. %A Memon, Atif M. %B Encyclopedia of Information Science and Technology %8 2009/// %G eng %0 Conference Paper %B Proceedings of the 14th ACM SIGPLAN international conference on Functional programming %D 2009 %T A theory of typed coercions and its applications %A Swamy,Nikhil %A Hicks, Michael W. %A Bierman,Gavin M. %K coercion insertion %K gradual typing %K nonambiguity %K provenance %K type-directed translation %X A number of important program rewriting scenarios can be recast as type-directed coercion insertion. These range from more theoretical applications such as coercive subtyping and supporting overloading in type theories, to more practical applications such as integrating static and dynamically typed code using gradual typing, and inlining code to enforce security policies such as access control and provenance tracking. In this paper we give a general theory of type-directed coercion insertion. We specifically explore the inherent tradeoff between expressiveness and ambiguity--the more powerful the strategy for generating coercions, the greater the possibility of several, semantically distinct rewritings for a given program. We consider increasingly powerful coercion generation strategies, work out example applications supported by the increased power (including those mentioned above), and identify the inherent ambiguity problems of each setting, along with various techniques to tame the ambiguities. %B Proceedings of the 14th ACM SIGPLAN international conference on Functional programming %S ICFP '09 %I ACM %C New York, NY, USA %P 329 - 340 %8 2009/// %@ 978-1-60558-332-7 %G eng %U http://doi.acm.org/10.1145/1596550.1596598 %R 10.1145/1596550.1596598 %0 Journal Article %J Appl Environ Microbiol %D 2009 %T Three genomes from the phylum Acidobacteria provide insight into the lifestyles of these microorganisms in soils. %A Ward, Naomi L %A Challacombe, Jean F %A Janssen, Peter H %A Henrissat, Bernard %A Coutinho, Pedro M %A Wu, Martin %A Xie, Gary %A Haft, Daniel H %A Sait, Michelle %A Badger, Jonathan %A Barabote, Ravi D %A Bradley, Brent %A Brettin, Thomas S %A Brinkac, Lauren M %A Bruce, David %A Creasy, Todd %A Daugherty, Sean C %A Davidsen, Tanja M %A DeBoy, Robert T %A Detter, J Chris %A Dodson, Robert J %A Durkin, A Scott %A Ganapathy, Anuradha %A Gwinn-Giglio, Michelle %A Han, Cliff S %A Khouri, Hoda %A Kiss, Hajnalka %A Kothari, Sagar P %A Madupu, Ramana %A Nelson, Karen E %A Nelson, William C %A Paulsen, Ian %A Penn, Kevin %A Ren, Qinghu %A Rosovitz, M J %A Jeremy D Selengut %A Shrivastava, Susmita %A Sullivan, Steven A %A Tapia, Roxanne %A Thompson, L Sue %A Watkins, Kisha L %A Yang, Qi %A Yu, Chunhui %A Zafar, Nikhat %A Zhou, Liwei %A Kuske, Cheryl R %K Anti-Bacterial Agents %K bacteria %K Biological Transport %K Carbohydrate Metabolism %K Cyanobacteria %K DNA, Bacterial %K Fungi %K Genome, Bacterial %K Macrolides %K Molecular Sequence Data %K Nitrogen %K Phylogeny %K Proteobacteria %K Sequence Analysis, DNA %K sequence homology %K Soil Microbiology %X

The complete genomes of three strains from the phylum Acidobacteria were compared. Phylogenetic analysis placed them as a unique phylum. They share genomic traits with members of the Proteobacteria, the Cyanobacteria, and the Fungi. The three strains appear to be versatile heterotrophs. Genomic and culture traits indicate the use of carbon sources that span simple sugars to more complex substrates such as hemicellulose, cellulose, and chitin. The genomes encode low-specificity major facilitator superfamily transporters and high-affinity ABC transporters for sugars, suggesting that they are best suited to low-nutrient conditions. They appear capable of nitrate and nitrite reduction but not N(2) fixation or denitrification. The genomes contained numerous genes that encode siderophore receptors, but no evidence of siderophore production was found, suggesting that they may obtain iron via interaction with other microorganisms. The presence of cellulose synthesis genes and a large class of novel high-molecular-weight excreted proteins suggests potential traits for desiccation resistance, biofilm formation, and/or contribution to soil structure. Polyketide synthase and macrolide glycosylation genes suggest the production of novel antimicrobial compounds. Genes that encode a variety of novel proteins were also identified. The abundance of acidobacteria in soils worldwide and the breadth of potential carbon use by the sequenced strains suggest significant and previously unrecognized contributions to the terrestrial carbon cycle. Combining our genomic evidence with available culture traits, we postulate that cells of these isolates are long-lived, divide slowly, exhibit slow metabolic rates under low-nutrient conditions, and are well equipped to tolerate fluctuations in soil hydration.

%B Appl Environ Microbiol %V 75 %P 2046-56 %8 2009 Apr %G eng %N 7 %R 10.1128/AEM.02294-08 %0 Conference Paper %B Empirical Software Engineering and Measurement, 2009. ESEM 2009. 3rd International Symposium on %D 2009 %T Tool supported detection and judgment of nonconformance in process execution %A Zazworka, N. %A Basili, Victor R. %A Shull, F. %K aerospace %K community;tool %K cycle;software %K detection;software %K development %K domain;large-scale %K engineering %K execution;process %K life %K management;software %K nonconformance %K project;nonfunctional %K requirement;process %K software %K supported %K tools; %X In the past decades the software engineering community has proposed a large collection of software development life cycles, models, and processes. The goal of a major set of these processes is to assure that the product is finished within time and budget, and that a predefined set of functional and nonfunctional requirements (e.g. quality goals) are satisfied at delivery time. Based upon the assumption that there is a real relationship between the process applied and the characteristics of the product developed from that process, we developed a tool supported approach that uses process nonconformance detection to identify potential risks in achieving the required process characteristics. In this paper we present the approach and a feasibility study that demonstrates its use on a large-scale software development project in the aerospace domain. We demonstrate that our approach, in addition to meeting the criteria above, can be applied to a real system of reasonable size; can represent a useful and adequate set of rules of relevance in such an environment; and can detect relevant examples of process nonconformance that provide useful insight to the project manager. %B Empirical Software Engineering and Measurement, 2009. ESEM 2009. 3rd International Symposium on %P 312 - 323 %8 2009/10// %G eng %R 10.1109/ESEM.2009.5315983 %0 Conference Paper %B Indo-US Workshop on International Trends in Digital Preservation %D 2009 %T Tools and Services for Long-Term Preservation of Digital Archives %A JaJa, Joseph F. %A Smorul,M. %A Song,S. %X We have been working on a technology model to support thepreservation and reliable access of long term digital archives. The model is built around a layered object architecture involving modular, extensible components that can gracefully adapt to the evolving technology, standards, and protocols. This has led to the development of methodologies, tools and services to handle a number of core requirements of long term digital archives. Specifically, we have built flexible tools for implementing general ingestion workflows, active monitoring and auditing of the archive’s collections to ensure their long-term availability and integrity, storage organization and indexing to optimize access. These tools are platform and architecture independent, and have been tested using a wide variety of collections on heterogeneous computing platforms. In this paper, we will primarily focus on describing the underpinnings of our software called ACE (Auditing Control Environment), and report on its performance on a large scale distributed environment called Chronopolis. Built on top of rigorous cryptographic techniques, ACE provides a policy- driven, scalable environment to monitor and audit the archive’s contents in a cost effective way. In addition, we will briefly introduce some our recent efforts to deal with storage organization and access of web archives. Long term preservation is a process that must begin before an object is ingested into the archive and must remain active throughout the lifetime of the archive. The ACE tool provides a very flexible environment to actively monitor and audit the contents of a digital archive throughout its lifetime, so as to ensure the availability and integrity of the archive’s holdings with extremely high probability. ACE is based on rigorous cryptographic techniques, and enables periodic auditing of the archive’s holdings at the granularity and frequency set by the manager of the archive. The scheme is cost effective and very general, does not depend on the archive’s architecture, and can detect any alterations, including alterations made by a malicious user. ACE can gracefully adapt to format migrations and changes to the archive’s policies. %B Indo-US Workshop on International Trends in Digital Preservation %8 2009/// %G eng %0 Conference Paper %B Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP: Volume 1-Volume 1 %D 2009 %T Topological ordering of function words in hierarchical phrase-based translation %A Setiawan,H. %A Kan,M.Y. %A Li,H. %A Resnik, Philip %B Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP: Volume 1-Volume 1 %P 324 - 332 %8 2009/// %G eng %0 Journal Article %J BMC Evol Biol %D 2009 %T Toward reconstructing the evolution of advanced moths and butterflies (Lepidoptera: Ditrysia): an initial molecular study %A Regier,J. C %A Zwick,A. %A Cummings, Michael P. %A Kawahara,A. Y %A Cho,S. %A Weller,S. %A Roe,A. %A Baixeras,J. %A Brown,J. W %A Parr,C. %A Davis,DR %A Epstein,M %A Hallwachs,W %A Hausmann,A %A Janzen,DH %A Kitching,IJ %A Solis,MA %A Yen,S-H %A Bazinet,A. L %A Mitter,C %X BACKGROUND: In the mega-diverse insect order Lepidoptera (butterflies and moths; 165,000 described species), deeper relationships are little understood within the clade Ditrysia, to which 98% of the species belong. To begin addressing this problem, we tested the ability of five protein-coding nuclear genes (6.7 kb total), and character subsets therein, to resolve relationships among 123 species representing 27 (of 33) superfamilies and 55 (of 100) families of Ditrysia under maximum likelihood analysis. RESULTS: Our trees show broad concordance with previous morphological hypotheses of ditrysian phylogeny, although most relationships among superfamilies are weakly supported. There are also notable surprises, such as a consistently closer relationship of Pyraloidea than of butterflies to most Macrolepidoptera. Monophyly is significantly rejected by one or more character sets for the putative clades Macrolepidoptera as currently defined (P < 0.05) and Macrolepidoptera excluding Noctuoidea and Bombycoidea sensu lato (P < or = 0.005), and nearly so for the superfamily Drepanoidea as currently defined (P < 0.08). Superfamilies are typically recovered or nearly so, but usually without strong support. Relationships within superfamilies and families, however, are often robustly resolved. We provide some of the first strong molecular evidence on deeper splits within Pyraloidea, Tortricoidea, Geometroidea, Noctuoidea and others.Separate analyses of mostly synonymous versus non-synonymous character sets revealed notable differences (though not strong conflict), including a marked influence of compositional heterogeneity on apparent signal in the third codon position (nt3). As available model partitioning methods cannot correct for this variation, we assessed overall phylogeny resolution through separate examination of trees from each character set. Exploration of "tree space" with GARLI, using grid computing, showed that hundreds of searches are typically needed to find the best-feasible phylogeny estimate for these data. CONCLUSION: Our results (a) corroborate the broad outlines of the current working phylogenetic hypothesis for Ditrysia, (b) demonstrate that some prominent features of that hypothesis, including the position of the butterflies, need revision, and (c) resolve the majority of family and subfamily relationships within superfamilies as thus far sampled. Much further gene and taxon sampling will be needed, however, to strongly resolve individual deeper nodes. %B BMC Evol Biol %V 9 %P 280 - 280 %8 2009/// %G eng %R 10.1186/1471-2148-9-280 %0 Report %D 2009 %T Towards an Internet Connectivity Market %A Feamster, Nick %A Hassan,U. %A Sundaresan,S. %A Valancius,V. %A Johari,R. %A Vazirani,V. %X Today’s Internet achieves end-to-end connectivity through bilateral contracts between neighboring networks; unfortunately, this “one size fits all” connectivity results in less efficient paths, unsold capacity and unmet demand, and sometimes catastrophic market failures that result in global disconnectivity. This paper presents the design and evaluation of MINT, a Market for Internet Transit. MINT is a connectivity market and corresponding set of protocols that allows ISPs to offer path segments on an open market. Edge networks bid for end-to-end paths, and a mediator matches bids for paths to collections of path segments that form end-to-end paths. MINT can be deployed using protocols that are present in today’s routers, and it operates in parallel with the existing routing infrastructure and connectivity market. We present MINT’s market model and protocol design; evaluate how MINT improves efficiency, the utility of edge networks, and the profits of transit networks; and how MINT can operate at Internet scale. %I Georgia Institute of Technology %V GT-CS-09-01 %8 2009/// %G eng %U http://hdl.handle.net/1853/30622 %0 Journal Article %J Evaluation and Usability of Programming Languages and Tools (PLATEAU) PLATEAU 2009 %D 2009 %T Triaging Checklists: a Substitute for a PhD in Static Analysis %A Phang,K.Y. %A Foster, Jeffrey S. %A Hicks, Michael W. %A Sazawal,V. %X Static analysis tools have achieved great success in recentyears in automating the process of detecting defects in soft- ware. However, these sophisticated tools have yet to gain widespread adoption, since many of these tools remain too difficult to understand and use. In previous work, we dis- covered that even with an effective code visualization tool, users still found it hard to determine if warnings reported by these tools were true errors or false warnings. The fun- damental problem users face is to understand enough of the underlying algorithm to determine if a warning is caused by imprecision in the algorithm, a challenge that even experts with PhDs may take a while to achieve. In our current work, we propose to use triaging checklists to provide users with systematic guidance to identify false warnings by taking into account specific sources of imprecision in the particular tool. Additionally, we plan to provide checklist assistants, which is a library of simple analyses designed to aid users in an- swering checklist questions. %B Evaluation and Usability of Programming Languages and Tools (PLATEAU) PLATEAU 2009 %8 2009/// %G eng %0 Book Section %B Passive and Active Network MeasurementPassive and Active Network Measurement %D 2009 %T Triangle Inequality and Routing Policy Violations in the Internet %A Lumezanu,Cristian %A Baden,Randy %A Spring, Neil %A Bhattacharjee, Bobby %E Moon,Sue %E Teixeira,Renata %E Uhlig,Steve %K Computer %K Science %X Triangle inequality violations (TIVs) are the effect of packets between two nodes being routed on the longer direct path between them when a shorter detour path through an intermediary is available. TIVs are a natural, widespread and persistent consequence of Internet routing policies. By exposing opportunities to improve the delay between two nodes, TIVs can help myriad applications that seek to minimize end-to-end latency. However, sending traffic along the detour paths revealed by TIVs may influence Internet routing negatively. In this paper we study the interaction between triangle inequality violations and policy routing in the Internet. We use measured and predicted AS paths between Internet nodes to show that 25% of the detour paths exposed by TIVs are in fact available to BGP but are simply deemed “less efficient”. We also compare the AS paths of detours and direct paths and find that detours use AS edges that are rarely followed by default Internet paths, while avoiding others that BGP seems to prefer. Our study is important both for understanding the various interactions that occur at the routing layer as well as their effects on applications that seek to use TIVs to minimize latency. %B Passive and Active Network MeasurementPassive and Active Network Measurement %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5448 %P 45 - 54 %8 2009/// %@ 978-3-642-00974-7 %G eng %U http://dx.doi.org/10.1007/978-3-642-00975-4_5 %0 Conference Paper %B Proceedings of the 9th ACM SIGCOMM conference on Internet measurement conference %D 2009 %T Triangle inequality variations in the internet %A Lumezanu,Cristian %A Baden,Randy %A Spring, Neil %A Bhattacharjee, Bobby %K latency %K tiv %K triangle inequality violation %K variation %X Triangle inequality violations (TIVs) are important for latency sensitive distributed applications. On one hand, they can expose opportunities to improve network routing by finding shorter paths between nodes. On the other hand, TIVs can frustrate network embedding or positioning systems that treat the Internet as a metric space where the triangle inequality holds. Even though triangle inequality violations are both significant and curious, their study has been limited to aggregate data sets that combine measurements taken over long periods of time. The limitations of these data sets open crucial questions in the design of systems that exploit (or avoid) TIVs: are TIVs stable or transient? Or are they illusions caused by aggregating measurements taken at different times? We collect latency matrices at varying sizes and time granularities and study dynamic properties of triangle inequality violations in the Internet. We show that TIVs are not results of measurement error and that their number varies with time. We examine how latency aggregates of data measured over longer periods of time preserve TIVs. Using medians to compute violations eliminates most of the TIVs that appear sporadically during the measurement but it misses many of the ones that are present for more than five hours. %B Proceedings of the 9th ACM SIGCOMM conference on Internet measurement conference %S IMC '09 %I ACM %C New York, NY, USA %P 177 - 183 %8 2009/// %@ 978-1-60558-771-4 %G eng %U http://doi.acm.org/10.1145/1644893.1644914 %R 10.1145/1644893.1644914 %0 Conference Paper %B Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems %D 2009 %T TwitterStand: news in tweets %A Sankaranarayanan,Jagan %A Samet, Hanan %A Teitler,Benjamin E. %A Lieberman,Michael D. %A Sperling,Jon %K geotagging %K news %K online clustering %K Twitter %X Twitter is an electronic medium that allows a large user populace to communicate with each other simultaneously. Inherent to Twitter is an asymmetrical relationship between friends and followers that provides an interesting social network like structure among the users of Twitter. Twitter messages, called tweets, are restricted to 140 characters and thus are usually very focused. We investigate the use of Twitter to build a news processing system, called TwitterStand, from Twitter tweets. The idea is to capture tweets that correspond to late breaking news. The result is analogous to a distributed news wire service. The difference is that the identities of the contributors/reporters are not known in advance and there may be many of them. Furthermore, tweets are not sent according to a schedule: they occur as news is happening, and tend to be noisy while usually arriving at a high throughput rate. Some of the issues addressed include removing the noise, determining tweet clusters of interest bearing in mind that the methods must be online, and determining the relevant locations associated with the tweets. %B Proceedings of the 17th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems %S GIS '09 %I ACM %C New York, NY, USA %P 42 - 51 %8 2009/// %@ 978-1-60558-649-6 %G eng %U http://doi.acm.org/10.1145/1653771.1653781 %R 10.1145/1653771.1653781 %0 Journal Article %J Genome Biology %D 2009 %T Ultrafast and memory-efficient alignment of short DNA sequences to the human genome %A Langmead,Ben %A Trapnell,Cole %A Pop, Mihai %A Salzberg,Steven L. %X Bowtie is an ultrafast, memory-efficient alignment program for aligning short DNA sequence reads to large genomes. For the human genome, Burrows-Wheeler indexing allows Bowtie to align more than 25 million reads per CPU hour with a memory footprint of approximately 1.3 gigabytes. Bowtie extends previous Burrows-Wheeler techniques with a novel quality-aware backtracking algorithm that permits mismatches. Multiple processor cores can be used simultaneously to achieve even greater alignment speeds. Bowtie is open source http://bowtie.cbcb.umd.edu. %B Genome Biology %V 10 %P R25 - R25 %8 2009/03/04/ %@ 1465-6906 %G eng %U http://genomebiology.com/2009/10/3/R25 %N 3 %R 10.1186/gb-2009-10-3-r25 %0 Journal Article %J ICIS 2009 Proceedings %D 2009 %T Understanding IT Innovations Through Computational Analysis of Discourse %A Tsui,Chia-jung %A Wang,Ping %A Fleischmann,Kenneth %A Oard, Douglas %A Sayeed,Asad %B ICIS 2009 Proceedings %8 2009/01/01/ %G eng %U http://aisel.aisnet.org/icis2009/102 %0 Conference Paper %B International Symposium on Collaborative Technologies and Systems, 2009. CTS '09 %D 2009 %T Understanding social computing participation with visual exploration tools %A Shneiderman, Ben %K Application software %K Books %K Collaborative tools %K Computer science %K Data visualization %K Educational institutions %K History %K International collaboration %K Social network services %K Sociotechnical systems %X The rapid growth of socio-technical systems, social media and social networking websites has raised the importance of understanding the determinants of their success. The pressure to understand success is increased by the shift from playful discretionary applications to mission critical applications in government, business, and civic settings. These include homeland defense, energy sustainability, environmental conservation, disaster response, and community safety. Information visualization tools and statistical methods can both be helpful, but their utility grows when they are well-integrated. This talk will demonstrate novel tools for network evolution and offer a framework for thinking about motivating technology-mediated social participation. %B International Symposium on Collaborative Technologies and Systems, 2009. CTS '09 %I IEEE %P xi-xii - xi-xii %8 2009/05/18/22 %@ 978-1-4244-4584-4 %G eng %R 10.1109/CTS.2009.5067426 %0 Conference Paper %B Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on %D 2009 %T Understanding videos, constructing plots learning a visually grounded storyline model from annotated videos %A Gupta,A. %A Srinivasan,P. %A Shi,Jianbo %A Davis, Larry S. %K (artificial %K action %K activity %K analysis;integer %K AND-OR %K annotation;video %K coding; %K constraint;video %K construction;semantic %K extraction;graph %K framework;plots %K graph;encoding;human %K grounded %K intelligence);spatiotemporal %K learning %K meaning;spatio-temporal %K model %K phenomena;video %K Programming %K programming;learning %K recognition;human %K representation;integer %K storyline %K theory;image %K understanding;visually %X Analyzing videos of human activities involves not only recognizing actions (typically based on their appearances), but also determining the story/plot of the video. The storyline of a video describes causal relationships between actions. Beyond recognition of individual actions, discovering causal relationships helps to better understand the semantic meaning of the activities. We present an approach to learn a visually grounded storyline model of videos directly from weakly labeled data. The storyline model is represented as an AND-OR graph, a structure that can compactly encode storyline variation across videos. The edges in the AND-OR graph correspond to causal relationships which are represented in terms of spatio-temporal constraints. We formulate an Integer Programming framework for action recognition and storyline extraction using the storyline model and visual groundings learned from training data. %B Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on %P 2012 - 2019 %8 2009/06// %G eng %R 10.1109/CVPR.2009.5206492 %0 Journal Article %J Proceedings of the VLDB Endowment %D 2009 %T A unified approach to ranking in probabilistic databases %A Li,Jian %A Saha,Barna %A Deshpande, Amol %X The dramatic growth in the number of application domains that naturally generate probabilistic, uncertain data has resulted in a need for efficiently supporting complex querying and decision-making over such data. In this paper, we present a unified approach to ranking and top-k query processing in probabilistic databases by viewing it as a multi-criteria optimization problem, and by deriving a set of features that capture the key properties of a probabilistic dataset that dictate the ranked result. We contend that a single, specific ranking function may not suffice for probabilistic databases, and we instead propose two parameterized ranking functions, called PRFω and PRFe, that generalize or can approximate many of the previously proposed ranking functions. We present novel generating functions-based algorithms for efficiently ranking large datasets according to these ranking functions, even if the datasets exhibit complex correlations modeled using probabilistic and/xor trees or Markov networks. We further propose that the parameters of the ranking function be learned from user preferences, and we develop an approach to learn those parameters. Finally, we present a comprehensive experimental study that illustrates the effectiveness of our parameterized ranking functions, especially PRFe, at approximating other ranking functions and the scalability of our proposed algorithms for exact or approximate ranking. %B Proceedings of the VLDB Endowment %V 2 %P 502 - 513 %8 2009/08// %@ 2150-8097 %G eng %U http://dl.acm.org/citation.cfm?id=1687627.1687685 %N 1 %0 Journal Article %J J. ACM %D 2009 %T A unified approach to scheduling on unrelated parallel machines %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %K Approximation algorithms %K Randomized rounding %K scheduling under multiple criteria %X We develop a single rounding algorithm for scheduling on unrelated parallel machines; this algorithm works well with the known linear programming-, quadratic programming-, and convex programming-relaxations for scheduling to minimize completion time, makespan, and other well-studied objective functions. This algorithm leads to the following applications for the general setting of unrelated parallel machines: (i) a bicriteria algorithm for a schedule whose weighted completion-time and makespan simultaneously exhibit the current-best individual approximations for these criteria; (ii) better-than-two approximation guarantees for scheduling to minimize the Lp norm of the vector of machine-loads, for all 1 < p < ∞; and (iii) the first constant-factor multicriteria approximation algorithms that can handle the weighted completion-time and any given collection of integer Lp norms. Our algorithm has a natural interpretation as a melding of linear-algebraic and probabilistic approaches. Via this view, it yields a common generalization of rounding theorems due to Karp et al. [1987] and Shmoys & Tardos [1993], and leads to improved approximation algorithms for the problem of scheduling with resource-dependent processing times introduced by Grigoriev et al. [2007]. %B J. ACM %V 56 %P 28:1–28:31 - 28:1–28:31 %8 2009/08// %@ 0004-5411 %G eng %U http://doi.acm.org/10.1145/1552285.1552289 %N 5 %R 10.1145/1552285.1552289 %0 Conference Paper %B Proceedings of the Fourth Workshop on Statistical Machine Translation %D 2009 %T The University of Maryland statistical machine translation system for the Fourth Workshop on Machine Translation %A Dyer,C. %A Setiawan,H. %A Marton,Y. %A Resnik, Philip %B Proceedings of the Fourth Workshop on Statistical Machine Translation %P 145 - 149 %8 2009/// %G eng %0 Report %D 2009 %T An Unreliable Convergence Criterion for Arnoldi’s Method %A Stewart, G.W. %I Technical Report, CMSC TR-4938, University of Maryland %8 2009/// %G eng %0 Journal Article %J ACM Computing Surveys %D 2009 %T Using formal specifications to support testing %A Hierons,Robert M. %A Krause,Paul %A Lüttgen,Gerald %A Simons,Anthony J. H. %A Vilkomir,Sergiy %A Woodward,Martin R. %A Zedan,Hussein %A Bogdanov,Kirill %A Bowen,Jonathan P. %A Cleaveland, Rance %A Derrick,John %A Dick,Jeremy %A Gheorghe,Marian %A Harman,Mark %A Kapoor,Kalpesh %B ACM Computing Surveys %V 41 %P 1 - 76 %8 2009/02/01/ %@ 03600300 %G eng %U http://dl.acm.org/citation.cfm?id=1459352.1459354 %R 10.1145/1459352.1459354 %0 Journal Article %J Computing in Science Engineering %D 2009 %T Using Graphics Processors for High-Performance Computation and Visualization of Plasma Turbulence %A Stantchev,G. %A Juba,D. %A Dorland,W. %A Varshney, Amitabh %K analysis;parallel %K computation;parallel %K computing;numerical %K direct %K engineering %K numerical %K PROCESSING %K processing;plasma %K processors;data %K simulation;graphics %K systems;nuclear %K turbulence %K turbulence; %K units;high-performance %K visualisation;multiprocessing %K visualization;single-program-multiple-data %X Direct numerical simulation (DNS) of turbulence is computationally intensive and typically relies on some form of parallel processing. The authors present techniques to map DNS computations to modern graphics processing units (GPUs), which are characterized by very high memory bandwidth and hundreds of SPMD (single-program-multiple-data) processors. %B Computing in Science Engineering %V 11 %P 52 - 59 %8 2009/04//march %@ 1521-9615 %G eng %N 2 %R 10.1109/MCSE.2009.42 %0 Conference Paper %B Proceedings of the 5th International Conference on Predictor Models in Software Engineering %D 2009 %T Using uncertainty as a model selection and comparison criterion %A Sarcia',Salvatore Alessandro %A Basili, Victor R. %A Cantone,Giovanni %K accuracy %K Bayesian prediction intervals %K Calibration %K cost estimation %K cost model %K model evaluation %K model selection %K prediction interval %K Uncertainty %X Over the last 25+ years, software estimation research has been searching for the best model for estimating variables of interest (e.g., cost, defects, and fault proneness). This research effort has not lead to a common agreement. One problem is that, they have been using accuracy as the basis for selection and comparison. But accuracy is not invariant; it depends on the test sample, the error measure, and the chosen error statistics (e.g., MMRE, PRED, Mean and Standard Deviation of error samples). Ideally, we would like an invariant criterion. In this paper, we show that uncertainty can be used as an invariant criterion to figure out which estimation model should be preferred over others. The majority of this work is empirically based, applying Bayesian prediction intervals to some COCOMO model variations with respect to a publicly available cost estimation data set coming from the PROMISE repository. %B Proceedings of the 5th International Conference on Predictor Models in Software Engineering %S PROMISE '09 %I ACM %C New York, NY, USA %P 18:1–18:9 - 18:1–18:9 %8 2009/// %@ 978-1-60558-634-2 %G eng %U http://doi.acm.org/10.1145/1540438.1540464 %R 10.1145/1540438.1540464 %0 Conference Paper %B Automated Software Engineering, International Conference on %D 2009 %T Validating Automotive Control Software Using Instrumentation-Based Verification %A Ray,Arnab %A Morschhaeuser,Iris %A Ackermann,Chris %A Cleaveland, Rance %A Shelton,Charles %A Martin,Chris %X This paper discusses the results of an application of a formally based verification technique, called Instrumentation-Based Verification (IBV), to a production automotive lighting controller. The goal of the study is to assess, from both a tools as well as a methodological perspective, the performance of IBV in an industrial setting. The insights obtained as a result of the project include a refinement of a previously developed architecture for requirements specifications; observations about changes to model-based design workflows; insights into the role of requirements during development; and the capability of automated verification to detect inconsistencies among requirements as well as between requirements and design models. %B Automated Software Engineering, International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %P 15 - 25 %8 2009/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/ASE.2009.98 %0 Conference Paper %B Visual Analytics Science and Technology, 2009. VAST 2009. IEEE Symposium on %D 2009 %T VAST 2009 challenge: An insider threat %A Grinstein,G. %A Scholtz,J. %A Whiting,M. %A Plaisant, Catherine %K 2009 %K analytics;data %K analytics;geospatial %K challenge;cyber %K data;security %K information %K information;human %K interaction;network %K interfaces; %K network;visual %K traffic %K user %K VAST %K video;social %K visualisation;graphical %X The 4th VAST Challenge centered on a cyber analytics scenario and offered three mini-challenges with datasets of badge and network traffic data, a social network including geospatial information, and security video. Teams could also enter the Grand challenge which combined all three datasets. In this paper, we summarize the dataset, the overall scenario and the questions asked in the challenges. We describe the judging process and new infrastructure developed to manage the submissions and compute accuracy measures in the social network mini challenge. We received 49 entries from 30 teams, and gave 23 different awards to a total of 16 teams. %B Visual Analytics Science and Technology, 2009. VAST 2009. IEEE Symposium on %P 243 - 244 %8 2009/10// %G eng %R 10.1109/VAST.2009.5334454 %0 Journal Article %J SIGPLAN Not. %D 2009 %T Verified enforcement of stateful information release policies %A Swamy,Nikhil %A Hicks, Michael W. %K affine types %K certified evaluation %K declassification %K dependent types %K singleton types %K state modifying policies %X Many organizations specify information release policies to describe the terms under which sensitive information may be released to other organizations. This paper presents a new approach for ensuring that security-critical software correctly enforces its information release policy. Our approach has two parts. First, an information release policy is specified as a security automaton written in a new language called AIR. Second, we enforce an AIR policy by translating it into an API for programs written in lAIR, a core formalism for a functional programming language. lAIR uses a novel combination of dependent, affine, and singleton types to ensure that the API is used correctly. As a consequence we can certify that programs written in lAIR meet the requirements of the original AIR policy specification. %B SIGPLAN Not. %V 43 %P 21 - 31 %8 2009/02// %@ 0362-1340 %G eng %U http://doi.acm.org/10.1145/1513443.1513448 %N 12 %R 10.1145/1513443.1513448 %0 Journal Article %J Image Analysis and Recognition %D 2009 %T Video Compression and Retrieval of Moving Object Location Applied to Surveillance %A Schwartz,W. %A Pedrini,H. %A Davis, Larry S. %X A major problem in surveillance systems is the storage requirements for video archival; videos are recorded continuously for long periods of time, resulting in large amounts of data. Therefore, it is essential to apply efficient compression techniques. Additionally, it is useful to be able to index the archived videos based on events. In general, such events are defined by the interaction among moving objects in the scene. Consequently, besides data compression, efficient ways of storing moving objects should be considered. We present a method that exploits both temporal and spatial redundancy of videos captured from static cameras to perform compression and subsequently allows fast retrieval of moving object locations directly from the compressed data. Experimental results show that the approach achieves high compression ratios compared to other existing video compression techniques without significant quality degradation and is fast due to the simplicity of the operations required for compression and decompression. %B Image Analysis and Recognition %P 906 - 916 %8 2009/// %G eng %0 Book %D 2009 %T Video-based Lane Detection using Boosting Principles %A Gopalan,R. %A Hong, T. %A Shneier, M. %A Chellapa, Rama %X Autonomous navigation of road vehicles is a challenging problem that has widespread applications in intelligent systems, and robotics. An integral componentof such a system is to understand how the road is structured. Detection of road lane markings assumes importance in this regard, and this problem has been approached with different visual input- based inference algorithms ([1], [2]), besides other sensing modalities such as the GPS sensor and the internal vehicle- state sensors. But the challenge still remains when there is considerable amount of shadows on the road, variations in outdoor lighting conditions of the scene (transition from day to night), among others. To address such issues, we propose a machine learning approach based on Real Adaboost [3], and train linear classifiers for both the appearance and edge cues of the training examplars. Additionally, we incorporate prior knowledge about the relative importance of the training samples by computing their weights using kernel discriminant analysis [4], before learning the classification function through boosting. The regions identified as lane markings are then analyzed for gradient direction consistency before making the final detection decision. We illustrate the effectiveness of our algorithm on challenging daylight and night- time road scenarios. %I Snowbird %8 2009/// %G eng %0 Journal Article %J Journal of the American Society for Information Science and Technology %D 2009 %T Visual overviews for discovering key papers and influences across research fronts %A Aris,Aleks %A Shneiderman, Ben %A Qazvinian,Vahed %A Radev,Dragomir %X Gaining a rapid overview of an emerging scientific topic, sometimes called research fronts, is an increasingly common task due to the growing amount of interdisciplinary collaboration. Visual overviews that show temporal patterns of paper publication and citation links among papers can help researchers and analysts to see the rate of growth of topics, identify key papers, and understand influences across subdisciplines. This article applies a novel network-visualization tool based on meaningful layouts of nodes to present research fronts and show citation links that indicate influences across research fronts. To demonstrate the value of two-dimensional layouts with multiple regions and user control of link visibility, we conducted a design-oriented, preliminary case study with 6 domain experts over a 4-month period. The main benefits were being able (a) to easily identify key papers and see the increasing number of papers within a research front, and (b) to quickly see the strength and direction of influence across related research fronts. %B Journal of the American Society for Information Science and Technology %V 60 %P 2219 - 2228 %8 2009/11/01/ %@ 1532-2890 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/asi.21160/abstract %N 11 %R 10.1002/asi.21160 %0 Journal Article %J Computer Graphics and Applications, IEEE %D 2009 %T Visual-Analytics Evaluation %A Plaisant, Catherine %A Grinstein,G. %A Scholtz,J. %K analytic %K analytics %K evaluation;cognition;data %K generation;user %K interactive %K interfaces;synthetic-data-set %K reasoning;visual %K systems; %K tools;visual %K visual %K visualisation;interactive %X Visual analytics (VA) is the science of analytical reasoning facilitated by interactive visual interfaces. Assessing VA technology's effectiveness is challenging because VA tools combine several disparate components, both low and high level, integrated in complex interactive systems used by analysts, emergency responders, and others. These components include analytical reasoning, visual representations, computer-human interaction techniques, data representations and transformations, collaboration tools, and especially tools for communicating the results of their use. VA tool users' activities can be exploratory and can take place over days, weeks, or months. Users might not follow a predefined or even linear work flow. They might work alone or in groups. To understand these complex behaviors, an evaluation can target the component level, the system level, or the work environment level, and requires realistic data and tasks. Traditional evaluation metrics such as task completion time, number of errors, or recall and precision are insufficient to quantify the utility of VA tools, and new research is needed to improve our VA evaluation methodology. %B Computer Graphics and Applications, IEEE %V 29 %P 16 - 17 %8 2009/06//may %@ 0272-1716 %G eng %N 3 %R 10.1109/MCG.2009.56 %0 Conference Paper %B Pattern Recognition, 2008. ICPR 2008. 19th International Conference on %D 2008 %T 3. Similarity searching: Indexing, nearest neighbor finding, dimensionality reduction, and embedding methods for applications in multimedia databases %A Samet, Hanan %K database;nearest %K databases;singular %K decomposition; %K DFT;FastMap %K EMBEDDING %K embedding;Lipschitz %K finding;similarity %K Fourier %K indexing;discrete %K indexing;multidimensional %K method;multimedia %K method;SVD;distance-based %K neighbor %K scaling %K searching;database %K transforms;multimedia %K value %X Similarity searching is usually achieved by means of nearest neighbor finding. Existing methods for handling similarity search in this setting fall into one of two classes. The first is based on mapping to a low-dimensional vector space which is then indexed using representations such as k-d trees, R-trees, quadtrees, etc. The second directly indexes the the objects based on distances using representations such as the vp-tree, M-tree, etc. Mapping from a high-dimensional space into a low-dimensional space is known as dimensionality reduction and is achieved using SVD, DFT, etc. At times, when we just have distance information, the data objects are embedded in a vector space so that the distances of the embedded objects as measured by the distance metric in the embedding space approximate the actual distance. The search in the embedding space uses conventional indexing methods which are often coupled with dimensionality reduction. Some commonly known embedding methods are multidimensional scaling, Lipschitz embeddings, and Fast Map. This tutorial is organized into five parts that cover the five basic concepts outlined above: indexing low and high dimensional spaces, distance-based indexing, dimensionality reduction, embedding methods, and nearest neighbor searching. %B Pattern Recognition, 2008. ICPR 2008. 19th International Conference on %P 1 - 1 %8 2008/12// %G eng %R 10.1109/ICPR.2008.4760937 %0 Journal Article %J SIGCOMM Comput. Commun. Rev. %D 2008 %T Accountable internet protocol (aip) %A Andersen,David G. %A Balakrishnan,Hari %A Feamster, Nick %A Koponen,Teemu %A Moon,Daekyeong %A Shenker,Scott %K accountability %K address %K internet architecture %K scalability %K Security %X This paper presents AIP (Accountable Internet Protocol), a network architecture that provides accountability as a first-order property. AIP uses a hierarchy of self-certifying addresses, in which each component is derived from the public key of the corresponding entity. We discuss how AIP enables simple solutions to source spoofing, denial-of-service, route hijacking, and route forgery. We also discuss how AIP's design meets the challenges of scaling, key management, and traffic engineering. %B SIGCOMM Comput. Commun. Rev. %V 38 %P 339 - 350 %8 2008/08// %@ 0146-4833 %G eng %U http://doi.acm.org/10.1145/1402946.1402997 %N 4 %R 10.1145/1402946.1402997 %0 Conference Paper %B In G. Visaggio (Ed.), 12 th International Conference on Evaluation and Assessment in Software Engineering. BCS eWIC %D 2008 %T Adopting Curvilinear Component Analysis to Improve Software Cost Estimation Accuracy. Model, Application Strategy, and an Experimental Verification %A Sarcia,Salvatore A. %A Cantone,Giovanni %A Basili, Victor R. %X Cost estimation is a critical issue for software organizations. Good estimates can help us make more informed decisions (controlling and planning software risks), if they are reliable (correct) and valid (stable). In this study, we apply a variable reduction technique (based on auto-associative feed--forward neural networks – called Curvilinear component analysis) to log-linear regression functions calibrated with ordinary least squares. Based on a COCOMO 81 data set, we show that Curvilinear component analysis can improve the estimation model accuracy by turning the initial input variables into an equivalent and more compact representation. We show that, the models obtained by applying Curvilinear component analysis are more parsimonious, correct, and reliable. %B In G. Visaggio (Ed.), 12 th International Conference on Evaluation and Assessment in Software Engineering. BCS eWIC %8 2008/// %G eng %0 Journal Article %J Lecture Notes in Computer Science %D 2008 %T Advances in multilingual and multimodal information retrieval %A Peters,C. %A Jijkoun,V. %A Mandl,T. %A Müller,H. %A Oard, Douglas %A Peñas,A. %A Santos,D. %B Lecture Notes in Computer Science %V 5152 %8 2008/// %G eng %0 Conference Paper %B Proceedings of the 23rd national conference on Artificial intelligence %D 2008 %T An AGM-based belief revision mechanism for probabilistic spatio-temporal logics %A Parker,A. %A Infantes,G. %A V.S. Subrahmanian %A Grant,J. %B Proceedings of the 23rd national conference on Artificial intelligence %P 511 - 516 %8 2008/// %G eng %0 Journal Article %J ACM Trans. Math. Softw. %D 2008 %T Algorithm 879: EIGENTEST—a test matrix generator for large-scale eigenproblems %A Lee,Che-Rung %A Stewart, G.W. %K Eigensystem %K test matrix generation %X Eigentest is a package that produces real test matrices with known eigensystems. A test matrix, called an eigenmat, is generated in a factored form, in which the user can specify the eigenvalues and has some control over the condition of the eigenvalues and eigenvectors. An eigenmat A of order n requires only O(n) storage for its representation. Auxiliary programs permit the computation of (A − sI)b, (A − sI)Tb, (A − sI)−1 b, and (A − sI)−T b in O(n) operations. A special routine computes specified eigenvectors of an eigenmat and the condition of its eigenvalue. Thus eigenmats are suitable for testing algorithms based on Krylov sequences, as well as others based on matrix-vector products. This article introduces the eigenmat and describes implementations in Fortran 77, Fortran 95, C, and Matlab. %B ACM Trans. Math. Softw. %V 35 %P 7:1–7:11 - 7:1–7:11 %8 2008/07// %@ 0098-3500 %G eng %U http://doi.acm.org/10.1145/1377603.1377610 %N 1 %R 10.1145/1377603.1377610 %0 Journal Article %J Image Processing, IEEE Transactions on %D 2008 %T Algorithmic and Architectural Optimizations for Computationally Efficient Particle Filtering %A Sankaranarayanan,A. C %A Srivastava, A. %A Chellapa, Rama %K Automated;Reproducibility of Results;Sensitivity and Specificity;Signal Processing %K Computer-Assisted;Models %K Computer-Assisted;Video Recording; %K convex program;independent Metropolis Hastings sampler;nonGaussian noise process;nonlinear dynamical system filtering;particle filtering algorithm;pipelined architectural optimization;video sequences;visual tracking;convex programming;image sequences;opti %K Statistical;Pattern Recognition %X In this paper, we analyze the computational challenges in implementing particle filtering, especially to video sequences. Particle filtering is a technique used for filtering nonlinear dynamical systems driven by non-Gaussian noise processes. It has found widespread applications in detection, navigation, and tracking problems. Although, in general, particle filtering methods yield improved results, it is difficult to achieve real time performance. In this paper, we analyze the computational drawbacks of traditional particle filtering algorithms, and present a method for implementing the particle filter using the Independent Metropolis Hastings sampler, that is highly amenable to pipelined implementations and parallelization. We analyze the implementations of the proposed algorithm, and, in particular, concentrate on implementations that have minimum processing times. It is shown that the design parameters for the fastest implementation can be chosen by solving a set of convex programs. The proposed computational methodology was verified using a cluster of PCs for the application of visual tracking. We demonstrate a linear speedup of the algorithm using the methodology proposed in the paper. %B Image Processing, IEEE Transactions on %V 17 %P 737 - 748 %8 2008/05// %@ 1057-7149 %G eng %N 5 %R 10.1109/TIP.2008.920760 %0 Conference Paper %B Proceedings of the twenty-sixth annual SIGCHI conference on Human factors in computing systems %D 2008 %T Aligning temporal data by sentinel events: discovering patterns in electronic health records %A Wang,Taowei David %A Plaisant, Catherine %A Quinn,Alexander J. %A Stanchak,Roman %A Murphy,Shawn %A Shneiderman, Ben %K electronic health record %K Evaluation %K Information Visualization %K search %K temporal data %K Uncertainty %X Electronic Health Records (EHRs) and other temporal databases contain hidden patterns that reveal important cause-and-effect phenomena. Finding these patterns is a challenge when using traditional query languages and tabular displays. We present an interactive visual tool that complements query formulation by providing operations to align, rank and filter the results, and to visualize estimates of the intervals of validity of the data. Display of patient histories aligned on sentinel events (such as a first heart attack) enables users to spot precursor, co-occurring, and aftereffect events. A controlled study demonstrates the benefits of providing alignment (with a 61% speed improvement for complex tasks). A qualitative study and interviews with medical professionals demonstrates that the interface can be learned quickly and seems to address their needs. %B Proceedings of the twenty-sixth annual SIGCHI conference on Human factors in computing systems %S CHI '08 %I ACM %C New York, NY, USA %P 457 - 466 %8 2008/// %@ 978-1-60558-011-1 %G eng %U http://doi.acm.org/10.1145/1357054.1357129 %R 10.1145/1357054.1357129 %0 Journal Article %J NSF Symposium on Semantic Knowledge Discovery, Organization and Use %D 2008 %T Applying automatically generated semantic knowledge: A case study in machine translation %A Madnani,N. %A Resnik, Philip %A Dorr, Bonnie J %A Schwartz,R. %X In this paper, we discuss how we apply automatically generated semantic knowledge to benefit statisticalmachine translation (SMT). Currently, almost all statistical machine translation systems rely heavily on memorizing translations of phrases. Some systems attempt to go further and generalize these learned phrase translations into templates using empirically derived information about word alignments and a small amount of syntactic information, if at all. There are several issues in a SMT pipeline that could be addressed by the application of semantic knowledge, if such knowledge were easily available. One such issue, an important one, is that of reference sparsity. The fundamental problem that translation systems have to face is that there is no such thing as the correct translation for any sentence. In fact, any given source sentence can often be translated into the target language in many valid ways. Since there can be many “correct answers,” almost all models employed by SMT systems require, in addition to a large bitext, a held-out development set comprised of multiple high-quality, human-authored reference translations in the target language in order to tune their parameters relative to a translation quality metric.1 There are several reasons that this requirement is not an easy one to satisfy. First, with a few exceptions—notably NIST’s annual MT evaluations—most new MT research data sets are provided with only a single reference translation. Second, obtaining multiple reference translations in rapid development, low-density source language scenarios (e.g. (Oard, 2003)) is likely to be severely limited (or made entirely impractical) by limitations of time, cost, and ready availability of qualified translators. %B NSF Symposium on Semantic Knowledge Discovery, Organization and Use %8 2008/// %G eng %0 Conference Paper %B Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on %D 2008 %T Approximate earth mover's distance in linear time %A Shirdhonkar,S. %A Jacobs, David W. %K algorithm;normal %K complexity;earth %K complexity;image %K constraint;Kantorovich-Rubinstein %K continuity %K distance;histograms;linear %K distance;weighted %K Euclidean %K Holder %K matching;wavelet %K movers %K problem;computational %K TIME %K transform;computational %K transforms; %K transshipment %K wavelet %X The earth moverpsilas distance (EMD) is an important perceptually meaningful metric for comparing histograms, but it suffers from high (O(N3 logN)) computational complexity. We present a novel linear time algorithm for approximating the EMD for low dimensional histograms using the sum of absolute values of the weighted wavelet coefficients of the difference histogram. EMD computation is a special case of the Kantorovich-Rubinstein transshipment problem, and we exploit the Holder continuity constraint in its dual form to convert it into a simple optimization problem with an explicit solution in the wavelet domain. We prove that the resulting wavelet EMD metric is equivalent to EMD, i.e. the ratio of the two is bounded. We also provide estimates for the bounds. The weighted wavelet transform can be computed in time linear in the number of histogram bins, while the comparison is about as fast as for normal Euclidean distance or chi2 statistic. We experimentally show that wavelet EMD is a good approximation to EMD, has similar performance, but requires much less computation. %B Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on %P 1 - 8 %8 2008/06// %G eng %R 10.1109/CVPR.2008.4587662 %0 Conference Paper %B IEEE INFOCOM 2008. The 27th Conference on Computer Communications %D 2008 %T Approximation Algorithms for Computing Capacity of Wireless Networks with SINR Constraints %A Chafekar,D. %A Kumart,V. S.A %A Marathe,M. V %A Parthasarathy,S. %A Srinivasan, Aravind %K Algorithm design and analysis %K approximation algorithm %K Approximation algorithms %K approximation theory %K Computer networks %K Computer science %K graph theory %K graph-based model %K Interference constraints %K polynomial time algorithm %K Propagation losses %K Protocols %K radio networks %K radiofrequency interference %K signal to interference plus noise ratio %K Signal to noise ratio %K Throughput %K wireless interference %K wireless network %K Wireless networks %X A fundamental problem in wireless networks is to estimate its throughput capacity - given a set of wireless nodes, and a set of connections, what is the maximum rate at which data can be sent on these connections. Most of the research in this direction has focused on either random distributions of points, or has assumed simple graph-based models for wireless interference. In this paper, we study capacity estimation problem using the more general Signal to Interference Plus Noise Ratio (SINR) model for interference, on arbitrary wireless networks. The problem becomes much harder in this setting, because of the non-locality of the SINR model. Recent work by Moscibroda et al. (2006) has shown that the throughput in this model can differ from graph based models significantly. We develop polynomial time algorithms to provably approximate the total throughput in this setting. %B IEEE INFOCOM 2008. The 27th Conference on Computer Communications %I IEEE %P 1166 - 1174 %8 2008/04/13/18 %@ 978-1-4244-2025-4 %G eng %R 10.1109/INFOCOM.2008.172 %0 Book Section %B Computational Linguistics and Intelligent Text Processing, Lecture Notes in Computer Science Volume 4919Computational Linguistics and Intelligent Text Processing, Lecture Notes in Computer Science Volume 4919 %D 2008 %T Arabic/English Multi-document Summarization with CLASSYThe Past and the Future %A Schlesinger,Judith D. %A O'Leary, Dianne P. %A Conroy,John M. %B Computational Linguistics and Intelligent Text Processing, Lecture Notes in Computer Science Volume 4919Computational Linguistics and Intelligent Text Processing, Lecture Notes in Computer Science Volume 4919 %I Springer Berlin %P 568 - 581 %8 2008/// %G eng %U http://dx.doi.org/10.1007/978-3-540-78135-6_49http://dx.doi.org/10.1007/978-3-540-78135-6_49 %0 Report %D 2008 %T Archiving Temporal Web Information: Organization of Web Contents for Fast Access and Compact Storage %A Song,Sangchul %A JaJa, Joseph F. %K Technical Report %X We address the problem of archiving dynamic web contents oversignificant time spans. Current schemes crawl the web contents at regular time intervals and archive the contents after each crawl regardless of whether or not the contents have changed between consecutive crawls. Our goal is to store newly crawled web contents only when they are different than the previous crawl, while ensuring accurate and quick retrieval of archived contents based on arbitrary temporal queries over the archived time period. In this paper, we develop a scheme that stores unique temporal web contents in containers following the widely used ARC/WARC format, and that provides quick access to the archived contents for arbitrary temporal queries. A novel component of our scheme is the use of a new indexing structure based on the concept of persistent or multi-version data structures. Our scheme can be shown to be asymptotically optimal both in storage utilization and insert/retrieval time. We illustrate the performance of our method on two very different data sets from the Stanford WebBase project, the first reflecting very dynamic web contents and the second relatively static web contents. The experimental results clearly illustrate the substantial storage savings achieved by eliminating duplicate contents detected between consecutive crawls, as well as the speed at which our method can find the archived contents specified through arbitrary temporal queries. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2008-08 %8 2008/04/07/ %G eng %U http://drum.lib.umd.edu/handle/1903/7569 %0 Journal Article %J Proceedings of the Eighth Conference of the Association for Machine Translation in the Americas, October %D 2008 %T Are multiple reference translations necessary? investigating the value of paraphrased reference translations in parameter optimization %A Madnani,N. %A Resnik, Philip %A Dorr, Bonnie J %A Schwartz,R. %X Most state-of-the-art statistical machine trans-lation systems use log-linear models, which are defined in terms of hypothesis features and weights for those features. It is standard to tune the feature weights in order to maxi- mize a translation quality metric, using held- out test sentences and their corresponding ref- erence translations. However, obtaining refer- ence translations is expensive. In our earlier work (Madnani et al., 2007), we introduced a new full-sentence paraphrase technique, based on English-to-English decoding with an MT system, and demonstrated that the resulting paraphrases can be used to cut the number of human reference translations needed in half. In this paper, we take the idea a step further, asking how far it is possible to get with just a single good reference translation for each item in the development set. Our analysis suggests that it is necessary to invest in four or more hu- man translations in order to significantly im- prove on a single translation augmented by monolingual paraphrases. %B Proceedings of the Eighth Conference of the Association for Machine Translation in the Americas, October %8 2008/// %G eng %0 Conference Paper %B Data Engineering Workshop, 2008. ICDEW 2008. IEEE 24th International Conference on %D 2008 %T Augmenting spatio-textual search with an infectious disease ontology %A Lieberman,M.D. %A Sankaranarayanan,J. %A Samet, Hanan %A Sperling,J. %K (artificial %K article;ontology;spatio-textual %K disease;map %K engines; %K information %K intelligence);search %K interface;newspaper %K search;classification;diseases;indexing;medical %K STEWARD %K system;classification;epidemics;indexing;infectious %K systems;ontologies %X A system is described that automatically categorizes and classifies infectious disease incidence reports by type and geographic location, to aid analysis by domain experts. It identifies references to infectious diseases by using a disease ontology. The system leverages the textual and spatial search capabilities of the STEWARD system to enable queries such as reports on "influenza" near "Hong Kong", possibly within a particular time period. Documents from the U.S. National Library of Medicine (http://www.pubmed.gov) and the World Health Organization (http://www.who.int) are tagged so that spatial relationships to specific disease occurrences can be presented graphically via a map interface. In addition, newspaper articles can be tagged and indexed to bolster the surveillance of ongoing epidemics. Examining past epidemics using this system may lead to improved understanding of the cause and spread of infectious diseases. %B Data Engineering Workshop, 2008. ICDEW 2008. IEEE 24th International Conference on %P 266 - 269 %8 2008/04// %G eng %R 10.1109/ICDEW.2008.4498330 %0 Journal Article %J Advances in Neural Information Processing Systems %D 2008 %T Automatic online tuning for fast Gaussian summation %A Morariu,V. %A Srinivasan,B.V. %A Raykar,V.C. %A Duraiswami, Ramani %A Davis, Larry S. %X Many machine learning algorithms require the summation of Gaussian kernelfunctions, an expensive operation if implemented straightforwardly. Several meth- ods have been proposed to reduce the computational complexity of evaluating such sums, including tree and analysis based methods. These achieve varying speedups depending on the bandwidth, dimension, and prescribed error, making the choice between methods difficult for machine learning tasks. We provide an algorithm that combines tree methods with the Improved Fast Gauss Transform (IFGT). As originally proposed the IFGT suffers from two problems: (1) the Taylor series expansion does not perform well for very low bandwidths, and (2) parameter se- lection is not trivial and can drastically affect performance and ease of use. We address the first problem by employing a tree data structure, resulting in four eval- uation methods whose performance varies based on the distribution of sources and targets and input parameters such as desired accuracy and bandwidth. To solve the second problem, we present an online tuning approach that results in a black box method that automatically chooses the evaluation method and its parameters to yield the best performance for the input data, desired accuracy, and bandwidth. In addition, the new IFGT parameter selection approach allows for tighter error bounds. Our approach chooses the fastest method at negligible additional cost, and has superior performance in comparisons with previous approaches. %B Advances in Neural Information Processing Systems %8 2008/// %G eng %0 Journal Article %J Intelligent Systems, IEEE %D 2008 %T AVA: Adjective-Verb-Adverb Combinations for Sentiment Analysis %A V.S. Subrahmanian %A Reforgiato,Diego %X Most research on determining the strength of subjective expressions in a sentence or document uses single, specific parts of speech such as adjectives, adverbs, or verbs. To date, almost no research covers the development of a single comprehensive framework in which we can analyze sentiment that takes all three into account. The authors propose the AVA (adjective verb adverb) framework for identifying opinions on any given topic. In AVA, a user can select any topic t of interest and any document d. AVA will return a score that d expresses topic t. The score is expressed on a #x02013;1 (maximally negative) to +1 (maximally positive) scale. %B Intelligent Systems, IEEE %V 23 %P 43 - 50 %8 2008/08//july %@ 1541-1672 %G eng %N 4 %R 10.1109/MIS.2008.57 %0 Journal Article %J The Journal of chemical physics %D 2008 %T A Bayesian statistics approach to multiscale coarse graining %A Liu,P. %A Shi,Q. %A Daumé, Hal %A Voth,G. A %B The Journal of chemical physics %V 129 %P 214114 - 214114 %8 2008/// %G eng %0 Conference Paper %B CHI '08 extended abstracts on Human factors in computing systems %D 2008 %T BELIV'08: Beyond time and errors: novel evaluation methods for information visualization %A Bertini,Enrico %A Perer,Adam %A Plaisant, Catherine %A Santucci,Giuseppe %K Evaluation %K Information Visualization %X Information visualization systems allow users to produce insights, innovations and discoveries. Evaluating such tools is a challenging task and the goal of BELIV'08 is to make a step ahead in the comprehension of such a complex activity. Current evaluation methods exhibit noticeable limitations and researchers in the area experiment some frustration with evaluation processes that are time consuming and too often leading to unsatisfactory results. The most used evaluation metrics such as task time completion and number of errors appear insufficient to quantify the quality of an information visualization system; thus the name of the workshop: "beyond time and errors". %B CHI '08 extended abstracts on Human factors in computing systems %S CHI EA '08 %I ACM %C New York, NY, USA %P 3913 - 3916 %8 2008/// %@ 978-1-60558-012-8 %G eng %U http://doi.acm.org/10.1145/1358628.1358955 %R 10.1145/1358628.1358955 %0 Journal Article %J Trends in Genetics %D 2008 %T Bioinformatics challenges of new sequencing technology %A Pop, Mihai %A Salzberg,Steven L. %X New DNA sequencing technologies can sequence up to one billion bases in a single day at low cost, putting large-scale sequencing within the reach of many scientists. Many researchers are forging ahead with projects to sequence a range of species using the new technologies. However, these new technologies produce read lengths as short as 35–40 nucleotides, posing challenges for genome assembly and annotation. Here we review the challenges and describe some of the bioinformatics systems that are being proposed to solve them. We specifically address issues arising from using these technologies in assembly projects, both de novo and for resequencing purposes, as well as efforts to improve genome annotation in the fragmented assemblies produced by short read lengths. %B Trends in Genetics %V 24 %P 142 - 149 %8 2008/03// %@ 0168-9525 %G eng %U http://www.sciencedirect.com/science/article/pii/S016895250800022X %N 3 %R 10.1016/j.tig.2007.12.006 %0 Conference Paper %B Proceedings of the ACM SIGCOMM 2008 conference on Data communication %D 2008 %T Bittorrent is an auction: analyzing and improving bittorrent's incentives %A Levin,Dave %A LaCurts,Katrina %A Spring, Neil %A Bhattacharjee, Bobby %K auctions %K bittorrent %K incentive systems %K proportional share %K tit-for-tat %X Incentives play a crucial role in BitTorrent, motivating users to upload to others to achieve fast download times for all peers. Though long believed to be robust to strategic manipulation, recent work has empirically shown that BitTorrent does not provide its users incentive to follow the protocol. We propose an auction-based model to study and improve upon BitTorrent's incentives. The insight behind our model is that BitTorrent uses, not tit-for-tat as widely believed, but an auction to decide which peers to serve. Our model not only captures known, performance-improving strategies, it shapes our thinking toward new, effective strategies. For example, our analysis demonstrates, counter-intuitively, that BitTorrent peers have incentive to intelligently under-report what pieces of the file they have to their neighbors. We implement and evaluate a modification to BitTorrent in which peers reward one another with proportional shares of bandwidth. Within our game-theoretic model, we prove that a proportional-share client is strategy-proof. With experiments on PlanetLab, a local cluster, and live downloads, we show that a proportional-share unchoker yields faster downloads against BitTorrent and BitTyrant clients, and that under-reporting pieces yields prolonged neighbor interest. %B Proceedings of the ACM SIGCOMM 2008 conference on Data communication %S SIGCOMM '08 %I ACM %C New York, NY, USA %P 243 - 254 %8 2008/// %@ 978-1-60558-175-0 %G eng %U http://doi.acm.org/10.1145/1402958.1402987 %R 10.1145/1402958.1402987 %0 Journal Article %J SIAM Journal on Scientific Computing %D 2008 %T Block Gram–Schmidt Orthogonalization %A Stewart, G.W. %K blocked algorithm %K GramSchmidt algorithm %K orthogonalization %K QR factorization %X The classical Gram–Schmidt algorithm for computing the QR factorization of a matrix $X$ requires at least one pass over the current orthogonalized matrix $Q$ as each column of $X$ is added to the factorization. When $Q$ becomes so large that it must be maintained on a backing store, each pass involves the costly transfer of data from the backing store to main memory. However, if one orthogonalizes the columns of $X$ in blocks of $m$ columns, the number of passes is reduced by a factor of $1/m$. Moreover, matrix-vector products are converted into matrix-matrix products, allowing level-3 BLAS cache performance. In this paper we derive such a block algorithm and give some experimental results that suggest it can be quite effective for large scale problems, even when the matrix $X$ is rank degenerate. %B SIAM Journal on Scientific Computing %V 31 %P 761 - 775 %8 2008/// %G eng %U http://link.aip.org/link/?SCE/31/761/1 %N 1 %R 10.1137/070682563 %0 Book Section %B Approximation, Randomization and Combinatorial Optimization. Algorithms and TechniquesApproximation, Randomization and Combinatorial Optimization. Algorithms and Techniques %D 2008 %T Budgeted Allocations in the Full-Information Setting %A Srinivasan, Aravind %E Goel,Ashish %E Jansen,Klaus %E Rolim,José %E Rubinfeld,Ronitt %X We build on the work of Andelman & Mansour and Azar, Birnbaum, Karlin, Mathieu & Thach Nguyen to show that the full-information (i.e., offline) budgeted-allocation problem can be approximated to within 4/3: we conduct a rounding of the natural LP relaxation, for which our algorithm matches the known lower-bound on the integrality gap. %B Approximation, Randomization and Combinatorial Optimization. Algorithms and TechniquesApproximation, Randomization and Combinatorial Optimization. Algorithms and Techniques %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5171 %P 247 - 253 %8 2008/// %@ 978-3-540-85362-6 %G eng %U http://dx.doi.org/10.1007/978-3-540-85363-3_20 %0 Conference Paper %B IEEE INFOCOM 2008. The 27th Conference on Computer Communications %D 2008 %T Capacity of Asynchronous Random-Access Scheduling in Wireless Networks %A Chafekar,D. %A Levin,D. %A Kumar,V. S.A %A Marathe,M. V %A Parthasarathy,S. %A Srinivasan, Aravind %K asynchronous random-access scheduling %K channel access probability %K Computer networks %K Computer science %K Educational institutions %K Interference %K Optimal scheduling %K Peer to peer computing %K probability %K Processor scheduling %K radio link %K radio links %K radio networks %K Routing %K scheduling %K Throughput %K throughput capacity %K wireless channels %K Wireless networks %X We study the throughput capacity of wireless networks which employ (asynchronous) random-access scheduling as opposed to deterministic scheduling. The central question we answer is: how should we set the channel-access probability for each link in the network so that the network operates close to its optimal throughput capacity? We design simple and distributed channel-access strategies for random-access networks which are provably competitive with respect to the optimal scheduling strategy, which is deterministic, centralized, and computationally infeasible. We show that the competitiveness of our strategies are nearly the best achievable via random-access scheduling, thus establishing fundamental limits on the performance of random- access. A notable outcome of our work is that random access compares well with deterministic scheduling when link transmission durations differ by small factors, and much worse otherwise. The distinguishing aspects of our work include modeling and rigorous analysis of asynchronous communication, asymmetry in link transmission durations, and hidden terminals under arbitrary link-conflict based wireless interference models. %B IEEE INFOCOM 2008. The 27th Conference on Computer Communications %I IEEE %P 1148 - 1156 %8 2008/04/13/18 %@ 978-1-4244-2025-4 %G eng %R 10.1109/INFOCOM.2008.170 %0 Conference Paper %B Engineering of Computer-Based Systems, IEEE International Conference on the %D 2008 %T CiCUTS: Combining System Execution Modeling Tools with Continuous Integration Environments %A Hill,James H. %A Schmidt,Douglas C. %A Porter, Adam %A Slaby,John M. %K continuous integration %K integration testing %K serialized phasing %K system execution modeling %X System execution modeling (SEM) tools provide an effective means to evaluate the quality of service (QoS) of enterprise distributed real-time and embedded (DRE) systems. SEM tools facilitate testing and resolving performance issues throughout the entire development life-cycle, rather than waiting until final system integration. SEM tools have not historically focused on effective testing. New techniques are therefore needed to help bridge the gap between the early integration capabilities of SEM tools and testing so developers can focus on resolving strategic integration and performance issues, as opposed to wrestling with tedious and error-prone low-level testing concerns. This paper provides two contributions to research on using SEM tools to address enterprise DRE system integration challenges. First, we evaluate several approaches for combining continuous integration environments with SEM tools and describe CiCUTS, which combines the CUTS SEM tool with the CruiseControl .NET continuous integration environment. Second, we present a case study that shows how CiCUTS helps reduce the time and effort required to manage and execute integration tests that evaluate QoS metrics for a representative DRE system from the domain of shipboard computing. The results of our case study show that CiCUTS helps developers and testers ensure the performance of an example enterprise DRE system is within its QoS specifications throughout development, instead of waiting until system integration time to evaluate QoS. %B Engineering of Computer-Based Systems, IEEE International Conference on the %I IEEE Computer Society %C Los Alamitos, CA, USA %P 66 - 75 %8 2008/// %@ 978-0-7695-3141-0 %G eng %R http://doi.ieeecomputersociety.org/10.1109/ECBS.2008.20 %0 Journal Article %J AI Magazine %D 2008 %T Collective Classification in Network Data %A Sen,Prithviraj %A Namata,Galileo %A Bilgic,Mustafa %A Getoor, Lise %A Galligher,Brian %A Eliassi-Rad,Tina %X Collective Classification in Network Data %B AI Magazine %V 29 %P 93 - 93 %8 2008/09/06/ %@ 0738-4602 %G eng %U http://www.aaai.org/ojs/index.php/aimagazine/article/viewArticle/2157 %N 3 %R 10.1609/aimag.v29i3.2157 %0 Book Section %B Studies in Computational Intelligence: Machine Learning in Document Analysis and RecognitionStudies in Computational Intelligence: Machine Learning in Document Analysis and Recognition %D 2008 %T Combining Classifiers with Informational Confidence %A Jaeger,Stefan %A Ma,Huanfeng %A David Doermann %E Simone Marinai,Hiromichi Fujisawa %X We propose a new statistical method for learning normalized confidence values in multiple classifier systems. Our main idea is to adjust confidence values so that their nominal values equal the information actually conveyed. In order to do so, we assume that information depends on the actual performance of each confidence value on an evaluation set. As information measure, we use Shannon's well-known logarithmic notion of information. With the confidence values matching their informational content, the classifier combination scheme reduces to the simple sum-rule, theoretically justifying this elementary combination scheme. In experimental evaluations for script identification, and both handwritten and printed character recognition, we achieve a consistent improvement on the best single recognition rate. We cherish the hope that our information-theoretical framework helps fill the theoretical gap we still experience in classifier combination, and puts the excellent practical performance of multiple classifier systems on a more solid basis. %B Studies in Computational Intelligence: Machine Learning in Document Analysis and RecognitionStudies in Computational Intelligence: Machine Learning in Document Analysis and Recognition %I Springer %P 163 - 192 %8 2008/// %G eng %0 Conference Paper %D 2008 %T Combining Domain-Independent Planning and HTN Planning: The Duet Planner %A Gerevini,Alfonso %A Kuter,Ugur %A Nau, Dana S. %A Saetti,Alessandro %A Waisbrot,Nathaniel %X Despite the recent advances in planning for classical domains, the question of how to use domain knowledge in planning is yet to be completely and clearly answered. Some of the existing planners use domain-independent search heuristics, and some others depend on intensively-engineered domain-specific knowledge to guide the planning process. In this paper, we describe an approach to combine ideas from both of the above schools of thought. We present Duet, our planning system that incorporates the ability of using hierarchical domain knowledge in the form of Hierarchical Task Networks (HTNs) as in SHOP2 [14] and using domain-independent local search techniques as in LPG [8]. In our experiments, Duet was able to solve much larger problems than LPG could solve, with only minimal domain knowledge encoded in HTNs (much less domain knowledge than SHOP2 needed to solve those problems by itself). %I IOS Press %C Amsterdam, The Netherlands, The Netherlands %P 573 - 577 %8 2008/// %@ 978-1-58603-891-5 %G eng %U http://dl.acm.org/citation.cfm?id=1567281.1567406 %0 Journal Article %J Proceedings of the 5th International ISCRAM Conference %D 2008 %T Community response grid (CRG) for a university campus: Design requirements and implications %A Wu,P.F. %A Qu,Y. %A Preece,J. %A Fleischmann,K. %A Golbeck,J. %A Jaeger,P. %A Shneiderman, Ben %X This paper describes the initial stages of the participatory design of a community-oriented emergency responsesystem for a university campus. After reviewing related work and the current University emergency response system, this paper describes our participatory design process, discusses initial findings from a design requirement survey and from our interactions with different stakeholders, and proposes a Web interface design for a community response grid system. The prototyping of the system demonstrates the possibility of fostering a social-network-based community participation in emergency response, and also identifies concerns raised by potential users and by the professional responder community. %B Proceedings of the 5th International ISCRAM Conference %P 34 - 43 %8 2008/// %G eng %0 Conference Paper %D 2008 %T On the Comparison of Network Attack Datasets: An Empirical Analysis %A Berthier,R. %A Korman,D. %A Michel Cukier %A Hiltunen,M. %A Vesonder,G. %A Sheleheda,D. %K ATLAS %K distributed network telescope %K Internet %K intrusion detection systems %K network attack datasets %K network malicious activity %K network security operators %K security of data %X Network malicious activity can be collected and reported by various sources using different attack detection solutions. The granularity of these solutions provides either very detailed information (intrusion detection systems, honeypots) or high-level trends (CAIDA, SANS). The problem for network security operators is often to select the sources of information to better protect their network. How much information from these sources is redundant and how much is unique? The goal of this paper is to show empirically that while some global attack events can be correlated across various sensors, the majority of incoming malicious activity has local specificities. This study presents a comparative analysis of four different attack datasets offering three different levels of granularity: 1) two high interaction honeynets deployed at two different locations (i.e., a corporate and an academic environment); 2) ATLAS which is a distributed network telescope from Arbor; and 3) Internet Protecttrade which is a global alerting service from AT amp;T. %P 39 - 48 %8 2008/12// %G eng %R 10.1109/HASE.2008.50 %0 Conference Paper %B International Conference on Computer Graphics and Interactive Techniques %D 2008 %T Complexity and Accessibility: Sorting in space: multidimensional, spatial, and metric data structures for CG applications %A Samet, Hanan %B International Conference on Computer Graphics and Interactive Techniques %8 2008/// %G eng %0 Conference Paper %B Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on %D 2008 %T Compressed sensing for multi-view tracking and 3-D voxel reconstruction %A Reddy, D. %A Sankaranarayanan,A. C %A Cevher, V. %A Chellapa, Rama %K 3D %K background-subtracted %K coding; %K Estimation %K image %K problems;multi-view %K projections;silhouette %K reconstruction;CS %K reconstruction;video %K sensing;multi-view %K silhouettes;image %K sparsity;sparse %K theory;compressed %K tracking;random %K voxel %X Compressed sensing (CS) suggests that a signal, sparse in some basis, can be recovered from a small number of random projections. In this paper, we apply the CS theory on sparse background-subtracted silhouettes and show the usefulness of such an approach in various multi-view estimation problems. The sparsity of the silhouette images corresponds to sparsity of object parameters (location, volume etc.) in the scene. We use random projections (compressed measurements) of the silhouette images for directly recovering object parameters in the scene coordinates. To keep the computational requirements of this recovery procedure reasonable, we tessellate the scene into a bunch of non-overlapping lines and perform estimation on each of these lines. Our method is scalable in the number of cameras and utilizes very few measurements for transmission among cameras. We illustrate the usefulness of our approach for multi-view tracking and 3-D voxel reconstruction problems. %B Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on %P 221 - 224 %8 2008/10// %G eng %R 10.1109/ICIP.2008.4711731 %0 Journal Article %J Computer Vision–ECCV 2008 %D 2008 %T Compressive sensing for background subtraction %A Cevher, V. %A Sankaranarayanan, A. %A Duarte, M. %A Reddy, D. %A Baraniuk, R. %A Chellapa, Rama %X Compressive sensing (CS) is an emerging field that provides a framework for image recovery using sub-Nyquist sampling rates. The CS theory shows that a signal can be reconstructed from a small set of random projections, provided that the signal is sparse in some basis, e.g., wavelets. In this paper, we describe a method to directly recover background subtracted images using CS and discuss its applications in some communication constrained multi-camera computer vision problems. We show how to apply the CS theory to recover object silhouettes (binary background subtracted images) when the objects of interest occupy a small portion of the camera view, i.e., when they are sparse in the spatial domain. We cast the background subtraction as a sparse approximation problem and provide different solutions based on convex optimization and total variation. In our method, as opposed to learning the background, we learn and adapt a low dimensional compressed representation of it, which is sufficient to determine spatial innovations; object silhouettes are then estimated directly using the compressive samples without any auxiliary image reconstruction. We also discuss simultaneous appearance recovery of the objects using compressive measurements. In this case, we show that it may be necessary to reconstruct one auxiliary image. To demonstrate the performance of the proposed algorithm, we provide results on data captured using a compressive single-pixel camera. We also illustrate that our approach is suitable for image coding in communication constrained problems by using data captured by multiple conventional cameras to provide 2D tracking and 3D shape reconstruction results with compressive measurements. %B Computer Vision–ECCV 2008 %P 155 - 168 %8 2008/// %G eng %0 Journal Article %J PLoS ONEPLoS ONE %D 2008 %T Computational Analysis of Constraints on Noncoding Regions, Coding Regions and Gene Expression in Relation to Plasmodium Phenotypic Diversity %A Essien,Kobby %A Hannenhalli, Sridhar %A Stoeckert,Christian J. %X Malaria-causing Plasmodium species exhibit marked differences including host choice and preference for invading particular cell types. The genetic bases of phenotypic differences between parasites can be understood, in part, by investigating constraints on gene expression and genic sequences, both coding and regulatory.We investigated the evolutionary constraints on sequence and expression of parasitic genes by applying comparative genomics approaches to 6 Plasmodium genomes and 2 genome-wide expression studies. We found that the coding regions of Plasmodium transcription factor and sexual development genes are relatively less constrained, as are those of genes encoding CCCH zinc fingers and invasion proteins, which all play important roles in these parasites. Transcription factors and genes with stage-restricted expression have conserved upstream regions and so do several gene classes critical to the parasite's lifestyle, namely, ion transport, invasion, chromatin assembly and CCCH zinc fingers. Additionally, a cross-species comparison of expression patterns revealed that Plasmodium-specific genes exhibit significant expression divergence. Overall, constraints on Plasmodium's protein coding regions confirm observations from other eukaryotes in that transcription factors are under relatively lower constraint. Proteins relevant to the parasite's unique lifestyle also have lower constraint on their coding regions. Greater conservation between Plasmodium species in terms of promoter motifs suggests tight regulatory control of lifestyle genes. However, an interspecies divergence in expression patterns of these genes suggests that either expression is controlled via genomic or epigenomic features not encoded in the proximal promoter sequence, or alternatively, the combinatorial interactions between motifs confer species-specific expression patterns. %B PLoS ONEPLoS ONE %V 3 %P e3122 - e3122 %8 2008/// %G eng %U UR - http://dx.doi.org/10.1371/journal.pone.0003122,http://dx.doi.org/10.1371/journal.pone.0003122 %N 9 %R 10.1371/journal.pone.0003122 %0 Journal Article %J IEEE intelligent systems %D 2008 %T Computational cultural dynamics %A Martinez,V. %A Simari,G. I %A Sliva,A. %A V.S. Subrahmanian %B IEEE intelligent systems %V 23 %P 51 - 57 %8 2008/// %G eng %N 4 %0 Journal Article %J Understanding Events %D 2008 %T Computational Vision Approaches for Event Modeling %A Chellapa, Rama %A Cuntoor, N.P. %A Joo, S.W. %A V.S. Subrahmanian %A Turaga,P. %X Event modeling systems provide a semantic interpretation of sequences of pixels that are captured by a video camera. The design of a practical system has to take into account the following three main factors: low-level preprocessing limitations, computational and storage complexity of the event model, and user interaction. The hidden Markov model (HMM) and its variants have been widely used to model both speech and video signals. Computational efficiency of the Baum-Welch and the Viterbi algorithms has been a leading reason for the popularity of the HMM. Since the objective is to detect events in video sequences that are meaningful to humans, one might want to provide space in the design loop for a user who can specify events of interest. This chapter explores this using semantic approaches that not only use features extracted from raw video streams but also incorporate metadata and ontologies of activities. It presents three approaches for applications such as event recognition: anomaly detection, temporal segmentation, and ontology evaluation. The three approaches discussed are statistical methods based on HMMs, formal grammars, and ontologies. The effectiveness of these approaches is illustrated using video sequences captured both indoors and outdoors: the indoor UCF human action dataset, the TSA airport tarmac surveillance dataset, and the bank monitoring dataset. %B Understanding Events %V 1 %P 473 - 522 %8 2008/// %G eng %N 9 %0 Journal Article %J Plasma Science, IEEE Transactions on %D 2008 %T Confluent Volumetric Visualization of Gyrokinetic Turbulence %A Stantchev,G. %A Juba,D. %A Dorland,W. %A Varshney, Amitabh %K flow;plasma %K geometry;plasma %K gyrokinetic %K simulation;nontrivial %K simulation;plasma %K turbulence; %K turbulence;nonlinear %K turbulence;volumetric %K visualisation;plasma %K visualization;flow %X Data from gyrokinetic turbulence codes are often difficult to visualize due their high dimensionality, the nontrivial geometry of the underlying grids, and the vast range of spatial scales. We present an interactive visualization framework that attempts to address these issues. Images from a nonlinear gyrokinetic simulation are presented. %B Plasma Science, IEEE Transactions on %V 36 %P 1112 - 1113 %8 2008/08// %@ 0093-3813 %G eng %N 4 %R 10.1109/TPS.2008.924509 %0 Journal Article %J Multimedia, IEEE Transactions on %D 2008 %T A Constrained Probabilistic Petri Net Framework for Human Activity Detection in Video %A Albanese, M. %A Chellapa, Rama %A Moscato, V. %A Picariello, A. %A V.S. Subrahmanian %A Turaga,P. %A Udrea,O. %K activity %K dataset;automated %K detection;human %K interactions;security %K net;human %K nets;image %K Petri %K probabilistic %K processing;multiagent %K processing;video %K representation;low-level %K representation;video %K signal %K Surveillance %K surveillance; %K systems;constrained %K systems;surveillance %K tarmac %K TSA %K videos;Petri %X Recognition of human activities in restricted settings such as airports, parking lots and banks is of significant interest in security and automated surveillance systems. In such settings, data is usually in the form of surveillance videos with wide variation in quality and granularity. Interpretation and identification of human activities requires an activity model that a) is rich enough to handle complex multi-agent interactions, b) is robust to uncertainty in low-level processing and c) can handle ambiguities in the unfolding of activities. We present a computational framework for human activity representation based on Petri nets. We propose an extension-Probabilistic Petri Nets (PPN)-and show how this model is well suited to address each of the above requirements in a wide variety of settings. We then focus on answering two types of questions: (i) what are the minimal sub-videos in which a given activity is identified with a probability above a certain threshold and (ii) for a given video, which activity from a given set occurred with the highest probability? We provide the PPN-MPS algorithm for the first problem, as well as two different algorithms (naive PPN-MPA and PPN-MPA) to solve the second. Our experimental results on a dataset consisting of bank surveillance videos and an unconstrained TSA tarmac surveillance dataset show that our algorithms are both fast and provide high quality results. %B Multimedia, IEEE Transactions on %V 10 %P 982 - 996 %8 2008/10// %@ 1520-9210 %G eng %N 6 %R 10.1109/TMM.2008.2001369 %0 Journal Article %J Computer-Aided Design %D 2008 %T Content-based assembly search: A step towards assembly reuse %A Deshmukh,Abhijit S. %A Banerjee,Ashis Gopal %A Gupta, Satyandra K. %A Sriram,Ram D. %K Assembly characteristics %K Assembly mating conditions %K Content-based assembly search %K Graph compatibility %X The increased use of CAD systems by product development organizations has resulted in the creation of large databases of assemblies. This explosion of assembly data is likely to continue in the future. In many situations, a text-based search alone may not be sufficient to search for assemblies and it may be desirable to search for assemblies based on the content of the assembly models. The ability to perform content-based searches on these databases is expected to help the designers in the following two ways. First, it can facilitate the reuse of existing assembly designs, thereby reducing the design time. Second, a lot of useful designs for manufacturing, and assembly knowledge are implicitly embedded in existing assemblies. Therefore a capability to locate existing assemblies and examine them can be used as a learning tool by designers to learn from the existing assembly designs. This paper describes a system for performing content-based searches on assembly databases. We identify templates for comprehensive search definitions and describe algorithms to perform content-based searches for mechanical assemblies. We also illustrate the capabilities of our system through several examples. %B Computer-Aided Design %V 40 %P 244 - 261 %8 2008/02// %@ 0010-4485 %G eng %U http://www.sciencedirect.com/science/article/pii/S0010448507002424 %N 2 %R 10.1016/j.cad.2007.10.012 %0 Journal Article %J Intelligent Systems, IEEE %D 2008 %T CONVEX: Similarity-Based Algorithms for Forecasting Group Behavior %A Martinez,V. %A Simari,G. I %A Sliva,A. %A V.S. Subrahmanian %K (artificial %K algorithm;action %K algorithm;behavioural %K algorithm;CONVEXk-NN %K BEHAVIOR %K computing;ontologies %K CONVEXMerge %K forecasting;high-dimensional %K intelligence); %K metric %K sciences %K space;ontology;similarity-based %K vector;context %K vector;group %X A proposed framework for predicting a group's behavior associates two vectors with that group. The context vector tracks aspects of the environment in which the group functions; the action vector tracks the group's previous actions. Given a set of past behaviors consisting of a pair of these vectors and given a query context vector, the goal is to predict the associated action vector. To achieve this goal, two families of algorithms employ vector similarity. CONVEXk _NN algorithms use k-nearest neighbors in high-dimensional metric spaces; CONVEXMerge algorithms look at linear combinations of distances of the query vector from context vectors. Compared to past prediction algorithms, these algorithms are extremely fast. Moreover, experiments on real-world data sets show that the algorithms are highly accurate, predicting actions with well over 95-percent accuracy. %B Intelligent Systems, IEEE %V 23 %P 51 - 57 %8 2008/08//july %@ 1541-1672 %G eng %N 4 %R 10.1109/MIS.2008.62 %0 Journal Article %J Science %D 2008 %T Copernican challenges face those who suggest that collaboration, not computation are the driving energy for socio-technical systems that characterize Web 2.0. %A Shneiderman, Ben %B Science %V 319 %P 1349 - 1350 %8 2008/// %G eng %0 Journal Article %J Algorithmica %D 2008 %T Cost-Sharing Mechanisms for Network Design %A Gupta,Anupam %A Srinivasan, Aravind %A Tardos,Éva %X We consider a single-source network design problem from a game-theoretic perspective. Gupta, Kumar and Roughgarden (Proc. 35th Annual ACM STOC, pp. 365–372, 2003 ) developed a simple method for a single-source rent-or-buy problem that also yields the best-known approximation ratio for the problem. We show how to use a variant of this method to develop an approximately budget-balanced and group strategyproof cost-sharing method for the problem. The novelty of our approach stems from our obtaining the cost-sharing methods for the rent-or-buy problem by carefully combining cost-shares for the simpler Steiner tree problem. Our algorithm is conceptually simpler than the previous such cost-sharing method due to Pál and Tardos (Proc. 44th Annual FOCS, pp. 584–593, 2003 ), and improves the previously-known approximation factor of 15 to 4.6. %B Algorithmica %V 50 %P 98 - 119 %8 2008/// %@ 0178-4617 %G eng %U http://dx.doi.org/10.1007/s00453-007-9065-y %N 1 %0 Conference Paper %B Proceedings of the 10th annual conference on Genetic and evolutionary computation %D 2008 %T CrossNet: a framework for crossover with network-based chromosomal representations %A Stonedahl,Forrest %A Rand, William %A Wilensky,Uri %K crossover %K Genetic algorithms %K graphs %K linkage %K NETWORKS %K recombination %X We propose a new class of crossover operators for genetic algorithms (CrossNet) which use a network-based (or graph-based) chromosomal representation. We designed CrossNet with the intent of providing a framework for creating crossover operators that take advantage of domain-specific knowledge for solving problems. Specifically, GA users supply a network which defines the epistatic relationships between genes in the genotype. CrossNet-based crossover uses this information with the goal of improving linkage. We performed two experiments that compared CrossNet-based crossover with one-point and uniform crossover. The first experiment involved the density classification problem for cellular automata (CA), and the second experiment involved fitting two randomly generated hyperplane-defined functions (hdf's). Both of these exploratory experiments support the hypothesis that CrossNet-based crossover can be useful, although performance improvements were modest. We discuss the results and remain hopeful about the successful application of CrossNet to other domains. We conjecture that future work with the CrossNet framework will provide a useful new perspective for investigating linkage and chromosomal representations. %B Proceedings of the 10th annual conference on Genetic and evolutionary computation %S GECCO '08 %I ACM %C New York, NY, USA %P 1057 - 1064 %8 2008/// %@ 978-1-60558-130-9 %G eng %U http://doi.acm.org/10.1145/1389095.1389290 %R 10.1145/1389095.1389290 %0 Journal Article %J ACM Trans. Comput.-Hum. Interact. %D 2008 %T Data Sonification for Users with Visual Impairment: A Case Study with Georeferenced Data %A Zhao,Haixia %A Plaisant, Catherine %A Shneiderman, Ben %A Lazar,Jonathan %K auditory user interfaces %K information seeking %K Interactive sonification %K universal usability %K users with visual impairment %X We describe the development and evaluation of a tool, iSonic, to assist users with visual impairment in exploring georeferenced data using coordinated maps and tables, augmented with nontextual sounds and speech output. Our in-depth case studies with 7 blind users during 42 hours of data collection, showed that iSonic enabled them to find facts and discover trends in georeferenced data, even in unfamiliar geographical contexts, without special devices. Our design was guided by an Action-by-Design-Component (ADC) framework, which was also applied to scatterplots to demonstrate its generalizability. Video and download is available at www.cs.umd.edu/hcil/iSonic/. %B ACM Trans. Comput.-Hum. Interact. %V 15 %P 4:1–4:28 - 4:1–4:28 %8 2008/05// %@ 1073-0516 %G eng %U http://doi.acm.org/10.1145/1352782.1352786 %N 1 %R 10.1145/1352782.1352786 %0 Conference Paper %B Proceedings of the 8th ACM workshop on Digital rights management %D 2008 %T A decision theoretic framework for analyzing binary hash-based content identification systems %A Varna,Avinash L. %A Swaminathan,Ashwin %A M. Wu %K content fingerprinting %K content identification %K decision theory %X Content identification has many applications, ranging from preventing illegal sharing of copyrighted content on video sharing websites, to automatic identification and tagging of content. Several content identification techniques based on watermarking or robust hashes have been proposed in the literature, but they have mostly been evaluated through experiments. This paper analyzes binary hash-based content identification schemes under a decision theoretic framework and presents a lower bound on the length of the hash required to correctly identify multimedia content that may have undergone modifications. A practical scheme for content identification is evaluated under the proposed framework. The results obtained through experiments agree very well with the performance suggested by the theoretical analysis. %B Proceedings of the 8th ACM workshop on Digital rights management %S DRM '08 %I ACM %C New York, NY, USA %P 67 - 76 %8 2008/// %@ 978-1-60558-290-0 %G eng %U http://doi.acm.org/10.1145/1456520.1456532 %R 10.1145/1456520.1456532 %0 Book Section %B Automata, Languages and Programming %D 2008 %T Delegating Capabilities in Predicate Encryption Systems %A Elaine Shi %A Waters,Brent %E Aceto,Luca %E Damgård,Ivan %E Goldberg,Leslie %E Halldórsson,Magnús %E Ingólfsdóttir,Anna %E Walukiewicz,Igor %K Computer science %X In predicate encryption systems, given a capability, one can evaluate one or more predicates on the plaintext encrypted, while all other information about the plaintext remains hidden. We consider the role of delegation in such predicate encryption systems. Suppose Alice has a capability, and she wishes to delegate to Bob a more restrictive capability allowing the decryption of a subset of the information Alice can learn about the plaintext encrypted. We formally define delegation in predicate encryption systems, propose a new security definition for delegation, and give an efficient construction supporting conjunctive queries. The security of our construction can be reduced to the general 3-party Bilinear Diffie-Hellman assumption, and the Bilinear Decisional Diffie-Hellman assumption in composite order bilinear groups. %B Automata, Languages and Programming %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5126 %P 560 - 578 %8 2008 %@ 978-3-540-70582-6 %G eng %U http://www.springerlink.com/content/w320422h15050004/abstract/ %0 Journal Article %J Proceedings of the American Society for Information Science and Technology %D 2008 %T Designing community-based emergency communication system: A preliminary study %A Fei Wu,P. %A Qu,Y. %A Fleischmann,K. %A Golbeck,J. %A Jaeger,P. %A Preece,J. %A Shneiderman, Ben %B Proceedings of the American Society for Information Science and Technology %V 45 %P 1 - 3 %8 2008/// %G eng %N 1 %0 Journal Article %J Information Forensics and Security, IEEE Transactions on %D 2008 %T Digital image forensics via intrinsic fingerprints %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K ACQUISITION %K analysis;intrinsic %K approximation;cameras;digital %K camera %K deconvolution;digital %K devices;blind %K fingerprints;time %K forensics;forensic %K identification;image %K image %K images;digital %K invariant %K photography;fingerprint %K sensors; %X Digital imaging has experienced tremendous growth in recent decades, and digital camera images have been used in a growing number of applications. With such increasing popularity and the availability of low-cost image editing software, the integrity of digital image content can no longer be taken for granted. This paper introduces a new methodology for the forensic analysis of digital camera images. The proposed method is based on the observation that many processing operations, both inside and outside acquisition devices, leave distinct intrinsic traces on digital images, and these intrinsic fingerprints can be identified and employed to verify the integrity of digital data. The intrinsic fingerprints of the various in-camera processing operations can be estimated through a detailed imaging model and its component analysis. Further processing applied to the camera captured image is modelled as a manipulation filter, for which a blind deconvolution technique is applied to obtain a linear time-invariant approximation and to estimate the intrinsic fingerprints associated with these postcamera operations. The absence of camera-imposed fingerprints from a test image indicates that the test image is not a camera output and is possibly generated by other image production processes. Any change or inconsistencies among the estimated camera-imposed fingerprints, or the presence of new types of fingerprints suggest that the image has undergone some kind of processing after the initial capture, such as tampering or steganographic embedding. Through analysis and extensive experimental studies, this paper demonstrates the effectiveness of the proposed framework for nonintrusive digital image forensics. %B Information Forensics and Security, IEEE Transactions on %V 3 %P 101 - 117 %8 2008/03// %@ 1556-6013 %G eng %N 1 %R 10.1109/TIFS.2007.916010 %0 Report %D 2008 %T Directing javascript with arrows (functional pearl) %A Khoo,Y.P. %A Hicks, Michael W. %A Foster, Jeffrey S. %A Sazawal,V. %X JavaScript, being a single-threaded language, makes extensive use of event-driven programming to enable responsive web applications. However, standard approaches to sequencing events are messy, and often lead to code that is difficult to understand and maintain. We have found that arrows, a generalization of monads, are an elegant solution to this problem. Arrows allow us to easily write asynchronous programs in small, modular units of code, and flexibly compose them in many different ways, while nicely abstracting the details of asynchronous program composition. In particular, we show how to use arrows to construct a variety of state machines, such as autoscrollers and drag-and-drop handlers. %B Technical Reports of the Computer Science Department %8 2008/// %G eng %0 Journal Article %J ACM SIGCOMM Computer Communication Review %D 2008 %T Discarte: a disjunctive internet cartographer %A Sherwood,R. %A Bender,A. %A Spring, Neil %B ACM SIGCOMM Computer Communication Review %V 38 %P 303 - 314 %8 2008/// %G eng %N 4 %0 Conference Paper %B ICPR 2008. 19th International Conference on Pattern Recognition, 2008. %D 2008 %T Document Zone Classification Using Partial Least Squares and Hybrid Classifiers %A Abd-Almageed, Wael %A Agrawal,Mudit %A Seo,W. %A David Doermann %X This paper introduces a novel document-zone classification algorithm. Low level image features are first extracted from document zones and partial least squares is used on pairs of classes to compute discriminating pairwise features. Rather than using the popular one-against-all and one-against-one voting schemes, we introduce a novel hybrid method which combines the benefits of the two schemes. The algorithm is applied on the University of Washington dataset and 97.3% classification accuracy is obtained. %B ICPR 2008. 19th International Conference on Pattern Recognition, 2008. %P 1 - 4 %8 2008/// %G eng %0 Journal Article %J Nature %D 2008 %T The draft genome of the transgenic tropical fruit tree papaya (Carica papaya Linnaeus) %A Ming,R. %A Hou,S. %A Feng,Y. %A Yu,Q. %A Dionne-Laporte,A. %A Saw,J.H. %A Senin,P. %A Wang,W. %A Ly,B.V. %A Lewis,K.L.T. %A others %X Papaya, a fruit crop cultivated in tropical and subtropical regions, is known for its nutritional benefits and medicinal applications. Here we report a 3 draft genome sequence of 'SunUp' papaya, the first commercial virus-resistant transgenic fruit tree to be sequenced. The papaya genome is three times the size of the Arabidopsis genome, but contains fewer genes, including significantly fewer disease-resistance gene analogues. Comparison of the five sequenced genomes suggests a minimal angiosperm gene set of 13,311. A lack of recent genome duplication, atypical of other angiosperm genomes sequenced so far, may account for the smaller papaya gene number in most functional groups. Nonetheless, striking amplifications in gene number within particular functional groups suggest roles in the evolution of tree-like habit, deposition and remobilization of starch reserves, attraction of seed dispersal agents, and adaptation to tropical daylengths. Transgenesis at three locations is closely associated with chloroplast insertions into the nuclear genome, and with topoisomerase I recognition sites. Papaya offers numerous advantages as a system for fruit-tree functional genomics, and this draft genome sequence provides the foundation for revealing the basis of Carica's distinguishing morpho-physiological, medicinal and nutritional properties. %B Nature %V 452 %P 991 - 996 %8 2008/// %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v452/n7190/full/nature06856.html %N 7190 %R 10.1038/nature06856 %0 Journal Article %J Journal of Functional Programming %D 2008 %T Dynamic Rebinding for Marshalling and Update, Via Redex-Time and Destruct-Time Reduction %A Sewell,Peter %A Stoyle,Gareth %A Hicks, Michael W. %A Bierman,Gavin %A Wansbrough,Keith %X Most programming languages adopt static binding, but for distributed programming an exclusive reliance on static binding is too restrictive: dynamic binding is required in various guises, for example, when a marshalled value is received from the network, containing identifiers that must be rebound to local resources. Typically, it is provided only by ad hoc mechanisms that lack clean semantics. In this paper, we adopt a foundational approach, developing core dynamic rebinding mechanisms as extensions to the simply typed call-by-value λ calculus. To do so, we must first explore refinements of the call-by-value reduction strategy that delay instantiation, to ensure computations make use of the most recent versions of rebound definitions. We introduce redex-time and destruct-time strategies. The latter forms the basis for a λmarsh calculus that supports dynamic rebinding of marshalled values, while remaining as far as possible statically typed. We sketch an extension of λmarsh with concurrency and communication, giving examples showing how wrappers for encapsulating untrusted code can be expressed. Finally, we show that a high-level semantics for dynamic updating can also be based on the destruct-time strategy, defining a λupdate calculus with simple primitives to provide type-safe updating of running code. We show how the ideas of this simple calculus extend to more real-world, module-level dynamic updating in the style of Erlang. We thereby establish primitives and a common semantic foundation for a variety of real-world dynamic rebinding requirements. %B Journal of Functional Programming %V 18 %P 437 - 502 %8 2008/// %G eng %N 04 %R 10.1017/S0956796807006600 %0 Journal Article %J Multimedia, IEEE Transactions on %D 2008 %T Dynamic Resource Allocation for Robust Distributed Multi-Point Video Conferencing %A Chen,Meng %A Su,Guan-Ming %A M. Wu %K allocation;error %K channels;packet-division %K communication; %K conferencing %K conferencing;multi-access %K distributed %K dynamic %K erasure %K error %K multiple-access-based %K multipoint %K Protection %K resource %K scheme;multi-point %K scheme;robust %K schemes;packet %K streams;multistream %K system;multiple %K systems;teleconferencing;video %K video %X This paper proposes a distributed multi-point video conferencing system over packet erasure channels, where the aggregation of multiple video streams and resource allocation are performed in a distributed manner. Video stream combiners, which are located in different geographical areas and serve as portals for conferees, aggregate incoming streams supplied by local users with other streams aggregated from nearby video stream combiners. A packet-division multiple-access (PDMA)-based error protection scheme is proposed to be performed at each video stream combiner to minimize the maximal expected video distortion among aggregated streams. The proposed error protection scheme for multi-stream aggregation also supports user preference. In order to deliver video streams to end users with different preferred quality, a consensus algorithm is proposed to adaptively perform resource allocation based on user preference. Simulation results show that the proposed multi-stream aggregation and error protection scheme has significant gains over traditional multi-stream error protection schemes for a multi-point video conferencing system. %B Multimedia, IEEE Transactions on %V 10 %P 910 - 925 %8 2008/08// %@ 1520-9210 %G eng %N 5 %R 10.1109/TMM.2008.922846 %0 Conference Paper %B Proceedings of the 2008 international symposium on Software testing and analysis %D 2008 %T Effective and scalable software compatibility testing %A Yoon,Il-Chul %A Sussman, Alan %A Memon, Atif M. %A Porter, Adam %K component-based software system %K software compatibility testing %X Today's software systems are typically composed of multiple components, each with different versions. Software compatibility testing is a quality assurance task aimed at ensuring that multi-component based systems build and/or execute correctly across all their versions' combinations, or configurations. Because there are complex and changing interdependencies between components and their versions, and because there are such a large number of configurations, it is generally infeasible to test all potential configurations. Consequently, in practice, compatibility testing examines only a handful of default or popular configurations to detect problems; as a result costly errors can and do escape to the field. This paper presents a new approach to compatibility testing, called Rachet. We formally model the entire configuration space for software systems and use the model to generate test plans to sample a portion of the space. In this paper, we test all direct dependencies between components and execute the test plan efficiently in parallel. We present empirical results obtained by applying our approach to two large-scale scientific middleware systems. The results show that for these systems Rachet scaled well and discovered incompatibilities between components, and that testing only direct dependences did not compromise test quality. %B Proceedings of the 2008 international symposium on Software testing and analysis %S ISSTA '08 %I ACM %C New York, NY, USA %P 63 - 74 %8 2008/// %@ 978-1-60558-050-0 %G eng %U http://doi.acm.org/10.1145/1390630.1390640 %R 10.1145/1390630.1390640 %0 Journal Article %J Mobile Computing, IEEE Transactions on %D 2008 %T Efficient and Resilient Backbones for Multihop Wireless Networks %A Lee,Seungjoon %A Bhattacharjee, Bobby %A Srinivasan, Aravind %A Khuller, Samir %K algorithm;protocols;wireless %K backbone %K channels; %K CONSTRUCTION %K distributed %K networks;parameterized %K protocol;multihop %K wireless %X We consider the problem of finding "backbones" in multihop wireless networks. The backbone provides end-to-end connectivity, allowing nonbackbone nodes to save energy since they do not have to route nonlocal data or participate in the routing protocol. Ideally, such a backbone would be small, consist primarily of high capacity nodes, and remain connected even when nodes are mobile or fail. Unfortunately, it is often infeasible to construct a backbone that has all of these properties; e.g., a small optimal backbone is often too sparse to handle node failures or high mobility. We present a parameterized backbone construction algorithm that permits explicit trade-offs between backbone size, resilience to node movement and failure, energy consumption, and path lengths. We prove that our scheme can construct essentially best possible backbones (with respect to energy consumption and backbone size) when the network is relatively static. We generalize our scheme to build more robust structures better suited to networks with higher mobility. We present a distributed protocol based upon our algorithm and show that this protocol builds and maintains a connected backbone in dynamic networks. Finally, we present detailed packet-level simulation results to evaluate and compare our scheme with existing energy-saving techniques. Our results show that, depending on the network environment, our scheme increases network lifetimes by 20 percent to 220 percent without adversely affecting delivery ratio or end-to-end latency. %B Mobile Computing, IEEE Transactions on %V 7 %P 1349 - 1362 %8 2008/11// %@ 1536-1233 %G eng %N 11 %R 10.1109/TMC.2008.69 %0 Journal Article %J Sparks of innovation in human-computer interaction %D 2008 %T Embedded menus: selecting items in context %A Koved,L. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %8 2008/// %G eng %0 Journal Article %J Computers & Education %D 2008 %T Enabling teachers to explore grade patterns to identify individual needs and promote fairer student assessment %A Friedler,Sorelle A. %A Tan,Yee Lin %A Peer,Nir J. %A Shneiderman, Ben %K Evaluation methodologies %K Human–computer interface %K Improving classroom teaching %K Pedagogical issues %K Teaching/learning strategies %X Exploring student test, homework, and other assessment scores is a challenge for most teachers, especially when attempting to identify cross-assessment weaknesses and produce final course grades. During the course, teachers need to identify subject weaknesses in order to help students who are struggling with a particular topic. This identification often needs to happen across multiple assessment data points and should be considered in comparison to the class’s progress as a whole. When determining grades, fairness to all is essential, but there are special needs for students who did poorly on one exam or had a steadily increasing grasp of the subject. We present eduViz, a visualization tool designed to help teachers explore and assign grades. Teachers can see the trajectory of student scores, the relationship of a particular student to the class, and use categories they have defined in order to filter their assessment information. Query response is immediate and all logical comparisons are possible. Teachers can easily compare their query to the class or per student average as well as view scores by raw point total or percentage. Additionally, eduViz provides a grade assignment interface which allows teachers to view sorted student scores in a scatterplot. This scatterplot is coupled with a unique partition slider which allows users to move color coordinated bands on the scatterplot to indicate grade ranges. As these grade ranges are set, a histogram is updated to show the number of students assigned to each grade range. These features give teachers new and powerful ways to explore and assign grades so that they can better understand student strengths and weaknesses and make the most of the time they have available. Interviews with 16 expert teachers indicate that eduViz is a success across fields, provides teachers with a useful tool to understand and help their classes, and encourages reflective practice. %B Computers & Education %V 51 %P 1467 - 1485 %8 2008/12// %@ 0360-1315 %G eng %U http://www.sciencedirect.com/science/article/pii/S0360131508000353 %N 4 %R 10.1016/j.compedu.2008.01.005 %0 Book Section %B Software Development %D 2008 %T An Environment for Conducting Families of Software Engineering Experiments %A Hochstein, Lorin %A Nakamura,Taiga %A Shull, Forrest %A Zazworka, Nico %A Basili, Victor R. %A Zelkowitz, Marvin V %E Zelkowitz, Marvin V %X The classroom is a valuable resource for conducting software engineering experiments. However, coordinating a family of experiments in classroom environments presents a number of challenges to researchers. Understanding how to run such experiments, developing procedures to collect accurate data, and collecting data that is consistent across multiple studies are major problems. This paper describes an environment, the Experiment Manager that simplifies the process of collecting, managing, and sanitizing data from classroom experiments, while minimizing disruption to natural subject behavior. We have successfully used this environment to study the impact of parallel programming languages in the high‐performance computing domain on programmer productivity at multiple universities across the United States. %B Software Development %I Elsevier %V Volume 74 %P 175 - 200 %8 2008/// %@ 0065-2458 %G eng %U http://www.sciencedirect.com/science/article/pii/S0065245808006050 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2008 %T Environmental signatures associated with cholera epidemics %A Constantin de Magny,G. %A Murtugudde,R. %A Sapiano,M. R. P. %A Nizam,A. %A Brown,C. W. %A Busalacchi,A. J. %A Yunus,M. %A Nair,G. B. %A Gil,A. I. %A Lanata,C. F. %A Rita R Colwell %X The causative agent of cholera, Vibrio cholerae, has been shown to be autochthonous to riverine, estuarine, and coastal waters along with its host, the copepod, a significant member of the zooplankton community. Temperature, salinity, rainfall and plankton have proven to be important factors in the ecology of V. cholerae, influencing the transmission of the disease in those regions of the world where the human population relies on untreated water as a source of drinking water. In this study, the pattern of cholera outbreaks during 1998–2006 in Kolkata, India, and Matlab, Bangladesh, and the earth observation data were analyzed with the objective of developing a prediction model for cholera. Satellite sensors were used to measure chlorophyll a concentration (CHL) and sea surface temperature (SST). In addition, rainfall data were obtained from both satellite and in situ gauge measurements. From the analyses, a statistically significant relationship between the time series for cholera in Kolkata, India, and CHL and rainfall anomalies was determined. A statistically significant one month lag was observed between CHL anomaly and number of cholera cases in Matlab, Bangladesh. From the results of the study, it is concluded that ocean and climate patterns are useful predictors of cholera epidemics, with the dynamics of endemic cholera being related to climate and/or changes in the aquatic ecosystem. When the ecology of V. cholerae is considered in predictive models, a robust early warning system for cholera in endemic regions of the world can be developed for public health planning and decision making.ecology epidemiology microbiology remote sensing %B Proceedings of the National Academy of Sciences %V 105 %P 17676 - 17676 %8 2008/// %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/105/46/17676 %N 46 %R 10.1073/pnas.0809654105 %0 Journal Article %J Information processing & management %D 2008 %T Evaluating Exploratory Search Systems %A Kules,B. %A Shneiderman, Ben %B Information processing & management %V 44 %P 463 - 484 %8 2008/// %G eng %N 2 %0 Thesis %D 2008 %T Evaluating predicates over encrypted data %A Elaine Shi %X The high-level goal of this thesis is to build predicate encryption systems that are efficient, support expressive queries and rich operations. Our contributions are summarized below: (1) We propose a predicate encryption scheme supporting multi-dimensional range queries. Prior to this work, researchers have constructed schemes support equality tests. Hence, our scheme supports more expressive queries than before. At the core of this construction is a technique to support conjunctive queries without leaking the outcome of each individual clause. (2) We study how to delegate capabilities in predicate encryption schemes. To demonstrate why delegation may be interesting, imagine that Alice has a capability, and she wishes to delegate to Bob a more restrictive capability allowing him to decrypt a subset of the information Alice can learn about the plaintext encrypted. We propose a security definition for delegation, and build a scheme supporting delegation and conjunctive queries. (3) Most prior work focuses on hiding the plaintext (encoded in the ciphertext), but does not provide guarantees about the secrecy of the queries (encoded in the capabilities). In other words, given a capability, one might be able to infer from it what the query predicate is. We study how to hide the query predicates, and propose a scheme supporting inner-product queries that hides the query predicates in addition to the plaintext. %I Carnegie Mellon University %8 2008 %G eng %U http://books.google.com/books?hl=en&lr=&id=jBOgjMdBeHAC&oi=fnd&pg=PA1&dq=+Elaine+Shi&ots=W3nTBSxa3H&sig=76cmlidxiCTBXbaFhfkl9zVOm6s %R Doctoral Dissertation %0 Journal Article %J Computer Graphics and Applications, IEEE %D 2008 %T Evaluating Visual Analytics at the 2007 VAST Symposium Contest %A Plaisant, Catherine %A Grinstein,G. %A Scholtz,J. %A Whiting,M. %A O'Connell,T. %A Laskowski,S. %A Chien,L. %A Tat,A. %A Wright,W. %A Gorg,C. %A Zhicheng Liu %A Parekh,N. %A Singhal,K. %A Stasko,J. %K 2007 %K analytics %K analytics;data %K and %K Contest;Visual %K Science %K Symposium %K Technology;data %K VAST %K visualisation; %K visualization;visual %X In this article, we report on the contest's data set and tasks, the judging criteria, the winning tools, and the overall lessons learned in the competition. We believe that by organizing these contests, we're creating useful resources for researchers and are beginning to understand how to better evaluate VA tools. Competitions encourage the community to work on difficult problems, improve their tools, and develop baselines for others to build or improve upon. We continue to evolve a collection of data sets, scenarios, and evaluation methodologies that reflect the richness of the many VA tasks and applications. %B Computer Graphics and Applications, IEEE %V 28 %P 12 - 21 %8 2008/04//march %@ 0272-1716 %G eng %N 2 %R 10.1109/MCG.2008.27 %0 Journal Article %J Software: Practice and Experience %D 2008 %T Experiences with building an intrusion-tolerant group communication system %A Ramasamy,HariGovind V. %A Pandey,Prashant %A Michel Cukier %A Sanders,William H. %K distributed protocols %K experimental evaluation %K Fault tolerance %K Group communication %K intrusion tolerance %X There are many group communication systems (GCSs) that provide consistent group membership and reliable, ordered multicast properties in the presence of crash faults. However, relatively few GCS implementations are able to provide these properties in the presence of malicious faults resulting from intrusions. We describe the systematic transformation of a crash-tolerant GCS, namely C-Ensemble, into an intrusion-tolerant GCS, the ITUA GCS. To perform the transformation, we devised intrusion-tolerant versions of key group communication protocols. We then inserted implementations of the protocols into C-Ensemble and made significant changes to the rest of the C-Ensemble protocol stack to make the stack intrusion tolerant. We quantify the cost of providing intrusion-tolerant group communication in two ways. First, we quantify the implementation effort by presenting a detailed analysis of the amount of change required to the original C-Ensemble system. In doing so, we provide insight into the choice of building an intrusion-tolerant GCS from scratch versus building one by leveraging a crash-tolerant implementation. Second, we quantify the run-time performance cost of tolerating intrusions by presenting results from an experimental evaluation of the main intrusion-tolerant microprotocols. The results are analyzed to identify the parts that contribute the most overhead while providing intrusion tolerance during both normal operation and recovery from intrusions. Copyright © 2007 John Wiley & Sons, Ltd. %B Software: Practice and Experience %V 38 %P 639 - 666 %8 2008/// %@ 1097-024X %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/spe.848/abstract %N 6 %R 10.1002/spe.848 %0 Conference Paper %B Data Integration in the Life Sciences %D 2008 %T Exploiting ontology structure and patterns of annotation to mine significant associations between pairs of controlled vocabulary terms %A Lee,W. J %A Raschid, Louiqa %A Sayyadi,H. %A Srinivasan,P. %B Data Integration in the Life Sciences %P 44 - 60 %8 2008/// %G eng %0 Journal Article %J Proceedings of the VLDB Endowment %D 2008 %T Exploiting shared correlations in probabilistic databases %A Sen,Prithviraj %A Deshpande, Amol %A Getoor, Lise %X There has been a recent surge in work in probabilistic databases, propelled in large part by the huge increase in noisy data sources --- from sensor data, experimental data, data from uncurated sources, and many others. There is a growing need for database management systems that can efficiently represent and query such data. In this work, we show how data characteristics can be leveraged to make the query evaluation process more efficient. In particular, we exploit what we refer to as shared correlations where the same uncertainties and correlations occur repeatedly in the data. Shared correlations occur mainly due to two reasons: (1) Uncertainty and correlations usually come from general statistics and rarely vary on a tuple-to-tuple basis; (2) The query evaluation procedure itself tends to re-introduce the same correlations. Prior work has shown that the query evaluation problem on probabilistic databases is equivalent to a probabilistic inference problem on an appropriately constructed probabilistic graphical model (PGM). We leverage this by introducing a new data structure, called the random variable elimination graph (rv-elim graph) that can be built from the PGM obtained from query evaluation. We develop techniques based on bisimulation that can be used to compress the rv-elim graph exploiting the presence of shared correlations in the PGM, the compressed rv-elim graph can then be used to run inference. We validate our methods by evaluating them empirically and show that even with a few shared correlations significant speed-ups are possible. %B Proceedings of the VLDB Endowment %V 1 %P 809 - 820 %8 2008/08// %@ 2150-8097 %G eng %U http://dx.doi.org/10.1145/1453856.1453944 %N 1 %R 10.1145/1453856.1453944 %0 Journal Article %J Proceedings of the American Society for Information Science and Technology %D 2008 %T Extending sense-making models with ideas from cognition and learning theories %A Zhang,P. %A Soergel,D. %A Klavans,J. L %A Oard, Douglas %B Proceedings of the American Society for Information Science and Technology %V 45 %P 23 - 23 %8 2008/// %G eng %N 1 %0 Journal Article %J SIGMOD Conference %D 2008 %T Extreme visualization: squeezing a billion records into a million pixels %A Shneiderman, Ben %X Database searches are usually performed with querylanguages and form fill in templates, with results displayed in tabular lists. However, excitement is building around dynamic queries sliders and other graphical selectors for query specification, with results displayed by information visualization techniques. These filtering techniques have proven to be effective for many tasks in which visual presentations enable discovery of relationships, clusters, outliers, gaps, and other patterns. Scaling visual presentations from millions to billions of records will require collaborative research efforts in information visualization and database management to enable rapid aggregation, meaningful coordinated windows, and effective summary graphics. This paper describes current and proposed solutions (atomic, aggregated, and density plots) that facilitate sense-making for interactive visual exploration of billion record data sets. %B SIGMOD Conference %P 3 - 12 %8 2008/// %G eng %0 Conference Paper %B IEEE Symposium on Security and Privacy, 2008. SP 2008 %D 2008 %T Fable: A Language for Enforcing User-defined Security Policies %A Swamy,N. %A Corcoran,B.J. %A Hicks, Michael W. %K Access control %K Automata %K Collaborative work %K Communication system security %K Computer languages %K computer security %K Data security %K enforcement policy %K FABLE %K Government %K high-level security goals %K information flow %K Information security %K Language-based security %K programming languages %K Programming profession %K provenance %K security automata %K security labels %K security of data %K user-defined security policies %K verified enforcement %K Web programming language %X This paper presents FABLE, a core formalism for a programming language in which programmers may specify security policies and reason that these policies are properly enforced. In FABLE, security policies can be expressed by associating security labels with the data or actions they protect. Programmers define the semantics of labels in a separate part of the program called the enforcement policy. FABLE prevents a policy from being circumvented by allowing labeled terms to be manipulated only within the enforcement policy; application code must treat labeled values abstractly. Together, these features facilitate straightforward proofs that programs implementing a particular policy achieve their high-level security goals. FABLE is flexible enough to implement a wide variety of security policies, including access control, information flow, provenance, and security automata. We have implemented FABLE as part of the LINKS web programming language; we call the resulting language SELlNKS. We report on our experience using SELlNKS to build two substantial applications, a wiki and an on-line store, equipped with a combination of access control and provenance policies. To our knowledge, no existing framework enables the enforcement of such a wide variety of security policies with an equally high level of assurance. %B IEEE Symposium on Security and Privacy, 2008. SP 2008 %I IEEE %P 369 - 383 %8 2008/05/18/22 %@ 978-0-7695-3168-7 %G eng %R 10.1109/SP.2008.29 %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on %D 2008 %T Factorized variational approximations for acoustic multi source localization %A Cevher, V. %A Sankaranarayanan,A. C %A Chellapa, Rama %K approximation;acoustic %K approximation;Bayes %K approximation;multisensor %K approximation;variational %K Bayes %K Chi %K detection;wireless %K distribution;Gaussian %K localization;factorized %K localization;sensor %K methods;Gaussian %K multisource %K networks; %K networks;stochastic %K processes;acoustic %K sensor %K signal %K strength;sensor %K systems;object %K tracking;received %K variational %X Estimation based on received signal strength (RSS) is crucial in sensor networks for sensor localization, target tracking, etc. In this paper, we present a Gaussian approximation of the Chi distribution that is applicable to general RSS source localization problems in sensor networks. Using our Gaussian approximation, we provide a factorized variational Bayes (VB) approximation to the location and power posterior of multiple sources using a sensor network. When the source signal and the sensor noise have uncorrelated Gaussian distributions, we demonstrate that the envelope of the sensor output can be accurately modeled with a multiplicative Gaussian noise model. In turn, our factorized VB approximations decrease the computational complexity and provide computational robustness as the number of targets increases. Simulations are provided to demonstrate the effectiveness of the proposed approximations. %B Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on %P 2409 - 2412 %8 2008/04/31/4 %G eng %R 10.1109/ICASSP.2008.4518133 %0 Conference Paper %B 8th International Web Archiving Workshop, Aarhus, Denmark. %D 2008 %T Fast browsing of archived Web contents %A Song,S. %A JaJa, Joseph F. %X The web is becoming the preferred medium for communicatingand storing information pertaining to almost any human activity. However it is an ephemeral medium whose contents are constantly changing, resulting in a permanent loss of part of our cultural and scientific heritage on a regular basis. Archiving important web contents is a very challenging technical problem due to its tremendous scale and complex structure, extremely dynamic nature, and its rich heterogeneous and deep contents. In this paper, we consider the problem of archiving a linked set of web objects into web containers in such a way as to minimize the number of containers accessed during a typical browsing session. We develop a method that makes use of the notion of PageRank and optimized graph partitioning to enable faster browsing of archived web contents. We include simulation results that illustrate the performance of our scheme and compare it to the common scheme currently used to organize web objects into web containers. %B 8th International Web Archiving Workshop, Aarhus, Denmark. %8 2008/// %G eng %0 Conference Paper %B Proceedings of the 8th ACM SIGCOMM conference on Internet measurement %D 2008 %T Fast monitoring of traffic subpopulations %A Ramachandran,Anirudh %A Seetharaman,Srinivasan %A Feamster, Nick %A Vazirani,Vijay %K counters %K flexsample %K sampling %K traffic statistics %K traffic subpopulations %X Network accounting, forensics, security, and performance monitoring applications often need to examine detailed traces from subsets of flows ("subpopulations"), where the application desires flexibility in specifying the subpopulation (e.g., to detect a portscan, the application must observe many packets between a source and a destination with one packet to each port). However, the dynamism and volume of network traffic on many high-speed links necessitates traffic sampling, which adversely affects subpopulation monitoring: because many subpopulations of interest to operators are low-volume flows, conventional sampling schemes (e.g., uniform random sampling) miss much of the subpopulation's traffic. Today's routers and network devices provide scant support for monitoring specific traffic subpopulations. This paper presents the design, implementation, and evaluation of FlexSample, a traffic monitoring engine that dynamically extracts traffic from subpopulations that operators define using conditions on packet header fields. FlexSample uses a fast, flexible counter array to provide rough estimates of packets' membership in respective subpopulations. Based on these coarse estimates, FlexSample then makes per-packet sampling decisions to sample proportionately from each subpopulation (as specified by a network operator), subject to an overall sampling constraint. We apply FlexSample to extract subpopulations such as port scans and traffic to high-degree nodes and find that it is able to capture significantly more packets from these subpopulations than conventional approaches. %B Proceedings of the 8th ACM SIGCOMM conference on Internet measurement %S IMC '08 %I ACM %C New York, NY, USA %P 257 - 270 %8 2008/// %@ 978-1-60558-334-1 %G eng %U http://doi.acm.org/10.1145/1452520.1452551 %R 10.1145/1452520.1452551 %0 Conference Paper %B Data Engineering, 2008. ICDE 2008. IEEE 24th International Conference on %D 2008 %T A Fast Similarity Join Algorithm Using Graphics Processing Units %A Lieberman,M.D. %A Sankaranarayanan,J. %A Samet, Hanan %K algorithm;graphics %K algorithms;quadtrees;set %K arrays %K curves;parallel %K join %K problem;space-filling %K PROCESSING %K searching;fast %K similarity %K theory;sorting; %K units;sort-and-search %X A similarity join operation A BOWTIEepsiv B takes two sets of points A, B and a value epsiv isin Ropf, and outputs pairs of points p isin A,q isin B, such that the distance D(p, q) les epsiv. Similarity joins find use in a variety of fields, such as clustering, text mining, and multimedia databases. A novel similarity join algorithm called LSS is presented that executes on a graphics processing unit (GPU), exploiting its parallelism and high data throughput. As GPUs only allow simple data operations such as the sorting and searching of arrays, LSS uses these two operations to cast a similarity join operation as a GPU sort-and-search problem. It first creates, on the fly, a set of space-filling curves on one of its input datasets, using a parallel GPU sort routine. Next, LSS processes each point p of the other dataset in parallel. For each p, it searches an interval of one of the space-filling curves guaranteed to contain all the pairs in which p participates. Using extensive theoretical and experimental analysis, LSS is shown to offer a good balance between time and work efficiency. Experimental results demonstrate that LSS is suitable for similarity joins in large high-dimensional datasets, and that it performs well when compared against two existing prominent similarity join methods. %B Data Engineering, 2008. ICDE 2008. IEEE 24th International Conference on %P 1111 - 1120 %8 2008/04// %G eng %R 10.1109/ICDE.2008.4497520 %0 Conference Paper %B Proceedings of the 9th international conference on Passive and active network measurement %D 2008 %T On the fidelity of 802.11 packet traces %A Schulman,A. %A Levin,D. %A Spring, Neil %B Proceedings of the 9th international conference on Passive and active network measurement %P 132 - 141 %8 2008/// %G eng %0 Book Section %B Computer Safety, Reliability, and Security %D 2008 %T Finding Corrupted Computers Using Imperfect Intrusion Prevention System Event Data %A Chrun,Danielle %A Michel Cukier %A Sneeringer,Gerry %E Harrison,Michael %E Sujan,Mark-Alexander %K Computer science %X With the increase of attacks on the Internet, a primary concern for organizations is how to protect their network. The objectives of a security team are 1) to prevent external attackers from launching successful attacks against organization computers that could become compromised, 2) to ensure that organization computers are not vulnerable (e.g., fully patched) so that in either case the organization computers do not start launching attacks. The security team can monitor and block malicious activity by using devices such as intrusion prevention systems. However, in large organizations, such monitoring devices could record a high number of events. The contributions of this paper are 1) to introduce a method that ranks potentially corrupted computers based on imperfect intrusion prevention system event data, and 2) to evaluate the method based on empirical data collected at a large organization of about 40,000 computers. The evaluation is based on the judgment of a security expert of which computers were indeed corrupted. On the one hand, we studied how many computers classified as of high concern or of concern were indeed corrupted (i.e., true positives). On the other hand, we analyzed how many computers classified as of lower concern were in fact corrupted (i.e., false negatives). %B Computer Safety, Reliability, and Security %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5219 %P 221 - 234 %8 2008/// %@ 978-3-540-87697-7 %G eng %U http://www.springerlink.com/content/4767u3017r613822/abstract/ %0 Report %D 2008 %T Fishing for Phishing from the Network Stream %A Ramachandran,A. %A Feamster, Nick %A Krishnamurthy,B. %A Spatscheck,O. %A Van der Merwe,J. %X Phishing is an increasingly prevalent social-engineering attack that attempts identity theft using spoofed Web pages of legitimate organizations. Unfortunately, current phishing detection methods are neither complete nor responsive because they rely on user reports, and many also require clientside software. Anti-phishing techniques could be more effective if they (1) could detect phishing attacks automatically from the network traffic; (2) could operate without cooperation from end-users. This paper performs a preliminary study to determine the feasibility of detecting phishing attacks in real-time, from the network traffic stream itself. We develop a model to identify the stages where in-network phishing detection is feasible and the data sources that can be analyzed to provide relevant information at each stage. Based on this model, we develop and evaluate a detection method based on features that exist in the network traffic it- self and are correlated with confirmed phishing attacks. %I Georgia Institute of Technology %V GT-CS-08-08 %8 2008/// %G eng %U http://hdl.handle.net/1853/25463 %0 Conference Paper %B Proceedings of the 8th ACM SIGCOMM conference on Internet measurement %D 2008 %T Fixing ally's growing pains with velocity modeling %A Bender,Adam %A Sherwood,Rob %A Spring, Neil %K alias resolution %K ally %K ip identifier %K velocity modeling %X Mapping the router topology is an important component of Internet measurement. Alias resolution, the process of mapping IP addresses to routers, is critical to accurate Internet mapping. Ally, a popular alias resolution tool, was developed to resolve aliases in individual ISPs, but its probabilistic accuracy and need to send O(n2) probes to infer aliases among n IP addresses make it unappealing for large-scale Internet mapping. In this paper, we present RadarGun, a tool that uses IP identifier velocity modeling to improve the accuracy and scalability of the Ally-based resolution technique. We provide analytical bounds on Ally's accuracy and validate our predicted aliases against Ally. Additionally, we show that velocity modeling requires only O(n) probes and thus scales to Internet-sized mapping efforts. %B Proceedings of the 8th ACM SIGCOMM conference on Internet measurement %S IMC '08 %I ACM %C New York, NY, USA %P 337 - 342 %8 2008/// %@ 978-1-60558-334-1 %G eng %U http://doi.acm.org/10.1145/1452520.1452560 %R 10.1145/1452520.1452560 %0 Conference Paper %B Engineering of Complex Computer Systems, 2008. ICECCS 2008. 13th IEEE International Conference on %D 2008 %T A Framework for Software Engineering Experimental Replications %A Mendonca,M.G. %A Maldonado,J.C. %A de Oliveira,M.C.F. %A Carver, J. %A Fabbri,C.P.F. %A Shull, F. %A Travassos,G.H. %A Hohn,E.N. %A Basili, Victor R. %K engineering %K engineering; %K evidence;experimental %K experimental %K Experiments;experimental %K for %K Framework %K Improving %K knowledge %K of %K replication %K replication;software %K sharing;software %K the %K transfer;knowledge %X Experimental replications are very important to the advancement of empirical software engineering. Replications are one of the key mechanisms to confirm previous experimental findings. They are also used to transfer experimental knowledge, to train people, and to expand a base of experimental evidence. Unfortunately, experimental replications are difficult endeavors. It is not easy to transfer experimental know-how and experimental findings. Based on our experience, this paper discusses this problem and proposes a Framework for Improving the Replication of Experiments (FIRE). The FIRE addresses knowledge sharing issues both at the intra-group (internal replications) and inter-group (external replications) levels. It encourages coordination of replications in order to facilitate knowledge transfer for lower cost, higher quality replications and more generalizable results. %B Engineering of Complex Computer Systems, 2008. ICECCS 2008. 13th IEEE International Conference on %P 203 - 212 %8 2008/04/31/3 %G eng %R 10.1109/ICECCS.2008.38 %0 Journal Article %J PLoS ONEPLoS ONE %D 2008 %T Functional Diversification of Paralogous Transcription Factors via Divergence in DNA Binding Site Motif and in Expression %A Singh,Larry N. %A Hannenhalli, Sridhar %X Gene duplication is a major driver of evolutionary innovation as it allows for an organism to elaborate its existing biological functions via specialization or diversification of initially redundant gene paralogs. Gene function can diversify in several ways. Transcription factor gene paralogs in particular, can diversify either by changes in their tissue-specific expression pattern or by changes in the DNA binding site motif recognized by their protein product, which in turn alters their gene targets. The relationship between these two modes of functional diversification of transcription factor paralogs has not been previously investigated, and is essential for understanding adaptive evolution of transcription factor gene families.Based on a large set of human paralogous transcription factor pairs, we show that when the DNA binding site motifs of transcription factor paralogs are similar, the expressions of the genes that encode the paralogs have diverged, so in general, at most one of the paralogs is highly expressed in a tissue. Moreover, paralogs with diverged DNA binding site motifs tend to be diverged in their function. Conversely, two paralogs that are highly expressed in a tissue tend to have dissimilar DNA binding site motifs. We have also found that in general, within a paralogous family, tissue-specific decrease in gene expression is more frequent than what is expected by chance. While previous investigations of paralogous gene diversification have only considered coding sequence divergence, by explicitly quantifying divergence in DNA binding site motif, our work presents a new paradigm for investigating functional diversification. Consistent with evolutionary expectation, our quantitative analysis suggests that paralogous transcription factors have survived extinction in part, either through diversification of their DNA binding site motifs or through alterations in their tissue-specific expression levels. %B PLoS ONEPLoS ONE %V 3 %P e2345 - e2345 %8 2008/06/04/ %G eng %U UR - http://dx.plos.org/10.1371/journal.pone.0002345,http://dx.plos.org/10.1371/journal.pone.0002345 %N 6 %R 10.1371/journal.pone.0002345 %0 Journal Article %J Evolution %D 2008 %T A GENEALOGICAL APPROACH TO QUANTIFYING LINEAGE DIVERGENCE %A Cummings, Michael P. %A Neel,Maile C %A Shaw,Kerry L %K Ancestral polymorphism %K congruence %K exclusivity %K genealogy %K lineage sorting %K monophyly %K paraphyly %K Phylogeny %K polyphyly %K speciation %K species %X We introduce a statistic, the genealogical sorting index (gsi), for quantifying the degree of exclusive ancestry of labeled groups on a rooted genealogy and demonstrate its application. The statistic is simple, intuitive, and easily calculated. It has a normalized range to facilitate comparisons among different groups, trees, or studies and it provides information on individual groups rather than a composite measure for all groups. It naturally handles polytomies and accommodates measures of uncertainty in phylogenetic relationships. We use coalescent simulations to explore the behavior of the gsi across a range of divergence times, with the mean value increasing to 1, the maximum value when exclusivity within a group reached monophyly. Simulations also demonstrate that the power to reject the null hypothesis of mixed genealogical ancestry increased markedly as sample size increased, and that the gsi provides a statistically more powerful measure of divergence than FST. Applications to data from published studies demonstrated that the gsi provides a useful way to detect significant exclusivity even when groups are not monophyletic. Although we describe this statistic in the context of divergence, it is more broadly applicable to quantify and assess the significance of clustering of observations in labeled groups on any tree. %B Evolution %V 62 %P 2411 - 2422 %8 2008/09/01/ %@ 1558-5646 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1558-5646.2008.00442.x/abstract %N 9 %R 10.1111/j.1558-5646.2008.00442.x %0 Journal Article %J Genome Biology %D 2008 %T Genome assembly forensics: finding the elusive mis-assembly %A Phillippy,Adam M %A Schatz,Michael C %A Pop, Mihai %X We present the first collection of tools aimed at automated genome assembly validation. This work formalizes several mechanisms for detecting mis-assemblies, and describes their implementation in our automated validation pipeline, called amosvalidate. We demonstrate the application of our pipeline in both bacterial and eukaryotic genome assemblies, and highlight several assembly errors in both draft and finished genomes. The software described is compatible with common assembly formats and is released, open-source, at http://amos.sourceforge.net. %B Genome Biology %V 9 %P R55 - R55 %8 2008/03/14/ %@ 1465-6906 %G eng %U http://genomebiology.com/2008/9/3/R55 %N 3 %R 10.1186/gb-2008-9-3-r55 %0 Journal Article %J PLoS ONEPLoS ONE %D 2008 %T Genome-Wide Analysis of Natural Selection on Human Cis-Elements %A Sethupathy,Praveen %A Giang,Hoa %A Plotkin,Joshua B. %A Hannenhalli, Sridhar %X It has been speculated that the polymorphisms in the non-coding portion of the human genome underlie much of the phenotypic variability among humans and between humans and other primates. If so, these genomic regions may be undergoing rapid evolutionary change, due in part to natural selection. However, the non-coding region is a heterogeneous mix of functional and non-functional regions. Furthermore, the functional regions are comprised of a variety of different types of elements, each under potentially different selection regimes.Using the HapMap and Perlegen polymorphism data that map to a stringent set of putative binding sites in human proximal promoters, we apply the Derived Allele Frequency distribution test of neutrality to provide evidence that many human-specific and primate-specific binding sites are likely evolving under positive selection. We also discuss inherent limitations of publicly available human SNP datasets that complicate the inference of selection pressures. Finally, we show that the genes whose proximal binding sites contain high frequency derived alleles are enriched for positive regulation of protein metabolism and developmental processes. Thus our genome-scale investigation provides evidence for positive selection on putative transcription factor binding sites in human proximal promoters. %B PLoS ONEPLoS ONE %V 3 %P e3137 - e3137 %8 2008/// %G eng %U UR - http://dx.doi.org/10.1371/journal.pone.0003137,http://dx.doi.org/10.1371/journal.pone.0003137 %N 9 %R 10.1371/journal.pone.0003137 %0 Journal Article %J Tropical Plant Biology %D 2008 %T Genome-Wide Analysis of Repetitive Elements in Papaya %A Nagarajan,Niranjan %A Navajas-Pérez,Rafael %A Pop, Mihai %A Alam,Maqsudul %A Ming,Ray %A Paterson,Andrew %A Salzberg,Steven %X Papaya ( Carica papaya L.) is an important fruit crop cultivated in tropical and subtropical regions worldwide. A first draft of its genome sequence has been recently released. Together with Arabidopsis , rice, poplar, grapevine and other genomes in the pipeline, it represents a good opportunity to gain insight into the organization of plant genomes. Here we report a detailed analysis of repetitive elements in the papaya genome, including transposable elements (TEs), tandemly-arrayed sequences, and high copy number genes. These repetitive sequences account for ∼56% of the papaya genome with TEs being the most abundant at 52%, tandem repeats at 1.3% and high copy number genes at 3%. Most common types of TEs are represented in the papaya genome with retrotransposons being the dominant class, accounting for 40% of the genome. The most prevalent retrotransposons are Ty3-gypsy (27.8%) and Ty1-copia (5.5%). Among the tandem repeats, microsatellites are the most abundant in number, but represent only 0.19% of the genome. Minisatellites and satellites are less abundant, but represent 0.68% and 0.43% of the genome, respectively, due to greater repeat length. Despite an overall smaller gene repertoire in papaya than many other angiosperms, a significant fraction of genes (>2%) are present in large gene families with copy number greater than 20. This repeat database clarified a major part of the papaya genome organization and partly explained the lower gene repertoire in papaya than in Arabidopsis . %B Tropical Plant Biology %V 1 %P 191 - 201 %8 2008/// %@ 1935-9756 %G eng %U http://dx.doi.org/10.1007/s12042-008-9015-0 %N 3 %0 Journal Article %J networks %D 2008 %T Group proximity measure for recommending groups in online social networks %A Saha,B. %A Getoor, Lise %X There are currently, thousands of online social networks (OSN)available, each of which hosts millions of users. Users make new friendship links, join groups on diverse topics, share opinions and thus help in building a big knowledge repository. In this paper we study the problem of defining proximity measure between groups (communities) of OSN. Understanding the proximity among the groups in OSN can reveal new insights into the structure of the network. We use this proximity measure in a novel way with other structural features of online groups to predict the evolving state of the network. One of the important task in the sphere of OSN is helping users in selecting groups of their interest, by making rec- ommendations. We utilize group proximity measure to propose a new framework for recommending groups to users of OSN. Finally we investigate the effectiveness of our methods in three real OSN: Flickr, Live Journal and You Tube. Our methods give promising results. %B networks %V 1 %P 5 - 5 %8 2008/// %G eng %N 6 %0 Journal Article %J IEEE Transactions on Image Processing %D 2008 %T Hardware and Software Systems for Image and Video Processing-Algorithmic and Architectural Optimizations for Computationally Efficient Particle Filtering %A Sankaranarayanan,A. C %A Srivastava, A. %A Chellapa, Rama %B IEEE Transactions on Image Processing %V 17 %P 737 - 748 %8 2008/// %G eng %N 5 %0 Conference Paper %B SIGCHI EA '08 %D 2008 %T Hci for community and international development %A Thomas, John %A Dearden, Andy %A Dray, Susan %A Light, Ann %A Best, Michael %A Arkin, Nuray %A Maunder, Andrew %A Kam, Mathew %A Marshini Chetty %A Sambasivan, Nithya %A Buckhalter, Celeste %A Krishnan, Gaurishankar %K community design %K ict4d %K information and communication technology %K international development %K participatory design %K ucd4id %K User centered design %X This workshop explores the challenges in applying, extending and inventing appropriate methods and contributions of Human Computer Interaction (HCI) to International economic and community Development. We address interaction design for parts of the world that are often marginalized by the Global North as well as people in the Global North who are themselves similarly marginalized by poverty or other barriers. We hope to extend the boundaries of the field of Human Computer Interaction by spurring a discussion on how existing methods and practices can be adapted and modified, and how new practices can be developed, to deal with the unique challenges posed by these contexts. %B SIGCHI EA '08 %S CHI EA '08 %I ACM %P 3909 - 3912 %8 2008/// %@ 978-1-60558-012-8 %G eng %U http://doi.acm.org/10.1145/1358628.1358954 %0 Journal Article %J Geospatial Services and Applications for the Internet %D 2008 %T Hierarchical infrastructure for internet mapping services %A Brabec,F. %A Samet, Hanan %B Geospatial Services and Applications for the Internet %P 1 - 30 %8 2008/// %G eng %R 10.1007/978-0-387-74674-6_1 %0 Conference Paper %B Similarity Search and Applications, 2008. SISAP 2008. First International Workshop on %D 2008 %T High-Dimensional Similarity Retrieval Using Dimensional Choice %A Tahmoush,D. %A Samet, Hanan %K data %K database %K function;high-dimensional %K management %K processing; %K processing;sequential %K reduction;database %K reduction;distance %K retrieval;query %K search;data %K search;similarity %K similarity %K system;dimension %K systems;query %X There are several pieces of information that can be utilized in order to improve the efficiency of similarity searches on high-dimensional data. The most commonly used information is the distribution of the data itself but the use of dimensional choice based on the information in the query as well as the parameters of the distribution can provide an effective improvement in the query processing speed and storage. The use of this method can produce dimension reduction by as much as a factor of n, the number of data points in the database, over sequential search. We demonstrate that the curse of dimensionality is not based on the dimension of the data itself, but primarily upon the effective dimension of the distance function. We also introduce a new distance function that utilizes fewer dimensions of the higher dimensional space to produce a maximal lower bound distance in order to approximate the full distance function. This work has demonstrated significant dimension reduction, up to 70% reduction with an improvement in accuracy or over 99% with only a 6% loss in accuracy on a prostate cancer data set. %B Similarity Search and Applications, 2008. SISAP 2008. First International Workshop on %P 35 - 42 %8 2008/04// %G eng %R 10.1109/SISAP.2008.20 %0 Report %D 2008 %T How Do Users Find Things with PubMed? Towards Automatic Utility Evaluation with User Simulations %A Jimmy Lin %A Smucker,Mark D. %K *BIOMEDICAL INFORMATION SYSTEMS %K *BROWSING %K *INFORMATION RETRIEVAL %K *PUBMED %K *SIMILARITY %K *SIMULATION %K *USER NEEDS %K DOCUMENTS %K FIND-SIMILAR %K HUMAN FACTORS ENGINEERING & MAN MACHINE SYSTEM %K INFORMATION SCIENCE %K MAN COMPUTER INTERFACE %K QUALITY %K queries %K searching %K TEST AND EVALUATION %K UTILITY %X In the context of document retrieval in the biomedical domain, this paper explores the complex relationship between the quality of initial query results and the overall utility of an interactive system. We demonstrate that a content-similarity browsing tool can compensate for poor retrieval results, and that the relationship between retrieval performance and overall utility is non-linear. Arguments are advanced with user simulations, which characterize the relevance of documents that a user might encounter with different browsing strategies. With broader implications to IR, this work provides a case study of how user simulations can be exploited as a formative tool for automatic utility evaluation. Simulation-based studies provide researchers with an additional evaluation tool to complement interactive and Cranfield-style experiments. %I Human-Computer Interaction Lab, University of Maryland, College Park %8 2008/02// %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA478703 %0 Journal Article %J University of Washington CSE Tech Report UW-CSE-08-04-02 %D 2008 %T ilearn on the iphone: Real-time human activity classification on commodity mobile phones %A Saponas,T. %A Lester,J. %A Jon Froehlich %A Fogarty,J. %A Landay,J. %B University of Washington CSE Tech Report UW-CSE-08-04-02 %8 2008 %G eng %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on %D 2008 %T Image acquisition forensics: Forensic analysis to identify imaging source %A McKay,C. %A Swaminathan,A. %A Gou,Hongmei %A M. Wu %K ACQUISITION %K acquisition;image %K analysis; %K analysis;image %K analysis;interpolation;statistical %K cameras;color %K cell %K coefficients;computer %K colour %K editing %K forensics;image %K graphics;digital %K identification;noise %K images;forensic %K Interpolation %K phone %K processing;data %K softwares;imaging %K source %K statistics;scanners;signal %X With widespread availability of digital images and easy-to-use image editing softwares, the origin and integrity of digital images has become a serious concern. This paper introduces the problem of image acquisition forensics and proposes a fusion of a set of signal processing features to identify the source of digital images. Our results show that the devices' color interpolation coefficients and noise statistics can jointly serve as good forensic features to help accurately trace the origin of the input image to its production process and to differentiate between images produced by cameras, cell phone cameras, scanners, and computer graphics. Further, the proposed features can also be extended to determining the brand and model of the device. Thus, the techniques introduced in this work provide a unified framework for image acquisition forensics. %B Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on %P 1657 - 1660 %8 2008/04/31/4 %G eng %R 10.1109/ICASSP.2008.4517945 %0 Conference Paper %B 5th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks, 2008. SECON '08 %D 2008 %T Immunity-Based Epidemic Routing in Intermittent Networks %A Mundur, Padma %A Seligman,M. %A Jin Na Lee %K Analytical models %K Delay %K delivered messages %K Disruption tolerant networking %K Educational institutions %K immunity-based epidemic routing %K information dissemination %K intermittent networks %K Mobile ad hoc networks %K Network topology %K Performance analysis %K Resource management %K resource utilization %K routing protocol %K routing protocols %K telecommunication network topology %X In this research, we propose to modify and extend epidemic routing used in intermittent networks. In particular, we propose to include immunity-based information disseminated in the reverse once messages get delivered to their destination. The goal is to design a more efficient routing protocol in terms of resource utilization. The idea is to analyze and evaluate the network performance using an immunity scheme in the context of epidemic routing and its variants. The reverse dissemination of such information requires minimal resources and the tradeoff in timely purging of delivered messages can be significant. We are using ns2 to implement a detailed simulation of the proposed immunity-based epidemic routing. %B 5th Annual IEEE Communications Society Conference on Sensor, Mesh and Ad Hoc Communications and Networks, 2008. SECON '08 %I IEEE %P 609 - 611 %8 2008/06/16/20 %@ 978-1-4244-1777-3 %G eng %R 10.1109/SAHCN.2008.86 %0 Book Section %B Information Systems SecurityInformation Systems Security %D 2008 %T Implicit Flows: Can’t Live with ‘Em, Can’t Live without ‘Em %A King,Dave %A Hicks,Boniface %A Hicks, Michael W. %A Jaeger,Trent %E Sekar,R. %E Pujari,Arun %X Verifying that programs trusted to enforce security actually do so is a practical concern for programmers and administrators. However, there is a disconnect between the kinds of tools that have been successfully applied to real software systems (such as taint mode in Perl and Ruby), and information-flow compilers that enforce a variant of the stronger security property of noninterference. Tools that have been successfully used to find security violations have focused on explicit flows of information, where high-security information is directly leaked to output. Analysis tools that enforce noninterference also prevent implicit flows of information, where high-security information can be inferred from a program’s flow of control. However, these tools have seen little use in practice, despite the stronger guarantees that they provide. To better understand why, this paper experimentally investigates the explicit and implicit flows identified by the standard algorithm for establishing noninterference. When applied to implementations of authentication and cryptographic functions, the standard algorithm discovers many real implicit flows of information, but also reports an extremely high number of false alarms, most of which are due to conservative handling of unchecked exceptions (e.g., null pointer exceptions). After a careful analysis of all sources of true and false alarms, due to both implicit and explicit flows, the paper concludes with some ideas to improve the false alarm rate, toward making stronger security analysis more practical. %B Information Systems SecurityInformation Systems Security %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5352 %P 56 - 70 %8 2008/// %@ 978-3-540-89861-0 %G eng %U http://dx.doi.org/10.1007/978-3-540-89862-7_4 %0 Conference Paper %B Proceedings of the nineteenth annual ACM-SIAM symposium on Discrete algorithms %D 2008 %T Improved algorithmic versions of the Lovász Local Lemma %A Srinivasan, Aravind %X The Lovász Local Lemma is a powerful tool in combinatorics and computer science. The original version of the lemma was nonconstructive, and efficient algorithmic versions have been developed by Beck, Alon, Molloy & Reed, et al. In particular, the work of Molloy & Reed lets us automatically extract efficient versions of essentially any application of the symmetric version of the Local Lemma. However, with some exceptions, there is a significant gap between what one can prove using the original Lemma nonconstructively, and what is possible through these efficient versions; also, some of these algorithmic versions run in super-polynomial time. Here, we lessen this gap, and improve the running time of all these applications (which cover all applications in the Molloy & Reed framework) to polynomial. We also improve upon the parallel algorithmic version of the Local Lemma for hypergraph coloring due to Alon, by allowing noticeably more overlap among the edges. %B Proceedings of the nineteenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '08 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 611 - 620 %8 2008/// %G eng %U http://dl.acm.org/citation.cfm?id=1347082.1347150 %0 Conference Paper %B Proc. 2008 Intl. Conference on Knowledge Representation and Reasoning (KR 2008) %D 2008 %T Inconsistency management policies %A Martinez,M. V %A Parisi,F. %A Pugliese, A. %A Simari,G. I %A V.S. Subrahmanian %X Though there is much work on how inconsistency indatabases should be managed, there is good reason to believe that end users will want to bring their domain expertise and needs to bear in how to deal with inconsistencies. In this paper, we propose the concept of inconsistency management policies (IMPs). We show that IMPs are rich enough to spec- ify many types of inconsistency management methods pro- posed previously, but provide end users with tools that allow them to use the policies that they want. Our policies are also capable of allowing inconsistency to persist in the database or of eliminating more than a minimal subset of tuples involved in the inconsistency. We present a formal axiomatic definition of IMPs and present appropriate complexity results, together with results linking different IMPs together. We extend the relational algebra (RA) to incorporate IMPs and present the- oretical results showing how IMPs and classical RA operators interact. %B Proc. 2008 Intl. Conference on Knowledge Representation and Reasoning (KR 2008) %8 2008/// %G eng %0 Conference Paper %B Proceedings of the 16th ACM SIGSPATIAL international conference on Advances in geographic information systems %D 2008 %T Indexing planar point quartets via geometric attributes %A Cranston,Charles B. %A Samet, Hanan %X An index is devised to support position-independent search for images containing quartets of icons. Point quartets exist that do not unambiguously map to quadrilaterals, however, four points do unambiguously determine a set of six interpoint line segments. Values for the "size", "shape", and "orientation" attributes of an icon quartet can be derived as functions of this interpoint line set, and can be used to construct a point-based index, in which each point quartet maps to a single point in the resulting hyperdimensional index space. Orientation can be represented by a single, spatially closed dimension. However, assignment of a reference direction for quartets possessing a k-fold rotational symmetry presents a significant challenge. Methods are described for determining shape and orientation attributes for point quartets, and for mapping these attributes onto a set of attribute axes to form a combined index. The orientation computation supplies, as a byproduct, one component of the shape attribute. All attributes are continuous with respect to small variations in the indexed point quartets. %B Proceedings of the 16th ACM SIGSPATIAL international conference on Advances in geographic information systems %S GIS '08 %I ACM %C New York, NY, USA %P 71:1–71:4 - 71:1–71:4 %8 2008/// %@ 978-1-60558-323-5 %G eng %U http://doi.acm.org/10.1145/1463434.1463516 %R 10.1145/1463434.1463516 %0 Journal Article %J Model-Driven Development of Reliable Automotive Services %D 2008 %T An instrumentation-based approach to controller model validation %A Cleaveland, Rance %A Smolka,S. %A Sims,S. %B Model-Driven Development of Reliable Automotive Services %P 84 - 97 %8 2008/// %G eng %0 Conference Paper %B Grid Computing, 2008 9th IEEE/ACM International Conference on %D 2008 %T Integrating categorical resource types into a P2P desktop grid system %A Kim,Jik-Soo %A Nam,Beomseok %A Marsh,M. %A Keleher,P. %A Bhattacharjee, Bobby %A Sussman, Alan %K computing;peer-to-peer %K computing;protocols; %K computing;protocols;scalability %K desktop %K Grid %K grid;distributed %K grid;peer-to-peer %K N-dimensional %K resource %K space;P2P %K system;categorical %K technique;grid %K types;decentralized %X We describe and evaluate a set of protocols that implement a distributed, decentralized desktop grid. Incoming jobs are matched with system nodes through proximity in an N-dimensional resource space. This work improves on prior work by (1) efficiently accommodating node and job characterizations that include both continuous and categorical resource types, and (2) scaling gracefully to large system sizes even with highly non-uniform distributions of job and node types. We use extensive simulation results to show that the resulting system handles both continuous and categorical constraints efficiently, and that the new scalability techniques are effective. %B Grid Computing, 2008 9th IEEE/ACM International Conference on %P 284 - 291 %8 2008/10/29/1 %G eng %R 10.1109/GRID.2008.4662810 %0 Conference Paper %B Proceedings of the twenty-sixth annual SIGCHI conference on Human factors in computing systems %D 2008 %T Integrating statistics and visualization: case studies of gaining clarity during exploratory data analysis %A Perer,Adam %A Shneiderman, Ben %K case studies %K Evaluation %K exploratory data analysis %K Information Visualization %K social networks %K Statistics %X Although both statistical methods and visualizations have been used by network analysts, exploratory data analysis remains a challenge. We propose that a tight integration of these technologies in an interactive exploratory tool could dramatically speed insight development. To test the power of this integrated approach, we created a novel social network analysis tool, SocialAction, and conducted four long-term case studies with domain experts, each working on unique data sets with unique problems. The structured replicated case studies show that the integrated approach in SocialAction led to significant discoveries by a political analyst, a bibliometrician, a healthcare consultant, and a counter-terrorism researcher. Our contributions demonstrate that the tight integration of statistics and visualizations improves exploratory data analysis, and that our evaluation methodology for long-term case studies captures the research strategies of data analysts. %B Proceedings of the twenty-sixth annual SIGCHI conference on Human factors in computing systems %S CHI '08 %I ACM %C New York, NY, USA %P 265 - 274 %8 2008/// %@ 978-1-60558-011-1 %G eng %U http://doi.acm.org/10.1145/1357054.1357101 %R 10.1145/1357054.1357101 %0 Journal Article %J ACM Transactions on Computer-Human Interaction %D 2008 %T Interactive auditory data exploration: A framework and evaluation with geo-referenced data sonification %A Zhao,H. %A Plaisant, Catherine %A Shneiderman, Ben %A Lazar,J. %X We describe an Action-by-Design-Component (ADC) framework to guide auditory interface designers for exploratory data analysis. The framework characterizes data interaction in the auditory mode as a set of Auditory Information Seeking Actions (AISAs). Contrasting AISAs with actions in visualizations, the framework also discusses design considerations for a set of Design Components to support AISAs. Applying the framework to geo-referenced data, we systematically explored and evaluated its design space. A data exploration tool, iSonic, was built for blind users. In depth case studies with 7 blind users, with over 42 hours of data collection, showed that iSonic enabled them to find facts and discover trends of geo-referenced data, even in unfamiliar geographical contexts, without special devices. The results also showed that blind users dramatically benefited from the rich set of task-oriented actions (AISAs) and the use of multiple highly coordinated data views provided by the ADC framework. Some widely used techniques in visualization, with appropriate adaptation, also work in the auditory mode. The application of the framework to scatterplots shows that the framework can be generalized and lead to the design of a unified auditory workspace for general exploratory data analysis. Readers can view a supplementary video demonstration of iSonic by visiting www.cs.umd.edu/hcil/iSonic/ %B ACM Transactions on Computer-Human Interaction %8 2008/// %G eng %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2008 %T Interactive Entity Resolution in Relational Data: A Visual Analytic Tool and Its Evaluation %A Kang,Hyunmo %A Getoor, Lise %A Shneiderman, Ben %A Bilgic,M. %A Licamele,L. %K algorithms %K Computer Graphics %K D-Dupe %K data visualisation %K database management systems %K Databases, Factual %K graphical user interface %K Graphical user interfaces %K human-centered computing %K Image Interpretation, Computer-Assisted %K Information Storage and Retrieval %K Information Visualization %K interactive entity resolution %K relational context visualization %K Relational databases %K relational entity resolution algorithm %K User interfaces %K user-centered design %K User-Computer Interface %K visual analytic tool %X Databases often contain uncertain and imprecise references to real-world entities. Entity resolution, the process of reconciling multiple references to underlying real-world entities, is an important data cleaning process required before accurate visualization or analysis of the data is possible. In many cases, in addition to noisy data describing entities, there is data describing the relationships among the entities. This relational data is important during the entity resolution process; it is useful both for the algorithms which determine likely database references to be resolved and for visual analytic tools which support the entity resolution process. In this paper, we introduce a novel user interface, D-Dupe, for interactive entity resolution in relational data. D-Dupe effectively combines relational entity resolution algorithms with a novel network visualization that enables users to make use of an entity's relational context for making resolution decisions. Since resolution decisions often are interdependent, D-Dupe facilitates understanding this complex process through animations which highlight combined inferences and a history mechanism which allows users to inspect chains of resolution decisions. An empirical study with 12 users confirmed the benefits of the relational context visualization on the performance of entity resolution tasks in relational data in terms of time as well as users' confidence and satisfaction. %B IEEE Transactions on Visualization and Computer Graphics %V 14 %P 999 - 1014 %8 2008/10//Sept %@ 1077-2626 %G eng %N 5 %R 10.1109/TVCG.2008.55 %0 Conference Paper %B Proceedings of the Conference on Empirical Methods in Natural Language Processing %D 2008 %T Language and translation model adaptation using comparable corpora %A Snover,Matthew %A Dorr, Bonnie J %A Schwartz,Richard %X Traditionally, statistical machine translation systems have relied on parallel bi-lingual data to train a translation model. While bi-lingual parallel data are expensive to generate, monolingual data are relatively common. Yet monolingual data have been under-utilized, having been used primarily for training a language model in the target language. This paper describes a novel method for utilizing monolingual target data to improve the performance of a statistical machine translation system on news stories. The method exploits the existence of comparable text---multiple texts in the target language that discuss the same or similar stories as found in the source language document. For every source document that is to be translated, a large monolingual data set in the target language is searched for documents that might be comparable to the source documents. These documents are then used to adapt the MT system to increase the probability of generating texts that resemble the comparable document. Experimental results obtained by adapting both the language and translation models show substantial gains over the baseline system. %B Proceedings of the Conference on Empirical Methods in Natural Language Processing %S EMNLP '08 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 857 - 866 %8 2008/// %G eng %U http://dl.acm.org/citation.cfm?id=1613715.1613825 %0 Conference Paper %B Proceedings of the 2008 ACM workshop on Search in social media %D 2008 %T Leveraging social context for searching social media %A Smith,Marc %A Barash,Vladimir %A Getoor, Lise %A Lauw,Hady W. %K social context %K social media %K social search %X The ability to utilize and benefit from today's explosion of social media sites depends on providing tools that allow users to productively participate. In order to participate, users must be able to find resources (both people and information) that they find valuable. Here, we argue that in order to do this effectively, we should make use of a user's "social context". A user's social context includes both their personal social context (their friends and the communities to which they belong) and their community social context (their role and identity in different communities). %B Proceedings of the 2008 ACM workshop on Search in social media %S SSM '08 %I ACM %C New York, NY, USA %P 91 - 94 %8 2008/// %@ 978-1-60558-258-0 %G eng %U http://doi.acm.org/10.1145/1458583.1458602 %R 10.1145/1458583.1458602 %0 Journal Article %J Circuits and Systems for Video Technology, IEEE Transactions on %D 2008 %T Machine Recognition of Human Activities: A Survey %A Turaga,P. %A Chellapa, Rama %A V.S. Subrahmanian %A Udrea,O. %K activity;low-level %K annotation;human %K cameras;video %K content-based %K factors;image %K processing;machine %K processing;rate-invariant %K recognition;image %K recognition;midlevel %K representation;image %K representation;video %K representations;semantic %K retrieval;human %K sequences; %K summarization;content-based %K video %X The past decade has witnessed a rapid proliferation of video cameras in all walks of life and has resulted in a tremendous explosion of video content. Several applications such as content-based video annotation and retrieval, highlight extraction and video summarization require recognition of the activities occurring in the video. The analysis of human activities in videos is an area with increasingly important consequences from security and surveillance to entertainment and personal archiving. Several challenges at various levels of processing-robustness against errors in low-level processing, view and rate-invariant representations at midlevel processing and semantic representation of human activities at higher level processing-make this problem hard to solve. In this review paper, we present a comprehensive survey of efforts in the past couple of decades to address the problems of representation, recognition, and learning of human activities from video and related applications. We discuss the problem at two major levels of complexity: 1) "actions" and 2) "activities." "Actions" are characterized by simple motion patterns typically executed by a single human. "Activities" are more complex and involve coordinated actions among a small number of humans. We will discuss several approaches and classify them according to their ability to handle varying degrees of complexity as interpreted above. We begin with a discussion of approaches to model the simplest of action classes known as atomic or primitive actions that do not require sophisticated dynamical modeling. Then, methods to model actions with more complex dynamics are discussed. The discussion then leads naturally to methods for higher level representation of complex activities. %B Circuits and Systems for Video Technology, IEEE Transactions on %V 18 %P 1473 - 1488 %8 2008/11// %@ 1051-8215 %G eng %N 11 %R 10.1109/TCSVT.2008.2005594 %0 Journal Article %J Development %D 2008 %T Maternal depletion of CTCF reveals multiple functions during oocyte and preimplantation embryo development %A Wan,Le-Ben %A Pan,Hua %A Hannenhalli, Sridhar %A Cheng,Yong %A Ma,Jun %A Fedoriw,Andrew %A Lobanenkov,Victor %A Latham,Keith E. %A Schultz,Richard M. %A Bartolomei,Marisa S. %X CTCF is a multifunctional nuclear factor involved in epigenetic regulation. Despite recent advances that include the systematic discovery of CTCF-binding sites throughout the mammalian genome, the in vivo roles of CTCF in adult tissues and during embryonic development are largely unknown. Using transgenic RNAi, we depleted maternal stores of CTCF from growing mouse oocytes, and identified hundreds of misregulated genes. Moreover, our analysis suggests that CTCF predominantly activates or derepresses transcription in oocytes. CTCF depletion causes meiotic defects in the egg, and mitotic defects in the embryo that are accompanied by defects in zygotic gene expression, and culminate in apoptosis. Maternal pronuclear transfer and CTCF mRNA microinjection experiments indicate that CTCF is a mammalian maternal effect gene, and that persistent transcriptional defects rather than persistent chromosomal defects perturb early embryonic development. This is the first study detailing a global and essential role for CTCF in mouse oocytes and preimplantation embryos. %B Development %V 135 %P 2729 - 2738 %8 2008/// %G eng %U http://dev.biologists.org/content/135/16/2729.abstract %N 16 %R 10.1242/dev.024539 %0 Conference Paper %B Distributed Computing Systems, International Conference on %D 2008 %T Measurement Manipulation and Space Selection in Network Coordinates %A Lumezanu,Cristian %A Spring, Neil %X Internet coordinate systems have emerged as an efficient method to estimate the latency between pairs of nodes without any communication between them. However, most coordinate systems have been evaluated solely on data sets built by their authors from measurements gathered over large periods of time. Although they show good prediction results, it is unclear whether the accuracy is the result of the system design properties or is more connected to the characteristics of the data sets. In this paper, we revisit a simple question: how do the features of the embedding space and the inherent attributes of the data sets interact in producing good embeddings? We adapt the Vivaldi algorithm to use Hyperbolic space for embedding and evaluate both Euclidean and Hyperbolic Vivaldi on seven sets of real-world latencies. Our results show that node filtering and latency distributions can significantly influence the accuracy of the predictions. For example, although Euclidean Vivaldi performs well on data sets that were chosen, constructed and filtered by the designers of the algorithm, its performance and robustness decrease considerably when run on third party data sets that were not filtered a priori. Our results offer important insight into designing and building coordinate systems that are both robust and accurate in Internet-like environments. %B Distributed Computing Systems, International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %P 361 - 368 %8 2008/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/ICDCS.2008.27 %0 Journal Article %J Journal of Signal Processing Systems %D 2008 %T Memory-constrained block processing for dsp software optimization %A Ko,M. Y %A Shen,C. C %A Bhattacharyya, Shuvra S. %B Journal of Signal Processing Systems %V 50 %P 163 - 177 %8 2008/// %G eng %N 2 %0 Journal Article %J ACM Trans. Database Syst. %D 2008 %T Metric space similarity joins %A Jacox,Edwin H. %A Samet, Hanan %K distance-based indexing %K external memory algorithms %K nearest neighbor queries %K range queries %K Ranking %K Similarity join %X Similarity join algorithms find pairs of objects that lie within a certain distance ε of each other. Algorithms that are adapted from spatial join techniques are designed primarily for data in a vector space and often employ some form of a multidimensional index. For these algorithms, when the data lies in a metric space, the usual solution is to embed the data in vector space and then make use of a multidimensional index. Such an approach has a number of drawbacks when the data is high dimensional as we must eventually find the most discriminating dimensions, which is not trivial. In addition, although the maximum distance between objects increases with dimension, the ability to discriminate between objects in each dimension does not. These drawbacks are overcome via the introduction of a new method called Quickjoin that does not require a multidimensional index and instead adapts techniques used in distance-based indexing for use in a method that is conceptually similar to the Quicksort algorithm. A formal analysis is provided of the Quickjoin method. Experiments show that the Quickjoin method significantly outperforms two existing techniques. %B ACM Trans. Database Syst. %V 33 %P 7:1–7:38 - 7:1–7:38 %8 2008/06// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/1366102.1366104 %N 2 %R 10.1145/1366102.1366104 %0 Journal Article %J Proceedings of SPIE %D 2008 %T Mixed state models for automatic target recognition and behavior analysis in video sequences %A Chellapa, Rama %A Sankaranarayanan,Aswin C. %A Veeraraghavan,Ashok %X Mixed state or hybrid state space systems are useful tools for various problems in computer vision. These systems model complicated system dynamics as a mixture of inherently simple sub-systems, with an additional mechanism to switch between the sub-systems. This approach of modeling using simpler systems allows for ease in learning the parameters of the system and in solving the inference problem. In this paper, we study the use of such mixed state space systems for problems in recognition and behavior analysis in video sequences. We begin with a dynamical system formulation for recognition of faces from a video. This system is used to introduce the simultaneous tracking and recognition paradigm that allows for improved performance in both tracking and recognition. We extend this framework to design a second system for verification of vehicles across non-overlapping views using structural and textural fingerprints for characterizing the identity of the target. Finally, we show the use of such modeling for tracking and behavior analysis of bees from video. %B Proceedings of SPIE %V 6967 %P 69670Q-69670Q-11 - 69670Q-69670Q-11 %8 2008/04/03/ %@ 0277786X %G eng %U http://spiedigitallibrary.org/proceedings/resource/2/psisdg/6967/1/69670Q_1?isAuthorized=no %N 1 %R doi:10.1117/12.777766 %0 Report %D 2008 %T Model Based Design Verification: A Monitor Based Approach %A Ackermann,Chris %A Ray,Arnab %A Cleaveland, Rance %A Heit,Jürgen %A Martin,Christopher %A Shelton,Charles %X This paper assesses the utility of an automated functional verification methodology in the design of a body electronics application. The use of design models as detailed specifications for software is gaining steady acceptance in the automotive industry because of the advantages it offers in terms of rapid system prototyping and automatic code generation. This paper argues that such modeling effort can also be used to provide support for design verification, so that errors may be pinpointed before they find their way into code, when they become much harder and more expensive to isolate, diagnose and repair. %I SAE International %C Warrendale, PA %V 2008-01-0741 %8 2008/04// %G eng %U http://papers.sae.org/2008-01-0741 %0 Journal Article %J Pattern Analysis and Machine Intelligence, IEEE Transactions on %D 2008 %T Model Driven Segmentation of Articulating Humans in Laplacian Eigenspace %A Sundaresan, A. %A Chellapa, Rama %K 3D voxel data segmentation %K algorithms %K Anatomic;Pattern Recognition %K Artificial intelligence %K Automated;Reproducibility of Results;Sensitivity and Specificity;Whole Body Imaging; %K Computer simulation %K Computer-Assisted;Imaging %K curve fitting %K Eigenvalues and eigenfunctions %K graph theory %K human articulation %K human body graphical model %K human motion analysis %K HUMANS %K Image Enhancement %K Image Interpretation %K Image motion analysis %K image registration %K Image segmentation %K Laplace transforms %K Laplacian eigenspace transform %K model driven segmentation %K probability %K representative graph %K skeleton registration %K Smoothing methods %K solid modelling %K spline fitting error %K splines (mathematics) %K Three-Dimensional;Joints;Models %K top-down probabilistic approach %K voxel neighborhood graph %X We propose a general approach using Laplacian Eigenmaps and a graphical model of the human body to segment 3D voxel data of humans into different articulated chains. In the bottom-up stage, the voxels are transformed into a high-dimensional (6D or less) Laplacian Eigenspace (LE) of the voxel neighborhood graph. We show that LE is effective at mapping voxels on long articulated chains to nodes on smooth 1D curves that can be easily discriminated, and prove these properties using representative graphs. We fit 1D splines to voxels belonging to different articulated chains such as the limbs, head and trunk, and determine the boundary between splines using the spline fitting error. A top-down probabilistic approach is then used to register the segmented chains, utilizing their mutual connectivity and individual properties. Our approach enables us to deal with complex poses such as those where the limbs form loops. We use the segmentation results to automatically estimate the human body models. While we use human subjects in our experiments, the method is fairly general and can be applied to voxel-based segmentation of any articulated object composed of long chains. We present results on real and synthetic data that illustrate the usefulness of this approach. %B Pattern Analysis and Machine Intelligence, IEEE Transactions on %V 30 %P 1771 - 1785 %8 2008/10// %@ 0162-8828 %G eng %N 10 %R 10.1109/TPAMI.2007.70823 %0 Journal Article %J IEEE Transactions on Software Engineering %D 2008 %T Modular Information Hiding and Type-Safe Linking for C %A Srivastava,S. %A Hicks, Michael W. %A Foster, Jeffrey S. %A Jenkins,P. %K C language %K CMOD %K Code design %K Coding Tools and Techniques %K compiler %K data encapsulation %K Information hiding %K modular information hiding %K modular reasoning %K modules %K object-oriented programming %K open source programs %K packages %K program compilers %K public domain software %K reliability %K software reusability %K type-safe linking %X This paper presents CMod, a novel tool that provides a sound module system for C. CMod works by enforcing four rules that are based on principles of modular reasoning and on current programming practice. CMod's rules flesh out the convention that .h header files are module interfaces and .c source files are module implementations. Although this convention is well-known, existing explanations of it are incomplete, omitting important subtleties needed for soundness. In contrast, we have proven formally that CMod's rules enforce both information hiding and type-safe linking. To use CMod, the programmer develops and builds their software as usual, redirecting the compiler and linker to CMod's wrappers. We evaluated CMod by applying it to 30 open source programs, totaling more than one million LoC. Violations to CMod's rules revealed more than a thousand information hiding errors, dozens of typing errors, and hundreds of cases that, although not currently bugs, make programming mistakes more likely as the code evolves. At the same time, programs generally adhere to the assumptions underlying CMod's rules, and so we could fix rule violations with a modest effort. We conclude that CMod can effectively support modular programming in C: it soundly enforces type-safe linking and information-hiding while being largely compatible with existing practice. %B IEEE Transactions on Software Engineering %V 34 %P 357 - 376 %8 2008/06//May %@ 0098-5589 %G eng %N 3 %R 10.1109/TSE.2008.25 %0 Journal Article %J J Mol Evol %D 2008 %T A molecular footprint of limb loss: sequence variation of the autopodial identity gene Hoxa-13 %A Kohlsdorf,T %A Cummings, Michael P. %A Lynch,VJ %A Stopper,GF %A Takahashi,K %A Wagner,GP %X The homeobox gene Hoxa-13 codes for a transcription factor involved in multiple functions, including body axis and hand/foot development in tetrapods. In this study we investigate whether the loss of one function (e.g., limb loss in snakes) left a molecular footprint in exon 1 of Hoxa-13 that could be associated with the release of functional constraints caused by limb loss. Fragments of the Hoxa-13 exon 1 were sequenced from 13 species and analyzed, with additional published sequences of the same region, using relative rates and likelihood-ratio tests. Five amino acid sites in exon 1 of Hoxa-13 were detected as evolving under positive selection in the stem lineage of snakes. To further investigate whether there is an association between limb loss and sequence variation in Hoxa-13, we used the random forest method on an alignment that included shark, basal fish lineages, and "eu-tetrapods" such as mammals, turtle, alligator, and birds. The random forest method approaches the problem as one of classification, where we seek to predict the presence or absence of autopodium based on amino acid variation in Hoxa-13 sequences. Different alignments tested were associated with similar error rates (18.42%). The random forest method suggested that phenotypic states (autopodium present and absent) can often be correctly predicted based on Hoxa-13 sequences. Basal, nontetrapod gnat-hostomes that never had an autopodium were consistently classified as limbless together with the snakes, while eu-tetrapods without any history of limb loss in their phylogeny were also consistently classified as having a limb. Misclassifications affected mostly lizards, which, as a group, have a history of limb loss and limb re-evolution, and the urodele and caecilian in our sample. We conclude that a molecular footprint can be detected in Hoxa-13 that is associated with the lack of an autopodium; groups with classification ambiguity (lizards) are characterized by a history of repeated limb loss and possible limb re-evolution. %B J Mol Evol %V 67 %P 581 - 593 %8 2008/12// %G eng %N 6 %R 10.1007/s00239-008-9156-7 %0 Conference Paper %B Proceedings of the 3rd international workshop on Economics of networked systems %D 2008 %T Motivating participation in internet routing overlays %A Levin,Dave %A Baden,Randolph %A Lumezanu,Cristian %A Spring, Neil %A Bhattacharjee, Bobby %K incentives %K internet routing overlays %K service-level agreements %X PeerWise is an Internet routing overlay that reduces end-to-end latencies by allowing peers to forward through a relay instead of connecting directly to their destinations. Fundamental to PeerWise is the notion of peering agreements between two peers, wherein they agree to forward for one another. In this paper, we consider the problem of motivating users to establish and maintain peerings in a completely decentralized, scalable manner. We show that routing overlays present unique challenges and goals. For instance, since participants can always "fall back" on standard Internet routing, we must encourage users to stay in the system and maintain long-lived peering agreements. To address these challenges, we propose two mechanisms: First, we use Service Level Agreements (SLAs) to expressively negotiate peers' demands and the recourses they will take when SLAs are violated. Second, we propose a mechanism to address SLA violations that differs from the standard notion of punishment via service degradation. Our simulation results demonstrate that our mechanism causes peers to avoid SLA violators in favor of long-lived peerings. Lastly, we discuss potential, emergent behaviors in a selfish routing overlay. %B Proceedings of the 3rd international workshop on Economics of networked systems %S NetEcon '08 %I ACM %C New York, NY, USA %P 91 - 96 %8 2008/// %@ 978-1-60558-179-8 %G eng %U http://doi.acm.org/10.1145/1403027.1403048 %R 10.1145/1403027.1403048 %0 Journal Article %J AAMAS 2008: ALAMAS + ALAg Workshop %D 2008 %T Multi-agent learning with a distributed genetic algorithm %A Stonedahl,F. %A Rand, William %A Wilensky,U. %X Lightweight agents distributed in space have the potential to solve many complex problems. In this paper, we examine a model where agents represent individuals in a genetic algorithm (GA) solving a shared problem. We examine two questions: (1) How does the network density of connections between agents a ffect the performance of the systems? (2) How does the interaction topology a affect the performance of the system? In our model, agents exist in either a random network topology with long-distance communication, or a location-based topology, where agents only communicate with near neighbors. We examine both fixed and dynamic networks. Within the context of our investigation, our initial results indicate that relatively low network density achievesthe same results as a panmictic, or fully connected, population. Additionally, we find that dynamic networks outperform fixed networks, and that random network topologies outperform proximity-based network topologies. We conclude by showing how this model can be useful not only for multi-agent learning, but also for genetic algorithms, agent-based simulation and models of diff usion of innovation. %B AAMAS 2008: ALAMAS + ALAg Workshop %8 2008/// %G eng %0 Journal Article %J Journal of Universal Computer Science %D 2008 %T NASDAQ Velocity and Forces: An Interactive Visualization of Activity and Change %A Dao,H.T. %A Bazinet,A. %A Berthier,R. %A Shneiderman, Ben %B Journal of Universal Computer Science %V 14 %P 1391 - 1410 %8 2008/// %G eng %N 9 %0 Journal Article %J Bangladesh Journal of Botany %D 2008 %T New records of phytoplankton for Bangladesh. 2. Cryptophyceae and Synurophyceae %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasmin,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X This study presents two species of Rhodomonas, four species of Chroomonas, six species of Cryptomonas and Cryptochrysis minor, Cyanomonas coeruleus, Chrysodidymus synuroideus and Mallomonas akrokomos. These species have been reported from some ponds of Mathbaria in Pirojpur and Bakerganj of Barisal district in Bangladesh. %B Bangladesh Journal of Botany %V 36 %8 2008/11/05/ %@ 0253-5416 %G eng %U http://www.banglajol.info/bd/index.php/BJB/article/viewArticle/1549 %N 1 %R 10.3329/bjb.v36i1.1549 %0 Journal Article %J Bangladesh Journal of Plant Taxonomy %D 2008 %T New records of phytoplankton for Bangladesh. 5. Euglena, Euglenocapsa %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasmin,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X This study presents 20 taxa of the genus Euglena and one species of the rare euglenoid genus Euglenocapsa. All these taxa are reported for the first time from some pond ecosystems of Mathbaria in Pirojpur and Bakerganj of Barisal districts of Bangladesh. %B Bangladesh Journal of Plant Taxonomy %V 15 %8 2008/06/16/ %@ 1028-2092 %G eng %U http://www.banglajol.info/index.php/BJPT/article/viewArticle/910 %N 1 %R 10.3329/bjpt.v15i1.910 %0 Journal Article %J Bangladesh Journal of Botany %D 2008 %T New records of phytoplankton for Bangladesh. 7. Phacus spp. %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasmin,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X Thirteen species of Phacus hitherto not reported from Bangladesh have been described and illustrated. Freshwater ponds at southern districts of Pirojpur and Barisal revealed these presence of the species. %B Bangladesh Journal of Botany %V 37 %8 2008/11/05/ %@ 0253-5416 %G eng %U http://www.banglajol.info/bd/index.php/BJB/article/viewArticle/1564 %N 1 %R 10.3329/bjb.v37i1.1564 %0 Journal Article %J Bangladesh Journal of Botany %D 2008 %T New records of phytoplankton for Bangladesh. 8. Trachelomonas Ehr. (Euglenophyceae) %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasmin,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X Investigation of pelagic plankton communities from some freshwater ponds of Pirojpur and Barisal districts revealed the presence of 17 species under the genus Trachelomonas Ehr. for the first time in Bangladesh. %B Bangladesh Journal of Botany %V 37 %8 2008/12/19/ %@ 0253-5416 %G eng %U http://www.banglajol.info/bd/index.php/BJB/article/viewArticle/1719 %N 2 %R 10.3329/bjb.v37i2.1719 %0 Conference Paper %B Proceedings of the 16th ACM SIGSPATIAL international conference on Advances in geographic information systems %D 2008 %T NewsStand: a new view on news %A Teitler,Benjamin E. %A Lieberman,Michael D. %A Panozzo,Daniele %A Sankaranarayanan,Jagan %A Samet, Hanan %A Sperling,Jon %K clustering %K geotagging %K knowledge discovery %K text mining %X News articles contain a wealth of implicit geographic content that if exposed to readers improves understanding of today's news. However, most articles are not explicitly geotagged with their geographic content, and few news aggregation systems expose this content to users. A new system named NewsStand is presented that collects, analyzes, and displays news stories in a map interface, thus leveraging on their implicit geographic content. NewsStand monitors RSS feeds from thousands of online news sources and retrieves articles within minutes of publication. It then extracts geographic content from articles using a custom-built geotagger, and groups articles into story clusters using a fast online clustering algorithm. By panning and zooming in NewsStand's map interface, users can retrieve stories based on both topical significance and geographic region, and see substantially different stories depending on position and zoom level. %B Proceedings of the 16th ACM SIGSPATIAL international conference on Advances in geographic information systems %S GIS '08 %I ACM %C New York, NY, USA %P 18:1–18:10 - 18:1–18:10 %8 2008/// %@ 978-1-60558-323-5 %G eng %U http://doi.acm.org/10.1145/1463434.1463458 %R 10.1145/1463434.1463458 %0 Report %D 2008 %T A Node Aggregation Strategy to Reduce Complexity of Network Visualization using Semantic Substrates %A Aris,A. %A Shneiderman, Ben %X Semantic substrates are spatial templates for networks, wherenodes are grouped into regions and laid out within each region according to one or more node attributes. Analysts’ ability to design their own substrates leads to a different approach than other more automatic approaches to layout nodes (force-directed, circular, etc.). While the semantic substrate approach provides more interpretable node locations, sometimes a set of nodes is compressed into a small space on the display, leading to node overlap. In this paper, we significantly improve this situation by introducing the node aggregation mechanism in the context of semantic substrates. We illustrate this functionality in a document citation network and provide pros and cons of the approach. We conclude with guidelines and future directions for this research. Throughout the paper, examples are illustrated with NVSS 3.0, the network visualization tool developed to explore the semantic substrate idea. %I Human-Computer Interaction Lab, University of Maryland %8 2008/// %G eng %0 Journal Article %J Information Processing Letters %D 2008 %T A note on the distribution of the number of prime factors of the integers %A Srinivasan, Aravind %K Chernoff bounds %K Dependent random variables %K Primes %K Probabilistic number theory %K randomized algorithms %K Tail bounds %X The Chernoff–Hoeffding bounds are fundamental probabilistic tools. An elementary approach is presented to obtain a Chernoff-type upper-tail bound for the number of prime factors of a random integer in { 1 , 2 , … , n } . The method illustrates tail bounds in negatively-correlated settings. %B Information Processing Letters %V 109 %P 133 - 135 %8 2008/12/31/ %@ 0020-0190 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020019008002688 %N 2 %R 10.1016/j.ipl.2008.09.010 %0 Journal Article %J Proceedings of the IEEE %D 2008 %T Object Detection, Tracking and Recognition for Multiple Smart Cameras %A Sankaranarayanan,A. C %A Veeraraghavan,A. %A Chellapa, Rama %K algorithm;geometric %K analysis;image %K association;distributed %K camera;visual %K cameras; %K cameras;object %K colour %K constraints;imaging %K data %K detection;object %K detection;sensor %K detection;three-dimiensional %K fusion;target %K model;video %K network;distributed %K recognition;object %K scene %K sensor %K sensor;multiple %K sensors;geometry;image %K sensors;object %K smart %K texture;intelligent %K tracking;target %K tracking;video %X Video cameras are among the most commonly used sensors in a large number of applications, ranging from surveillance to smart rooms for videoconferencing. There is a need to develop algorithms for tasks such as detection, tracking, and recognition of objects, specifically using distributed networks of cameras. The projective nature of imaging sensors provides ample challenges for data association across cameras. We first discuss the nature of these challenges in the context of visual sensor networks. Then, we show how real-world constraints can be favorably exploited in order to tackle these challenges. Examples of real-world constraints are (a) the presence of a world plane, (b) the presence of a three-dimiensional scene model, (c) consistency of motion across cameras, and (d) color and texture properties. In this regard, the main focus of this paper is towards highlighting the efficient use of the geometric constraints induced by the imaging devices to derive distributed algorithms for target detection, tracking, and recognition. Our discussions are supported by several examples drawn from real applications. Lastly, we also describe several potential research problems that remain to be addressed. %B Proceedings of the IEEE %V 96 %P 1606 - 1624 %8 2008/10// %@ 0018-9219 %G eng %N 10 %R 10.1109/JPROC.2008.928758 %0 Journal Article %J Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %D 2008 %T Occurrence and Expression of Luminescence in Vibrio Cholerae %A Grim,Christopher J. %A Taviani,Elisa %A Alam,Munirul %A Huq,Anwar %A Sack,R. Bradley %A Rita R Colwell %X Several species of the genus Vibrio, including Vibrio cholerae, are bioluminescent or contain bioluminescent strains. Previous studies have reported that only 10% of V. cholerae strains are luminescent. Analysis of 224 isolates of non-O1/non-O139 V. cholerae collected from Chesapeake Bay, MD, revealed that 52% (116/224) were luminescent when an improved assay method was employed and 58% (130/224) of isolates harbored the luxA gene. In contrast, 334 non-O1/non-O139 V. cholerae strains isolated from two rural provinces in Bangladesh yielded only 21 (6.3%) luminescent and 35 (10.5%) luxA+ isolates. An additional 270 clinical and environmental isolates of V. cholerae serogroups O1 and O139 were tested, and none were luminescent or harbored luxA. These results indicate that bioluminescence may be a trait specific for non-O1/non-O139 V. cholerae strains that frequently occur in certain environments. Luminescence expression patterns of V. cholerae were also investigated, and isolates could be grouped based on expression level. Several strains with defective expression of the lux operon, including natural K variants, were identified. %B Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %V 74 %P 708 - 715 %8 2008/02/01/ %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/74/3/708 %N 3 %R 10.1128/AEM.01537-07 %0 Book Section %B Automata, Languages and Programming %D 2008 %T Optimal Cryptographic Hardness of Learning Monotone Functions %A Dana Dachman-Soled %A Lee, Homin K. %A Malkin, Tal %A Servedio, Rocco A. %A Wan, Andrew %A Wee, Hoeteck %E Aceto, Luca %E Damgård, Ivan %E Goldberg, Leslie Ann %E Halldórsson, Magnús M. %E Ingólfsdóttir, Anna %E Walukiewicz, Igor %K Data structures %K Data Structures, Cryptology and Information Theory %K Discrete Mathematics in Computer Science %K Numeric Computing %K Software Engineering/Programming and Operating Systems %K Theory of computation %X A wide range of positive and negative results have been established for learning different classes of Boolean functions from uniformly distributed random examples. However, polynomial-time algorithms have thus far been obtained almost exclusively for various classes of monotone functions, while the computational hardness results obtained to date have all been for various classes of general (nonmonotone) functions. Motivated by this disparity between known positive results (for monotone functions) and negative results (for nonmonotone functions), we establish strong computational limitations on the efficient learnability of various classes of monotone functions. We give several such hardness results which are provably almost optimal since they nearly match known positive results. Some of our results show cryptographic hardness of learning polynomial-size monotone circuits to accuracy only slightly greater than 1/2+1/n‾‾√1/2 + 1/\sqrt{n} ; this accuracy bound is close to optimal by known positive results (Blum et al., FOCS ’98). Other results show that under a plausible cryptographic hardness assumption, a class of constant-depth, sub-polynomial-size circuits computing monotone functions is hard to learn; this result is close to optimal in terms of the circuit size parameter by known positive results as well (Servedio, Information and Computation ’04). Our main tool is a complexity-theoretic approach to hardness amplification via noise sensitivity of monotone functions that was pioneered by O’Donnell (JCSS ’04). %B Automata, Languages and Programming %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 36 - 47 %8 2008/01/01/ %@ 978-3-540-70574-1, 978-3-540-70575-8 %G eng %U http://link.springer.com/chapter/10.1007/978-3-540-70575-8_4 %0 Conference Paper %B Motion and video Computing, 2008. WMVC 2008. IEEE Workshop on %D 2008 %T Optimal Multi-View Fusion of Object Locations %A Sankaranarayanan,A. C %A Chellapa, Rama %K control;surveillance;tracking; %K detection;position %K estimates;optimal %K estimation;location %K estimation;multicamera %K estimation;object %K estimator;motion %K fusion;surveillance;cameras;estimation %K location %K multiview %K theory;motion %K tracking;minimum %K variance %X In surveillance applications, it is common to have multiple cameras observing targets exhibiting motion on a ground plane. Tracking and estimation of the location of a target on the plane becomes an important inference problem. In this paper, we study the problem of combining estimates of location obtained from multiple cameras. We model the relation between the uncertainty in the location estimation to the position and location of the camera with respect to the plane (which is encoded by a 2D projective transformation). This is addressed by a theoretical study of the properties of a random variable under a projective transformation and analysis of the geometric setting when the moments of the transformed random variable exist. In this context, we prove that ground plane tracking near the horizon line is often inaccurate. Using suitable approximations to compute the moments, a minimum variance estimator is designed to fuse the multi-camera location estimates. Finally, we present experimental results that illustrate the importance of such modeling in location estimation and tracking. %B Motion and video Computing, 2008. WMVC 2008. IEEE Workshop on %P 1 - 8 %8 2008/01// %G eng %R 10.1109/WMVC.2008.4544048 %0 Report %D 2008 %T Parallel Algorithms for Volumetric Surface Construction %A JaJa, Joseph F. %A Shi,Qingmin %A Varshney, Amitabh %X Large scale scientific data sets are appearing at an increasing rate whose sizes can range from hundreds of gigabytes to tens of terabytes. Isosurface extraction and rendering is an important visualization technique that enables the visual exploration of such data sets using surfaces. However the computational requirements of this approach are substantial which in general prevent the interactive rendering of isosurfaces for large data sets. Therefore, parallel and distributed computing techniques offer a promising direction to deal with the corresponding computational challenges. In this chapter, we give a brief historical perspective of the isosurface visualization approach, and describe the basic sequential and parallel techniques used to extract and render isosurfaces with a particular focus on out-of-core techniques. For parallel algorithms, we assume a distributed memory model in which each processor has its own local disk, and processors communicate and exchange data through an interconnection network. We present a general framework for evaluating parallel isosurface extraction algorithms and describe the related best known parallel algorithms. We also describe the main parallel strategies used to handle isosurface rendering, pointing out the limitations of these strategies. 1. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %8 2008/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/summary?doi=?doi=10.1.1.122.4472 %0 Conference Paper %B Proceedings of the 8th ACM SIGPLAN-SIGSOFT workshop on Program analysis for software tools and engineering %D 2008 %T Path projection for user-centered static analysis tools %A Khoo,Yit Phang %A Foster, Jeffrey S. %A Hicks, Michael W. %A Sazawal,Vibha %X The research and industrial communities have made great strides in developing sophisticated defect detection tools based on static analysis. To date most of the work in this area has focused on developing novel static analysis algorithms, but has neglected study of other aspects of static analysis tools, particularly user interfaces. In this work, we present a novel user interface toolkit called Path Projection that helps users visualize, navigate, and understand program paths, a common component of many tools' error reports. We performed a controlled user study to measure the benefit of Path Projection in triaging error reports from Locksmith, a data race detection tool for C. We found that Path Projection improved participants' time to complete this task without affecting accuracy, while participants felt Path Projection was useful and strongly preferred it to a more standard viewer. %B Proceedings of the 8th ACM SIGPLAN-SIGSOFT workshop on Program analysis for software tools and engineering %S PASTE '08 %I ACM %C New York, NY, USA %P 57 - 63 %8 2008/// %@ 978-1-60558-382-2 %G eng %U http://doi.acm.org/10.1145/1512475.1512488 %R 10.1145/1512475.1512488 %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on %D 2008 %T A pattern classification framework for theoretical analysis of component forensics %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K analysis;information %K classification; %K classification;parameter %K component %K estimation;pattern %K forensics;component %K forensics;pattern %K identification;forensic %K parameter %K processing;nonintrusive %X Component forensics is an emerging methodology for forensic analysis that aims at estimating the algorithms and parameters in each component of a digital device. This paper proposes a theoretical foundation to examine the performance limits of component forensics. Using ideas from pattern classification theory, we define formal notions of identifiability of components in the information processing chain. We show that the parameters of certain device components can be accurately identified only in controlled settings through semi non-intrusive forensics, while the parameters of some others can be computed directly from the available sample data via complete non-intrusive analysis. We then extend the proposed theoretical framework to quantify and improve the accuracies and confidence in component parameter identification for several forensic applications. %B Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on %P 1665 - 1668 %8 2008/04/31/4 %G eng %R 10.1109/ICASSP.2008.4517947 %0 Conference Paper %B Proceedings of the theory and applications of cryptographic techniques 27th annual international conference on Advances in cryptology %D 2008 %T Predicate encryption supporting disjunctions, polynomial equations, and inner products %A Katz, Jonathan %A Sahai,Amit %A Waters,Brent %X Predicate encryption is a new paradigm generalizing, among other things, identity-based encryption. In a predicate encryption scheme, secret keys correspond to predicates and ciphertexts are associated with attributes; the secret key SKf corresponding to a predicate f can be used to decrypt a ciphertext associated with attribute I if and only if f(I) = 1. Constructions of such schemes are currently known for relatively few classes of predicates. We construct such a scheme for predicates corresponding to the evaluation of inner products over ZN (for some large integer N). This, in turn, enables constructions in which predicates correspond to the evaluation of disjunctions, polynomials, CNF/DNF formulae, or threshold predicates (among others). Besides serving as a significant step forward in the theory of predicate encryption, our results lead to a number of applications that are interesting in their own right. %B Proceedings of the theory and applications of cryptographic techniques 27th annual international conference on Advances in cryptology %S EUROCRYPT'08 %I Springer-Verlag %C Berlin, Heidelberg %P 146 - 162 %8 2008/// %@ 3-540-78966-9, 978-3-540-78966-6 %G eng %U http://dl.acm.org/citation.cfm?id=1788414.1788423 %0 Book Section %B Automata, Languages and ProgrammingAutomata, Languages and Programming %D 2008 %T The Randomized Coloring Procedure with Symmetry-Breaking %A Pemmaraju,Sriram %A Srinivasan, Aravind %E Aceto,Luca %E Damgård,Ivan %E Goldberg,Leslie %E Halldórsson,Magnús %E Ingólfsdóttir,Anna %E Walukiewicz,Igor %X A basic randomized coloring procedure has been used in probabilistic proofs to obtain remarkably strong results on graph coloring. These results include the asymptotic version of the List Coloring Conjecture due to Kahn, the extensions of Brooks’ Theorem to sparse graphs due to Kim and Johansson, and Luby’s fast parallel and distributed algorithms for graph coloring. The most challenging aspect of a typical probabilistic proof is showing adequate concentration bounds for key random variables. In this paper, we present a simple symmetry-breaking augmentation to the randomized coloring procedure that works well in conjunction with Azuma’s Martingale Inequality to easily yield the requisite concentration bounds. We use this approach to obtain a number of results in two areas: frugal coloring and weighted equitable coloring . A β-frugal coloring of a graph G is a proper vertex-coloring of G in which no color appears more than β times in any neighborhood. Let G = ( V , E ) be a vertex-weighted graph with weight function w : V →[0, 1] and let W = ∑ v ∈ V w ( v ). A weighted equitable coloring of G is a proper k -coloring such that the total weight of every color class is “large”, i.e., “not much smaller” than W / k ; this notion is useful in obtaining tail bounds for sums of dependent random variables. %B Automata, Languages and ProgrammingAutomata, Languages and Programming %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 5125 %P 306 - 319 %8 2008/// %@ 978-3-540-70574-1 %G eng %U http://dx.doi.org/10.1007/978-3-540-70575-8_26 %0 Conference Paper %B Software Testing, Verification, and Validation, 2008 1st International Conference on %D 2008 %T Relationships between Test Suites, Faults, and Fault Detection in GUI Testing %A Strecker,J. %A Memon, Atif M. %K Fault detection %K fault-related factors %K Graphical user interfaces %K GUI testing %K program testing %K software-testing %K test suites %K test-suite-related factors %X Software-testing researchers have long sought recipes for test suites that detect faults well. In the literature, empirical studies of testing techniques abound, yet the ideal technique for detecting the desired kinds of faults in a given situation often remains unclear. This work shows how understanding the context in which testing occurs, in terms of factors likely to influence fault detection, can make evaluations of testing techniques more readily applicable to new situations. We present a methodology for discovering which factors do statistically affect fault detection, and we perform an experiment with a set of test-suite- and fault-related factors in the GUI testing of two fielded, open-source applications. Statement coverage and GUI-event coverage are found to be statistically related to the likelihood of detecting certain kinds of faults. %B Software Testing, Verification, and Validation, 2008 1st International Conference on %P 12 - 21 %8 2008/04// %G eng %R 10.1109/ICST.2008.26 %0 Journal Article %J Information Seeking Support Systems %D 2008 %T Research Agenda: Visual Overviews for Exploratory Search %A Shneiderman, Ben %X Exploratory search is necessary when users knowledge ofthe domain is incomplete or when initial user goals do not match available data or metadata that is the basis for search indexing attributes. Such mismatches mean that users need to learn more in order to develop a better understanding of the domain or to revise their search goals. Exploratory search processes may take weeks or months, so interfaces that support prolonged exploration are necessary. The attraction of exploratory search is that users can take on more ambitious goals that require substantial learning and creative leaps to bridge the gaps between what they know and that they seek. %B Information Seeking Support Systems %V 11 %P 4 - 4 %8 2008/// %G eng %0 Journal Article %J Syst Biol %D 2008 %T Resolving arthropod phylogeny: exploring phylogenetic signal within 41 kb of protein-coding nuclear gene sequence %A Regier,J. C %A Shultz,J. W %A Ganley,A. R.D %A Hussey,A. %A Shi,D. %A Ball,B. %A Zwick,A. %A Stajich,J. E %A Cummings, Michael P. %A Martin,J. W %A Cunningham,CW %X This study attempts to resolve relationships among and within the four basal arthropod lineages (Pancrustacea, Myriapoda, Euchelicerata, Pycnogonida) and to assess the widespread expectation that remaining phylogenetic problems will yield to increasing amounts of sequence data. Sixty-eight regions of 62 protein-coding nuclear genes (approximately 41 kilobases (kb)/taxon) were sequenced for 12 taxonomically diverse arthropod taxa and a tardigrade outgroup. Parsimony, likelihood, and Bayesian analyses of total nucleotide data generally strongly supported the monophyly of each of the basal lineages represented by more than one species. Other relationships within the Arthropoda were also supported, with support levels depending on method of analysis and inclusion/exclusion of synonymous changes. Removing third codon positions, where the assumption of base compositional homogeneity was rejected, altered the results. Removing the final class of synonymous mutations–first codon positions encoding leucine and arginine, which were also compositionally heterogeneous–yielded a data set that was consistent with a hypothesis of base compositional homogeneity. Furthermore, under such a data-exclusion regime, all 68 gene regions individually were consistent with base compositional homogeneity. Restricting likelihood analyses to nonsynonymous change recovered trees with strong support for the basal lineages but not for other groups that were variably supported with more inclusive data sets. In a further effort to increase phylogenetic signal, three types of data exploration were undertaken. (1) Individual genes were ranked by their average rate of nonsynonymous change, and three rate categories were assigned–fast, intermediate, and slow. Then, bootstrap analysis of each gene was performed separately to see which taxonomic groups received strong support. Five taxonomic groups were strongly supported independently by two or more genes, and these genes mostly belonged to the slow or intermediate categories, whereas groups supported only by a single gene region tended to be from genes of the fast category, arguing that fast genes provide a less consistent signal. (2) A sensitivity analysis was performed in which increasing numbers of genes were excluded, beginning with the fastest. The number of strongly supported nodes increased up to a point and then decreased slightly. Recovery of Hexapoda required removal of fast genes. Support for Mandibulata (Pancrustacea + Myriapoda) also increased, at times to "strong" levels, with removal of the fastest genes. (3) Concordance selection was evaluated by clustering genes according to their ability to recover Pancrustacea, Euchelicerata, or Myriapoda and analyzing the three clusters separately. All clusters of genes recovered the three concordance clades but were at times inconsistent in the relationships recovered among and within these clades, a result that indicates that the a priori concordance criteria may bias phylogenetic signal in unexpected ways. In a further attempt to increase support of taxonomic relationships, sequence data from 49 additional taxa for three slow genes (i.e., EF-1 alpha, EF-2, and Pol II) were combined with the various 13-taxon data sets. The 62-taxon analyses supported the results of the 13-taxon analyses and provided increased support for additional pancrustacean clades found in an earlier analysis including only EF-1 alpha, EF-2, and Pol II. %B Syst Biol %V 57 %P 920 - 938 %8 2008/12// %G eng %N 6 %R 10.1080/10635150802570791 %0 Journal Article %J The Eighth International Workshop on Visual Surveillance %D 2008 %T Resource allocation for tracking multiple targets using particle filters %A Kembhavi,A. %A Schwartz, W.R. %A Davis, Larry S. %A others %X Particle filters have been very widely used to track targets in video sequences. However, they suffer from an exponential rise in the number of particles needed to jointly track multiple targets. On the other hand, using multiple independent filters to track in crowded scenes often leads to erroneous results. We present a new particle filtering framework which uses an intelligent resource allocation scheme allowing us to track a large number of targets using a small set of particles. First, targets with overlapping posterior distributions and similar appearance models are clustered into interaction groups and tracked jointly, but independent of other targets in the scene. Second, different number of particles are allocated to different groups based on the following observations. Groups with higher associations (quantifying spatial proximity and pairwise appearance similarity) are given more particles. Groups with larger number of targets are given a larger number of particles. Finally, groups with ineffective proposal distributions are assigned more particles. Our experiments demonstrate the effectiveness of this framework over the commonly used joint particle filter with Markov Chain Monte Carlo (MCMC) sampling. %B The Eighth International Workshop on Visual Surveillance %8 2008/// %G eng %0 Book Section %B Studies in Computational Intelligence: Machine Learning in Document Analysis and RecognitionStudies in Computational Intelligence: Machine Learning in Document Analysis and Recognition %D 2008 %T Review of Classifier Combination Methods %A Tulyakov,Sergey %A Jaeger,Stefan %A Govindaraju,Venu %A David Doermann %E Simone Marinai,Hiromichi Fujisawa %X Classifier combination methods have proved to be an effective tool to increase the performance of pattern recognition applications. In this chapter we review and categorize major advancements in this field. Despite a significant number of publications describing successful classifier combination implementations, the theoretical basis is still missing and achieved improvements are inconsistent. By introducing different categories of classifier combinations in this review we attempt to put forward more specific directions for future theoretical research.We also introduce a retraining effect and effects of locality based training as important properties of classifier combinations. Such effects have significant influence on the performance of combinations, and their study is necessary for complete theoretical understanding of combination algorithms. %B Studies in Computational Intelligence: Machine Learning in Document Analysis and RecognitionStudies in Computational Intelligence: Machine Learning in Document Analysis and Recognition %I Springer %P 361 - 386 %8 2008/// %G eng %0 Conference Paper %B AAAI-08 Workshop on Metareasoning,(Chicago, IL) %D 2008 %T The role of metacognition in robust AI systems %A Schmill,M. D %A Oates,T. %A Anderson,M. %A Fults,S. %A Josyula,D. %A Perlis, Don %A Wilson,S. %B AAAI-08 Workshop on Metareasoning,(Chicago, IL) %8 2008/// %G eng %0 Conference Paper %B Proceedings of the 2008 ACM SIGMOD international conference on Management of data %D 2008 %T Scalable network distance browsing in spatial databases %A Samet, Hanan %A Sankaranarayanan,Jagan %A Alborzi,Houman %K decoupling %K nearest neighbor %K scalability %K shortest path quadtree %K spatial networks %X An algorithm is presented for finding the k nearest neighbors in a spatial network in a best-first manner using network distance. The algorithm is based on precomputing the shortest paths between all possible vertices in the network and then making use of an encoding that takes advantage of the fact that the shortest paths from vertex u to all of the remaining vertices can be decomposed into subsets based on the first edges on the shortest paths to them from u. Thus, in the worst case, the amount of work depends on the number of objects that are examined and the number of links on the shortest paths to them from q, rather than depending on the number of vertices in the network. The amount of storage required to keep track of the subsets is reduced by taking advantage of their spatial coherence which is captured by the aid of a shortest path quadtree. In particular, experiments on a number of large road networks as well as a theoretical analysis have shown that the storage has been reduced from O(N3) to O(N1.5) (i.e., by an order of magnitude equal to the square root). The precomputation of the shortest paths along the network essentially decouples the process of computing shortest paths along the network from that of finding the neighbors, and thereby also decouples the domain S of the query objects and that of the objects from which the neighbors are drawn from the domain V of the vertices of the spatial network. This means that as long as the spatial network is unchanged, the algorithm and underlying representation of the shortest paths in the spatial network can be used with different sets of objects. %B Proceedings of the 2008 ACM SIGMOD international conference on Management of data %S SIGMOD '08 %I ACM %C New York, NY, USA %P 43 - 54 %8 2008/// %@ 978-1-60558-102-6 %G eng %U http://doi.acm.org/10.1145/1376616.1376623 %R 10.1145/1376616.1376623 %0 Conference Paper %B Proceedings of the 17th international conference on World Wide Web %D 2008 %T Scaling RDF with Time %A Pugliese,Andrea %A Udrea,Octavian %A V.S. Subrahmanian %K RDF indexing %K resource description framework %K temporal RDF %X The World Wide Web Consortium's RDF standard primarily consists of (subject, property, object) triples that specify the value that a given subject has for a given property. However, it is frequently the case that even for a fixed subject and property, the value varies with time. As a consequence, efforts have been made to annotate RDF triples with "valid time" intervals. However, to date, no proposals exist for efficient indexing of such temporal RDF databases. It is clearly beneficial to store RDF data in a relational DB - however, standard relational indexes are inadequately equipped to handle RDF's graph structure. In this paper, we propose the tGRIN index structure that builds a specialized index for temporal RDF that is physically stored in an RDBMS. Past efforts to store RDF in relational stores include Jena2 from HP, Sesame from OpenRDF.org, and 3store from the University of Southampton. We show that even when these efforts are augmented with well known temporal indexes like R+ trees, SR-trees, ST-index, and MAP21, the tGRIN index exhibits superior performance. In terms of index build time, tGRIN takes two thirds or less of the time used by any other system, and it uses a comparable amount of memory and less disk space than Jena, Sesame and 3store. More importantly, tGRIN can answer queries three to six times faster for average query graph patterns and five to ten times faster for complex queries than these systems. %B Proceedings of the 17th international conference on World Wide Web %S WWW '08 %I ACM %C New York, NY, USA %P 605 - 614 %8 2008/// %@ 978-1-60558-085-2 %G eng %U http://doi.acm.org/10.1145/1367497.1367579 %R 10.1145/1367497.1367579 %0 Journal Article %J AMIA Annual Symposium ProceedingsAMIA Annu Symp Proc %D 2008 %T Searching Electronic Health Records for Temporal Patterns in Patient Histories: A Case Study with Microsoft Amalga %A Plaisant, Catherine %A Lam,Stanley %A Shneiderman, Ben %A Smith,Mark S. %A Roseman,David %A Marchand,Greg %A Gillam,Michael %A Feied,Craig %A Handler,Jonathan %A Rappaport,Hank %X As electronic health records (EHR) become more widespread, they enable clinicians and researchers to pose complex queries that can benefit immediate patient care and deepen understanding of medical treatment and outcomes. However, current query tools make complex temporal queries difficult to pose, and physicians have to rely on computer professionals to specify the queries for them. This paper describes our efforts to develop a novel query tool implemented in a large operational system at the Washington Hospital Center (Microsoft Amalga, formerly known as Azyxxi). We describe our design of the interface to specify temporal patterns and the visual presentation of results, and report on a pilot user study looking for adverse reactions following radiology studies using contrast. %B AMIA Annual Symposium ProceedingsAMIA Annu Symp Proc %V 2008 %P 601 - 605 %8 2008/// %@ 1942-597X %G eng %0 Journal Article %J Computer Vision–ECCV 2008 %D 2008 %T Searching the world’s herbaria: A system for visual identification of plant species %A Belhumeur,P. %A Chen,D. %A Feiner,S. %A Jacobs, David W. %A Kress,W. %A Ling,H. %A Lopez,I. %A Ramamoorthi,R. %A Sheorey,S. %A White,S. %X We describe a working computer vision system that aids in the identification of plant species. A user photographs an isolated leaf on a blank background, and the system extracts the leaf shape and matches it to the shape of leaves of known species. In a few seconds, the system displays the top matching species, along with textual descriptions and additional images. This system is currently in use by botanists at the Smithsonian Institution National Museum of Natural History. The primary contributions of this paper are: a description of a working computer vision system and its user interface for an important new application area; the introduction of three new datasets containing thousands of single leaf images, each labeled by species and verified by botanists at the US National Herbarium; recognition results for two of the three leaf datasets; and descriptions throughout of practical lessons learned in constructing this system. %B Computer Vision–ECCV 2008 %P 116 - 129 %8 2008/// %G eng %R 10.1007/978-3-540-88693-8_9 %0 Journal Article %J Emerging Infectious DiseasesEmerg Infect Dis %D 2008 %T Seasonal Cholera from Multiple Small Outbreaks, Rural Bangladesh %A Stine,O. Colin %A Alam,Munirul %A Tang,Li %A Nair,G. Balakrish %A Siddique,A. Kasem %A Faruque,Shah M. %A Huq,Anwar %A Rita R Colwell %A Sack,R. Bradley %A Morris,J. Glenn %X Clinical and environmental Vibrio cholerae organisms collected from February 2004 through April 2005 were systematically isolated from 2 rural Bangladeshi locales. Their genetic relatedness was evaluated at 5 loci that contained a variable number of tandem repeats (VNTR). The observed minimal overlap in VNTR patterns between the 2 communities was consistent with sequential, small outbreaks from local sources. %B Emerging Infectious DiseasesEmerg Infect Dis %V 14 %P 831 - 833 %8 2008/05// %@ 1080-6040 %G eng %N 5 %R 10.3201/eid1405.071116 %0 Book %D 2008 %T Shape Analysis and Structuring %A De Floriani, Leila %A Spagnuolo,Michela %K Computer Graphics %K Computer vision %K Computers / Computer Graphics %K Computers / Image Processing %K Geometrical models %K Geometry %K Geometry, Analytic %K Image analysis %K IMAGE PROCESSING %K Mathematics / Functional Analysis %K Mathematics / Geometry / General %K Mathematics / Graphic Methods %K Mathematics / Mathematical Analysis %K shapes %K Technology & Engineering / Engineering (General) %K Visualization %X With a lot of recent developments in the field, this much-needed book has come at just the right time. It covers a variety of topics related to preserving and enhancing shape information at a geometric level. The contributors also cover subjects that are relevant to effectively capturing the structure of a shape by identifying relevant shape components and their mutual relationships. %I Springer %8 2008/// %@ 9783540332640 %G eng %0 Journal Article %J Proc. of NIPS %D 2008 %T A “Shape Aware” Model for semi-supervised Learning of Objects and its Context %A Gupta,A. %A Shi,J. %A Davis, Larry S. %X We present an approach that combines bag-of-words and spatial models to performsemantic and syntactic analysis for recognition of an object based on its internal appearance and its context. We argue that while object recognition requires mod- eling relative spatial locations of image features within the object, a bag-of-word is sufficient for representing context. Learning such a model from weakly labeled data involves labeling of features into two classes: foreground(object) or “infor- mative” background(context). We present a “shape-aware” model which utilizes contour information for efficient and accurate labeling of features in the image. Our approach iterates between an MCMC-based labeling and contour based la- beling of features to integrate co-occurrence of features and shape similarity. %B Proc. of NIPS %8 2008/// %G eng %0 Journal Article %J Pattern Analysis and Machine Intelligence, IEEE Transactions on %D 2008 %T Shape-and-Behavior Encoded Tracking of Bee Dances %A Veeraraghavan,A. %A Chellapa, Rama %A Srinivasan, M. %K Automated;Reproducibility of Results;Sensitivity and Specificity;Social Behavior; %K Computer-Assisted;Imaging %K Markov model;bee dance;behavior analysis;shape model;social insect;three-tier hierarchical motion model;tracking;video sequence;waggle dance;Markov processes;image sequences;optical tracking;video signal processing;zoology;Algorithms;Animal Communication; %K Three-Dimensional;Information Storage and Retrieval;Male;Pattern Recognition %X Behavior analysis of social insects has garnered impetus in recent years and has led to some advances in fields like control systems and flight navigation. Manual labeling of insect motions required for analyzing the behaviors of insects requires significant investment of time and effort. In this paper, we propose certain general principles that help in simultaneous automatic tracking and behavior analysis, with applications in tracking bees and recognizing specific behaviors that they exhibit. The state space for tracking is defined using the position, orientation, and current behavior of the insect being tracked. The position and the orientation are parameterized using a shape model, whereas the behavior is explicitly modeled using a three-tier hierarchical motion model. The first tier (dynamics) models the local motions exhibited, and the models built in this tier act as a vocabulary for behavior modeling. The second tier is a Markov motion model built on top of the local motion vocabulary, which serves as the behavior model. The third tier of the hierarchy models the switching between behaviors, and this is also modeled as a Markov model. We address issues in learning the three-tier behavioral model, in discriminating between models, and in detecting and modeling abnormal behaviors. Another important aspect of this work is that it leads to joint tracking and behavior analysis instead of the traditional "track-and-then-recognize" approach. We apply these principles for tracking bees in a hive while they are executing the waggle dance and the round dance. %B Pattern Analysis and Machine Intelligence, IEEE Transactions on %V 30 %P 463 - 476 %8 2008/03// %@ 0162-8828 %G eng %N 3 %R 10.1109/TPAMI.2007.70707 %0 Report %D 2008 %T SNARE: Spatio-temporal Network-level Automatic Reputation Engine %A Feamster, Nick %A Gray,A.G. %A Krasser,S. %A Syed,N.A. %X Current spam filtering techniques classify email based oncontent and IP reputation blacklists or whitelists. Unfortu- nately, spammers can alter spam content to evade content- based filters, and spammers continually change the IP ad- dresses from which they send spam. Previous work has sug- gested that filters based on network-level behavior might be more efficient and robust, by making decisions based on how messages are sent, as opposed to what is being sent or who is sending them. This paper presents a technique to identify spammers based on features that exploit the network-level spatio- temporal behavior of email senders to differentiate the spam- ming IPs from legitimate senders. Our behavioral classifier has two benefits: (1) it is early (i.e., it can automatically detect spam without seeing a large amount of email from a sending IP address—sometimes even upon seeing only a single packet); (2) it is evasion-resistant (i.e., it is based on spatial and temporal features that are difficult for a sender to change). We build classifiers based on these features us- ing two different machine learning methods, support vec- tor machine and decision trees, and we study the efficacy of these classifiers using labeled data from a deployed com- mercial spam-filtering system. Surprisingly, using only fea- tures from a single IP packet header (i.e., without looking at packet contents), our classifier can identify spammers with about 93% accuracy and a reasonably low false-positive rate (about 7%). After looking at a single message spammer identification accuracy improves to more than 94% with a false rate of just over 5%. These suggest an effective sender reputation mechanism. %I Georgia Institute of Technology %V GT-CSE-08-02 %8 2008/// %G eng %U http://hdl.handle.net/1853/25135 %0 Conference Paper %B First Intl. Workshop on Social Computing, Behavioral Modeling and Prediction %D 2008 %T The soma terror organization portal (stop): Social network and analytic tools for the real-time analysis of terror groups %A Martinez,V. %A Simari,G. %A Sliva,A. %A V.S. Subrahmanian %X Stochastic Opponent Modeling Agents (SOMA) have been proposed as a paradigm for reasoning about cultural groups, terror groups, and other socioeconomic-political-military organizations worldwide. In this paper, we describe the SOMA Terror Organization Portal (STOP). STOP provides a single point of contact through which analysts may access data about terror groups world wide. In order to analyze this data, SOMA provides three major components: the SOMA Extraction Engine (SEE), the SOMA Adversarial Forecast Engine (SAFE), and the SOMA Analyst NEtwork (SANE) that allows analysts to find other analysts doing similar work, share findings with them, and let consensus findings emerge. This paper describes the STOP framework. %B First Intl. Workshop on Social Computing, Behavioral Modeling and Prediction %8 2008/// %G eng %R 10.1007/978-0-387-77672-9_2 %0 Journal Article %J Social Computing, Behavioral Modeling, and Prediction %D 2008 %T The SOMA Terror Organization Portal (STOP): social network and analytic tools for the real-time analysis of terror groups %A Sliva,A. %A V.S. Subrahmanian %A Martinez,V. %A Simari,G. I %X Stochastic Opponent Modeling Agents (SOMA) have been proposed as a paradigm for reasoning about cultural groups, terror groups, and other socioeconomic-political-military organizations worldwide. In this paper, we describe the SOMA Terror Organization Portal (STOP). STOP provides a single point of contact through which analysts may access data about terror groups world wide. In order to analyze this data, SOMA provides three major components: the SOMA Extraction Engine (SEE), the SOMA Adversarial Forecast Engine (SAFE), and the SOMA Analyst NEtwork (SANE) that allows analysts to find other analysts doing similar work, share findings with them, and let consensus findings emerge. This paper describes the STOP framework. %B Social Computing, Behavioral Modeling, and Prediction %P 9 - 18 %8 2008/// %G eng %R 10.1007/978-0-387-77672-9_2 %0 Conference Paper %B ACM SIGGRAPH 2008 classes %D 2008 %T A sorting approach to indexing spatial data %A Samet, Hanan %X Spatial data is distinguished from conventional data by having extent. Therefore, spatial queries involve both the objects and the space that they occupy. The handling of queries that involve spatial data is facilitated by building an index on the data. The traditional role of the index is to sort the data, which means that it orders the data. However, since generally no ordering exists in dimensions greater than 1 without a transformation of the data to one dimension, the role of the sort process is one of differentiating between the data and what is usually done is to sort the spatial objects with respect to the space that they occupy. The resulting ordering is usually implicit rather than explicit so that the data need not be resorted (i.e., the index need not be rebuilt) when the queries change (e.g., the query reference objects). The index is said to order the space and the characteristics of such indexes are explored further. %B ACM SIGGRAPH 2008 classes %S SIGGRAPH '08 %I ACM %C New York, NY, USA %P 91:1–91:15 - 91:1–91:15 %8 2008/// %G eng %U http://doi.acm.org/10.1145/1401132.1401250 %R 10.1145/1401132.1401250 %0 Conference Paper %B Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on %D 2008 %T Stochastic fusion of multi-view gradient fields %A Sankaranarayanan,A. C %A Chellapa, Rama %K application;brightness;cameras;gradient %K application;image %K camera %K estimators;multiview %K fields;projective %K fusion;image %K fusion;textured %K gradient %K gradients;linear %K gradients;scene %K imaging;scene %K map;stochastic %K methods;image %K noise;graphics %K planar %K radiance;scene %K reconstruction;image %K reconstruction;scene %K scene;vision %K TEXTURE %K texture; %K view;corrupting %X Image gradients form powerful cues in a host of vision and graphics applications. In this paper, we consider multiple views of a textured planar scene and consider the problem of estimating the scene texture map using these multi-view inputs. Modeling each camera view as a projective transformation of the scene, we show that the problem is equivalent to that of studying the effect of noise (and the projective imaging) on the gradient fields induced by this texture map. We show that these noisy gradient fields can be modeled as complete observers of the scene radiance. Further, the corrupting noise can be shown to be additive and linear, although spatially varying. However, the specific form of the noise term can be exploited to design linear estimators that fuse the gradient fields obtained from each of the individual views. The fused gradient field forms a robust estimate of the scene gradients and can be used for scene reconstruction. %B Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on %P 1324 - 1327 %8 2008/10// %G eng %R 10.1109/ICIP.2008.4712007 %0 Journal Article %J Proceedings of ICCCD %D 2008 %T Stochastic opponent modeling agents: A case study with Hamas %A Mannes,A. %A Sliva,A. %A V.S. Subrahmanian %A Wilkenfeld,J. %X Stochastic Opponent Modeling Agents (SOMA) have beenproposed as a paradigm for reasoning about cultural groups, terror groups, and other socio-economic-political-military organizations worldwide. In this paper, we describe a case study that shows how SOMA was used to model the behavior of the terrorist organization, Hamas. Our team, consisting of a mix of computer scientists, policy experts, and political scientists, were able to understand new facts about Hamas of which even seasoned Hamas experts may not have been aware. This paper briefly overviews SOMA rules, explains how several thousand SOMA rules for Hamas were automatically derived, and then describes a few key findings about Hamas, enabled by this framework. %B Proceedings of ICCCD %8 2008/// %G eng %0 Journal Article %J Probabilistic, Logical and Relational Learning-A Further Synthesis %D 2008 %T Structural Sampling for Statistical Software Testing %A Baskiotis,N. %A Sebag,M. %A De Raedt,L. %A Dietterich,T. %A Getoor, Lise %A Kersting,K. %A Muggleton,S. H %X Structural Statistical Software Testing exploits the control flow graph of the program being tested to construct test cases. While test cases can easily be extracted from {em feasible paths} in the control flow graph, that is, paths which are actually exerted for some values of the program input, the feasible path region is a tiny fraction of the graph paths (less than $10^{-5}]$ for medium size programs). The S4T algorithm presented in this paper aims to address this limitation; as an Active Relational Learning Algorithm, it uses the few feasible paths initially available to sample new feasible paths. The difficulty comes from the non-Markovian nature of the feasible path concept, due to the long-range dependencies between the nodes in the control flow graph. Experimental validation on real-world and artificial problems demonstrates significant improvements compared to the state of the art. %B Probabilistic, Logical and Relational Learning-A Further Synthesis %8 2008/// %G eng %0 Conference Paper %D 2008 %T A Survey of the Virtual Environments-based Assembly Training Applications %A Gupta,S.K. %A Anand,D. K. %A Brough,J. E. %A Kavetsky,R. A. %A Schwartz,M. %A Thakur,A. %X The advent of virtual environments is presenting new ways of trainingtomorrow’s workforce. Virtual environments offer numerous benefits in training applications. First, virtual environments allow extensive user interactions in a very convenient and natural manner. This interaction is greatly beneficial for increasing the user’s retention of spatial information compared to text-based or video-based instructions that are non-interactive in nature. Second, virtual environments provide users with a 3D immersive experience. This feature helps users gain a better understanding of spatial relationships compared to 2D displays. Third, virtual environments support multi-media instructions. One can watch standard videos, view 3D animations, view text instructions, listen to audio instructions, and interact with 3D objects in the scene. This paper describes representative research and associated systems that use of virtual environments in assembly training applications. %8 2008/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.151.2892&rep=rep1&type=pdf %0 Conference Paper %B Proceedings of the 13th international conference on Intelligent user interfaces %D 2008 %T Systematic yet flexible discovery: guiding domain experts through exploratory data analysis %A Perer,Adam %A Shneiderman, Ben %K exploratory data analysis %K guides %K Information Visualization %K social networks %K systematic yet flexible %K wizards %X During exploratory data analysis, visualizations are often useful for making sense of complex data sets. However, as data sets increase in size and complexity, static information visualizations decrease in comprehensibility. Interactive techniques can yield valuable discoveries, but current data analysis tools typically support only opportunistic exploration that may be inefficient and incomplete. We present a refined architecture that uses systematic yet flexible (SYF) design goals to guide domain expert users through complex exploration of data over days, weeks and months. The SYF system aims to support exploratory data analysis with some of the simplicity of an e-commerce check-out while providing added flexibility to pursue insights. The SYF system provides an overview of the analysis process, suggests unexplored states, allows users to annotate useful states, supports collaboration, and enables reuse of successful strategies. The affordances of the SYF system are demonstrated by integrating it into a social network analysis tool employed by social scientists and intelligence analysts. The SYF system is a tool-independent component and can be incorporated into other data analysis tools. %B Proceedings of the 13th international conference on Intelligent user interfaces %S IUI '08 %I ACM %C New York, NY, USA %P 109 - 118 %8 2008/// %@ 978-1-59593-987-6 %G eng %U http://doi.acm.org/10.1145/1378773.1378788 %R 10.1145/1378773.1378788 %0 Conference Paper %D 2008 %T System-Level Optimization Model for a Snake-Inspired Robot Based on a Rectilinear Gait %A Hopkins,James K. %A Spranklin,Brent W. %A Gupta, Satyandra K. %X Physical parameters of modules and gait parameters affect the overall snake-inspired robot performance. Hence the system-level optimization model has to concurrently optimize the module parameters and the gait. The equations of motion associated with the rectilinear gait are quite complex due to the changing topology of the rectilinear gait. Embedding these equations in the system-level optimization model leads to a computationally challenging formulation. This paper presents a system-level optimization model that utilizes a hierarchical optimization approach and meta-models of the pre-computed optimal gaits to reduce the complexity of the optimization model. This approach enabled us to use an experimentally validated physics-based model of the rectilinear gait and yet at the same time enabled us to create a system-level optimization model with a manageable complexity. A detailed case study is presented to show the importance of concurrently optimizing the module parameters and the gait using our model to obtain the optimal performance for a given mission. %I ASME %P 913 - 924 %8 2008/// %@ 978-0-7918-4326-0 %G eng %U http://link.aip.org/link/ASMECP/v2008/i43260/p913/s1&Agg=doi %R 10.1115/DETC2008-49972 %0 Journal Article %J MetricsMATR workshop at AMTA %D 2008 %T TERp system description %A Snover,M. %A Madnani,N. %A Dorr, Bonnie J %A Schwartz,R. %X This paper describes TER-Plus (TERp) the University of Maryland / BBN Technologies submission for the NIST Metric MATR 2008 workshop on automatic machine translation evaluation metrics. TERp is an extension of Translation Edit Rate (TER) that builds off of the success of TER as an evaluation metric and alignment tool while addressing several of its weaknesses through the use of paraphrases, morphological stemming, and synonyms, as well as edit costs that are optimized to correlate better with various types of human judgments. %B MetricsMATR workshop at AMTA %8 2008/// %G eng %0 Conference Paper %D 2008 %T Towards Design and Fabrication of a Miniature MRI-Compatible Robot for Applications in Neurosurgery %A Pappafotis,Nicholas %A Bejgerowski,Wojciech %A Gullapalli,Rao %A Simard,J. Marc %A Gupta, Satyandra K. %A Desai,Jaydev P. %X Brain tumors are among the most feared complications of cancer and they occur in 20–40% of adult cancer patients. Despite numerous advances in treatment, the prognosis for these patients is poor, with a median survival of 4–8 months. The primary reasons for poor survival rate are the lack of good continuous imaging modality for intraoperative intracranial procedures and the inability to remove the complete tumor tissue due to its placement in the brain and the corresponding space constraints to reach it. Intraoperative magnetic resonance imaging (MRI) supplements the surgeon's visual and tactile senses in a way that no other imaging device can achieve resulting in less trauma to surrounding healthy brain tissue during surgery. To minimize the trauma to surrounding healthy brain tissue, it would be beneficial to operate through a narrow surgical corridor dissected by the neurosurgeon. Facilitating tumor removal by accessing regions outside the direct “line-of-sight” of the neurosurgical corridor will require a highly dexterous, small cross section, and MRI-compatible robot. Developing such a robot is extremely challenging task. In this paper we report a preliminary design of 6-DOF robot for possible application in neurosurgery. The robot actuators and body parts are constructed from MRI compatible materials. The current prototype is 0.36” in diameter and weighs only 0.0289 N (2.95 grams). The device was actuated using Flexinol® which is a shape memory alloy manufactured by Dynalloy, Inc. The end-effector forces ranged from 12 mN to 50 mN depending on the robot configuration. The end-effector force to robot weight ratio varied from 0.41 to 1.73. During trials the robot motion was repeatable and the range of motion of the robot was about 90 degrees for the end-effector when one side shape memory alloy (SMA) channel was actuated. The actuation time from the start to finish was about 2.5 s. %I ASME %P 747 - 754 %8 2008/// %@ 978-0-7918-4326-0 %G eng %U http://link.aip.org/link/ASMECP/v2008/i43260/p747/s1&Agg=doi %R 10.1115/DETC2008-49587 %0 Journal Article %J Future Generation Computer Systems %D 2008 %T Trade-offs in matching jobs and balancing load for distributed desktop grids %A Kim,Jik-Soo %A Nam,Beomseok %A Keleher,Peter %A Marsh,Michael %A Bhattacharjee, Bobby %A Sussman, Alan %K Desktop grid %K load balancing %K Matchmaking %K peer-to-peer computing %K Resource discovery %X Desktop grids can achieve tremendous computing power at low cost through opportunistic sharing of resources. However, traditional client–server Grid architectures do not deal with all types of failures, and do not always cope well with very dynamic environments. This paper describes the design of a desktop grid implemented over a modified Peer-to-Peer (P2P) architecture. The underlying P2P system is decentralized and inherently adaptable, giving the Grid robustness, scalability, and the ability to cope with dynamic environments, while still efficiently mapping application instances to available resources throughout the system.We use simulation to compare three different types of matching algorithms under differing workloads. Overall, the P2P approach produces significantly lower wait times than prior approaches, while adapting efficiently to the dynamic environment. %B Future Generation Computer Systems %V 24 %P 415 - 424 %8 2008/05// %@ 0167-739X %G eng %U http://www.sciencedirect.com/science/article/pii/S0167739X07001240 %N 5 %R 10.1016/j.future.2007.07.007 %0 Journal Article %J Antonie van Leeuwenhoek %D 2008 %T Transesterification activity of a novel lipase from Acinetobacter venetianus RAG-1 %A Snellman,E. A. %A Rita R Colwell %X Transesterification activity and the industrial potential of a novel lipase prepared from Acinetobacter ventiatus RAG-1 were evaluated. Purified lipase samples were dialyzed against pH 9.0 buffer in a single optimization step prior to lyophilization. The enzyme and organic phase were pre-equilibrated (separately) to the same thermodynamic water activities (a w) ranging from a w 0.33 to 0.97. Production of 1-octyl butyrate by lipase-catalyzed transesterification of vinyl butyrate with 1-octanol in hexane was monitored by gas chromatography. Production of 1-octyl butyrate and initial rate of reaction depended on water activity. Product synthesis and rate of transesterification increased sharply with increase from a w 0.33 to 0.55. Highest product concentration (218 mM) and rate of reaction (18.7 μmol h−1 · 10 μg protein) were measured at a w 0.86. Transesterification activity in hexane represented 32% of comparable hydrolytic activity in aqueous buffer. %B Antonie van Leeuwenhoek %V 94 %P 621 - 625 %8 2008/// %G eng %N 4 %R 10.1007/s10482-008-9276-5 %0 Journal Article %J Advances in Bioinformatics %D 2008 %T A Tutorial of the Poisson Random Field Model in Population Genetics %A Sethupathy,Praveen %A Hannenhalli, Sridhar %X Population genetics is the study of allele frequency changes driven by various evolutionary forces such as mutation, natural selection, and random genetic drift. Although natural selection is widely recognized as a bona-fide phenomenon, the extent to which it drives evolution continues to remain unclear and controversial. Various qualitative techniques, or so-called “tests of neutrality”, have been introduced to detect signatures of natural selection. A decade and a half ago, Stanley Sawyer and Daniel Hartl provided a mathematical framework, referred to as the Poisson random field (PRF), with which to determine quantitatively the intensity of selection on a particular gene or genomic region. The recent availability of large-scale genetic polymorphism data has sparked widespread interest in genome-wide investigations of natural selection. To that end, the original PRF model is of particular interest for geneticists and evolutionary genomicists. In this article, we will provide a tutorial of the mathematical derivation of the original Sawyer and Hartl PRF model. %B Advances in Bioinformatics %V 2008 %P 1 - 9 %8 2008/// %@ 1687-8027, 1687-8035 %G eng %U http://www.hindawi.com/journals/abi/2008/257864/ %R 10.1155/2008/257864 %0 Journal Article %J Software, IEEE %D 2008 %T Understanding the High-Performance-Computing Community: A Software Engineer's Perspective %A Basili, Victor R. %A Carver, J. C %A Cruzes, D. %A Hochstein, L. M %A Hollingsworth, Jeffrey K %A Shull, F. %A Zelkowitz, Marvin V %K community;software %K computational %K engineering; %K engineering;software %K Science %K software;high-performance-computing %X Computational scientists developing software for HPC systems face unique software engineering issues. Attempts to transfer SE technologies to this domain must take these issues into account. %B Software, IEEE %V 25 %P 29 - 36 %8 2008/08//july %@ 0740-7459 %G eng %N 4 %R 10.1109/MS.2008.103 %0 Journal Article %J Networking, IEEE/ACM Transactions on %D 2008 %T A Unified Framework for Multipath Routing for Unicast and Multicast Traffic %A Guven,T. %A La,R.J. %A Shayman,M.A. %A Bhattacharjee, Bobby %K application-layer %K approximation;unicast %K balancing;multicast %K channels;stochastic %K communication;multipath %K cost %K function;optimization %K network %K overlay;load %K perturbation %K problem;simultaneous %K processes;telecommunication %K routing;network %K routing;telecommunication %K Stochastic %K traffic; %K traffic;multicast %K traffic;multipath %X We study the problem of load balancing the traffic from a set of unicast and multicast sessions. The problem is formulated as an optimization problem. However, we assume that the gradient of the network cost function is not available and needs to be estimated. Multiple paths are provided between a source and a destination using application-layer overlay. We propose a novel algorithm that is based on what is known as simultaneous perturbation stochastic approximation and utilizes only noisy measurements collected and reported to the sources, using an overlay architecture. We consider three network models that reflect different sets of assumptions regarding multicast capabilities of the network. Using an analytical model we first prove the almost sure convergence of the algorithm to a corresponding optimal solution under each network model considered in this paper with decreasing step sizes. Then, we establish the weak convergence (or convergence in distribution) with a fixed step size. In addition, we investigate the benefits acquired from implementing additional multicast capabilities by studying the relative performance of our algorithm under the three network models. %B Networking, IEEE/ACM Transactions on %V 16 %P 1038 - 1051 %8 2008/10// %@ 1063-6692 %G eng %N 5 %R 10.1109/TNET.2007.909686 %0 Conference Paper %D 2008 %T On the Use of Security Metrics Based on Intrusion Prevention System Event Data: An Empirical Analysis %A Chrun,D. %A Michel Cukier %A Sneeringer,G. %K empirical analysis %K Internet %K Internet attack group %K intrusion prevention system event data %K network traffic monitoring %K organization security metrics %K security of data %X With the increasing number of attacks on the Internet, a primary concern for organizations is the protection of their network. To do so, organizations install security devices such as intrusion prevention systems to monitor network traffic. However, data that are collected by these devices are often imperfect. The contribution of this paper is to try to define some practical metrics based on imperfect data collected by an intrusion prevention system. Since attacks greatly differ, we propose to group the attacks into several attack type groups. We then define a set of metrics for each attack type group. We introduce an approach that consists in analyzing the evolution of these metrics per attack type group by focusing on outliers in order to give an insight into an organizationpsilas security. The method is assessed for an organization of about 40,000 computers. The results were encouraging: outliers could be related to security issues that, in some cases, had not been previously flagged. %P 49 - 58 %8 2008/12// %G eng %R 10.1109/HASE.2008.52 %0 Journal Article %J Information Processing & Management %D 2008 %T Users can change their web search tactics: Design guidelines for categorized overviews %A Kules,Bill %A Shneiderman, Ben %K categorization %K Categorized overviews %K Categorized search results %K Exploratory search %K Human-computer interaction %K information seeking %K Search result visualization %K Search user interfaces %K Web search %X Categorized overviews of web search results are a promising way to support user exploration, understanding, and discovery. These search interfaces combine a metadata-based overview with the list of search results to enable a rich form of interaction. A study of 24 sophisticated users carrying out complex tasks suggests how searchers may adapt their search tactics when using categorized overviews. This mixed methods study evaluated categorized overviews of web search results organized into thematic, geographic, and government categories. Participants conducted four exploratory searches during a 2-hour session to generate ideas for newspaper articles about specified topics such as “human smuggling.” Results showed that subjects explored deeper while feeling more organized, and that the categorized overview helped subjects better assess their results, although no significant differences were detected in the quality of the article ideas. A qualitative analysis of searcher comments identified seven tactics that participants reported adopting when using categorized overviews. This paper concludes by proposing a set of guidelines for the design of exploratory search interfaces. An understanding of the impact of categorized overviews on search tactics will be useful to web search researchers, search interface designers, information architects and web developers. %B Information Processing & Management %V 44 %P 463 - 484 %8 2008/03// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306457307001574 %N 2 %R 10.1016/j.ipm.2007.07.014 %0 Journal Article %J Probabilistic, Logical and Relational Learning-A Further Synthesis %D 2008 %T Variational Bayes via Propositionalization %A Sato,T. %A Kameya,Y. %A Kurihara,K. %A De Raedt,L. %A Dietterich,T. %A Getoor, Lise %A Kersting,K. %A Muggleton,S. H %B Probabilistic, Logical and Relational Learning-A Further Synthesis %8 2008/// %G eng %0 Conference Paper %B Visual Analytics Science and Technology, 2008. VAST '08. IEEE Symposium on %D 2008 %T VAST 2008 Challenge: Introducing mini-challenges %A Grinstein,G. %A Plaisant, Catherine %A Laskowski,S. %A O'Connell,T. %A Scholtz,J. %A Whiting,M. %K 2008 %K analytics;data %K Challenge;heterogeneous %K collections;visual %K data %K VAST %K visualisation; %X Visual analytics experts realize that one effective way to push the field forward and to develop metrics for measuring the performance of various visual analytics components is to hold an annual competition. The VAST 2008 Challenge is the third year that such a competition was held in conjunction with the IEEE Visual Analytics Science and Technology (VAST) symposium. The authors restructured the contest format used in 2006 and 2007 to reduce the barriers to participation and offered four mini-challenges and a Grand Challenge. Mini Challenge participants were to use visual analytic tools to explore one of four heterogeneous data collections to analyze specific activities of a fictitious, controversial movement. Questions asked in the Grand Challenge required the participants to synthesize data from all four data sets. In this paper we give a brief overview of the data sets, the tasks, the participation, the judging, and the results. %B Visual Analytics Science and Technology, 2008. VAST '08. IEEE Symposium on %P 195 - 196 %8 2008/10// %G eng %R 10.1109/VAST.2008.4677383 %0 Journal Article %J Proceedings of the 2008 Workshop on Programming Languages and Analysis for Security. ACM Press %D 2008 %T Verified enforcement of automaton-based information release policies %A Swamy,N. %A Hicks, Michael W. %X Many organizations specify information release policies to describe the terms under which sensitive information may bereleased to other organizations. This paper presents a new approach for ensuring that security-critical software correctly enforces its information release policy. Our approach has two parts. First, an information release policy is specified as a security automaton written in a new language called AIR. Second, we enforce an AIR policy by translating it into an API for programs written in λAIR, a core formalism for a functional programming language. λAIR uses a novel combination of dependent, affine, and singleton types to ensure that the API is used correctly. As a consequence we can certify that programs written in λAIR meet the requirements of the original AIR policy specification. %B Proceedings of the 2008 Workshop on Programming Languages and Analysis for Security. ACM Press %8 2008/// %G eng %0 Journal Article %J Environmental Microbiology %D 2008 %T Vibrio cholerae non‐O1, non‐O139 strains isolated before 1992 from Varanasi, India are multiple drug resistant, contain intSXT, dfr18 and aadA5 genes %A Mohapatra,Harapriya %A Mohapatra,Saswat S. %A Mantri,Chinmay K. %A Rita R Colwell %A Singh,Durg V. %X In this study, we report the presence of the SXT element and Class I integron in Vibrio cholerae non-O1, non-O139 strains isolated from Varanasi, India. Isolates were resistant to cotrimoxazole, trimethoprim and/or streptomycin, furazolidone and ampicillin. None contained plasmids. Polymerase chain reaction (PCR) and DNA sequencing revealed the presence of antibiotic resistance gene cassettes, aadA1, aadA2, aadA5 and dfrA15, in the Class I integron and SXT, an integrative conjugative element containing dfr18, sulII and strAB, in three and six of the isolates respectively. Conjugation experiments, followed by PCR analysis of transconjugants, provided evidence for the transferable nature of intSXT and associated antibiotic resistance gene cassettes. This is the first report of the occurrence of SXT ICE, dfr18, sulII, strAB and aadA5 genes in environmental V. cholerae non-O1, non-O139 strains from Varanasi, India, that had been isolated before 1992. %B Environmental Microbiology %V 10 %P 866 - 873 %8 2008/04/01/ %@ 1462-2920 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1462-2920.2007.01502.x/abstract?userIsAuthenticated=false&deniedAccessCustomisedMessage= %N 4 %R 10.1111/j.1462-2920.2007.01502.x %0 Book Section %B Handbook of Data VisualizationHandbook of Data Visualization %D 2008 %T Visualizing Functional Data with an Application to eBay’s Online Auctions %A Chen,Chun-houh %A Härdle,Wolfgang %A Unwin,Antony %A Jank,Wolfgang %A Shmueli,Galit %A Plaisant, Catherine %A Shneiderman, Ben %X Technological advances in the measurement, collection, and storage of data have led tomore andmore complex data structures. Examples of such structures include measurements of the behavior of individuals over time, digitized two- or three-dimensional images of the brain, and recordings of three- or even four-dimensional movements of objects traveling through space and time. Such data, although recorded in a discrete fashion, are usually thought of as continuous objects that are represented by functional relationships. This gave rise to functional data analysis (FDA), which was made popular by the monographs of Ramsay and Silverman (1997, 2002), where the center of interest is a set of curves, shapes, objects, or, more generally, a set of functional observations , in contrast to classical statistics where interest centers on a set of data vectors. In that sense, functional data is not only different from the data structure studied in classical statistics, but it actually generalizes it. Many of these new data structures require new statistical methods to unveil the information that they carry. %B Handbook of Data VisualizationHandbook of Data Visualization %S Springer Handbooks of Computational Statistics %I Springer Berlin Heidelberg %P 873 - 898 %8 2008/// %@ 978-3-540-33037-0 %G eng %U http://dx.doi.org/10.1007/978-3-540-33037-0_34 %0 Journal Article %J Nature biotechnology %D 2008 %T What are decision trees? %A Kingsford, Carl %A Salzberg,S. L %B Nature biotechnology %V 26 %P 1011 - 1011 %8 2008/// %G eng %N 9 %0 Journal Article %J Discrete Optimization %D 2008 %T Worst case examples of an exterior point algorithm for the assignment problem %A Charalampos Papamanthou %A Paparrizos, Konstantinos %A Samaras, Nikolaos %A Stergiou, Konstantinos %K Assignment problem %K Exterior point algorithm %K Worst case examples %X An efficient exterior point simplex type algorithm for the assignment problem has been developed by Paparrizos [K. Paparrizos, An infeasible (exterior point) simplex algorithm for assignment problems, Math. Program. 51 (1991) 45–54]. This algorithm belongs to the category of forest algorithms and solves an n × n assignment problem in at most n ( n − 1 ) 2 iterations and in at most O ( n 3 ) time. In this paper worst case examples are presented. Specifically, a systematic procedure to construct worst case assignment problems is presented for the exterior point algorithm. The algorithm applied to these examples executes exactly n ( n − 1 ) 2 iterations. This result verifies that the bound O ( n 3 ) is the best possible for the above-mentioned algorithm. %B Discrete Optimization %V 5 %P 605 - 614 %8 2008/08// %@ 1572-5286 %G eng %U http://www.sciencedirect.com/science/article/pii/S1572528608000030 %N 3 %! Discrete Optimization %0 Journal Article %J Magazine interactions - 25 years of CHI conferences: a photographic essay %D 2007 %T 25 years of CHI conferences: capturing the exchange of ideas %A Shneiderman, Ben %B Magazine interactions - 25 years of CHI conferences: a photographic essay %V 14 %P 24 - 31 %8 2007/03// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/1229863.1229865 %N 2 %R 10.1145/1229863.1229865 %0 Journal Article %J Science(Washington) %D 2007 %T 911. gov %A Shneiderman, Ben %A Preece,J. %B Science(Washington) %V 315 %P 944 - 944 %8 2007/// %G eng %N 5814 %0 Conference Paper %B Proceedings of the 3rd USENIX workshop on Steps to reducing unwanted traffic on the internet %D 2007 %T Accountability as a service %A Bender,Adam %A Spring, Neil %A Levin,Dave %A Bhattacharjee, Bobby %X We propose that accountability be a first-class network service, independent of addressing and routing. We design a scheme for allowing accountability services, rather than connectivity-providing ISPs, to vouch for traffic, allowing victims to report abuse, filter abusive traffic, and isolate malicious senders. We discuss how accountability services may evolve, how they may facilitate new applications, and the implications of shifting the burden of network policing to a dedicated service. %B Proceedings of the 3rd USENIX workshop on Steps to reducing unwanted traffic on the internet %S SRUTI'07 %I USENIX Association %C Berkeley, CA, USA %P 5:1–5:6 - 5:1–5:6 %8 2007/// %G eng %U http://dl.acm.org/citation.cfm?id=1361436.1361441 %0 Report %D 2007 %T ACE: A Novel Software Platform to Ensure the Integrity of Long Term Archives %A Song,Sangchul %A JaJa, Joseph F. %K Technical Report %X We develop a new methodology to address the integrity of long term archives using rigorous cryptographic techniques. A prototype system called ACE (Auditing Control Environment) was designed and developed based on this methodology. ACE creates a small-size integrity token for each digital object and some cryptographic summary information based on all the objects handled within a dynamic time period. ACE continuously audits the contents of the various objects according to the policy set by the archive, and provides mechanisms for an independent third-party auditor to certify the integrity of any object. In fact, our approach will allow an independent auditor to verify the integrity of every version of an archived digital object as well as link the current version to the original form of the object when it was ingested into the archive. We show that ACE is very cost effective and scalable while making no assumptions about the archive architecture. We include in this paper some preliminary results on the validation and performance of ACE on a large image collection. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-07 %8 2007/01/31/ %G eng %U http://drum.lib.umd.edu/handle/1903/4047 %0 Journal Article %J Parallel Computing %D 2007 %T Active semantic caching to optimize multidimensional data analysis in parallel and distributed environments %A Andrade,Henrique %A Kurc,Tahsin %A Sussman, Alan %A Saltz,Joel %K Active semantic caching %K Parallel databases %K Query optimization %K Scientific data analysis %X In this paper, we present a multi-query optimization framework based on the concept of active semantic caching. The framework permits the identification and transparent reuse of data and computation in the presence of multiple queries (or query batches) that specify user-defined operators and aggregations originating from scientific data-analysis applications. We show how query scheduling techniques, coupled with intelligent cache replacement policies, can further improve the performance of query processing by leveraging the active semantic caching operators. We also propose a methodology for functionally decomposing complex queries in terms of primitives so that multiple reuse sites are exposed to the query optimizer, to increase the amount of reuse. The optimization framework and the database system implemented with it are designed to be efficient irrespective of the underlying parallel and/or distributed machine configuration. We present experimental results highlighting the performance improvements obtained by our methods using real scientific data-analysis applications on multiple parallel and distributed processing configurations (e.g., single symmetric multiprocessor (SMP) machine, cluster of SMP nodes, and a Grid computing configuration). %B Parallel Computing %V 33 %P 497 - 520 %8 2007/08// %@ 0167-8191 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167819107000506 %N 7-8 %R 10.1016/j.parco.2007.03.001 %0 Book Section %B Scalable Uncertainty ManagementScalable Uncertainty Management %D 2007 %T Aggregates in Generalized Temporally Indeterminate Databases %A Udrea,Octavian %A Majkić,Zoran %A Subrahmanian,V. %E Prade,Henri %E Subrahmanian,V. %K Computer %K Science %X Dyreson and Snodgrass as well as Dekhtyar et. al. have provided a probabilistic model (as well as compelling example applications) for why there may be temporal indeterminacy in databases. In this paper, we first propose a formal model for aggregate computation in such databases when there is uncertainty not just in the temporal attribute, but also in the ordinary (non-temporal) attributes. We identify two types of aggregates: event correlated aggregates, and non event correlated aggregations, and provide efficient algorithms for both of them. We prove that our algorithms are correct, and we present experimental results showing that the algorithms work well in practice. %B Scalable Uncertainty ManagementScalable Uncertainty Management %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4772 %P 171 - 186 %8 2007/// %@ 978-3-540-75407-7 %G eng %U http://dx.doi.org/10.1007/978-3-540-75410-7_13 %0 Conference Paper %B Image Processing, 2007. ICIP 2007. IEEE International Conference on %D 2007 %T Analysis of Nonlinear Collusion Attacks on Fingerprinting Systems for Compressed Multimedia %A Varna,A.L. %A He,Shan %A Swaminathan,A. %A M. Wu %K attacks;data %K coding;multimedia %K coding;security %K collusion %K compressed %K compression;image %K data;watermarking; %K fingerprinting %K multimedia;data %K of %K systems;nonlinear %X In this paper, we analyze the effect of various collusion attacks on fingerprinting systems for compressed multimedia. We evaluate the effectiveness of the collusion attacks in terms of the probability of detection and accuracy in estimating the host signal. Our analysis shows that applying averaging collusion on copies of moderately compressed content gives a highly accurate estimation of the host, and can effectively remove the embedded fingerprints. Averaging is thus the best choice for an attacker as the probability of detection and the distortion introduced are the lowest. %B Image Processing, 2007. ICIP 2007. IEEE International Conference on %V 2 %P II -133 -II -136 - II -133 -II -136 %8 2007/10/16/19 %G eng %R 10.1109/ICIP.2007.4379110 %0 Report %D 2007 %T Analysis of the Residual Arnoldi Method %A Lee,Che-Rung %A Stewart, G.W. %K Technical Report %X The Arnoldi method generates a nested squences of orthonormal bases$U_{1},U_{2}, \ldots$ by orthonormalizing $Au_{k}$ against $U_{k}$. Frequently these bases contain increasingly accurate approximations of eigenparis from the periphery of the spectrum of $A$. However, the convergence of these approximations stagnates if $u_{k}$ is contaminated by error. It has been observed that if one chooses a Rayleigh--Ritz approximation $(\mu_{k}, z_{k})$ to a chosen target eigenpair $(\lambda, x)$ and orthonormalizes the residual $Az_{k - }\mu_{k} z_{k}$, the approximations to $x$ (but not the other eigenvectors) continue to converge, even when the residual is contaminated by error. The same is true of the shift-invert variant of Arnoldi's method. In this paper we give a mathematical analysis of these new methods. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-45 %8 2007/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/7428 %0 Journal Article %J Technical Reports of the Computer Science Department %D 2007 %T Appendix to CMod: Modular Information Hiding and Type-Safe Linking for C %A Srivastava,S. %A Hicks, Michael W. %A Foster, Jeffrey S. %X This brief note is an appendix to the paper "CMod: Modular Information Hiding and Type-Safe Linking for C." It consists of the proof of soundness for the formal language presented in that paper. %B Technical Reports of the Computer Science Department %8 2007/06/30/undef %G eng %0 Conference Paper %B Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms %D 2007 %T Approximation algorithms for stochastic and risk-averse optimization %A Srinivasan, Aravind %X We present improved approximation algorithms in stochastic optimization. We prove that the multi-stage stochastic versions of covering integer programs (such as set cover and vertex cover) admit essentially the same approximation algorithms as their standard (non-stochastic) counterparts; this improves upon work of Swamy & Shmoys that shows an approximability which depends multiplicatively on the number of stages. We also present approximation algorithms for facility location and some of its variants in the 2-stage recourse model, improving on previous approximation guarantees. %B Proceedings of the eighteenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '07 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 1305 - 1313 %8 2007/// %@ 978-0-898716-24-5 %G eng %U http://dl.acm.org/citation.cfm?id=1283383.1283523 %0 Journal Article %J IEEE internet computing %D 2007 %T Autonomic computing %A Brabec,F. %A Samet, Hanan %B IEEE internet computing %V 11 %P 52 - 59 %8 2007/// %G eng %N 1 %0 Journal Article %J interactions %D 2007 %T BELIV'06: beyond time and errors: novel evaluation methods for information visualization %A Bertini,Enrico %A Plaisant, Catherine %A Santucci,Giuseppe %B interactions %V 14 %P 59 - 60 %8 2007/05// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/1242421.1242460 %N 3 %R 10.1145/1242421.1242460 %0 Conference Paper %B Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on %D 2007 %T Bilattice-based Logical Reasoning for Human Detection %A Shet,V. D %A Neumann, J. %A Ramesh,V. %A Davis, Larry S. %K automated %K detection;parts-based %K detectors;static %K histograms;human %K images;formal %K interactions;gradient %K logic;image %K logical %K mechanisms;surveillance;video %K processing; %K reasoning;complex %K recognition;inference %K signal %K Surveillance %K systems;bilattice-based %K visual %X The capacity to robustly detect humans in video is a critical component of automated visual surveillance systems. This paper describes a bilattice based logical reasoning approach that exploits contextual information and knowledge about interactions between humans, and augments it with the output of different low level detectors for human detection. Detections from low level parts-based detectors are treated as logical facts and used to reason explicitly about the presence or absence of humans in the scene. Positive and negative information from different sources, as well as uncertainties from detections and logical rules, are integrated within the bilattice framework. This approach also generates proofs or justifications for each hypothesis it proposes. These justifications (or lack thereof) are further employed by the system to explain and validate, or reject potential hypotheses. This allows the system to explicitly reason about complex interactions between humans and handle occlusions. These proofs are also available to the end user as an explanation of why the system thinks a particular hypothesis is actually a human. We employ a boosted cascade of gradient histograms based detector to detect individual body parts. We have applied this framework to analyze the presence of humans in static images from different datasets. %B Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on %P 1 - 8 %8 2007/06// %G eng %R 10.1109/CVPR.2007.383133 %0 Journal Article %J Nucleic Acids Research %D 2007 %T BIPASS: BioInformatics Pipeline Alternative Splicing Services %A Legendre,C. %A Raschid, Louiqa %A Snyder,B. %B Nucleic Acids Research %V 35 %P W292 - W292 %8 2007/// %G eng %N Web Server issue %0 Journal Article %J Proc. Joint Workshop on The Economics of Networked Systems and Incentive-Based Computing (NetEcon) %D 2007 %T Bitstore: An incentive-compatible solution for blocked downloads in bittorrent %A Ramachandran,A. %A Sarma,A.D. %A Feamster, Nick %X As many as 30% of all files shared on public BitTorrentnetworks suffer from the lack of “seeders” (peers that have complete copies of the file being shared); peers attempting to download such a file (“leechers”) may have to wait indef- initely to obtain certain file chunks that are not distributed in the file’s network of peers (the “swarm”). We call this the Blocked Leecher Problem (BLP). To alleviate BLP, we propose BitStore, a larger, secure network of BitTorrent users (not necessarily all sharing the same content) where nodes offer their resources (such as disk space and bandwidth) for public use. Peers sharing any file can use the storage network to maintain replicas for each chunk of the file. Any leecher seeking chunks that are absent from in its own swarm can query the public network, locate the node storing the said chunks, and retrieve them. BitStore also provides robust incentives for nodes contributing re- sources: In return for storing and serving chunks, such nodes can negotiate micropayments using a second-price auction. Peers who receive these credits may later use them to re- trieve blocks they need from the storage network. This paper quantifies the BLP, presents an overview of the BitStore design, and discusses various challenges related to storage management and incentives. %B Proc. Joint Workshop on The Economics of Networked Systems and Incentive-Based Computing (NetEcon) %8 2007/// %G eng %0 Journal Article %J NetE-con+ IBC %D 2007 %T Boycotting and extorting nodes in an internetwork %A Levin,D. %A Bender,A. %A Lumezanu,C. %A Spring, Neil %A Bhattacharjee, Bobby %B NetE-con+ IBC %8 2007/// %G eng %0 Conference Paper %B Proceedings of the 6th ACM SIGCHI conference on Creativity & cognition %D 2007 %T Bridging art and science with creativity support tools %A Shneiderman, Ben %A Rita R Colwell %A Diamond,Sara %A Greenhalgh,Paul %A Wulf,William %B Proceedings of the 6th ACM SIGCHI conference on Creativity & cognition %S C&C '07 %I ACM %C New York, NY, USA %P 309 - 309 %8 2007/// %@ 978-1-59593-712-4 %G eng %U http://doi.acm.org/10.1145/1254960.1255044 %R 10.1145/1254960.1255044 %0 Book Section %B Document Understanding Conference DUC 2007Document Understanding Conference DUC 2007 %D 2007 %T Bridging the ROUGE/Human Evaluation Gap in Multi-Document Summarization %A Conroy,John M. %A Schlesinger,Judith D. %A O'Leary, Dianne P. %B Document Understanding Conference DUC 2007Document Understanding Conference DUC 2007 %I HLT-NAACL %C Rochester, NY %8 2007/04/26/ %G eng %0 Journal Article %J Information Systems %D 2007 %T Browsing large online data tables using generalized query previews %A Tanin,Egemen %A Shneiderman, Ben %A Xie,Hairuo %K Information Visualization %K Internet %K Online querying %K World-wide web %X Companies, government agencies, and other organizations are making their data available to the world over the Internet. They often use large online relational tables for this purpose. Users query such tables with front-ends that typically use menus or form fillin interfaces, but these interfaces rarely give users information about the contents and distribution of the data. Such a situation leads users to waste time and network/server resources posing queries that have zero- or mega-hit results. Generalized query previews enable efficient browsing of large online data tables by supplying data distribution information to users. The data distribution information provides continuous feedback about the size of the result set as the query is being formed. Our paper presents a new user interface architecture and discusses three controlled experiments (with 12, 16, and 48 participants). Our prototype systems provide flexible user interfaces for research and testing of the ideas. The user studies show that for exploratory querying tasks, generalized query previews can speed user performance for certain user domains and can reduce network/server load. %B Information Systems %V 32 %P 402 - 423 %8 2007/05// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306437905001213 %N 3 %R 10.1016/j.is.2005.12.006 %0 Report %D 2007 %T Building a Better Mousetrap %A Ramachandran,A. %A Seetharaman,S. %A Feamster, Nick %A Vazirani,V. %X Routers in the network core are unable to maintain detailed statistics for every packet; thus, traffic statistics are often based on packet sampling, which reduces accuracy. Because tracking large ("heavy-hitter") traffic flows is important both for pricing and for traffic engineering, much attention has focused on maintaining accurate statistics for such flows, often at the expense of small-volume flows. Eradicating these smaller flows makes it difficult to observe communication structure, which is sometimes more important than maintaining statistics about flow sizes. This paper presents FlexSample, a sampling framework that allows network operators to get the best of both worlds: For a fixed sampling budget, FlexSample can capture significantly more small-volume flows for only a small increase in relative error of large traffic flows. FlexSample uses a fast, lightweight counter array that provides a coarse estimate of the size ("class") of each traffic flow; a router then can sample at different rates according to the class of the traffic using any existing sampling strategy. Given a fixed sampling rate and a target fraction of sampled packets to allocate across traffic classes, FlexSample computes packet sampling rates for each class that achieve these allocations online. Through analysis and trace-based experiments, we find that FlexSample captures at least 50% more mouse flows than strategies that do not perform class-dependent packet sampling. We also show how FlexSample can be used to capture unique flows for specific applications. %I Georgia Institute of Technology. College of Computing %V GIT-CSS-07-01 %8 2007/// %G eng %U http://hdl.handle.net/1853/14350 %0 Journal Article %J International Journal of Human-] Computer Interaction %D 2007 %T Capture, annotate, browse, find, share: novel interfaces for personal photo management %A Kang,H. %A Bederson, Benjamin B. %A Suh,B. %B International Journal of Human-] Computer Interaction %V 23 %P 315 - 337 %8 2007/// %G eng %N 3 %0 Journal Article %J IEEE Intelligent Systems %D 2007 %T CARA: A Cultural-Reasoning Architecture %A V.S. Subrahmanian %A Albanese,Massimiliano %A Martinez,Mar?a Vanina %A Nau, Dana S. %A Reforgiato,Diego %A Simari,Gerardo I. %A Sliva,Amy %A Udrea,Octavian %A Wilkenfeld,Jonathan %K cultural reasoning %K opinion extraction %K opponent modeling %X The Cognitive Architecture for Reasoning about Adversaries supports methods to gather data about different cultural groups and learn the intensity of those groups' opinions on various topics. It also lets users build and extract models of those groups' behavior and continuously refine those models through shared multiperson learning experiences. Researchers have applied the architecture to develop cultural-reasoning applications for example, a game that teaches soldiers about cultural norms in an Afghan village. %B IEEE Intelligent Systems %V 22 %P 12 - 16 %8 2007/// %@ 1541-1672 %G eng %N 2 %0 Journal Article %J IEEE Transactions on Mobile Computing %D 2007 %T Cell breathing in wireless LANs: Algorithms and evaluation %A Hajiaghayi, Mohammad T. %A Mirrokni,S. V. %A Saberi,A. %A Bahl,P. %A Jain,K. %A Qiu,L. %B IEEE Transactions on Mobile Computing %V 6 %P 164 - 178 %8 2007/// %G eng %N 2 %0 Journal Article %J Journal of Biological Chemistry %D 2007 %T Characterization of Ehp, a Secreted Complement Inhibitory Protein from Staphylococcus aureus %A Hammel,Michal %A Sfyroera,Georgia %A Pyrpassopoulos,Serapion %A Ricklin,Daniel %A Ramyar,Kasra X. %A Pop, Mihai %A Jin,Zhongmin %A Lambris,John D. %A Geisbrecht,Brian V. %X We report here the discovery and characterization of Ehp, a new secreted Staphylococcus aureus protein that potently inhibits the alternative complement activation pathway. Ehp was identified through a genomic scan as an uncharacterized secreted protein from S. aureus, and immunoblotting of conditioned S. aureus culture medium revealed that the Ehp protein was secreted at the highest levels during log-phase bacterial growth. The mature Ehp polypeptide is composed of 80 residues and is 44% identical to the complement inhibitory domain of S. aureus Efb (extracellular fibrinogen-binding protein). We observed preferential binding by Ehp to native and hydrolyzed C3 relative to fully active C3b and found that Ehp formed a subnanomolar affinity complex with these various forms of C3 by binding to its thioester-containing C3d domain. Site-directed mutagenesis demonstrated that Arg75 and Asn82 are important in forming the Ehp·C3d complex, but loss of these side chains did not completely disrupt Ehp/C3d binding. This suggested the presence of a second C3d-binding site in Ehp, which was mapped to the proximity of Ehp Asn63. Further molecular level details of the Ehp/C3d interaction were revealed by solving the 2.7-Å crystal structure of an Ehp·C3d complex in which the low affinity site had been mutationally inactivated. Ehp potently inhibited C3b deposition onto sensitized surfaces by the alternative complement activation pathway. This inhibition was directly related to Ehp/C3d binding and was more potent than that seen for Efb-C. An altered conformation in Ehp-bound C3 was detected by monoclonal antibody C3-9, which is specific for a neoantigen exposed in activated forms of C3. Our results suggest that increased inhibitory potency of Ehp relative to Efb-C is derived from the second C3-binding site in this new protein. %B Journal of Biological Chemistry %V 282 %P 30051 - 30061 %8 2007/10/12/ %G eng %U http://www.jbc.org/content/282/41/30051.abstract %N 41 %R 10.1074/jbc.M704247200 %0 Report %D 2007 %T Chit-based access control %A Keleher,P. J %A Spring, Neil %A Bhattacharjee, Bobby %I Technical Report CS-TR-4878, University of Maryland at College Park %8 2007/// %G eng %0 Conference Paper %B Content-Based Multimedia Indexing, 2007. CBMI '07. International Workshop on %D 2007 %T Classifying Computer Generated Charts %A Prasad,V. S.N %A Siddiquie,B. %A Golbeck,J. %A Davis, Larry S. %K algorithm;scale %K analysis;visual %K classification;image %K database;image %K databases; %K feature %K Internet;bar-chart;curve-plot;image %K invariant %K match %K matching;image %K relationship;surface-plot;Internet;image %K representation;image %K segmentation;pie-chart;pyramid %K segmentation;statistical %K transform;scatter-plot;spatial %X We present an approach for classifying images of charts based on the shape and spatial relationships of their primitives. Five categories are considered: bar-charts, curve-plots, pie-charts, scatter-plots and surface-plots. We introduce two novel features to represent the structural information based on (a) region segmentation and (b) curve saliency. The local shape is characterized using the Histograms of Oriented Gradients (HOG) and the Scale Invariant Feature Transform (SIFT) descriptors. Each image is represented by sets of feature vectors of each modality. The similarity between two images is measured by the overlap in the distribution of the features -measured using the Pyramid Match algorithm. A test image is classified based on its similarity with training images from the categories. The approach is tested with a database of images collected from the Internet. %B Content-Based Multimedia Indexing, 2007. CBMI '07. International Workshop on %P 85 - 92 %8 2007/06// %G eng %R 10.1109/CBMI.2007.385396 %0 Journal Article %J IEEE Internet Computing %D 2007 %T Client-Based Spatial Browsing on the World Wide Web %A Brabec,Frantisek %A Samet, Hanan %K client-server spatial browsing %K geographic information systems (GIS) %K Spatial databases %K web access %K web-based mapping %X Being able to visualize both spatial and nonspatial data is becoming increasingly important to today's Internet users. Spatial data viewers and query tools can aid in visualization, but they should also let users access data instantly and with minimal effort. The authors explore new ways to allow visualization of data stored in a central server database on a simple client. They also consider usage scenarios in which transferring the whole database to the client for processing isn't feasible due to the amount of data on the server, insufficient computing power on the client, and a slow link between the two. %B IEEE Internet Computing %V 11 %P 52 - 59 %8 2007/// %@ 1089-7801 %G eng %N 1 %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference on %D 2007 %T Collusion-Resistant Fingerprinting for Compressed Multimedia Signals %A Varna,A.L. %A He,Shan %A Swaminathan,A. %A M. Wu %A Lu,Haiming %A Lu,Zengxiang %K based %K compression;Gaussian %K compression;multimedia %K dithering;collusion-resistant %K fingerprinting;multimedia %K Gaussian %K processes;data %K sequences;anticollusion %K signals %K spectrum %K spread %K systems; %X Most existing collusion-resistant fingerprinting techniques are for fingerprinting uncompressed signals. In this paper, we first study the performance of the traditional Gaussian based spread spectrum sequences for fingerprinting compressed signals and show that the system can be easily defeated by averaging or taking the median of a few copies. To overcome the collusion problem for compressed multimedia host signals, we propose a technique called anti-collusion dithering to mimic an uncompressed signal. Results show higher probability of catching a colluder using the proposed scheme compared to using Gaussian based fingerprints. %B Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference on %V 2 %P II-165 -II-168 - II-165 -II-168 %8 2007/04// %G eng %R 10.1109/ICASSP.2007.366198 %0 Journal Article %J Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Proceedings of the Main Conference %D 2007 %T Combining outputs from multiple machine translation systems %A Rosti,A.V.I. %A Ayan,N.F. %A Xiang,B. %A Matsoukas,S. %A Schwartz,R. %A Dorr, Bonnie J %X Currently there are several approaches tomachine translation (MT) based on differ- ent paradigms; e.g., phrasal, hierarchical and syntax-based. These three approaches yield similar translation accuracy despite using fairly different levels of linguistic knowledge. The availability of such a variety of systems has led to a growing interest toward finding better translations by combining outputs from multiple sys- tems. This paper describes three differ- ent approaches to MT system combina- tion. These combination methods oper- ate on sentence, phrase and word level exploiting information from N -best lists, system scores and target-to-source phrase alignments. The word-level combination provides the most robust gains but the best results on the development test sets (NIST MT05 and the newsgroup portion of GALE 2006 dry-run) were achieved by combining all three methods. %B Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Proceedings of the Main Conference %P 228 - 235 %8 2007/// %G eng %0 Journal Article %J On-line Proceedings of the Workshop on Principles of Provenance (PrOPr) %D 2007 %T Combining provenance and security policies in a web-based document management system %A Corcoran,B.J. %A Swamy,N. %A Hicks, Michael W. %B On-line Proceedings of the Workshop on Principles of Provenance (PrOPr) %8 2007/// %G eng %0 Journal Article %J Telecommunications Policy %D 2007 %T Community response grids: E-government, social networks, and effective emergency management %A Jaeger,Paul T. %A Shneiderman, Ben %A Fleischmann,Kenneth R. %A Preece,Jennifer %A Qu,Yan %A Fei Wu,Philip %K Community response grid %K E-government %K Emergency response %K Mobile communications %K Public policy %K social networks %X This paper explores the concept of developing community response grids (CRGs) for community emergency response and the policy implications of such a system. CRGs make use of the Internet and mobile communication devices, allowing residents and responders to share information, communicate, and coordinate activities in response to a major disaster. This paper explores the viability of using mobile communication technologies and the Web, including e-government, to develop response systems that would aid communities before, during, and after a major disaster, providing channels for contacting residents and responders, uploading information, distributing information, coordinating the responses of social networks, and facilitating resident-to-resident assistance. Drawing upon research from computer science, information studies, public policy, emergency management, and several other disciplines, the paper elaborates on the concept of and need for CRGs, examines related current efforts that can inform the development of CRGs, discusses how research about community networks can be used to instill trust and social capital in CRGs, and examines the issues of public policy, telecommunications, and e-government related to such a system. %B Telecommunications Policy %V 31 %P 592 - 604 %8 2007/11// %@ 0308-5961 %G eng %U http://www.sciencedirect.com/science/article/pii/S0308596107000699 %N 10–11 %R 10.1016/j.telpol.2007.07.008 %0 Journal Article %J Proceedings of the 13th Americas Conference on Information Systems %D 2007 %T Community response grids for older adults: Motivations, usability, and sociability %A Wu,P.F. %A Preece,J. %A Shneiderman, Ben %A Jaeger,P. T %A Qu,Y. %X This paper discusses the motivation for a Community Response Grid (CRG) to helpolder adults improve their capability for coping with emergency situations. We define and discuss the concept of a CRG, briefly review the limits of current emergency response systems, and identify usability and sociability guidelines for CRGs for older adults based on existing research. The paper ends with a call to action and suggestions for future research directions. %B Proceedings of the 13th Americas Conference on Information Systems %8 2007/// %G eng %0 Journal Article %J Biosecurity and Bioterrorism: Biodefense Strategy, Practice, and Science %D 2007 %T Community Response Grids: Using Information Technology to Help Communities Respond to Bioterror Emergencies %A Jaeger,Paul T. %A Fleischmann,Kenneth R. %A Preece,Jennifer %A Shneiderman, Ben %A Fei Wu,Philip %A Qu,Yan %X Access to accurate and trusted information is vital in preparing for, responding to, and recovering from an mergency. To facilitate response in large-scale emergency situations, Community Response Grids (CRGs) integrate Internet and mobile technologies to enable residents to report information, professional emergency responders to disseminate instructions, and residents to assist one another. CRGs use technology to help residents and professional emergency responders to work together in community response to emergencies, including bioterrorism events. In a time of increased danger from bioterrorist threats, the application of advanced information and communication technologies to community response is vital in confronting such threats. This article describes CRGs, their underlying concepts, development efforts, their relevance to biosecurity and bioterrorism, and future research issues in the use of technology to facilitate community response. %B Biosecurity and Bioterrorism: Biodefense Strategy, Practice, and Science %V 5 %P 335 - 346 %8 2007/12// %@ 1538-7135, 1557-850X %G eng %U http://www.liebertonline.com/doi/abs/10.1089/bsp.2007.0034 %N 4 %R 10.1089/bsp.2007.0034 %0 Conference Paper %B Circuits and Systems, 2007. ISCAS 2007. IEEE International Symposium on %D 2007 %T Compact, low power wireless sensor network system for line crossing recognition %A Shen,C. C %A Kupershtok,R. %A Yang,B. %A Vanin,F. M %A Shao,X. %A Sheth,D. %A Goldsman,N. %A Balzano,Q. %A Bhattacharyya, Shuvra S. %B Circuits and Systems, 2007. ISCAS 2007. IEEE International Symposium on %P 2506 - 2509 %8 2007/// %G eng %0 Conference Paper %B Multimedia Signal Processing, 2007. MMSP 2007. IEEE 9th Workshop on %D 2007 %T A Component Estimation Framework for Information Forensics %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K component %K data %K data; %K Estimation %K estimation;image %K forensics;parameter %K framework;digital %K integrity;imaging %K of %K processing;security %K technologies;information %X With a rapid growth of imaging technologies and an increasingly widespread usage of digital images and videos for a large number of high security and forensic applications, there is a strong need for techniques to verify the source and integrity of digital data. Component forensics is new approach for forensic analysis that aims to estimate the algorithms and parameters in each component of the digital device. In this paper, we develop a novel theoretical foundation to understand the fundamental performance limits of component forensics. We define formal notions of identifiability of components in the information processing chain, and present methods to quantify the accuracies at which the component parameters can be estimated. Building upon the proposed theoretical framework, we devise methods to improve the accuracies of component parameter estimation for a wide range of forensic applications. %B Multimedia Signal Processing, 2007. MMSP 2007. IEEE 9th Workshop on %P 397 - 400 %8 2007/10// %G eng %R 10.1109/MMSP.2007.4412900 %0 Journal Article %J Nucleic acids research %D 2007 %T COMPUTATIONAL BIOLOGY %A Leparc,GG %A Mitra,RD %A Vardhanabhuti,S. %A Wang,J. %A Hannenhalli, Sridhar %A Smit,S. %A Widmann,J. %A Knight,R. %A Wu,S. %A Zhang,Y. %A others %B Nucleic acids research %V 35 %P 3163 - 3163 %8 2007/// %G eng %N 9-12 %0 Journal Article %J BMC Bioinformatics %D 2007 %T A computational survey of candidate exonic splicing enhancer motifs in the model plant Arabidopsis thaliana %A Pertea,Mihaela %A Mount, Stephen M. %A Salzberg,Steven L. %X Algorithmic approaches to splice site prediction have relied mainly on the consensus patterns found at the boundaries between protein coding and non-coding regions. However exonic splicing enhancers have been shown to enhance the utilization of nearby splice sites. %B BMC Bioinformatics %V 8 %P 159 - 159 %8 2007/05/21/ %@ 1471-2105 %G eng %U http://www.biomedcentral.com/1471-2105/8/159/abstract %N 1 %R 10.1186/1471-2105-8-159 %0 Journal Article %J Annals of Mathematics and Artificial Intelligence %D 2007 %T Computing most probable worlds of action probabilistic logic programs: scalable estimation for 10^30,000 worlds %A Khuller, Samir %A Martinez,M. %A Nau, Dana S. %A Sliva,Amy %A Simari,Gerardo %A Subrahmanian,V. %K Computer science %X The semantics of probabilistic logic programs (PLPs) is usually given through a possible worlds semantics. We propose a variant of PLPs called action probabilistic logic programs or -programs that use a two-sorted alphabet to describe the conditions under which certain real-world entities take certain actions. In such applications, worlds correspond to sets of actions these entities might take. Thus, there is a need to find the most probable world (MPW) for -programs. In contrast, past work on PLPs has primarily focused on the problem of entailment. This paper quickly presents the syntax and semantics of -programs and then shows a naive algorithm to solve the MPW problem using the linear program formulation commonly used for PLPs. As such linear programs have an exponential number of variables, we present two important new algorithms, called and to solve the MPW problem exactly. Both these algorithms can significantly reduce the number of variables in the linear programs. Subsequently, we present a “binary” algorithm that applies a binary search style heuristic in conjunction with the Naive, and algorithms to quickly find worlds that may not be “most probable.” We experimentally evaluate these algorithms both for accuracy (how much worse is the solution found by these heuristics in comparison to the exact solution) and for scalability (how long does it take to compute). We show that the results of are very accurate and also very fast: more than 10 30,000 worlds can be handled in a few minutes. Subsequently, we develop parallel versions of these algorithms and show that they provide further speedups. %B Annals of Mathematics and Artificial Intelligence %V 51 %P 295 - 331 %8 2007/// %@ 1012-2443 %G eng %U http://www.springerlink.com/content/t70g873h613273pu/abstract/ %N 2 %R 10.1007/s10472-008-9089-2 %0 Journal Article %J Proceedings of the Workshop on Language Technology for Cultural Heritage Data (LaTeCH 2007) %D 2007 %T Concept disambiguation for improved subject access using multiple knowledge sources %A Sidhu,T. %A Klavans,J. %A Jimmy Lin %X We address the problem of mining text forrelevant image metadata. Our work is situ- ated in the art and architecture domain, where highly specialized technical vocabu- lary presents challenges for NLP tech- niques. To extract high quality metadata, the problem of word sense disambiguation must be addressed in order to avoid leading the searcher to the wrong image as a result of ambiguous—and thus faulty—meta- data. In this paper, we present a disam- biguation algorithm that attempts to select the correct sense of nouns in textual de- scriptions of art objects, with respect to a rich domain-specific thesaurus, the Art and Architecture Thesaurus (AAT). We per- formed a series of intrinsic evaluations us- ing a data set of 600 subject terms ex- tracted from an online National Gallery of Art (NGA) collection of images and text. Our results showed that the use of external knowledge sources shows an improvement over a baseline. %B Proceedings of the Workshop on Language Technology for Cultural Heritage Data (LaTeCH 2007) %P 25 - 25 %8 2007/// %G eng %0 Journal Article %J International Journal of Human-Computer Interaction %D 2007 %T Conducting In Situ Evaluations for and With Ubiquitous Computing Technologies %A Consolvo,Sunny %A Harrison,Beverly %A Smith,Ian %A Chen,Mike Y. %A Everitt,Katherine %A Jon Froehlich %A Landay,James A. %X Abstract To evaluate ubiquitous computing technologies, which may be embedded in the environment, embedded in objects, worn, or carried by the user throughout everyday life, it is essential to use methods that accommodate the often unpredictable, real-world environments in which the technologies are used. This article discusses how we have adapted and applied traditional methods from psychology and human-computer interaction, such as Wizard of Oz and Experience Sampling, to be more amenable to the in situ evaluations of ubiquitous computing applications, particularly in the early stages of design. The way that ubiquitous computing technologies can facilitate the in situ collection of self-report data is also discussed. Although the focus is on ubiquitous computing applications and tools for their assessment, it is believed that the in situ evaluation tools that are proposed will be generally useful for field trials of other technology, applications, or formative studies that are concerned with collecting data in situ.Abstract To evaluate ubiquitous computing technologies, which may be embedded in the environment, embedded in objects, worn, or carried by the user throughout everyday life, it is essential to use methods that accommodate the often unpredictable, real-world environments in which the technologies are used. This article discusses how we have adapted and applied traditional methods from psychology and human-computer interaction, such as Wizard of Oz and Experience Sampling, to be more amenable to the in situ evaluations of ubiquitous computing applications, particularly in the early stages of design. The way that ubiquitous computing technologies can facilitate the in situ collection of self-report data is also discussed. Although the focus is on ubiquitous computing applications and tools for their assessment, it is believed that the in situ evaluation tools that are proposed will be generally useful for field trials of other technology, applications, or formative studies that are concerned with collecting data in situ. %B International Journal of Human-Computer Interaction %V 22 %P 103 - 118 %8 2007/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/10447310709336957 %N 1-2 %R 10.1080/10447310709336957 %0 Conference Paper %B Proceedings of the 2007 ACM workshop on Storage security and survivability %D 2007 %T Confidentiality-preserving rank-ordered search %A Swaminathan,Ashwin %A Mao,Yinian %A Su,Guan-Ming %A Gou,Hongmei %A Varna,Avinash L. %A He,Shan %A M. Wu %A Oard, Douglas %K encrypted domain search %K Ranked retrieval %K secure index %X This paper introduces a new framework for confidentiality preserving rank-ordered search and retrieval over large document collections. The proposed framework not only protects document/query confidentiality against an outside intruder, but also prevents an untrusted data center from learning information about the query and the document collection. We present practical techniques for proper integration of relevance scoring methods and cryptographic techniques, such as order preserving encryption, to protect data collections and indices and provide efficient and accurate search capabilities to securely rank-order documents in response to a query. Experimental results on the W3C collection show that these techniques have comparable performance to conventional search systems designed for non-encrypted data in terms of search accuracy. The proposed methods thus form the first steps to bring together advanced information retrieval and secure search capabilities for a wide range of applications including managing data in government and business operations, enabling scholarly study of sensitive data, and facilitating the document discovery process in litigation. %B Proceedings of the 2007 ACM workshop on Storage security and survivability %S StorageSS '07 %I ACM %C New York, NY, USA %P 7 - 12 %8 2007/// %@ 978-1-59593-891-6 %G eng %U http://doi.acm.org/10.1145/1314313.1314316 %R 10.1145/1314313.1314316 %0 Conference Paper %B IEEE International Conference on Services Computing, 2007. SCC 2007 %D 2007 %T Configuration Reasoning and Ontology For Web %A Dana Dachman-Soled %A Sreedhar, V.C. %K Apache server %K Application software %K configuration management %K configuration reasoning %K configuration space %K File servers %K Information security %K Internet %K knowledge representation languages %K logical framework %K logical reasoning %K ontologies %K ontologies (artificial intelligence) %K Orbital robotics %K OWL %K path planning %K Robot kinematics %K Runtime environment %K Taxonomy %K Web infrastructures management %K Web ontology language %K Web server %X Configuration plays a central role in the deployment and management of Web infrastructures and applications. A configuration often consists of assigning "values" to a pre-defined set of parameters defined in one or more files. Although the task of assigning values to (configuration) parameters looks simple, configuring infrastructures and applications is a very complex process. In this paper we present a framework for defining and analyzing configuration of an Apache server. We define the notion of "configuration space" of an Apache server as a set of possible values that can be assigned to configuration parameters. We then define the notion of an "obstacle" and "forbidden region" in the configuration space that should be avoided. We model configuration space using a logical framework based on OWL (Web ontology language). The obstacles and forbidden regions in the configuration space are modeled as constraints in the logical framework. These obstacles and forbidden regions are essentially "anti-patterns" that a typical installation should avoid. Given an instance of a configuration (that is, a "point" in the configuration space) we then check if the instance is "obstacle free" using logical reasoning. %B IEEE International Conference on Services Computing, 2007. SCC 2007 %P 387 - 394 %8 2007/07// %G eng %0 Journal Article %J IEEE Intelligent Systems %D 2007 %T CONVEX: Context vectors as a paradigm for learning group behaviors based on similarity %A Martinez,V. %A Simari,G. I %A Sliva,A. %A V.S. Subrahmanian %B IEEE Intelligent Systems %V 23 %P 51 - 57 %8 2007/// %G eng %N 4 %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2007. IPDPS 2007. IEEE International %D 2007 %T Creating a Robust Desktop Grid using Peer-to-Peer Services %A Kim,J.-S. %A Nam,B. %A Marsh,M. %A Keleher,P. %A Bhattacharjee, Bobby %A Richardson,D. %A Wellnitz,D. %A Sussman, Alan %K allocation; %K architecture;load %K balancing;peer-to-peer %K client-server %K computing;peer-to-peer %K computing;resource %K desktop %K grid;scalable %K infrastructure;client-server %K system;robust %K systems;grid %X The goal of the work described in this paper is to design and build a scalable infrastructure for executing grid applications on a widely distributed set of resources. Such grid infrastructure must be decentralized, robust, highly available, and scalable, while efficiently mapping application instances to available resources in the system. However, current desktop grid computing platforms are typically based on a client-server architecture, which has inherent shortcomings with respect to robustness, reliability and scalability. Fortunately, these problems can be addressed through the capabilities promised by new techniques and approaches in peer-to-peer (P2P) systems. By employing P2P services, our system allows users to submit jobs to be run in the system and to run jobs submitted by other users on any resources available in the system, essentially allowing a group of users to form an ad-hoc set of shared resources. The initial target application areas for the desktop grid system are in astronomy and space science simulation and data analysis. %B Parallel and Distributed Processing Symposium, 2007. IPDPS 2007. IEEE International %P 1 - 7 %8 2007/03// %G eng %R 10.1109/IPDPS.2007.370505 %0 Journal Article %J Commun. ACM %D 2007 %T Creativity support tools: accelerating discovery and innovation %A Shneiderman, Ben %X How can designers of programming interfaces, interactive tools, and rich social environments enable more people to be more creative more often? %B Commun. ACM %V 50 %P 20 - 32 %8 2007/12// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/1323688.1323689 %N 12 %R 10.1145/1323688.1323689 %0 Conference Paper %B Proceedings of the 8th ACM international symposium on Mobile ad hoc networking and computing %D 2007 %T Cross-layer latency minimization in wireless networks with SINR constraints %A Chafekar,Deepti %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %K cross-layer design %K end-to-end scheduling %K Interference %K SINR model %K Wireless networks %X Recently, there has been substantial interest in the design of cross-layer protocols for wireless networks. These protocols optimize certain performance metric(s) of interest (e.g. latency, energy, rate) by jointly optimizing the performance of multiple layers of the protocol stack. Algorithm designers often use geometric-graph-theoretic models for radio interference to design such cross-layer protocols. In this paper we study the problem of designing cross-layer protocols for multi-hop wireless networks using a more realistic Signal to Interference plus Noise Ratio (SINR) model for radio interference. The following cross-layer latency minimization problem is studied: Given a set V of transceivers, and a set of source-destination pairs, (i) choose power levels for all the transceivers, (ii) choose routes for all connections, and (iii) construct an end-to-end schedule such that the SINR constraints are satisfied at each time step so as to minimize the make-span of the schedule (the time by which all packets have reached their respective destinations). We present a polynomial-time algorithm with provable worst-case performance guarantee for this cross-layer latency minimization problem. As corollaries of the algorithmic technique we show that a number of variants of the cross-layer latency minimization problem can also be approximated efficiently in polynomial time. Our work extends the results of Kumar et al. (Proc. SODA, 2004) and Moscibroda et al. (Proc. MOBIHOC, 2006). Although our algorithm considers multiple layers of the protocol stack, it can naturally be viewed as compositions of tasks specific to each layer --- this allows us to improve the overall performance while preserving the modularity of the layered structure. %B Proceedings of the 8th ACM international symposium on Mobile ad hoc networking and computing %S MobiHoc '07 %I ACM %C New York, NY, USA %P 110 - 119 %8 2007/// %@ 978-1-59593-684-4 %G eng %U http://doi.acm.org/10.1145/1288107.1288123 %R 10.1145/1288107.1288123 %0 Journal Article %J Science %D 2007 %T Cultural modeling in real time %A V.S. Subrahmanian %X Computer models are being developed that can help policy-makers predict the behavior of political, economic, and social groups. %B Science %V 317 %P 1509 - 1509 %8 2007/// %G eng %N 5844 %R 10.1126/science.1148306 %0 Journal Article %J IEEE Pervasive Computing %D 2007 %T Data Management in the Worldwide Sensor Web %A Balazinska,Magdalena %A Deshpande, Amol %A Franklin,Michael J. %A Gibbons,Phillip B. %A Gray,Jim %A Hansen,Mark %A Liebhold,Michael %A Nath,Suman %A Szalay,Alexander %A Tao,Vincent %K data modeling %K data streams %K data uncertainty %K distributed systems %K interoperability %K sensor networks %X Advances in hardware and miniaturization technologies have led to a rapid increase in the number of large-scale sensor network deployments around the world, bringing us closer to the vision of a worldwide sensor web. Exploiting the sensor web to its full potential, however, raises several hard data management challenges. Addressing these challenges requires a concerted multidisciplinary effort encompassing many computer science fields. This article outlines some of the key data management challenges of a worldwide sensor web. It also presents recent advances in data management research that aim to address these challenges. Hopefully, the scientific, ubiquitous computing, and sensor network communities will adopt some of these solutions for managing their data. The article also attempts to bring many of the remaining open issues to the attention of the database community. This article is part of a special issue on Building a Sensor-Rich World. %B IEEE Pervasive Computing %V 6 %P 30 - 40 %8 2007/// %@ 1536-1268 %G eng %N 2 %0 Journal Article %J EURASIP Journal on Embedded Systems %D 2007 %T Dataflow-based mapping of computer vision algorithms onto FPGAs %A Sen,M. %A Corretjer,I. %A Haim,F. %A Saha,S. %A Schlessman,J. %A Lv,T. %A Bhattacharyya, Shuvra S. %A Wolf,W. %B EURASIP Journal on Embedded Systems %V 2007 %P 29 - 29 %8 2007/// %G eng %N 1 %0 Conference Paper %B Proceedings of the 16th international conference on World Wide Web %D 2007 %T Defeating script injection attacks with browser-enforced embedded policies %A Jim,Trevor %A Swamy,Nikhil %A Hicks, Michael W. %K cross-site scripting %K script injection %K web application security %X Web sites that accept and display content such as wiki articles or comments typically filter the content to prevent injected script code from running in browsers that view the site. The diversity of browser rendering algorithms and the desire to allow rich content make filtering quite difficult, however, and attacks such as the Samy and Yamanner worms have exploited filtering weaknesses. This paper proposes a simple alternative mechanism for preventing script injection called Browser-Enforced Embedded Policies (BEEP). The idea is that a web site can embed a policy in its pages that specifies which scripts are allowed to run. The browser, which knows exactly when it will run a script, can enforce this policy perfectly. We have added BEEP support to several browsers, and built tools to simplify adding policies to web applications. We found that supporting BEEP in browsers requires only small and localized modifications, modifying web applications requires minimal effort, and enforcing policies is generally lightweight. %B Proceedings of the 16th international conference on World Wide Web %S WWW '07 %I ACM %C New York, NY, USA %P 601 - 610 %8 2007/// %@ 978-1-59593-654-7 %G eng %U http://doi.acm.org/10.1145/1242572.1242654 %R 10.1145/1242572.1242654 %0 Journal Article %J Information VisualizationInformation Visualization %D 2007 %T Designing Semantic Substrates for Visual Network Exploration %A Aris,Aleks %A Shneiderman, Ben %K data exploration and analysis %K Graphical user interfaces %K Information Visualization %K network visualization %K semantic substrate design %X A semantic substrate is a spatial template for a network, where nodes are grouped into regions and laid out within each region according to one or more node attributes. This paper shows how users can be given control in designing their own substrates and how this ability leads to a different approach to network data exploration. Users can create a semantic substrate, enter their data, get feedback from domain experts, edit the semantic substrate, and iteratively continue this procedure until the domain experts are satisfied with the insights they have gained. We illustrate this process in two case studies with domain experts working with legal precedents and food webs. Guidelines for designing substrates are provided, including how to locate, size, and align regions in a substrate, which attributes to choose for grouping nodes into regions, how to select placement methods and which attributes to set as parameters of the selected placement method. Throughout the paper, examples are illustrated with NVSS 2.0, the network visualization tool developed to explore the semantic substrate idea. %B Information VisualizationInformation Visualization %V 6 %P 281 - 300 %8 2007/12/21/ %@ 1473-8716, 1473-8724 %G eng %U http://ivi.sagepub.com/content/6/4/281 %N 4 %R 10.1057/palgrave.ivs.9500162 %0 Conference Paper %B International Joint Conference on Artificial Intelligence %D 2007 %T Detecting stochastically scheduled activities in video %A Albanese, M. %A Moscato, V. %A Picariello, A. %A V.S. Subrahmanian %A Udrea,O. %X The ability to automatically detect activities invideo is of increasing importance in applications such as bank security, airport tarmac security, bag- gage area security and building site surveillance. We present a stochastic activity model composed of atomic actions which are directly observable through image understanding primitives. We focus on answering two types of questions: (i) what are the minimal sub-videos in which a given action is identified with probability above a certain thresh- old and (ii) for a given video, can we decide which activity from a given set most likely occurred? We provide the MPS algorithm for the first problem, as well as two different algorithms (naiveMPA and MPA) to solve the second. Our experimental re- sults on a dataset consisting of staged bank robbery videos (described in [Vu et al., 2003]) show that our algorithms are both fast and provide high qual- ity results when compared to human reviewers. %B International Joint Conference on Artificial Intelligence %P 1802 - 1807 %8 2007/// %G eng %0 Conference Paper %B Proceedings of the twenty-second IEEE/ACM international conference on Automated software engineering %D 2007 %T Direct-dependency-based software compatibility testing %A Yoon,Il-Chul %A Sussman, Alan %A Memon, Atif M. %A Porter, Adam %K compatibility testing %K component-based software system %X Software compatibility testing is an important quality assurance task aimed at ensuring that component-based software systems build and/or execute properly across a broad range of user system configurations. Because each configuration can involve multiple components with different versions, and because there are complex and changing interdependencies between components and their versions, it is generally infeasible to test all potential configurations. Therefore, compatibility testing usually means examining only a handful of default or popular configurations to detect problems, and as a result costly errors can and do escape to the field This paper presents an improved approach to compatibility testing called RACHET. We formally model the configuration space for component-based systems and use the model to generate test plans covering user-specified portion of the space - the example in this paper is covering all it direct dependencies between components. The test plan is executed efficiently in parallel, by distributing work so as to best utilize test resources. We conducted experimentsand simulation studies applying our approach to a large-scale data management middleware system. The results showed that for this system RACHET discovered incompatibilities between components at a small fraction of the cost for exhaustive testing without compromising test quality %B Proceedings of the twenty-second IEEE/ACM international conference on Automated software engineering %S ASE '07 %I ACM %C New York, NY, USA %P 409 - 412 %8 2007/// %@ 978-1-59593-882-4 %G eng %U http://doi.acm.org/10.1145/1321631.1321696 %R 10.1145/1321631.1321696 %0 Conference Paper %B Proceedings of the sixteenth ACM conference on Conference on information and knowledge management %D 2007 %T Discovering interesting usage patterns in text collections: integrating text mining with visualization %A Don,Anthony %A Zheleva,Elena %A Gregory,Machon %A Tarkan,Sureyya %A Auvil,Loretta %A Clement,Tanya %A Shneiderman, Ben %A Plaisant, Catherine %K digital humanities %K frequent closed itemsets %K n-grams %K text mining %K user interface %X This paper addresses the problem of making text mining results more comprehensible to humanities scholars, journalists, intelligence analysts, and other researchers, in order to support the analysis of text collections. Our system, FeatureLens1, visualizes a text collection at several levels of granularity and enables users to explore interesting text patterns. The current implementation focuses on frequent itemsets of n-grams, as they capture the repetition of exact or similar expressions in the collection. Users can find meaningful co-occurrences of text patterns by visualizing them within and across documents in the collection. This also permits users to identify the temporal evolution of usage such as increasing, decreasing or sudden appearance of text patterns. The interface could be used to explore other text features as well. Initial studies suggest that FeatureLens helped a literary scholar and 8 users generate new hypotheses and interesting insights using 2 text collections. %B Proceedings of the sixteenth ACM conference on Conference on information and knowledge management %S CIKM '07 %I ACM %C New York, NY, USA %P 213 - 222 %8 2007/// %@ 978-1-59593-803-9 %G eng %U http://doi.acm.org/10.1145/1321440.1321473 %R 10.1145/1321440.1321473 %0 Journal Article %J High Performance Computing–HiPC 2007 %D 2007 %T Distributed ranked search %A Gopalakrishnan,V. %A Morselli,R. %A Bhattacharjee, Bobby %A Keleher,P. %A Srinivasan, Aravind %X P2P deployments are a natural infrastructure for building distributed search networks. Proposed systems support locating and retrieving all results, but lack the information necessary to rank them. Users, however, are primarily interested in the most relevant results, not necessarily all possible results.Using random sampling, we extend a class of well-known information retrieval ranking algorithms such that they can be applied in this decentralized setting. We analyze the overhead of our approach, and quantify how our system scales with increasing number of documents, system size, document to node mapping (uniform versus non-uniform), and types of queries (rare versus popular terms). Our analysis and simulations show that a) these extensions are efficient, and scale with little overhead to large systems, and b) the accuracy of the results obtained using distributed ranking is comparable to that of a centralized implementation. %B High Performance Computing–HiPC 2007 %P 7 - 20 %8 2007/// %G eng %0 Book Section %B High Performance Computing – HiPC 2007High Performance Computing – HiPC 2007 %D 2007 %T Distributed Ranked Search %A Gopalakrishnan,Vijay %A Morselli,Ruggero %A Bhattacharjee, Bobby %A Keleher,Pete %A Srinivasan, Aravind %E Aluru,Srinivas %E Parashar,Manish %E Badrinath,Ramamurthy %E Prasanna,Viktor %X P2P deployments are a natural infrastructure for building distributed search networks. Proposed systems support locating and retrieving all results, but lack the information necessary to rank them. Users, however, are primarily interested in the most relevant results, not necessarily all possible results. Using random sampling, we extend a class of well-known information retrieval ranking algorithms such that they can be applied in this decentralized setting. We analyze the overhead of our approach, and quantify how our system scales with increasing number of documents, system size, document to node mapping (uniform versus non-uniform), and types of queries (rare versus popular terms). Our analysis and simulations show that a) these extensions are efficient, and scale with little overhead to large systems, and b) the accuracy of the results obtained using distributed ranking is comparable to that of a centralized implementation. %B High Performance Computing – HiPC 2007High Performance Computing – HiPC 2007 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4873 %P 7 - 20 %8 2007/// %@ 978-3-540-77219-4 %G eng %U http://dx.doi.org/10.1007/978-3-540-77220-0_6 %0 Book Section %B Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques %D 2007 %T Distribution-Free Testing Lower Bounds for Basic Boolean Functions %A Dana Dachman-Soled %A Servedio, Rocco A. %E Charikar, Moses %E Jansen, Klaus %E Reingold, Omer %E Rolim, José D. P. %K Algorithm Analysis and Problem Complexity %K Discrete Mathematics in Computer Science %K Numeric Computing %X In the distribution-free property testing model, the distance between functions is measured with respect to an arbitrary and unknown probability distribution \mathcal{D} over the input domain. We consider distribution-free testing of several basic Boolean function classes over {0,1} n , namely monotone conjunctions, general conjunctions, decision lists, and linear threshold functions. We prove that for each of these function classes, Ω((n/logn)1/5) oracle calls are required for any distribution-free testing algorithm. Since each of these function classes is known to be distribution-free properly learnable (and hence testable) using Θ(n) oracle calls, our lower bounds are within a polynomial factor of the best possible. %B Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 494 - 508 %8 2007/01/01/ %@ 978-3-540-74207-4, 978-3-540-74208-1 %G eng %U http://link.springer.com/chapter/10.1007/978-3-540-74208-1_36 %0 Conference Paper %B International Conference on Image Analysis and Processing (ICIAP'07) %D 2007 %T Document Image Mosaicing with Mobile Phones %A Hannuksela,J. %A Sangi,P. %A Heikkila,J. %A Liu,X. %A David Doermann %X This paper presents a novel user interaction oncept for document image scanning with mobile phones. A high resolution mosaic image is constructed in two main stages. Firstly, online camera motion estimation is applied on the phone to assist the user to capture small image batches of the document page. Automatic image stitching process with a help of estimated device motion is carried out to reconstruct the full view of the document. Experiments on document images captured and processed with the mosaicing software clearly show the feasibility of the approach. %B International Conference on Image Analysis and Processing (ICIAP'07) %P 1 - 8 %8 2007/09// %G eng %0 Journal Article %J Science %D 2007 %T Draft Genome of the Filarial Nematode Parasite Brugia Malayi %A Ghedin,Elodie %A Wang,Shiliang %A Spiro,David %A Caler,Elisabet %A Zhao,Qi %A Crabtree,Jonathan %A Allen,Jonathan E %A Delcher,Arthur L. %A Guiliano,David B %A Miranda-Saavedra,Diego %A Angiuoli,Samuel V %A Creasy,Todd %A Amedeo,Paolo %A Haas,Brian %A El‐Sayed, Najib M. %A Wortman,Jennifer R. %A Feldblyum,Tamara %A Tallon,Luke %A Schatz,Michael %A Shumway,Martin %A Koo,Hean %A Salzberg,Steven L. %A Schobel,Seth %A Pertea,Mihaela %A Pop, Mihai %A White,Owen %A Barton,Geoffrey J %A Carlow,Clotilde K. S %A Crawford,Michael J %A Daub,Jennifer %A Dimmic,Matthew W %A Estes,Chris F %A Foster,Jeremy M %A Ganatra,Mehul %A Gregory,William F %A Johnson,Nicholas M %A Jin,Jinming %A Komuniecki,Richard %A Korf,Ian %A Kumar,Sanjay %A Laney,Sandra %A Li,Ben-Wen %A Li,Wen %A Lindblom,Tim H %A Lustigman,Sara %A Ma,Dong %A Maina,Claude V %A Martin,David M. A %A McCarter,James P %A McReynolds,Larry %A Mitreva,Makedonka %A Nutman,Thomas B %A Parkinson,John %A Peregrín-Alvarez,José M %A Poole,Catherine %A Ren,Qinghu %A Saunders,Lori %A Sluder,Ann E %A Smith,Katherine %A Stanke,Mario %A Unnasch,Thomas R %A Ware,Jenna %A Wei,Aguan D %A Weil,Gary %A Williams,Deryck J %A Zhang,Yinhua %A Williams,Steven A %A Fraser-Liggett,Claire %A Slatko,Barton %A Blaxter,Mark L %A Scott,Alan L %X Parasitic nematodes that cause elephantiasis and river blindness threaten hundreds of millions of people in the developing world. We have sequenced the ∼90 megabase (Mb) genome of the human filarial parasite Brugia malayi and predict ∼11,500 protein coding genes in 71 Mb of robustly assembled sequence. Comparative analysis with the free-living, model nematode Caenorhabditis elegans revealed that, despite these genes having maintained little conservation of local synteny during ∼350 million years of evolution, they largely remain in linkage on chromosomal units. More than 100 conserved operons were identified. Analysis of the predicted proteome provides evidence for adaptations of B. malayi to niches in its human and vector hosts and insights into the molecular basis of a mutualistic relationship with its Wolbachia endosymbiont. These findings offer a foundation for rational drug design. %B Science %V 317 %P 1756 - 1760 %8 2007/09/21/ %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/317/5845/1756 %N 5845 %R 10.1126/science.1145406 %0 Conference Paper %B Proceedings of the sixteenth ACM conference on Conference on information and knowledge management %D 2007 %T A dual-view approach to interactive network visualization %A Namata,Galileo Mark %A Staats,Brian %A Getoor, Lise %A Shneiderman, Ben %K coordinated views %K Information Visualization %K interactive graph visualization %K social networks %X Visualizing network data, from tree structures to arbitrarily connected graphs, is a difficult problem in information visualization. A large part of the problem is that in network data, users not only have to visualize the attributes specific to each data item, but also the links specifying how those items are connected to each other. Past approaches to resolving these difficulties focus on zooming, clustering, filtering and applying various methods of laying out nodes and edges. Such approaches, however, focus only on optimizing a network visualization in a single view, limiting the amount of information that can be shown and explored in parallel. Moreover, past approaches do not allow users to cross reference different subsets or aspects of large, complex networks. In this paper, we propose an approach to these limitations using multiple coordinated views of a given network. To illustrate our approach, we implement a tool called DualNet and evaluate the tool with a case study using an email communication network. We show how using multiple coordinated views improves navigation and provides insight into large networks with multiple node and link properties and types. %B Proceedings of the sixteenth ACM conference on Conference on information and knowledge management %S CIKM '07 %I ACM %C New York, NY, USA %P 939 - 942 %8 2007/// %@ 978-1-59593-803-9 %G eng %U http://doi.acm.org/10.1145/1321440.1321580 %R 10.1145/1321440.1321580 %0 Journal Article %J IEEE Journal on Selected Areas in Communications %D 2007 %T Efficient lookup on unstructured topologies %A Morselli,R. %A Bhattacharjee, Bobby %A Marsh,M.A. %A Srinivasan, Aravind %K Computer science %K DHT %K distributed algorithms %K Distributed computing %K distributed hash table %K Least squares approximation %K LMS %K local minima search %K lookup protocol %K Network topology %K node failures %K Peer to peer computing %K Performance analysis %K Protocols %K replication strategy %K Resilience %K Robustness %K table lookup %K telecommunication network topology %K unstructured network topology %X We present LMS, a protocol for efficient lookup on unstructured networks. Our protocol uses a virtual namespace without imposing specific topologies. It is more efficient than existing lookup protocols for unstructured networks, and thus is an attractive alternative for applications in which the topology cannot be structured as a Distributed Hash Table (DHT). We present analytic bounds for the worst-case performance of LMS. Through detailed simulations (with up to 100,000 nodes), we show that the actual performance on realistic topologies is significantly better. We also show in both simulations and a complete implementation (which includes over five hundred nodes) that our protocol is inherently robust against multiple node failures and can adapt its replication strategy to optimize searches according to a specific heuristic. Moreover, the simulation demonstrates the resilience of LMS to high node turnover rates, and that it can easily adapt to orders of magnitude changes in network size. The overhead incurred by LMS is small, and its performance approaches that of DHTs on networks of similar size %B IEEE Journal on Selected Areas in Communications %V 25 %P 62 - 72 %8 2007/01// %@ 0733-8716 %G eng %N 1 %R 10.1109/JSAC.2007.07007 %0 Journal Article %J Red %D 2007 %T Efficient parallel memory organization for turbo decoders %A Salmela,P. %A Gu,R. %A Bhattacharyya, Shuvra S. %A Takala,J. %B Red %V 5 %8 2007/// %G eng %0 Journal Article %J American Journal of Preventive Medicine %D 2007 %T eHealth Research from the User’s Perspective %A Hesse,Bradford W. %A Shneiderman, Ben %X The application of information technology (IT) to issues of healthcare delivery has had a long and tortuous history in the United States. Within the field of eHealth, vanguard applications of advanced computing techniques, such as applications in artificial intelligence or expert systems, have languished in spite of a track record of scholarly publication and decisional accuracy. The problem is one of purpose, of asking the right questions for the science to solve. Historically, many computer science pioneers have been tempted to ask “what can the computer do?” New advances in eHealth are prompting developers to ask “what can people do?” How can eHealth take part in national goals for healthcare reform to empower relationships between healthcare professionals and patients, healthcare teams and families, and hospitals and communities to improve health equitably throughout the population? To do this, eHealth researchers must combine best evidence from the user sciences (human factors engineering, human–computer interaction, psychology, and usability) with best evidence in medicine to create transformational improvements in the quality of care that medicine offers. These improvements should follow recommendations from the Institute of Medicine to create a healthcare system that is (1) safe, (2) effective (evidence based), (3) patient centered, and (4) timely. Relying on the eHealth researcher’s intuitive grasp of systems issues, improvements should be made with considerations of users and beneficiaries at the individual (patient–physician), group (family–staff), community, and broad environmental levels. %B American Journal of Preventive Medicine %V 32 %P S97-S103 - S97-S103 %8 2007/05// %@ 0749-3797 %G eng %U http://www.sciencedirect.com/science/article/pii/S0749379707000487 %N 5, Supplement %R 10.1016/j.amepre.2007.01.019 %0 Journal Article %J Journal of Vacuum Science & Technology B %D 2007 %T Electron beam and optical proximity effect reduction for nanolithography: New results %A Peckerar,Martin %A Sander,David %A Srivastava,Ankur %A Foli,Adakou %A Vishkin, Uzi %K electron beam lithography %K Linear programming %K masks %K nanolithography %K proximity effect (lithography) %X Proximity effect correction by dose modulation is widely practiced in electron-beam lithography. Optical proximity control is also possible using a combination of shape adjustment and phase control. Assigning “the right” dose (or fill factor and phase for optics) is a well known mathematical inverse problem. Linear programming, by definition, is the appropriate method for determining dose. In the past, the technique was too slow for full-scale implementation in mask making. Here, the authors discuss how recent developments in computer speed and architecture have improved the prospects for full-scale implementation. In addition, the authors discuss some numerical techniques, analogous to gridding and relaxation, that make linear programming more attractive in mask making. %B Journal of Vacuum Science & Technology B %V 25 %P 2288 - 2294 %8 2007/// %G eng %U http://link.aip.org/link/?JVB/25/2288/1 %N 6 %R 10.1116/1.2806967 %0 Conference Paper %B Proceedings of RTSS %D 2007 %T An energy-driven design methodology for distributing DSP applications across wireless sensor networks %A Shen,C. C %A Plishker,W. %A Bhattacharyya, Shuvra S. %A Goldsman,N. %B Proceedings of RTSS %P 214 - 226 %8 2007/// %G eng %0 Report %D 2007 %T An Environment of Conducting Families of Software Engineering Experiments %A Hochstein, Lorin %A Nakamura,Taiga %A Shull, Forrest %A Zazworka, Nico %A Voelp,Martin %A Zelkowitz, Marvin V %A Basili, Victor R. %K collecting data %K managing data %K parallel programming languages %K sanitizing data %K software engineering experiments %K universities %X The classroom is a valuable resource for conducting software engineering experiments. However, coordinating a family of experiments in classroom environments presents a number of challenges to researchers. This paper describes an environment that simplifies the process of collecting, managing and sanitizing data from classroom experiments, while minimizing disruption to natural subject behavior. We have successfully used this environment to study the impact of parallel programming languages on programmer productivity at multiple universities across the United States. %I University of Maryland, College Park %8 2007/05// %G eng %U http://drum.lib.umd.edu/handle/1903/7545 %0 Journal Article %J Journal of the Optical Society of America AJ. Opt. Soc. Am. A %D 2007 %T Estimation of contour motion and deformation for nonrigid object tracking %A Jie Shao %A Porikli,Fatih %A Chellapa, Rama %K Motion detection %K VISION %X We present an algorithm for nonrigid contour tracking in heavily cluttered background scenes. Based on the properties of nonrigid contour movements, a sequential framework for estimating contour motion and deformation is proposed. We solve the nonrigid contour tracking problem by decomposing it into three subproblems: motion estimation, deformation estimation, and shape regulation. First, we employ a particle filter to estimate the global motion parameters of the affine transform between successive frames. Then we generate a probabilistic deformation map to deform the contour. To improve robustness, multiple cues are used for deformation probability estimation. Finally, we use a shape prior model to constrain the deformed contour. This enables us to retrieve the occluded parts of the contours and accurately track them while allowing shape changes specific to the given object types. Our experiments show that the proposed algorithm significantly improves the tracker performance. %B Journal of the Optical Society of America AJ. Opt. Soc. Am. A %V 24 %P 2109 - 2121 %8 2007/// %G eng %U http://josaa.osa.org/abstract.cfm?URI=josaa-24-8-2109 %N 8 %R 10.1364/JOSAA.24.002109 %0 Journal Article %J Nature %D 2007 %T Evolution of genes and genomes on the Drosophila phylogeny %A Clark,Andrew G. %A Eisen,Michael B. %A Smith,Douglas R. %A Bergman,Casey M. %A Oliver,Brian %A Markow,Therese A. %A Kaufman,Thomas C. %A Kellis,Manolis %A Gelbart,William %A Iyer,Venky N. %A Pollard,Daniel A. %A Sackton,Timothy B. %A Larracuente,Amanda M. %A Singh,Nadia D. %A Abad,Jose P. %A Abt,Dawn N. %A Adryan,Boris %A Aguade,Montserrat %A Akashi,Hiroshi %A Anderson,Wyatt W. %A Aquadro,Charles F. %A Ardell,David H. %A Arguello,Roman %A Artieri,Carlo G. %A Barbash,Daniel A. %A Barker,Daniel %A Barsanti,Paolo %A Batterham,Phil %A Batzoglou,Serafim %A Begun,Dave %A Bhutkar,Arjun %A Blanco,Enrico %A Bosak,Stephanie A. %A Bradley,Robert K. %A Brand,Adrianne D. %A Brent,Michael R. %A Brooks,Angela N. %A Brown,Randall H. %A Butlin,Roger K. %A Caggese,Corrado %A Calvi,Brian R. %A Carvalho,A. Bernardo de %A Caspi,Anat %A Castrezana,Sergio %A Celniker,Susan E. %A Chang,Jean L. %A Chapple,Charles %A Chatterji,Sourav %A Chinwalla,Asif %A Civetta,Alberto %A Clifton,Sandra W. %A Comeron,Josep M. %A Costello,James C. %A Coyne,Jerry A. %A Daub,Jennifer %A David,Robert G. %A Delcher,Arthur L. %A Delehaunty,Kim %A Do,Chuong B. %A Ebling,Heather %A Edwards,Kevin %A Eickbush,Thomas %A Evans,Jay D. %A Filipski,Alan %A Findei|[szlig]|,Sven %A Freyhult,Eva %A Fulton,Lucinda %A Fulton,Robert %A Garcia,Ana C. L. %A Gardiner,Anastasia %A Garfield,David A. %A Garvin,Barry E. %A Gibson,Greg %A Gilbert,Don %A Gnerre,Sante %A Godfrey,Jennifer %A Good,Robert %A Gotea,Valer %A Gravely,Brenton %A Greenberg,Anthony J. %A Griffiths-Jones,Sam %A Gross,Samuel %A Guigo,Roderic %A Gustafson,Erik A. %A Haerty,Wilfried %A Hahn,Matthew W. %A Halligan,Daniel L. %A Halpern,Aaron L. %A Halter,Gillian M. %A Han,Mira V. %A Heger,Andreas %A Hillier,LaDeana %A Hinrichs,Angie S. %A Holmes,Ian %A Hoskins,Roger A. %A Hubisz,Melissa J. %A Hultmark,Dan %A Huntley,Melanie A. %A Jaffe,David B. %A Jagadeeshan,Santosh %A Jeck,William R. %A Johnson,Justin %A Jones,Corbin D. %A Jordan,William C. %A Karpen,Gary H. %A Kataoka,Eiko %A Keightley,Peter D. %A Kheradpour,Pouya %A Kirkness,Ewen F. %A Koerich,Leonardo B. %A Kristiansen,Karsten %A Kudrna,Dave %A Kulathinal,Rob J. %A Kumar,Sudhir %A Kwok,Roberta %A Lander,Eric %A Langley,Charles H. %A Lapoint,Richard %A Lazzaro,Brian P. %A Lee,So-Jeong %A Levesque,Lisa %A Li,Ruiqiang %A Lin,Chiao-Feng %A Lin,Michael F. %A Lindblad-Toh,Kerstin %A Llopart,Ana %A Long,Manyuan %A Low,Lloyd %A Lozovsky,Elena %A Lu,Jian %A Luo,Meizhong %A Machado,Carlos A. %A Makalowski,Wojciech %A Marzo,Mar %A Matsuda,Muneo %A Matzkin,Luciano %A McAllister,Bryant %A McBride,Carolyn S. %A McKernan,Brendan %A McKernan,Kevin %A Mendez-Lago,Maria %A Minx,Patrick %A Mollenhauer,Michael U. %A Montooth,Kristi %A Mount, Stephen M. %A Mu,Xu %A Myers,Eugene %A Negre,Barbara %A Newfeld,Stuart %A Nielsen,Rasmus %A Noor,Mohamed A. F. %A O'Grady,Patrick %A Pachter,Lior %A Papaceit,Montserrat %A Parisi,Matthew J. %A Parisi,Michael %A Parts,Leopold %A Pedersen,Jakob S. %A Pesole,Graziano %A Phillippy,Adam M %A Ponting,Chris P. %A Pop, Mihai %A Porcelli,Damiano %A Powell,Jeffrey R. %A Prohaska,Sonja %A Pruitt,Kim %A Puig,Marta %A Quesneville,Hadi %A Ram,Kristipati Ravi %A Rand,David %A Rasmussen,Matthew D. %A Reed,Laura K. %A Reenan,Robert %A Reily,Amy %A Remington,Karin A. %A Rieger,Tania T. %A Ritchie,Michael G. %A Robin,Charles %A Rogers,Yu-Hui %A Rohde,Claudia %A Rozas,Julio %A Rubenfield,Marc J. %A Ruiz,Alfredo %A Russo,Susan %A Salzberg,Steven L. %A Sanchez-Gracia,Alejandro %A Saranga,David J. %A Sato,Hajime %A Schaeffer,Stephen W. %A Schatz,Michael C %A Schlenke,Todd %A Schwartz,Russell %A Segarra,Carmen %A Singh,Rama S. %A Sirot,Laura %A Sirota,Marina %A Sisneros,Nicholas B. %A Smith,Chris D. %A Smith,Temple F. %A Spieth,John %A Stage,Deborah E. %A Stark,Alexander %A Stephan,Wolfgang %A Strausberg,Robert L. %A Strempel,Sebastian %A Sturgill,David %A Sutton,Granger %A Sutton,Granger G. %A Tao,Wei %A Teichmann,Sarah %A Tobari,Yoshiko N. %A Tomimura,Yoshihiko %A Tsolas,Jason M. %A Valente,Vera L. S. %A Venter,Eli %A Venter,J. Craig %A Vicario,Saverio %A Vieira,Filipe G. %A Vilella,Albert J. %A Villasante,Alfredo %A Walenz,Brian %A Wang,Jun %A Wasserman,Marvin %A Watts,Thomas %A Wilson,Derek %A Wilson,Richard K. %A Wing,Rod A. %A Wolfner,Mariana F. %A Wong,Alex %A Wong,Gane Ka-Shu %A Wu,Chung-I %A Wu,Gabriel %A Yamamoto,Daisuke %A Yang,Hsiao-Pei %A Yang,Shiaw-Pyng %A Yorke,James A. %A Yoshida,Kiyohito %A Zdobnov,Evgeny %A Zhang,Peili %A Zhang,Yu %A Zimin,Aleksey V. %A Baldwin,Jennifer %A Abdouelleil,Amr %A Abdulkadir,Jamal %A Abebe,Adal %A Abera,Brikti %A Abreu,Justin %A Acer,St Christophe %A Aftuck,Lynne %A Alexander,Allen %A An,Peter %A Anderson,Erica %A Anderson,Scott %A Arachi,Harindra %A Azer,Marc %A Bachantsang,Pasang %A Barry,Andrew %A Bayul,Tashi %A Berlin,Aaron %A Bessette,Daniel %A Bloom,Toby %A Blye,Jason %A Boguslavskiy,Leonid %A Bonnet,Claude %A Boukhgalter,Boris %A Bourzgui,Imane %A Brown,Adam %A Cahill,Patrick %A Channer,Sheridon %A Cheshatsang,Yama %A Chuda,Lisa %A Citroen,Mieke %A Collymore,Alville %A Cooke,Patrick %A Costello,Maura %A D'Aco,Katie %A Daza,Riza %A Haan,Georgius De %A DeGray,Stuart %A DeMaso,Christina %A Dhargay,Norbu %A Dooley,Kimberly %A Dooley,Erin %A Doricent,Missole %A Dorje,Passang %A Dorjee,Kunsang %A Dupes,Alan %A Elong,Richard %A Falk,Jill %A Farina,Abderrahim %A Faro,Susan %A Ferguson,Diallo %A Fisher,Sheila %A Foley,Chelsea D. %A Franke,Alicia %A Friedrich,Dennis %A Gadbois,Loryn %A Gearin,Gary %A Gearin,Christina R. %A Giannoukos,Georgia %A Goode,Tina %A Graham,Joseph %A Grandbois,Edward %A Grewal,Sharleen %A Gyaltsen,Kunsang %A Hafez,Nabil %A Hagos,Birhane %A Hall,Jennifer %A Henson,Charlotte %A Hollinger,Andrew %A Honan,Tracey %A Huard,Monika D. %A Hughes,Leanne %A Hurhula,Brian %A Husby,M Erii %A Kamat,Asha %A Kanga,Ben %A Kashin,Seva %A Khazanovich,Dmitry %A Kisner,Peter %A Lance,Krista %A Lara,Marcia %A Lee,William %A Lennon,Niall %A Letendre,Frances %A LeVine,Rosie %A Lipovsky,Alex %A Liu,Xiaohong %A Liu,Jinlei %A Liu,Shangtao %A Lokyitsang,Tashi %A Lokyitsang,Yeshi %A Lubonja,Rakela %A Lui,Annie %A MacDonald,Pen %A Magnisalis,Vasilia %A Maru,Kebede %A Matthews,Charles %A McCusker,William %A McDonough,Susan %A Mehta,Teena %A Meldrim,James %A Meneus,Louis %A Mihai,Oana %A Mihalev,Atanas %A Mihova,Tanya %A Mittelman,Rachel %A Mlenga,Valentine %A Montmayeur,Anna %A Mulrain,Leonidas %A Navidi,Adam %A Naylor,Jerome %A Negash,Tamrat %A Nguyen,Thu %A Nguyen,Nga %A Nicol,Robert %A Norbu,Choe %A Norbu,Nyima %A Novod,Nathaniel %A O'Neill,Barry %A Osman,Sahal %A Markiewicz,Eva %A Oyono,Otero L. %A Patti,Christopher %A Phunkhang,Pema %A Pierre,Fritz %A Priest,Margaret %A Raghuraman,Sujaa %A Rege,Filip %A Reyes,Rebecca %A Rise,Cecil %A Rogov,Peter %A Ross,Keenan %A Ryan,Elizabeth %A Settipalli,Sampath %A Shea,Terry %A Sherpa,Ngawang %A Shi,Lu %A Shih,Diana %A Sparrow,Todd %A Spaulding,Jessica %A Stalker,John %A Stange-Thomann,Nicole %A Stavropoulos,Sharon %A Stone,Catherine %A Strader,Christopher %A Tesfaye,Senait %A Thomson,Talene %A Thoulutsang,Yama %A Thoulutsang,Dawa %A Topham,Kerri %A Topping,Ira %A Tsamla,Tsamla %A Vassiliev,Helen %A Vo,Andy %A Wangchuk,Tsering %A Wangdi,Tsering %A Weiand,Michael %A Wilkinson,Jane %A Wilson,Adam %A Yadav,Shailendra %A Young,Geneva %A Yu,Qing %A Zembek,Lisa %A Zhong,Danni %A Zimmer,Andrew %A Zwirko,Zac %A Jaffe,David B. %A Alvarez,Pablo %A Brockman,Will %A Butler,Jonathan %A Chin,CheeWhye %A Gnerre,Sante %A Grabherr,Manfred %A Kleber,Michael %A Mauceli,Evan %A MacCallum,Iain %X Comparative analysis of multiple genomes in a phylogenetic framework dramatically improves the precision and sensitivity of evolutionary inference, producing more robust results than single-genome analyses can provide. The genomes of 12 Drosophila species, ten of which are presented here for the first time (sechellia, simulans, yakuba, erecta, ananassae, persimilis, willistoni, mojavensis, virilis and grimshawi), illustrate how rates and patterns of sequence divergence across taxa can illuminate evolutionary processes on a genomic scale. These genome sequences augment the formidable genetic tools that have made Drosophila melanogaster a pre-eminent model for animal genetics, and will further catalyse fundamental research on mechanisms of development, cell biology, genetics, disease, neurobiology, behaviour, physiology and evolution. Despite remarkable similarities among these Drosophila species, we identified many putatively non-neutral changes in protein-coding genes, non-coding RNA genes, and cis-regulatory regions. These may prove to underlie differences in the ecology and behaviour of these diverse species. %B Nature %V 450 %P 203 - 218 %8 2007/11/08/ %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v450/n7167/full/nature06341.html %N 7167 %R 10.1038/nature06341 %0 Journal Article %J Nature %D 2007 %T Evolution of genes and genomes on the Drosophila phylogeny %A Clark, A.G. %A Eisen,M. B %A Smith,D. R %A Bergman, C.M. %A Oliver, B. %A Markow, T.A. %A Kaufman, T.C. %A Kellis, M. %A Gelbart, W. %A Iyer, V.N. %A others %B Nature %V 450 %P 203 - 218 %8 2007 %G eng %N 7167 %0 Conference Paper %B Proceedings of the 9th annual conference on Genetic and evolutionary computation %D 2007 %T Evolution of non-uniform cellular automata using a genetic algorithm: diversity and computation %A Sondahl,Forrest %A Rand, William %K diversity %K Genetic algorithms %K non-uniform cellular automata %K search spaces %X We used a genetic algorithm to evaluate the cost / benefit of diversity in evolving sets of rules for non-uniform cellular automata solving the density classification problem. %B Proceedings of the 9th annual conference on Genetic and evolutionary computation %S GECCO '07 %I ACM %C New York, NY, USA %P 1531 - 1531 %8 2007/// %@ 978-1-59593-697-4 %G eng %U http://doi.acm.org/10.1145/1276958.1277260 %R 10.1145/1276958.1277260 %0 Journal Article %J Information Processing Letters %D 2007 %T Execution time analysis of a top-down R-tree construction algorithm %A Alborzi,Houman %A Samet, Hanan %K Bulk loading %K Data structures %K Packing %K R-trees %K Spatial databases %X A detailed CPU execution-time analysis and implementation are given for a bulk loading algorithm to construct R-trees due to García et al. [Y.J. García, M.A. López, S.T. Leutenegger, A greedy algorithm for bulk loading R-trees, in: GIS'98: Proc. of the 6th ACM Intl. Symp. on Advances in Geographic Information Systems, Washington, DC, 1998, pp. 163–164] which is known as the top-down greedy split (TGS) bulk loading algorithm. The TGS algorithm makes use of a classical bottom-up packing approach. In addition, an alternative packing approach termed top-down packing is introduced which may lead to improved query performance, and it is shown how to incorporate it into the TGS algorithm. A discussion is also presented of the tradeoffs of using the bottom-up and top-down packing approaches. %B Information Processing Letters %V 101 %P 6 - 12 %8 2007/01/16/ %@ 0020-0190 %G eng %U http://www.sciencedirect.com/science/article/pii/S002001900600233X %N 1 %R 10.1016/j.ipl.2006.07.010 %0 Journal Article %J Protist %D 2007 %T The Expression of a Plant-type Ferredoxin Redox System provides Molecular Evidence for a Plastid in the Early Dinoflagellate Perkinsus marinus %A Stelter,Kathrin %A El‐Sayed, Najib M. %A Seeber,Frank %K Apicomplexa %K ferredoxin %K Perkinsozoa %K plastid %K transit peptide %X Perkinsus marinus is a parasitic protozoan with a phylogenetic positioning between Apicomplexa and dinoflagellates. It is thus of interest for reconstructing the early evolution of eukaryotes, especially with regard to the acquisition of secondary plastids in these organisms. It is also an important pathogen of oysters, and the definition of parasite-specific metabolic pathways would be beneficial for the identification of efficient treatments for infected mollusks. Although these different scientific interests have resulted in the start of a genome project for this organism, it is still unknown whether P. marinus contains a plastid or plastid-like organelle like the related dinoflagellates and Apicomplexa. Here, we show that in vitro-cultivated parasites contain transcripts of the plant-type ferredoxin and its associated reductase. Both proteins are nuclear-encoded and possess N-terminal targeting sequences similar to those characterized in dinoflagellates. Since this redox pair is exclusively found in cyanobacteria and plastid-harboring organisms its presence also in P. marinus is highly indicative of a plastid. We also provide additional evidence for such an organelle by demonstrating pharmacological sensitivity to inhibitors of plastid-localized enzymes involved in fatty acid biosynthesis (e.g. acetyl-CoA carboxylase) and by detection of genes for three enzymes of plastid-localized isoprenoid biosynthesis (1-deoxy-D-xylulose 5-phosphate reductoisomerase, (E)-4-hydroxy-3-methyl-but-2-enyl diphosphate reductase, and (E)-4-hydroxy-3-methyl-but-2-enyl diphosphate synthase). %B Protist %V 158 %P 119 - 130 %8 2007/01/22/ %@ 1434-4610 %G eng %U http://www.sciencedirect.com/science/article/pii/S1434461006001015 %N 1 %R 16/j.protis.2006.09.003 %0 Journal Article %J Computers & Graphics %D 2007 %T A fast all nearest neighbor algorithm for applications involving large point-clouds %A Sankaranarayanan,Jagan %A Samet, Hanan %A Varshney, Amitabh %K All nearest neighbor algorithm %K Disk-based data structures %K Incremental neighbor finding algorithm %K k Nearest neighbors %K kNN Algorithm %K Locality %K Neighbor finding %K Neighborhood %K Point-cloud graphics %K Point-cloud operations %X Algorithms that use point-cloud models make heavy use of the neighborhoods of the points. These neighborhoods are used to compute the surface normals for each point, mollification, and noise removal. All of these primitive operations require the seemingly repetitive process of finding the k nearest neighbors (kNNs) of each point. These algorithms are primarily designed to run in main memory. However, rapid advances in scanning technologies have made available point-cloud models that are too large to fit in the main memory of a computer. This calls for more efficient methods of computing the kNNs of a large collection of points many of which are already in close proximity. A fast kNN algorithm is presented that makes use of the locality of successive points whose k nearest neighbors are sought to reduce significantly the time needed to compute the neighborhood needed for the primitive operation as well as enable it to operate in an environment where the data is on disk. Results of experiments demonstrate an order of magnitude improvement in the time to perform the algorithm and several orders of magnitude improvement in work efficiency when compared with several prominent existing methods. %B Computers & Graphics %V 31 %P 157 - 174 %8 2007/04// %@ 0097-8493 %G eng %U http://www.sciencedirect.com/science/article/pii/S0097849306002378 %N 2 %R 10.1016/j.cag.2006.11.011 %0 Patent %D 2007 %T Fast and scalable approximation methods for finding minimum cost flows with ... %A Fleischer,Lisa Karen %A Saniee,Iraj %A Shepherd,Frederick Bruce %A Srinivasan, Aravind %E Lucent Technologies Inc. %X Broadly, techniques for solving network routing within a predetermined error are disclosed. These techniques may be applied to networks supporting dedicated reserve capacity, where reserved capacity on links in the network is dedicated for a particular commodity (generally, a source and sink pair of computers), and shared recovery, where reserved capacity on links is shared amongst two or more commodities. These techniques use an iterative process to determine flows on each of the links in a network. Costs are set for each commodity, and primary and secondary (i.e., backup) flows are initialized. A commodity is selected and demand for the commodity is routed through the shortest path. Costs are updated for each potential failure mode. For each commodity, the flows and costs are updated. Once all flows and costs are updated, then it is determined if a function is less than a predetermined value. If the function is less than a predetermined value, then the commodity selection, and... %V 10/053,079 %8 2007/10/09/ %G eng %U http://www.google.com/patents?id=rkWpAAAAEBAJ %N 7280526 %0 Conference Paper %B Fourth international workshop on Software quality assurance: in conjunction with the 6th ESEC/FSE joint meeting %D 2007 %T Faults' context matters %A Strecker,Jaymie %A Memon, Atif M. %X When choosing a testing technique, practitioners want to know which one will detect the faults that matter most to them in the programs that they plan to test. Do empirical evaluations of testing techniques provide this information? More often than not, they report how many faults in a carefully chosen "representative" sample the evaluated techniques detect. But the population of faults that such a sample would represent depends heavily on the faults' context or environment---as does the cost of failing to detect those faults. If empirical studies are to provide information that a practitioner can apply outside the context of the study, they must characterize the faults studied in a way that translates across contexts. A testing technique's fault-detecting abilities could then be interpreted relative to the fault characterization. In this paper, we present a list of criteria that a fault characterization must meet in order to be fit for this task, and we evaluate several well-known fault characterizations against the criteria. Two families of characterizations are found to satisfy the criteria: those based on graph models of programs and those based on faults' detection by testing techniques. %B Fourth international workshop on Software quality assurance: in conjunction with the 6th ESEC/FSE joint meeting %S SOQUA '07 %I ACM %C New York, NY, USA %P 112 - 115 %8 2007/// %@ 978-1-59593-724-7 %G eng %U http://doi.acm.org/10.1145/1295074.1295096 %R 10.1145/1295074.1295096 %0 Journal Article %J Scalable Uncertainty Management %D 2007 %T Finding most probable worlds of probabilistic logic programs %A Khuller, Samir %A Martinez,V. %A Nau, Dana S. %A Simari,G. %A Sliva,A. %A Subrahmanian,V. %X Probabilistic logic programs have primarily studied the problem of entailment of probabilistic atoms. However, there are some interesting applications where we are interested in finding a possible world that is most probable. Our first result shows that the problem of computing such ”maximally probable worlds” (MPW) is intractable. We subsequently show that we can often greatly reduce the size of the linear program used in past work (by Ng and Subrahmanian) and yet solve the problem exactly. However, the intractability results still make computational efficiency quite impossible. We therefore also develop several heuristics to solve the MPW problem and report extensive experimental results on the accuracy and efficiency of such heuristics. %B Scalable Uncertainty Management %P 45 - 59 %8 2007/// %G eng %R 10.1007/978-3-540-75410-7_4 %0 Book Section %B Scalable Uncertainty Management %D 2007 %T Finding Most Probable Worlds of Probabilistic Logic Programs %A Khuller, Samir %A Martinez,Vanina %A Nau, Dana S. %A Simari,Gerardo %A Sliva,Amy %A Subrahmanian,V. %E Prade,Henri %E Subrahmanian,V. %K Computer science %X Probabilistic logic programs have primarily studied the problem of entailment of probabilistic atoms. However, there are some interesting applications where we are interested in finding a possible world that is most probable. Our first result shows that the problem of computing such ”maximally probable worlds” (MPW) is intractable. We subsequently show that we can often greatly reduce the size of the linear program used in past work (by Ng and Subrahmanian) and yet solve the problem exactly. However, the intractability results still make computational efficiency quite impossible. We therefore also develop several heuristics to solve the MPW problem and report extensive experimental results on the accuracy and efficiency of such heuristics. %B Scalable Uncertainty Management %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4772 %P 45 - 59 %8 2007/// %@ 978-3-540-75407-7 %G eng %U http://www.springerlink.com/content/e4463p4rv4k01u93/abstract/ %0 Book %D 2007 %T Fundamentals of Multidimensional and Metric data structure %A Samet, Hanan %I Morgan-Kaufmann, San Francisco %8 2007/// %G eng %0 Conference Paper %B Proceedings of the International Joint Conference on Artificial Intelligence %D 2007 %T A general framework for reasoning about inconsistency %A V.S. Subrahmanian %A Amgoud,L. %X Numerous logics have been developed for reason-ing about inconsistency which differ in (i) the logic to which they apply, and (ii) the criteria used to draw inferences. In this paper, we propose a gen- eral framework for reasoning about inconsistency in a wide variety of logics including ones for which inconsistency resolution methods have not yet been studied (e.g. various temporal and epistemic log- ics). We start with Tarski and Scott’s axiomatiza- tion of logics, but drop their monotonicity require- ments that we believe are too strong for AI. For such a logic L, we define the concept of an option. Options are sets of formulas in L that are closed and consistent according to the notion of consequence and consistency in L. We show that by defining an appropriate preference relation on options, we can capture several existing works such as Brewka’s subtheories. We also provide algorithms to com- pute most preferred options. %B Proceedings of the International Joint Conference on Artificial Intelligence %P 599 - 604 %8 2007/// %G eng %0 Journal Article %J Emerging Infectious DiseasesEmerg Infect Dis %D 2007 %T Genome Analysis Linking Recent European and African Influenza (H5N1) Viruses %A Salzberg,Steven L. %A Kingsford, Carl %A Cattoli,Giovanni %A Spiro,David J. %A Janies,Daniel A. %A Aly,Mona Mehrez %A Brown,Ian H. %A Couacy-Hymann,Emmanuel %A De Mia,Gian Mario %A Dung,Do Huu %A Guercio,Annalisa %A Joannis,Tony %A Ali,Ali Safar Maken %A Osmani,Azizullah %A Padalino,Iolanda %A Saad,Magdi D. %A Savić,Vladimir %A Sengamalay,Naomi A. %A Yingst,Samuel %A Zaborsky,Jennifer %A Zorman-Rojs,Olga %A Ghedin,Elodie %A Capua,Ilaria %X Although linked, these viruses are distinct from earlier outbreak strains., To better understand the ecology and epidemiology of the highly pathogenic avian influenza virus in its transcontinental spread, we sequenced and analyzed the complete genomes of 36 recent influenza A (H5N1) viruses collected from birds in Europe, northern Africa, and southeastern Asia. These sequences, among the first complete genomes of influenza (H5N1) viruses outside Asia, clearly depict the lineages now infecting wild and domestic birds in Europe and Africa and show the relationships among these isolates and other strains affecting both birds and humans. The isolates fall into 3 distinct lineages, 1 of which contains all known non-Asian isolates. This new Euro-African lineage, which was the cause of several recent (2006) fatal human infections in Egypt and Iraq, has been introduced at least 3 times into the European-African region and has split into 3 distinct, independently evolving sublineages. One isolate provides evidence that 2 of these sublineages have recently reassorted. %B Emerging Infectious DiseasesEmerg Infect Dis %V 13 %P 713 - 718 %8 2007/05// %@ 1080-6040 %G eng %N 5 %R 10.3201/eid1305.070013 %0 Conference Paper %B Information Visualization, 2007. IV '07. 11th International Conference %D 2007 %T GeoDDupe: A Novel Interface for Interactive Entity Resolution in Geospatial Data %A Kang,Hyunmo %A Sehgal,V. %A Getoor, Lise %K algorithms;data %K cleaning;geospatial %K data %K entity %K GeoDDupe;automatic %K interfaces; %K mining %K mining;geography;user %K mining;geospatial %K resolution;interactive %K tool;network %K visualization;data %X Due to the growing interest in geospatial data mining and analysis, data cleaning and integration in geospatial data is becoming an important issue. Geospatial entity resolution is the process of reconciling multiple location references to the same real world location within a single data source (deduplication) or across multiple data sources (integration). In this paper, we introduce an interactive tool called GeoDDupe which effectively combines automatic data mining algorithms for geospatial entity resolution with a novel network visualization supporting users' resolution analysis and decisions. We illustrate the GeoDDupe interface with an example geospatial dataset and show how users can efficiently and accurately resolve location entities. Finally, the case study with two real-world geospatial datasets demonstrates the potential of GeoDDupe. %B Information Visualization, 2007. IV '07. 11th International Conference %P 489 - 496 %8 2007/07// %G eng %R 10.1109/IV.2007.55 %0 Conference Paper %B Proceedings of the 4th workshop on Data management for sensor networks: in conjunction with 33rd International Conference on Very Large Data Bases %D 2007 %T A graph-based approach to vehicle tracking in traffic camera video streams %A Shahri,Hamid Haidarian %A Namata,Galileo %A Navlakha,Saket %A Deshpande, Amol %A Roussopoulos, Nick %X Vehicle tracking has a wide variety of applications from law enforcement to traffic planning and public safety. However, the image resolution of the videos available from most traffic camera systems, make it difficult to track vehicles based on unique identifiers like license plates. In many cases, vehicles with similar attributes are indistinguishable from one another due to image quality issues. Often, network bandwidth and power constraints limit the frame rate, as well. In this paper, we discuss the challenges of performing vehicle tracking queries over video streams from ubiquitous traffic cameras. We identify the limitations of tracking vehicles individually in such conditions and provide a novel graph-based approach using the identity of neighboring vehicles to improve the performance. We evaluate our approach using streaming video feeds from live traffic cameras available on the Internet. The results show that vehicle tracking is feasible, even for low quality and low frame rate traffic cameras. Additionally, exploitation of the attributes of neighboring vehicles significantly improves the performance. %B Proceedings of the 4th workshop on Data management for sensor networks: in conjunction with 33rd International Conference on Very Large Data Bases %S DMSN '07 %I ACM %C New York, NY, USA %P 19 - 24 %8 2007/// %@ 978-159593-911-1 %G eng %U http://doi.acm.org/10.1145/1286380.1286386 %R 10.1145/1286380.1286386 %0 Conference Paper %B Proceedings of the National Conference on Artificial Intelligence %D 2007 %T GRIN: A graph based RDF index %A Udrea,O. %A Pugliese, A. %A V.S. Subrahmanian %X RDF (“Resource Description Framework”) is now a widelyused World Wide Web Consortium standard. However, methods to index large volumes of RDF data are still in their infancy. In this paper, we focus on providing a very lightweight indexing mechanism for certain kinds of RDF queries, namely graph-based queries where there is a need to traverse edges in the graph determined by an RDF database. Our approach uses the idea of drawing circles around selected “center” vertices in the graph where the circle would encom- pass those vertices in the graph that are within a given dis- tance of the “center” vertex. We come up with methods of finding such “center” vertices and identifying the radius of the circles and then leverage this to build an index called GRIN. We compare GRIN with three existing RDF indexex: Jena, Sesame, and RDFBroker. We compared (i) the time to an- swer graph based queries, (ii) memory needed to store the index, and (iii) the time to build the index. GRIN outper- forms Jena, Sesame and RDFBroker on all three measures for graph based queries (for other types of queries, it may be worth building one of these other indexes and using it), at the expense of using a larger amount of memory when answering queries. %B Proceedings of the National Conference on Artificial Intelligence %V 22 %P 1465 - 1465 %8 2007/// %G eng %0 Journal Article %J Genome Biology %D 2007 %T Hawkeye: an interactive visual analytics tool for genome assemblies %A Schatz,Michael C %A Phillippy,Adam M %A Shneiderman, Ben %A Salzberg,Steven L. %X Genome sequencing remains an inexact science, and genome sequences can contain significant errors if they are not carefully examined. Hawkeye is our new visual analytics tool for genome assemblies, designed to aid in identifying and correcting assembly errors. Users can analyze all levels of an assembly along with summary statistics and assembly metrics, and are guided by a ranking component towards likely mis-assemblies. Hawkeye is freely available and released as part of the open source AMOS project http://amos.sourceforge.net/hawkeye. %B Genome Biology %V 8 %P R34 - R34 %8 2007/03/09/ %@ 1465-6906 %G eng %U http://genomebiology.com/2007/8/3/R34 %N 3 %R 10.1186/gb-2007-8-3-r34 %0 Journal Article %J BMC Bioinformatics %D 2007 %T High-throughput sequence alignment using Graphics Processing Units %A Schatz,Michael C %A Trapnell,Cole %A Delcher,Arthur L. %A Varshney, Amitabh %B BMC Bioinformatics %V 8 %P 474 - 474 %8 2007/// %@ 1471-2105 %G eng %U http://www.biomedcentral.com/1471-2105/8/474 %N 1 %R 10.1186/1471-2105-8-474 %0 Conference Paper %B Proc. 6th ACM Workshop on Hot Topics in Networks (Hotnets-VI) %D 2007 %T Holding the Internet Accountable %A Andersen,David %A Balakrishnan,Hari %A Feamster, Nick %A Koponen,Teemu %A Moon,Daekyong %A Shenker,Scott %X Today’s IP network layer provides little to no protection against misconfiguration or malice. Despite some progress in improving the robustness and security of the IP layer, misconfigurations and attacks still occur frequently. We show how a network layer that provides accountability, i.e., the ability to associate each action with the responsible entity, provides a firm foundation for defenses against misconfiguration and malice. We present the design of a network layer that incorporates accountability called AIP (Accountable Internet Protocol) and show how its features—notably, its use of self-certifying addresses— can improve both source accountability (the ability to trace actions to a particular end host and stop that host from misbehaving) and control-plane accountability (the ability to pinpoint and prevent attacks on routing). %B Proc. 6th ACM Workshop on Hot Topics in Networks (Hotnets-VI) %8 2007/11/01/ %G eng %U http://repository.cmu.edu/compsci/66 %0 Conference Paper %B Proceedings of the Workshop on Metareasoning in Agent-Based Systems %D 2007 %T Hood College, Master of Business Administration, 2005 Hood College, Master of Science (Computer Science), 2001 Hood College, Bachelor of Science (Computer Science), 1998 Frederick Community College, Associate in Arts (Business Administration), 1993 %A Anderson,M. L %A Schmill,M. %A Oates,T. %A Perlis, Don %A Josyula,D. %A Wright,D. %A Human,S. W.T.D.N %A Metacognition,L. %A Fults,S. %A Josyula,D. P %B Proceedings of the Workshop on Metareasoning in Agent-Based Systems %8 2007/// %G eng %0 Book Section %B UbiComp 2007: Ubiquitous Computing %D 2007 %T How Smart Homes Learn: The Evolution of the Networked Home and Household %A Marshini Chetty %A Sung, Ja-Young %A Grinter, Rebecca E. %E Krumm, John %E Abowd, Gregory D. %E Seneviratne, Aruna %E Strang, Thomas %K Computer Communication Networks %K computers and society %K home networking %K Information Systems Applications (incl.Internet) %K infrastructure %K smart home %K software engineering %K Systems and Data Security %K User Interfaces and Human Computer Interaction %X Despite a growing desire to create smart homes, we know little about how networked technologies interact with a house’s infrastructure. In this paper, we begin to close this gap by presenting findings from a study that examined the relationship between home networking and the house itself—and the work that results for householders as a consequence of this interaction. We discuss four themes that emerged: an ambiguity in understanding the virtual boundaries created by wireless networks, the home network control paradox, a new home network access paradox, and the relationship between increased responsibilities and the possibilities of wireless networking. %B UbiComp 2007: Ubiquitous Computing %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 127 - 144 %8 2007/01/01/ %@ 978-3-540-74852-6, 978-3-540-74853-3 %G eng %U http://link.springer.com/chapter/10.1007/978-3-540-74853-3_8 %0 Journal Article %J IEEE Intelligent Systems %D 2007 %T Human Responsibility for Autonomous Agents %A Shneiderman, Ben %K Automatic control %K Autonomous agents %K autonomous systems %K Bandwidth %K Computer bugs %K Computer errors %K Control systems %K data privacy %K Human-computer interaction %K HUMANS %K Robots %K Safety %K Software design %X Automated or autonomous systems can sometimes fail harmlessly, but they can also destroy data, compromise privacy, and consume resources, such as bandwidth or server capacity. What's more troubling is that automated systems embedded in vital systems can cause financial losses, destruction of property, and loss of life. Controlling these dangers will increase trust while enabling broader use of these systems with higher degrees of safety. Obvious threats stem from design errors and software bugs, but we can't overlook mistaken assumptions by designers, unanticipated actions by humans, and interference from other computerized systems. This article is part of a special issue on Interacting with Autonomy. %B IEEE Intelligent Systems %V 22 %P 60 - 61 %8 2007/04//March %@ 1541-1672 %G eng %N 2 %R 10.1109/MIS.2007.32 %0 Book Section %B Human-Computer Interaction – INTERACT 2007Human-Computer Interaction – INTERACT 2007 %D 2007 %T Human Values for Shaping the Made World %A Shneiderman, Ben %E Baranauskas,Cécilia %E Palanque,Philippe %E Abascal,Julio %E Barbosa,Simone %X Interface design principles have been effective in shaping new desktop applications, web-based resources, and mobile devices. Usability and sociability promote successful online communities and social network services. The contributions of human-computer interaction researchers have been effective in raising the quality of design of many products and services. As our influence grows, we can play an even more profound role in guaranteeing that enduring human values are embedded in the next generation of technology. This talk identifies which goals are realistic, such as universality, responsibility, trust, empathy, and privacy, and how we might ensure that they become part of future services and systems. %B Human-Computer Interaction – INTERACT 2007Human-Computer Interaction – INTERACT 2007 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4662 %P 1 - 1 %8 2007/// %@ 978-3-540-74794-9 %G eng %U http://dx.doi.org/10.1007/978-3-540-74796-3_1 %0 Journal Article %J ACM Transactions on Mathematical Software %D 2007 %T IFISS: A Matlab toolbox for modelling incompressible flow %A Elman, Howard %A Ramage, A. %A Silvester, D. J %B ACM Transactions on Mathematical Software %V 33 %8 2007/// %G eng %N 2 %0 Journal Article %J Proceedings of SPIE %D 2007 %T An improved asymmetry measure to detect breast cancer %A Tahmoush,Dave %A Samet, Hanan %X Radiologists can use the differences between the left and right breasts, or asymmetry, in mammograms to help detect certain malignant breast cancers. An image similarity method has been improved to make use of this knowledge base to recognize breast cancer. Image similarity is determined using computer-aided detection (CAD) prompts as the features, and then a cluster comparison is done to determine whether there is asymmetry. We develop the analysis through a combination of clustering and supervised learning of model parameters. This process correctly classifies cancerous mammograms 95% of the time, and all mammograms 84% of the time, and thus asymmetry is a measure that can play an important role in significantly improving computer-aided breast cancer detection systems. This technique represents an improvement in accuracy of 121% over commercial techniques on non-cancerous cases. Most computer-aided detection (CAD) systems are tested on images which contain cancer on the assumption that images without cancer would produce the same number of false positives. However, a pre-screening system is designed to remove the normal cases from consideration, and so the inclusion of a pre-screening system into CAD dramatically reduces the number of false positives reported by the CAD system. We define three methods for the inclusion of pre-screening into CAD, and improve the performance of the CAD system by over 70% at low levels of false positives. %B Proceedings of SPIE %V 6514 %P 65141Q-65141Q-9 - 65141Q-65141Q-9 %8 2007/03/08/ %@ 0277786X %G eng %U http://spiedigitallibrary.org/proceedings/resource/2/psisdg/6514/1/65141Q_1?isAuthorized=no %N 1 %R doi:10.1117/12.708327 %0 Conference Paper %B Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on %D 2007 %T In Situ Evaluation of Tracking Algorithms Using Time Reversed Chains %A Wu,Hao %A Sankaranarayanan,A. C %A Chellapa, Rama %K (numerical %K algorithm;Markov %K chain;tracking %K decision %K density;time %K detection;particle %K Evaluation %K evaluation;object %K filter;performance %K Filtering %K Markov %K methods);tracking;visual %K processes;decision %K reversed %K servoing; %K situ %K statistics;in %K strategy;posterior %K systems;visual %K theory;object %K tracking %K tracking;particle %X Automatic evaluation of visual tracking algorithms in the absence of ground truth is a very challenging and important problem. In the context of online appearance modeling, there is an additional ambiguity involving the correctness of the appearance model. In this paper, we propose a novel performance evaluation strategy for tracking systems based on particle filter using a time reversed Markov chain. Starting from the latest observation, the time reversed chain is propagated back till the starting time t = 0 of the tracking algorithm. The posterior density of the time reversed chain is also computed. The distance between the posterior density of the time reversed chain (at t = 0) and the prior density used to initialize the tracking algorithm forms the decision statistic for evaluation. It is postulated that when the data is generated true to the underlying models, the decision statistic takes a low value. We empirically demonstrate the performance of the algorithm against various common failure modes in the generic visual tracking problem. Finally, we derive a small frame approximation that allows for very efficient computation of the decision statistic. %B Computer Vision and Pattern Recognition, 2007. CVPR '07. IEEE Conference on %P 1 - 8 %8 2007/06// %G eng %R 10.1109/CVPR.2007.382992 %0 Conference Paper %B Current Trends in Computer Science, 2007. ENC 2007. Eighth Mexican International Conference on %D 2007 %T Indexing Methods for Similarity Searching %A Samet, Hanan %K content-based %K database;indexing %K database;pattern %K databases; %K indexing;multimedia %K methods;multimedia %K recognition;similarity %K retrieval;data %K searching;data %K structures;database %K structures;image %X An overview is given of the various techniques and issues involved in providing indexing support for similarity searching. Similarity searching is a crucial part of retrieval in multimedia databases used for applications such as pattern recognition, image databases, and content-based retrieval. It involves finding objects in a data set S that are similar to a query object q based on some distance measure d which is usually a distance metric. The search process is usually achieved by means of nearest neighbor finding. Existing methods for handling similarity search in this setting fall into one of two classes. The first is based on mapping to a vector space. The vector space is usually of high dimension which requires special handling due to the fact indexing methods do not discriminate well in such spaces. In particular, the query regions often overlap all of the blocks that result from the decomposition of the underlying space. This has led to some special solutions that make use of a sequential scan. An alternative is to use dimension reduction to find a mapping from a high-dimensional space into a low-dimensional space by finding the most discriminating dimensions and then index the data using one of a number of different data structures such as k-d trees, R-trees, quadtrees, etc. The second directly indexes the objects based on distances making use of data structures such as the vp-tree, M-tree, etc. %B Current Trends in Computer Science, 2007. ENC 2007. Eighth Mexican International Conference on %P xv - xv %8 2007/09// %G eng %R 10.1109/ENC.2007.9 %0 Conference Paper %B Data Engineering, 2007. ICDE 2007. IEEE 23rd International Conference on %D 2007 %T Indexing Point Triples Via Triangle Geometry %A Cranston,C.B. %A Samet, Hanan %K database %K databases; %K dimension;spatial %K geometry;database %K hyperdimensional %K index %K index;single %K index;triangle %K indexing;query %K linear %K point;k-fold %K processing;visual %K relationships;structured %K rotational %K search;indexing %K space;image %K symmetry;point-based %X Database search for images containing icons with specific mutual spatial relationships can be facilitated by an appropriately structured index. For the case of images containing subsets each of which consist of three icons, the one-to-one correspondence between (distinct) point triples and triangles allows the use of such triangle attributes as position, size, orientation, and "shape" in constructing a point-based index, in which each triangle maps to a single point in a resulting hyperdimensional index space. Size (based on the triangle perimeter) can be represented by a single linear dimension. The abstract "shape" of a triangle induces a space that is inherently two-dimensional, and a number of alternative definitions of a basis for this space are examined. Within a plane, orientation reduces to rotation, and (after assignment of a reference direction for the triangle) can be represented by a single, spatially closed dimension. However, assignment of a reference direction for triangles possessing a k-fold rotational symmetry presents a significant challenge. Methods are described for characterizing shape and orientation of triangles, and for mapping these attributes onto a set of linear axes to form a combined index. The shape attribute is independent of size, orientation, and position, and the characterization of shape and orientation is stable with respect to small variations in the indexed triangles. %B Data Engineering, 2007. ICDE 2007. IEEE 23rd International Conference on %P 936 - 945 %8 2007/04// %G eng %R 10.1109/ICDE.2007.367939 %0 Journal Article %J Proc. SPIE Conf. Security, Steganography, Watermarking of Multimedia Contents %D 2007 %T Intrinsic fingerprints for image authentication and steganalysis %A Swaminathan,A. %A Wu,M. %A Liu,K. J.R %X With growing popularity of digital imaging devices and low-cost image editing software, the integrity of imagecontent can no longer be taken for granted. This paper introduces a methodology for forensic analysis of digital camera images, based on the observation that many in-camera and post-camera processing operations leave distinct traces on digital images. We present methods to identify these intrinsic fingerprint traces of the various processing operations and employ them to verify the authenticity of digital data. We develop an explicit imaging model to characterize the properties that should be satisfied by a direct camera output, and model any further processing applied to the camera captured image by a manipulation filter. Utilizing the manipulation filter coefficients and reference patterns estimated from direct camera outputs using blind deconvolution techniques, the proposed methods are capable of detecting manipulations made by previously unseen operations and steganographic embedding. %B Proc. SPIE Conf. Security, Steganography, Watermarking of Multimedia Contents %8 2007/// %G eng %0 Journal Article %J Proceedings of the 8th International Symposium on Mathematical Morphology %D 2007 %T Jmin-image based color-texture segmentation using watershed and hierarchical clustering %A Santos, T.T. %A Morimoto, CH %A Chellapa, Rama %B Proceedings of the 8th International Symposium on Mathematical Morphology %V 2 %P 35 - 36 %8 2007/10//undefin %G eng %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference on %D 2007 %T Joint Acoustic-Video Fingerprinting of Vehicles, Part II %A Cevher, V. %A Guo, F. %A Sankaranarayanan,A. C %A Chellapa, Rama %K acoustic-video %K analysis;video %K approximations;acoustic %K Bayesian %K colour %K density %K efficiency;joint %K estimation;Bayes %K fingerprinting;metrology %K framework;Laplacian %K functions;performance %K fusion;color %K identification;image %K invariants;computational %K methods;acoustic %K metrology;acoustic %K processing; %K processing;fingerprint %K signal %K video %X In this second paper, we first show how to estimate the wheelbase length of a vehicle using line metrology in video. We then address the vehicle fingerprinting problem using vehicle silhouettes and color invariants. We combine the acoustic metrology and classification results discussed in Part I with the video results to improve estimation performance and robustness. The acoustic video fusion is achieved in a Bayesian framework by assuming conditional independence of the observations of each modality. For the metrology density functions, Laplacian approximations are used for computational efficiency. Experimental results are given using field data %B Acoustics, Speech and Signal Processing, 2007. ICASSP 2007. IEEE International Conference on %V 2 %P II-749 -II-752 - II-749 -II-752 %8 2007/04// %G eng %R 10.1109/ICASSP.2007.366344 %0 Journal Article %J Selected Topics in Signal Processing, IEEE Journal of %D 2007 %T Joint Uplink and Downlink Optimization for Real-Time Multiuser Video Streaming Over WLANs %A Su,Guan-Ming %A Han,Zhu %A M. Wu %A Liu,K. J.R %K 802.11 %K allocation;image %K allocation;real-time %K communications;maximal %K distortion;multiuser %K end-to-end %K error %K expected %K IEEE %K LAN; %K mechanism;downlink %K multiuser %K networks;WLAN;cross-layer %K optimization;interactive %K optimization;video %K Protection %K quality;video %K resource %K sequences;video %K streaming %K streaming;uplink %K streaming;wireless %K system;bandwidth %K video %X In this paper, a network-aware and source-aware video streaming system is proposed to support interactive multiuser communications within single-cell and multicell IEEE 802.11 networks. Unlike the traditional streaming video services, the strict delay constraints of an interactive video streaming system pose more challenges. These challenges include the heterogeneity of uplink and downlink channel conditions experienced by different users, the multiuser resource allocation of limited radio resources, the incorporation of the cross-layer design, and the diversity of content complexities exhibited by different video sequences. With the awareness of video content and network resources, the proposed system integrates cross-layer error protection mechanism and performs dynamic resource allocation across multiple users. We formulate the proposed system as to minimize the maximal end-to-end expected distortion received by all users, subject to maximal transmission power and delay constraints. To reduce the high dimensionality of the search space, fast multiuser algorithms are proposed to find the near-optimal solutions. Compared to the strategy without dynamically and jointly allocating bandwidth resource for uplinks and downlinks, the proposed framework outperforms by 2.18 7.95 dB in terms of the average received PSNR of all users and by 3.82 11.50 dB in terms of the lowest received PSNR among all users. Furthermore, the proposed scheme can provide more uniform video quality for all users and lower quality fluctuation for each received video sequence. %B Selected Topics in Signal Processing, IEEE Journal of %V 1 %P 280 - 294 %8 2007/08// %@ 1932-4553 %G eng %N 2 %R 10.1109/JSTSP.2007.901518 %0 Journal Article %J IEEE transactions on pattern analysis and machine intelligence %D 2007 %T K-nearest neighbor finding using MaxNearestDist %A Samet, Hanan %B IEEE transactions on pattern analysis and machine intelligence %P 243 - 252 %8 2007/// %G eng %0 Conference Paper %B Proceedings of the 8th annual international conference on Digital government research: bridging disciplines & domains %D 2007 %T Knowledge discovery using the sand spatial browser %A Samet, Hanan %A Phillippy,Adam %A Sankaranarayanan,Jagan %K distance semi-join %K knowledge discovery %K sand database system %K snow cholera map %X The use of the SAND Internet Browser as a knowledge discovery tool for epidemiological cartography is highlighted by recreating the results of Dr. John Snow's study of the 1854 Cholera epidemic in Soho, London. %B Proceedings of the 8th annual international conference on Digital government research: bridging disciplines & domains %S dg.o '07 %I Digital Government Society of North America %P 284 - 285 %8 2007/// %@ 1-59593-599-1 %G eng %U http://dl.acm.org/citation.cfm?id=1248460.1248521 %0 Journal Article %J SIAM J. Sci. Comput %D 2007 %T Least squares preconditioners for stabilized discretizations of the Navier-Stokes equations %A Elman, Howard %A Howle, V. E %A Shadid, J. %A Silvester, D. %A Tuminaro, R. %B SIAM J. Sci. Comput %V 30 %P 290 - 311 %8 2007/// %G eng %N 1 %0 Report %D 2007 %T Link-based Classification %A Sen,Prithviraj %A Getoor, Lise %K Technical Report %X Over the past few years, a number of approximate inference algorithms fornetworked data have been put forth. We empirically compare the performance of three of the popular algorithms: loopy belief propagation, mean field relaxation labeling and iterative classification. We rate each algorithm in terms of its robustness to noise, both in attribute values and correlations across links. We also compare them across varying types of correlations across links. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-11 %8 2007/02/19/ %G eng %U http://drum.lib.umd.edu/handle/1903/4298 %0 Conference Paper %B Signal Processing Systems, 2007 IEEE Workshop on %D 2007 %T Low-overhead run-time scheduling for fine-grained acceleration of signal processing systems %A Boutellier,J. %A Bhattacharyya, Shuvra S. %A Silvén,O. %B Signal Processing Systems, 2007 IEEE Workshop on %P 457 - 462 %8 2007/// %G eng %0 Conference Paper %B Information Reuse and Integration, 2007. IRI 2007. IEEE International Conference on %D 2007 %T MAGIC: A Multi-Activity Graph Index for Activity Detection %A Albanese, M. %A Pugliese, A. %A V.S. Subrahmanian %A Udrea,O. %K complexity;graph %K detection;complexity %K graph %K index;probability %K MAGIC;activity %K reducing %K restrictions;exponential %K theory;probability; %K threshold;computational %K time;multiactivity %X Suppose we are given a set A of activities of interest, a set O of observations, and a probability threshold p. We are interested in finding the set of all pairs (a, O'), where a epsi A and O' sube O, that minimally validate the fact that an instance of activity a occurs in O with probability p or more. The novel contribution of this paper is the notion of the multi-activity graph index (MAGIC), which can index very large numbers of observations from interleaved activities and quickly retrieve completed instances of the monitored activities. We introduce two complexity reducing restrictions of the problem (which takes exponential time) and develop algorithms for each. We experimentally evaluate our exponential algorithm as well as the restricted algorithms on both synthetic data and a real (depersonalized) travel data set consisting of 5.5 million observations. Our experiments show that MAGIC consumes reasonable amounts of memory and can retrieve completed instances of activities in just a few seconds. We also report appropriate statistical significance results validating our experimental hypotheses. %B Information Reuse and Integration, 2007. IRI 2007. IEEE International Conference on %P 267 - 272 %8 2007/08// %G eng %R 10.1109/IRI.2007.4296632 %0 Journal Article %J The International Journal of Advanced Manufacturing Technology %D 2007 %T Manufacturing multi-material articulated plastic products using in-mold assembly %A Priyadarshi,Alok %A Gupta, Satyandra K. %A Gouker,Regina %A Krebs,Florian %A Shroeder,Martin %A Warth,Stefan %K engineering %X In-mold assembly can be used to create plastic products with articulated joints. This process eliminates the need for post-molding assembly and reduces the number of parts being used in the product, hence improving the product quality. However, designing both products and molds is significantly more challenging in case of in-mold assembly. Currently, a systematic methodology does not exist for developing product and processes to exploit potential benefits of in-mold assembly for creating articulated joints. This paper is a step towards creating such a methodology and reports the following three results. First, it presents a model for designing assemblies and molding process so that the joint clearances and variation in the joint clearances can meet the performance goals. Second, it describes proven mold design templates for realizing revolute, prismatic, and spherical joints. Third, it describes a mold design methodology for designing molds for products that contain articulated joints and will be produced using in-mold assembly process. Three case studies are also presented to illustrate how in-mold assembly process can be used to create articulated devices. %B The International Journal of Advanced Manufacturing Technology %V 32 %P 350 - 365 %8 2007/// %@ 0268-3768 %G eng %U http://www.springerlink.com/content/q77nl78827r7w461/abstract/ %N 3 %R 10.1007/s00170-005-0343-z %0 Conference Paper %B Proceedings of the Eleventh European Workshop on Natural Language Generation %D 2007 %T Measuring variability in sentence ordering for news summarization %A Madnani,Nitin %A Passonneau,Rebecca %A Ayan,Necip Fazil %A Conroy,John M. %A Dorr, Bonnie J %A Klavans,Judith L. %A O'Leary, Dianne P. %A Schlesinger,Judith D. %X The issue of sentence ordering is an important one for natural language tasks such as multi-document summarization, yet there has not been a quantitative exploration of the range of acceptable sentence orderings for short texts. We present results of a sentence reordering experiment with three experimental conditions. Our findings indicate a very high degree of variability in the orderings that the eighteen subjects produce. In addition, the variability of reorderings is significantly greater when the initial ordering seen by subjects is different from the original summary. We conclude that evaluation of sentence ordering should use multiple reference orderings. Our evaluation presents several metrics that might prove useful in assessing against multiple references. We conclude with a deeper set of questions: (a) what sorts of independent assessments of quality of the different reference orderings could be made and (b) whether a large enough test set would obviate the need for such independent means of quality assessment. %B Proceedings of the Eleventh European Workshop on Natural Language Generation %S ENLG '07 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 81 - 88 %8 2007/// %G eng %U http://dl.acm.org/citation.cfm?id=1610163.1610177 %0 Patent %D 2007 %T Method of three-dimensional object reconstruction from a video sequence using a generic model %A Chellapa, Rama %A Chowdhury,Amit K. Roy %A Srinivasan,Sridhar %E University of Maryland %X In a novel method of 3D modeling of an object from a video sequence using an SfM algorithm and a generic object model, the generic model is incorporated after the SfM algorithm generates a 3D estimate of the object model purely and directly from the input video sequence. An optimization framework provides for comparison of the local trends of the 3D estimate and the generic model so that the errors in the 3D estimate are corrected. The 3D estimate is obtained by fusing intermediate 3D reconstructions of pairs of frames of the video sequence after computing the uncertainty of the two frame solutions. The quality of the fusion algorithm is tracked using a rate-distortion function. In order to combine the generic model with the 3D estimate, an energy function minimization procedure is applied to the 3D estimate. The optimization is performed using a Metropolis-Hasting sampling strategy. %V 10/644,817 %8 2007/02/27/ %G eng %U http://www.google.com/patents?id=4At_AAAAEBAJ %N 7184071 %0 Journal Article %J BMC Bioinformatics %D 2007 %T Minimus: a fast, lightweight genome assembler %A Sommer,Daniel D %A Delcher,Arthur L. %A Salzberg,Steven L. %A Pop, Mihai %X Genome assemblers have grown very large and complex in response to the need for algorithms to handle the challenges of large whole-genome sequencing projects. Many of the most common uses of assemblers, however, are best served by a simpler type of assembler that requires fewer software components, uses less memory, and is far easier to install and run. %B BMC Bioinformatics %V 8 %P 64 - 64 %8 2007/02/26/ %@ 1471-2105 %G eng %U http://www.biomedcentral.com/1471-2105/8/64 %N 1 %R 10.1186/1471-2105-8-64 %0 Journal Article %J Information Processing & Management %D 2007 %T Multi-candidate reduction: Sentence compression as a tool for document summarization tasks %A Zajic, David %A Dorr, Bonnie J %A Jimmy Lin %A Schwartz,Richard %K Headline generation %K Hidden Markov model %K Parse-and-trim %K Summarization %X This article examines the application of two single-document sentence compression techniques to the problem of multi-document summarization—a “parse-and-trim” approach and a statistical noisy-channel approach. We introduce the multi-candidate reduction (MCR) framework for multi-document summarization, in which many compressed candidates are generated for each source sentence. These candidates are then selected for inclusion in the final summary based on a combination of static and dynamic features. Evaluations demonstrate that sentence compression is a valuable component of a larger multi-document summarization framework. %B Information Processing & Management %V 43 %P 1549 - 1570 %8 2007/11// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306457307000295 %N 6 %R 10.1016/j.ipm.2007.01.016 %0 Conference Paper %D 2007 %T Multi-Dimensional Range Query over Encrypted Data %A Elaine Shi %A Bethencourt, J. %A Chan, T.-H.H. %A Song, Dawn %A Perrig, A. %K auditing %K cryptography %K data privacy %K decision bilinear Diffie-Hellman %K decision linear assumptions %K encrypted data %K multi-dimensional range query %K network audit logs %K network gateway %K network intrusions %X We design an encryption scheme called multi-dimensional range query over encrypted data (MRQED), to address the privacy concerns related to the sharing of network audit logs and various other applications. Our scheme allows a network gateway to encrypt summaries of network flows before submitting them to an untrusted repository. When network intrusions are suspected, an authority can release a key to an auditor, allowing the auditor to decrypt flows whose attributes (e.g., source and destination addresses, port numbers, etc.) fall within specific ranges. However, the privacy of all irrelevant flows are still preserved. We formally define the security for MRQED and prove the security of our construction under the decision bilinear Diffie-Hellman and decision linear assumptions in certain bilinear groups. We study the practical performance of our construction in the context of network audit logs. Apart from network audit logs, our scheme also has interesting applications for financial audit logs, medical privacy, untrusted remote storage, etc. In particular, we show that MRQED implies a solution to its dual problem, which enables investors to trade stocks through a broker in a privacy-preserving manner. %P 350 - 364 %8 2007 %G eng %0 Journal Article %J ACM Trans. Program. Lang. Syst. %D 2007 %T Mutatis Mutandis: Safe and predictable dynamic software updating %A Stoyle,Gareth %A Hicks, Michael W. %A Bierman,Gavin %A Sewell,Peter %A Neamtiu,Iulian %K capability %K dynamic software updating %K proteus %K Type inference %K updateability analysis %X This article presents Proteus, a core calculus that models dynamic software updating, a service for fixing bugs and adding features to a running program. Proteus permits a program's type structure to change dynamically but guarantees the updated program remains type-correct by ensuring a property we call con-freeness. We show how con-freeness can be enforced dynamically, and how it can be approximated via a novel static analysis. This analysis can be used to assess the implications of a program's structure on future updates in order to make update success more predictable. We have implemented Proteus for C, and briefly discuss our implementation which we have tested on several well-known programs. %B ACM Trans. Program. Lang. Syst. %V 29 %8 2007/08// %@ 0164-0925 %G eng %U http://doi.acm.org/10.1145/1255450.1255455 %N 4 %R 10.1145/1255450.1255455 %0 Conference Paper %B Proceedings of SPIE-Medical Imaging %D 2007 %T A new database for medical images and information %A Tahmoush,D. %A Samet, Hanan %X We present a medical image and medical record database for the storage, research,transmission, and evaluation of medical images, as well as tele-medicine applications. Any medical image from a source that supports the DICOM standard can be stored and accessed, as well as associated analysis and annotations. Information and image retrieval can be done based on patient info, date, doctor’s annotations, features in the images, or a spatial combination of features. Secure access and transmission is addressed for tele- medicine applications. This database application follows all HIPAA regulations. %B Proceedings of SPIE-Medical Imaging %8 2007/// %G eng %0 Journal Article %J Bangladesh Journal of Plant Taxonomy %D 2007 %T New records of phytoplankton for Bangladesh. 3. Volvocales %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasmin,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X This study presents 21 species of Chlamydomonas, four species of Carteria, two species of each of Nephroselmis, Pyramidomonas and Scherffelia, and Collodictyon triciliatum, Polytoma minus, Tetrachloridium ? allorgei and Tetraselmis cordiformis. These species have been reported from some ponds of Mathbaria of Pirojpur and Bakerganj of Barisal districts in Bangladesh. %B Bangladesh Journal of Plant Taxonomy %V 14 %8 2007/12/01/ %@ 1028-2092 %G eng %U http://www.banglajol.info/bd/index.php/BJPT/article/viewArticle/518 %N 1 %R 10.3329/bjpt.v14i1.518 %0 Journal Article %J Bangladesh Journal of Plant Taxonomy %D 2007 %T New records of phytoplankton for Bangladesh. 4. Chlorococcales %A Khondker,Moniruzzaman %A Bhuiyan,Rauf Ahmed %A Yeasim,Jenat %A Alam,Munirul %A Sack,R. Bradley %A Huq,Anwar %A Rita R Colwell %X This study presents three species from each of Schroederia, Monoraphidium and Ankistrodesmus, two species and one variety of Dictyosphaerium, two varieties of Pediastrum, and Tetraedron arthrodesmiforme var. contorta, Chlorotetraedron polymorphum, Myrmecia aquatica, Oocystis tainoensis, Nephrocytium spirale, Kirchneriella irregularis, Coelastrum indicum and Scenedesmus similagineus. These taxa have been reported from some ponds of Mathbaria of Pirojpur and Bakerganj of Barisal Districts in Bangladesh. %B Bangladesh Journal of Plant Taxonomy %V 14 %8 2007/12/02/ %@ 1028-2092 %G eng %U http://www.banglajol.info/bd/index.php/BJPT/article/viewArticle/528 %N 2 %R 10.3329/bjpt.v14i2.528 %0 Conference Paper %B Proceedings of the 8th annual international conference on Digital government research %D 2007 %T New techniques for ensuring the long term integrity of digital archives %A Song,S. %A JaJa, Joseph F. %X A large portion of the government, business, cultural, andscientific digital data being created today needs to be archived and preserved for future use of periods ranging from a few years to decades and sometimes centuries. A fundamental requirement of a long term archive is to ensure the integrity of its holdings. In this paper, we develop a new methodology to address the integrity of long term archives using rigorous cryptographic techniques. Our approach involves the generation of a small-size integrity token for each digital object to be archived, and some cryptographic summary information based on all the objects handled within a dynamic time period. We present a framework that enables the continuous auditing of the holdings of the archive depending on the policy set by the archive. Moreover, an independent auditor will be able to verify the integrity of every version of an archived digital object as well as link the current version to the original form of the object when it was ingested into the archive. We built a - prototype system that is completely independent of the archive’s underlying architecture, and tested it on large scale data. We include in this paper some preliminary results on the validation and performance of our prototype. %B Proceedings of the 8th annual international conference on Digital government research %P 57 - 65 %8 2007/// %G eng %0 Journal Article %J Eukaryotic Cell %D 2007 %T New Trypanosoma cruzi Repeated Element That Shows Site Specificity for Insertion %A Souza,Renata T. %A Santos,Marcia R. M. %A Lima,Fabio M. %A El‐Sayed, Najib M. %A Myler,Peter J. %A Ruiz,Jeronimo C. %A da Silveira,Jose Franco %X A new family of site-specific repeated elements identified in Trypanosoma cruzi, which we named TcTREZO, is described here. TcTREZO appears to be a composite repeated element, since three subregions may be defined within it on the basis of sequence similarities with other T. cruzi sequences. Analysis of the distribution of TcTREZO in the genome clearly indicates that it displays site specificity for insertion. Most TcTREZO elements are flanked by conserved sequences. There is a highly conserved 68-bp sequence at the 5' end of the element and a sequence domain of [~]500 bp without a well-defined borderline at the 3' end. Northern blot hybridization and reverse transcriptase PCR analyses showed that TcTREZO transcripts are expressed as oligo(A)-terminated transcripts whose length corresponds to the unit size of the element (1.6 kb). Transcripts of [~]0.2 kb derived from a small part of TcTREZO are also detected in steady-state RNA. TcTREZO transcripts are unspliced and not translated. The copy number of TcTREZO sequences was estimated to be [~]173 copies per haploid genome. TcTREZO appears to have been assembled by insertions of sequences into a progenitor element. Once associated with each other, these subunits were amplified as a new transposable element. TcTREZO shows site specificity for insertion, suggesting that a sequence-specific endonuclease could be responsible for its insertion at a unique site. %B Eukaryotic Cell %V 6 %P 1228 - 1238 %8 2007/07/01/ %G eng %U http://ec.asm.org/cgi/content/abstract/6/7/1228 %N 7 %R

10.1128/EC.00036-07

%0 Conference Paper %B Image Processing, 2007. ICIP 2007. IEEE International Conference on %D 2007 %T Noise Features for Image Tampering Detection and Steganalysis %A Gou,Hongmei %A Swaminathan,A. %A M. Wu %K analysis;cryptography;data %K analysis;image %K analysis;wavelet %K authenticity;hidden %K computing;statistical %K data;image %K denoising %K denoising;multimedia %K detection;low-cost %K editing %K encapsulation;feature %K extraction;image %K features;steganalysis;wavelet %K forensic %K forensics;neighborhood %K image %K NOISE %K operations;digital %K prediction;statistical %K softwares;multimedia %K tampering %K transforms; %X With increasing availability of low-cost image editing softwares, the authenticity of digital images can no longer be taken for granted. Digital images have also been used as cover data for transmitting secret information in the field of steganography. In this paper, we introduce a new set of features for multimedia forensics to determine if a digital image is an authentic camera output or if it has been tampered or embedded with hidden data. We perform such image forensic analysis employing three sets of statistical noise features, including those from denoising operations, wavelet analysis, and neighborhood prediction. Our experimental results demonstrate that the proposed method can effectively distinguish digital images from their tampered or stego versions. %B Image Processing, 2007. ICIP 2007. IEEE International Conference on %V 6 %P VI -97 -VI -100 - VI -97 -VI -100 %8 2007/10/16/19 %G eng %R 10.1109/ICIP.2007.4379530 %0 Journal Article %J Information Forensics and Security, IEEE Transactions on %D 2007 %T Nonintrusive component forensics of visual sensors using output images %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K ACQUISITION %K array;color %K authenticating %K component %K filter %K forensics;patent %K infringements;visual %K Interpolation %K manipulations;intellectual %K modules;content %K property %K property; %K protection;nonintrusive %K rights %K sensors;image %K sensors;industrial %K sources;color %X Rapid technology development and the widespread use of visual sensors have led to a number of new problems related to protecting intellectual property rights, handling patent infringements, authenticating acquisition sources, and identifying content manipulations. This paper introduces nonintrusive component forensics as a new methodology for the forensic analysis of visual sensing information, aiming to identify the algorithms and parameters employed inside various processing modules of a digital device by only using the device output data without breaking the device apart. We propose techniques to estimate the algorithms and parameters employed by important camera components, such as color filter array and color interpolation modules. The estimated interpolation coefficients provide useful features to construct an efficient camera identifier to determine the brand and model from which an image was captured. The results obtained from such component analysis are also useful to examine the similarities between the technologies employed by different camera models to identify potential infringement/licensing and to facilitate studies on technology evolution %B Information Forensics and Security, IEEE Transactions on %V 2 %P 91 - 106 %8 2007/03// %@ 1556-6013 %G eng %N 1 %R 10.1109/TIFS.2006.890307 %0 Conference Paper %B Proceedings from the Workshop on Metareasoning in Agent Based Systems at the Sixth International Joint Conference on Autonomous Agents and Multiagent Sytems %D 2007 %T Ontologies for reasoning about failures in AI systems %A Schmill,M. %A Josyula,D. %A Anderson,M. L %A Wilson,S. %A Oates,T. %A Perlis, Don %A Fults,S. %B Proceedings from the Workshop on Metareasoning in Agent Based Systems at the Sixth International Joint Conference on Autonomous Agents and Multiagent Sytems %8 2007/// %G eng %0 Journal Article %J Spatial data on the Web: modeling and management %D 2007 %T Out-of-core Multi-resolution Terrain Modeling %A Danovaro,E. %A De Floriani, Leila %A Puppo,E. %A Samet, Hanan %B Spatial data on the Web: modeling and management %P 43 - 43 %8 2007/// %G eng %0 Journal Article %J Computers and Software Engineering %D 2007 %T Password Changes: Empirical Results %A Michel Cukier %A Sharma,A. %X This paper focuses on a detailed analysis of oneaspect of password evolution based on empirical data: password changes. Passwords can be divided into weak and strong based on how easy it is to crack them. We present a model of password changes and analyze passwords collected during 21 months from a large network of an average of 770 users. The results include tracking the percentage of users with weak passwords over time and the percentage of users changing between weak and strong passwords. Based on the data analysis, the distribution of users switching between weak and strong passwords was characterized and two parameters of the model were estimated. %B Computers and Software Engineering %8 2007/// %G eng %U http://isastorganization.org/ISAST_CS_1_2007.pdf#page=10 %0 Conference Paper %B Proc. Workshop on Hot Topics in Networks (HotNets) %D 2007 %T PeerWise discovery and negotiation of faster paths %A Lumezanu,C. %A Levin,D. %A Spring, Neil %B Proc. Workshop on Hot Topics in Networks (HotNets) %8 2007/// %G eng %0 Journal Article %J Bulletin of the American Physical Society %D 2007 %T Plasma Turbulence Simulation and Visualization on Graphics Processors: Efficient Parallel Computing on the Desktop %A Stantchev,George %A Juba,Derek %A Dorland,William %A Varshney, Amitabh %X Direct numerical simulation (DNS) of turbulence is computationally very intensive and typically relies on some form of parallel processing. Spectral kernels used for spatial discretization are a common computational bottleneck on distributed memory architectures. One way to increase the efficiency of DNS algorithms is to parallelize spectral kernels using tightly-coupled SPMD multiprocessor hardware architecture with minimal inter-processor communication latency. In this poster we present techniques to take advantage of the recent programmable interfaces for modern Graphics Processing Units (GPUs) to carefully map DNS computations to GPU architectures that are characterized by a very high memory bandwidth and hundreds of SPMD processors. We compare and contrast the performance of our parallel algorithm on a modern GPU versus a CPU implementation of several turbulence simulation codes. We also demonstrate a prototype of a scalable computational steering framework based on turbulence simulation and visualization coupling on the GPU. %B Bulletin of the American Physical Society %V Volume 52, Number 11 %8 2007/11/12/ %G eng %U http://meetings.aps.org/Meeting/DPP07/Event/70114 %0 Conference Paper %B Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series %D 2007 %T Plasmonics and the parallel programming problem %A Vishkin, Uzi %A Smolyaninov,I. %A Davis,C. %X While many parallel computers have been built, it has generally been too difficult to program them. Now, all computersare effectively becoming parallel machines. Biannual doubling in the number of cores on a single chip, or faster, over the coming decade is planned by most computer vendors. Thus, the parallel programming problem is becoming more critical. The only known solution to the parallel programming problem in the theory of computer science is through a parallel algorithmic theory called PRAM. Unfortunately, some of the PRAM theory assumptions regarding the bandwidth between processors and memories did not properly reflect a parallel computer that could be built in previous decades. Reaching memories, or other processors in a multi-processor organization, required off-chip connections through pins on the boundary of each electric chip. Using the number of transistors that is becoming available on chip, on-chip architectures that adequately support the PRAM are becoming possible. However, the bandwidth of off-chip connections remains insufficient and the latency remains too high. This creates a bottleneck at the boundary of the chip for a PRAM-On-Chip architecture. This also prevents scalability to larger “supercomputing” organizations spanning across many processing chips that can handle massive amounts of data. Instead of connections through pins and wires, power-efficient CMOS-compatible on-chip conversion to plasmonic nanowaveguides is introduced for improved latency and bandwidth. Proper incorporation of our ideas offer exciting avenues to resolving the parallel programming problem, and an alternative way for building faster, more useable and much more compact supercomputers. %B Society of Photo-Optical Instrumentation Engineers (SPIE) Conference Series %V 6477 %P 19 - 19 %8 2007/// %G eng %0 Conference Paper %D 2007 %T Portcullis: protecting connection setup from denial-of-capability attacks %A Parno, Bryan %A Wendlandt, Dan %A Elaine Shi %A Perrig, Adrian %A Maggs, Bruce %A Hu, Yih-Chun %K network capability %K per-computation fairness %X Systems using capabilities to provide preferential service to selected flows have been proposed as a defense against large-scale network denial-of-service attacks. While these systems offer strong protection for established network flows, the Denial-of-Capability (DoC) attack, which prevents new capability-setup packets from reaching the destination, limits the value of these systems. Portcullis mitigates DoC attacks by allocating scarce link bandwidth for connection establishment packets based on per-computation fairness. We prove that a legitimate sender can establish a capability with high probability regardless of an attacker's resources or strategy and that no system can improve on our guarantee. We simulate full and partial deployments of Portcullis on an Internet-scale topology to confirm our theoretical results and demonstrate the substantial benefits of using per-computation fairness. %S SIGCOMM '07 %I ACM %P 289 - 300 %8 2007 %@ 978-1-59593-713-1 %G eng %U http://doi.acm.org/10.1145/1282380.1282413 %0 Journal Article %J Computational Statistics & Data Analysis %D 2007 %T A practical approximation algorithm for the LMS line estimator %A Mount, Dave %A Netanyahu,Nathan S. %A Romanik,Kathleen %A Silverman,Ruth %A Wu,Angela Y. %K Approximation algorithms %K least median-of-squares regression %K line arrangements %K line fitting %K randomized algorithms %K robust estimation %X The problem of fitting a straight line to a finite collection of points in the plane is an important problem in statistical estimation. Robust estimators are widely used because of their lack of sensitivity to outlying data points. The least median-of-squares (LMS) regression line estimator is among the best known robust estimators. Given a set of n points in the plane, it is defined to be the line that minimizes the median squared residual or, more generally, the line that minimizes the residual of any given quantile q, where 0 < q ⩽ 1 . This problem is equivalent to finding the strip defined by two parallel lines of minimum vertical separation that encloses at least half of the points.The best known exact algorithm for this problem runs in O ( n 2 ) time. We consider two types of approximations, a residual approximation, which approximates the vertical height of the strip to within a given error bound ε r ⩾ 0 , and a quantile approximation, which approximates the fraction of points that lie within the strip to within a given error bound ε q ⩾ 0 . We present two randomized approximation algorithms for the LMS line estimator. The first is a conceptually simple quantile approximation algorithm, which given fixed q and ε q > 0 runs in O ( n log n ) time. The second is a practical algorithm, which can solve both types of approximation problems or be used as an exact algorithm. We prove that when used as a quantile approximation, this algorithm's expected running time is O ( n log 2 n ) . We present empirical evidence that the latter algorithm is quite efficient for a wide variety of input distributions, even when used as an exact algorithm. %B Computational Statistics & Data Analysis %V 51 %P 2461 - 2486 %8 2007/02/01/ %@ 0167-9473 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167947306002921 %N 5 %R 10.1016/j.csda.2006.08.033 %0 Journal Article %J Journal of Parallel and Distributed Computing %D 2007 %T Principles for designing data-/compute-intensive distributed applications and middleware systems for heterogeneous environments %A Kim,Jik-Soo %A Andrade,Henrique %A Sussman, Alan %K Computational science applications %K Data-/compute-intensive applications %K Heterogeneous environments %K Middleware systems %X The nature of distributed systems is constantly and steadily changing as the hardware and software landscape evolves. Porting applications and adapting existing middleware systems to ever changing computational platforms has become increasingly complex and expensive. Therefore, the design of applications, as well as the design of next generation middleware systems, must follow a set of guiding principles in order to insure long-term “survivability” without costly re-engineering. From our practical experience, the key determinants to success in this endeavor are adherence to the following principles: (1) Design for change; (2) Provide for storage subsystem I/O coordination; (3) Employ workload partitioning and load balancing techniques; (4) Employ caching; (5) Schedule the workload; and (6) Understand the workload. In order to support these principles, we have collected extensive experimental results comparing three middleware systems targeted at data- and compute-intensive applications implemented by our research group during the course of the last decade, on a single data- and compute-intensive application. The main contribution of this work is the analysis of a level playing field, where we discuss and quantify how adherence to these guiding principles impacts overall system throughput and response time. %B Journal of Parallel and Distributed Computing %V 67 %P 755 - 771 %8 2007/07// %@ 0743-7315 %G eng %U http://www.sciencedirect.com/science/article/pii/S0743731507000603 %N 7 %R 10.1016/j.jpdc.2007.04.006 %0 Conference Paper %D 2007 %T Probabilistic go theories %A Parker,A. %A Yaman,F. %A Nau, Dana S. %A V.S. Subrahmanian %X There are numerous cases where we need to rea- son about vehicles whose intentions and itineraries are not known in advance to us. For example, Coast Guard agents tracking boats don’t always know where they are headed. Likewise, in drug en- forcement applications, it is not always clear where drug-carrying airplanes (which do often show up on radar) are headed, and how legitimate planes with an approved flight manifest can avoid them. Likewise, traffic planners may want to understand how many vehicles will be on a given road at a given time. Past work on reasoning about vehi- cles (such as the “logic of motion” by Yaman et. al. [Yaman et al., 2004]) only deals with vehicles whose plans are known in advance and don’t cap- ture such situations. In this paper, we develop a for- mal probabilistic extension of their work and show that it captures both vehicles whose itineraries are known, and those whose itineraries are not known. We show how to correctly answer certain queries against a set of statements about such vehicles. A prototype implementation shows our system to work efficiently in practice. %8 2007/// %G eng %U http://www.aaai.org/Papers/IJCAI/2007/IJCAI07-079.pdf %0 Conference Paper %B Proceedings of the 33rd international conference on Very large data bases %D 2007 %T Probabilistic graphical models and their role in databases %A Deshpande, Amol %A Sarawagi,Sunita %X Probabilistic graphical models provide a framework for compact representation and efficient reasoning about the joint probability distribution of several interdependent variables. This is a classical topic with roots in statistical physics. In recent years, spurred by several applications in unstructured data integration, sensor networks, image processing, bio-informatics, and code design, the topic has received renewed interest in the machine learning, data mining, and database communities. Techniques from graphical models have also been applied to many topics directly of interest to the database community including information extraction, sensor data analysis, imprecise data representation and querying, selectivity estimation for query optimization, and data privacy. As database research continues to expand beyond the confines of traditional enterprise domains, we expect both the need and applicability of probabilistic graphical models to increase dramatically over the next few years. With this tutorial, we are aiming to provide a foundational overview of probabilistic graphical models to the database community, accompanied by a brief overview of some of the recent research literature on the role of graphical models in databases. %B Proceedings of the 33rd international conference on Very large data bases %S VLDB '07 %I VLDB Endowment %P 1435 - 1436 %8 2007/// %@ 978-1-59593-649-3 %G eng %U http://dl.acm.org/citation.cfm?id=1325851.1326038 %0 Journal Article %J ACM Transactions on Computational Logic (TOCL) %D 2007 %T Probabilistic interval XML %A Hung,Edward %A Getoor, Lise %A V.S. Subrahmanian %K Semistructured Databases %K XML %X Interest in XML databases has been expanding rapidly over the last few years. In this paper, we study the problem of incorporating probabilistic information into XML databases. We propose the Probabilistic Interval XML (PIXML for short) data model in this paper. Using this data model, users can express probabilistic information within XML markups. In addition, we provide two alternative formal model-theoretic semantics for PIXML data. The first semantics is a “global” semantics which is relatively intuitive, but is not directly amenable to computation. The second semantics is a “local” semantics which supports efficient computation. We prove several correspondence results between the two semantics. To our knowledge, this is the first formal model theoretic semantics for probabilistic interval XML. We then provide an operational semantics that may be used to compute answers to queries and that is correct for a large class of probabilistic instances. %B ACM Transactions on Computational Logic (TOCL) %V 8 %8 2007/08// %@ 1529-3785 %G eng %U http://doi.acm.org/10.1145/1276920.1276926 %N 4 %R 10.1145/1276920.1276926 %0 Book Section %B Empirical Software EngineeringEmpirical Software Engineering %D 2007 %T Protocols in the use of empirical software engineering artifacts %A Basili, Victor R. %A Zelkowitz, Marvin V %A Sjøberg,D. I.K %A Johnson,P. %A Cowling,A. J %X If empirical software engineering is to grow as a valid scientific endeavor, the ability to acquire, use, share, and compare data collected from a variety of sources must be encouraged. This is necessary to validate the formal models being developed within computer science. However, within the empirical software engineering community this has not been easily accomplished. This paper analyses experiences from a number of projects, and defines the issues, which include the following: (1) How should data, testbeds, and artifacts be shared? (2) What limits should be placed on who can use them and how? How does one limit potential misuse? (3) What is the appropriate way to credit the organization and individual that spent the effort collecting the data, developing the testbed, and building the artifact? (4) Once shared, who owns the evolved asset? As a solution to these issues, the paper proposes a framework for an empirical software engineering artifact agreement. Such an agreement is intended to address the needs for both creator and user of such artifacts and should foster a market in making available and using such artifacts. If this framework for sharing software engineering artifacts is commonly accepted, it should encourage artifact owners to make the artifacts accessible to others (gaining credit is more likely and misuse is less likely). It may be easier for other researchers to request artifacts since there will be a well-defined protocol for how to deal with relevant matters. %B Empirical Software EngineeringEmpirical Software Engineering %I Springer %V 12 %P 107 - 119 %8 2007/// %G eng %0 Journal Article %J Proc. DialM-POMC Workshop on Foundations of Mobile Computing %D 2007 %T Provable algorithms for joint optimization of transport, routing and MAC layers in wireless ad hoc networks %A Kumar,V. S.A %A Marathe,M. V %A Parthasarathy,S. %A Srinivasan, Aravind %B Proc. DialM-POMC Workshop on Foundations of Mobile Computing %8 2007/// %G eng %0 Journal Article %J Information Processing & Management %D 2007 %T QCS: A system for querying, clustering and summarizing documents %A Dunlavy,Daniel M. %A O’Leary,Dianne P. %A Conroy,John M. %A Schlesinger,Judith D. %K clustering %K Information retrieval %K latent semantic indexing %K Sentence trimming %K Summarization %K TEXT PROCESSING %X Information retrieval systems consist of many complicated components. Research and development of such systems is often hampered by the difficulty in evaluating how each particular component would behave across multiple systems. We present a novel integrated information retrieval system—the Query, Cluster, Summarize (QCS) system—which is portable, modular, and permits experimentation with different instantiations of each of the constituent text analysis components. Most importantly, the combination of the three types of methods in the QCS design improves retrievals by providing users more focused information organized by topic.We demonstrate the improved performance by a series of experiments using standard test sets from the Document Understanding Conferences (DUC) as measured by the best known automatic metric for summarization system evaluation, ROUGE. Although the DUC data and evaluations were originally designed to test multidocument summarization, we developed a framework to extend it to the task of evaluation for each of the three components: query, clustering, and summarization. Under this framework, we then demonstrate that the QCS system (end-to-end) achieves performance as good as or better than the best summarization engines. Given a query, QCS retrieves relevant documents, separates the retrieved documents into topic clusters, and creates a single summary for each cluster. In the current implementation, Latent Semantic Indexing is used for retrieval, generalized spherical k-means is used for the document clustering, and a method coupling sentence “trimming” and a hidden Markov model, followed by a pivoted QR decomposition, is used to create a single extract summary for each cluster. The user interface is designed to provide access to detailed information in a compact and useful format. Our system demonstrates the feasibility of assembling an effective IR system from existing software libraries, the usefulness of the modularity of the design, and the value of this particular combination of modules. %B Information Processing & Management %V 43 %P 1588 - 1605 %8 2007/11// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306457307000246 %N 6 %R 10.1016/j.ipm.2007.01.003 %0 Book Section %B Stochastic Algorithms: Foundations and ApplicationsStochastic Algorithms: Foundations and Applications %D 2007 %T Randomized Algorithms and Probabilistic Analysis in Wireless Networking %A Srinivasan, Aravind %E Hromkovic,Juraj %E Královic,Richard %E Nunkesser,Marc %E Widmayer,Peter %X Devices connected wirelessly, in various forms including computers, hand-held devices, ad hoc networks, and embedded systems, are expected to become ubiquitous all around us. Wireless networks pose interesting new challenges, some of which do not arise in standard (wired) networks. This survey discusses some key probabilistic notions – both randomized algorithms and probabilistic analysis – in wireless networking. %B Stochastic Algorithms: Foundations and ApplicationsStochastic Algorithms: Foundations and Applications %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4665 %P 54 - 57 %8 2007/// %@ 978-3-540-74870-0 %G eng %U http://dx.doi.org/10.1007/978-3-540-74871-7_5 %0 Journal Article %J Software Engineering, IEEE Transactions on %D 2007 %T Reliable Effects Screening: A Distributed Continuous Quality Assurance Process for Monitoring Performance Degradation in Evolving Software Systems %A Yilmaz,C. %A Porter, Adam %A Krishna,A. S %A Memon, Atif M. %A Schmidt,D. C %A Gokhale,A.S. %A Natarajan,B. %K configuration subset %K distributed continuous quality assurance process %K evolving software systems %K in house testing %K main effects screening %K performance bottlenecks %K performance degradation monitoring %K performance intensive software systems %K process configuration %K process execution %K program testing %K regression testing %K reliable effects screening %K software benchmarks %K Software performance %K software performance evaluation %K Software quality %K software reliability %K tool support %X Developers of highly configurable performance-intensive software systems often use in-house performance-oriented "regression testing" to ensure that their modifications do not adversely affect their software's performance across its large configuration space. Unfortunately, time and resource constraints can limit in-house testing to a relatively small number of possible configurations, followed by unreliable extrapolation from these results to the entire configuration space. As a result, many performance bottlenecks escape detection until systems are fielded. In our earlier work, we improved the situation outlined above by developing an initial quality assurance process called "main effects screening". This process 1) executes formally designed experiments to identify an appropriate subset of configurations on which to base the performance-oriented regression testing, 2) executes benchmarks on this subset whenever the software changes, and 3) provides tool support for executing these actions on in-the-field and in-house computing resources. Our initial process had several limitations, however, since it was manually configured (which was tedious and error-prone) and relied on strong and untested assumptions for its accuracy (which made its use unacceptably risky in practice). This paper presents a new quality assurance process called "reliable effects screening" that provides three significant improvements to our earlier work. First, it allows developers to economically verify key assumptions during process execution. Second, it integrates several model-driven engineering tools to make process configuration and execution much easier and less error prone. Third, we evaluate this process via several feasibility studies of three large, widely used performance-intensive software frameworks. Our results indicate that reliable effects screening can detect performance degradation in large-scale systems more reliably and with significantly less resources than conventional t- echniques %B Software Engineering, IEEE Transactions on %V 33 %P 124 - 141 %8 2007/02// %@ 0098-5589 %G eng %N 2 %R 10.1109/TSE.2007.20 %0 Conference Paper %B Data Engineering, 2007. ICDE 2007. IEEE 23rd International Conference on %D 2007 %T Representing and Querying Correlated Tuples in Probabilistic Databases %A Sen,Prithviraj %A Deshpande, Amol %X Probabilistic databases have received considerable attention recently due to the need for storing uncertain data produced by many real world applications. The widespread use of probabilistic databases is hampered by two limitations: (1) current probabilistic databases make simplistic assumptions about the data (e.g., complete independence among tuples) that make it difficult to use them in applications that naturally produce correlated data, and (2) most probabilistic databases can only answer a restricted subset of the queries that can be expressed using traditional query languages. We address both these limitations by proposing a framework that can represent not only probabilistic tuples, but also correlations that may be present among them. Our proposed framework naturally lends itself to the possible world semantics thus preserving the precise query semantics extant in current probabilistic databases. We develop an efficient strategy for query evaluation over such probabilistic databases by casting the query processing problem as an inference problem in an appropriately constructed probabilistic graphical model. We present several optimizations specific to probabilistic databases that enable efficient query evaluation. We validate our approach by presenting an experimental evaluation that illustrates the effectiveness of our techniques at answering various queries using real and synthetic datasets. %B Data Engineering, 2007. ICDE 2007. IEEE 23rd International Conference on %I IEEE %P 596 - 605 %8 2007/// %@ 1-4244-0802-4 %G eng %U zotero://attachment/12299/ %R 10.1109/ICDE.2007.367905 %0 Conference Paper %B Seventh IEEE International Conference on Data Mining Workshops, 2007. ICDM Workshops 2007 %D 2007 %T Representing Tuple and Attribute Uncertainty in Probabilistic Databases %A Sen,P. %A Deshpande, Amol %A Getoor, Lise %K attribute uncertainty %K Computer science %K Conferences %K correlation structures %K data mining %K Data models %K database management systems %K Educational institutions %K inference mechanisms %K noisy data sources %K probabilistic database %K probabilistic inference %K Probability distribution %K Query processing %K Relational databases %K Sensor phenomena and characterization %K tuple representation %K Uncertainty %K uncertainty handling %X There has been a recent surge in work in probabilistic databases, propelled in large part by the huge increase in noisy data sources-sensor data, experimental data, data from uncurated sources, and many others. There is a growing need to be able to flexibly represent the uncertainties in the data, and to efficiently query the data. Building on existing probabilistic database work, we present a unifying framework which allows a flexible representation of correlated tuple and attribute level uncertainties. An important capability of our representation is the ability to represent shared correlation structures in the data. We provide motivating examples to illustrate when such shared correlation structures are likely to exist. Representing shared correlations structures allows the use of sophisticated inference techniques based on lifted probabilistic inference that, in turn, allows us to achieve significant speedups while computing probabilities for results of user-submitted queries. %B Seventh IEEE International Conference on Data Mining Workshops, 2007. ICDM Workshops 2007 %I IEEE %P 507 - 512 %8 2007/10/28/31 %@ 978-0-7695-3019-2 %G eng %R 10.1109/ICDMW.2007.11 %0 Report %D 2007 %T A Residual Inverse Power Method %A Stewart, G.W. %K Technical Report %X The inverse power method involves solving shifted equations of theform $(A -\sigma I)v = u$. This paper describes a variant method in which shifted equations may be solved to a fixed reduced accuracy without affecting convergence. The idea is to alter the right-hand side to produce a correction step to be added to the current approximations. The digits of this step divide into two parts: leading digits that correct the solution and trailing garbage. Hence the step can be be evaluated to a reduced accuracy corresponding to the correcting digits. The cost is an additional multiplication by $A$ at each step to generate the right-hand side. Analysis and experiments show that the method is suitable for normal and mildly nonnormal problems. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-09 %8 2007/02// %G eng %U http://drum.lib.umd.edu/handle/1903/4260 %0 Conference Paper %B INFOCOM 2007. 26th IEEE International Conference on Computer Communications. IEEE %D 2007 %T Robust Routing with Unknown Traffic Matrices %A Tabatabaee,V. %A Kashyap,A. %A Bhattacharjee, Bobby %A La,R.J. %A Shayman,M.A. %K engineering;linear %K inequalities;maximum %K intradomain %K linear %K link %K matrices;linear %K matrix %K network %K nodes;polynomial %K of %K point %K presence %K problem;traffic %K Programming %K programming;semiinfinite %K programming;telecommunication %K region;unknown %K routing;telecommunication %K size %K traffic %K traffic; %K utilization;network %X In this paper, we present an algorithm for intra-domain traffic engineering. We assume that the traffic matrix, which specifies traffic load between every source-destination pair in the network, is unknown and varies with time, but that always lies inside an explicitly defined region. Our goal is to compute a fixed robust routing with best worst case performance for all traffic matrices inside the bounding region. We formulate this problem as a semi-infinite programming problem. Then, we focus on a special case with practical merits, where (1) the traffic matrix region is assumed to be a polytope specified by a finite set of linear inequalities, and (2) our objective is to find the routing that minimizes the maximum link utilization. Under these assumptions, the problem can be formulated as a polynomial size linear programming (LP) problem with finite number of constraints. We further consider two specific set of constraints for the traffic matrix region. The first set is based on the hose model and limits the total traffic rate of network point of presence (PoP) nodes. The second set is based on the pipe model and limits the traffic between source-destination pairs. We study the effectiveness of each set of constraints using extensive simulations. %B INFOCOM 2007. 26th IEEE International Conference on Computer Communications. IEEE %P 2436 - 2440 %8 2007/05// %G eng %R 10.1109/INFCOM.2007.296 %0 Journal Article %J Proc. SPIE, Electronic Imaging, Security, Steganography, and Watermarking of Multimedia Contents IX %D 2007 %T Robust scanner identification based on noise features %A Gou,H. %A Swaminathan,A. %A Wu,M. %X A large portion of digital image data available today is acquired using digital cameras or scanners. While camerasallow digital reproduction of natural scenes, scanners are often used to capture hardcopy art in more controlled scenarios. This paper proposes a new technique for non-intrusive scanner model identification, which can be further extended to perform tampering detection on scanned images. Using only scanned image samples that contain arbitrary content, we construct a robust scanner identifier to determine the brand/model of the scanner used to capture each scanned image. The proposed scanner identifier is based on statistical features of scanning noise. We first analyze scanning noise from several angles, including through image de-noising, wavelet analysis, and neighborhood prediction, and then obtain statistical features from each characterization. Experimental results demonstrate that the proposed method can effectively identify the correct scanner brands/models with high accuracy. %B Proc. SPIE, Electronic Imaging, Security, Steganography, and Watermarking of Multimedia Contents IX %V 6505 %P 0S–0T - 0S–0T %8 2007/// %G eng %0 Conference Paper %B Computer Vision, 2007. ICCV 2007. IEEE 11th International Conference on %D 2007 %T Robust Visual Tracking Using the Time-Reversibility Constraint %A Wu,Hao %A Chellapa, Rama %A Sankaranarayanan,A. C %A Zhou,S. K %K backward %K constraint;video %K criterion;state %K forward %K frame %K KLT %K processing; %K processing;video %K processing;visual %K signal %K tracker;minimization %K tracking;minimisation;video %K vectors;time-reversibility %X Visual tracking is a very important front-end to many vision applications. We present a new framework for robust visual tracking in this paper. Instead of just looking forward in the time domain, we incorporate both forward and backward processing of video frames using a novel time-reversibility constraint. This leads to a new minimization criterion that combines the forward and backward similarity functions and the distances of the state vectors between the forward and backward states of the tracker. The new framework reduces the possibility of the tracker getting stuck in local minima and significantly improves the tracking robustness and accuracy. Our approach is general enough to be incorporated into most of the current tracking algorithms. We illustrate the improvements due to the proposed approach for the popular KLT tracker and a search based tracker. The experimental results show that the improved KLT tracker significantly outperforms the original KLT tracker. The time-reversibility constraint used for tracking can be incorporated to improve the performance of optical flow, mean shift tracking and other algorithms. %B Computer Vision, 2007. ICCV 2007. IEEE 11th International Conference on %P 1 - 8 %8 2007/10// %G eng %R 10.1109/ICCV.2007.4408956 %0 Journal Article %J Proceedings of 4th USENIX Symposium on Networked Systems Design & Implementation %D 2007 %T SAAR: A shared control plane for overlay multicast %A Nandi,A. %A Ganjam,A. %A Druschel,P. %A Ng,T. S.E %A Stoica,I. %A Zhang,H. %A Bhattacharjee, Bobby %X Many cooperative overlay multicast systems of diverse designs have been implemented and deployed. In this paper, we explore a new architecture for overlay multicast: we factor out the control plane into a separate overlay that provides a single primitive: a configurable anycast for peer selection. This separation of control and data overlays has several advantages. Data overlays can be optimized for efficient content delivery, while the control overlay can be optimized for flexible and efficient peer selection. Several data channels can share a control plane for fast switching among content channels, which is particularly important for IPTV. And, the control overlay can be reused in multicast systems with different data plane organizations.We designed and evaluated a decentralized control overlay for endsystem multicast. The overlay proactively aggregates system state and implements a powerful anycast primitive for peer selection. We demonstrate that SAAR's efficiency in locating peers reduces channel switching time, improves the quality of content delivery, and reduces overhead, even under dynamic conditions and at scale. An experimental evaluation demonstrates that the system can efficiently support single-tree, multi-tree and block-based multicast systems. %B Proceedings of 4th USENIX Symposium on Networked Systems Design & Implementation %P 57 - 72 %8 2007/// %G eng %0 Book %D 2007 %T Scalable Uncertainty Management: First International Conference, SUM 2007, Washington, DC, USA, October 10-12, 2007 : Proceedings %A Prade,Henri %A V.S. Subrahmanian %K Artificial intelligence %K Computers / Information Theory %K Computers / Intelligence (AI) & Semantics %K Electronic books %K Mathematics / Logic %K Uncertainty (Information theory) %X This book constitutes the refereed proceedings of the First International Conference on Scalable Uncertainty Management, SUM 2007, held in Washington, DC, USA, in Oktober 2007.The 20 revised full papers presented were carefully reviewed and selected from numerous submissions for inclusion in the book. The papers address artificial intelligence researchers, database researchers, and practitioners to demonstrate theoretical techniques required to manage the uncertainty that arises in large scale real world applications. The papers deal with the following topics: uncertainty models, probabilistic logics, fuzzy logics, and annotated logics, inconsistency logics, database algebras and calculi, scalable database systems, spatial, temporal, mobile and multimedia databases, as well as implementations, and applications. %I Springer %8 2007/11/14/ %@ 9783540754077 %G eng %0 Conference Paper %B Proceedings of the nineteenth annual ACM symposium on Parallel algorithms and architectures %D 2007 %T Scheduling to minimize gaps and power consumption %A Demaine,E. D %A Ghodsi,M. %A Hajiaghayi, Mohammad T. %A Sayedi-Roshkhar,A. S %A Zadimoghaddam,M. %B Proceedings of the nineteenth annual ACM symposium on Parallel algorithms and architectures %P 46 - 54 %8 2007/// %G eng %0 Journal Article %J Technical Reports of the Computer Science Department, CS-TR-4884 %D 2007 %T A Secure DHT via the Pigeonhole Principle %A Baden,Randy %A Bender,Adam %A Levin,Dave %A Sherwood,Rob %A Spring, Neil %A Bhattacharjee, Bobby %K Technical Report %X The standard Byzantine attack model assumes no more than some fixedfraction of the participants are faulty. This assumption does not accurately apply to peer-to-peer settings, where Sybil attacks and botnets are realistic threats. We propose an attack model that permits an arbitrary number of malicious nodes under the assumption that each node can be classified based on some of its attributes, such as autonomous system number or operating system, and that the number of classes with malicious nodes is bounded (e.g., an attacker may exploit at most a few operating systems at a time). In this model, we present a secure DHT, evilTwin, which replaces a single, large DHT with sufficiently many smaller instances such that it is impossible for an adversary to corrupt every instance. Our system ensures high availability and low-latency lookups, is easy to implement, does not require a complex Byzantine agreement protocol, and its proof of security is a straightforward application of the pigeonhole principle. The cost of security comes in the form of increased storage and bandwidth overhead; we show how to reduce these costs by replicating data and adaptively querying participants who historically perform well. We use implementation and simulation to show that evilTwin imposes a relatively small additional cost compared to conventional DHTs. %B Technical Reports of the Computer Science Department, CS-TR-4884 %8 2007/09/24/ %G eng %U http://drum.lib.umd.edu/handle/1903/7136 %0 Journal Article %J Software Engineering, IEEE Transactions on %D 2007 %T Semantics-Preserving Design of Embedded Control Software from Synchronous Models %A Mangeruca,L. %A Baleani,M. %A Ferrari,A. %A Sangiovanni-Vincentelli,A. %K embedded control software %K embedded systems %K multiprocessing programs %K multiprocessor %K multitasking %K semantics-preserving design %K software engineering %K software implementation %K systems analysis %X The design of embedded controllers is experiencing a growth in complexity as embedded systems increase their functionality while they become ubiquitous in electronic appliances, cars, airplanes, etc. As requirements become more challenging, mathematical models gain importance for mastering complexity. Among the different computational models proposed, synchronous models have proved to be the most widely used for control dominated applications. While synchronous models simplify the way of dealing with concurrency by decoupling functional and timing aspects, their software implementation on multitasking and multiprocessor platforms is far from straightforward, because of the asynchronous nature of most industrial software platforms. Known solutions in the literature either restrict the solution space or focus on special cases. We present a method for preserving the synchronous semantics through buffer-based intertask communication mechanisms, grounded on an abstraction of the target platform. This allows us to deal with any task set and, most importantly, being independent of the implementation, to explore the design space effectively. %B Software Engineering, IEEE Transactions on %V 33 %P 497 - 509 %8 2007/08// %@ 0098-5589 %G eng %N 8 %R 10.1109/TSE.2007.70718 %0 Journal Article %J Interacting with Computers %D 2007 %T Semi-automatic photo annotation strategies using event based clustering and clothing based person recognition %A Suh,B. %A Bederson, Benjamin B. %B Interacting with Computers %V 19 %P 524 - 544 %8 2007/// %G eng %N 4 %0 Journal Article %J Numerische Mathematik %D 2007 %T On Sensitivity of Gauss-Christoffel Quadrature %A O'Leary, Dianne P. %A Strakoš,Zdeněk %A Tichý,Petr %X In numerical computations the question how much does a function change under perturbations of its arguments is of central importance. In this work, we investigate sensitivity of Gauss–Christoffel quadrature with respect to small perturbations of the distribution function. In numerical quadrature, a definite integral is approximated by a finite sum of functional values evaluated at given quadrature nodes and multiplied by given weights. Consider a sufficiently smooth integrated function uncorrelated with the perturbation of the distribution function. Then it seems natural that given the same number of function evaluations, the difference between the quadrature approximations is of the same order as the difference between the (original and perturbed) approximated integrals. That is perhaps one of the reasons why, to our knowledge, the sensitivity question has not been formulated and addressed in the literature, though several other sensitivity problems, motivated, in particular, by computation of the quadrature nodes and weights from moments, have been thoroughly studied by many authors. We survey existing particular results and show that even a small perturbation of a distribution function can cause large differences in Gauss–Christoffel quadrature estimates. We then discuss conditions under which the Gauss–Christoffel quadrature is insensitive under perturbation of the distribution function, present illustrative examples, and relate our observations to known conjectures on some sensitivity problems. %B Numerische Mathematik %V 107 %P 147 - 174 %8 2007/// %G eng %U doi:10.1007/s00211-007-0078-xdoi:10.1007/s00211-007-0078-x %0 Conference Paper %B Proceedings of the International Conference on Weblogs and Social Media (ICWSM) %D 2007 %T Sentiment analysis: Adjectives and adverbs are better than adjectives alone %A Benamara,F. %A Cesarano,C. %A Picariello, A. %A Reforgiato,D. %A V.S. Subrahmanian %X Most past work on determining the strength of subjective expres-sions within a sentence or a document use specific parts of speech such as adjectives, verbs and nouns. To date, there is almost no work on the use of adverbs in sentiment analysis, nor has there been any work on the use of adverb-adjective combinations (AACs). We propose an AAC-based sentiment analysis technique that uses a lin- guistic analysis of adverbs of degree. We define a set of general axioms (based on a classification of adverbs of degree into five cat- egories) that all adverb scoring techniques must satisfy. Instead of aggregating scores of both adverbs and adjectives using simple scor- ing functions, we propose an axiomatic treatment of AACs based on the linguistic classification of adverbs. Three specific AAC scor- ing methods that satisfy the axioms are presented. We describe the results of experiments on an annotated set of 200 news articles (an- notated by 10 students) and compare our algorithms with some exist- ing sentiment analysis algorithms. We show that our results lead to higher accuracy based on Pearson correlation with human subjects. %B Proceedings of the International Conference on Weblogs and Social Media (ICWSM) %8 2007/// %G eng %0 Conference Paper %B Statistical Signal Processing, 2007. SSP '07. IEEE/SP 14th Workshop on %D 2007 %T A Signal Processing and Randomization Perspective of Robust and Secure Image Hashing %A M. Wu %A Mao,Yinian %A Swaminathan,Ashwin %X An image hash is a content-based compact representation of an image and finds applications in image authentication, watermarking, and image similarity comparison. This paper reviews representative approaches developed by the research community in constructing robust and secure image hash. A unified framework and case studies on representative image hash schemes are presented to demonstrate important roles of signal processing and controlled randomization in constructing and evaluating robust and secure hashing for visual data. %B Statistical Signal Processing, 2007. SSP '07. IEEE/SP 14th Workshop on %P 166 - 170 %8 2007/08// %G eng %R 10.1109/SSP.2007.4301240 %0 Conference Paper %B Information Visualization, 2007. IV '07. 11th International Conference %D 2007 %T Similarity-Based Forecasting with Simultaneous Previews: A River Plot Interface for Time Series Forecasting %A Buono,P. %A Plaisant, Catherine %A Simeone,A. %A Aris,A. %A Shneiderman, Ben %A Shmueli,G. %A Jank,W. %K data driven forecasting method %K data visualisation %K Data visualization %K Economic forecasting %K forecasting preview interface %K Graphical user interfaces %K historical time series dataset %K Laboratories %K new stock offerings %K partial time series %K pattern matching %K pattern matching search %K Predictive models %K river plot interface %K Rivers %K similarity-based forecasting %K Smoothing methods %K Technological innovation %K Testing %K time series %K time series forecasting %K Weather forecasting %X Time-series forecasting has a large number of applications. Users with a partial time series for auctions, new stock offerings, or industrial processes desire estimates of the future behavior. We present a data driven forecasting method and interface called similarity-based forecasting (SBF). A pattern matching search in an historical time series dataset produces a subset of curves similar to the partial time series. The forecast is displayed graphically as a river plot showing statistical information about the SBF subset. A forecasting preview interface allows users to interactively explore alternative pattern matching parameters and see multiple forecasts simultaneously. User testing with 8 users demonstrated advantages and led to improvements. %B Information Visualization, 2007. IV '07. 11th International Conference %I IEEE %P 191 - 196 %8 2007/07/04/6 %@ 0-7695-2900-3 %G eng %R 10.1109/IV.2007.101 %0 Journal Article %J Software Engineering, IEEE Transactions on %D 2007 %T Skoll: A Process and Infrastructure for Distributed Continuous Quality Assurance %A Porter, Adam %A Yilmaz,C. %A Memon, Atif M. %A Schmidt,D. C %A Natarajan,B. %K ACE+TAO communication software package %K component assembly %K component deployment %K distributed continuous quality assurance %K distributed development teams %K distributed processing %K end-user customization %K flexible product design %K incremental development %K object-oriented programming %K Skoll %K software engineering %K Software quality %K systems analysis %X Software engineers increasingly emphasize agility and flexibility in their designs and development approaches. They increasingly use distributed development teams, rely on component assembly and deployment rather than green field code writing, rapidly evolve the system through incremental development and frequent updating, and use flexible product designs supporting extensive end-user customization. While agility and flexibility have many benefits, they also create an enormous number of potential system configurations built from rapidly changing component implementations. Since today's quality assurance (QA) techniques do not scale to handle highly configurable systems, we are developing and validating novel software QA processes and tools that leverage the extensive computing resources of user and developer communities in a distributed, continuous manner to improve software quality significantly. This paper provides several contributions to the study of distributed, continuous QA (DCQA). First, it shows the structure and functionality of Skoll, which is an environment that defines a generic around-the-world, around-the-clock QA process and several sophisticated tools that support this process. Second, it describes several novel QA processes built using the Skoll environment. Third, it presents two studies using Skoll: one involving user testing of the Mozilla browser and another involving continuous build, integration, and testing of the ACE+TAO communication software package. The results of our studies suggest that the Skoll environment can manage and control distributed continuous QA processes more effectively than conventional QA processes. For example, our DCQA processes rapidly identified problems that had taken the ACE+TAO developers much longer to find and several of which they had not found. Moreover, the automatic analysis of QA results provided developers information that enabled them to quickly find the root causes of problems %B Software Engineering, IEEE Transactions on %V 33 %P 510 - 525 %8 2007/08// %@ 0098-5589 %G eng %N 8 %R 10.1109/TSE.2007.70719 %0 Conference Paper %B 3rd International Workshop on Semantic Web Enabled Software Engineering (SWESE 2007), Innsubruk, Austria %D 2007 %T Software configuration management using ontologies %A Shahri,H. H %A Hendler,J. A %A Porter, Adam %B 3rd International Workshop on Semantic Web Enabled Software Engineering (SWESE 2007), Innsubruk, Austria %8 2007/// %G eng %0 Conference Paper %B Proc. 2007 Intl. Conf. on Computational Cultural Dynamics %D 2007 %T Soma models of the behaviors of stakeholders in the afghan drug economy: A preliminary report %A Sliva,A. %A Martinez,V. %A Simari,G. I %A V.S. Subrahmanian %X Most cultural reasoning today is done by anthropologists andsociologists who use their detailed knowledge of culture to make predictions about how a given group will respond to a given situation. The main problem with this is that experts in a particular culture or subculture are few and not readily accessible to the many who might suddenly need to tap their expertise (e.g. in cases of war or conflict). In this paper, we briefly describe how the SOMA (Stochastic Opponent Mod- eling Agents) paradigm proposed by the authors can and has been used to model the behaviors of various stake-holders in the drug trade in Afghanistan. %B Proc. 2007 Intl. Conf. on Computational Cultural Dynamics %P 78 - 86 %8 2007/// %G eng %0 Conference Paper %B International Conference on Computer Graphics and Interactive Techniques %D 2007 %T Sorting in space: multidimensional, spatial, and metric data structures for computer graphics applications %A Samet, Hanan %B International Conference on Computer Graphics and Interactive Techniques %8 2007/// %G eng %0 Journal Article %J ACM Trans. Database Syst. %D 2007 %T Spatial join techniques %A Jacox,Edwin H. %A Samet, Hanan %K external memory algorithms %K plane-sweep %K Spatial join %X A variety of techniques for performing a spatial join are reviewed. Instead of just summarizing the literature and presenting each technique in its entirety, distinct components of the different techniques are described and each is decomposed into an overall framework for performing a spatial join. A typical spatial join technique consists of the following components: partitioning the data, performing internal-memory spatial joins on subsets of the data, and checking if the full polygons intersect. Each technique is decomposed into these components and each component addressed in a separate section so as to compare and contrast similar aspects of each technique. The goal of this survey is to describe the algorithms within each component in detail, comparing and contrasting competing methods, thereby enabling further analysis and experimentation with each component and allowing the best algorithms for a particular situation to be built piecemeal, or, even better, enabling an optimizer to choose which algorithms to use. %B ACM Trans. Database Syst. %V 32 %8 2007/03// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/1206049.1206056 %N 1 %R 10.1145/1206049.1206056 %0 Conference Paper %B Proceedings of the 15th annual ACM international symposium on Advances in geographic information systems %D 2007 %T STEWARD: architecture of a spatio-textual search engine %A Lieberman,Michael D. %A Samet, Hanan %A Sankaranarayanan,Jagan %A Sperling,Jon %K geocoding %K spatio-textual search engine %K STEWARD %X STEWARD ("Spatio-Textual Extraction on the Web Aiding Retrieval of Documents"), a system for extracting, querying, and visualizing textual references to geographic locations in unstructured text documents, is presented. Methods for retrieving and processing web documents, extracting and disambiguating georeferences, and identifying geographic focus are described. A brief overview of STEWARD's querying capabilities, as well as the design of an intuitive user interface, are provided. Finally, several application scenarios and future extensions to STEWARD are discussed. %B Proceedings of the 15th annual ACM international symposium on Advances in geographic information systems %S GIS '07 %I ACM %C New York, NY, USA %P 25:1–25:8 - 25:1–25:8 %8 2007/// %@ 978-1-59593-914-2 %G eng %U http://doi.acm.org/10.1145/1341012.1341045 %R 10.1145/1341012.1341045 %0 Conference Paper %B Proceedings of the 8th annual international conference on Digital government research: bridging disciplines & domains %D 2007 %T STEWARD: demo of spatio-textual extraction on the web aiding the retrieval of documents %A Samet, Hanan %A Lieberman,Michael D. %A Sankaranarayanan,Jagan %A Sperling,Jon %K geocoding %K spatio-textual search engine %K STEWARD %X A spatio-textual search engine, termed "STEWARD" is demonstrated where document similarity is based on both the textual similarity as well as the spatial proximity of the locations in the document to the spatial search input. STEWARD's performance is enhanced by the presence of a document tagger that is able to identify textual references to geographical entities. The user-interface of STEWARD provides the ability to browse results, thereby making it a valuable "knowledge discovery" tool. %B Proceedings of the 8th annual international conference on Digital government research: bridging disciplines & domains %S dg.o '07 %I Digital Government Society of North America %P 300 - 301 %8 2007/// %@ 1-59593-599-1 %G eng %U http://dl.acm.org/citation.cfm?id=1248460.1248529 %0 Journal Article %J Wireless communications and mobile computing %D 2007 %T Storage routing for dtn congestion control %A Seligman,M. %A Fall,K. %A Mundur, Padma %B Wireless communications and mobile computing %V 7 %P 1183 - 1196 %8 2007/// %G eng %N 10 %0 Journal Article %J Multimedia Tools and Applications %D 2007 %T Story creation from heterogeneous data sources %A Fayzullin,M. %A V.S. Subrahmanian %A Albanese, M. %A Cesarano,C. %A Picariello, A. %X There are numerous applications where there is a need to rapidly infer a story about a given subject from a given set of potentially heterogeneous data sources. In this paper, we formally define a story to be a set of facts about a given subject that satisfies a “story length” constraint. An optimal story is a story that maximizes the value of an objective function measuring the goodness of a story. We present algorithms to extract stories from text and other data sources. We also develop an algorithm to compute an optimal story, as well as three heuristic algorithms to rapidly compute a suboptimal story. We run experiments to show that constructing stories can be efficiently performed and that the stories constructed by these heuristic algorithms are high quality stories. We have built a prototype STORY system based on our model—we briefly describe the prototype as well as one application in this paper. %B Multimedia Tools and Applications %V 33 %P 351 - 377 %8 2007/// %G eng %N 3 %R 10.1007/s11042-007-0100-4 %0 Journal Article %J Nature %D 2007 %T Structural Biology: Analysis of 'downhill' protein folding; Analysis of protein-folding cooperativity (Reply) %A Sadqi,Mourad %A Fushman, David %A Muñoz,Victor %K Astronomy %K astrophysics %K Biochemistry %K Bioinformatics %K Biology %K biotechnology %K cancer %K cell cycle %K cell signalling. %K climate change %K Computational Biology %K development %K developmental biology %K DNA %K drug discovery %K earth science %K ecology %K environmental science %K Evolution %K evolutionary biology %K functional genomics %K Genetics %K Genomics %K geophysics %K immunology %K interdisciplinary science %K life %K marine biology %K materials science %K medical research %K medicine %K metabolomics %K molecular biology %K molecular interactions %K nanotechnology %K Nature %K neurobiology %K neuroscience %K palaeobiology %K pharmacology %K Physics %K proteomics %K quantum physics %K RNA %K Science %K science news %K science policy %K signal transduction %K structural biology %K systems biology %K transcriptomics %X Ferguson et al. and Zhou and Bai criticize the quality of our nuclear magnetic resonance (NMR) data and atom-by-atom analysis of global 'downhill' folding, also claiming that the data are compatible with two-state folding. %B Nature %V 445 %P E17-E18 - E17-E18 %8 2007/02/15/ %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v445/n7129/full/nature05645.html?lang=en %N 7129 %R 10.1038/nature05645 %0 Conference Paper %B Computer Vision, 2007. ICCV 2007. IEEE 11th International Conference on %D 2007 %T A Study of Face Recognition as People Age %A Ling,Haibin %A Soatto,S. %A Ramanathan,N. %A Jacobs, David W. %K Bayesian %K difference;passport %K machine;face %K machines; %K magnitude;gradient %K orientation %K photo %K pyramid;hierarchical %K recognition;face %K recognition;support %K representation;face %K task;support %K technique;discriminative %K techniques;intensity %K vector %K verification %K verification;gradient %X In this paper we study face recognition across ages within a real passport photo verification task. First, we propose using the gradient orientation pyramid for this task. Discarding the gradient magnitude and utilizing hierarchical techniques, we found that the new descriptor yields a robust and discriminative representation. With the proposed descriptor, we model face verification as a two-class problem and use a support vector machine as a classifier. The approach is applied to two passport data sets containing more than 1,800 image pairs from each person with large age differences. Although simple, our approach outperforms previously tested Bayesian technique and other descriptors, including the intensity difference and gradient with magnitude. In addition, it works as well as two commercial systems. Second, for the first time, we empirically study how age differences affect recognition performance. Our experiments show that, although the aging process adds difficulty to the recognition task, it does not surpass illumination or expression as a confounding factor. %B Computer Vision, 2007. ICCV 2007. IEEE 11th International Conference on %P 1 - 8 %8 2007/10// %G eng %R 10.1109/ICCV.2007.4409069 %0 Journal Article %J Advances in Complex Systems %D 2007 %T Swarm Intelligence Systems Using Guided Self-Organization for Collective Problem Solving %A RodrIGuez,A. %A Grushin,A. %A Reggia, James A. %A HAUPTMAN,AMI %A SIPPER,M. %A PAN,Z. %A Reggia, James A. %A GAO,D. %A DARABOS,C. %A GIACOBINI,M. %A others %B Advances in Complex Systems %V 10 %P 5 - 34 %8 2007/// %G eng %N 1 %0 Conference Paper %B Parallel and Distributed Processing Symposium, International %D 2007 %T Taking Advantage of Collective Operation Semantics for Loosely Coupled Simulations %A Wu,Joe Shang-Chieh %A Sussman, Alan %X Although a loosely coupled component-based framework offers flexibility and versatility for building and deploying large-scale multi-physics simulation systems, the performance of such a system can suffer from excessive buffering of data objects which may or may not be transferred between components. By taking advantages of the collective properties of parallel simulation components, which is common for data-parallel scientific applications, an optimization method, which we call buddy-help, can greatly enhance overall performance. Buddy-Help can reduce the time taken for buffering operations in an exporting component, when there are timing differences across processes in the exporting component. The optimization enables skipping unnecessary buffering operations, once another process, which has already performed the collective export operation, has determined that the buffered data will never be needed. Because an analytical study would be very difficult due to the complexity of the overall coupled simulation system, the performance improvement enabled by buddy-help is investigated via a micro-benchmark specifically designed to illustrate the behavior of coupled simulation scenarios under which buddy-help can provide performance gains. %B Parallel and Distributed Processing Symposium, International %I IEEE Computer Society %C Los Alamitos, CA, USA %P 128 - 128 %8 2007/// %@ 1-4244-0909-8 %G eng %R http://doi.ieeecomputersociety.org/10.1109/IPDPS.2007.370318 %0 Journal Article %J Multimedia, IEEE Transactions on %D 2007 %T Target Tracking Using a Joint Acoustic Video System %A Cevher, V. %A Sankaranarayanan,A. C %A McClellan, J.H. %A Chellapa, Rama %K (numerical %K acoustic %K adaptive %K appearance %K approach;synchronization;time-delay %K data %K delay;acoustic %K divergence;acoustic %K estimate;joint %K estimation;hidden %K feature %K filter;sliding %K Filtering %K fusion;multitarget %K fusion;synchronisation;target %K highways;direction-of-arrival %K Kullback-Leibler %K methods);sensor %K model;particle %K processing; %K processing;automated %K propagation %K removal;optical %K signal %K system;multimodal %K tracking;acoustic %K tracking;direction-of-arrival %K tracking;occlusion;online %K tracking;particle %K tracking;video %K variable;visual %K video %K window;state-space %X In this paper, a multitarget tracking system for collocated video and acoustic sensors is presented. We formulate the tracking problem using a particle filter based on a state-space approach. We first discuss the acoustic state-space formulation whose observations use a sliding window of direction-of-arrival estimates. We then present the video state space that tracks a target's position on the image plane based on online adaptive appearance models. For the joint operation of the filter, we combine the state vectors of the individual modalities and also introduce a time-delay variable to handle the acoustic-video data synchronization issue, caused by acoustic propagation delays. A novel particle filter proposal strategy for joint state-space tracking is introduced, which places the random support of the joint filter where the final posterior is likely to lie. By using the Kullback-Leibler divergence measure, it is shown that the joint operation of the filter decreases the worst case divergence of the individual modalities. The resulting joint tracking filter is quite robust against video and acoustic occlusions due to our proposal strategy. Computer simulations are presented with synthetic and field data to demonstrate the filter's performance %B Multimedia, IEEE Transactions on %V 9 %P 715 - 727 %8 2007/06// %@ 1520-9210 %G eng %N 4 %R 10.1109/TMM.2007.893340 %0 Journal Article %J Information Processing & Management %D 2007 %T Task-based evaluation of text summarization using Relevance Prediction %A Hobson,Stacy President %A Dorr, Bonnie J %A Monz,Christof %A Schwartz,Richard %K Relevance prediction %K Summarization evaluation %K Summary usefulness %X This article introduces a new task-based evaluation measure called Relevance Prediction that is a more intuitive measure of an individual’s performance on a real-world task than interannotator agreement. Relevance Prediction parallels what a user does in the real world task of browsing a set of documents using standard search tools, i.e., the user judges relevance based on a short summary and then that same user—not an independent user—decides whether to open (and judge) the corresponding document. This measure is shown to be a more reliable measure of task performance than LDC Agreement, a current gold-standard based measure used in the summarization evaluation community. Our goal is to provide a stable framework within which developers of new automatic measures may make stronger statistical statements about the effectiveness of their measures in predicting summary usefulness. We demonstrate—as a proof-of-concept methodology for automatic metric developers—that a current automatic evaluation measure has a better correlation with Relevance Prediction than with LDC Agreement and that the significance level for detected differences is higher for the former than for the latter. %B Information Processing & Management %V 43 %P 1482 - 1499 %8 2007/11// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306457307000234 %N 6 %R 10.1016/j.ipm.2007.01.002 %0 Conference Paper %B Proceedings of the 7th ACM/IEEE-CS joint conference on Digital libraries %D 2007 %T Task-based interaction with an integrated multilingual, multimedia information system: a formative evaluation %A Zhang,Pengyi %A Plettenberg,Lynne %A Klavans,Judith L. %A Oard, Douglas %A Soergel,Dagobert %K Cross-language information retrieval %K multimedia %K User studies %X This paper describes a formative evaluation of an integrated multilingual, multimedia information system, a series of user studies designed to guide system development. The system includes automatic speech recognition for English, Chinese, and Arabic, automatic translation from Chinese and Arabic into English, and query-based and profile-based search options. The study design emphasizes repeated evaluation with the same (increasingly experienced) participants, exploration of alternative task designs, rich qualitative and quantitative data collection, and rapid analysis to provide the timely feedback needed to support iterative and responsive development. Results indicate that users presented with materials in a language that they do not know can generate remarkably useful work products, but that integration of transcription, translation, search and profile management poses challenges that would be less evident were each technology to be evaluated in isolation. %B Proceedings of the 7th ACM/IEEE-CS joint conference on Digital libraries %S JCDL '07 %I ACM %C New York, NY, USA %P 117 - 126 %8 2007/// %@ 978-1-59593-644-8 %G eng %U http://doi.acm.org/10.1145/1255175.1255199 %R 10.1145/1255175.1255199 %0 Conference Paper %B Life Science Systems and Applications Workshop, 2007. LISA 2007. IEEE/NIH %D 2007 %T A taxonomy for medical image registration acceleration techniques %A Plishker,W. %A Dandekar,O. %A Bhattacharyya, Shuvra S. %A Shekhar,R. %B Life Science Systems and Applications Workshop, 2007. LISA 2007. IEEE/NIH %P 160 - 163 %8 2007/// %G eng %0 Journal Article %J IEEE Transactions on Software Engineering %D 2007 %T Techniques for Classifying Executions of Deployed Software to Support Software Engineering Tasks %A Haran,Murali %A Karr,Alan %A Last,Michael %A Orso,Alessandro %A Porter, Adam %A Sanil,Ashish %A Fouch?,Sandro %K execution classification %K remote analysis/measurement. %X There is an increasing interest in techniques that support analysis and measurement of fielded software systems. These techniques typically deploy numerous instrumented instances of a software system, collect execution data when the instances run in the field, and analyze the remotely collected data to better understand the system's in-the-field behavior. One common need for these techniques is the ability to distinguish execution outcomes (e.g., to collect only data corresponding to some behavior or to determine how often and under which condition a specific behavior occurs). Most current approaches, however, do not perform any kind of classification of remote executions and either focus on easily observable behaviors (e.g., crashes) or assume that outcomes' classifications are externally provided (e.g., by the users). To address the limitations of existing approaches, we have developed three techniques for automatically classifying execution data as belonging to one of several classes. In this paper, we introduce our techniques and apply them to the binary classification of passing and failing behaviors. Our three techniques impose different overheads on program instances and, thus, each is appropriate for different application scenarios. We performed several empirical studies to evaluate and refine our techniques and to investigate the trade-offs among them. Our results show that 1) the first technique can build very accurate models, but requires a complete set of execution data; 2) the second technique produces slightly less accurate models, but needs only a small fraction of the total execution data; and 3) the third technique allows for even further cost reductions by building the models incrementally, but requires some sequential ordering of the software instances' instrumentation. %B IEEE Transactions on Software Engineering %V 33 %P 287 - 304 %8 2007/// %@ 0098-5589 %G eng %N 5 %R http://doi.ieeecomputersociety.org/10.1109/TSE.2007.1004 %0 Conference Paper %B AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning %D 2007 %T Toward domain-neutral human-level metacognition %A Anderson,M. L %A Schmill,M. %A Oates,T. %A Perlis, Don %A Josyula,D. %A Wright,D. %A Wilson,S. %B AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning %8 2007/// %G eng %0 Journal Article %J Technical Reports from UMIACS %D 2007 %T Toward Specifying and Validating Cross-Domain Policies %A Hicks, Michael W. %A Swamy,N. %A Tsang,S. %X Formal security policies are extremely useful for two related reasons. First, they allow a policy to be considered in isolation, separate from programs under the purview of the policy and separate from the implementation of the policy's enforcement. Second, policies can be checked for compliance against higher-level security goals by using automated analyses. By contrast, ad hoc enforcement mechanisms (for which no separate policies are specified) enjoy neither benefit, and non-formal policies enjoy the first but not the second. We would like to understand how best to define (and enforce) multi-level security policies when information must be shared across domains that have varying levels of trust (the so-called "cross domain" problem). Because we wish to show such policies meet higher-level security goals with high assurance, we are interested in specifying cross domain policies formally, and then reasoning about them using automated tools. In this report, we briefly survey work that presents formal security policies with cross-domain concerns, in particular with respect to the problem of downgrading. We also describe correctness properties for such policies, all based on noninterference. Finally, we briefly discuss recently-developed tools for analyzing formal security policies; though no existing tools focus on the analysis of downgrading-oriented policies, existing research points the way to providing such support. %B Technical Reports from UMIACS %8 2007/04/17/undef %G eng %0 Conference Paper %B Biomedical Circuits and Systems Conference, 2007. BIOCAS 2007. IEEE %D 2007 %T Towards a heterogeneous medical image registration acceleration platform %A Plishker,W. %A Dandekar,O. %A Bhattacharyya, Shuvra S. %A Shekhar,R. %B Biomedical Circuits and Systems Conference, 2007. BIOCAS 2007. IEEE %P 231 - 234 %8 2007/// %G eng %0 Journal Article %J Virtual Reality %D 2007 %T Towards the development of a virtual environment-based training system for mechanical assembly operations %A Brough,John %A Schwartz,Maxim %A Gupta, Satyandra K. %A Anand,Davinder %A Kavetsky,Robert %A Pettersen,Ralph %K Computer science %X In this paper, we discuss the development of Virtual Training Studio (VTS), a virtual environment-based training system that allows training supervisors to create training instructions and allows trainees to learn assembly operations in a virtual environment. Our system is mainly focused on the cognitive side of training so that trainees can learn to recognize parts, remember assembly sequences, and correctly orient the parts during assembly operations. Our system enables users to train using the following three training modes: (1) Interactive Simulation, (2) 3D Animation, and (3) Video. Implementing these training modes required us to develop several new system features. This paper presents an overview of the VTS system and describes a few main features of the system. We also report user test results that show how people train using our system. The user test results indicate that the system is able to support a wide variety of training preferences and works well to support training for assembly operations. %B Virtual Reality %V 11 %P 189 - 206 %8 2007/// %@ 1359-4338 %G eng %U http://www.springerlink.com/content/g7760422m13l83x1/abstract/ %N 4 %R 10.1007/s10055-007-0076-4 %0 Journal Article %J Nucleic Acids Research %D 2007 %T TREMOR—a tool for retrieving transcriptional modules by incorporating motif covariance %A Singh,Larry N. %A Wang,Li-San %A Hannenhalli, Sridhar %X A transcriptional module (TM) is a collection of transcription factors (TF) that as a group, co-regulate multiple, functionally related genes. The task of identifying TMs poses an important biological challenge. Since TFs belong to evolutionarily and structurally related families, TF family members often bind to similar DNA motifs and can confound sequence-based approaches to TM identification. A previous approach to TM detection addresses this issue by pre-selecting a single representative from each TF family. One problem with this approach is that closely related transcription factors can still target sufficiently distinct genes in a biologically meaningful way, and thus, pre-selecting a single family representative may in principle miss certain TMs. Here we report a method—TREMOR (Transcriptional Regulatory Module Retriever). This method uses the Mahalanobis distance to assess the validity of a TM and automatically incorporates the inter-TF binding similarity without resorting to pre-selecting family representatives. The application of TREMOR on human muscle-specific, liver-specific and cell-cycle-related genes reveals TFs and TMs that were validated from literature and also reveals additional related genes. %B Nucleic Acids Research %V 35 %P 7360 - 7371 %8 2007/12/01/ %G eng %U http://nar.oxfordjournals.org/content/35/21/7360.abstract %N 21 %R 10.1093/nar/gkm885 %0 Conference Paper %B Proceedings of the First International Conference on Computational Cultural Dynamics (ICCCD 2007) %D 2007 %T T-rex: A domain-independent system for automated cultural information extraction %A Albanese, M. %A V.S. Subrahmanian %X RDF (Resource Description Framework) is a web standarddefined by the World Wide Web Consortium. In RDF, we can define schemas of interest. For example, we can define a schema about tribes on the Pakistan-Afghanistan borderland, or a schema about violent events. An RDF instance is a set of facts that are compatible with the schema. The principal contribution of this paper is the development of a scalable system called T-REX (short for “The RDF EXtractor”) that allows us to extract instances associated with a user-specified schema, independently of the domain about which we wish to extract data. Using T-REX, we have successfully extracted information about various aspects of about 20 tribes living in the Pakistan-Afghanistan border. Moreover, we have used T-REX to successfully extract occurrences of violent events from a set of 80 news sites in approximately 50 countries. T-REX scales well – it has processed approximately 45,000 web pages per day for the last 6 months. %B Proceedings of the First International Conference on Computational Cultural Dynamics (ICCCD 2007) %8 2007/// %G eng %0 Journal Article %J Association for Logic Programming Newsletter %D 2007 %T Uncertainty in logic programming: Some recollections %A V.S. Subrahmanian %B Association for Logic Programming Newsletter %V 20 %8 2007/// %G eng %N 2 %0 Journal Article %J Molecular biology and evolution %D 2007 %T A unified model explaining the offsets of overlapping and near-overlapping prokaryotic genes %A Kingsford, Carl %A Delcher,A. L %A Salzberg,S. L %B Molecular biology and evolution %V 24 %P 2091 - 2091 %8 2007/// %G eng %N 9 %0 Journal Article %J Mathematical Programming A %D 2007 %T Universal Duality in Conic Convex Optimization %A Schurr,Simon P. %A Tits,André %A O'Leary, Dianne P. %X Given a primal-dual pair of linear programs, it is well known that if their optimal values are viewed as lying on the extended real line, then the duality gap is zero, unless both problems are infeasible, in which case the optimal values are +∞ and −∞. In contrast, for optimization problems over nonpolyhedral convex cones, a nonzero duality gap can exist when either the primal or the dual is feasible.For a pair of dual conic convex programs, we provide simple conditions on the ``constraint matrices'' and cone under which the duality gap is zero for every choice of linear objective function and constraint right-hand side. We refer to this property as ``universal duality''. Our conditions possess the following properties: (i) they are necessary and sufficient, in the sense that if (and only if) they do not hold, the duality gap is nonzero for some linear objective function and constraint right-hand side; (ii) they are metrically and topologically generic; and (iii) they can be verified by solving a single conic convex program. We relate to universal duality the fact that the feasible sets of a primal convex program and its dual cannot both be bounded, unless they are both empty. Finally we illustrate our theory on a class of semidefinite programs that appear in control theory applications. %B Mathematical Programming A %V 109 %P 69 - 88 %8 2007/// %G eng %N 1 %0 Journal Article %J Universal usability: Designing computer interfaces for user diversity %D 2007 %T Universal Usability: A grand challenge for HCI %A Shneiderman, Ben %B Universal usability: Designing computer interfaces for user diversity %8 2007/// %G eng %0 Conference Paper %B Proceedings of the 7th ACM SIGCOMM conference on Internet measurement %D 2007 %T Usage-based dhcp lease time optimization %A Khadilkar,Manas %A Feamster, Nick %A Sanders,Matt %A Clark,Russ %K dhcp %K network management %K OPTIMIZATION %K usage %X The Dynamic Host Configuration Protocol (DHCP) is used to dynamically allocate address space to hosts on a local area network. Despite its widespread usage, few studies exist on DHCP usage patterns, and even less is known about the importance of setting the lease time (the time that a client retains ownership over some IP address) to an appropriate value. Lease time can greatly affect the tradeoff between address space utilization and the number of both renewal messages and client session expirations. In this paper, using a DHCP trace for 5 weekdays from the Georgia Tech campus network, we present the largest known study of DHCP utilization. We also explore how various strategies for setting lease times can dramatically reduce the number of renewals and expirations without prohibitively increasing address space utilization. %B Proceedings of the 7th ACM SIGCOMM conference on Internet measurement %S IMC '07 %I ACM %C New York, NY, USA %P 71 - 76 %8 2007/// %@ 978-1-59593-908-1 %G eng %U http://doi.acm.org/10.1145/1298306.1298315 %R 10.1145/1298306.1298315 %0 Journal Article %J The VLDB Journal %D 2007 %T Using a distributed quadtree index in peer-to-peer networks %A Tanin,Egemen %A Harwood,Aaron %A Samet, Hanan %K Distributed data structures %K peer-to-peer networks %K quadtrees %K spatial data structures %X Peer-to-peer (P2P) networks have become a powerful means for online data exchange. Currently, users are primarily utilizing these networks to perform exact-match queries and retrieve complete files. However, future more data intensive applications, such as P2P auction networks, P2P job-search networks, P2P multiplayer games, will require the capability to respond to more complex queries such as range queries involving numerous data types including those that have a spatial component. In this paper, a distributed quadtree index that adapts the MX-CIF quadtree is described that enables more powerful accesses to data in P2P networks. This index has been implemented for various prototype P2P applications and results of experiments are presented. Our index is easy to use, scalable, and exhibits good load-balancing properties. Similar indices can be constructed for various multidimensional data types with both spatial and non-spatial components. %B The VLDB Journal %V 16 %P 165 - 178 %8 2007/04// %@ 1066-8888 %G eng %U http://dx.doi.org/10.1007/s00778-005-0001-y %N 2 %R 10.1007/s00778-005-0001-y %0 Conference Paper %B Proceedings of the 4th international conference on Data integration in the life sciences %D 2007 %T Using annotations from controlled vocabularies to find meaningful associations %A Lee,W. J %A Raschid, Louiqa %A Srinivasan,P. %A Shah,N. %A Rubin,D. %A Noy,N. %B Proceedings of the 4th international conference on Data integration in the life sciences %P 247 - 263 %8 2007/// %G eng %0 Report %D 2007 %T Using Content-Addressable Networks for Load Balancing in Desktop Grids %A Kim,Jik-Soo %A Keleher,Peter %A Marsh,Michael %A Bhattacharjee, Bobby %A Sussman, Alan %K Technical Report %X Desktop grids combine Peer-to-Peer and Grid computing techniques to improvethe robustness, reliability and scalability of job execution infrastructures. However, efficiently matching incoming jobs to available system resources and achieving good load balance in a fully decentralized and heterogeneous computing environment is a challenging problem. In this paper, we extend our prior work with a new decentralized algorithm for maintaining approximate global load information, and a job pushing mechanism that uses the global information to push jobs towards underutilized portions of the system. The resulting system more effectively balances load and improves overall system throughput. Through a comparative analysis of experimental results across different system configurations and job profiles, performed via simulation, we show that our system can reliably execute Grid applications on a distributed set of resources both with low cost and with good load balance. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-16 %8 2007/03/29/ %G eng %U http://drum.lib.umd.edu//handle/1903/4355 %0 Conference Paper %B Proceedings of the 22nd national conference on Artificial intelligence - Volume 2 %D 2007 %T Using multiresolution learning for transfer in image classification %A Eaton,Eric %A desJardins, Marie %A Stevenson,John %X Our work explores the transfer of knowledge at multiple levels of abstraction to improve learning. By exploiting the similarities between objects at various levels of detail, multiresolution learning can facilitate transfer between image classification tasks.We extract features from images at multiple levels of resolution, then use these features to create models at different resolutions. Upon receiving a new task, the closest-matching stored model can be generalized (adapted to the appropriate resolution) and transferred to the new task. %B Proceedings of the 22nd national conference on Artificial intelligence - Volume 2 %S AAAI'07 %I AAAI Press %P 1852 - 1853 %8 2007/// %@ 978-1-57735-323-2 %G eng %U http://dl.acm.org/citation.cfm?id=1619797.1619927 %0 Conference Paper %D 2007 %T Using virtual demonstrations for creating multi-media training instructions %A Schwartz,M. %A Gupta,S.K. %A Anand,D. K. %A Brough,J. E. %A Kavetsky,R. %X Often generating training instructions for virtual environments is a long and tedious process. In thispaper, we discuss the development of a virtual environment (VE) instruction generating tool called Virtual Author which is the main component of the Virtual Training Studio (VTS). VTS is a virtual environment-based training system that provides instructors with a tool to create training instructions and allows trainees to learn assembly operations in a personal virtual environment. The Virtual Author tool is designed to allow an instructor to perform virtual demonstrations using CAD models in the virtual environment in order to quickly generate VE-based training instructions for use in VTS. This paper describes the algorithms used to carry out motion smoothening of instructor’s actions, automated text instruction generation based on part and assembly motions, and extraction of alignment constraints from 3D CAD models to support instruction generation. We also present examples to illustrate how the use of the Virtual Author tool leads to a significant reduction in the training instruction generation time. %8 2007/// %G eng %U http://www.cadanda.com/CAD_4_1-4__11.PDF %0 Conference Paper %B Visual Analytics Science and Technology, 2007. VAST 2007. IEEE Symposium on %D 2007 %T VAST 2007 Contest - Blue Iguanodon %A Grinstein,G. %A Plaisant, Catherine %A Laskowski,S. %A O'Connell,T. %A Scholtz,J. %A Whiting,M. %K activities;large %K activities;visual %K analytic %K collection;terrorist %K data %K heterogeneous %K illegal %K interfaces; %K tools;data %K user %K visualisation;graphical %X Visual analytics experts realize that one effective way to push the field forward and to develop metrics for measuring the performance of various visual analytics components is to hold an annual competition. The second visual analytics science and technology (VAST) contest was held in conjunction with the 2007 IEEE VAST symposium. In this contest participants were to use visual analytic tools to explore a large heterogeneous data collection to construct a scenario and find evidence buried in the data of illegal and terrorist activities that were occurring. A synthetic data set was made available as well as tasks. In this paper we describe some of the advances we have made from the first competition held in 2006. %B Visual Analytics Science and Technology, 2007. VAST 2007. IEEE Symposium on %P 231 - 232 %8 2007/11/30/1 %G eng %R 10.1109/VAST.2007.4389032 %0 Conference Paper %B IEEE Military Communications Conference, 2007. MILCOM 2007 %D 2007 %T Verified Enforcement of Security Policies for Cross-Domain Information Flows %A Swamy,Nikhil %A Hicks, Michael W. %A Tsang,Simon %K Collaboration %K Computer languages %K Data security %K Educational institutions %K Government %K Information analysis %K Information filtering %K Information filters %K Information security %K Vehicles %X We describe work in progress that uses program analysis to show that security-critical programs, such as cross-domain guards, correctly enforce cross-domain security policies. We are enhancing existing techniques from the field of Security-oriented Programming Languages to construct a new language for the construction of secure networked applications, SELINKS. In order to specify and enforce expressive and fine-grained policies, we advocate dynamically associating security labels with sensitive entities. Programs written in SELINKS are statically guaranteed to correctly manipulate an entity's security labels and to ensure that the appropriate policy checks mediate all operations that are performed on the entity. We discuss the design of our main case study : a web-based Collaborative Planning Application that will permit a collection of users, with varying security requirements and clearances, to access sensitive data sources and collaboratively create documents based on these sources. %B IEEE Military Communications Conference, 2007. MILCOM 2007 %I IEEE %P 1 - 7 %8 2007/10/29/31 %@ 978-1-4244-1513-7 %G eng %R 10.1109/MILCOM.2007.4455189 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2007 %T Viable but nonculturable Vibrio cholerae O1 in biofilms in the aquatic environment and their role in cholera transmission %A Alam,M. %A Sultana,M. %A Nair,G. B. %A Siddique,A. K. %A Hasan,N. A. %A Sack,R. B. %A Sack,D. A. %A Ahmed,K. U. %A Sadique,A. %A Watanabe,H. %A Rita R Colwell %X Vibrio cholerae persists in aquatic environments predominantly in a nonculturable state. In this study coccoid, nonculturable V. cholerae O1 in biofilms maintained for 495 days in Mathbaria, Bangladesh, pond water became culturable upon animal passage. Culturability, biofilm formation, and the wbe, ctxA, and rstR2 genes were monitored by culture, direct fluorescent antibody (DFA), and multiplex PCR. DFA counts were not possible after formation of biofilm. Furthermore, wbe, but not ctxA, were amplifiable, even after incubation for 54 and 68 days at room temperature (≈25°C) and 4°C, respectively, when no growth was detectable. Slower biofilm formation and extended culturability were observed for cultures incubated at 4°C, compared with ≈25°C, suggesting biofilm production to be temperature dependent and linked to loss of culturability. Small colonies appearing after incubation in microcosms for 54 and 68 days at 25°C and 4°C, respectively, were wbe positive and ctxA and rstR2 negative, indicating loss of bacteriophage CTXΦ. The coccoid V. cholerae O1 observed as free cells in microcosms incubated for 495 days could not be cultured, but biofilms in the same microcosms yielded culturable cells. It is concluded that biofilms can act as a reservoir for V. cholerae O1 between epidemics because of its long-term viability in biofilms. In contrast to biofilms produced in Mathbaria pond water, V. cholerae O1 in biofilms present in cholera stools and incubated under identical conditions as the Mathbaria pond water biofilms could not be cultured after 2 months, indicating that those V. cholerae cells freshly discharged into the environment are significantly less robust than cells adapted to environmental conditions.Bangladesh bacteriophage CTXΦ DFA multiplex-PCR ctxA %B Proceedings of the National Academy of Sciences %V 104 %P 17801 - 17801 %8 2007/// %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/104/45/17801 %N 45 %R 10.1073/pnas.0705599104 %0 Journal Article %J ACM SIGKDD Explorations Newsletter %D 2007 %T Visual analysis of dynamic group membership in temporal social networks %A Kang,Hyunmo %A Getoor, Lise %A Singh,Lisa %K dynamic group detection %K group visualization %K visual data mining %X C-Group is a tool for analyzing dynamic group membership in temporal social networks over time. Unlike most network visualization tools, which show the group structure within an entire network, or the group membership for a single actor, C-Group allows users to focus their analysis on a pair of individuals. While C-Group allows for viewing the addition and deletion of nodes (actors) and edges (relationships) over time, its major contribution is its focus on changing group memberships over time. By doing so, users can investigate the context of temporal group memberships for the pair. C-Group provides users with a flexible interface for defining (and redefining) groups interactively, and supports two novel visual representations of the evolving group memberships. This flexibility gives users alternate views that are appropriate for different network sizes and provides users with different insights into the grouping behavior. We demonstrate the utility of the tool on a scientific publication network. %B ACM SIGKDD Explorations Newsletter %V 9 %P 13 - 21 %8 2007/12// %@ 1931-0145 %G eng %U http://doi.acm.org/10.1145/1345448.1345452 %N 2 %R 10.1145/1345448.1345452 %0 Conference Paper %B Information Visualization, 2007. IV '07. 11th International Conference %D 2007 %T Visual Mining of Multi-Modal Social Networks at Different Abstraction Levels %A Singh,L. %A Beard,M. %A Getoor, Lise %A Blake,M.B. %K computing; %K database;interactive %K graph %K management %K mining;data %K network;structural %K property;visual %K sciences %K social %K systems;graph %K systems;social %K theory;interactive %K visualisation;database %K visualization;multimodal %X Social networks continue to become more and more feature rich. Using local and global structural properties and descriptive attributes are necessary for more sophisticated social network analysis and support for visual mining tasks. While a number of visualization tools for social network applications have been developed, most of them are limited to uni-modal graph representations. Some of the tools support a wide range of visualization options, including interactive views. Others have better support for calculating structural graph properties such as the density of the graph or deploying traditional statistical social network analysis. We present Invenio, a new tool for visual mining of socials. Invenio integrates a wide range of interactive visualization options from Prefuse, with graph mining algorithm support from JUNG. While the integration expands the breadth of functionality within the core engine of the tool, our goal is to interactively explore multi-modal, multi-relational social networks. Invenio also supports construction of views using both database operations and basic graph mining operations. %B Information Visualization, 2007. IV '07. 11th International Conference %P 672 - 679 %8 2007/07// %G eng %R 10.1109/IV.2007.126 %0 Report %D 2007 %T Web Archiving: Organizing Web Objects into Web Containers to Optimize Access %A Song,Sangchul %A JaJa, Joseph F. %K Technical Report %X The web is becoming the preferred medium for communicating and storinginformation pertaining to almost any human activity. However it is an ephemeral medium whose contents are constantly changing, resulting in a permanent loss of part of our cultural and scientific heritage on a regular basis. Archiving important web contents is a very challenging technical problem due to its tremendous scale and complex structure, extremely dynamic nature, and its rich heterogeneous and deep contents. In this paper, we consider the problem of archiving a linked set of web objects into web containers in such a way as to minimize the number of containers accessed during a typical browsing session. We develop a method that makes use of the notion of PageRank and optimized graph partitioning to enable faster browsing of archived web contents. We include simulation results that illustrate the performance of our scheme and compare it to the common scheme currently used to organize web objects into web containers. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-42 %8 2007/10/09/ %G eng %U http://drum.lib.umd.edu/handle/1903/7426 %0 Conference Paper %B Proceedings of SPIE–Medical Imaging %D 2007 %T A web collaboration system for content-based image retrieval of medical images %A Tahmoush,D. %A Samet, Hanan %X Building effective content-based image retrieval (CBIR) systemsinvolves the combination of image creation, storage, security, transmission, analysis, evaluation feature extraction, and feature combination in order to store and retrieve medical images effectively. This requires the involvement of a large community of experts across several fields. We have created a CBIR system called Archimedes which integrates the community together without requiring disclosure of sensitive details. Archimedes' system design enables researchers to upload their feature sets and quickly compare the effectiveness of their methods against other stored feature sets. Additionally, research into the techniques used by radiologists is possible in Archimedes through double- blind radiologist comparisons based on their annotations and feature markups. This research archive contains the essential technologies of secure transmission and storage, textual and feature searches, spatial searches, annotation searching, filtering of result sets, feature creation, and bulk loading of features, while creating a repository and testbed for the community. %B Proceedings of SPIE–Medical Imaging %V 6516 %8 2007/// %G eng %0 Journal Article %J Communications of the ACM %D 2007 %T Web science: a provocative invitation to computer science %A Shneiderman, Ben %X Here's how it can awaken computer science to the interdisciplinary possibilities of the Web's socially embedded computing technology. %B Communications of the ACM %V 50 %P 25 - 27 %8 2007/06// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/1247001.1247022 %N 6 %R 10.1145/1247001.1247022 %0 Book Section %B Articulated Motion and Deformable ObjectsArticulated Motion and Deformable Objects %D 2006 %T Acquisition of Articulated Human Body Models Using Multiple Cameras %A Sundaresan,Aravind %A Chellapa, Rama %E Perales,Francisco %E Fisher,Robert %X Motion capture is an important application in different areas such as biomechanics, computer animation, and human-computer interaction. Current motion capture methods typically use human body models in order to guide pose estimation and tracking. We model the human body as a set of tapered super-quadrics connected in an articulated structure and propose an algorithm to automatically estimate the parameters of the model using video sequences obtained from multiple calibrated cameras. Our method is based on the fact that the human body is constructed of several articulated chains that can be visualised as essentially 1-D segments embedded in 3-D space and connected at specific joint locations. The proposed method first computes a voxel representation from the images and maps the voxels to a high dimensional space in order to extract the 1-D structure. A bottom-up approach is then suggested in order to build a parametric (spline-based) representation of a general articulated body in the high dimensional space followed by a top-down probabilistic approach that registers the segments to the known human body model. We then present an algorithm to estimate the parameters of our model using the segmented and registered voxels. %B Articulated Motion and Deformable ObjectsArticulated Motion and Deformable Objects %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4069 %P 78 - 89 %8 2006/// %@ 978-3-540-36031-5 %G eng %U http://dx.doi.org/10.1007/11789239_9 %0 Journal Article %J Proceedings of the Third International WordNet Conference %D 2006 %T Adding dense, weighted connections to wordnet %A Jordan Boyd-Graber %A Fellbaum,C. %A Osherson,D. %A Schapire,R. %X WORDNET, a ubiquitous tool for natural language processing, suffers from sparsity of connectionsbetween its component concepts (synsets). Through the use of human annotators, a subset of the connections between 1000 hand-chosen synsets was assigned a value of “evocation” representing how much the first concept brings to mind the second. These data, along with existing similarity measures, constitute the basis of a method for predicting evocation between previously unrated pairs. %B Proceedings of the Third International WordNet Conference %P 29 - 36 %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 2006 SIGCOMM workshop on Challenged networks %D 2006 %T Alternative custodians for congestion control in delay tolerant networks %A Seligman,Matthew %A Fall,Kevin %A Mundur, Padma %K delay tolerant network %K Routing %B Proceedings of the 2006 SIGCOMM workshop on Challenged networks %S CHANTS '06 %I ACM %C New York, NY, USA %P 229 - 236 %8 2006/// %@ 1-59593-572-X %G eng %U http://doi.acm.org/10.1145/1162654.1162660 %R 10.1145/1162654.1162660 %0 Journal Article %J IACR ePrint report %D 2006 %T Analyzing the HB and HB+ protocols in the “large error” case %A Katz, Jonathan %A Smith,A. %X HB and HB+ are two shared-key, unidirectional authentication protocols whose extremelylow computational cost makes them potentially well-suited for severely resource-constrained devices. Security of these protocols is based on the conjectured hardness of learning parity with noise; that is, learning a secret s given “noisy” dot products of s that are incorrect with probability ε. Although the problem of learning parity with noise is meaningful for any constant ε < 1/2, existing proofs of security for HB and HB+ only imply security when ε < 1/4. In this note, we show how to extend these proofs to the case of arbitrary ε < 1/2. %B IACR ePrint report %V 326 %P 2006 - 2006 %8 2006/// %G eng %0 Journal Article %J The Semantic Web: Research and Applications %D 2006 %T Annotated rdf %A Udrea,O. %A Recupero,D. %A V.S. Subrahmanian %X There are numerous extensions of RDF that support temporal reasoning, reasoning about pedigree, reasoning about uncertainty, and so on. In this paper, we present Annotated RDF (or aRDF for short) in which RDF triples are annotated by members of a partially ordered set (with bottom element) that can be selected in any way desired by the user. We present a formal declarative semantics (model theory) for annotated RDF and develop algorithms to check consistency of aRDF theories and to answer queries to aRDF theories. We show that annotated RDF captures versions of all the forms of reasoning mentioned above within a single unified framework. We develop a prototype aRDF implementation and show that our algorithms work very fast indeed – in fact, in just a matter of seconds for theories with over 100,000 nodes. %B The Semantic Web: Research and Applications %V 4011 %P 487 - 501 %8 2006/// %G eng %R 10.1007/11762256_36 %0 Conference Paper %B Proceedings of the Workshop on Frontiers in Linguistically Annotated Corpora 2006 %D 2006 %T Annotation compatibility working group report %A Meyers,A. %A Fang,A. C. %A Ferro,L. %A Kübler,S. %A Jia-Lin,T. %A Palmer,M. %A Poesio,M. %A Dolbey,A. %A Schuler,K. K. %A Loper,E. %A Zinsmeister,H. %A Penn,G. %A Xue,N. %A Hinrichs,E. %A Wiebe,J. %A Pustejovsky,J. %A Farwell,D. %A Hajicova,E. %A Dorr, Bonnie J %A Hovy,E. %A Onyshkevych,B. A. %A Levin,L. %X This report explores the question of compatibility between annotation projects including translating annotation formalisms to each other or to common forms. Compatibility issues are crucial for systems that use the results of multiple annotation projects. We hope that this report will begin a concerted effort in the field to track the compatibility of annotation schemes for part of speech tagging, time annotation, treebanking, role labeling and other phenomena. %B Proceedings of the Workshop on Frontiers in Linguistically Annotated Corpora 2006 %S LAC '06 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 38 - 53 %8 2006/// %@ 1-932432-78-7 %G eng %U http://dl.acm.org/citation.cfm?id=1641991.1641997 %0 Report %D 2006 %T Anonymous multi-attribute encryption with range query and conditional decryption %A Bethencourt, J. %A Chan, H. %A Perrig, A. %A Elaine Shi %A Song,D. %X We introduce the concept of Anonymous Multi-Attribute Encryption with Range Query and Con-ditional Decryption (AMERQCD). In AMERQCD, a plaintext is encrypted under a point in multi- dimensional space. To a computationally bounded adversary, the ciphertext hides both the plaintext and the point under which it is encrypted. In a range query, a master key owner releases the decryp- tion key for an arbitrary hyper-rectangle in space, thus allowing decryption of ciphertexts previ- ously encrypted under any point within the hyper-rectangle. However, a computationally bounded adversary cannot learn any information on ciphertexts outside the range covered by the decryption key (except the fact that they do not lie within this range). We give an efficient construction based on the Decision Bilinear Diffie-Hellman (D-BDH) and Decision Linear (D-Linear) assumption. %I Carnegie Mellon University %8 2006 %@ CMU-CS-06-135 %G eng %U http://reports-archive.adm.cs.cmu.edu/anon/anon/home/ftp/2006/CMU-CS-06-135.pdf %R Technical Report %0 Journal Article %J Networks %D 2006 %T Approximation algorithms for channel allocation problems in broadcast networks %A Gandhi,Rajiv %A Khuller, Samir %A Srinivasan, Aravind %A Wang,Nan %K channel allocation problem %K distributed approximation algorithm %K edge partitioning problem %X We study two packing problems that arise in the area of dissemination-based information systems; a second theme is the study of distributed approximation algorithms. The problems considered have the property that the space occupied by a collection of objects together could be significantly less than the sum of the sizes of the individual objects. In the Channel Allocation Problem, there are requests that are subsets of topics. There are a fixed number of channels that can carry an arbitrary number of topics. All the topics of each request must be broadcast on some channel. The load on any channel is the number of topics that are broadcast on that channel; the objective is to minimize the maximum load on any channel. We present approximation algorithms for this problem, and also show that the problem is MAX-SNP hard. The second problem is the Edge Partitioning Problem addressed by Goldschmidt, Hochbaum, Levin, and Olinick (Networks, 41:13–23, 2003). Each channel here can deliver topics for at most k requests, and we aim to minimize the total load on all channels. We present an O(n1/3)–approximation algorithm, and also show that the algorithm can be made fully distributed with the same approximation guarantee; we also generalize the (nondistributed) Edge Partitioning Problem of graphs to the case of hypergraphs. © 2006 Wiley Periodicals, Inc. NETWORKS, Vol. 47(4), 225–236 2006 %B Networks %V 47 %P 225 - 236 %8 2006/03/09/ %@ 1097-0037 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/net.20111/abstract %N 4 %R 10.1002/net.20111 %0 Conference Paper %B American Medical Informatics Association 2006 Annual Symposium Proceedings Biomedical and Health Informatics: From Foundations to Applications to Policy, Washington, DC, DW Bates, JH Holmes, and G. Kuperman, editors %D 2006 %T Archimedes, an archive of medical images %A Tahmoush,D. %A Samet, Hanan %X We present a medical image and medical recorddatabase for the storage, research, transmission, and evaluation of medical images. Medical images from any source that supports the DICOM standard can be stored and accessed, as well as associated analysis and annotations. Retrieval is based on patient info, date, doctor’s annotations, features in the images, or a spatial combination. This database supports the secure transmission of sensitive data for tele-medicine and follows all HIPPA regulations. %B American Medical Informatics Association 2006 Annual Symposium Proceedings Biomedical and Health Informatics: From Foundations to Applications to Policy, Washington, DC, DW Bates, JH Holmes, and G. Kuperman, editors %8 2006/// %G eng %0 Journal Article %J Software: Practice and Experience %D 2006 %T An architecture for adaptive intrusion-tolerant applications %A Pal,Partha %A Rubel,Paul %A Atighetchi,Michael %A Webber,Franklin %A Sanders,William H. %A Seri,Mouna %A Ramasamy,HariGovind %A Lyons,James %A Courtney,Tod %A Agbaria,Adnan %A Michel Cukier %A Gossett,Jeanna %A Keidar,Idit %K adaptive defense %K adaptive middleware %K Byzantine fault tolerance %K intrusion tolerance %K redundancy %K survivability architecture %X Applications that are part of a mission-critical information system need to maintain a usable level of key services through ongoing cyber-attacks. In addition to the well-publicized denial of service (DoS) attacks, these networked and distributed applications are increasingly threatened by sophisticated attacks that attempt to corrupt system components and violate service integrity. While various approaches have been explored to deal with DoS attacks, corruption-inducing attacks remain largely unaddressed. We have developed a collection of mechanisms based on redundancy, Byzantine fault tolerance, and adaptive middleware that help distributed, object-based applications tolerate corruption-inducing attacks. In this paper, we present the ITUA architecture, which integrates these mechanisms in a framework for auto-adaptive intrusion-tolerant systems, and we describe our experience in using the technology to defend a critical application that is part of a larger avionics system as an example. We also motivate the adaptive responses that are key to intrusion tolerance, and explain the use of the ITUA architecture to support them in an architectural framework. Copyright © 2006 John Wiley & Sons, Ltd. %B Software: Practice and Experience %V 36 %P 1331 - 1354 %8 2006/// %@ 1097-024X %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/spe.747/abstract %N 11-12 %R 10.1002/spe.747 %0 Journal Article %J ECOOP 2002—Object-Oriented Programming %D 2006 %T Atomic instructions in java %A Hovemeyer,D. %A Pugh, William %A Spacco,J. %B ECOOP 2002—Object-Oriented Programming %P 5 - 16 %8 2006/// %G eng %0 Report %D 2006 %T Automatic web services composition using Shop2 %A Hendler,J. %A Nau, Dana S. %A Parsia,B. %A Sirin,E. %A Wu,D. %X Semantic markup of Web services will enable the automation of various kinds of tasks, including discovery, composition, and execution of Web services. We describe how an AI planning system (SHOP2) can be used with DAML-S Web service descriptions to automatically compose Web services. %I Department of Computer Science, University of Maryland, College Park %8 2006/// %G eng %U http://www.stormingmedia.us/96/9697/A969744.html %0 Book Section %B DUC 06 Conference ProceedingsDUC 06 Conference Proceedings %D 2006 %T Back to Basics: CLASSY 2006 %A Conroy,John M. %A Schlesinger,Judith D. %A O'Leary, Dianne P. %A Goldstein,Jade %X The IDA/CCS summarization system, CLASSY, underwent significant change for this year’sDUC. Two changes made processing simpler and faster: 1) we eliminated the use of a POS (part of speech) tagger for sentence splitting and to assist sentence trimming, and 2) we simplified the scoring of sentences for inclusion in the summary by introducing a new “approximate oracle” score. An additional change introduced a modest amount of extra computation: we ordered sentences in the summary using a new Traveling Salesperson (TSP) formulation. These changes improved ROUGE scores on the DUC 2005 data from last year and gave strong performance in the DUC 2006 competition. %B DUC 06 Conference ProceedingsDUC 06 Conference Proceedings %I U.S. National Inst. of Standards and Technology %8 2006/// %G eng %U \tt http://duc.nist.gov/ %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2006 %T Balancing Systematic and Flexible Exploration of Social Networks %A Perer,A. %A Shneiderman, Ben %K Aggregates %K algorithms %K attribute ranking %K Cluster Analysis %K Computer Graphics %K Computer simulation %K Coordinate measuring machines %K coordinated views %K Data analysis %K data visualisation %K Data visualization %K exploratory data analysis %K Filters %K Gain measurement %K graph theory %K Graphical user interfaces %K Information Storage and Retrieval %K interactive graph visualization %K matrix algebra %K matrix overview %K Models, Biological %K Navigation %K network visualization %K Pattern analysis %K Population Dynamics %K Social Behavior %K social network analysis %K Social network services %K social networks %K social sciences computing %K Social Support %K SocialAction %K software %K statistical analysis %K statistical methods %K User-Computer Interface %X Social network analysis (SNA) has emerged as a powerful method for understanding the importance of relationships in networks. However, interactive exploration of networks is currently challenging because: (1) it is difficult to find patterns and comprehend the structure of networks with many nodes and links, and (2) current systems are often a medley of statistical methods and overwhelming visual output which leaves many analysts uncertain about how to explore in an orderly manner. This results in exploration that is largely opportunistic. Our contributions are techniques to help structural analysts understand social networks more effectively. We present SocialAction, a system that uses attribute ranking and coordinated views to help users systematically examine numerous SNA measures. Users can (1) flexibly iterate through visualizations of measures to gain an overview, filter nodes, and find outliers, (2) aggregate networks using link structure, find cohesive subgroups, and focus on communities of interest, and (3) untangle networks by viewing different link types separately, or find patterns across different link types using a matrix overview. For each operation, a stable node layout is maintained in the network visualization so users can make comparisons. SocialAction offers analysts a strategy beyond opportunism, as it provides systematic, yet flexible, techniques for exploring social networks %B IEEE Transactions on Visualization and Computer Graphics %V 12 %P 693 - 700 %8 2006/10//Sept %@ 1077-2626 %G eng %N 5 %R 10.1109/TVCG.2006.122 %0 Conference Paper %D 2006 %T Bio-inspired, Modular, and Multifunctional Thermal and Impact Protected (TIPed) Embedded Sensing Controls Actuation Power Element (ESCAPE) Structures %A Gyger Jr,L. S. %A Spranklin,B. W. %A Gupta,S.K. %A Bruck,H. A. %X There is a great deal of interest in creating new structural concepts, such as snake robots, from analogousbiological systems. These biological systems are typically composed of materials that serve multiple functions. For example, the skin of a snake is a system consisting of a soft, hyperelastic covering, with nerves providing thermal and pressure sensitivity, and a hard, scaly coating to resist wear and tear from hard particles such as sand. Therefore, bio-inspired structures can also be composed of multifunctional materials. There are currently many approaches to fabricating bio-inspired structures with conventional materials. These concepts typically utilize a system-of-systems design approach where the structure will be composed of a structural framework, an actuation system, a power source for the actuation system, controllers, and external sensors. The drawbacks to these structures include the need to assemble all of the components, interfaces between components that can compromise reliability, constraints on scaling down the structure, and very high power consumption and power generation requirements. These problems have been addressed by developing a new scalable approach to creating multifunctional materials for bio-inspired structures where the controllers, actuators, and sensors are integrated into a modular, multifunctional structure. This approach is facilitated by using multi-material multi-stage molding processes with fully embedded electronic systems, and has resulted in new Thermal and Impact Protected (TIPed) Embedded Sensing Controls Actuation Power Element (ESCAPE) Structures for compact and rugged robotic applications. %8 2006/// %G eng %U http://sem-proceedings.com/06s/sem.org-2006-SEM-Ann-Conf-s89p07-Bio-inspired-Modular-Multifunctional-Thermal-Impact-Protected-TIPed.pdf %0 Journal Article %J SIAM Journal on Scientific Computing %D 2006 %T Block preconditioners based on approximate commutators %A Elman, Howard %A Howle, V. E %A Shadid, J. %A Shuttleworth, R. %A Tuminaro, R. %B SIAM Journal on Scientific Computing %V 27 %P 1651 - 1668 %8 2006/// %G eng %N 5 %0 Journal Article %J Geoinformatica %D 2006 %T Building and querying a P2P virtual world %A Tanin,E. %A Harwood,A. %A Samet, Hanan %A Nayar,D. %A Nutanong,S. %X Peer-to-Peer (P2P) systems are known to provide excellent scalability in a networked environment. One peer is introduced to the system by each participant. However current P2P applications can only provide file sharing and other forms of relatively simple data communications, and, in this paper, we demonstrate how this limitation can be bridged by indexing and querying a 3D virtual-world on a dynamic distributed network. We present an algorithm for 3D range queries as well as an algorithm for nearest neighbor queries. We also show how to build such a complex application from the ground level of a P2P routing algorithm. %B Geoinformatica %V 10 %P 91 - 116 %8 2006/// %G eng %N 1 %R 10.1007/s10707-005-4887-8 %0 Journal Article %J Engineering with Computers %D 2006 %T Building complex coupled physical simulations on the grid with InterComm %A Sussman, Alan %B Engineering with Computers %V 22 %P 311 - 323 %8 2006/// %G eng %N 3 %0 Conference Paper %B Global Telecommunications Conference, 2006. GLOBECOM '06. IEEE %D 2006 %T CAM04-6: Single-Path Routing of Time-varying Traffic %A Kashyap,A. %A Bhattacharjee, Bobby %A La,R. %A Shayman,M. %A Tabatabaee,V. %K algorithm;linear %K algorithm;time-varying %K algorithms;telecommunication %K bound;IP %K case %K complexity;iterative %K intra-domain %K IP %K methods;linear %K multipath %K network %K network;ISP %K networks;Internet;computational %K optimal %K performance %K problem;heuristic %K profile;worst %K programming;optimal %K programming;probability;randomised %K rounding %K routing;iterated %K routing;probability;randomized %K routing;telecommunication %K single-path %K topology;NP-hard %K topology;telecommunication %K traffic %K traffic; %X We consider the problem of finding a single-path intra-domain routing for time-varying traffic. We characterize the traffic variations by a finite set of traffic profiles with given non-zero fractions of occurrence. Our goal is to optimize the average performance over all of these traffic profiles. We solve the optimal multi-path version of this problem using linear programming and develop heuristic single-path solutions using randomized rounding and iterated rounding. We analyze our single-path heuristic (finding the optimal single-path routing is NP-hard), and prove that the randomized rounding algorithm has a worst case performance bound of O(log(KN)/log(log(KN))) compared to the optimal multi-path routing with a high probability, where K is the number of traffic profiles, and N the number of nodes in the network. Further, our simulations show the iterated rounding heuristics perform close to the optimal multi-path routing on a wide range of measured ISP topologies, in both the average and the worst-case. Overall, these results are extremely positive since they show that in a wide-range of practical situations, it is not necessary to deploy multi-path routing; instead, an appropriately computed single-path routing is sufficient to provide good performance. %B Global Telecommunications Conference, 2006. GLOBECOM '06. IEEE %P 1 - 6 %8 2006/12/27/1 %G eng %R 10.1109/GLOCOM.2006.29 %0 Conference Paper %B Proceedings of the 2006 conference on Interaction design and children %D 2006 %T A case study of tangible flags: a collaborative technology to enhance field trips %A Chipman,G. %A Druin, Allison %A Beer,D. %A Fails,J. A %A Guha,M.L. %A Simms,S. %B Proceedings of the 2006 conference on Interaction design and children %P 1 - 8 %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 6th ACM/IEEE-CS joint conference on Digital libraries %D 2006 %T Categorizing web search results into meaningful and stable categories using fast-feature techniques %A Kules,Bill %A Kustanowitz,Jack %A Shneiderman, Ben %K Browsing %K categorization %K classification %K metadata %K open directory %K taxonomies %X When search results against digital libraries and web resources have limited metadata, augmenting them with meaningful and stable category information can enable better overviews and support user exploration. This paper proposes six fast-feature techniques that use only features available in the search result list, such as title, snippet, and URL, to categorize results into meaningful categories. They use credible knowledge resources, including a US government organizational hierarchy, a thematic hierarchy from the Open Directory Project (ODP) web directory, and personal browse histories, to add valuable metadata to search results. In three tests the percent of results categorized for five representative queries was high enough to suggest practical benefits: general web search (76-90%), government web search (39-100%), and the Bureau of Labor Statistics website (48-94%). An additional test submitted 250 TREC queries to a search engine and successfully categorized 66% of the top 100 using the ODP and 61% of the top 350. Fast-feature techniques have been implemented in a prototype search engine. We propose research directions to improve categorization rates and make suggestions about how web site designers could re-organize their sites to support fast categorization of search results. %B Proceedings of the 6th ACM/IEEE-CS joint conference on Digital libraries %S JCDL '06 %I ACM %C New York, NY, USA %P 210 - 219 %8 2006/// %@ 1-59593-354-9 %G eng %U http://doi.acm.org/10.1145/1141753.1141801 %R 10.1145/1141753.1141801 %0 Journal Article %J AIAA Paper %D 2006 %T Characterization of the turbulence structure in supersonic boundary layers using DNS data %A Ringuette,M. J %A Martin, M.P %A Smits,A. J. %A Wu,M. %B AIAA Paper %V 3539 %8 2006/// %G eng %0 Journal Article %J J. Am. Soc. Inf. Sci. Technol. %D 2006 %T Classifying science: Phenomena, data, theory, method, practice: Book Reviews %A Perer,Adam %A Shneiderman, Ben %A Oard, Douglas %K corpus_analysis %K email %K hci %K project--email %K text_analysis %K Visualization %X Due to e-mail's ubiquitous nature, millions of users are intimate with the technology; however, most users are only familiar with managing their own e-mail, which is an inherently different task from exploring an e-mail archive. Historians and social scientists believe that e-mail archives are important artifacts for understanding the individuals and communities they represent. To understand the conversations evidenced in an archive, context is needed. In this article, we present a new way to gain this necessary context: analyzing the temporal rhythms of social relationships. We provide methods for constructing meaningful rhythms from the e-mail headers by identifying relationships and interpreting their attributes. With these visualization techniques, e-mail archive explorers can uncover insights that may have been otherwise hidden in the archive. We apply our methods to an individual's 15-year e-mail archive, which consists of about 45,000 messages and over 4,000 relationships. © 2006 Wiley Periodicals, Inc. %B J. Am. Soc. Inf. Sci. Technol. %V 57 %P 1977 - 1978 %8 2006/12// %@ 1532-2882 %G eng %U http://dx.doi.org/10.1002/asi.v57:14 %N 14 %0 Book Section %B Multi-Lingual Summarization Evaluation 2006Multi-Lingual Summarization Evaluation 2006 %D 2006 %T CLASSYArabic and English Multi-Document Summarization %A Conroy,John M. %A O'Leary, Dianne P. %A Schlesinger,Judith D. %B Multi-Lingual Summarization Evaluation 2006Multi-Lingual Summarization Evaluation 2006 %8 2006/// %G eng %U http://www.isi.edu/ cyl/MTSE2006/MSE2006/papers/index.htmlhttp://www.isi.edu/ cyl/MTSE2006/MSE2006/papers/index.html %0 Journal Article %J IEEE Infocom %D 2006 %T A client-driven approach for channel management in wireless LANs %A Mishra,A. %A Brik,V. %A Banerjee,S. %A Srinivasan, Aravind %A Arbaugh, William A. %X We propose an efficient client-based approach for channel man-agement (channel assignment and load balancing) in 802.11-based WLANs that lead to better usage of the wireless spectrum. This approach is based on a “conflict set coloring” formulation that jointly performs load balancing along with channel assignment. Such a formulation has a number of advantages. First, it explicitly captures interference effects at clients. Next, it intrinsically ex- poses opportunities for better channel re-use. Finally, algorithms based on this formulation do not depend on specific physical RF models and hence can be applied efficiently to a wide-range of in- building as well as outdoor scenarios. We have performed extensive packet-level simulations and mea- surements on a deployed wireless testbed of 70 APs to validate the performance of our proposed algorithms. We show that in addi- tion to single network scenarios, the conflict set coloring formu- lation is well suited for channel assignment where multiple wire- less networks share and contend for spectrum in the same physical space. Our results over a wide range of both simulated topologies and in-building testbed experiments indicate that our approach improves application level performance at the clients by upto three times (and atleast 50%) in comparison to current best-known tech- niques. %B IEEE Infocom %V 6 %8 2006/// %G eng %0 Journal Article %J SIGMOBILE Mob. Comput. Commun. Rev. %D 2006 %T Client-driven channel management for wireless LANs %A Mishra,Arunesh %A Brik,Vladimir %A Banerjee,Suman %A Srinivasan, Aravind %A Arbaugh, William A. %X With the explosive growth in the density of 802.11 access points (APs) in the form of hotspots and public/home access networks, coordinating the shared use of spectrum has become an important problem. The irregular coverage topologies present in WLANs due to the vagaries of the indoor RF environment make the channel assignment algorithms in cellular networks inapplicable [1, 2]. %B SIGMOBILE Mob. Comput. Commun. Rev. %V 10 %P 8 - 10 %8 2006/10// %@ 1559-1662 %G eng %U http://doi.acm.org/10.1145/1215976.1215981 %N 4 %R 10.1145/1215976.1215981 %0 Conference Paper %B Proceedings of the seventeenth annual ACM-SIAM symposium on Discrete algorithm %D 2006 %T Combination can be hard: Approximability of the unique coverage problem %A Demaine,E. D %A Hajiaghayi, Mohammad T. %A Feige,U. %A Salavatipour,M. R %B Proceedings of the seventeenth annual ACM-SIAM symposium on Discrete algorithm %P 162 - 171 %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 2006 IEEE/ACM/IFIP Workshop on Embedded Systems for Real Time Multimedia %D 2006 %T A Communication Interface for Multiprocessor Signal Processing Systems %A Sankalita Saha %A Bhattacharyya, Shuvra S. %A Wayne Wolf %K application program interfaces %K Application software %K coarse-grain dataflow modeling %K Computer applications %K Concurrent computing %K Context %K data flow graphs %K Embedded computing %K Embedded software %K face detection system %K Hardware %K Message passing %K message passing interface %K MPI %K Multiprocessing systems %K multiprocessor signal processing system %K PARALLEL PROCESSING %K signal passing interface %K Signal processing %K SPI %X Parallelization of embedded software is often desirable for power/performance-related considerations for computation-intensive applications that frequently occur in the signal-processing domain. Although hardware support for parallel computation is increasingly available in embedded processing platforms, there is a distinct lack of effective software support. One of the most widely known efforts in support of parallel software is the message passing interface (MPI). However, MPI suffers from several drawbacks with regards to customization to specialized parallel processing contexts, and performance degradation for communication-intensive applications. In this paper, we propose a new interface, the signal passing interface (SPI), that is targeted toward signal processing applications and addresses the limitations of MPI for this important domain of embedded software by integrating relevant properties of MPI and coarse-grain dataflow modeling. SPI is much easier and more intuitive to use, and due to its careful specialization, more performance-efficient for the targeted application domain. We present our preliminary version of SPI, along with experiments using SPI on a practical face detection system that demonstrate the capabilities of SPI %B Proceedings of the 2006 IEEE/ACM/IFIP Workshop on Embedded Systems for Real Time Multimedia %I IEEE %P 127 - 132 %8 2006/10// %@ 0-7803-9783-5 %G eng %R 10.1109/ESTMED.2006.321285 %0 Journal Article %J ACM SIGCOMM Computer Communication Review %D 2006 %T Community-oriented network measurement infrastructure (CONMI) workshop report %A claffy,kc %A Crovella,Mark %A Friedman,Timur %A Shannon,Colleen %A Spring, Neil %B ACM SIGCOMM Computer Communication Review %V 36 %P 41 - 41 %8 2006/04/28/ %@ 01464833 %G eng %U http://dl.acm.org/citation.cfm?id=1129594 %R 10.1145/1129582.1129594 %0 Conference Paper %B Combinatorial Pattern Matching %D 2006 %T A compact mathematical programming formulation for DNA motif finding %A Kingsford, Carl %A Zaslavsky,E. %A Singh,M. %B Combinatorial Pattern Matching %P 233 - 245 %8 2006/// %G eng %0 Conference Paper %B Information Sciences and Systems, 2006 40th Annual Conference on %D 2006 %T Component Forensics of Digital Cameras: A Non-Intrusive Approach %A Swaminathan,A. %A Wu,M. %A Liu,K. J.R %K analysis;image %K approach;processing %K array %K camera;image %K Color %K colour %K detection;optical %K filter %K filters; %K forensics;detecting %K identification;intellectual %K infringement;digital %K interpolation;component %K intrusive %K module;cameras;image %K pattern;color %K property %K protection;non %K right %K sensors;interpolation;object %K tampering %K technology %X This paper considers the problem of component forensics and proposes a methodology to identify the algorithms and parameters employed by various processing modules inside a digital camera. The proposed analysis techniques are non-intrusive, using only sample output images collected from the camera to find the color filter array pattern; and the algorithm and parameters of color interpolation employed in cameras. As demonstrated by various case studies in the paper, the features obtained from component forensic analysis provide useful evidence for such applications as detecting technology infringement, protecting intellectual property rights, determining camera source, and identifying image tampering. %B Information Sciences and Systems, 2006 40th Annual Conference on %P 1194 - 1199 %8 2006/03// %G eng %R 10.1109/CISS.2006.286646 %0 Journal Article %J Computer Networks %D 2006 %T Cooperative peer groups in NICE %A Sherwood,Rob %A Lee,Seungjoon %A Bhattacharjee, Bobby %K distributed algorithms %K P2P %K Reputation base trust %X We present a distributed scheme for trust inference in peer-to-peer networks. Our work is in the context of the NICE system, which is a platform for implementing cooperative applications over the Internet. We describe a technique for efficiently storing user reputation information in a completely decentralized manner, and show how this information can be used to efficiently identify non-cooperative users in NICE. We present a simulation-based study of our algorithms, in which we show our scheme scales to thousands of users using modest amounts of storage, processing, and bandwidth at any individual node. Lastly we show that our scheme is robust and can form cooperative groups in systems where the vast majority of users are malicious. %B Computer Networks %V 50 %P 523 - 544 %8 2006/03/15/ %@ 1389-1286 %G eng %U http://www.sciencedirect.com/science/article/pii/S1389128605002185 %N 4 %R 10.1016/j.comnet.2005.07.012 %0 Conference Paper %B Proceedings of the 23rd international conference on Machine learning %D 2006 %T Cost-sensitive learning with conditional Markov networks %A Sen,Prithviraj %A Getoor, Lise %X There has been a recent, growing interest in classification and link prediction in structured domains. Methods such as CRFs (Lafferty et al., 2001) and RMNs (Taskar et al., 2002) support flexible mechanisms for modeling correlations due to the link structure. In addition, in many structured domains, there is an interesting structure in the risk or cost function associated with different misclassifications. There is a rich tradition of cost-sensitive learning applied to unstructured (IID) data. Here we propose a general framework which can capture correlations in the link structure and handle structured cost functions. We present a novel cost-sensitive structured classifier based on Maximum Entropy principles that directly determines the cost-sensitive classification. We contrast this with an approach which employs a standard 0/1 loss structured classifier followed by minimization of the expected cost of misclassification. We demonstrate the utility of our proposed classifier with experiments on both synthetic and real-world data. %B Proceedings of the 23rd international conference on Machine learning %S ICML '06 %I ACM %C New York, NY, USA %P 801 - 808 %8 2006/// %@ 1-59593-383-2 %G eng %U http://doi.acm.org/10.1145/1143844.1143945 %R 10.1145/1143844.1143945 %0 Journal Article %J International Journal of Human-Computer Interaction %D 2006 %T Creativity Support Tools: Report From a U.S. National Science Foundation Sponsored Workshop %A Shneiderman, Ben %A Fischer,Gerhard %A Czerwinski,Mary %A Resnick,Mitch %A Myers,Brad %A Candy,Linda %A Edmonds,Ernest %A Eisenberg,Mike %A Giaccardi,Elisa %A Hewett,Tom %A Jennings,Pamela %A Kules,Bill %A Nakakoji,Kumiyo %A Nunamaker,Jay %A Pausch,Randy %A Selker,Ted %A Sylvan,Elisabeth %A Terry,Michael %X Creativity support tools is a research topic with high risk but potentially very high payoff. The goal is to develop improved software and user interfaces that empower users to be not only more productive but also more innovative. Potential users include software and other engineers, diverse scientists, product and graphic designers, architects, educators, students, and many others. Enhanced interfaces could enable more effective searching of intellectual resources, improved collaboration among teams, and more rapid discovery processes. These advanced interfaces should also provide potent support in hypothesis formation, speedier evaluation of alternatives, improved understanding through visualization, and better dissemination of results. For creative endeavors that require composition of novel artifacts (e.g., computer programs, scientific papers, engineering diagrams, symphonies, artwork), enhanced interfaces could facilitate exploration of alternatives, prevent unproductive choices, and enable easy backtracking. This U.S. National Science Foundation sponsored workshop brought together 25 research leaders and graduate students to share experiences, identify opportunities, and formulate research challenges. Two key outcomes emerged: (a) encouragement to evaluate creativity support tools through multidimensional in-depth longitudinal case studies and (b) formulation of 12 principles for design of creativity support tools.Creativity support tools is a research topic with high risk but potentially very high payoff. The goal is to develop improved software and user interfaces that empower users to be not only more productive but also more innovative. Potential users include software and other engineers, diverse scientists, product and graphic designers, architects, educators, students, and many others. Enhanced interfaces could enable more effective searching of intellectual resources, improved collaboration among teams, and more rapid discovery processes. These advanced interfaces should also provide potent support in hypothesis formation, speedier evaluation of alternatives, improved understanding through visualization, and better dissemination of results. For creative endeavors that require composition of novel artifacts (e.g., computer programs, scientific papers, engineering diagrams, symphonies, artwork), enhanced interfaces could facilitate exploration of alternatives, prevent unproductive choices, and enable easy backtracking. This U.S. National Science Foundation sponsored workshop brought together 25 research leaders and graduate students to share experiences, identify opportunities, and formulate research challenges. Two key outcomes emerged: (a) encouragement to evaluate creativity support tools through multidimensional in-depth longitudinal case studies and (b) formulation of 12 principles for design of creativity support tools. %B International Journal of Human-Computer Interaction %V 20 %P 61 - 77 %8 2006/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1207/s15327590ijhc2002_1 %N 2 %R 10.1207/s15327590ijhc2002_1 %0 Book Section %B Text, Speech and DialogueText, Speech and Dialogue %D 2006 %T Cross-Language Access to Recorded Speech in the MALACH Project %A Oard, Douglas %A Demner-Fushman,Dina %A Hajič,Jan %A Ramabhadran,Bhuvana %A Gustman,Samuel %A Byrne,William %A Soergel,Dagobert %A Dorr, Bonnie J %A Resnik, Philip %A Picheny,Michael %E Sojka,Petr %E Kopecek,Ivan %E Pala,Karel %X The MALACH project seeks to help users find information in a vast multilingual collections of untranscribed oral history interviews. This paper introduces the goals of the project and focuses on supporting access by users who are unfamiliar with the interview language. It begins with a review of the state of the art in crosslanguage speech retrieval; approaches that will be investigated in the project are then described. Czech was selected as the first non-English language to be supported, so results of an initial experiment with Czech/English cross-language retrieval are reported. %B Text, Speech and DialogueText, Speech and Dialogue %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2448 %P 197 - 212 %8 2006/// %@ 978-3-540-44129-8 %G eng %U http://dx.doi.org/10.1007/3-540-46154-X_8 %0 Conference Paper %B Image Processing, 2006 IEEE International Conference on %D 2006 %T Cross-Path PDMA-Based Error Protection for Streaming Multiuser Video over Multiple Paths %A Su,Guan-Ming %A Chen,Meng %A M. Wu %K access;error %K codes;forward %K coding;multiple %K coding;video %K correction %K correction;multi-access %K division %K error %K error-prone %K forward %K multiplexing %K path;multistream %K PDMA;cross-path %K Protection %K scheme;multiuser %K streaming; %K streaming;packet %K systems;video %K video %X In this paper, we consider aggregating multiple video streams and transmitting the merged stream over multiple error-prone paths. We propose a novel multi-stream error protection scheme based on packet division multiplexing access (PDMA) and cross-path forward error coding. Compared with the traditional time division multiplexing access (TDMA)-based error protection scheme, the proposed scheme outperforms 1.43 1.88 dB for the averaged PSNR of all received streams %B Image Processing, 2006 IEEE International Conference on %P 21 - 24 %8 2006/10// %G eng %R 10.1109/ICIP.2006.313151 %0 Journal Article %J Journal of Parallel and Distributed Computing %D 2006 %T Data redistribution and remote method invocation for coupled components %A Bertrand,Felipe %A Bramley,Randall %A Bernholdt,David E. %A Kohl,James A. %A Sussman, Alan %A Larson,Jay W. %A Damevski,Kostadin B. %K Common component architecture %K Model coupling %K Parallel data redistribution %K Parallel remote method invocation %X With the increasing availability of high-performance massively parallel computer systems, the prevalence of sophisticated scientific simulation has grown rapidly. The complexity of the scientific models being simulated has also evolved, leading to a variety of coupled multi-physics simulation codes. Such cooperating parallel programs require fundamentally new interaction capabilities, to efficiently exchange parallel data structures and collectively invoke methods across programs. So-called “ M × N ” research, as part of the common component architecture (CCA) effort, addresses these special and challenging needs, to provide generalized interfaces and tools that support flexible parallel data redistribution and parallel remote method invocation. Using this technology, distinct simulation codes with disparate distributed data decompositions can work together to achieve greater scientific discoveries. %B Journal of Parallel and Distributed Computing %V 66 %P 931 - 946 %8 2006/07// %@ 0743-7315 %G eng %U http://www.sciencedirect.com/science/article/pii/S0743731506000700 %N 7 %R 10.1016/j.jpdc.2005.12.009 %0 Conference Paper %B Proceedings of the International Symposium on System-on-Chip, Tampere, Finland %D 2006 %T Dataflow transformations in high-level DSP system design %A Saha,S. %A Puthenpurayil,S. %A Bhattacharyya, Shuvra S. %B Proceedings of the International Symposium on System-on-Chip, Tampere, Finland %P 131 - 136 %8 2006/// %G eng %0 Conference Paper %B Visual Analytics Science And Technology, 2006 IEEE Symposium On %D 2006 %T D-Dupe: An Interactive Tool for Entity Resolution in Social Networks %A Bilgic,M. %A Licamele,L. %A Getoor, Lise %A Shneiderman, Ben %K algorithm;data %K analysis;social %K computing;user %K D-Dupe %K interactive %K interface;data %K interfaces; %K mining %K mining;interactive %K network %K problem;entity %K QUALITY %K resolution;entity-resolution;social %K sciences %K systems;social %K tool;data %K visual %K visualization;task-specific %X Visualizing and analyzing social networks is a challenging problem that has been receiving growing attention. An important first step, before analysis can begin, is ensuring that the data is accurate. A common data quality problem is that the data may inadvertently contain several distinct references to the same underlying entity; the process of reconciling these references is called entity-resolution. D-Dupe is an interactive tool that combines data mining algorithms for entity resolution with a task-specific network visualization. Users cope with complexity of cleaning large networks by focusing on a small subnetwork containing a potential duplicate pair. The subnetwork highlights relationships in the social network, making the common relationships easy to visually identify. D-Dupe users resolve ambiguities either by merging nodes or by marking them distinct. The entity resolution process is iterative: as pairs of nodes are resolved, additional duplicates may be revealed; therefore, resolution decisions are often chained together. We give examples of how users can flexibly apply sequences of actions to produce a high quality entity resolution result. We illustrate and evaluate the benefits of D-Dupe on three bibliographic collections. Two of the datasets had already been cleaned, and therefore should not have contained duplicates; despite this fact, many duplicates were rapidly identified using D-Dupe's unique combination of entity resolution algorithms within a task-specific visual interface %B Visual Analytics Science And Technology, 2006 IEEE Symposium On %P 43 - 50 %8 2006/11/31/2 %G eng %R 10.1109/VAST.2006.261429 %0 Conference Paper %B Proceedings of the ACM/IFIP/USENIX 2006 International Conference on Middleware %D 2006 %T Decentralized message ordering for publish/subscribe systems %A Lumezanu,Cristian %A Spring, Neil %A Bhattacharjee, Bobby %X We describe a method to order messages across groups in a publish/subscribe system without centralized control or large vector timestamps. We show that our scheme is practical---little state is required; that it is scalable---the maximum message load is limited by receivers; and that it performs well---the paths messages traverse to be ordered are not made much longer than necessary. Our insight is that only messages to groups that overlap in membership can be observed to arrive out of order: sequencing messages to these groups is sufficient to provide a consistent order, and when publishers subscribe to the groups to which they send, this message order is a causal order. %B Proceedings of the ACM/IFIP/USENIX 2006 International Conference on Middleware %S Middleware '06 %I Springer-Verlag New York, Inc. %C New York, NY, USA %P 162 - 179 %8 2006/// %G eng %U http://dl.acm.org/citation.cfm?id=1515984.1515997 %0 Journal Article %J Journal of the ACM %D 2006 %T Dependent rounding and its applications to approximation algorithms %A Gandhi,Rajiv %A Khuller, Samir %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %K broadcast scheduling %K Randomized rounding %X We develop a new randomized rounding approach for fractional vectors defined on the edge-sets of bipartite graphs. We show various ways of combining this technique with other ideas, leading to improved (approximation) algorithms for various problems. These include:---low congestion multi-path routing;---richer random-graph models for graphs with a given degree-sequence;---improved approximation algorithms for: (i) throughput-maximization in broadcast scheduling, (ii) delay-minimization in broadcast scheduling, as well as (iii) capacitated vertex cover; and---fair scheduling of jobs on unrelated parallel machines. %B Journal of the ACM %V 53 %P 324 - 360 %8 2006/05// %@ 0004-5411 %G eng %U http://doi.acm.org/10.1145/1147954.1147956 %N 3 %R 10.1145/1147954.1147956 %0 Journal Article %J Learning & MemoryLearn. Mem. %D 2006 %T Differential Transcriptional Response to Nonassociative and Associative Components of Classical Fear Conditioning in the Amygdala and Hippocampus %A Keeley,Michael B %A Wood,Marcelo A %A Isiegas,Carolina %A Stein,Joel %A Hellman,Kevin %A Hannenhalli, Sridhar %A Abel,Ted %X Classical fear conditioning requires the recognition of conditioned stimuli (CS) and the association of the CS with an aversive stimulus. We used Affymetrix oligonucleotide microarrays to characterize changes in gene expression compared to naive mice in both the amygdala and the hippocampus 30 min after classical fear conditioning and 30 min after exposure to the CS in the absence of an aversive stimulus. We found that in the hippocampus, levels of gene regulation induced by classical fear conditioning were not significantly greater than those induced by CS alone, whereas in the amygdala, classical fear conditioning did induce significantly greater levels of gene regulation compared to the CS. Computational studies suggest that transcriptional changes in the hippocampus and amygdala are mediated by large and overlapping but distinct combinations of molecular events. Our results demonstrate that an increase in gene regulation in the amygdala was partially correlated to associative learning and partially correlated to nonassociative components of the task, while gene regulation in the hippocampus was correlated to nonassociative components of classical fear conditioning, including configural learning. %B Learning & MemoryLearn. Mem. %V 13 %P 135 - 142 %8 2006/03/01/ %@ 1072-0502, 1549-5485 %G eng %U http://learnmem.cshlp.org/content/13/2/135 %N 2 %R 10.1101/lm.86906 %0 Journal Article %J B-EYE-Network-Boulder, CO, USA, April %D 2006 %T Discovering business intelligence using treemap visualizations %A Shneiderman, Ben %B B-EYE-Network-Boulder, CO, USA, April %V 11 %8 2006/// %G eng %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2006. IPDPS 2006. 20th International %D 2006 %T DiST: fully decentralized indexing for querying distributed multidimensional datasets %A Nam,Beomseok %A Sussman, Alan %K Computer network management %K distributed multidimensional dataset querying %K failure recovery %K fault tolerant computing %K fully decentralized multidimensional indexing %K grid computing %K Indexing %K large scale distributed resource management %K Large-scale systems %K Multidimensional systems %K Network servers %K P2P systems %K Peer to peer computing %K peer-to-peer computing %K peer-to-peer systems %K Publishing %K Query processing %K query routing %K resource allocation %K Resource management %K telecommunication network routing %K wide area networks %X Grid computing and peer-to-peer (P2P) systems are emerging as new paradigms for managing large scale distributed resources across wide area networks. While grid computing focuses on managing heterogeneous resources and relies on centralized managers for resource and data discovery, P2P systems target scalable, decentralized methods for publishing and searching for data. In large distributed systems, a centralized resource manager is a potential performance bottleneck and decentralization can help avoid this bottleneck, as is done in P2P systems. However, the query functionality provided by most existing P2P systems is very rudimentary, and is not directly applicable to grid resource management. In this paper, we propose a fully decentralized multidimensional indexing structure, called DiST, that operates in a fully distributed environment with no centralized control. In DiST, each data server only acquires information about data on other servers from executing and routing queries. We describe the DiST algorithms for maintaining the decentralized network of data servers, including adding and deleting servers, the query routing algorithm, and failure recovery algorithms. We also evaluate the performance of the decentralized scheme against a more structured hierarchical indexing scheme that we have previously shown to perform well in distributed grid environments %B Parallel and Distributed Processing Symposium, 2006. IPDPS 2006. 20th International %I IEEE %8 2006/04// %@ 1-4244-0054-6 %G eng %R 10.1109/IPDPS.2006.1639280 %0 Conference Paper %B Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems %D 2006 %T Distance join queries on spatial networks %A Sankaranarayanan,Jagan %A Alborzi,Houman %A Samet, Hanan %K location-based services %K path coherence %K Query processing %K SILC framework %K Spatial databases %K spatial networks %X The result of a distance join operation on two sets of objects R, S on a spatial network G is a set P of object pairs pq, p É R, q É S such that the distance of an object pair pq is the shortest distance from p to q in G. Several variations to the distance join operation such as UnOrdered, Incremental, topk, Semi-Join impose additional constraints on the distance between the object pairs in P, the ordering of object pairs in P, and on the cardinality of P. A distance join algorithm on spatial networks is proposed that works in conjunction with the SILC framework, which is a new approach to query processing on spatial networks. Experimental results demonstrate up to an order of magnitude speed up when compared with a prominent existing technique. %B Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems %S GIS '06 %I ACM %C New York, NY, USA %P 211 - 218 %8 2006/// %@ 1-59593-529-0 %G eng %U http://doi.acm.org/10.1145/1183471.1183506 %R 10.1145/1183471.1183506 %0 Conference Paper %B International Conference on Document Recognition and Retrieval XIII %D 2006 %T DOCLIB: a Software Library for Document Processing %A Jaeger,Stefan %A Zhu,Guangyu %A David Doermann %A Chen,Kevin %A Sampat,Summit %X Most researchers would agree that research in the field of document processing can benefit tremendously from a common software library through which institutions are able to develop and share research-related software and applications across academic, business, and government domains. However, despite several attempts in the past, the research community still lacks a widely-accepted standard software library for document processing. This paper describes a new library called DOCLIB, which tries to overcome the drawbacks of earlier approaches. Many of DOCLIB’s features are unique either in themselves or in their combination with others, e.g. the factory concept for support of different image types, the juxtaposition of image data and metadata, or the add-on mechanism. We cherish the hope that DOCLIB serves the needs of researchers better than previous approaches and will readily be accepted by a larger group of scientists. %B International Conference on Document Recognition and Retrieval XIII %I San Jose, CA %P 1 - 9 %8 2006/// %G eng %0 Patent %D 2006 %T Dwarf cube architecture for reducing storage sizes of multidimensional data %A Roussopoulos, Nick %A Sismanis,John %A Deligiannakis,Antonios %E The University of Maryland College Park %X The invention relates to data warehouses and the ability to create and maintain data cubes of multi-dimensional data. More specifically, the invention pertains to data cube architectures that permit significant reduction of storage, exhibit very efficient retrieval and provide a very efficient incremental update of the data cubes. %V 10/157,960 %8 2006/11/07/ %G eng %U http://www.google.com/patents?id=PGl7AAAAEBAJ %N 7133876 %0 Journal Article %J Applied and environmental microbiology %D 2006 %T Effect of transport at ambient temperature on detection and isolation of Vibrio cholerae from environmental samples %A Alam,M. %A Sadique,A. %A Bhuiyan,N. A. %A Nair,G. B. %A Siddique,A. K. %A Sack,D. A. %A Ahsan,S. %A Huq,A. %A Sack,R. B. %A Rita R Colwell %A others %X It has long been assumed that prolonged holding of environmental samples at the ambient air temperature prior to bacteriological analysis is detrimental to isolation and detection of Vibrio cholerae, the causative agent of pandemic cholera. The present study was aimed at understanding the effect of transporting environmental samples at the ambient air temperature on isolation and enumeration of V. cholerae. For water and plankton samples held at ambient temperatures ranging from 31°C to 35°C for 20 h, the total counts did not increase significantly but the number of culturable V. cholerae increased significantly compared to samples processed within 1 h of collection, as measured by culture, acridine orange direct count, direct fluorescent-antibody-direct viable count (DFA-DVC), and multiplex PCR analyses. For total coliform counts, total bacterial counts, and DFA-DVC counts, the numbers did not increase significantly, but the culturable plate counts for V. cholerae increased significantly after samples were held at the ambient temperature during transport to the laboratory for analysis. An increase in the recovery of V. cholerae O1 and improved detection of V. cholerae O1 rfb and ctxA also occurred when samples were enriched after they were kept for 20 h at the ambient temperature during transport. Improved detection and isolation of toxigenic V. cholerae from freshwater ecosystems can be achieved by holding samples at the ambient temperature, an observation that has significant implications for tracking this pathogen in diverse aquatic environments. %B Applied and environmental microbiology %V 72 %P 2185 - 2190 %8 2006/// %G eng %N 3 %0 Report %D 2006 %T Efficient Isosurface Extraction for Large Scale Time-Varying Data Using the Persistent Hyperoctree (PHOT) %A Shi,Qingmin %A JaJa, Joseph F. %K Technical Report %X We introduce the Persistent HyperOcTree (PHOT) to handle the4D isocontouring problem for large scale time-varying data sets. This novel data structure is provably space efficient and optimal in retrieving active cells. More importantly, the set of active cells for any possible isovalue are already organized in a Compact Hyperoctree, which enables very efficient slicing of the isocontour along spatial and temporal dimensions. Experimental results based on the very large Richtmyer-Meshkov instability data set demonstrate the effectiveness of our approach. This technique can also be used for other isosurfacing schemes such as view-dependent isosurfacing and ray-tracing, which will benefit from the inherent hierarchical structure associated with the active cells. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2006-01 %8 2006/01/13/T20:3 %G eng %U http://drum.lib.umd.edu/handle/1903/3035 %0 Conference Paper %B Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems %D 2006 %T Efficient position-independent iconic search using an R-theta index %A Cranston,Charles B. %A Samet, Hanan %K iconic database %K position-independent %K spatial indexing %X An iconic image database is a collection of symbolic images where each image is a collection of labeled point features called icons. A method is presented to support fast position-independent similarity search in an iconic database for symbolic images where the similarity condition involves finding icon pairs that satisfy a specific spatial relationship. This is achieved by introducing an index data structure based on r-θ space, which corresponds to the Cartesian product of separation (i.e., inter-icon distance) and (some representation of) relative spatial orientation. In this space, each pairing of two icons is represented by a single point, and all pairs with the same separation and relative orientation (regardless of absolute position) map to the same point. Similarly, all icon pairs with the same separation but different relative orientations map to points on a line parallel to the θ axis, while all pairs with different separations but the same relative orientation map to points on a line parallel to the r axis. Using such an index, database search for icon pairs with a given spatial relationship or range is accomplished by examining the subarea of the index space into which desired pairs would map. This r-θ index space can be organized using well-known spatial database techniques, such as quadtrees or R-trees. Although the size of such an index grows only linearly with respect to the number of images in the collection, it grows quadratically with the average number of icons in an image. A scheme is described to reduce the size of the index by pruning away a subset of the pairs, at the cost of incurring additional work when searching the database. This pruning is governed by a parameter φ, whose variation provides a continuous range of trade-offs between index size and search time. %B Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems %S GIS '06 %I ACM %C New York, NY, USA %P 27 - 34 %8 2006/// %@ 1-59593-529-0 %G eng %U http://doi.acm.org/10.1145/1183471.1183478 %R 10.1145/1183471.1183478 %0 Report %D 2006 %T EIGENTEST: A Test Matrix Generator for Large-Scale Eigenproblems %A Lee,Che Rung %A Stewart, G.W. %K Technical Report %X Eigentest is a package that produces real test matrices with knowneigensystems. A test matrix, called an eigenmat, is generated in a factored form, in which the user can specify the eigenvalues and has some control over the condition of the eigenvalues and eigenvectors. An eigenmat A of order n requires only O(n) storage for its representation. Auxiliary programs permit the computation of (A - sI)*b, (A - sI)'*b, inv(A - sI)*b, and inv(A - sI)*b in O(n) operations. A special routine computes specified eigenvectors of an eigenmat and the condition of its eigenvalue. Thus eigenmats are suitable for testing algorithms based on Krylov sequences, as well as others based on matrix-vector products. This paper introduces the eigenmat and describes implementations in Fortran~77, Fortran~95, C, and Matlab. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2006-07 %8 2006/02/13/T19:4 %G eng %U http://drum.lib.umd.edu/handle/1903/3328 %0 Journal Article %J Open Problems in Statistical Relational Learning: Papers from the ICML Workshop. Pittsburgh, PA: www. cs. umd. edu/projects/srl2006 %D 2006 %T Empirical comparison of approximate inference algorithms for networked data %A Sen,P. %A Getoor, Lise %X Over the past few years, a number of approx-imate inference algorithms for networked data have been put forth. We empirically compare the performance of three of the popular algorithms: loopy belief propagation, mean field relaxation labeling and iterative classification. We rate each algorithm in terms of its robustness to noise, both in attribute values and correlations across links. A novel observation from our experiments is that loopy belief propagation faces difficulty when in- ferring over data with homophily, a common type of link correlation observed in relational data. %B Open Problems in Statistical Relational Learning: Papers from the ICML Workshop. Pittsburgh, PA: www. cs. umd. edu/projects/srl2006 %8 2006/// %G eng %0 Journal Article %J Journal of Multimedia %D 2006 %T Employing the Hand as an Interface Device %A Sepehri,Afshin %A Yacoob,Yaser %A Davis, Larry S. %X We propose algorithms and applications for using the hand as an interface device in virtual and physical spaces. In virtual drawing, by tracking the hand in 3-D and estimating a virtual plane in space, the intended drawing of user is recognized. In a virtual marble game, the instantaneous orientation of the hand is simulated to render a graphical scene of the game board. Real-time visual feedback allows the user to navigate a virtual ball in a maze. In 3-D model construction, the system tracks the hand motion in space while the user is traversing edges of a physical object. The object is then rendered virtually by the computer. These applications involve estimating the 3-D absolute position and/or orientation of the hand in space. We propose parametric modelling of the central region of the hand to extract this information. A stereo camera is used to first build a preliminary disparity map of the hand. Then, the best fitting plane to the disparity points is computed using robust estimation. The 3-D hand plane is calculated based on the disparity plane and the position and orientation parameters of the hand. Tracking the hand region over a sequence of frames and coping with noise using robust modelling of the hand motion enables estimating the trajectory of the hand in space. The algorithms are real-time and experiments are presented to demonstrate the proposed applications of using the hand as an interface device. %B Journal of Multimedia %V 1 %8 2006/11/01/ %@ 1796-2048 %G eng %U http://www.ojs.academypublisher.com/index.php/jmm/article/viewArticle/01071829 %N 7 %R 10.4304/jmm.1.7.18-29 %0 Journal Article %J Journal of Multimedia %D 2006 %T Employing the hand as an interface device %A Sepehri,A. %A Yacoob,Yaser %A Davis, Larry S. %B Journal of Multimedia %V 1 %P 18 - 29 %8 2006/// %G eng %N 7 %0 Conference Paper %B Data Engineering, 2006. ICDE '06. Proceedings of the 22nd International Conference on %D 2006 %T Enabling Query Processing on Spatial Networks %A Sankaranarayanan,J. %A Alborzi,H. %A Samet, Hanan %X A system that enables real time query processing on large spatial networks is demonstrated. The system provides functionality for processing a wide range of spatial queries such as nearest neighbor searches and spatial joins on spatial networks of sufficiently large sizes. %B Data Engineering, 2006. ICDE '06. Proceedings of the 22nd International Conference on %P 163 - 163 %8 2006/04// %G eng %R 10.1109/ICDE.2006.60 %0 Journal Article %J ICML Workshop on Statistical Network Analysis %D 2006 %T Entity and relationship labeling in affiliation networks %A Zhao,B. %A Sen,P. %A Getoor, Lise %X Many domains are best characterized as anaffiliation network describing a set of actors and a set of events interlinked together in a variety of relationships. Relational clas- sification in these domains requires the col- lective classification of both entities (actors and events) and relationships. We investigate the use of relational Markov networks (RMN) for relational classification in affiliation net- works. In this paper, we introduce a novel dataset, Profile in Terror (PIT) knowledge base, that provides a rich source of various af- filiation networks. We study two tasks, entity labeling and relationship labeling. We high- light several important issues concerning the effectiveness of relational classification. Our results show that the PIT dataset has a rich source of relational structure and therefore it is a useful dataset for statisical relational network learning community. %B ICML Workshop on Statistical Network Analysis %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems %D 2006 %T Entity resolution in geospatial data integration %A Sehgal,Vivek %A Getoor, Lise %A Viechnicki,Peter D %X Due to the growing availability of geospatial data from a wide variety of sources, there is a pressing need for robust, accurate and automatic merging and matching techniques. Geospatial Entity Resolution is the process of determining, from a collection of database sources referring to geospatial locations, a single consolidated collection of 'true' locations. At the heart of this process is the problem of determining when two locations references match---i.e., when they refer to the same underlying location. In this paper, we introduce a novel method for resolving location entities in geospatial data. A typical geospatial database contains heterogeneous features such as location name, spatial coordinates, location type and demographic information. We investigate the use of all of these features in algorithms for geospatial entity resolution. Entity resolution is further complicated by the fact that the different sources may use different vocabularies for describing the location types and a semantic mapping is required. We propose a novel approach which learns how to combine the different features to perform accurate resolutions. We present experimental results showing that methods combining spatial and non-spatial features (e.g., location-name, location-type, etc.) together outperform methods based on spatial or name information alone. %B Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems %S GIS '06 %I ACM %C New York, NY, USA %P 83 - 90 %8 2006/// %@ 1-59593-529-0 %G eng %U http://doi.acm.org/10.1145/1183471.1183486 %R 10.1145/1183471.1183486 %0 Conference Paper %B Seventh Workshop of the Cross-Language Evaluation Forum, CLEF 2006 %D 2006 %T Evaluation of Multilingual and Multi-modal Information Retrieval-Seventh Workshop of the Cross-Language Evaluation Forum, CLEF 2006, Alicante, Spain, September 19-21, 2006, Revised Selected Papers %A Peters,C. %A Clough,P. %A Gey,F. %A Karlgren,J. %A Magnini,B. %A Oard, Douglas %A De Rijke,M. %A Stempfhuber,M. %B Seventh Workshop of the Cross-Language Evaluation Forum, CLEF 2006 %C Alicante, Spain %8 2006/// %G eng %0 Journal Article %J Proceedings of the Workshop on Statistical Network Analysis (SNA) at the 23rd International Conference on Machine Learning (ICML) %D 2006 %T Event classification and relationship labeling in affiliation networks %A Zhao,B. %A Sen,P. %A Getoor, Lise %X Many domains are best described as an affiliation network in which there are entities such as actors, events and organizations linked together in a variety of relationships. Relational classification in these domains requires the collective classification of both entities and relationships. In this paper, we investigate the use of relational Markov networks (RMN) for relational classification in affiliation networks. We study two tasks, event classification and relationship labeling, and discuss general issues in constructing an appropriate RMN from an affiliation network. We evaluate our methods on a novel dataset describing terrorist affiliation networks which includes data about actors (terrorists), events (terrorist attacks) and organizations (terrorist organizations). Our results highlight several important issues concerning the effectiveness of relational classification and our experiments show that the relational structure significantly helps relationship labeling. %B Proceedings of the Workshop on Statistical Network Analysis (SNA) at the 23rd International Conference on Machine Learning (ICML) %8 2006/// %G eng %0 Journal Article %J CTWatch Quarterly %D 2006 %T Experiments to understand HPC time to development %A Hochstein, Lorin %A Nakamura,Taiga %A Basili, Victor R. %A Asgari, Sima %A Zelkowitz, Marvin V %A Hollingsworth, Jeffrey K %A Shull, Forrest %A Carver,Jeffrey %A Voelp,Martin %A Zazworka, Nico %A Johnson,Philip %K hackystat %K HPC %K publications-journals %X In order to understand how high performance computing (HPC) programs are developed, a series of experiments, using students in graduate level HPC classes, have been conducted at many universities in the US. In this paper we discuss the general process of conducting those experiments, give some of the early results of those experiments, and describe a web-based process we are developing that will allow us to run additional experiments at other universities and laboratories that will be easier to conduct and generate results that more accurately reflect the process of building HPC programs. %B CTWatch Quarterly %8 2006/// %G eng %U http://csdl.ics.hawaii.edu/techreports/06-08/06-08.pdf %0 Conference Paper %B Proceedings of the Third Conference on Email and Anti-Spam, Mountain View, CA, July %D 2006 %T An exploratory study of the W3C mailing list test collection for retrieval of emails with pro/con arguments %A Wu,Y. %A Oard, Douglas %A Soboroff,I. %B Proceedings of the Third Conference on Email and Anti-Spam, Mountain View, CA, July %8 2006/// %G eng %0 Journal Article %J Decision Support Systems %D 2006 %T Exploring auction databases through interactive visualization %A Shmueli,Galit %A Jank,Wolfgang %A Aris,Aleks %A Plaisant, Catherine %A Shneiderman, Ben %K Auction dynamics %K Bid history %K Online auctions %K time series %K user interface %X We introduce AuctionExplorer, a suite of tools for exploring databases of online auctions. The suite combines tools for collecting, processing, and interactively exploring auction attributes (e.g., seller rating), and the bid history (price evolution represented as a time series). Part of AuctionExplorer's power comes from its coupling of the two information structures, thereby allowing exploration of relationships between them. Exploration can be directed by hypothesis testing or exploratory data analysis. We propose a process for visual data analysis and illustrate AuctionExplorer's operations with a dataset of eBay auctions. Insights may improve seller, bidder, auction house, and other vendors' understanding of the market, thereby assisting their decision making process. %B Decision Support Systems %V 42 %P 1521 - 1538 %8 2006/12// %@ 0167-9236 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167923606000042 %N 3 %R 10.1016/j.dss.2006.01.001 %0 Conference Paper %B Proceedings of the 6th ACM/IEEE-CS joint conference on Digital libraries %D 2006 %T Exploring erotics in Emily Dickinson's correspondence with text mining and visual interfaces %A Plaisant, Catherine %A Rose,James %A Yu,Bei %A Auvil,Loretta %A Kirschenbaum,Matthew G. %A Smith,Martha Nell %A Clement,Tanya %A Lord,Greg %K case studies %K humanities %K literary criticism %K text mining %K user interface %K Visualization %X This paper describes a system to support humanities scholars in their interpretation of literary work. It presents a user interface and web architecture that integrates text mining, a graphical user interface and visualization, while attempting to remain easy to use by non specialists. Users can interactively read and rate documents found in a digital libraries collection, prepare training sets, review results of classification algorithms and explore possible indicators and explanations. Initial evaluation steps suggest that there is a rationale for "provocational" text mining in literary interpretation. %B Proceedings of the 6th ACM/IEEE-CS joint conference on Digital libraries %S JCDL '06 %I ACM %C New York, NY, USA %P 141 - 150 %8 2006/// %@ 1-59593-354-9 %G eng %U http://doi.acm.org/10.1145/1141753.1141781 %R 10.1145/1141753.1141781 %0 Journal Article %J Journal of Visual Languages & Computing %D 2006 %T Exploring personal media: A spatial interface supporting user-defined semantic regions %A Kang,Hyunmo %A Shneiderman, Ben %K dynamic queries %K Fling-and-flock %K Personal media management %K Spatial information management %K User interfaces %X Graphical mechanisms for spatially organizing personal media data could enable users to fruitfully apply their conceptual models. This paper introduces Semantic regions, an innovative way for users to construct display representations of their conceptual models by drawing regions on 2D space and specifying the semantics for each region. Then users can apply personal categorizations to personal media data using the fling-and-flock metaphor. This allows personal media to be dragged to the spatially organized display and automatically grouped according to time, geography, family trees, groups of friends, or other spatially organized display representations of conceptual models. The prototype implementation for semantic regions, MediaFinder, was refined based on two small usability tests for usage and construction of user-defined conceptual models. %B Journal of Visual Languages & Computing %V 17 %P 254 - 283 %8 2006/06// %@ 1045-926X %G eng %U http://www.sciencedirect.com/science/article/pii/S1045926X06000048 %N 3 %R 10.1016/j.jvlc.2006.01.001 %0 Journal Article %J Proceedings of SPIE %D 2006 %T Exploring QIM-based anti-collusion fingerprinting for multimedia %A Swaminathan,A. %A He,S. %A Wu,M. %X Digital fingerprinting is an emerging technology to protect multimedia from unauthorizeduse by embedding a unique fingerprint signal into each user’s copy. A robust embedding algorithm is an important building block in order to make the fingerprint resilient to vari- ous distortions and collusion attacks. Spread spectrum embedding has been widely used for multimedia fingerprinting. In this paper, we explore another class of embedding methods – Quantization Index Modulation (QIM) for fingerprinting applications. We first employ Dither Modulation (DM) technique and extend it for embedding multiple symbols through a basic dither sequence design. We then develop a theoretical model and propose a new algorithm to improve the collusion resistance of the basic scheme. Simulation results show that the improvement algorithm enhances the collusion resistance, while there is still a performance gap with the existing spread spectrum based fingerprinting. We then explore coded fingerprinting based on spread transform dither modulation (STDM) embedding. Simulation results show that this coded STDM based fingerprinting has significant advantages over spread spectrum based fingerprinting under blind detection. %B Proceedings of SPIE %V 6072 %P 698 - 709 %8 2006/// %G eng %0 Journal Article %J Geographical Analysis %D 2006 %T Extending the SAND Spatial Database System for the Visualization of Three-Dimensional Scientific Data %A Samet, Hanan %A Webber,R. E %X The three-dimensional extension of the SAND (Spatial and Nonspatial Data) spatial database system is described as is its use for data found in scientific visualization applications. The focus is on surface data. Some of the principal operations supported by SAND involve locating spatial objects in the order of their distance from other spatial objects in an incremental manner so that the number of objects that are needed is not known a priori. These techniques are shown to be useful in enabling users to visualize the results of certain proximity queries without having to execute algorithms to completion as is the case when performing a nearest-neighbor query where a Voronoi diagram (i.e., Thiessen polygon) would be computed as a preprocessing step before any attempt to respond to the query could be made. This is achieved by making use of operations such as the spatial join and the distance semijoin. Examples of the utility of such operations is demonstrated in the context of posing meteorological queries to a spatial database with a visualization component. %B Geographical Analysis %V 38 %P 87 - 101 %8 2006/// %G eng %N 1 %0 Conference Paper %B 2006 10th IEEE International Symposium on Wearable Computers %D 2006 %T Eye of the Beholder: Phone-Based Text-Recognition for the Visually-Impaired %A Tudor Dumitras %A Lee, M. %A Quinones, P. %A Smailagic, A. %A Siewiorek, Dan %A Narasimhan, P. %K CAMERAS %K data mining %K handicapped aids %K Hardware %K Image quality %K Internet %K Land mobile radio cellular systems %K Mobile computing %K mobile handsets %K mobile text-recognition system %K Optical character recognition software %K phone-based text-recognition %K Product safety %K Text recognition %K visually-impaired %X Blind and visually-impaired people cannot access essential information in the form of written text in our environment (e.g., on restaurant menus, street signs, door labels, product names and instructions, expiration dates). In this paper, we present and evaluate a mobile text-recognition system capable of extracting written information from a wide variety of sources and communicating it on-demand to the user. The user needs no additional hardware except an ordinary, Internet- enabled mobile camera-phone - a device that many visually-impaired individuals already own. This approach fills a gap in assistive technologies for the visually- impaired because it makes users aware of textual information not available to them through any other means. %B 2006 10th IEEE International Symposium on Wearable Computers %P 145 - 146 %8 2006/// %G eng %0 Conference Paper %B Proc. 5th IPTPS %D 2006 %T Fair file swarming with FOX %A Levin,D. %A Sherwood,R. %A Bhattacharjee, Bobby %X File swarming is a popular method of coordi-nated download by which peers obtain a file from an under- provisioned server. Critical problems arise within this domain when users act selfishly, yet most systems are built with altru- ism assumed. Working under the assumption that all peers are greedy, we introduce the Fair, Optimal eXchange (FOX) protocol. FOX, in addition to effective and robust application of the tit-for-tat incentive mechanism, provides theoretically optimal download times when everyone cooperates. Under our assumption of server and peer capabilities, we develop a strong threat model that provides peers with the incentives to not deviate from the protocol. From a theoretical perspective, we prove FOX’s optimality and incentive properties, even when the network consists only of purely self-interested peers. We also discuss issues in implementing and deploying such a system, and address the cost of ensuring fairness in a domain where efficiency is so important. %B Proc. 5th IPTPS %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 3rd IEEE/Eurographics Symposium on Point-Based Graphics. ACM, Boston, MA, USA %D 2006 %T A Fast k-Neighborhood Algorithm for Large Point-Clouds %A Sankaranarayanan,J. %A Samet, Hanan %A Varshney, Amitabh %X Algorithms that use point-cloud models make heavy use of the neighborhoods of the points. These neighborhoodsare used to compute the surface normals for each point, mollification, and noise removal. All of these primitive operations require the seemingly repetitive process of finding the k nearest neighbors of each point. These algo- rithms are primarily designed to run in main memory. However, rapid advances in scanning technologies have made available point-cloud models that are too large to fit in the main memory of a computer. This calls for more efficient methods of computing the k nearest neighbors of a large collection of points many of which are already in close proximity. A fast k nearest neighbor algorithm is presented that makes use of the locality of successive points whose k nearest neighbors are sought to significantly reduce the time needed to compute the neighborhood needed for the primitive operation as well as enable it to operate in an environment where the data is on disk. Results of experiments demonstrate an order of magnitude improvement in the time to perform the algorithm and several orders of magnitude improvement in work efficiency when compared with several prominent existing method. %B Proceedings of the 3rd IEEE/Eurographics Symposium on Point-Based Graphics. ACM, Boston, MA, USA %8 2006/// %G eng %0 Journal Article %J Communications of the ACM - Supporting exploratory search %D 2006 %T Find that photo!: interface strategies to annotate, browse, and share %A Shneiderman, Ben %A Bederson, Benjamin B. %A Drucker,Steven M. %B Communications of the ACM - Supporting exploratory search %V 49 %P 69 - 71 %8 2006/04// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/1121949.1121985 %N 4 %R 10.1145/1121949.1121985 %0 Book Section %B Geometric Modeling and Processing - GMP 2006 %D 2006 %T Finding Mold-Piece Regions Using Computer Graphics Hardware %A Priyadarshi,Alok %A Gupta, Satyandra K. %E Kim,Myung-Soo %E Shimada,Kenji %K Computer science %X An important step in the mold design process that ensures disassembly of mold pieces consists of identifying various regions on the part that will be formed by different mold pieces. This paper presents an efficient and robust algorithm to find and highlight the mold-piece regions on a part. The algorithm can be executed on current-generation computer graphics hardware. The complexity of the algorithm solely depends on the time to render the given part. By using a system that can quickly find various mold-piece regions on a part, designers can easily optimize the part and mold design and if needed make appropriate corrections upfront, streamlining the subsequent design steps. %B Geometric Modeling and Processing - GMP 2006 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4077 %P 655 - 662 %8 2006/// %@ 978-3-540-36711-6 %G eng %U http://www.springerlink.com/content/t8t3588924263p12/abstract/ %0 Journal Article %J Taxon %D 2006 %T First Steps Toward an Electronic Field Guide for Plants %A Haibin,Gaurav Agarwal %A Agarwal,Gaurav %A Ling,Haibin %A Jacobs, David W. %A Shirdhonkar,Sameer %A Kress,W. John %A Russell,Rusty %A Belhumeur,Peter %A Dixit,An %A Feiner,Steve %A Mahajan,Dhruv %A Sunkavalli,Kalyan %A Ramamoorthi,Ravi %A White,Sean %X this paper, we will describe our progress towards building a digital collection of the Smithsonian's type specimens, developing recognition algorithms that can match an image of a leaf to the species of plant from which it comes, and designing user interfaces for interacting with an electronic field guide. To start, we are developing a prototype electronic field guide for the flora of Plummers Island, a small, well-studied island in the Potomac River. This prototype system contains multiple images for each of about 130 species of plants on the island, and should soon grow to cover all 200+ species currently recorded (Shetler et al., 2005). Images of full specimens are available, as well as images of isolated leaves of each species. A zoomable user interface allows a user to browse these images, zooming in on ones of interest. Visual recognition algorithms assist a botanist in locating the specimens that are most relevant to identify the species of a plant. The system currently runs on a small hand-held computer. We will describe the components of this prototype, and also describe some of the future challenges we anticipate if we are to provide botanists in the field with all the resources that are now currently available in the world's museums and herbaria. Type Specimen Digital Collection The first challenge in producing our electronic field guide is to create a digital collection covering all of the Smithsonian's 85,000 vascular plant type specimens. For each type specimen, the database should eventually include systematically acquired high-resolution digital images of the specimen, textual descriptions, links to decision trees, images of live plants, and 3D models. Figure 1: On the left, our set-up at the Smithsonian for digitally photographing type specimens. On the... %B Taxon %V 55 %P 597 - 610 %8 2006/// %G eng %0 Book %D 2006 %T Foundations of Multidimensional And Metric Data Structures %A Samet, Hanan %K Computers / Data Modeling & Design %K Computers / Data Processing %K Computers / Databases / General %X The field of multidimensional data structures is large and growing very quickly. Here, for the first time, is a thorough treatment of multidimensional point data, object and image-based representations, intervals and small rectangles, and high-dimensional datasets. The book includes a thorough introduction; a comprehensive survey to spatial and multidimensional data structures and algorithms; and implementation details for the most useful data structures. Along with the hundreds of worked exercises and hundreds of illustrations, the result is an excellent and valuable reference tool for professionals in many areas, including computer graphics, databases, geographic information systems (GIS), game programming, image processing, pattern recognition, solid modeling, similarity retrieval, and VLSI design. Award Winner in 2006 Best Book competition in Professional and Scholarly Publishing from the Association of American Publishers. Morgan Kaufmann would like to congratulate Hanan Samet on receiving the UCGIS 2009 Research Award! Read the announcement here: http: //www.ucgis.org/summer2009/researchaward.htm * First comprehensive work on multidimensional data structures available, a thorough and authoritative treatment. * An algorithmic rather than mathematical approach, with a liberal use of examples that allows the readers to easily see the possible implementation and use. * Each section includes a large number of exercises and solutions to self-test and confirm the reader's understanding and suggest future directions. * Written by a well-known authority in the area of spatial data structures who has made many significant contributions to the field. The author's website includes: Spatial Index Demos %I Morgan Kaufmann %P 1023 %8 2006 %@ 9780123694461 %G eng %0 Book %D 2006 %T Foundations of Multidimensional And Metric Data Structures %A Samet, Hanan %K Computer algorithms %K Computers / Computer Graphics %K Computers / Computer Vision & Pattern Recognition %K Computers / Data Modeling & Design %K Computers / Data Processing %K Computers / Database Management / General %K Computers / Programming / General %K Computers / User Interfaces %K Data structures (Computer science) %K Science / Physics / General %X The field of multidimensional data structures is large and growing very quickly. Here, for the first time, is a thorough treatment of multidimensional point data, object and image-based representations, intervals and small rectangles, and high-dimensional datasets. The book includes a thorough introduction; a comprehensive survey to spatial and multidimensional data structures and algorithms; and implementation details for the most useful data structures. Along with the hundreds of worked exercises and hundreds of illustrations, the result is an excellent and valuable reference tool for professionals in many areas, including computer graphics, databases, geographic information systems (GIS), game programming, image processing, pattern recognition, solid modeling, similarity retrieval, and VLSI design. Award Winner in 2006 "Best Book" competition in Professional and Scholarly Publishing from the Association of American Publishers. Morgan Kaufmann would like to congratulate Hanan Samet on receiving the UCGIS 2009 Research Award! Read the announcement here: http://www.ucgis.org/summer2009/researchaward.htm * First comprehensive work on multidimensional data structures available, a thorough and authoritative treatment. * An algorithmic rather than mathematical approach, with a liberal use of examples that allows the readers to easily see the possible implementation and use. * Each section includes a large number of exercises and solutions to self-test and confirm the reader's understanding and suggest future directions. * Written by a well-known authority in the area of spatial data structures who has made many significant contributions to the field. The author's website includes: Spatial Index Demos %I Morgan Kaufmann %8 2006/08/08/ %@ 9780123694461 %G eng %0 Journal Article %J Molecular EndocrinologyMolecular Endocrinology %D 2006 %T Functional Analysis of Hes-1 in Preadipocytes %A Ross,David A %A Hannenhalli, Sridhar %A Tobias,John W %A Cooch,Neil %A Shiekhattar,Ramin %A Kadesch,Tom %X Notch signaling blocks differentiation of 3T3-L1 preadipocytes, and this can be mimicked by constitutive expression of the Notch target gene Hes-1. Although considered initially to function only as a repressor, recent evidence indicates that Hes-1 can also activate transcription. We show here that the domains of Hes-1 needed to block adipogenesis coincide with those necessary for transcriptional repression. HRT1, another basic-helix-loop-helix protein and potential Hes-1 partner, was also induced by Notch in 3T3-L1 cells but did not block adipogenesis, suggesting that Hes-1 functions primarily as a homodimer or possibly as a heterodimer with an unknown partner. Purification of Hes-1 identified the Groucho/transducin-like enhancer of split family of corepressors as the only significant Hes-1 interacting proteins in vivo. An evaluation of global gene expression in preadipocytes identified approximately 200 Hes-1-responsive genes comprising roughly equal numbers of up-regulated and down-regulated genes. However, promoter analyses indicated that the down-regulated genes were significantly more likely to contain Hes-1 binding sites, indicating that Hes-1 is more likely to repress transcription of its direct targets. We conclude that Notch most likely blocks adipogenesis through the induction of Hes-1 homodimers, which repress transcription of key target genes. %B Molecular EndocrinologyMolecular Endocrinology %V 20 %P 698 - 705 %8 2006/03/01/ %@ 0888-8809, 1944-9917 %G eng %U http://mend.endojournals.org/content/20/3/698 %N 3 %R 10.1210/me.2005-0325 %0 Report %D 2006 %T The Gram-Schmidt Algorithm and Its Variations %A Stewart, G.W. %K Technical Report %X The Gram--Schmidt algorithm is a widely used method fororthogonalizing a sequence of vectors. It comes in two forms: classical Gram--Schmidt and modified Gram--Schmidt, each of whose operations can be ordered in different ways. This expository paper gives a systematic treatment of this confusing variety of algorithms. It also treats the numerical issue of loss of orthogonality and reorthogonalization as well as the implementation of column pivoting. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2004-84 %8 2006/01/13/ %G eng %U http://drum.lib.umd.edu/handle/1903/3036 %0 Journal Article %J IEEE Multimedia %D 2006 %T Hierarchical Layouts for Photo Libraries %A Kustanowitz,J. %A Shneiderman, Ben %K annotated digital photo collection %K auto-layout technique %K bi-level hierarchies %K Computer science %K data visualisation %K digital libraries %K document image processing %K Information Visualization %K interactive algorithms %K interactive displays %K Libraries %K Lifting equipment %K Organization Charts %K photo collections %K photo layouts %K photo library %K Photography %K quantum content %K Silver %K Springs %K User interfaces %K Web pages %X We use an annotated digital photo collection to demonstrate a two-level auto-layout technique consisting of a central primary region with secondary regions surrounding it. Because the object sizes within regions can only be changed in discrete units, we refer to them as quantum content. Our real-time algorithms enable a compelling interactive display as users resize the canvas, or move and resize the primary region %B IEEE Multimedia %V 13 %P 62 - 72 %8 2006/12//Oct %@ 1070-986X %G eng %N 4 %R 10.1109/MMUL.2006.83 %0 Journal Article %J Computer %D 2006 %T High-confidence medical device software and systems %A Lee,I. %A Pappas,G. J %A Cleaveland, Rance %A Hatcliff,J. %A Krogh,B. H %A Lee,P. %A Rubin,H. %A Sha,L. %K Aging %K biomedical equipment %K Clinical software engineering %K Costs %K FDA device %K Food and Drug Administration device %K health and safety %K health care %K health information system %K health safety %K healthcare delivery %K Healthcare technology %K Information systems %K medical computing %K medical device manufacturing %K medical device software development %K medical device systems development %K medical information systems %K Medical services %K Medical software %K Medical tests %K networked medical devices %K Production systems %K Software design %K Software safety %K Software systems %K Software testing %K US healthcare quality %X Given the shortage of caregivers and the increase in an aging US population, the future of US healthcare quality does not look promising and definitely is unlikely to be cheaper. Advances in health information systems and healthcare technology offer a tremendous opportunity for improving the quality of care while reducing costs. The development and production of medical device software and systems is a crucial issue, both for the US economy and for ensuring safe advances in healthcare delivery. As devices become increasingly smaller in physical terms but larger in software terms, the design, testing, and eventual Food and Drug Administration (FDA) device approval is becoming much more expensive for medical device manufacturers both in terms of time and cost. Furthermore, the number of devices that have recently been recalled due to software and hardware problems is increasing at an alarming rate. As medical devices are becoming increasingly networked, ensuring even the same level of health safety seems a challenge. %B Computer %V 39 %P 33 - 38 %8 2006/04// %@ 0018-9162 %G eng %N 4 %R 10.1109/MC.2006.127 %0 Journal Article %J Proceedings of the IEEE %D 2006 %T How Multirobot Systems Research will Accelerate our Understanding of Social Animal Behavior %A Balch, T. %A Dellaert, F. %A Feldman, A. %A Guillory, A. %A Isbell, C.L. %A Zia Khan %A Pratt, S.C. %A Stein, A.N. %A Wilde, H. %K Acceleration %K Animal behavior %K ant movement tracking %K Artificial intelligence %K biology computing %K Computer vision %K control engineering computing %K Insects %K Intelligent robots %K Labeling %K monkey movement tracking %K multi-robot systems %K multirobot systems %K robotics algorithms %K Robotics and automation %K social animal behavior %K social animals %K social insect behavior %K Speech recognition %K tracking %X Our understanding of social insect behavior has significantly influenced artificial intelligence (AI) and multirobot systems' research (e.g., ant algorithms and swarm robotics). In this work, however, we focus on the opposite question: "How can multirobot systems research contribute to the understanding of social animal behavior?" As we show, we are able to contribute at several levels. First, using algorithms that originated in the robotics community, we can track animals under observation to provide essential quantitative data for animal behavior research. Second, by developing and applying algorithms originating in speech recognition and computer vision, we can automatically label the behavior of animals under observation. In some cases the automatic labeling is more accurate and consistent than manual behavior identification. Our ultimate goal, however, is to automatically create, from observation, executable models of behavior. An executable model is a control program for an agent that can run in simulation (or on a robot). The representation for these executable models is drawn from research in multirobot systems programming. In this paper we present the algorithms we have developed for tracking, recognizing, and learning models of social animal behavior, details of their implementation, and quantitative experimental results using them to study social insects %B Proceedings of the IEEE %V 94 %P 1445 - 1463 %8 2006/07// %@ 0018-9219 %G eng %N 7 %0 Book %D 2006 %T Image Database Systems and Techniques: A Symbolic Approach %A Samet, Hanan %A Soffer,Aya %I Morgan Kaufmann Publishers Inc. %C San Francisco, CA, USA %8 2006/// %@ 1558605495 %G eng %0 Journal Article %J Digital Mammography %D 2006 %T Image Similarity and Asymmetry to Improve Computer-Aided Detection of Breast Cancer %A Tahmoush,D. %A Samet, Hanan %X An improved image similarity method is introduced to recognize breast cancer, and it is incorporated into a computer-aided breast cancer detection system through Bayes Theorem. Radiologists can use the differences between the left and right breasts, or asymmetry, in mammograms to help detect certain malignant breast cancers. Image similarity is used to determine asymmetry using a contextual and then a spatial comparison. The mammograms are filtered to find the most contextually significant points, and then the resulting point set is analyzed for spatial similarity. We develop the analysis through a combination of modeling and supervised learning of model parameters. This process correctly classifies mammograms 84% of the time, and significantly improves the accuracy of a computer-aided breast cancer detection system by 71%. %B Digital Mammography %P 221 - 228 %8 2006/// %G eng %R 10.1007/11783237_31 %0 Conference Paper %B Image Processing, 2006 IEEE International Conference on %D 2006 %T Image Tampering Identification using Blind Deconvolution %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K (access %K approximation;blind %K authentication;image %K coding; %K compression;deconvolution;filtering %K control);data %K deconvolution;camera;consumer %K diagnosis;surveillance;tampering %K editing %K identification;approximation %K images;filter %K photography;digital %K process;image %K softwares;medical %K theory;biometrics %K theory;image %X Digital images have been used in growing number of applications from law enforcement and surveillance, to medical diagnosis and consumer photography. With such widespread popularity and the presence of low-cost image editing softwares, the integrity of image content can no longer be taken for granted. In this paper, we propose a novel technique based on blind deconvolution to verify image authenticity. We consider the direct output images of a camera as authentic, and introduce algorithms to detect further processing such as tampering applied to the image. Our proposed method is based on the observation that many tampering operations can be approximated as a combination of linear and non-linear components. We model the linear part of the tampering process as a filter, and obtain its coefficients using blind deconvolution. These estimated coefficients are then used to identify possible manipulations. We demonstrate the effectiveness of the proposed image authentication technique and compare our results with existing works %B Image Processing, 2006 IEEE International Conference on %P 2309 - 2312 %8 2006/10// %G eng %R 10.1109/ICIP.2006.312848 %0 Conference Paper %B Data Engineering Workshops, 2006. Proceedings. 22nd International Conference on %D 2006 %T Implementing a bioinformatics pipeline (bip) on a mediator platform: Comparing cost and quality of alternate choices %A Eckman,B. A %A Gaasterland,T. %A Lacroix,Z. %A Raschid, Louiqa %A Snyder,B. %A Vidal,M. E %B Data Engineering Workshops, 2006. Proceedings. 22nd International Conference on %P 67 - 67 %8 2006/// %G eng %0 Journal Article %J Journal of Computer and System Sciences %D 2006 %T An improved approximation algorithm for vertex cover with hard capacities %A Gandhi,Rajiv %A Halperin,Eran %A Khuller, Samir %A Kortsarz,Guy %A Srinivasan, Aravind %K Approximation algorithms %K Capacitated covering %K Linear programming %K Randomized rounding %K Set cover %K Vertex cover %X We study the capacitated vertex cover problem, a generalization of the well-known vertex-cover problem. Given a graph G = ( V , E ) , the goal is to cover all the edges by picking a minimum cover using the vertices. When we pick a vertex, we can cover up to a pre-specified number of edges incident on this vertex (its capacity). The problem is clearly NP-hard as it generalizes the well-known vertex-cover problem. Previously, approximation algorithms with an approximation factor of 2 were developed with the assumption that an arbitrary number of copies of a vertex may be chosen in the cover. If we are allowed to pick at most a fixed number of copies of each vertex, the approximation algorithm becomes much more complex. Chuzhoy and Naor (FOCS, 2002) have shown that the weighted version of this problem is at least as hard as set cover; in addition, they developed a 3-approximation algorithm for the unweighted version. We give a 2-approximation algorithm for the unweighted version, improving the Chuzhoy–Naor bound of three and matching (up to lower-order terms) the best approximation ratio known for the vertex-cover problem. %B Journal of Computer and System Sciences %V 72 %P 16 - 33 %8 2006/02// %@ 0022-0000 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022000005000747 %N 1 %R 10.1016/j.jcss.2005.06.004 %0 Journal Article %J Proceedings of 7th Annual International Conference on Digital Libraries (DG06) %D 2006 %T Integrating data and interfaces to enhance understanding of government statistics: toward the national statistical knowledge network project briefing %A Marchionini,G. %A Haas,S. %A Plaisant, Catherine %A Shneiderman, Ben %X This paper reports the results of work in the final no-costextension year of a digital government project that developed user interface models and prototypes to help people find and understand government statistics; proposed a Statistical Knowledge Network architecture that supports cross agency information access; and demonstrated models for government- academic collaboration. %B Proceedings of 7th Annual International Conference on Digital Libraries (DG06) %P 21 - 24 %8 2006/// %G eng %0 Journal Article %J Visualization and Computer Graphics, IEEE Transactions on %D 2006 %T Isosurface Extraction and Spatial Filtering using Persistent Octree (POT) %A Shi,Q. %A JaJa, Joseph F. %K 4D isocontour slicing;Richtmyer-Meshkov instability dataset;branch-on-need octree;hybrid data structure;isosurface extraction;persistent octree;spatial filtering;data visualisation;database indexing;feature extraction;octrees;spatial data structures;Algor %K Computer-Assisted;Imaging %K Three-Dimensional;Information Storage and Retrieval;User-Computer Interface; %X We propose a novel persistent octree (POT) indexing structure for accelerating isosurface extraction and spatial filtering from volumetric data. This data structure efficiently handles a wide range of visualization problems such as the generation of view-dependent isosurfaces, ray tracing, and isocontour slicing for high dimensional data. POT can be viewed as a hybrid data structure between the interval tree and the branch-on-need octree (BONO) in the sense that it achieves the asymptotic bound of the interval tree for identifying the active cells corresponding to an isosurface and is more efficient than BONO for handling spatial queries. We encode a compact octree for each isovalue. Each such octree contains only the corresponding active cells, in such a way that the combined structure has linear space. The inherent hierarchical structure associated with the active cells enables very fast filtering of the active cells based on spatial constraints. We demonstrate the effectiveness of our approach by performing view-dependent isosurfacing on a wide variety of volumetric data sets and 4D isocontour slicing on the time-varying Richtmyer-Meshkov instability dataset %B Visualization and Computer Graphics, IEEE Transactions on %V 12 %P 1283 - 1290 %8 2006/10//sept %@ 1077-2626 %G eng %N 5 %R 10.1109/TVCG.2006.157 %0 Journal Article %J IEEE/ACM Trans. Netw. %D 2006 %T JET: dynamic join-exit-tree amortization and scheduling for contributory key management %A Mao,Yinian %A Sun,Yan %A M. Wu %A Liu,K. J.R %K contributory key management %K dynamic tree topology %K secure group communications %K time efficiency %X In secure group communications, the time cost associated with key updates in the events of member join and departure is an important aspect of quality of service, especially in large groups with highly dynamic membership. To achieve better time efficiency, we propose a join-exit-tree (JET) key management framework. First, a special key tree topology with join and exit subtrees is introduced to handle key updates for dynamic membership. Then, optimization techniques are employed to determine the capacities of join and exit subtrees for achieving the best time efficiency, and algorithms are designed to dynamically update the join and exit trees. We show that, on average, the asymptotic time cost for each member join/departure event is reduced to O(log (log n)) from the previous cost of O(log n), where n is the group size. Our experimental results based on simulated user activities as well as the real MBone data demonstrate that the proposed JET scheme can significantly improve the time efficiency, while maintaining low communication and computation cost, of tree-based contributory key management. %B IEEE/ACM Trans. Netw. %V 14 %P 1128 - 1140 %8 2006/10// %@ 1063-6692 %G eng %U http://dx.doi.org/10.1109/TNET.2006.882851 %N 5 %R 10.1109/TNET.2006.882851 %0 Book Section %B Computer Vision – ACCV 2006Computer Vision – ACCV 2006 %D 2006 %T Key Frame-Based Activity Representation Using Antieigenvalues %A Cuntoor,Naresh %A Chellapa, Rama %E Narayanan,P. %E Nayar,Shree %E Shum,Heung-Yeung %X Many activities may be characterized by a sequence of key frames that are related to important changes in motion rather than dominant characteristics that persist over a long sequence of frames. To detect such changes, we define a transformation operator at every time instant, which relates the past to the future states. One of the useful quantities associated with numerical range of an operator is the eigenvalue. In the literature, eigenvalue-based approaches have been studied extensively for many modeling tasks. These rely on gross properties of the data and are not suitable to detect subtle changes. We propose an antieigenvalue – based measure to detect key frames. Antieigenvalues depend critically on the turning of the operator, whereas eigenvalues represent the amount of dilation along the eigenvector directions aligned with the direction of maximum variance. We demonstrate its application to activity modeling and recognition using two datasets: a motion capture dataset and the UCF human action dataset. %B Computer Vision – ACCV 2006Computer Vision – ACCV 2006 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3852 %P 499 - 508 %8 2006/// %@ 978-3-540-31244-4 %G eng %U http://dx.doi.org/10.1007/11612704_50 %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2006 %T Knowledge discovery in high-dimensional data: case studies and a user survey for the rank-by-feature framework %A Seo,Jinwook %A Shneiderman, Ben %K case study %K Computer aided software engineering %K Computer Society %K Data analysis %K data mining %K data visualisation %K Data visualization %K database management systems %K e-mail user survey %K Genomics %K Helium %K Hierarchical Clustering Explorer %K hierarchical clustering explorer. %K high-dimensional data %K Histograms %K Information visualization evaluation %K interactive systems %K interactive tool %K knowledge discovery %K multivariate data %K Rank-by-feature framework %K Scattering %K Testing %K user interface %K User interfaces %K user survey %K visual analytic tools %K visual analytics %K visualization tools %X Knowledge discovery in high-dimensional data is a challenging enterprise, but new visual analytic tools appear to offer users remarkable powers if they are ready to learn new concepts and interfaces. Our three-year effort to develop versions of the hierarchical clustering explorer (HCE) began with building an interactive tool for exploring clustering results. It expanded, based on user needs, to include other potent analytic and visualization tools for multivariate data, especially the rank-by-feature framework. Our own successes using HCE provided some testimonial evidence of its utility, but we felt it necessary to get beyond our subjective impressions. This paper presents an evaluation of the hierarchical clustering explorer (HCE) using three case studies and an e-mail user survey (n=57) to focus on skill acquisition with the novel concepts and interface for the rank-by-feature framework. Knowledgeable and motivated users in diverse fields provided multiple perspectives that refined our understanding of strengths and weaknesses. A user survey confirmed the benefits of HCE, but gave less guidance about improvements. Both evaluations suggested improved training methods %B IEEE Transactions on Visualization and Computer Graphics %V 12 %P 311 - 322 %8 2006/06//May %@ 1077-2626 %G eng %N 3 %R 10.1109/TVCG.2006.50 %0 Journal Article %J Probabilistic, Logical and Relational Learning-Towards a Synthesis %D 2006 %T Learning through failure %A Sato,T. %A Kameya,Y. %A De Raedt,L. %A Dietterich,T. %A Getoor, Lise %A Muggleton,S. H %X PRISM, a symbolic-statistical modeling language we have been developing since '97, recently incorporated a program transformation technique to handle failure in generative modeling. I'll show this feature opens a way to new breeds of symbolic models, including EM learning from negative observations, constrained HMMs and finite PCFGs. %B Probabilistic, Logical and Relational Learning-Towards a Synthesis %8 2006/// %G eng %N 05051 %0 Book Section %B Algorithms and ComputationAlgorithms and Computation %D 2006 %T Lower Bounds on the Deterministic and Quantum Communication Complexities of Hamming-Distance Problems %A Ambainis,Andris %A Gasarch,William %A Srinivasan, Aravind %A Utis,Andrey %E Asano,Tetsuo %X A sub-area of discrepancy theory that has received much attention in computer science re-cently, is that of explicit constructions of low-discrepancy point sets for various types of rectangle families in high dimension. This research has led to interesting applications in error-control cod- ing, distributed protocols, Web document filtering, derandomization, and other areas. We give a short survey of this area here. %B Algorithms and ComputationAlgorithms and Computation %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4288 %P 628 - 637 %8 2006/// %@ 978-3-540-49694-6 %G eng %U http://dx.doi.org/10.1007/11940128_63 %0 Journal Article %J ACM Trans. Database Syst. %D 2006 %T Maintenance of K-nn and spatial join queries on continuously moving points %A Iwerks,Glenn S. %A Samet, Hanan %A Smith,Kenneth P. %K k-nearest neighbor %K continuously moving objects %K materialized view maintenance %K Moving object databases %K Spatial join %K temporal databases %X Cars, aircraft, mobile cell phones, ships, tanks, and mobile robots all have the common property that they are moving objects. A kinematic representation can be used to describe the location of these objects as a function of time. For example, a moving point can be represented by the function p(t) = &OV0489;0 + (t − t0)&OV0505;, where &OV0489;0 is the start location, t0 is the start time, and &OV0505; is its velocity vector. Instead of storing the location of the object at a given time in a database, the coefficients of the function are stored. When an object's behavior changes enough so that the function describing its location is no longer accurate, the function coefficients for the object are updated. Because the location of each object is represented as a function of time, spatial query results can change even when no transactions update the database. We present efficient algorithms to maintain k-nearest neighbor, and spatial join queries in this domain as time advances and updates occur. We assume no previous knowledge of what the updates will be before they occur. We experimentally compare these new algorithms with more straight forward adaptations of previous work to support updates. Experiments are conducted using synthetic uniformly distributed data, and real aircraft flight data. The primary metric of comparison is the number of I/O disk accesses needed to maintain the query results and the supporting data structures. %B ACM Trans. Database Syst. %V 31 %P 485 - 536 %8 2006/06// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/1138394.1138396 %N 2 %R 10.1145/1138394.1138396 %0 Conference Paper %B CHI '06 extended abstracts on Human factors in computing systems %D 2006 %T Making a difference: integrating socially relevant projects into HCI teaching %A Shneiderman, Ben %A Bishop,Ann %A Friedman,Batya %A Lazar,Jonathan %A Marsden,Gary %A Nass,Cliff %K education %K making a difference %K relevance %K service learning %K social impact %K team projects %X Enriching courses on human-computer interaction with socially-relevant projects provides a compelling opportunity for students to improve their education and make socially beneficial contributions. By having clearly defined user communities outside the classroom, students have the chance to practice their interview, observation, and usability testing skills, while developing projects that continue beyond the semester. These projects often give students life-changing exposure to genuine needs and impressive results to include in their portfolio when seeking employment. Educators will present their strategies for arranging, supervising, and grading these projects. Students will describe their experience and how it influenced them. %B CHI '06 extended abstracts on Human factors in computing systems %S CHI EA '06 %I ACM %C New York, NY, USA %P 41 - 44 %8 2006/// %@ 1-59593-298-4 %G eng %U http://doi.acm.org/10.1145/1125451.1125463 %R 10.1145/1125451.1125463 %0 Conference Paper %B 19th IEEE Computer Security Foundations Workshop, 2006 %D 2006 %T Managing policy updates in security-typed languages %A Swamy,N. %A Hicks, Michael W. %A Tse,S. %A Zdancewic,S. %K Access control %K Computer languages %K Data security %K Database systems %K dynamic queries %K dynamic semantics %K Educational institutions %K high level languages %K Information security %K information-flow policy management %K Lattices %K Network servers %K Operating systems %K policy update management %K Robustness %K role-based security policies %K RT role-based trust-management framework %K Rx security-typed programming language %K security of data %K statically verified transactions %K transitive flows %X This paper presents Rx, a new security-typed programming language with features intended to make the management of information-flow policies more practical. Security labels in Rx, in contrast to prior approaches, are defined in terms of owned roles, as found in the RT role-based trust-management framework. Role-based security policies allow flexible delegation, and our language Rx provides constructs through which programs can robustly update policies and react to policy updates dynamically. Our dynamic semantics use statically verified transactions to eliminate illegal information flows across updates, which we call transitive flows. Because policy updates can be observed through dynamic queries, policy updates can potentially reveal sensitive information. As such, Rx considers policy statements themselves to be potentially confidential information and subject to information-flow metapolicies %B 19th IEEE Computer Security Foundations Workshop, 2006 %I IEEE %P 13 pp.-216 - 13 pp.-216 %8 2006/// %@ 0-7695-2615-2 %G eng %R 10.1109/CSFW.2006.17 %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on %D 2006 %T Mapping Multimedia Applications Onto Configurable Hardware With Parameterized Cyclo-Static Dataflow Graphs %A Haim,F. %A Sen,M. %A Ko,Dong-Ik %A Bhattacharyya, Shuvra S. %A Wolf,W. %K configurable hardware %K Data communication %K Gesture recognition %K gesture recognition application %K graph theory %K hardware mapping %K image processing applications %K mapping multimedia applications %K meta-modeling technique %K model-based design %K Multimedia communication %K parameterized cyclo-static dataflow graphs %X This paper develops methods for model-based design and implementation of image processing applications. We apply our previously developed meta-modeling technique of homogeneous parameterized dataflow (HPDF) (M. Sen et al., 2005) to the framework of cyclostatic dataflow (CSDF) (G. Bilsen et al., 1996), and demonstrate this integrated modeling methodology through hardware mapping of a gesture recognition application. We also provide a comparative study between HPDF/CSDF-based representation of the gesture recognition application, and a previously developed version based on applying HPDF in conjunction with conventional synchronous dataflow (SDF) semantics (M. Sen et al., 2005) %B Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on %V 3 %P III - III %8 2006/05// %G eng %R 10.1109/ICASSP.2006.1660838 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2006-15 %D 2006 %T Matching Jobs to Resources in Distributed Desktop Grid Environments %A Kim,Jik-Soo %A Bhattacharjee, Bobby %A Keleher, Peter J. %A Sussman, Alan %K Technical Report %X Desktop grids use opportunistic sharing to exploit large collectionsof personal computers and workstations across the Internet and can achieve tremendous computing power with low cost. However, current systems are typically based on a traditional client-server architecture, which has inherent shortcomings with respect to robustness, reliability and scalability. In this paper, we propose a decentralized, robust, highly available, and scalable infrastructure to match incoming jobs to available resources. The key idea behind our proposed system is to leverage information provided by an underlying peer-to-peer system to create a hierarchical Rendezvous Node Tree, which performs the matching efficiently. Our experimental results obtained via simulation show that we can effectively match jobs with varying levels of resource constraints to available nodes and maintain good load balance in a fully decentralized heterogeneous computational environment. %B Technical Reports from UMIACS, UMIACS-TR-2006-15 %8 2006/04// %G eng %U http://drum.lib.umd.edu/handle/1903/3669 %0 Journal Article %J Computer Networks %D 2006 %T Measurement-based optimal routing on overlay architectures for unicast sessions %A La,R.J. %A Shayman,M.A. %A Bhattacharjee, Bobby %X We propose a measurement-based routing algorithm to load-balance intradomain traffic along multiple paths for multi-ple unicast sources. Multiple paths are established using overlay nodes. The distributed algorithm is derived from simul- taneous perturbation stochastic approximation (SPSA) and does not assume that the gradient of an analytical cost function is known. Instead, we rely on (potentially) noisy estimates from local measurements. Using the analytical model presented in the paper, we first show the almost sure convergence of the algorithm to the optimal solution under a decreas- ing step size policy (as with a standard SPSA model). Motivated by practical concerns, we next consider the constant step size case, for which we establish weak convergence. We provide simulation results to demonstrate the advantages of our proposed algorithm under various network scenarios, and also present a comparative study with MATE (an existing opti- mal routing algorithm) %B Computer Networks %V 50 %P 1938 - 1951 %8 2006/// %G eng %N 12 %0 Journal Article %J Proceedings of SPIE %D 2006 %T Measuring human movement for biomechanical applications using markerless motion capture %A Mündermann, L. %A Corazza, S. %A Chaudhari, A.M. %A Andriacchi, T.P. %A Sundaresan, A. %A Chellapa, Rama %X Modern biomechanical and clinical applications require the accurate capture of normal and pathological humanmovement without the artifacts associated with standard marker-based motion capture techniques such as soft tissue artifacts and the risk of artificial stimulus of taped-on or strapped-on markers. In this study, the need for new markerless human motion capture methods is discussed in view of biomechanical applications. Three different approaches for estimating human movement from multiple image sequences were explored. The first two approaches tracked a 3D articulated model in 3D representations constructed from the image sequences, while the third approach tracked a 3D articulated model in multiple 2D image planes. The three methods are systematically evaluated and results for real data are presented. The role of choosing appropriate technical equipment and algorithms for accurate markerless motion capture is critical. The implementation of this new methodology offers the promise for simple, time-efficient, and potentially more meaningful assessments of human movement in research and clinical practice. %B Proceedings of SPIE %V 6056 %P 60560R - 60560R %8 2006/// %G eng %0 Conference Paper %B Embedded Computer Systems: Architectures, Modeling and Simulation, 2006. IC-SAMOS 2006. International Conference on %D 2006 %T Memory-constrained block processing optimization for synthesis of DSP software %A Ko,M. Y %A Shen,C. C %A Bhattacharyya, Shuvra S. %B Embedded Computer Systems: Architectures, Modeling and Simulation, 2006. IC-SAMOS 2006. International Conference on %P 137 - 143 %8 2006/// %G eng %0 Journal Article %J Science %D 2006 %T Metagenomic Analysis of the Human Distal Gut Microbiome %A Gill,Steven R. %A Pop, Mihai %A DeBoy,Robert T. %A Eckburg,Paul B %A Turnbaugh,Peter J %A Samuel,Buck S %A Gordon,Jeffrey I %A Relman,David A %A Fraser-Liggett,Claire M %A Nelson,Karen E. %X The human intestinal microbiota is composed of 1013 to 1014 microorganisms whose collective genome (“microbiome”) contains at least 100 times as many genes as our own genome. We analyzed ∼78 million base pairs of unique DNA sequence and 2062 polymerase chain reaction–amplified 16S ribosomal DNA sequences obtained from the fecal DNAs of two healthy adults. Using metabolic function analyses of identified genes, we compared our human genome with the average content of previously sequenced microbial genomes. Our microbiome has significantly enriched metabolism of glycans, amino acids, and xenobiotics; methanogenesis; and 2-methyl-d-erythritol 4-phosphate pathway–mediated biosynthesis of vitamins and isoprenoids. Thus, humans are superorganisms whose metabolism represents an amalgamation of microbial and human attributes. %B Science %V 312 %P 1355 - 1359 %8 2006/06/02/ %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/312/5778/1355 %N 5778 %R 10.1126/science.1124234 %0 Patent %D 2006 %T Methods for the electronic annotation, retrieval, and use of electronic images %A Shneiderman, Ben %E University of Maryland, College Park %X The present invention provides software for electronically annotating electronic images, such as drawings, photographs, video, etc., through the drag-and-drop of annotations from a pre-defined, but extendable, list. The annotations are placed at a user-selected X, Y location on the image, and stored in a searchable database. Thus, they can be searched in order to retrieve, organize, group, or display desired electronic images or collections of such images. The annotations may be text, images, sounds, etc. The invention provides a flexible, easy to learn, rapid, low error rate and satisfying interface for accomplishing such tasks. %V 09/976,362 %8 2006/03/07/ %G eng %U http://www.google.com/patents?id=iBt4AAAAEBAJ %N 7010751 %0 Journal Article %J Computers in Human Behavior %D 2006 %T A model for computer frustration: the role of instrumental and dispositional factors on incident, session, and post-session frustration and mood %A Bessière,Katie %A Newhagen,John E. %A Robinson,John P. %A Shneiderman, Ben %K Efficacy %K Emotion %K Frustration %K Mood %K problem solving %K usability %X Frustration is almost universally accepted as the emotional outcome of a negative computing experience. Despite the wide use of the term, however, it has not been rigorously conceptualized as a factor in the study of the human–computer interface. This project sets out to explicate frustration as a pre-emotional state generated by the user’s appraisal of the interface as an impediment to goal attainment, and looks at how user characteristics, such as self-efficacy, relate to it. This project employed episode report methodology to capture data from 144 computer users’ reports of actual frustrating events as they took place. Diaries taken as users worked at their everyday tasks yield detailed data about the problems they encountered and included information about session length and an estimate of the time lost due to the experiences. Outcomes were measured as either situational or dispositional factors. Situational factors, having to do with specific events, predicted incident frustration. However, disposition variables, especially user self-efficacy, were much stronger, predicting incident and session frustration, and post-session mood. One surprising outcome was the failure of demographic variables as predictors of frustration. %B Computers in Human Behavior %V 22 %P 941 - 961 %8 2006/11// %@ 0747-5632 %G eng %U http://www.sciencedirect.com/science/article/pii/S0747563204000615 %N 6 %R 10.1016/j.chb.2004.03.015 %0 Conference Paper %B Signals, Systems and Computers, 2006. ACSSC'06. Fortieth Asilomar Conference on %D 2006 %T Model-based mapping of image registration applications onto configurable hardware %A Hemaraj,Y. %A Sen,M. %A Shekhar,R. %A Bhattacharyya, Shuvra S. %B Signals, Systems and Computers, 2006. ACSSC'06. Fortieth Asilomar Conference on %P 1453 - 1457 %8 2006/// %G eng %0 Conference Paper %B 2006 International Conference on Parallel Processing Workshops, 2006. ICPP 2006 Workshops %D 2006 %T Model-based OpenMP implementation of a 3D facial pose tracking system %A Saha,S. %A Chung-Ching Shen %A Chia-Jui Hsu %A Aggarwal,G. %A Veeraraghavan,A. %A Sussman, Alan %A Bhattacharyya, Shuvra S. %K 3D facial pose tracking system %K application modeling %K application program interfaces %K application scheduling %K coarse-grain dataflow graphs %K Concurrent computing %K data flow graphs %K Educational institutions %K face recognition %K IMAGE PROCESSING %K image processing applications %K Inference algorithms %K Message passing %K OpenMP platform %K parallel implementation %K PARALLEL PROCESSING %K parallel programming %K Particle tracking %K Processor scheduling %K SHAPE %K shared memory systems %K shared-memory systems %K Solid modeling %K tracking %X Most image processing applications are characterized by computation-intensive operations, and high memory and performance requirements. Parallelized implementation on shared-memory systems offer an attractive solution to this class of applications. However, we cannot thoroughly exploit the advantages of such architectures without proper modeling and analysis of the application. In this paper, we describe our implementation of a 3D facial pose tracking system using the OpenMP platform. Our implementation is based on a design methodology that uses coarse-grain dataflow graphs to model and schedule the application. We present our modeling approach, details of the implementation that we derived based on this modeling approach, and associated performance results. The parallelized implementation achieves significant speedup, and meets or exceeds the target frame rate under various configurations %B 2006 International Conference on Parallel Processing Workshops, 2006. ICPP 2006 Workshops %I IEEE %P 8 pp.-73 - 8 pp.-73 %8 2006/// %@ 0-7695-2637-3 %G eng %R 10.1109/ICPPW.2006.55 %0 Journal Article %J Infect. Immun. %D 2006 %T Molecular Characterization of Serine-, Alanine-, and Proline-Rich Proteins of Trypanosoma cruzi and Their Possible Role in Host Cell Infection %A Baida,Renata C. P. %A Santos,Marcia R. M. %A Carmo,Mirian S. %A Yoshida,Nobuko %A Ferreira,Danielle %A Ferreira,Alice Teixeira %A El Sayed,Najib M. %A Andersson,Bjorn %A da Silveira,Jose Franco %X We previously reported the isolation of a novel protein gene family, termed SAP (serine-, alanine-, and proline-rich protein), from Trypanosoma cruzi. Aided by the availability of the completed genome sequence of T. cruzi, we have now identified 39 full-length sequences of SAP, six pseudogenes and four partial genes. SAPs share a central domain of about 55 amino acids and can be divided into four groups based on their amino (N)- and carboxy (C)-terminal sequences. Some SAPs have conserved N- and C-terminal domains encoding a signal peptide and a glycosylphosphatidylinositol anchor addition site, respectively. Analysis of the expression of SAPs in metacyclic trypomastigotes by two-dimensional electrophoresis and immunoblotting revealed that they are likely to be posttranslationally modified in vivo. We have also demonstrated that some SAPs are shed into the extracellular medium. The recombinant SAP exhibited an adhesive capacity toward mammalian cells, where binding was dose dependent and saturable, indicating a possible ligand-receptor interaction. SAP triggered the host cell Ca2+ response required for parasite internalization. A cell invasion assay performed in the presence of SAP showed inhibition of internalization of the metacyclic forms of the CL strain. Taken together, these results show that SAP is involved in the invasion of mammalian cells by metacyclic trypomastigotes, and they confirm the hypothesis that infective trypomastigotes exploit an arsenal of surface glycoproteins and shed proteins to induce signaling events required for their internalization. %B Infect. Immun. %V 74 %P 1537 - 1546 %8 2006/03/01/ %G eng %U http://iai.asm.org/cgi/content/abstract/74/3/1537 %N 3 %R

10.1128/IAI.74.3.1537-1546.2006

%0 Conference Paper %B Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on %D 2006 %T Motion Based Correspondence for 3D Tracking of Multiple Dim Objects %A Veeraraghavan,A. %A Srinivasan, M. %A Chellapa, Rama %A Baird, E. %A Lamont, R. %K 3D %K analysis;motion %K analysis;video %K based %K cameras;feature %K correspondence;motion %K dim %K extraction;image %K extraction;multiple %K features %K MOTION %K objects;video %K processing; %K signal %K tracking;motion %X Tracking multiple objects in a video is a demanding task that is frequently encountered in several systems such as surveillance and motion analysis. Ability to track objects in 3D requires the use of multiple cameras. While tracking multiple objects using multiples video cameras, establishing correspondence between objects in the various cameras is a nontrivial task. Specifically, when the targets are dim or are very far away from the camera, appearance cannot be used in order to establish this correspondence. Here, we propose a technique to establish correspondence across cameras using the motion features extracted from the targets, even when the relative position of the cameras is unknown. Experimental results are provided for the problem of tracking multiple bees in natural flight using two cameras. The reconstructed 3D flight paths of the bees show some interesting flight patterns %B Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on %V 2 %P II - II %8 2006/05// %G eng %R 10.1109/ICASSP.2006.1660431 %0 Book Section %B Computer Vision – ACCV 2006Computer Vision – ACCV 2006 %D 2006 %T Multi-camera Tracking of Articulated Human Motion Using Motion and Shape Cues %A Sundaresan,Aravind %A Chellapa, Rama %E Narayanan,P. %E Nayar,Shree %E Shum,Heung-Yeung %X We present a framework and algorithm for tracking articulated motion for humans. We use multiple calibrated cameras and an articulated human shape model. Tracking is performed using motion cues as well as image-based cues (such as silhouettes and “motion residues” hereafter referred to as spatial cues,) as opposed to constructing a 3D volume image or visual hulls. Our algorithm consists of a predictor and corrector: the predictor estimates the pose at the t + 1 using motion information between images at t and t + 1. The error in the estimated pose is then corrected using spatial cues from images at t + 1. In our predictor, we use robust multi-scale parametric optimisation to estimate the pixel displacement for each body segment. We then use an iterative procedure to estimate the change in pose from the pixel displacement of points on the individual body segments. We present a method for fusing information from different spatial cues such as silhouettes and “motion residues” into a single energy function. We then express this energy function in terms of the pose parameters, and find the optimum pose for which the energy is minimised. %B Computer Vision – ACCV 2006Computer Vision – ACCV 2006 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3852 %P 131 - 140 %8 2006/// %@ 978-3-540-31244-4 %G eng %U http://dx.doi.org/10.1007/11612704_14 %0 Conference Paper %B Proceedings of the 2006 ACM/IEEE conference on Supercomputing %D 2006 %T Multiple range query optimization with distributed cache indexing %A Nam,B. %A Andrade,H. %A Sussman, Alan %B Proceedings of the 2006 ACM/IEEE conference on Supercomputing %P 100–es - 100–es %8 2006/// %G eng %0 Journal Article %J Network, IEEE %D 2006 %T Multiuser cross-layer resource allocation for video transmission over wireless networks %A Su,Guan-Ming %A Man,Zhu %A M. Wu %A Liu,K. J.R %K 3G %K allocation;multiuser %K allocation;video %K cellular %K channels;quality %K codecs;video %K coding;video %K communication;4G %K communication;data %K compression;diversity %K cross-layer %K diversity;physical %K layer;multiple %K layer;quality %K mobile %K networks;3G %K of %K reception;multiuser %K resource %K service;resource %K service;video %K streaming; %K streams;multiuser %K system;4G %K system;QoS;application %K technology;wireless %K transmission;video-compression %K video %X With the advancement of video-compression technology and the wide deployment of wireless networks, there is an increasing demand for wireless video communication services, and many design challenges remain to be overcome. In this article, we discuss how to dynamically allocate resources according to the changing environments and requirements, so as to improve the overall system performance and ensure individual quality of service (QoS). Specifically, we consider two aspects with regard to design issues: cross-layer design, which jointly optimizes resource utilization from the physical layer to the application layer, and multiuser diversity, which explores source and channel heterogeneity for different users. We study how to efficiently transmit multiple video streams, encoded by current and future video codecs, over resource-limited wireless networks such as 3G/4G cellular system and future wireless local/metropolitan area networks (WLANs/WMANs) %B Network, IEEE %V 20 %P 21 - 27 %8 2006/04//march %@ 0890-8044 %G eng %N 2 %R 10.1109/MNET.2006.1607892 %0 Journal Article %J Wireless Communications, IEEE Transactions on %D 2006 %T Multiuser distortion management of layered video over resource limited downlink multicode-CDMA %A Han,Z. %A Su,Guan-Ming %A Kwasinski,A. %A M. Wu %A Liu,K. J.R %K (mobile %K access;combined %K allocation;broadband %K case;layered %K CDMA %K cellular %K code %K codes;radio %K codes;real-time %K codes;telecommunication %K coding;mobility %K coding;video %K coding;wireless %K communication; %K control;pseudo-random %K control;pseudonoise %K control;video %K distortion %K division %K downlink %K encoded %K limited %K links;random %K management %K management;power %K multicode-CDMA;source-channel %K multiple %K networks;broadband %K networks;cellular %K radio);power %K radio;code %K source-channel %K technology;code-unlimited %K video;multiuser %K videos;resource %X Transmitting multiple real-time encoded videos to multiple users over wireless cellular networks is a key driving force for developing broadband technology. We propose a new framework to transmit multiple users' video programs encoded by MPEG-4 FGS codec over downlink multicode CDMA networks in real time. The proposed framework jointly manages the rate adaptation of source and channel coding, CDMA code allocation, and power control. Subject to the limited system resources, such as the number of pseudo-random codes and the maximal power for CDMA transmission, we develop an adaptive scheme of distortion management to ensure baseline video quality for each user and further reduce the overall distortion received by all users. To efficiently utilize system resources, the proposed scheme maintains a balanced ratio between the power and code usages. We also investigate three special scenarios where demand, power, or code is limited, respectively. Compared with existing methods in the literature, the proposed algorithm can reduce the overall system's distortion by 14% to 26%. In the demand-limited case and the code-limited but power-unlimited case, the proposed scheme achieves the optimal solutions. In the power-limited but code-unlimited case, the proposed scheme has a performance very close to a performance upper bound %B Wireless Communications, IEEE Transactions on %V 5 %P 3056 - 3067 %8 2006/11// %@ 1536-1276 %G eng %N 11 %R 10.1109/TWC.2006.258231 %0 Journal Article %J Computer Vision–ECCV 2006 %D 2006 %T Multivalued default logic for identity maintenance in visual surveillance %A Shet,V. %A Harwood,D. %A Davis, Larry S. %X Recognition of complex activities from surveillance video requires detection and temporal ordering of its constituent “atomic” events. It also requires the capacity to robustly track individuals and maintain their identities across single as well as multiple camera views. Identity maintenance is a primary source of uncertainty for activity recognition and has been traditionally addressed via different appearance matching approaches. However these approaches, by themselves, are inadequate. In this paper, we propose a prioritized, multivalued, default logic based framework that allows reasoning about the identities of individuals. This is achieved by augmenting traditional appearance matching with contextual information about the environment and self identifying traits of certain actions. This framework also encodes qualitative confidence measures for the identity decisions it takes and finally, uses this information to reason about the occurrence of certain predefined activities in video. %B Computer Vision–ECCV 2006 %P 119 - 132 %8 2006/// %G eng %0 Journal Article %J Probabilistic, Logical and Relational Learning-Towards a Synthesis %D 2006 %T Multi-View Learning and Link Farm Discovery %A Scheffer,T. %A De Raedt,L. %A Dietterich,T. %A Getoor, Lise %A Muggleton,S. H %X The first part of this abstract focuses on estimation of mixture models for problems in which multiple views of the instances are available. Examples of this setting include clustering web pages or research papers that have intrinsic (text) and extrinsic (references) attributes. Mixture model estimation is a key problem for both semi-supervised and unsupervised learning. An appropriate optimization criterion quantifies the likelihood and the consensus among models in the individual views; maximizing this consensus minimizes a bound on the risk of assigning an instance to an incorrect mixture component. An EM algorithm maximizes this criterion. The second part of this abstract focuses on the problem of identifying link spam. Search engine optimizers inflate the page rank of a target site by spinning an artificial web for the sole purpose of providing inbound links to the target. Discriminating natural from artificial web sites is a difficult multi-view problem. %B Probabilistic, Logical and Relational Learning-Towards a Synthesis %8 2006/// %G eng %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2006 %T Network Visualization by Semantic Substrates %A Shneiderman, Ben %A Aris,A. %K Automatic control %K data visualisation %K Data visualization %K Displays %K Filters %K Graphical user interfaces %K Information Visualization %K information visualization designers %K Law %K legal citations %K Legal factors %K legal precedent data %K network visualization %K NVSS 1.0 %K scalability %K semantic substrate %K Terminology %K Tunneling %K user-defined semantic substrates %X Networks have remained a challenge for information visualization designers because of the complex issues of node and link layout coupled with the rich set of tasks that users present. This paper offers a strategy based on two principles: (1) layouts are based on user-defined semantic substrates, which are non-overlapping regions in which node placement is based on node attributes, (2) users interactively adjust sliders to control link visibility to limit clutter and thus ensure comprehensibility of source and destination. Scalability is further facilitated by user control of which nodes are visible. We illustrate our semantic substrates approach as implemented in NVSS 1.0 with legal precedent data for up to 1122 court cases in three regions with 7645 legal citations %B IEEE Transactions on Visualization and Computer Graphics %V 12 %P 733 - 740 %8 2006/10//Sept %@ 1077-2626 %G eng %N 5 %R 10.1109/TVCG.2006.166 %0 Conference Paper %B Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on %D 2006 %T Non-Intrusive Forensic Analysis of Visual Sensors Using Output Images %A Swaminathan,A. %A M. Wu %A Liu,K. J.R %K algorithms;interpolation %K analysis;image %K analysis;output %K array %K cameras;forensic %K Color %K colour %K engineering;forensic %K forensic %K images;visual %K methods;nonintrusive %K PROCESSING %K sensor;digital %K sensors;cameras;image %K sensors;interpolation; %K signal %X This paper considers the problem of non-intrusive forensic analysis of the individual components in visual sensors and its implementation. As a new addition to the emerging area of forensic engineering, we present a framework for analyzing technologies employed inside digital cameras based on output images, and develop a set of forensic signal processing algorithms for visual sensors based on color array sensor and interpolation methods. We show through simulations that the proposed method is robust against compression and noise, and can help identify various processing components inside the camera. Such a non-intrusive forensic framework would provide useful evidence for analyzing technology infringement and evolution for visual sensors %B Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on %V 5 %P V - V %8 2006/05// %G eng %R 10.1109/ICASSP.2006.1661297 %0 Journal Article %J Linear Algebra and its Applications %D 2006 %T A note on generalized and hypergeneralized projectors %A Stewart, G.W. %K Generalized projection %K projection %K Spectral decomposition %X Groß and Trenkler [Generalized and hypergeneralized projectors, Linear Algebra Appl. 264 (1997) 463–474] have introduced two generalizations of orthogonal projectors called generalized projectors and hypergeneralized projectors. In this note we characterize these generalizations by their spectral decompositions. %B Linear Algebra and its Applications %V 412 %P 408 - 411 %8 2006/01/15/ %@ 0024-3795 %G eng %U http://www.sciencedirect.com/science/article/pii/S0024379505003836 %N 2–3 %R 10.1016/j.laa.2005.07.022 %0 Journal Article %J BrainBrain %D 2006 %T Nuclear Envelope Dystrophies Show a Transcriptional Fingerprint Suggesting Disruption of Rb–MyoD Pathways in Muscle Regeneration %A Bakay,Marina %A Wang,Zuyi %A Melcon,Gisela %A Schiltz,Louis %A Xuan,Jianhua %A Zhao,Po %A Sartorelli,Vittorio %A Seo,Jinwook %A Pegoraro,Elena %A Angelini,Corrado %A Shneiderman, Ben %A Escolar,Diana %A Chen,Yi-Wen %A Winokur,Sara T %A Pachman,Lauren M %A Fan,Chenguang %A Mandler,Raul %A Nevo,Yoram %A Gordon,Erynn %A Zhu,Yitan %A Dong,Yibin %A Wang,Yue %A Hoffman,Eric P %K EDMD = Emery–Dreifuss muscular dystrophy %K emerin %K Emery-Dreifuss muscular dystrophy %K FSHD = fascioscapulohumeral muscular dystrophy %K IDG = individual discriminatory genes %K JDG = jointly discriminatory genes %K lamin A/C %K LGMD = limb-girdle muscular dystrophy %K LOO = leave-one-out %K RT–PCR = reverse transcription–polymerase chain reaction; VISDA = Visual and Statistical Data Analyzer %K Skeletal muscle %K wFC = weighted Fisher criterion %X Mutations of lamin A/C (LMNA) cause a wide range of human disorders, including progeria, lipodystrophy, neuropathies and autosomal dominant Emery–Dreifuss muscular dystrophy (EDMD). EDMD is also caused by X-linked recessive loss-of-function mutations of emerin, another component of the inner nuclear lamina that directly interacts with LMNA. One model for disease pathogenesis of LMNA and emerin mutations is cell-specific perturbations of the mRNA transcriptome in terminally differentiated cells. To test this model, we studied 125 human muscle biopsies from 13 diagnostic groups (125 U133A, 125 U133B microarrays), including EDMD patients with LMNA and emerin mutations. A Visual and Statistical Data Analyzer (VISDA) algorithm was used to statistically model cluster hierarchy, resulting in a tree of phenotypic classifications. Validations of the diagnostic tree included permutations of U133A and U133B arrays, and use of two probe set algorithms (MAS5.0 and MBEI). This showed that the two nuclear envelope defects (EDMD LMNA, EDMD emerin) were highly related disorders and were also related to fascioscapulohumeral muscular dystrophy (FSHD). FSHD has recently been hypothesized to involve abnormal interactions of chromatin with the nuclear envelope. To identify disease-specific transcripts for EDMD, we applied a leave-one-out (LOO) cross-validation approach using LMNA patient muscle as a test data set, with reverse transcription–polymerase chain reaction (RT–PCR) validations in both LMNA and emerin patient muscle. A high proportion of top-ranked and validated transcripts were components of the same transcriptional regulatory pathway involving Rb1 and MyoD during muscle regeneration (CRI-1, CREBBP, Nap1L1, ECREBBP/p300), where each was specifically upregulated in EDMD. Using a muscle regeneration time series (27 time points) we develop a transcriptional model for downstream consequences of LMNA and emerin mutations. We propose that key interactions between the nuclear envelope and Rb and MyoD fail in EDMD at the point of myoblast exit from the cell cycle, leading to poorly coordinated phosphorylation and acetylation steps. Our data is consistent with mutations of nuclear lamina components leading to destabilization of the transcriptome in differentiated cells. %B BrainBrain %V 129 %P 996 - 1013 %8 2006/04/01/ %@ 0006-8950, 1460-2156 %G eng %U http://brain.oxfordjournals.org/content/129/4/996 %N 4 %R 10.1093/brain/awl023 %0 Journal Article %J Optics ExpressOpt. Express %D 2006 %T Numerical computation of the Green?s function for two-dimensional finite-size photonic crystals of infinite length %A Seydou,F. %A Ramahi,Omar M. %A Duraiswami, Ramani %A Seppänen,T. %K Diffraction and gratings %K Multiple scattering %X We develop a numerical algorithm that computes the Green’s function of Maxwell equation for a 2D finite-size photonic crystal, composed of rods of arbitrary shape. The method is based on the boundary integral equation, and a Nyström discretization is used for the numerical solution. To provide an exact solution that validates our code we derive multipole expansions for circular cylinders using our integral equation approach. The numerical method performs very well on the test case. We then apply it to crystals of arbitrary shape and discuss the convergence. %B Optics ExpressOpt. Express %V 14 %P 11362 - 11371 %8 2006/11/13/ %G eng %U http://www.opticsexpress.org/abstract.cfm?URI=oe-14-23-11362 %N 23 %R 10.1364/OE.14.011362 %0 Conference Paper %B Proc. AAAI Spring Symposium on Computational Approaches to Analyzing Weblogs, Stanford, CA %D 2006 %T Opinion Analysis in Document Databases %A Cesarano,C. %A Dorr, Bonnie J %A Picariello, A. %A Reforgiato,D. %A Sagoff,A. %A V.S. Subrahmanian %X There are numerous applications in which we would like toassess what opinions are being expressed in text documents. Forr example, Martha Stewart’s company may have wished to assess the degree of harshness of news articles about her in the recent past. Likewise, a World Bank official may wish to as- sess the degree of criticism of a proposed dam in Bangladesh. The ability to gauge opinion on a given topic is therefore of critical interest. In this paper, we develop a suite of algo- rithms which take as input, a set D of documents as well as a topic t, and gauge the degree of opinion expressed about topic t in the set D of documents. Our algorithms can return both a number (larger the number, more positive the opinion) as well as a qualitative opinion (e.g. harsh, complimentary). We as- sess the accuracy of these algorithms via human experiments and show that the best of these algorithms can accurately re- flect human opinions. We have also conducted performance experiments showing that our algorithms are computationally fast. %B Proc. AAAI Spring Symposium on Computational Approaches to Analyzing Weblogs, Stanford, CA %8 2006/// %G eng %0 Journal Article %J Modeling and Management of Geographical Data over Distributed Architectures. Springer–Verlag %D 2006 %T Out–of–Core Multiresolution Terrain Modeling %A Danovaro,E. %A De Floriani, Leila %A Puppo,E. %A Samet, Hanan %B Modeling and Management of Geographical Data over Distributed Architectures. Springer–Verlag %8 2006/// %G eng %0 Conference Paper %B PROCEEDINGS OF THE NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE %D 2006 %T Overconfidence or paranoia? search in imperfect-information games %A Parker,A. %A Nau, Dana S. %A V.S. Subrahmanian %X We derive a recursive formula for expected utility valuesin imperfect- information game trees, and an imperfect- information game tree search algorithm based on it. The for- mula and algorithm are general enough to incorporate a wide variety of opponent models. We analyze two opponent mod- els. The “paranoid” model is an information-set analog of the minimax rule used in perfect-information games. The “over- confident” model assumes the opponent moves randomly. Our experimental tests in the game of kriegspiel chess (an imperfect-information variant of chess) produced surpris- ing results: (1) against each other, and against one of the kriegspiel algorithms presented at IJCAI-05, the overconfi- dent model usually outperformed the paranoid model; (2) the performance of both models depended greatly on how well the model corresponded to the opponent’s behavior. These re- sults suggest that the usual assumption of perfect-information game tree search—that the opponent will choose the best pos- sible move—isn’t as useful in imperfect-information games. %B PROCEEDINGS OF THE NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE %V 21 %P 1045 - 1045 %8 2006/// %G eng %U http://www.aaai.org/Papers/AAAI/2006/AAAI06-164.pdf %0 Report %D 2006 %T OverDoSe: A generic DDoS protection service using an overlay network %A Elaine Shi %A Stoica,I. %A Andersen,D.G. %A Perrig, A. %X We present the design and implementation of OverDoSe, an overlay network offering generic DDoS protection for targeted sites. OverDoSe clients and servers are isolated at the IP level. Overlay nodes route packets between a client and a server, and regulate traffic according to the server’s instructions. Through the use of light-weight security primitives, OverDoSe achieves resilience against compromised overlay nodes with a minimal performance overhead. OverDoSe can be deployed by a single ISP who wishes to offer DDoS protection as a value-adding service to its customers. %I School of Computer Science, Carnegie Mellon University %8 2006 %@ CMU-CS-06-114 %G eng %U http://repository.cmu.edu/cgi/viewcontent.cgi?article=1073&context=compsci %R Technical Report %0 Conference Paper %B Computer Vision Systems, 2006 ICVS '06. IEEE International Conference on %D 2006 %T Parametric Hand Tracking for Recognition of Virtual Drawings %A Sepehri,A. %A Yacoob,Yaser %A Davis, Larry S. %X A hand tracking system for recognition of virtual spatial drawings is presented. Using a stereo camera, the 3D position of the hand in space is estimated. Then, by tracking the central region of the hand in 3D and estimating a virtual plane in space, the intended drawing of the user is recognized. Experimental results demonstrate the accuracy and effectiveness of this technique. The system can be used to communicate drawings and alphabets to a computer where a classifier can transform the drawn alphabets into interpretable characters. %B Computer Vision Systems, 2006 ICVS '06. IEEE International Conference on %P 6 - 6 %8 2006/01// %G eng %R 10.1109/ICVS.2006.50 %0 Conference Paper %B Proceedings of the joint international conference on Measurement and modeling of computer systems %D 2006 %T Partially overlapped channels not considered harmful %A Mishra,A. %A Shrivastava,V. %A Banerjee,S. %A Arbaugh, William A. %B Proceedings of the joint international conference on Measurement and modeling of computer systems %P 63 - 74 %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 21st International Conference on Computational Linguistics and the 44th annual meeting of the Association for Computational Linguistics %D 2006 %T PCFGs with syntactic and prosodic indicators of speech repairs %A Hale,John %A Shafran,Izhak %A Yung,Lisa %A Dorr, Bonnie J %A Harper,Mary %A Krasnyanskaya,Anna %A Lease,Matthew %A Liu,Yang %A Roark,Brian %A Snover,Matthew %A Stewart,Robin %X A grammatical method of combining two kinds of speech repair cues is presented. One cue, prosodic disjuncture, is detected by a decision tree-based ensemble classifier that uses acoustic cues to identify where normal prosody seems to be interrupted (Lickley, 1996). The other cue, syntactic parallelism, codifies the expectation that repairs continue a syntactic category that was left unfinished in the reparandum (Levelt, 1983). The two cues are combined in a Treebank PCFG whose states are split using a few simple tree transformations. Parsing performance on the Switchboard and Fisher corpora suggests that these two cues help to locate speech repairs in a synergistic way. %B Proceedings of the 21st International Conference on Computational Linguistics and the 44th annual meeting of the Association for Computational Linguistics %S ACL-44 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 161 - 168 %8 2006/// %G eng %U http://dx.doi.org/10.3115/1220175.1220196 %R 10.3115/1220175.1220196 %0 Conference Paper %B Proceedings of the 3rd conference on USENIX Workshop on Real, Large Distributed Systems - Volume 3 %D 2006 %T A platform for unobtrusive measurements on PlanetLab %A Sherwood,Rob %A Spring, Neil %X TCP Sidecar is a network measurement platform for injecting probes transparently into externally generated TCP streams. By coupling measurement probes with nonmeasurement traffic, we are able to obtain measurements behind NATs and firewalls without alerting intrusion detection systems. In this paper, we discuss Sidecar's design and our deployment experience on PlanetLab. We present preliminary results from Sidecar-based tools for RTT estimation ("sideping") and receiver-side bottleneck location ("artrat"). %B Proceedings of the 3rd conference on USENIX Workshop on Real, Large Distributed Systems - Volume 3 %S WORLDS'06 %I USENIX Association %C Berkeley, CA, USA %P 2 - 2 %8 2006/// %G eng %U http://dl.acm.org/citation.cfm?id=1254840.1254842 %0 Journal Article %J ITTC-FY2006-TR-45030-01, Information and Telecommunication Center, The University of Kansas %D 2006 %T Postmodern internetwork architecture %A Bhattacharjee, Bobby %A Calvert,K. %A Griffioen,J. %A Spring, Neil %A Sterbenz,J. P.G %X Network-layer innovation has proven surprisingly difficult, in part because internetworking protocols ignore com-peting economic interests and because a few protocols dominate, enabling layer violations that entrench technologies. Many shortcomings of today’s internetwork layer result from its inflexibility with respect to the policies of the stake- holders: users and service providers. The consequences of these failings are well-known: various hacks, layering violations, and overloadings are introduced to enforce policies and attempt to get the upper hand in various “tus- sles”. The result is a network that is increasingly brittle, hostile to innovation, vulnerable to attack, and insensitive to concerns about accountability and privacy. Our project aims to design, implement, and evaluate through daily use a minimalist internetwork layer and aux- iliary functionality that anticipates tussles and allows them to be played out in policy space, as opposed to in the packet-forwarding path. We call our approach postmodern internetwork architecture, because it is a reaction against many established network layer design concepts. The overall goal of the project is to make a larger portion of the network design space accessible without sacrificing the economy of scale offered by the unified Internet. We will use the postmodern architecture to explore basic architectural questions. These include: • What mechanisms should be supported by the network such that any foreseeable policy requirement can be explicitly addressed? • To what extent can routing and forwarding be isolated from each other while maintaining an efficient and usable network? • What forms of identity should be visible within the network, and what forms of accountability do different identities enable? • What mechanisms are needed to enable efficient access to cross-layer information and mechanisms such that lower layers can express their characteristics and upper layers can exert control downward? We plan to build and evaluate a complete end-to-end networking layer to help us understand feasible solutions to these questions. The Internet has fulfilled the potential of a complete generation of networking research by producing a global platform for innovation, commerce, and democracy. Unfortunately, the Internet also amply demonstrates the com- plexity and architectural ugliness that ensue when competing interests vie for benefits beyond those envisioned in the original design. This project is about redesigning the waist of the architectural hourglass to foster innovation, enhance security and accountability, and accomodate competing interests. %B ITTC-FY2006-TR-45030-01, Information and Telecommunication Center, The University of Kansas %8 2006/02// %G eng %0 Journal Article %J SIGPLAN Not. %D 2006 %T Practical dynamic software updating for C %A Neamtiu,Iulian %A Hicks, Michael W. %A Stoyle,Gareth %A Oriol,Manuel %K dynamic software updating %K function indirection %K loop extraction %K type wrapping %X Software updates typically require stopping and restarting an application, but many systems cannot afford to halt service, or would prefer not to. Dynamic software updating (DSU) addresses this difficulty by permitting programs to be updated while they run. DSU is appealing compared to other approaches for on-line upgrades because it is quite general and requires no redundant hardware. The challenge is in making DSU practical: it should be flexible, and yet safe, efficient, and easy to use.In this paper, we present Ginseng, a DSU implementation for C that aims to meet this challenge. We compile programs specially so that they can be dynamically patched, and generate most of a dynamic patch automatically. Ginseng performs a series of analyses that when combined with some simple runtime support ensure that an update will not violate type-safety while guaranteeing that data is kept up-to-date. We have used Ginseng to construct and dynamically apply patches to three substantial open-source server programs---Very Secure FTP daemon, OpenSSH sshd daemon, and GNU Zebra. In total, we dynamically patched each program with three years' worth of releases. Though the programs changed substantially, the majority of updates were easy to generate. Performance experiments show that all patches could be applied in less than 5 ms, and that the overhead on application throughput due to updating support ranged from 0 to at most 32%. %B SIGPLAN Not. %V 41 %P 72 - 83 %8 2006/06// %@ 0362-1340 %G eng %U http://doi.acm.org/10.1145/1133255.1133991 %N 6 %R 10.1145/1133255.1133991 %0 Journal Article %J Information Systems %D 2006 %T The priority curve algorithm for video summarization %A Albanese, M. %A Fayzullin,M. %A Picariello, A. %A V.S. Subrahmanian %K Content based retrieval %K Video databases %K Video summarization %X In this paper, we introduce the concept of a priority curve associated with a video. We then provide an algorithm that can use the priority curve to create a summary (of a desired length) of any video. The summary thus created exhibits nice continuity properties and also avoids repetition. We have implemented the priority curve algorithm (PriCA) and compared it with other summarization algorithms in the literature with respect to both performance and the output quality. The quality of summaries was evaluated by a group of 200 students in Naples, Italy, who watched soccer videos. We show that PriCA is faster than existing algorithms and also produces better quality summaries. We also briefly describe a soccer video summarization system we have built on using the PriCA architecture and various (classical) image processing algorithms. %B Information Systems %V 31 %P 679 - 695 %8 2006/11// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306437905001250 %N 7 %R 10.1016/j.is.2005.12.003 %0 Journal Article %J SIAM Journal on Computing %D 2006 %T A probabilistic analysis of trie-based sorting of large collections of line segments in spatial databases %A Lindenbaum,M. %A Samet, Hanan %A Hjaltason,G. R %B SIAM Journal on Computing %V 35 %P 22 - 58 %8 2006/// %G eng %N 1 %0 Journal Article %J CONCUR 2006–Concurrency Theory %D 2006 %T Probabilistic I/O automata: Theories of two equivalences %A Stark,E. %A Cleaveland, Rance %A Smolka,S. %B CONCUR 2006–Concurrency Theory %P 343 - 357 %8 2006/// %G eng %0 Journal Article %J Dagstuhl Seminar Proceedings %D 2006 %T Probabilistic, Logical and Relational Learning-Towards a Synthesis %A De Raedt,L. %A Dietterich,T. %A Getoor, Lise %A Muggleton,S. H %A Lloyd,J. W %A Sears,T. D %A Milch,B. %A Marthi,B. %A Russell,S. %A Sontag,D. %B Dagstuhl Seminar Proceedings %V 5051 %8 2006/// %G eng %0 Journal Article %J Journal of Parallel and Distributed Computing %D 2006 %T Provable algorithms for parallel generalized sweep scheduling %A Anil Kumar,V. S %A Marathe,M. V %A Parthasarathy,S. %A Srinivasan, Aravind %A Zust,S. %B Journal of Parallel and Distributed Computing %V 66 %P 807 - 821 %8 2006/// %G eng %N 6 %0 Journal Article %J Technical Report, CR-TR-4779, Department of Computer Science, University of Maryland %D 2006 %T Ranking search results in P2P systems %A Gopalakrishnan,V. %A Morselli,R. %A Bhattacharjee, Bobby %A Keleher,P. %A Srinivasan, Aravind %X P2P deployments are a natural infrastructure for buildingdistributed search networks. Proposed systems support locating and retrieving all results, but lack the information necessary to rank them. Users, however, are primarily interested in the most relevant, and not all possible results. Using random sampling, we extend a class of well- known information retrieval ranking algorithms such that they can be applied in this distributed setting. We analyze the overhead of our approach, and quantify exactly how our system scales with increasing number of documents, system size, document to node mapping (uniform versus non-uniform), and types of queries (rare versus popular terms). Our analysis and simulations show that a) these extensions are efficient, and can scale with little overhead to large systems, and b) the accuracy of the results ob- tained using distributed ranking is comparable to a cen- tralized implementation. %B Technical Report, CR-TR-4779, Department of Computer Science, University of Maryland %8 2006/// %G eng %0 Journal Article %J Technical Reports from UMIACS %D 2006 %T Ranking search results in peer-to-peer systems %A Gopalakrishnan,V. %A Morselli,R. %A Bhattacharjee, Bobby %A Keleher,P. %A Srinivasan, Aravind %X P2P deployments are a natural infrastructure for building distributed search networks. Proposed systems support locating and retrieving all results, but lack the information necessary to rank them. Users, however, are primarily interested in the most relevant, and not all possible results. Using random sampling, we extend a class of well-known information retrieval ranking algorithms such that they can be applied in this distributed setting. We analyze the overhead of our approach, and quantify exactly how our system scales with increasing number of documents, system size, document to node mapping (uniform versus non-uniform), and types of queries (rare versus popular terms). Our analysis and simulations show that a) these extensions are efficient, and can scale with little overhead to large systems, and b) the accuracy of the results obtained using distributed ranking is comparable to a centralized implementation. %B Technical Reports from UMIACS %8 2006/// %G eng %0 Book %D 2006 %T Real-Time Distributed Algorithms for Visual and Battlefield Reasoning %A Davis, Larry S. %A Basili, Victor R. %A V.S. Subrahmanian %A Reggia, James A. %A Aloimonos, J. %X Information is key to the success of the next generation battlefield. There is a critical need to determine, in real-time, what the enemy is doing, and to interpret that information in the context of past related events. In this project we examined two aspects of this issue: development of a high-level task definition language for tasking a network of sensors to carry out given objectives, and interpreting recounted events so that past related scenarios could be automatically identified from a case database. %I MARYLAND UNIV COLLEGE PARK OFFICE OF RESEARCH ADMINISTRATION AND ADVANCEMENT %8 2006/// %G eng %0 Conference Paper %B Biomedical Circuits and Systems Conference, 2006. BioCAS 2006. IEEE %D 2006 %T Reconfigurable image registration on FPGA platforms %A Sen,M. %A Hemaraj,Y. %A Bhattacharyya, Shuvra S. %A Shekhar,R. %B Biomedical Circuits and Systems Conference, 2006. BioCAS 2006. IEEE %P 154 - 157 %8 2006/// %G eng %0 Conference Paper %B System-on-Chip, 2006. International Symposium on %D 2006 %T Register File Partitioning with Constraint Programming %A Salmela,P. %A Shen,C. C %A Bhattacharyya, Shuvra S. %A Takala,J. %B System-on-Chip, 2006. International Symposium on %P 1 - 4 %8 2006/// %G eng %0 Conference Paper %B INFOCOM 2006. 25th IEEE International Conference on Computer Communications. Proceedings %D 2006 %T Relay Placement for Higher Order Connectivity in Wireless Sensor Networks %A Kashyap,A. %A Khuller, Samir %A Shayman,M. %X Not available %B INFOCOM 2006. 25th IEEE International Conference on Computer Communications. Proceedings %P 1 - 12 %8 2006/04// %G eng %R 10.1109/INFOCOM.2006.273 %0 Conference Paper %B 2006 IEEE International Conference on Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings %D 2006 %T Reranking for Sentence Boundary Detection in Conversational Speech %A Roark,B. %A Liu,Yang %A Harper,M. %A Stewart,R. %A Lease,M. %A Snover,M. %A Shafran,I. %A Dorr, Bonnie J %A Hale,J. %A Krasnyanskaya,A. %A Yung,L. %K Automatic speech recognition %K conversational speech %K data mining %K Ear %K EARS metadata extraction tasks %K Feature extraction %K hidden Markov models %K meta data %K Model driven engineering %K NIST %K NIST RT-04F community evaluation %K oracle accuracy %K performance evaluation %K reranking %K sentence-like unit boundary detection %K Speech processing %K Speech recognition %K Telephony %X We present a reranking approach to sentence-like unit (SU) boundary detection, one of the EARS metadata extraction tasks. Techniques for generating relatively small n-best lists with high oracle accuracy are presented. For each candidate, features are derived from a range of information sources, including the output of a number of parsers. Our approach yields significant improvements over the best performing system from the NIST RT-04F community evaluation %B 2006 IEEE International Conference on Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings %I IEEE %V 1 %P I-I - I-I %8 2006/05/14/19 %@ 1-4244-0469-X %G eng %R 10.1109/ICASSP.2006.1660078 %0 Journal Article %J Background and Methodology. US Department of Health and Human Services. Washington %D 2006 %T Research-based web design & usability guidelines %A Leavitt,M.O. %A Shneiderman, Ben %B Background and Methodology. US Department of Health and Human Services. Washington %8 2006/// %G eng %0 Journal Article %J IEEE/ACM Transactions on Networking %D 2006 %T Resilient multicast using overlays %A Banerjee,S. %A Lee,Seungjoon %A Bhattacharjee, Bobby %A Srinivasan, Aravind %K application-layer multicast protocols %K Computer science %K Data communication %K Delay %K Internet %K Internet-like topologies %K IP networks %K loss recovery technique %K Multicast %K multicast data recovery scheme %K Multicast protocols %K Network topology %K NETWORKS %K overlays %K Performance loss %K probabilistic forwarding %K probabilistic resilient multicast %K Protocols %K Resilience %K Streaming media %K telecommunication network topology %K Terminology %X We introduce Probabilistic Resilient Multicast (PRM): a multicast data recovery scheme that improves data delivery ratios while maintaining low end-to-end latencies. PRM has both a proactive and a reactive components; in this paper we describe how PRM can be used to improve the performance of application-layer multicast protocols especially when there are high packet losses and host failures. Through detailed analysis in this paper, we show that this loss recovery technique has efficient scaling properties-the overheads at each overlay node asymptotically decrease to zero with increasing group sizes. As a detailed case study, we show how PRM can be applied to the NICE application-layer multicast protocol. We present detailed simulations of the PRM-enhanced NICE protocol for 10 000 node Internet-like topologies. Simulations show that PRM achieves a high delivery ratio (>97%) with a low latency bound (600 ms) for environments with high end-to-end network losses (1%-5%) and high topology change rates (5 changes per second) while incurring very low overheads (<5%). %B IEEE/ACM Transactions on Networking %V 14 %P 237 - 248 %8 2006/04// %@ 1063-6692 %G eng %N 2 %R 10.1109/TNET.2006.872579 %0 Conference Paper %B Proceedings of the 7th IEEE/ACM International Conference on Grid Computing %D 2006 %T Resource Discovery Techniques in Distributed Desktop Grid Environments %A Kim,Jik-Soo %A Nam,Beomseok %A Keleher,Peter %A Marsh,Michael %A Bhattacharjee, Bobby %A Sussman, Alan %X Desktop grids use opportunistic sharing to exploit large collections of personal computers and workstations across the Internet, achieving tremendous computing power at low cost. Traditional desktop grid systems are typically based on a client-server architecture, which has inherent shortcomings with respect to robustness, reliability and scalability. In this paper, we propose a decentralized, robust, highly available, and scalable infrastructure to match incoming jobs to available resources. Through a comparative analysis on the experimental results obtained via simulation of three different types of matchmaking algorithms under different workload scenarios, we show the trade-offs between effcient matchmaking and good load balancing in a fully decentralized, heterogeneous computational environment. %B Proceedings of the 7th IEEE/ACM International Conference on Grid Computing %S GRID '06 %I IEEE Computer Society %C Washington, DC, USA %P 9 - 16 %8 2006/// %@ 1-4244-0343-X %G eng %U http://dx.doi.org/10.1109/ICGRID.2006.310992 %R 10.1109/ICGRID.2006.310992 %0 Journal Article %J PLoS PathogPLoS Pathog %D 2006 %T Retroviral DNA Integration: Viral and Cellular Determinants of Target-Site Selection %A Lewinski,Mary K %A Yamashita,Masahiro %A Emerman,Michael %A Ciuffi,Angela %A Marshall,Heather %A Crawford,Gregory %A Collins,Francis %A Shinn,Paul %A Leipzig,Jeremy %A Hannenhalli, Sridhar %A Berry,Charles C %A Ecker,Joseph R %A Bushman,Frederic D. %X Retroviruses differ in their preferences for sites for viral DNA integration in the chromosomes of infected cells. Human immunodeficiency virus (HIV) integrates preferentially within active transcription units, whereas murine leukemia virus (MLV) integrates preferentially near transcription start sites and CpG islands. We investigated the viral determinants of integration-site selection using HIV chimeras with MLV genes substituted for their HIV counterparts. We found that transferring the MLV integrase (IN) coding region into HIV (to make HIVmIN) caused the hybrid to integrate with a specificity close to that of MLV. Addition of MLV gag (to make HIVmGagmIN) further increased the similarity of target-site selection to that of MLV. A chimeric virus with MLV Gag only (HIVmGag) displayed targeting preferences different from that of both HIV and MLV, further implicating Gag proteins in targeting as well as IN. We also report a genome-wide analysis indicating that MLV, but not HIV, favors integration near DNase I–hypersensitive sites (i.e., +/− 1 kb), and that HIVmIN and HIVmGagmIN also favored integration near these features. These findings reveal that IN is the principal viral determinant of integration specificity; they also reveal a new role for Gag-derived proteins, and strengthen models for integration targeting based on tethering of viral IN proteins to host proteins. %B PLoS PathogPLoS Pathog %V 2 %P e60 - e60 %8 2006/06/23/ %G eng %U UR - http://dx.plos.org/10.1371/journal.ppat.0020060,http://dx.plos.org/10.1371/journal.ppat.0020060 %N 6 %R 10.1371/journal.ppat.0020060 %0 Journal Article %J SIGACT News %D 2006 %T Review of "The Random Projection Method by Santosh Vempala" %A Srinivasan, Aravind %B SIGACT News %V 37 %P 41 - 43 %8 2006/12// %@ 0163-5700 %G eng %U http://doi.acm.org/10.1145/1189056.1189066 %N 4 %R 10.1145/1189056.1189066 %0 Report %D 2006 %T Revisiting Internet addressing: Back to the future %A Vutukuru,M. %A Feamster, Nick %A Walfish,M. %A Balakrishnan,H. %A Shenker,S. %X IP prefixes undermine three goals of Internet routing:accurate reflection of network-layer reachability, secure routing messages, and effective traffic control. This pa- per presents Atomic IP (AIP), a simple change to Inter- net addressing (which in fact reverts to how addressing once worked), that allows Internet routing to achieve these goals. %I Massachusetts Institute of Technology Computer Science and Artificial Intelligence Laboratory %V MIT-CSAIL-TR-2006-025 %8 2006/// %G eng %0 Journal Article %J Information Forensics and Security, IEEE Transactions on %D 2006 %T Robust and secure image hashing %A Swaminathan,A. %A Mao,Yinian %A M. Wu %K content-preserving %K cryptography; %K differential %K distortions; %K entropy; %K Filtering %K Fourier %K functions; %K hash %K hashing; %K image %K modifications; %K processing; %K secure %K theory; %K transform; %K transforms; %X Image hash functions find extensive applications in content authentication, database search, and watermarking. This paper develops a novel algorithm for generating an image hash based on Fourier transform features and controlled randomization. We formulate the robustness of image hashing as a hypothesis testing problem and evaluate the performance under various image processing operations. We show that the proposed hash function is resilient to content-preserving modifications, such as moderate geometric and filtering distortions. We introduce a general framework to study and evaluate the security of image hashing systems. Under this new framework, we model the hash values as random variables and quantify its uncertainty in terms of differential entropy. Using this security framework, we analyze the security of the proposed schemes and several existing representative methods for image hashing. We then examine the security versus robustness tradeoff and show that the proposed hashing methods can provide excellent security and robustness. %B Information Forensics and Security, IEEE Transactions on %V 1 %P 215 - 230 %8 2006/06// %@ 1556-6013 %G eng %N 2 %R 10.1109/TIFS.2006.873601 %0 Journal Article %J Advances in Cryptology-CRYPTO 2006 %D 2006 %T Robust fuzzy extractors and authenticated key agreement from close secrets %A Dodis,Y. %A Katz, Jonathan %A Reyzin,L. %A Smith,A. %X Consider two parties holding correlated random variables W and W′, respectively, that are within distance t of each other in some metric space. These parties wish to agree on a uniformly distributed secret key R by sending a single message over an insecure channel controlled by an all-powerful adversary. We consider both the keyless case, where the parties share no additional secret information, and the keyed case, where the parties share a long-term secret SK that they can use to generate a sequence of session keys {R j } using multiple pairs {(W j , W′ j )}. The former has applications to, e.g., biometric authentication, while the latter arises in, e.g., the bounded storage model with errors.Our results improve upon previous work in several respects: – The best previous solution for the keyless case with no errors (i.e., t=0) requires the min-entropy of W to exceed 2|W|/3. We show a solution when the min-entropy of W exceeds the minimal threshold |W|/2. – Previous solutions for the keyless case in the presence of errors (i.e., t>0) required random oracles. We give the first constructions (for certain metrics) in the standard model. – Previous solutions for the keyed case were stateful. We give the first stateless solution. %B Advances in Cryptology-CRYPTO 2006 %P 232 - 250 %8 2006/// %G eng %R 10.1007/11818175_14 %0 Conference Paper %B Proceedings of the IEEE MILCOM %D 2006 %T Routing for data delivery in dynamic networks %A Mundur, Padma %A Lee,S. %A Seligman,M. %B Proceedings of the IEEE MILCOM %P 1 - 7 %8 2006/// %G eng %0 Report %D 2006 %T Routing in Delay Tolerant Networks Using Storage Domains %A Mundur, Padma %A Lee,Sookyoung %A Seligman,Matthew %K Technical Report %X In this paper, we present a routing algorithm for a class ofdynamic networks called the Delay Tolerant Networks (DTNs). The proposed algorithm takes into account the quintessential DTN characteristic namely, intermittent link connectivity. We modify the breadth first search (BFS) algorithm to take into account link state changes and find the quickest route between source and destination nodes. We adopt a message drop policy at intermediate nodes to incorporate storage constraint. We also introduce the idea of time-varying storage domains where all nodes connected for a length of time act as a single storage unit by sharing the aggregated storage capacity of the nodes. We evaluate the routing algorithm with and without storage domain in an extensive simulation. We analyze the performance using metrics such as delivery ratio, incomplete transfers with no routes and dropped messages. The DTN topology dynamics are analyzed by varying: number of nodes generating traffic, link probability, link availability through combinations of downtime/uptime values, storage per node, message size, and traffic. The delay performance of the proposed algorithms is conceptually the same as flooding-based algorithms but without the penalty of multiple copies. More significantly, we show that the Quickest Storage Domain (Quickest SD) algorithm distributes the storage demand across many nodes in the network topology, enabling balanced load and higher network utilization. In fact, we show that for the same level of performance, we can actually cut the storage requirement in half using the Quickest SD algorithm. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2007-01 %8 2006/11/20/ %G eng %U http://drum.lib.umd.edu//handle/1903/4024 %0 Conference Paper %B Proceedings of the 9th ACM international symposium on Modeling analysis and simulation of wireless and mobile systems %D 2006 %T Routing in intermittent network topologies %A Mundur, Padma %A Lee,Sookyoung %A Seligman,Matthew %K delay tolerant network (DTN) %K performance evaluation %K routing algorithms %B Proceedings of the 9th ACM international symposium on Modeling analysis and simulation of wireless and mobile systems %S MSWiM '06 %I ACM %C New York, NY, USA %P 385 - 389 %8 2006/// %@ 1-59593-477-4 %G eng %U http://doi.acm.org/10.1145/1164717.1164782 %R 10.1145/1164717.1164782 %0 Journal Article %J Science of Computer Programming %D 2006 %T Safe manual memory management in Cyclone %A Swamy,Nikhil %A Hicks, Michael W. %A Morrisett,Greg %A Grossman,Dan %A Jim,Trevor %K cyclone %K Memory management %K memory safety %K Reaps %K Reference counting %K regions %K unique pointers %X The goal of the Cyclone project is to investigate how to make a low-level C-like language safe. Our most difficult challenge has been providing programmers with control over memory management while retaining safety. This paper describes our experience trying to integrate and use effectively two previously-proposed, safe memory-management mechanisms: statically-scoped regions and tracked pointers. We found that these typing mechanisms can be combined to build alternative memory-management abstractions, such as reference counted objects and arenas with dynamic lifetimes, and thus provide a flexible basis. Our experience — porting C programs and device drivers, and building new applications for resource-constrained systems — confirms that experts can use these features to improve memory footprint and sometimes to improve throughput when used instead of, or in combination with, conservative garbage collection. %B Science of Computer Programming %V 62 %P 122 - 144 %8 2006/10/01/ %@ 0167-6423 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167642306000785 %N 2 %R 10.1016/j.scico.2006.02.003 %0 Conference Paper %B Proceedings of the 2006 international conference on Digital government research %D 2006 %T Scalable data collection and retrieval infrastructure for digital government applications %A Samet, Hanan %A Golubchik,Leana %X In this paper we describe highlights of the project titled "Scalable data collection infrastructure for digital government applications" under the auspices of the Digital Government Research Program of the National Science Foundation. Our research is focused on taking advantage of the distributed nature of data and the interaction with it. Our efforts have been directed at both the systems/theoretical and applications levels. On the systems and theoretical levels, we have continued our development of the BISTRO system (Section 2). On the applications level, work has commenced on the development of a mechanism for spatially tagging text documents for retrieval by search engines based on both content and spatial proximity (Section 3). %B Proceedings of the 2006 international conference on Digital government research %S dg.o '06 %I ACM %C New York, NY, USA %P 301 - 302 %8 2006/// %G eng %U http://doi.acm.org/10.1145/1146598.1146679 %R 10.1145/1146598.1146679 %0 Journal Article %J Circuits and Systems for Video Technology, IEEE Transactions on %D 2006 %T A Scalable Multiuser Framework for Video Over OFDM Networks: Fairness and Efficiency %A Su,Guan-Ming %A Han,Zhu %A M. Wu %A Liu,K. J.R %K algorithm;video %K codec;video %K coding;video %K communication; %K deviation;downlink %K deviation;OFDM %K distortion;multidimensional %K diversity;multiple %K division %K end-to-end %K framework;time %K Frequency %K modulation;diversity %K multiple %K multiplex %K multiuser %K network;efficiency %K networks;video %K objectives;optimization %K of %K OFDM %K orthogonal %K over %K problem;fairness %K problem;maximal %K problems;quality %K programs;multiuser %K PSNR %K QUALITY %K reception;image %K resolution;optimisation;quality %K scalable %K service %K service;scalable %K service;video %K systems;optimal %K video %X In this paper, we propose a framework to transmit multiple scalable video programs over downlink multiuser orthogonal frequency division multiplex (OFDM) networks in real time. The framework explores the scalability of the video codec and multidimensional diversity of multiuser OFDM systems to achieve the optimal service objectives subject to constraints on delay and limited system resources. We consider two essential service objectives, namely, the fairness and efficiency. Fairness concerns the video quality deviation among users who subscribe the same quality of service, and efficiency relates to how to attain the highest overall video quality using the available system resources. We formulate the fairness problem as minimizing the maximal end-to-end distortion received among all users and the efficiency problem as minimizing total end-to-end distortion of all users. Fast suboptimal algorithms are proposed to solve the above two optimization problems. The simulation results demonstrated that the proposed fairness algorithm outperforms a time division multiple (TDM) algorithm by 0.5 3 dB in terms of the worst received video quality among all users. In addition, the proposed framework can achieve a desired tradeoff between fairness and efficiency. For achieving the same average video quality among all users, the proposed framework can provide fairer video quality with 1 1.8 dB lower PSNR deviation than a TDM algorithm %B Circuits and Systems for Video Technology, IEEE Transactions on %V 16 %P 1217 - 1231 %8 2006/10// %@ 1051-8215 %G eng %N 10 %R 10.1109/TCSVT.2006.883513 %0 Journal Article %J Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %D 2006 %T Seasonal Cholera Caused by Vibrio Cholerae Serogroups O1 and O139 in the Coastal Aquatic Environment of Bangladesh %A Alam,Munirul %A Hasan,Nur A. %A Sadique,Abdus %A Bhuiyan,N. A. %A Ahmed,Kabir U. %A Nusrin,Suraia %A Nair,G. Balakrish %A Siddique,A. K. %A Sack,R. Bradley %A Sack,David A. %A Huq,Anwar %A Rita R Colwell %X Since Vibrio cholerae O139 first appeared in 1992, both O1 El Tor and O139 have been recognized as the epidemic serogroups, although their geographic distribution, endemicity, and reservoir are not fully understood. To address this lack of information, a study of the epidemiology and ecology of V. cholerae O1 and O139 was carried out in two coastal areas, Bakerganj and Mathbaria, Bangladesh, where cholera occurs seasonally. The results of a biweekly clinical study (January 2004 to May 2005), employing culture methods, and of an ecological study (monthly in Bakerganj and biweekly in Mathbaria from March 2004 to May 2005), employing direct and enrichment culture, colony blot hybridization, and direct fluorescent-antibody methods, showed that cholera is endemic in both Bakerganj and Mathbaria and that V. cholerae O1, O139, and non-O1/non-O139 are autochthonous to the aquatic environment. Although V. cholerae O1 and O139 were isolated from both areas, most noteworthy was the isolation of V. cholerae O139 in March, July, and September 2004 in Mathbaria, where seasonal cholera was clinically linked only to V. cholerae O1. In Mathbaria, V. cholerae O139 emerged as the sole cause of a significant outbreak of cholera in March 2005. V. cholerae O1 reemerged clinically in April 2005 and established dominance over V. cholerae O139, continuing to cause cholera in Mathbaria. In conclusion, the epidemic potential and coastal aquatic reservoir for V. cholerae O139 have been demonstrated. Based on the results of this study, the coastal ecosystem of the Bay of Bengal is concluded to be a significant reservoir for the epidemic serogroups of V. cholerae. %B Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %V 72 %P 4096 - 4104 %8 2006/06/01/ %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/72/6/4096 %N 6 %R 10.1128/AEM.00066-06 %0 Conference Paper %B Pattern Recognition, 2006. ICPR 2006. 18th International Conference on %D 2006 %T Segmentation and Probabilistic Registration of Articulated Body Models %A Sundaresan, A. %A Chellapa, Rama %K (mathematics); %K and %K approach;eigenspace;image %K articulated %K body %K data;voxel-based %K eigenfunctions;graph %K estimation;probabilistic %K fitting;voxel %K graph;pose %K models;bottom-up %K registration;eigenvalues %K registration;image %K registration;spline %K segmentation;neighbourhood %K segmentation;probability;splines %K theory;image %X There are different approaches to pose estimation and registration of different body parts using voxel data. We propose a general bottom-up approach in order to segment the voxels into different body parts. The voxels are first transformed into a high dimensional space which is the eigenspace of the Laplacian of the neighbourhood graph. We exploit the properties of this transformation and fit splines to the voxels belonging to different body segments in eigenspace. The boundary of the splines is determined by examination of the error in spline fitting. We then use a probabilistic approach to register the segmented body segments by utilizing their connectivity and prior knowledge of the general structure of the subjects. We present results on real data, containing both simple and complex poses. While we use human subjects in our experiment, the method is fairly general and can be applied to voxel-based registration of any articulated or non-rigid object composed of primarily 1-D parts %B Pattern Recognition, 2006. ICPR 2006. 18th International Conference on %V 2 %P 92 - 96 %8 2006/// %G eng %R 10.1109/ICPR.2006.1034 %0 Conference Paper %B Proceedings of the 5th international conference on Information processing in sensor networks %D 2006 %T A sensory grammar for inferring behaviors in sensor networks %A Lymberopoulos, Dimitrios %A Ogale, Abhijit S. %A Savvides, Andreas %A Aloimonos, J. %K algorithms %K behavior identification %K human activity %K human factors %K pcfg %K real-time and embedded systems %K sensor grammars %K sensor networks %X The ability of a sensor network to parse out observable activities into a set of distinguishable actions is a powerful feature that can potentially enable many applications of sensor networks to everyday life situations. In this paper we introduce a framework that uses a hierarchy of Probabilistic Context Free Grammars (PCFGs) to perform such parsing. The power of the framework comes from the hierarchical organization of grammars that allows the use of simple local sensor measurements for reasoning about more macroscopic behaviors. Our presentation describes how to use a set of phonemes to construct grammars and how to achieve distributed operation using a messaging model. The proposed framework is flexible. It can be mapped to a network hierarchy or can be applied sequentially and across the network to infer behaviors as they unfold in space and time. We demonstrate this functionality by inferring simple motion patterns using a sequence of simple direction vectors obtained from our camera sensor network testbed. %B Proceedings of the 5th international conference on Information processing in sensor networks %S IPSN '06 %I ACM %C Nashville, Tennessee, USA %P 251 - 259 %8 2006/// %@ 1-59593-334-4 %G eng %R 10.1145/1127777.1127817 %0 Journal Article %J Proceedings of the 2006 Document Understanding Workshop, New York %D 2006 %T Sentence compression as a component of a multi-document summarization system %A Zajic, David %A Dorr, Bonnie J %A Jimmy Lin %A Schwartz,R. %X We applied a single-document sentence-trimming approach (Trimmer) to the problem of multi-document summariza- tion. Trimmer was designed with the in- tention of compressing a lead sentence into a space consisting of tens of char- acters. In our Multi-Document Trimmer (MDT), we use Trimmer to generate multiple trimmed candidates for each sentence. Sentence selection is used to determine which trimmed candidates provide the best combination of topic coverage and brevity. We demonstrate that we were able to port Trimmer easily to this new problem. We also show that MDT generally ranked higher for recall than for precision, suggesting that MDT is currently more successful at finding relevant content than it is at weeding out irrelevant content. Finally, we present an error analysis that shows that, while sentence compressions is making space for additional sentences, more work is needed in the area of generating and se- lecting the right candidates. %B Proceedings of the 2006 Document Understanding Workshop, New York %8 2006/// %G eng %0 Book Section %B DUC 06 Conference Proceedings %D 2006 %T Sentence Trimming and Selection: Mixing and Matching %A Zajic, David %A Dorr, Bonnie J %A Jimmy Lin %A O'Leary, Dianne P. %A Conroy,John M. %A Schlesinger,Judith D. %X We describe how components from two distinct multi-document summarization systems were combined. Twenty four possible combinations of components were considered. We observed some contrasts between conservative and aggressive sentence compression (i.e., trimming) in the context of multi-document summarization. %B DUC 06 Conference Proceedings %I U.S. National Inst. of Standards and Technology %8 2006/// %G eng %U http://duc.nist.gov/ %0 Journal Article %J FEMS Microbiology Letters %D 2006 %T Septaplex PCR assay for rapid identification of Vibrio cholerae including detection of virulence and int SXT genes %A Mantri,Chinmay K. %A Mohapatra,Saswat S. %A Ramamurthy,Thandavarayan %A Ghosh,Raikamal %A Rita R Colwell %A Singh,Durg V. %K DETECTION %K intsxt %K septaplex PCR %K Vibrio cholerae %K virulence %X In this study, we describe a septaplex PCR assay for rapid identification of Vibrio cholerae including detection of the virulence and intsxt genes. Conditions were optimized to amplify fragments of ISRrRNA (encoding for 16S–23S rRNA gene, Intergenic spacer regions), O1rfb (O1 serogroup specific rfb), O139rfb (O139 serogroup specific rfb), ctxA (cholera toxin subunit A), tcpA (toxin coregulated pilus), and intsxt (sxt integron) simultaneously in a single PCR. The septaplex PCR was evaluated using 211 strains of V. cholerae and six water samples for in situ testing. PCR results were correlated with genotype data obtained by individual PCR and slot-blot assays. The one-step PCR described here can be used to identify V. cholerae accurately and rapidly. Also, the virulence and intsxt genes can be simultaneously detected, providing a useful method for monitoring pathogenic, intsxt-positive and nonpathogenic, intsxt-negative V. cholerae serogroups both in the environment and clinical settings. %B FEMS Microbiology Letters %V 265 %P 208 - 214 %8 2006/12/01/ %@ 1574-6968 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1574-6968.2006.00491.x/full %N 2 %R 10.1111/j.1574-6968.2006.00491.x %0 Journal Article %J Lecture Notes in Computer Science %D 2006 %T Session 8A-Combinatorial Optimization and Quantum Computing-Lower Bounds on the Deterministic and Quantum Communication Complexities of Hamming-Distance Problems %A Ambainis,A. %A Gasarch,W. %A Srinivasan, Aravind %A Utis,A. %B Lecture Notes in Computer Science %V 4288 %P 628 - 637 %8 2006/// %G eng %0 Journal Article %J Interacting with Computers %D 2006 %T Severity and impact of computer user frustration: A comparison of student and workplace users %A Lazar,Jonathan %A Jones,Adam %A Hackley,Mary %A Shneiderman, Ben %K Computer anxiety %K Computer experience %K Helpdesk %K Training %K User frustration %K user interface design %X User frustration with information and computing technology is a pervasive and persistent problem. When computers crash, network congestion causes delays, and poor user interfaces trigger confusion there are dramatic consequences for individuals, organizations, and society. These frustrations, not only cause personal dissatisfaction and loss of self-efficacy, but may disrupt workplaces, slow learning, and reduce participation in local and national communities. Our exploratory study of 107 student computer users and 50 workplace computer users shows high levels of frustration and loss of 1/3–1/2 of time spent. This paper reports on the incident and individual factors that cause of frustration, and how they raise frustration severity. It examines the frustration impacts on the daily interactions of the users. The time lost and time to fix problem, and importance of task, strongly correlate with frustration levels for both student and workplace users. Differences between students and workplace users are discussed in the paper, as are implications for researchers. %B Interacting with Computers %V 18 %P 187 - 207 %8 2006/03// %@ 0953-5438 %G eng %U http://www.sciencedirect.com/science/article/pii/S0953543805000561 %N 2 %R 10.1016/j.intcom.2005.06.001 %0 Conference Paper %B Image Processing, 2006 IEEE International Conference on %D 2006 %T Shape-Regulated Particle Filtering for Tracking Non-Rigid Objects %A Jie Shao %A Chellapa, Rama %A Porikli, F. %K (numerical %K 2D %K algorithm;affine %K based %K contour %K detection;dynamic %K detection;motion %K detection;parameter %K estimation;active %K estimation;object %K estimation;particle %K Filtering %K filtering;probabilistic %K filters; %K map;shape-regulation;affine %K methods);tracking %K model;nonrigid %K MOTION %K object %K scenes;deformation %K tracking;parameter %K transform;cluttered %K transforms;clutter;edge %X This paper presents an active contour based algorithm for tracking non-rigid objects in heavily cluttered scenes. We decompose the non-rigid contour tracking problem into three subproblems: 2D motion estimation, deformation detection, and shape regulation. First, we employ a particle filter to estimate the affine transform parameters between successive frames. Second, by using a dynamic object model, we generate a probabilistic map of deformation to reshape its contour. Finally, we project the updated model onto a trained shape subspace to constrain deformations to be within possible object appearances. Our experiments show that the proposed algorithm significantly improves the performance of the tracker %B Image Processing, 2006 IEEE International Conference on %P 2813 - 2816 %8 2006/10// %G eng %R 10.1109/ICIP.2006.312993 %0 Book Section %B Geometric Modeling and Processing - GMP 2006 %D 2006 %T A Step Towards Automated Design of Side Actions in Injection Molding of Complex Parts %A Banerjee,Ashis %A Gupta, Satyandra K. %E Kim,Myung-Soo %E Shimada,Kenji %K Computer science %X Side actions contribute to mold cost by resulting in an additional manufacturing and assembly cost as well as by increasing the molding cycle time. Therefore, generating shapes of side actions requires solving a complex geometric optimization problem. Different objective functions may be needed depending upon different molding scenarios (e.g., prototyping versus large production runs). Manually designing side actions is a challenging task and requires considerable expertise. Automated design of side actions will significantly reduce mold design lead times. This paper describes algorithms for generating shapes of side actions to minimize a customizable molding cost function. %B Geometric Modeling and Processing - GMP 2006 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4077 %P 500 - 513 %8 2006/// %@ 978-3-540-36711-6 %G eng %U http://www.springerlink.com/content/w7q3212607500635/abstract/ %0 Conference Paper %B Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems %D 2006 %T A stochastic language for modelling opponent agents %A Simari,G. %A Sliva,A. %A Nau, Dana S. %A V.S. Subrahmanian %X There are numerous cases where a reasoning agent needs toreason about the behavior of an opponent agent. In this pa- per, we propose a hybrid probabilistic logic language within which we can express what actions an opponent may take in a given situation. We present the syntaxis and semantics of the language, and the concept of a Maximally Probable Course of Action. %B Proceedings of the fifth international joint conference on Autonomous agents and multiagent systems %P 244 - 246 %8 2006/// %G eng %0 Conference Paper %B Proceedings of the 2006 AVI workshop on BEyond time and errors: novel evaluation methods for information visualization %D 2006 %T Strategies for evaluating information visualization tools: multi-dimensional in-depth long-term case studies %A Shneiderman, Ben %A Plaisant, Catherine %X After an historical review of evaluation methods, we describe an emerging research method called Multi-dimensional In-depth Long-term Case studies (MILCs) which seems well adapted to study the creative activities that users of information visualization systems engage in. We propose that the efficacy of tools can be assessed by documenting 1) usage (observations, interviews, surveys, logging etc.) and 2) expert users' success in achieving their professional goals. We summarize lessons from related ethnography methods used in HCI and provide guidelines for conducting MILCs for information visualization. We suggest ways to refine the methods for MILCs in modest sized projects and then envision ambitious projects with 3-10 researchers working over 1-3 years to understand individual and organizational use of information visualization by domain experts working at the frontiers of knowledge in their fields. %B Proceedings of the 2006 AVI workshop on BEyond time and errors: novel evaluation methods for information visualization %S BELIV '06 %I ACM %C New York, NY, USA %P 1 - 7 %8 2006/// %@ 1-59593-562-2 %G eng %U http://doi.acm.org/10.1145/1168149.1168158 %R 10.1145/1168149.1168158 %0 Conference Paper %B Proceedings of the 6th ACM SIGCOMM conference on Internet measurement %D 2006 %T A structural approach to latency prediction %A Madhyastha,Harsha V. %A Anderson,Thomas %A Krishnamurthy,Arvind %A Spring, Neil %A Venkataramani,Arun %K internet topology %K latency prediction %K route measurements %X Several models have been recently proposed for predicting the latency of end to end Internet paths. These models treat the Internet as a black-box, ignoring its internal structure. While these models are simple, they can often fail systematically; for example, the most widely used models use metric embeddings that predict no benefit to detour routes even though half of all Internet routes can benefit from detours.In this paper, we adopt a structural approach that predicts path latency based on measurements of the Internet's routing topology, PoP connectivity, and routing policy. We find that our approach outperforms Vivaldi, the most widely used black-box model. Furthermore, unlike metric embeddings, our approach successfully predicts 65% of detour routes in the Internet. The number of measurements used in our approach is comparable with that required by black box techniques, but using traceroutes instead of pings. %B Proceedings of the 6th ACM SIGCOMM conference on Internet measurement %S IMC '06 %I ACM %C New York, NY, USA %P 99 - 104 %8 2006/// %@ 1-59593-561-4 %G eng %U http://doi.acm.org/10.1145/1177080.1177092 %R 10.1145/1177080.1177092 %0 Journal Article %J DIMACS Series in Discrete Mathematics and Theoretical Computer Science %D 2006 %T Structure of social contact networks and their impact on epidemics %A Eubank,S. %A Kumar,V. S.A %A Marathe,M. V %A Srinivasan, Aravind %A Wang,N. %X Traditional epidemiological research has focused on rate-based differential-equation models on completely mixing populations. In this paper, we outline an approach based on a combination of net- work theory and discrete-event simulations to study epidemics in large urban areas. We survey some of our results published in Nature (2004) and the Proc. ACM-SIAM Symposium on Discrete Algorithms (2004), and present some new results on: (i) mathematical properties of large social contact networks, as well as (ii) simulation-based dynamic anal- ysis of disease-spread in such networks. We identify a number of new measures that are significant for understanding epidemics and for devel- oping new strategies in policy planning. We also perform a very detailed structural analysis of the social contact networks produced by TRAN- SIMS, a simulator for detailed transportation/traffic studies, and study two random graph models to generate realistic social networks: Chung- Lu’s model and the configuration model. We also develop combinatorial formulations and approximation algorithms for quarantining, vaccina- tion and sensor placement, as aids to decision-making. %B DIMACS Series in Discrete Mathematics and Theoretical Computer Science %V 70 %P 181 - 181 %8 2006/// %G eng %0 Journal Article %J Proceedings of Association for Machine Translation in the Americas %D 2006 %T A study of translation edit rate with targeted human annotation %A Snover,M. %A Dorr, Bonnie J %A Schwartz,R. %A Micciulla,L. %A Makhoul,J. %X We examine a new, intuitive measurefor evaluating machine-translation output that avoids the knowledge intensiveness of more meaning-based approaches, and the labor-intensiveness of human judg- ments. Translation Edit Rate (TER) mea- sures the amount of editing that a hu- man would have to perform to change a system output so it exactly matches a reference translation. We show that the single-reference variant of TER correlates as well with human judgments of MT quality as the four-reference variant of BLEU. We also define a human-targeted TER (or HTER) and show that it yields higher correlations with human judgments than BLEU—even when BLEU is given human-targeted references. Our results in- dicate that HTER correlates with human judgments better than HMETEOR and that the four-reference variants of TER and HTER correlate with human judg- ments as well as—or better than—a sec- ond human judgment does. %B Proceedings of Association for Machine Translation in the Americas %P 223 - 231 %8 2006/// %G eng %0 Journal Article %J Software: Practice and Experience %D 2006 %T Synthetic‐perturbation techniques for screening shared memory programs %A Snelick,Robert %A JaJa, Joseph F. %A Kacker,Raghu %A Lyon,Gordon %K Design of experiments %K parallel programs %K performance %K Shared memory programming model %K Synthetic perturbation %X The synthetic-perturbation screening (SPS) methodology is based on an empirical approach; SPS introduces artificial perturbations into the MIMD program and captures the effects of such perturbations by using the modern branch of statistics called design of experiments. SPS can provide the basis of a powerful tool for screening MIMD programs for performance bottlenecks. This technique is portable across machines and architectures, and scales extremely well on massively parallel processors. The purpose of this paper is to explain the general approach and to extend it to address specific features that are the main source of poor performance on the shared memory programming model. These include performance degradation due to load imbalance and insufficient parallelism, and overhead introduced by synchronizations and by accessing shared data structures. We illustrate the practicality of SPS by demonstrating its use on two very different case studies: a large image understanding benchmark and a parallel quicksort. %B Software: Practice and Experience %V 24 %P 679 - 701 %8 2006/10/30/ %@ 1097-024X %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/spe.4380240802/abstract %N 8 %R 10.1002/spe.4380240802 %0 Patent %D 2006 %T System and method for locating a closest server in response to a client ... %A Andrews,Matthew %A Hofmann,Markus %A Shepherd,Bruce %A Srinivasan, Aravind %A Winkler,Peter %A Zane,Francis %E Lucent Technologies Inc. %X A scalable system and method for locating a closest server in response to a client request via an interactive distribution network, such as the Internet, are provided. A closest content server is defined as having the least round trip time for responding to a client request. The system including a plurality of content servers; and a local server in communication with a plurality of clients, the local server acting as a proxy for communicating client requests from clients to a redirection server. Client network distance and load information is periodically collected at each content server in the network from clients communicating with each of the respective content servers. The redirection server periodically aggregates the network distance and load information from each content server to create client clusters from both current and previously aggregated network distance and load information. Each client cluster represents a division or partition of the total IP address space. Each... %V 09/726,192 %8 2006/03/28/ %G eng %U http://www.google.com/patents?id=cE54AAAAEBAJ %N 7020698 %0 Book %D 2006 %T System synthesis for optically-connected, multiprocessors on-chip %A Bambha,N. K %A Bhattacharyya, Shuvra S. %A STUDIES,MARYLAND UNIV COLLEGE PARK INST FOR ADVANCED COMPUTER %I Citeseer %8 2006/// %G eng %0 Journal Article %J Software Process: Improvement and Practice %D 2006 %T Techniques and processes for improving the quality and performance of open-source software %A Porter, Adam %A Yilmaz,Cemal %A Memon, Atif M. %A Krishna,Arvind S. %A Schmidt,Douglas C. %A Gokhale,Aniruddha %K distributed continuous quality assurance %K open-source software development and testing %X Open-source development processes have emerged as an effective approach to reduce cycle-time and decrease design, implementation, and quality assurance (QA) costs for certain types of software, particularly systems infrastructure software, such as operating systems (OS), compilers and language processing tools, text and drawing editors, and middleware. This article presents two contributions to the study of open-source software processes. First, we describe key challenges of open-source software and illustrate how QA processes—specifically those tailored to open-source development—help mitigate these challenges better than traditional closed-source processes do. Second, we summarize results of empirical studies that evaluate how our Skoll distributed continuous quality assurance (DCQA) techniques and processes help to resolve key challenges in developing and validating open-source software. Our results show that: (a) using models to configure and guide the DCQA process improves developer understanding of open-source software, (b) improving the diversity of platform configurations helps QA engineers find defects missed during conventional testing, and (c) centralizing control of QA activities helps to eliminate redundant work. Copyright © 2006 John Wiley & Sons, Ltd. %B Software Process: Improvement and Practice %V 11 %P 163 - 176 %8 2006/// %@ 1099-1670 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/spip.260/abstract %N 2 %R 10.1002/spip.260 %0 Journal Article %J Computing in Science & Engineering %D 2006 %T A telescope for high-dimensional data %A Shneiderman, Ben %K Chemistry %K Data analysis %K data mining %K data visualisation %K degenerative disease %K Degenerative diseases %K diseases %K Finance %K genetic process %K Genetics %K high-dimensional data %K interactive data analysis %K medical computing %K Meteorology %K muscle %K muscle development %K Muscles %K muscular dystrophy %K Rank-by-feature framework %K software %K software libraries %K software tools %K Telescopes %K visual data %K visual data analysis %X Muscular dystrophy is a degenerative disease that destroys muscles and ultimately kills its victims. Researchers worldwide are racing to find a cure by trying to uncover the genetic processes that cause it. Given that a key process is muscle development, researchers at a consortium of 10 institutions are studying 1,000 men and women, ages 18 to 40 years, to see how their muscles enlarge with exercise. The 150 variables collected for each participant will make this data analysis task challenging for users of traditional statistical software tools. However, a new approach to visual data analysis is helping these researchers speed up their work. At the University of Maryland's Human-Computer Interaction Library, we developed an interactive approach to let researchers explore high-dimensional data in an orderly manner, focusing on specific features one at a time. The rank-by-feature framework lets them adjust controls to specify what they're looking for, and then, with only a glance, they can spot strong relationships among variables, find tight data clusters, or identify unexpected gaps. Sometimes surprising outliers invite further study as to whether they represent errors or an unusual outcome. Similar data analysis problems come up in meteorology, finance, chemistry, and other sciences in which complex relationships among many variables govern outcomes. The rank-by-feature framework could be helpful to many researchers, engineers, and managers because they can then steer their analyses toward the action %B Computing in Science & Engineering %V 8 %P 48 - 53 %8 2006/04//March %@ 1521-9615 %G eng %N 2 %R 10.1109/MCSE.2006.21 %0 Conference Paper %B Proceedings of the 4th ACM international workshop on Video surveillance and sensor networks %D 2006 %T Top-down, bottom-up multivalued default reasoning for identity maintenance %A Shet,Vinay D. %A Harwood,David %A Davis, Larry S. %K default logic %K identity maintenance %K nonmonotonic reasoning %K tracking %K visual surveillance %X Persistent tracking systems require the capacity to track individuals by maintaining identity across visibility gaps caused by occlusion events. In traditional computer vision systems, the flow of information is typically bottom-up. The low level image processing modules take video input, perform early vision tasks such as background subtraction and object detection,and pass this information to the high level reasoning module. This paper describes the architecture of a system that uses top-down information flow to perform identity maintenance across occlusion events. This system uses the high level reasoning module to provide control feedback to the low level image processing module to perform forensic analysis of archival video and actively acquire information required to arrive at identity decisions. This functionality is in addition to traditional bottom-up reasoning about identity, employing contextual cues and appearance matching, within the multivalued default logic framework proposed in [18]. This framework, in addition to bestowing upon the system the property of nonmonotonicity, also allows for it to qualitatively encode its confidence in the identity decisions it takes. %B Proceedings of the 4th ACM international workshop on Video surveillance and sensor networks %S VSSN '06 %I ACM %C New York, NY, USA %P 79 - 86 %8 2006/// %@ 1-59593-496-0 %G eng %U http://doi.acm.org/10.1145/1178782.1178795 %R 10.1145/1178782.1178795 %0 Conference Paper %B Proceedings of the COLING/ACL on Main conference poster sessions %D 2006 %T Topic-focused multi-document summarization using an approximate oracle score %A Conroy,John M. %A Schlesinger,Judith D. %A O'Leary, Dianne P. %X We consider the problem of producing a multi-document summary given a collection of documents. Since most successful methods of multi-document summarization are still largely extractive, in this paper, we explore just how well an extractive method can perform. We introduce an "oracle" score, based on the probability distribution of unigrams in human summaries. We then demonstrate that with the oracle score, we can generate extracts which score, on average, better than the human summaries, when evaluated with ROUGE. In addition, we introduce an approximation to the oracle score which produces a system with the best known performance for the 2005 Document Understanding Conference (DUC) evaluation. %B Proceedings of the COLING/ACL on Main conference poster sessions %S COLING-ACL '06 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 152 - 159 %8 2006/// %G eng %U http://dl.acm.org/citation.cfm?id=1273073.1273093 %0 Conference Paper %B Proceedings of the 6th ACM SIGCOMM conference on Internet measurement %D 2006 %T Touring the internet in a TCP sidecar %A Sherwood,Rob %A Spring, Neil %K network topology discovery %K passenger %K record route %K sidecar %X An accurate router-level topology of the Internet would benefit many research areas, including network diagnosis, inter-domain traffic engineering, and overlay construction. We present TCP Sidecar and Passenger, two elements of a system for router-level Internet topology discovery. Sidecar transparently injects measurement probes into non-measurement TCP streams, while Passenger combines TTL-limited probes with the often-ignored IP record route option. The combined approach mitigates problems associated with traceroute-based topology discovery, including abuse reports, spurious edge inference from multi-path routing, unresolved IP aliases, long network timeouts, and link discovery behind NATs and firewalls. We believe that we are the first mapping project to measure MPLS use with ICMP extensions and record route behavior when the TTL is not decremented. We are able to discover NATs when monitoring TCP connections that tunnel through them. In this paper, we present preliminary results for TCP Sidecar and Passenger on PlanetLab. Our experiments inject measurement probes into traffic generated both from the CoDeeN Web proxy project and from a custom web crawler to 166,745 web sites. %B Proceedings of the 6th ACM SIGCOMM conference on Internet measurement %S IMC '06 %I ACM %C New York, NY, USA %P 339 - 344 %8 2006/// %@ 1-59593-561-4 %G eng %U http://doi.acm.org/10.1145/1177080.1177093 %R 10.1145/1177080.1177093 %0 Journal Article %J Bulletin of the American Society for Information Science and Technology %D 2006 %T Toward an enriched (and revitalized) sense of help: Summary of an ASIS&T 2005 panel session %A Haas,Stephanie W %A Denn,Sheila %A Locke,David %A Shneiderman, Ben %A Brown,Laurie %B Bulletin of the American Society for Information Science and Technology %V 32 %P 23 - 26 %8 2006/06/01/ %@ 1550-8366 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/bult.2006.1720320509/abstract %N 5 %R 10.1002/bult.2006.1720320509 %0 Journal Article %J Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %D 2006 %T Toxigenic Vibrio Cholerae in the Aquatic Environment of Mathbaria, Bangladesh %A Alam,Munirul %A Sultana,Marzia %A Nair,G. Balakrish %A Sack,R. Bradley %A Sack,David A. %A Siddique,A. K. %A Ali,Afsar %A Huq,Anwar %A Rita R Colwell %X Toxigenic Vibrio cholerae, rarely isolated from the aquatic environment between cholera epidemics, can be detected in what is now understood to be a dormant stage, i.e., viable but nonculturable when standard bacteriological methods are used. In the research reported here, biofilms have proved to be a source of culturable V. cholerae, even in nonepidemic periods. Biweekly environmental surveillance for V. cholerae was carried out in Mathbaria, an area of cholera endemicity adjacent to the Bay of Bengal, with the focus on V. cholerae O1 and O139 Bengal. A total of 297 samples of water, phytoplankton, and zooplankton were collected between March and December 2004, yielding eight V. cholerae O1 and four O139 Bengal isolates. A combination of culture methods, multiplex-PCR, and direct fluorescent antibody (DFA) counting revealed the Mathbaria aquatic environment to be a reservoir for V. cholerae O1 and O139 Bengal. DFA results showed significant clumping of the bacteria during the interepidemic period for cholera, and the fluorescent micrographs revealed large numbers of V. cholerae O1 in thin films of exopolysaccharides (biofilm). A similar clumping of V. cholerae O1 was also observed in samples collected from Matlab, Bangladesh, where cholera also is endemic. Thus, the results of the study provided in situ evidence for V. cholerae O1 and O139 in the aquatic environment, predominantly as viable but nonculturable cells and culturable cells in biofilm consortia. The biofilm community is concluded to be an additional reservoir of cholera bacteria in the aquatic environment between seasonal epidemics of cholera in Bangladesh. %B Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %V 72 %P 2849 - 2855 %8 2006/04/01/ %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/72/4/2849 %N 4 %R 10.1128/AEM.72.4.2849-2855.2006 %0 Journal Article %J BMC Genomics %D 2006 %T Trypanosoma cruzi mitochondrial maxicircles display species- and strain-specific variation and a conserved element in the non-coding region %A Westenberger,Scott %A Cerqueira,Gustavo %A El‐Sayed, Najib M. %A Zingales,Bianca %A Campbell,David %A Sturm,Nancy %X BACKGROUND:The mitochondrial DNA of kinetoplastid flagellates is distinctive in the eukaryotic world due to its massive size, complex form and large sequence content. Comprised of catenated maxicircles that contain rRNA and protein-coding genes and thousands of heterogeneous minicircles encoding small guide RNAs, the kinetoplast network has evolved along with an extreme form of mRNA processing in the form of uridine insertion and deletion RNA editing. Many maxicircle-encoded mRNAs cannot be translated without this post-transcriptional sequence modification.RESULTS:We present the complete sequence and annotation of the Trypanosoma cruzi maxicircles for the CL Brener and Esmeraldo strains. Gene order is syntenic with Trypanosoma brucei and Leishmania tarentolae maxicircles. The non-coding components have strain-specific repetitive regions and a variable region that is unique for each strain with the exception of a conserved sequence element that may serve as an origin of replication, but shows no sequence identity with L. tarentolae or T. brucei. Alternative assemblies of the variable region demonstrate intra-strain heterogeneity of the maxicircle population. The extent of mRNA editing required for particular genes approximates that seen in T. brucei. Extensively edited genes were more divergent among the genera than non-edited and rRNA genes. Esmeraldo contains a unique 236-bp deletion that removes the 5'-ends of ND4 and CR4 and the intergenic region. Esmeraldo shows additional insertions and deletions outside of areas edited in other species in ND5, MURF1, and MURF2, while CL Brener has a distinct insertion in MURF2.CONCLUSION:The CL Brener and Esmeraldo maxicircles represent two of three previously defined maxicircle clades and promise utility as taxonomic markers. Restoration of the disrupted reading frames might be accomplished by strain-specific RNA editing. Elements in the non-coding region may be important for replication, transcription, and anchoring of the maxicircle within the kinetoplast network. %B BMC Genomics %V 7 %P 60 - 60 %8 2006/// %@ 1471-2164 %G eng %U http://www.biomedcentral.com/1471-2164/7/60 %N 1 %R 10.1186/1471-2164-7-60 %0 Conference Paper %B 44th AIAA Aerospace Meeting and Exhibit %D 2006 %T The turbulence structure of shockwave and boundary layer interaction in a compression corner %A Martin, M.P %A Smits,A. %A Wu,M. %A Ringuette,M. %X Shockwave and turbulent boundary layer interactions (STBLI) result in intense localizedheating rates and pressure loads, making them extremely important flow features that must be identified for engineering design. The absence of detailed and complete experimental and numerical data at the same flow and boundary conditions is one of the major stumbling blocks in the development of accurate turbulence models for the prediction of STBLI. We use a set of direct numerical simulation data (Wu & Martin, 2006) that has been validated against experiments (Bookey et al., 2005) at the same conditions to present detailed flow features of the STBLI over a compression corner at Mach 3 and low Reynolds number with Reθ=2100. Details regarding the evolution of the turbulence structure angle, characteristic streamwise length scales, and hairpin packets through the interaction are presented. The three-dimensionality of the turbulence field and main shock are illustrated and the strength of shocks and shocklets through the interaction are considered. %B 44th AIAA Aerospace Meeting and Exhibit %C Reno, NV %V 2006 %8 2006/// %G eng %0 Conference Paper %B Proc. of SPIE Vol %D 2006 %T Using image similarity and asymmetry to detect breast cancer %A Tahmoush,D. %A Samet, Hanan %X Radiologists can use the differences between the left and right breasts, or asymmetry, in mammograms to helpdetect certain malignant breast cancers. An image similarity method is introduced to make use of this knowledge base to recognize breast cancer. Image similarity is determined using a contextual and then a spatial comparison. The mammograms are filtered to find the most contextually significant points, and then the resulting point set is analyzed for spatial similarity. We develop the analysis through a combination of modeling and supervised learning of model parameters. This process correctly classifies mammograms 80% of the time and thus asymmetry is a measure that can play an important role in significantly improving computer-aided breast cancer detection systems. %B Proc. of SPIE Vol %V 6144 %P 61441S–1 - 61441S–1 %8 2006/// %G eng %0 Journal Article %J ACM SIGOPS Operating Systems Review %D 2006 %T Using PlanetLab for network research %A Spring, Neil %A Peterson,Larry %A Bavier,Andy %A Pait,Vivek %B ACM SIGOPS Operating Systems Review %V 40 %P 17 - 17 %8 2006/01/01/ %@ 01635980 %G eng %U http://dl.acm.org/citation.cfm?id=1113368 %R 10.1145/1113361.1113368 %0 Journal Article %J Journal of the American Society for Information Science and Technology %D 2006 %T Using rhythms of relationships to understand e‐mail archives %A Perer,Adam %A Shneiderman, Ben %A Oard, Douglas %X Due to e-mail's ubiquitous nature, millions of users are intimate with the technology; however, most users are only familiar with managing their own e-mail, which is an inherently different task from exploring an e-mail archive. Historians and social scientists believe that e-mail archives are important artifacts for understanding the individuals and communities they represent. To understand the conversations evidenced in an archive, context is needed. In this article, we present a new way to gain this necessary context: analyzing the temporal rhythms of social relationships. We provide methods for constructing meaningful rhythms from the e-mail headers by identifying relationships and interpreting their attributes. With these visualization techniques, e-mail archive explorers can uncover insights that may have been otherwise hidden in the archive. We apply our methods to an individual's 15-year e-mail archive, which consists of about 45,000 messages and over 4,000 relationships. %B Journal of the American Society for Information Science and Technology %V 57 %P 1936 - 1948 %8 2006/12/01/ %@ 1532-2890 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/asi.20387/abstract %N 14 %R 10.1002/asi.20387 %0 Conference Paper %B Visual Analytics Science And Technology, 2006 IEEE Symposium On %D 2006 %T VAST 2006 Contest - A Tale of Alderwood %A Grinstein,G. %A O'Connell,T. %A Laskowski,S. %A Plaisant, Catherine %A Scholtz,J. %A Whiting,M. %K Alderwood;human %K analysis;data %K analytics %K and %K contest;data %K information %K interaction;sense %K making;visual %K Science %K technology %K visualisation; %X Visual analytics experts realize that one effective way to push the field forward and to develop metrics for measuring the performance of various visual analytics components is to hold an annual competition. The first visual analytics science and technology (VAST) contest was held in conjunction with the 2006 IEEE VAST Symposium. The competition entailed the identification of possible political shenanigans in the fictitious town of Alderwood. A synthetic data set was made available as well as tasks. We summarize how we prepared and advertised the contest, developed some initial metrics for evaluation, and selected the winners. The winners were invited to participate at an additional live competition at the symposium to provide them with feedback from senior analysts %B Visual Analytics Science And Technology, 2006 IEEE Symposium On %P 215 - 216 %8 2006/11/31/2 %G eng %R 10.1109/VAST.2006.261420 %0 Report %D 2006 %T Vehicle Fingerprinting Using Drive-By Sounds %A Cevher, V. %A Chellapa, Rama %A Gurbuz, A.C. %A Shah, F. %A McClellan, J.H. %K *ACOUSTIC DETECTORS %K *GROUND VEHICLES %K *SIGNAL PROCESSING %K ACOUSTIC DETECTION AND DETECTORS %K AMPLITUDE %K Interference %K PASSIVE SYSTEMS %K spatial distribution %K STATISTICAL PROCESSES %K SURFACE TRANSPORTATION AND EQUIPMENT %K SYMPOSIA %K WAVE PROPAGATION %X We estimate a vehicle's speed, width, and length by jointly estimating its acoustic wave-pattern using a single passive acoustic sensor that records the vehicle's drive-by noise. The acoustic wave-pattern is estimated using three envelope shape (ES) components, which approximate the shape of the received signal's power envelope. We incorporate the parameters of the ES components along with estimates of the vehicle engine RPM and number of cylinders to form a vehicle profile vector. This vector provides a compressed statistics that can be used for vehicle identification and classification. Vehicle speed estimation and classification results are provided using field data. %I CENTER FOR AUTOMATION RESEARCH, UNIVERSITY OF MARYLAND COLLEGE PARK %8 2006/11// %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA481471 %0 Conference Paper %B Visual Analytics Science And Technology, 2006 IEEE Symposium On %D 2006 %T A Visual Interface for Multivariate Temporal Data: Finding Patterns of Events across Multiple Histories %A Fails,J. A %A Karlson,A. %A Shahamat,L. %A Shneiderman, Ben %K ball-and-chain visualization %K Chromium %K Computer science %K Data analysis %K data visualisation %K Data visualization %K Database languages %K event pattern discovery %K Graphical user interfaces %K History %K Information Visualization %K Medical treatment %K multivariate temporal data %K Pattern analysis %K pattern recognition %K PatternFinder integrated interface %K Query processing %K query visualization %K result-set visualization %K Spatial databases %K tabular visualization %K temporal pattern discovery %K temporal pattern searching %K Temporal query %K user interface %K User interfaces %K visual databases %K visual interface %X Finding patterns of events over time is important in searching patient histories, Web logs, news stories, and criminal activities. This paper presents PatternFinder, an integrated interface for query and result-set visualization for search and discovery of temporal patterns within multivariate and categorical data sets. We define temporal patterns as sequences of events with inter-event time spans. PatternFinder allows users to specify the attributes of events and time spans to produce powerful pattern queries that are difficult to express with other formalisms. We characterize the range of queries PatternFinder supports as users vary the specificity at which events and time spans are defined. Pattern Finder's query capabilities together with coupled ball-and-chain and tabular visualizations enable users to effectively query, explore and analyze event patterns both within and across data entities (e.g. patient histories, terrorist groups, Web logs, etc.) %B Visual Analytics Science And Technology, 2006 IEEE Symposium On %I IEEE %P 167 - 174 %8 2006/11/31/Oct. %@ 1-4244-0591-2 %G eng %R 10.1109/VAST.2006.261421 %0 Book Section %B UbiComp 2006: Ubiquitous Computing %D 2006 %T Voting with Your Feet: An Investigative Study of the Relationship Between Place Visit Behavior and Preference %A Jon Froehlich %A Chen,Mike %A Smith,Ian %A Potter,Fred %E Dourish,Paul %E Friday,Adrian %X Real world recommendation systems, personalized mobile search, and online city guides could all benefit from data on personal place preferences. However, collecting explicit rating data of locations as users travel from place to place is impractical. This paper investigates the relationship between explicit place ratings and implicit aspects of travel behavior such as visit frequency and travel time. We conducted a four-week study with 16 participants using a novel sensor-based experience sampling tool, called My Experience (Me), which we developed for mobile phones. Over the course of the study Me was used to collect 3,458 in-situ questionnaires on 1,981 place visits. Our results show that, first, sensor-triggered experience sampling is a useful methodology for collecting targeted information in situ. Second, despite the complexities underlying travel routines and visit behavior, there exist positive correlations between place preference and automatically detectable features like visit frequency and travel time. And, third, we found that when combined, visit frequency and travel time result in stronger correlations with place rating than when measured individually. Finally, we found no significant difference in place ratings due to the presence of others. %B UbiComp 2006: Ubiquitous Computing %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 4206 %P 333 - 350 %8 2006 %@ 978-3-540-39634-5 %G eng %U http://dx.doi.org/10.1007/11853565_20 %0 Journal Article %J Proceedings of the IEEE %D 2006 %T Wireless Network Security and Interworking %A Shin,M. %A Ma,J. %A Mishra,A. %A Arbaugh, William A. %K 3G mobile communication %K 3G systems %K Authentication %K Bandwidth %K Communication system security %K computer network security %K computer security %K Data security %K internetworking %K Land mobile radio cellular systems %K Paper technology %K security architectures %K security of data %K telecommunication security %K wireless communication %K wireless communications %K Wireless LAN %K wireless network security %K Wireless networks %K wireless technologies %K WLAN systems %X A variety of wireless technologies have been standardized and commercialized, but no single technology is considered the best because of different coverage and bandwidth limitations. Thus, interworking between heterogeneous wireless networks is extremely important for ubiquitous and high-performance wireless communications. Security in interworking is a major challenge due to the vastly different security architectures used within each network. The goal of this paper is twofold. First, we provide a comprehensive discussion of security problems and current technologies in 3G and WLAN systems. Second, we provide introductory discussions about the security problems in interworking, the state-of-the-art solutions, and open problems. %B Proceedings of the IEEE %V 94 %P 455 - 466 %8 2006/02// %@ 0018-9219 %G eng %N 2 %R 10.1109/JPROC.2005.862322 %0 Journal Article %J Behaviour & Information Technology %D 2006 %T Workplace user frustration with computers: an exploratory investigation of the causes and severity %A Lazar,Jonathan %A Jones,Adam %A Shneiderman, Ben %X When hard-to-use computers cause users to become frustrated, it can affect workplace productivity, user mood and interactions with other co-workers. Previous research has examined the frustration that students and their families face in using computers. To learn more about the causes and measure the severity of user frustration with computers in the workplace, we collected modified time diaries from 50 workplace users, who spent an average of 5.1 hours on the computer. In this exploratory research, users reported wasting on average, 42 ? 43% of their time on the computer due to frustrating experiences. The largest number of frustrating experiences occurred while using word processors, email and web browsers. The causes of the frustrating experiences, the time lost due to the frustrating experiences, and the effects of the frustrating experiences on the mood of the users are discussed in this paper. Implications for designers, managers, users, information technology staff and policymakers are discussed.When hard-to-use computers cause users to become frustrated, it can affect workplace productivity, user mood and interactions with other co-workers. Previous research has examined the frustration that students and their families face in using computers. To learn more about the causes and measure the severity of user frustration with computers in the workplace, we collected modified time diaries from 50 workplace users, who spent an average of 5.1 hours on the computer. In this exploratory research, users reported wasting on average, 42 ? 43% of their time on the computer due to frustrating experiences. The largest number of frustrating experiences occurred while using word processors, email and web browsers. The causes of the frustrating experiences, the time lost due to the frustrating experiences, and the effects of the frustrating experiences on the mood of the users are discussed in this paper. Implications for designers, managers, users, information technology staff and policymakers are discussed. %B Behaviour & Information Technology %V 25 %P 239 - 251 %8 2006/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449290500196963 %N 3 %R 10.1080/01449290500196963 %0 Book Section %B Discovery ScienceDiscovery Science %D 2005 %T Active Constrained Clustering by Examining Spectral Eigenvectors %A Xu,Qianjun %A desJardins, Marie %A Wagstaff,Kiri %E Hoffmann,Achim %E Motoda,Hiroshi %E Scheffer,Tobias %X This work focuses on the active selection of pairwise constraints for spectral clustering. We develop and analyze a technique for Active Constrained Clustering by Examining Spectral eigenvectorS (ACCESS) derived from a similarity matrix. The ACCESS method uses an analysis based on the theoretical properties of spectral decomposition to identify data items that are likely to be located on the boundaries of clusters, and for which providing constraints can resolve ambiguity in the cluster descriptions. Empirical results on three synthetic and five real data sets show that ACCESS significantly outperforms constrained spectral clustering using randomly selected constraints. %B Discovery ScienceDiscovery Science %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3735 %P 294 - 307 %8 2005/// %@ 978-3-540-29230-2 %G eng %U http://dx.doi.org/10.1007/11563983_25 %0 Journal Article %J J. ACM %D 2005 %T Aggregate operators in probabilistic databases %A Ross,Robert %A V.S. Subrahmanian %A Grant,John %K Aggregates %K probabilistic relational databases %X Though extensions to the relational data model have been proposed in order to handle probabilistic information, there has been very little work to date on handling aggregate operators in such databases. In this article, we present a very general notion of an aggregate operator and show how classical aggregation operators (such as COUNT, SUM, etc.) as well as statistical operators (such as percentiles, variance, etc.) are special cases of this general definition. We devise a formal linear programming based semantics for computing aggregates over probabilistic DBMSs, develop algorithms that satisfy this semantics, analyze their complexity, and introduce several families of approximation algorithms that run in polynomial time. We implemented all of these algorithms and tested them on a large set of data to help determine when each one is preferable. %B J. ACM %V 52 %P 54 - 101 %8 2005/01// %@ 0004-5411 %G eng %U http://doi.acm.org/10.1145/1044731.1044734 %N 1 %R 10.1145/1044731.1044734 %0 Journal Article %J Electronic Notes in Theoretical Computer Science %D 2005 %T An Algebraic Theory Of Boundary Crossing Transitions %A Ray,Arnab %A Cleaveland, Rance %A Skou,Arne %K Compositional Semantics %K Formal Methods %K Process algebra %K Statecharts %X This paper gives a process-algebraic semantics for the hierarchical state machine (HSM) fragment of Statecharts, in which state transitions are permitted to cross state boundaries. Although frowned upon by researchers as promoting unstructured modeling, such transitions are used extensively in practice to model parameterized start states and conditional exit states. The purpose of this work is to develop a compositional semantics for HSMs that may be fit together with compositional semantic accounts for Statecharts without boundary-crossing transitions in order to arrive at a compositional theory for virtually the whole Statecharts language. Our technical development consists of a process algebra for HSMs that is equipped with an operational semantics, an argument that bisimulation is a congruence for the algebra, a syntax-directed translation procedure for HSMs into the process algebra, and an equational axiomatization of the algebra. %B Electronic Notes in Theoretical Computer Science %V 115 %P 69 - 88 %8 2005/01/18/ %@ 1571-0661 %G eng %U http://www.sciencedirect.com/science/article/pii/S1571066104053186 %R 10.1016/j.entcs.2004.09.029 %0 Journal Article %J ACM Transactions on Mathematical Software-TOMS %D 2005 %T Algorithm 844: Computing sparse reduced-rank approximations to sparse matrices %A Berry,M. W %A Pulatova,S. A %A Stewart, G.W. %B ACM Transactions on Mathematical Software-TOMS %V 31 %P 252 - 269 %8 2005/// %G eng %N 2 %0 Conference Paper %B Computer Design: VLSI in Computers and Processors, 2005. ICCD 2005. Proceedings. 2005 IEEE International Conference on %D 2005 %T Algorithmic and architectural design methodology for particle filters in hardware %A Sankaranarayanan,A. C %A Chellapa, Rama %A Srivastava, A. %K (numerical %K algorithmic %K architectural %K architectures; %K bearing %K complexity; %K computational %K design %K digital %K evolution; %K Filtering %K filtering; %K filters; %K implementation; %K methodology; %K methods); %K nonGaussian %K nonlinear %K only %K Parallel %K particle %K pipeline %K pipelined %K problem; %K processing; %K state %K tracking %K VLSI %K VLSI; %X In this paper, we present algorithmic and architectural methodology for building particle filters in hardware. Particle filtering is a new paradigm for filtering in presence of nonGaussian nonlinear state evolution and observation models. This technique has found wide-spread application in tracking, navigation, detection problems especially in a sensing environment. So far most particle filtering implementations are not lucrative for real time problems due to excessive computational complexity involved. In this paper, we re-derive the particle filtering theory to make it more amenable to simplified VLSI implementations. Furthermore, we present and analyze pipelined architectural methodology for designing these computational blocks. Finally, we present an application using the bearing only tracking problem and evaluate the proposed architecture and algorithmic methodology. %B Computer Design: VLSI in Computers and Processors, 2005. ICCD 2005. Proceedings. 2005 IEEE International Conference on %P 275 - 280 %8 2005/10// %G eng %R 10.1109/ICCD.2005.20 %0 Conference Paper %B Proceedings of the 2005 ACM SIGMETRICS international conference on Measurement and modeling of computer systems %D 2005 %T Algorithmic aspects of capacity in wireless networks %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %K capacity modeling %K end-to-end scheduling %K Linear programming %K Wireless networks %X This paper considers two inter-related questions: (i) Given a wireless ad-hoc network and a collection of source-destination pairs {(si,ti)}, what is the maximum throughput capacity of the network, i.e. the rate at which data from the sources to their corresponding destinations can be transferred in the network? (ii) Can network protocols be designed that jointly route the packets and schedule transmissions at rates close to the maximum throughput capacity? Much of the earlier work focused on random instances and proved analytical lower and upper bounds on the maximum throughput capacity. Here, in contrast, we consider arbitrary wireless networks. Further, we study the algorithmic aspects of the above questions: the goal is to design provably good algorithms for arbitrary instances. We develop analytical performance evaluation models and distributed algorithms for routing and scheduling which incorporate fairness, energy and dilation (path-length) requirements and provide a unified framework for utilizing the network close to its maximum throughput capacity.Motivated by certain popular wireless protocols used in practice, we also explore "shortest-path like" path selection strategies which maximize the network throughput. The theoretical results naturally suggest an interesting class of congestion aware link metrics which can be directly plugged into several existing routing protocols such as AODV, DSR, etc. We complement the theoretical analysis with extensive simulations. The results indicate that routes obtained using our congestion aware link metrics consistently yield higher throughput than hop-count based shortest path metrics. %B Proceedings of the 2005 ACM SIGMETRICS international conference on Measurement and modeling of computer systems %S SIGMETRICS '05 %I ACM %C New York, NY, USA %P 133 - 144 %8 2005/// %@ 1-59593-022-1 %G eng %U http://doi.acm.org/10.1145/1064212.1064228 %R 10.1145/1064212.1064228 %0 Conference Paper %B Proceedings of the second international workshop on Software engineering for high performance computing system applications %D 2005 %T And away we go: understanding the complexity of launching complex HPC applications %A Yoon,Il-Chul %A Sussman, Alan %A Porter, Adam %K development %K high-performance %K Productivity %K software %X Although not well-studied, launching HPC applications is extremely complex. To better understand the launching process, we conducted a simple case study. Based in part on this study and an examination of existing toolkits, we have begun to develop a prototype environment to support HPC application launching. %B Proceedings of the second international workshop on Software engineering for high performance computing system applications %S SE-HPCS '05 %I ACM %C New York, NY, USA %P 45 - 49 %8 2005/// %@ 1-59593-117-1 %G eng %U http://doi.acm.org/10.1145/1145319.1145333 %R 10.1145/1145319.1145333 %0 Conference Paper %B Multi-Agent Security and Survivability, 2005 IEEE 2nd Symposium on %D 2005 %T Approximation results for probabilistic survivability %A Zhang,Y. %A Manister,E. %A Kraus,S. %A V.S. Subrahmanian %K application; %K computing; %K failure; %K fault %K heuristic %K heuristics %K industrial %K model; %K multi-agent %K multiagent %K node %K probabilistic %K probability; %K programming; %K survivability; %K system; %K systems; %K testing; %K tolerant %X As multiagent systems (MASs) are increasingly used in industrial applications, the need to make them more robust and resilient against disruption increases dramatically. The author has developed a probabilistic model (assuming complete ignorance of dependencies between node failures) of survivability based on deploying each agent in a MAS on one or more nodes. Finding a deployment that maximizes survivability is highly intractable for two reasons: firstly, computing the survivability of any deployment is intractable, and secondly, going through an exponential number of deployments to find the best one adds another layer of intractability. In this paper, we study what happens when node failures are independent. We show that computing survivability in this environment is still intractable. We propose various heuristics to compute the survivability of a given deployment. We have implemented and tested all these heuristics. We report on the advantages and disadvantages of different heuristics in different environmental settings. %B Multi-Agent Security and Survivability, 2005 IEEE 2nd Symposium on %P 1 - 10 %8 2005/08// %G eng %R 10.1109/MASSUR.2005.1507042 %0 Book Section %B Architecting Dependable Systems III %D 2005 %T Architecting and Implementing Versatile Dependability %A Tudor Dumitras %A Srivastava, Deepti %A Narasimhan, Priya %E Lemos, Rogério de %E Gacek, Cristina %E Romanovsky, Alexander %K Operating systems %K software engineering %X Distributed applications must often consider and select the appropriate trade-offs among three important aspects – fault-tolerance, performance and resources. We introduce a novel concept, called versatile dependability, that provides a framework for analyzing and reasoning about these trade-offs in dependable software architectures. We present the architecture of a middleware framework that implements versatile dependability by providing the appropriate ”knobs” to tune and re-calibrate the trade-offs. Our framework can adjust the properties and the behavior of the system at development-time, at deployment-time, and throughout the application’s life-cycle. This renders the versatile dependability approach useful both to applications that require static fault-tolerance configurations supporting the loss/addition of resources and changing workloads, as well as to applications that evolve in terms of their dependability requirements. Through a couple of specific examples, one on adapting the replication style at runtime and the other on tuning the system scalability under given constraints, we demonstrate concretely how versatile dependability can provide an extended coverage of the design space of dependable distributed systems. %B Architecting Dependable Systems III %S Lecture Notes in Computer Science %I Springer Berlin Heidelberg %P 212 - 231 %8 2005/01/01/ %@ 978-3-540-28968-5, 978-3-540-31648-0 %G eng %U http://link.springer.com/chapter/10.1007/11556169_10 %0 Conference Paper %B Proceedings of the 3rd IEEE/ACM/IFIP international conference on Hardware/software codesign and system synthesis %D 2005 %T An architectural level design methodology for embedded face detection %A Kianzad,V. %A Saha,S. %A Schlessman,J. %A Aggarwal,G. %A Bhattacharyya, Shuvra S. %A Wolf,W. %A Chellapa, Rama %K design space exploration %K Face detection %K platforms %K reconfigurable %K system-level models %X Face detection and recognition research has attracted great attention in recent years. Automatic face detection has great potential in a large array of application areas, including banking and security system access control, video surveillance, and multimedia information retrieval. In this paper, we discuss an architectural level design methodology for implementation of an embedded face detection system on a reconfigurable system on chip. We present models for performance estimation and validate these models with experimental values obtained from implementing our system on an FPGA platform. This modeling approach is shown to be efficient, accurate, and intuitive for designers to work with. Using this approach, we present several design options that trade-off various architectural features. %B Proceedings of the 3rd IEEE/ACM/IFIP international conference on Hardware/software codesign and system synthesis %S CODES+ISSS '05 %I ACM %C New York, NY, USA %P 136 - 141 %8 2005/// %@ 1-59593-161-9 %G eng %U http://doi.acm.org/10.1145/1084834.1084872 %R 10.1145/1084834.1084872 %0 Conference Paper %B Mobile Computing Systems and Applications, 2006. WMCSA'06. Proceedings. 7th IEEE Workshop on %D 2005 %T Are GSM phones THE solution for localization? %A Varshavsky,A. %A Chen,M.Y. %A de Lara,E. %A Jon Froehlich %A Haehnel,D. %A Hightower,J. %A LaMarca,A. %A Potter,F. %A Sohn,T. %A Tang,K. %A others %B Mobile Computing Systems and Applications, 2006. WMCSA'06. Proceedings. 7th IEEE Workshop on %P 34 - 42 %8 2005/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T An Augmented Visual Query Mechanism for Finding Patterns in Time Series Data (2002) %A Keogh,Eamonn %A Hochheiser,Harry %A Shneiderman, Ben %K Technical Report %X Relatively few query tools exist for data exploration and pattern identification in time series data sets. In previous work we introduced Timeboxes. Timeboxes are rectangular, direct-manipulation queries for studying time-series datasets. We demonstrated how Timeboxes can be used to support interactive exploration via dynamic queries, along with overviews of query results and drag-and-drop support for query-by-example. In this paper, we extend our work by introducing Variable Time Timeboxes (VTT). VTTs are a natural generalization of Timeboxes, which permit the specification of queries that allow a degree of uncertainty in the time axis. We carefully motivate the need for these more expressive queries, and demonstrate the utility of our approach on several data sets. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6495 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Bi-Level Hierarchical Layouts for Photo Libraries: Algorithms for Design Optimization with Quantum Content %A Kustanowitz,Jack %A Shneiderman, Ben %K Societal Infrastructure Systems %X A frequently-used layout for a collection of two-dimensional, fixed aspect-ratio objects, such as photo thumbnails, is the grid, in which rows and columns are configured to match the allowed space. However, in cases where these objects have some group relationship among them, it can be advantageous to show this relationship in the layout, rather than in textual captions. We use an annotated digital photo collection as a case study of an auto-layout technique in which a two-level hierarchy is generated, consisting of a primary, central region with secondary regions (typically 2-12 regions) surrounding it. We show that given specific requirements, this technique is also optimal, in the sense that it will generate the largest size for the objects. Since all objects are the same size we refer to them as quantum content. These algorithms are designed to be real-time, enabling a compelling interactive display as users resize the canvas, or move and resize the primary region. The interactive redisplay also occurs as users add regions or objects to a secondary region. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6567 %0 Conference Paper %D 2005 %T BIND: a fine-grained attestation service for secure distributed systems %A Elaine Shi %A Perrig, A. %A Van Doorn, L. %K BIND %K binding instructions and data %K code attestation %K data integrity %K digital signatures %K distributed processing %K fine-grained attestation service %K input data integrity %K program verification %K sandboxing mechanism %K secure distributed systems %K signature %K time-of-attestation %K time-of-use %K transitive integrity verification %K trusted computing %X In this paper we propose BIND (binding instructions and data), a fine-grained attestation service for securing distributed systems. Code attestation has recently received considerable attention in trusted computing. However, current code attestation technology is relatively immature. First, due to the great variability in software versions and configurations, verification of the hash is difficult. Second, the time-of-use and time-of-attestation discrepancy remains to be addressed, since the code may be correct at the time of the attestation, but it may be compromised by the time of use. The goal of BIND is to address these issues and make code attestation more usable in securing distributed systems. BIND offers the following properties: (1) BIND performs fine-grained attestation. Instead of attesting to the entire memory content, BIND attests only to the piece of code we are concerned about. This greatly simplifies verification. (2) BIND narrows the gap between time-of-attestation and time-of-use. BIND measures a piece of code immediately before it is executed and uses a sandboxing mechanism to protect the execution of the attested code. (3) BIND ties the code attestation with the data that the code produces, such that we can pinpoint what code has been run to generate that data. In addition, by incorporating the verification of input data integrity into the attestation, BIND offers transitive integrity verification, i.e., through one signature, we can vouch for the entire chain of processes that have performed transformations over a piece of data. BIND offers a general solution toward establishing a trusted environment for distributed system designers. %P 154 - 168 %8 2005 %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Bridging the Digital Divide with Universal Usability (2001) %A Shneiderman, Ben %K Technical Report %X How do you explain a trashcan to a culture that doesn have one? How do you describe a top loss limit orderto retirees managing their funds? Can you design a text-only interface that conveys the contents and experience of an animated Flash presentation?

These puzzles emerged during the first ACM Conference on Universal Usability (http://www.acm.org/sigchi/cuu/), held on November 15-17, 2000 near Washington, DC. The international group of organizers, presenters, and attendees of this conference shared an unusual commitment and passion for making information and communications services accessible, usable, and useful. They want to see effective healthcare services and appealing distance education. They want to create successful e-commerce and accessible government services for all. Realizing these possibilities requires more than low-cost hardware or broadband networks. These mass- market services are often too complex, unusable, or irrelevant for too many users; usability and design become the keys to success.

The source of these problems was often attributed to designers who make incorrect assumptions about user knowledge. This leads to difficulties with technical terminology and advanced concepts that are not balanced by adequate online help or live assistance. Unfortunately, most designers never see the pain they inflict on novice and even expert users. These problems have contributed to the growing digital divide in internet technology adoption levels between lowincome poorly-educated and high-income well-educated users [2]. Even as the gap between men and women internet users has been eliminated and the gap between young and old is shrinking, the slow adoption rates by poor and poorly educated users remains a problem. Low-cost equipment is needed, but progress in design will help make internet services more accessible to more people. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6478 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Broadening Access to Large Online Databases by Generalizing Query Previews (2000) %A Tanin,Egemen %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X Companies, government agencies, and other types of organizations are making their large databases available to the world over the Internet. Current database front-ends do not give users information about the distribution of data. This leads many users to waste time and network resources posing queries that have either zero-hit or mega-hit result sets. Query previews form a novel visual approach for browsing large databases. Query previews supply data distribution information about the database that is being searched and give continuous feedback about the size of the result set for the query as it is being formed. On the other hand, query previews use only a few pre-selected attributes of the database. The distribution information is displayed only on these attributes. Unfortunately, many databases are formed of numerous relations and attributes. This paper introduces a generalization of query previews. We allow users to browse all of the relations and attributes of a database using a hierarchical browser. Any of the attributes can be used to display the distribution information, making query previews applicable to many public online databases. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6474 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Browsing Large Online Data Tables Using Generalized Query Previews (2001) %A Tanin,Egemen %A Shneiderman, Ben %K Technical Report %X Companies, government agencies, and other organizations are making their data available to the world over the Internet. They often use large online relational tables for this purpose. Users query such tables with front-ends that typically use menus or form fillin interfaces, but these interfaces rarely give users information about the contents and distribution of the data. Such a situation leads users to waste time and network resources posing queries that have zero-hit or mega-hit results. Generalized query previews enable efficient browsing of large online databases by supplying data distribution information to the users. The data distribution information provides continuous feedback about the size of the result set as the query is being formed. Our paper presents a user interface architecture and discusses recent experimental findings. Our prototype system, ExpO, provides a flexible user interface for research and testing. The user study shows that for exploratory querying tasks, generalized query previews speed user performance and reduce network load. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6483 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Building a Coherent Data Pipeline in Microarray Data Analyses: Optimization of Signal/Noise Ratios Using an Interactive Visualization Tool and a Novel Noise Filtering Method (2003) %A Seo,Jinwook %A Bakay,Marina %A Chen,Yi-Wen %A Hilmer,Sara %A Shneiderman, Ben %A Hoffman,Eric P %K Technical Report %X Motivation: Sources of uncontrolled noise strongly influence data analysis in microarray studies, yet signal/noise ratios are rarely considered in microarray data analyses. We hypothesized that different research projects would have different sources and levels of confounding noise, and built an interactive visual analysis tool to test and define parameters in Affymetrix analyses that optimize the ratio of signal (desired biological variable) versus noise (confounding uncontrolled variables). Results: Five probe set algorithms were studied with and without statistical weighting of probe sets using Microarray Suite (MAS) 5.0 probe set detection p values. The signal/noise optimization method was tested in two large novel microarray datasets with different levels of confounding noise; a 105 sample U133A human muscle biopsy data set (11 groups) (extensive noise), and a 40 sample U74A inbred mouse lung data set (8 groups) (little noise). Success was measured using F-measure value of success of unsupervised clustering into appropriate biological groups (signal). We show that both probe set signal algorithm and probe set detection p-value weighting have a strong effect on signal/noise ratios, and that the different methods performed quite differently in the two data sets. Among the signal algorithms tested, dChip difference model with p-value weighting was the most consistent at maximizing the effect of the target biological variables on data interpretation of the two data sets. Availability: The Hierarchical Clustering Explorer 2.0 is [url=http://www.cs.umd.edu/hcil/hce/]available[/url] online and the improved version of the Hierarchical Clustering Explorer 2.0 with p-value weighting and Fmeasure is available upon request to the first author. Murine arrays (40 samples) are publicly available at the [url=http://microarray.cnmcresearch.org/pgadatatable.asp]PEPR resource.[/url] (Chen et al., 2004). %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6511 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Categorized Graphical Overviews for Web Search Results: An Exploratory Study using U.S. Government Agencies as a Meaningful and Stable Structure (2004) %A Kules,Bill %A Shneiderman, Ben %K Global Communication Systems %X Search engines are very effective at generating long lists of results that are highly relevant to user-provided query terms . But the lack of effective overviews presents challenges to users who seek to understand these results, especially for a complex task such as learning about a topic area, which requires gaining overviews of and exploring large sets of search results, identifying unusual documents, and understanding their context. Categorizing the results into comprehensible visual displays using meaningful and stable classifications can support user exploration and understanding of large sets of search results. This extended abstract presents a set of principles that we are developing for search result visualization. It also describes an exploratory study that investigated categorized overviews of search results for complex search tasks within the domain of U. S. government web sites, using a hierarchy based on the federal government organization. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6534 %0 Journal Article %J Technical Reports from UMIACS %D 2005 %T Catergorized Overviews for Web Search Results: Two Exploratory Studies in the US Government Domain %A Kules,B. %A Shneiderman, Ben %X Search engines are very effective at generating extensive lists of results that are highly relevant to user-provided query terms. But the lack of effective overviews challenges users who seek to understand these results, especially for complex tasks such as learning about a new topic, which require gaining overviews of and exploring large sets of search results, identifying unusual documents, and understanding their context. Categorizing the results into comprehensible visual displays using meaningful and stable classifications can support user exploration and understanding of large sets of search results. We conducted two exploratory studies to investigate categorized overviews of search results for complex search tasks within the domain of U. S. government web sites, using a hierarchy based on the federal government organization and a hierarchy generated by automated clustering. First we compared two overview conditions vs. a control (N=18) and found that participants performed better with and preferred the overview conditions. Then we compared automated clustering vs. an overview based on a government hierarchy (N=12) for several task types, and found that several factors appeared to influence the preferred overview. The studies motivated new requirements for the organization of and interaction with web search results: user-selectable organization of results and a lightweight mechanism for customizing hierarchies. The results were also used to refine a set of principles that we are developing for categorized search result visualization. %B Technical Reports from UMIACS %8 2005/// %G eng %0 Conference Paper %B Proceedings of the 2005 conference on Interaction design and children %D 2005 %T Child's play: a comparison of desktop and physical interactive environments %A Fails,J. A %A Druin, Allison %A Guha,M.L. %A Chipman,G. %A Simms,S. %A Churaman,W. %B Proceedings of the 2005 conference on Interaction design and children %P 48 - 55 %8 2005/// %G eng %0 Journal Article %J IEEE Transactions on Circuits and Systems for Video Technology %D 2005 %T Class-based access control for distributed video-on-demand systems %A Mundur, Padma %A Sood,A. K %A Simon,R. %K Access control %K Admission control %K Analytical models %K blocking performance %K class-based access control %K Computational modeling %K Computer architecture %K Computer science %K Distributed control %K Distributed video-on-demand (VoD) system %K distributed video-on-demand system %K multimedia systems %K multirate service model %K Performance analysis %K QoS %K quality of service %K request handling policy %K resource allocation %K resource capacity %K telecommunication congestion control %K threshold-based admission control %K video on demand %X The focus of this paper is the analysis of threshold-based admission control policies for distributed video-on-demand (VoD) systems. Traditionally, admission control methods control access to a resource based on the resource capacity. We have extended that concept to include the significance of an arriving request to the VoD system by enforcing additional threshold restrictions in the admission control process on request classes deemed less significant. We present an analytical model for computing blocking performance of the VoD system under threshold-based admission control. Extending the same methodology to a distributed VoD architecture we show through simulation that the threshold performance conforms to the analytical model. We also show that threshold-based analysis can work in conjunction with other request handling policies and are useful for manipulating the VoD performance since we are able to distinguish between different request classes based on their merit. Enforcing threshold restrictions with the option of downgrading blocked requests in a multirate service environment results in improved performance at the same time providing different levels of quality of service (QoS). In fact, we show that the downgrade option combined with threshold restrictions is a powerful tool for manipulating an incoming request mix over which we have no control into a workload that the VoD system can handle. %B IEEE Transactions on Circuits and Systems for Video Technology %V 15 %P 844 - 853 %8 2005/07// %@ 1051-8215 %G eng %N 7 %R 10.1109/TCSVT.2005.848351 %0 Conference Paper %B Visualization Conference, IEEE %D 2005 %T Clustering Techniques for Out-of-Core Multi-resolution Modeling %A Danovaro,Emanuele %A De Floriani, Leila %A Puppo,Enrico %A Samet, Hanan %X Thanks to improvements in simulation tools, high resolution scanning facilities and multidimensional medical imaging, huge datasets are commonly available. Multi-resolution models manage the complexity of such data sets, by varying resolution and focusing detail in specific areas of interests. Since many currently available data sets cannot fit in main memory, the need arises to design data structures, construction and query algorithms for multi-resolution models which work in secondary memory. %B Visualization Conference, IEEE %I IEEE Computer Society %C Los Alamitos, CA, USA %P 113 - 113 %8 2005/// %@ 0-7803-9462-3 %G eng %R http://doi.ieeecomputersociety.org/10.1109/VIS.2005.15 %0 Journal Article %J International Journal of Geographical Information Science %D 2005 %T Colour‐coded pixel‐based highly interactive Web mapping for georeferenced data exploration %A Zhao,Haixia %A Shneiderman, Ben %X This paper describes a pixel?based technique that enables highly interactive Web choropleth maps for georeferenced data publishing and visual exploration. Instead of delivering geographic knowledge to the client in polygon?based vector formats, we encode geographic object IDs and shape information into highly compact pixel images (decoding maps). This allows the combination of raster and vector characteristics while avoiding the problems in the currently existing pixel?based (raster?image?based) or vector?based techniques. Differing from traditional pixel?based techniques that are static and allow very little user interaction, our pixel?based technique allows varieties of sub?second (less than 1?s) interface controls such as dynamic query, dynamic classification, geographic object data identification, user?setting adjusting, as well as turning on/off layers, panning and zooming, with no or minimum server support. Compared with Web GIS techniques that use vector geographic data, our technique avoids transferring over the network large vector geographic data. It also avoids the non?trivial client?side computation to interpret the vector data and render the maps. Our technique features a short initial download time, near?constant performance scalability for larger numbers of geographic objects, and download?map?segment?only?when?necessary which potentially reduces the overall data transfer over the network. As a result, it accommodates general public users with slow modem network connections and low?end machines, as well as users with fast T?1 connections and fast machines. The client?side (browser) is implemented as lightweight Java applets. YMap, an easy?to?use, user?task?oriented highly interactive mapping tool for visual georeferenced data exploration is implemented using this technique. The performance comparison of YMap to some other vector?based Web GIS demonstrates the feasibility and benefits of this technique.This paper describes a pixel?based technique that enables highly interactive Web choropleth maps for georeferenced data publishing and visual exploration. Instead of delivering geographic knowledge to the client in polygon?based vector formats, we encode geographic object IDs and shape information into highly compact pixel images (decoding maps). This allows the combination of raster and vector characteristics while avoiding the problems in the currently existing pixel?based (raster?image?based) or vector?based techniques. Differing from traditional pixel?based techniques that are static and allow very little user interaction, our pixel?based technique allows varieties of sub?second (less than 1?s) interface controls such as dynamic query, dynamic classification, geographic object data identification, user?setting adjusting, as well as turning on/off layers, panning and zooming, with no or minimum server support. Compared with Web GIS techniques that use vector geographic data, our technique avoids transferring over the network large vector geographic data. It also avoids the non?trivial client?side computation to interpret the vector data and render the maps. Our technique features a short initial download time, near?constant performance scalability for larger numbers of geographic objects, and download?map?segment?only?when?necessary which potentially reduces the overall data transfer over the network. As a result, it accommodates general public users with slow modem network connections and low?end machines, as well as users with fast T?1 connections and fast machines. The client?side (browser) is implemented as lightweight Java applets. YMap, an easy?to?use, user?task?oriented highly interactive mapping tool for visual georeferenced data exploration is implemented using this technique. The performance comparison of YMap to some other vector?based Web GIS demonstrates the feasibility and benefits of this technique. %B International Journal of Geographical Information Science %V 19 %P 413 - 428 %8 2005/// %@ 1365-8816 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/1365881051233125120 %N 4 %R 10.1080/1365881051233125120 %0 Journal Article %J Science %D 2005 %T Comparative Genomics of Trypanosomatid Parasitic Protozoa %A El‐Sayed, Najib M. %A Myler,Peter J. %A Blandin,Gaëlle %A Berriman,Matthew %A Crabtree,Jonathan %A Aggarwal,Gautam %A Caler,Elisabet %A Renauld,Hubert %A Worthey,Elizabeth A. %A Hertz-Fowler,Christiane %A Ghedin,Elodie %A Peacock,Christopher %A Bartholomeu,Daniella C. %A Haas,Brian J. %A Tran,Anh-Nhi %A Wortman,Jennifer R. %A Alsmark,U. Cecilia M. %A Angiuoli,Samuel %A Anupama,Atashi %A Badger,Jonathan %A Bringaud,Frederic %A Cadag,Eithon %A Carlton,Jane M. %A Cerqueira,Gustavo C. %A Creasy,Todd %A Delcher,Arthur L. %A Djikeng,Appolinaire %A Embley,T. Martin %A Hauser,Christopher %A Ivens,Alasdair C. %A Kummerfeld,Sarah K. %A Pereira-Leal,Jose B. %A Nilsson,Daniel %A Peterson,Jeremy %A Salzberg,Steven L. %A Shallom,Joshua %A Silva,Joana C. %A Sundaram,Jaideep %A Westenberger,Scott %A White,Owen %A Melville,Sara E. %A Donelson,John E. %A Andersson,Björn %A Stuart,Kenneth D. %A Hall,Neil %X A comparison of gene content and genome architecture of Trypanosoma brucei, Trypanosoma cruzi, and Leishmania major, three related pathogens with different life cycles and disease pathology, revealed a conserved core proteome of about 6200 genes in large syntenic polycistronic gene clusters. Many species-specific genes, especially large surface antigen families, occur at nonsyntenic chromosome-internal and subtelomeric regions. Retroelements, structural RNAs, and gene family expansion are often associated with syntenic discontinuities that—along with gene divergence, acquisition and loss, and rearrangement within the syntenic regions—have shaped the genomes of each parasite. Contrary to recent reports, our analyses reveal no evidence that these species are descended from an ancestor that contained a photosynthetic endosymbiont. %B Science %V 309 %P 404 - 409 %8 2005/07/15/ %G eng %U http://www.sciencemag.org/content/309/5733/404.abstract %N 5733 %R 10.1126/science.1112181 %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International %D 2005 %T Comparing the Performance of High-Level Middleware Systems in Shared and Distributed Memory Parallel Environments %A Kim,Jik-Soo %A Andrade,H. %A Sussman, Alan %K Application software %K Computer science %K Computer vision %K Data analysis %K Distributed computing %K distributed computing environment %K distributed memory parallel environment %K distributed shared memory systems %K Educational institutions %K high-level middleware system %K I/O-intensive data analysis application %K Libraries %K Middleware %K parallel computing environment %K parallel library support %K parallel memories %K programming language %K programming languages %K Runtime environment %K shared memory parallel environment %K Writing %X The utilization of toolkits for writing parallel and/or distributed applications has been shown to greatly enhance developer's productivity. Such an approach hides many of the complexities associated with writing these applications, rather than relying solely on programming language aids and parallel library support, such as MPI or PVM. In this work, we evaluate three different middleware systems that have been used to implement a computation and I/O-intensive data analysis application from the domain of computer vision. This study shows the benefits and overheads associated with each of the middleware systems, in different homogeneous computational environments and with different workloads. Our results lead the way toward being able to make better decisions for tuning the application environment, for selecting the appropriate middleware, and also for designing more powerful middleware systems to efficiently build and run highly complex applications in both parallel and distributed computing environments. %B Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International %I IEEE %P 30 - 30 %8 2005/04// %@ 0-7695-2312-9 %G eng %R 10.1109/IPDPS.2005.144 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T A Comparison of Voice Controlled and Mouse Controlled Web Browsing (2000) %A Christian,Kevin %A Kules,Bill %A Shneiderman, Ben %A Youssef,Adel %K Technical Report %X Voice controlled web browsers allow users to navigate by speaking the text of a link or an associated number instead of clicking with a mouse. One such browser is Conversa, by Conversational Computing. This within subjects study with 18 subjects compared voice browsing with traditional mouse-based browsing. It attempted to identify which of three common hypertext forms (linear slide show, grid/tiled map, and hierarchical menu) are well suited to voice navigation, and whether voice navigation is helped by numbering links. The study shows that voice control adds approximately 50 percent to the performance time for certain types of tasks. Subjective satisfaction measures indicate that for voice browsing, textual links are preferable to numbered links. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6475 %0 Conference Paper %B Computer Vision and Pattern Recognition-Workshops, 2005. CVPR Workshops. IEEE Computer Society Conference on %D 2005 %T Computer vision on FPGAs: Design methodology and its application to gesture recognition %A Sen,M. %A Corretjer,I. %A Haim,F. %A Saha,S. %A Bhattacharyya, Shuvra S. %A Schlessman,J. %A Wolf,W. %B Computer Vision and Pattern Recognition-Workshops, 2005. CVPR Workshops. IEEE Computer Society Conference on %P 133 - 133 %8 2005/// %G eng %0 Journal Article %J INTERNATIONAL JOURNAL OF HUMAN–COMPUTER INTERACTION %D 2005 %T Content Index to Volume 18 %A Kuniavsky,M. %A Vaughan,M. %A Bederson, Benjamin B. %A Shneiderman, Ben %A Rau,P.L.P. %A from Menus,T. %A Lane,D.M. %A Napier,H.A. %A Peres,S.C. %A Sándor,A. %B INTERNATIONAL JOURNAL OF HUMAN–COMPUTER INTERACTION %V 18 %P 367 - 368 %8 2005/// %G eng %N 3 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Coordinating Overviews and Detail Views of WWW Log Data (2000) %A Hochheiser,Harry %A Shneiderman, Ben %K Technical Report %X Web server log analysis tools provide site operators with useful information regarding the visitors to their sites. Unfortunately, the utility of these tools is often limited by the use of aggregate summaries that hide the information associated with individual requests, and by the absence of contextual data that might help users interpret those summaries. Building upon earlier work in the use of starfield visualizations to display web site requests as individual data points [8], this paper describes the use of multiple-coordinated visualizations of web log data at varying granularities, and alongside additional related displays of appropriate contextual information. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6476 %0 Journal Article %J Multimedia Tools and Applications %D 2005 %T The CPR model for summarizing video %A Fayzullin,M. %A V.S. Subrahmanian %A Picariello, A. %A Sapino,M. L %X Most past work on video summarization has been based on selecting key frames from videos. We propose a model of video summarization based on three important parameters: Priority (of frames), Continuity (of the summary), and non-Repetition (of the summary). In short, a summary must include high priority frames and must be continuous and non-repetitive. An optimal summary is one that maximizes an objective function based on these three parameters. We show examples of how CPR parameters can be computed and provide algorithms to find optimal summaries based on the CPR approach. Finally, we briefly report on the performance of these algorithms. %B Multimedia Tools and Applications %V 26 %P 153 - 173 %8 2005/// %G eng %N 2 %R 10.1007/s11042-005-0451-7 %0 Journal Article %J Commun. ACM %D 2005 %T Crisis and opportunity in computer science %A Klawe,Maria %A Shneiderman, Ben %X The future of the field depends on winning back student enrollment, public interest in technology, and government research funding. %B Commun. ACM %V 48 %P 27 - 28 %8 2005/11// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/1096000.1096024 %N 11 %R 10.1145/1096000.1096024 %0 Journal Article %J Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %D 2005 %T Critical Factors Influencing the Occurrence of Vibrio Cholerae in the Environment of Bangladesh %A Huq,Anwar %A Sack,R. Bradley %A Nizam,Azhar %A Longini,Ira M. %A Nair,G. Balakrish %A Ali,Afsar %A Morris,J. Glenn %A Khan,M. N. Huda %A Siddique,A. Kasem %A Yunus,Mohammed %A Albert,M. John %A Sack,David A. %A Rita R Colwell %X The occurrence of outbreaks of cholera in Africa in 1970 and in Latin America in 1991, mainly in coastal communities, and the appearance of the new serotype Vibrio cholerae O139 in India and subsequently in Bangladesh have stimulated efforts to understand environmental factors influencing the growth and geographic distribution of epidemic Vibrio cholerae serotypes. Because of the severity of recent epidemics, cholera is now being considered by some infectious disease investigators as a “reemerging” disease, prompting new work on the ecology of vibrios. Epidemiological and ecological surveillance for cholera has been under way in four rural, geographically separated locations in Bangladesh for the past 4 years, during which both clinical and environmental samples were collected at biweekly intervals. The clinical epidemiology portion of the research has been published (Sack et al., J. Infect. Dis. 187:96-101, 2003). The results of environmental sampling and analysis of the environmental and clinical data have revealed significant correlations of water temperature, water depth, rainfall, conductivity, and copepod counts with the occurrence of cholera toxin-producing bacteria (presumably V. cholerae). The lag periods between increases or decreases in units of factors, such as temperature and salinity, and occurrence of cholera correlate with biological parameters, e.g., plankton population blooms. The new information on the ecology of V. cholerae is proving useful in developing environmental models for the prediction of cholera epidemics. %B Applied and Environmental MicrobiologyAppl. Environ. Microbiol. %V 71 %P 4645 - 4654 %8 2005/08/01/ %@ 0099-2240, 1098-5336 %G eng %U http://aem.asm.org/content/71/8/4645 %N 8 %R 10.1128/AEM.71.8.4645-4654.2005 %0 Journal Article %J Information Hiding %D 2005 %T Data hiding in compiled program binaries for enhancing computer system performance %A Swaminathan,A. %A Mao,Y. %A Wu,M. %A Kailas,K. %X Information hiding has been studied in many security applications such as authentication, copyright management and digital forensics. In this work, we introduce a new application where successful information hiding in compiled program binaries could bring system-wide performance improvements. Our goal is to enhance computer system performance by providing additional information to the processor, without changing the instruction set architecture. We first analyze the statistics of typical programs to demonstrate the feasibility of hiding data in them. We then propose several techniques to hide a large amount of data in the operand fields with very low computation and storage requirements during the extraction process. The data embedding is made reversible to recover the original instructions and to ensure the correct execution of the computer program. Our experiments on the SPEC CPU2000 benchmark programs show that up to 110K bits of information can be embedded in large programs with as little as 3K bits of additional run-time memory in the form of a simple look-up table. %B Information Hiding %P 357 - 371 %8 2005/// %G eng %R 10.1007/11558859_26 %0 Conference Paper %B Parallel and Distributed Processing Symposium, International %D 2005 %T Data Redistribution and Remote Method Invocation in Parallel Component Architectures %A Bertrand,Felipe %A Bramley,Randall %A Sussman, Alan %A Bernholdt,David E. %A Kohl,James A. %A Larson,Jay W. %A Damevski,Kostadin B. %X With the increasing availability of high-performance massively parallel computer systems, the prevalence of sophisticated scientific simulation has grown rapidly. The complexity of the scientific models being simulated has also evolved, leading to a variety of coupled multi-physics simulation codes. Such cooperating parallel programs require fundamentally new interaction capabilities, to efficiently exchange parallel data structures and collectively invoke methods across programs. So-called "M?N" research, as part of the Common Component Architecture (CCA) effort, addresses these special and challenging needs, to provide generalized interfaces and tools that support flexible parallel data redistribution and parallel remote method invocation. Using this technology, distinct simulation codes with disparate distributed data decompositions can work together to achieve greater scientific discoveries. %B Parallel and Distributed Processing Symposium, International %I IEEE Computer Society %C Los Alamitos, CA, USA %V 1 %P 40b - 40b %8 2005/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/IPDPS.2005.159 %0 Journal Article %J Algorithms–ESA 2005 %D 2005 %T On degree constrained shortest paths %A Khuller, Samir %A Lee,K. %A Shayman,M. %X Traditional shortest path problems play a central role in both the design and use of communication networks and have been studied extensively. In this work, we consider a variant of the shortest path problem. The network has two kinds of edges, “actual” edges and “potential” edges. In addition, each vertex has a degree/interface constraint. We wish to compute a shortest path in the graph that maintains feasibility when we convert the potential edges on the shortest path to actual edges. The central difficulty is when a node has only one free interface, and the unconstrained shortest path chooses two potential edges incident on this node. We first show that this problem can be solved in polynomial time by reducing it to the minimum weighted perfect matching problem. The number of steps taken by this algorithm is O(|E|2 log |E|) for the single-source single-destination case. In other words, for each v we compute the shortest path P v such that converting the potential edges on P v to actual edges, does not violate any degree constraint. We then develop more efficient algorithms by extending Dijkstra’s shortest path algorithm. The number of steps taken by the latter algorithm is O(|E||V|), even for the single-source all destination case. %B Algorithms–ESA 2005 %P 259 - 270 %8 2005/// %G eng %R 10.1007/11561071_25 %0 Conference Paper %B Proceedings of the 2nd conference on Symposium on Networked Systems Design & Implementation - Volume 2 %D 2005 %T Design and implementation of a routing control platform %A Caesar,Matthew %A Caldwell,Donald %A Feamster, Nick %A Rexford,Jennifer %A Shaikh,Aman %A van der Merwe,Jacobus %X The routers in an Autonomous System (AS) must distribute the information they learn about how to reach external destinations. Unfortunately, today's internal Border Gateway Protocol (iBGP) architectures have serious problems: a "full mesh" iBGP configuration does not scale to large networks and "route reflection" can introduce problems such as protocol oscillations and persistent loops. Instead, we argue that a Routing Control Platform (RCP) should collect information about external destinations and internal topology and select the BGP routes for each router in an AS. RCP is a logically-centralized platform, separate from the IP forwarding plane, that performs route selection on behalf of routers and communicates selected routes to the routers using the unmodified iBGP protocol. RCP provides scalability without sacrificing correctness. In this paper, we present the design and implementation of an RCP prototype on commodity hardware. Using traces of BGP and internal routing data from a Tier-1 backbone, we demonstrate that RCP is fast and reliable enough to drive the BGP routing decisions for a large network. We show that RCP assigns routes correctly, even when the functionality is replicated and distributed, and that networks using RCP can expect comparable convergence delays to those using today's iBGP architectures. %B Proceedings of the 2nd conference on Symposium on Networked Systems Design & Implementation - Volume 2 %S NSDI'05 %I USENIX Association %C Berkeley, CA, USA %P 15 - 28 %8 2005/// %G eng %U http://dl.acm.org/citation.cfm?id=1251203.1251205 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Designing a Metadata-Driven Visual Information Browser for Federal Statistics (2003) %A Kules,Bill %A Shneiderman, Ben %K Technical Report %X When looking for federal statistics, finding the right table, chart or report can be a daunting task for anyone not thoroughly familiar with the federal statistical system. Search tools help, but differing terminologies within the statistical agencies and a lack of familiarity of terms by information seekers limit their effectiveness. The FedStats Browser is a design for visually browsing federal agency statistical products and publications, using techniques that allow users to reformulate queries and iteratively refine results via simple, reversible actions with immediate feedback. This paper also discusses the characteristics of metadata needed for such a browser and the challenges inherent in acquiring that metadata. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6500 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Designing for Fun: Can we design user interfaces to be more fun? (2004) %A Shneiderman, Ben %K Technical Report %X User interfaces are taking their place in the world of fashion and style, which is great news. Just as dining out is more than getting a balanced diet and wearing clothes is more than staying warm, using interfaces is becoming a personal statement. It great that designers are turning attention to fun, as a separate design space, distinct from functionality, usability, and reliability. Did anyone notice that fun is part of functionality? %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6520 %0 Book %D 2005 %T Designing The User Interface %A Shneiderman, Ben %A Plaisant, Catherine %I Pearson Addison Wesley, USA %8 2005/// %G eng %0 Conference Paper %B Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on %D 2005 %T Detecting rotational symmetries %A Shiv Naga Prasad,V. %A Davis, Larry S. %K axial %K computational %K detection; %K field; %K flow; %K geometry; %K gradient %K graph %K graph; %K image %K image; %K magnitude %K methods; %K multiple %K n-sided %K object %K polygons; %K recognition; %K rotational %K symmetries; %K symmetry; %K theory; %K tire %K tyres; %K vector %X We present an algorithm for detecting multiple rotational symmetries in natural images. Given an image, its gradient magnitude field is computed, and information from the gradients is spread using a diffusion process in the form of a gradient vector flow (GVF) field. We construct a graph whose nodes correspond to pixels in tire image, connecting points that are likely to be rotated versions of one another The n-cycles present in tire graph are made to vote for Cn symmetries, their votes being weighted by the errors in transformation between GVF in the neighborhood of the voting points, and the irregularity of the n-sided polygons formed by the voters. The votes are accumulated at tire centroids of possible rotational symmetries, generating a confidence map for each order of symmetry. We tested the method with several natural images. %B Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on %V 2 %P 954 - 961 Vol. 2 - 954 - 961 Vol. 2 %8 2005/10// %G eng %R 10.1109/ICCV.2005.71 %0 Conference Paper %D 2005 %T Detection of denial-of-message attacks on sensor network broadcasts %A McCune, J.M. %A Elaine Shi %A Perrig, A. %A Reiter, M.K. %K authenticated acknowledgments %K broadcast channels %K broadcast protocols %K broadcasting base station %K countermeasures %K denial-of-message attacks %K DoM %K game theory %K game-theoretic approach %K malicious sensor nodes %K Mobile computing %K optimal attacker %K probabilistic detection %K probability %K Protocols %K Sampling methods %K secure implicit sampling %K sensor network broadcasts %K SIS %K telecommunication security %K Wireless sensor networks %X So far sensor network broadcast protocols assume a trustworthy environment. However in safety and mission-critical sensor networks this assumption may not be valid and some sensor nodes might be adversarial. In these environments, malicious sensor nodes can deprive other nodes from receiving a broadcast message. We call this attack a denial-of-message attack (DoM). In this paper we model and analyze this attack, and present countermeasures. We present SIS, a secure implicit sampling scheme that permits a broadcasting base station to probabilistically detect the failure of nodes to receive its broadcast, even if these failures result from an attacker motivated to induce these failures undetectably. SIS works by eliciting authenticated acknowledgments from a subset of nodes per broadcast, where the subset is unpredictable to the attacker and tunable so as to mitigate acknowledgment implosion on the base station. We use a game-theoretic approach to evaluate this scheme in the face of an optimal attacker that attempts to maximize the number of nodes it denies the broadcast while remaining undetected by the base station, and show that SIS significantly constrains such an attacker even in sensor networks exhibiting high intrinsic loss rates. We also discuss extensions that permit more targeted detection capabilities. %P 64 - 78 %8 2005 %G eng %0 Report %D 2005 %T Determining Causes and Severity of End-User Frustration (2002) %A Ceaparu,Irina %A Lazar,Jonathan %A Bessiere,Katie %A Robinson,John %A Shneiderman, Ben %K Technical Report %X While computers are beneficial to individuals and society, frequently, users encounter frustrating experiences when using computers. This study attempts to measure, through 111 participants, the frequency, cause, and the level of severity of frustrating experiences. The data showed that frustrating experiences happen on a frequent basis. The applications in which the frustrating experiences happened most frequently were web browsing, e-mail, and word processing. The most-cited causes of frustrating experiences were error messages, dropped network connections, long download times, and hard-to-find features. The time lost due to frustrating experiences ranged from 47-53% of time spent on a computer depending on the location and study method. After discarding extreme cases the time lost was still above 38%. These disturbing results should be a basis for future study. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6493 %0 Conference Paper %B INFOCOM 2005. 24th Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings IEEE %D 2005 %T Differentiated traffic engineering for QoS provisioning %A Tabatabaee,V. %A Bhattacharjee, Bobby %A La,R.J. %A Shayman,M.A. %K based %K Computer %K differentiated %K DiffServ %K DTE; %K engineering; %K evaluation; %K link; %K links; %K management; %K multipath %K network %K networks; %K nonconvex %K of %K optimisation; %K OPTIMIZATION %K packet %K performance %K problem; %K provisioning; %K QoS %K QUALITY %K routing; %K service; %K simulation-based %K source %K Telecommunication %K traffic %K traffic; %X We introduce a new approach for QoS provisioning in packet networks based on the notion of differentiated traffic engineering (DTE). We consider a single AS network capable of source based multi-path routing. We do not require sophisticated queuing or per-class scheduling at individual routers; instead, if a link is used to forward QoS sensitive packets, we maintain its utilization below a threshold. As a consequence, DTE eliminates the need for per-flow (IntServ) or per-class (DiffServ) packet processing tasks such as traffic classification, queueing, shaping, policing and scheduling in the core and hence poses a lower burden on the network management unit. Conversely, DTE utilizes network bandwidth much more efficiently than simple over-provisioning. In this paper, we propose a complete architecture and an algorithmic structure for DTE. We show that our scheme can be formulated as a non-convex optimization problem, and we present an optimal solution framework based on simulated annealing. We present a simulation-based performance evaluation of DTE, and compare our scheme to existing (gradient projection) methods. %B INFOCOM 2005. 24th Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings IEEE %V 4 %P 2349 - 2359 vol. 4 - 2349 - 2359 vol. 4 %8 2005/03// %G eng %R 10.1109/INFCOM.2005.1498521 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Direct Annotation: A Drag-and-Drop Strategy for Labeling Photos (2000) %A Shneiderman, Ben %A Kang,Hyunmo %K Technical Report %X Annotating photos is such a time-consuming, tedious and error-prone data entry task that it discourages most owners of personal photo libraries. By allowing users to drag labels such as personal names from a scrolling list and drop them on a photo, we believe we can make the task faster, easier and more appealing. Since the names are entered in a database, searching for all photos of a friend or family member is dramatically simplified. We describe the user interface design and the database schema to support direct annotation, as implemented in our PhotoFinder prototype. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6471 %0 Book Section %B Computational Logic in Multi-Agent SystemsComputational Logic in Multi-Agent Systems %D 2005 %T Distributed Algorithms for Dynamic Survivability of Multiagent Systems %A Subrahmanian,V. %A Kraus,Sarit %A Zhang,Yingqian %E Dix,Jürgen %E Leite,João %K Computer %K Science %X Though multiagent systems (MASs) are being increasingly used, few methods exist to ensure survivability of MASs. All existing methods suffer from two flaws. First, a centralized survivability algorithm (CSA) ensures survivability of the MAS – unfortunately, if the node on which the CSA exists goes down, the survivability of the MAS is questionable. Second, no mechanism exists to change how the MAS is deployed when external factors trigger a re-evaluation of the survivability of the MAS. In this paper, we present three algorithms to address these two important problems. Our algorithms can be built on top of any CSA. Our algorithms are completely distributed and can handle external triggers to compute a new deployment. We report on experiments assessing the efficiency of these algorithms. %B Computational Logic in Multi-Agent SystemsComputational Logic in Multi-Agent Systems %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3259 %P 139 - 144 %8 2005/// %@ 978-3-540-24010-5 %G eng %U http://dx.doi.org/10.1007/978-3-540-30200-1_1 %0 Conference Paper %B PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON DATA ENGINEERING %D 2005 %T A distributed quadtree index for peer-to-peer settings %A Tanin,E. %A Harwood,A. %A Samet, Hanan %B PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON DATA ENGINEERING %V 21 %P 254 - 254 %8 2005/// %G eng %0 Report %D 2005 %T Dynamic Aggregation to Support Pattern Discovery: A Case Study with Web Logs (2001) %A Tang,Lida %A Shneiderman, Ben %K Technical Report %X Rapid growth of digital data collections is overwhelming the capabilities of humans to comprehend them without aid. The extraction of useful data from large raw data sets is something that humans do poorly because of the overwhelming amount of information. Aggregation is a technique that extracts important aspect from groups of data thus reducing the amount that the user has to deal with at one time, thereby enabling them to discover patterns, outliers, gaps, and clusters. Previous mechanisms for interactive exploration with aggregated data was either too complex to use or too limited in scope. This paper proposes a new technique for dynamic aggregation that can combine with dynamic queries to support most of the tasks involved in data manipulation. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6487 %0 Journal Article %J Science of Computer Programming %D 2005 %T Dynamic inference of polymorphic lock types %A Rose,James %A Swamy,Nikhil %A Hicks, Michael W. %K Data races %K Dynamic analysis %K java %K Multithreading %K Type inference %X We present FindLocks, an approach for automatically proving the absence of data races in multi-threaded Java programs, using a combination of dynamic and static analysis. The program in question is instrumented so that when executed it will gather information about locking relationships. This information is then used to automatically generate annotations needed to type check the program using the Race-Free Java type system. Programs that type check are sure to be free from races. We call this technique dynamic annotation inference. We describe the design and implementation of our approach, and our experience applying the tool to a variety of Java programs. We have found that when using a reasonably comprehensive test suite, which is easy for small programs but harder for larger ones, the approach generates useful annotations. %B Science of Computer Programming %V 58 %P 366 - 383 %8 2005/12// %@ 0167-6423 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167642305000705 %N 3 %R 10.1016/j.scico.2005.03.003 %0 Report %D 2005 %T Dynamic Layout Management in a Multimedia Bulletin Board (2002) %A Kang,Hyunmo %A Shneiderman, Ben %A Wolff,Gregory J %K Technical Report %X This paper proposes a novel user interface to manage the dynamic layout of multimedia objects in the Multimedia Bulletin Board (MBB) system. The MBB has been designed and implemented as a prototype of an asynchronous communication system that enables rich communication and collaboration among users of multimedia objects such as text, image, moving picture, sound, voice, web, office document, and other files. The layout properties of the multimedia objects on a board (e.g. x-y position, size, z-order, partial occlusion, explicit and implicit links, etc.) show important and useful information on the user dynamics occurring within a board. However, a fixed layout created and edited by multiple users may prevent users from recognizing and identifying useful information. This paper resolves this problem with a novel user-controlled layout strategy made visible with dynamic layout templates (DLT). Users can reorganize the objects to extract meaningful information related to time, source, geographic location, or topic. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6489 %0 Report %D 2005 %T Dynamic Queries and Brushing on Choropleth Maps (2001) %A Dang,Gunjan %A North,Chris %A Shneiderman, Ben %K Technical Report %X Users who must combine demographic, economic or other data in a geographic context are often hampered by the integration of tabular and map representations. Static, paper-based solutions limit the amount of data that can be placed on a single map or table. By providing an effective user interface, we believe that researchers, journalists, teachers, and students can explore complex data sets more rapidly and effectively. This paper presents Dynamaps, a generalized map-based information-visualization tool for dynamic queries and brushing on choropleth maps. Users can use color coding to show a variable on each geographic region, and then filter out areas that do not meet the desired criteria. In addition, a scattergram view and a detailson- demand window support overviews and specific fact finding. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6481 %0 Report %D 2005 %T Dynamic Query Chloropleth Maps for Information Seeking and Decision Making (2003) %A Norman,Kent L %A Zhao,Haixia %A Shneiderman, Ben %A Golub,Evan %K Technical Report %X Information retrieval and visualization can be combined in dynamic query systems that allow users unparalleled access to information for decision making. In this paper, we report on the development and evaluation of a dynamic query system (YMap) that displays information on a chloropleth map using double thumb sliders to select ranges of query variables. The YMap prototype is a Java-Applet that supports panning and zooming. Several usability studies were conducted on early prototypes that resulted in the current version. Applications of YMap for decision making tasks are discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6505 %0 Report %D 2005 %T Dynamic query tools for time series data sets: Timebox widgets for interactive exploration (2004) %A Hochheiser,Harry %A Shneiderman, Ben %K Technical Report %X Timeboxes are rectangular widgets that can be used in direct-manipulation graphical user interfaces (GUIs) to specify query constraints on time series data sets. Timeboxes are used to specify simultaneously two sets of constraints: given a set of N time series profiles, a timebox covering time periods x1...x2 (x1rx2) and values y1...y2 (y1ry2) will retrieve only those nAN that have values y1ryry2 during all times x1rxrx2. TimeSearcher is an information visualization tool that combines timebox queries with overview displays, query-byexample facilities, and support for queries over multiple time-varying attributes. Query manipulation tools including pattern inversion and eaders & laggards graphical bookmarks provide additional support for interactive exploration of data sets. Extensions to the basic timebox model that provide additional expressivity include variable time timeboxes, which can be used to express queries with variability in the time interval, and angular queries, which search for ranges of differentials, rather than absolute values. Analysis of the algorithmic requirements for providing dynamic query performance for timebox queries showed that a sequential search outperformed searches based on geometric indices. Design studies helped identify the strengths and weaknesses of the query tools. Extended case studies involving the analysis of two different types of data from molecular biology experiments provided valuable feedback and validated the utility of both the timebox model and the TimeSearcher tool.

Timesearcher is available [url=http://www.cs.umd.edu/hcil/ timesearcher]online.[/url] %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6521 %0 Report %D 2005 %T Dynamic Query Visualizations on World Wide Web Clients: A DHTML Solution for Maps and Scattergrams (2002) %A Golub,Evan %A Shneiderman, Ben %K Technical Report %X Dynamic queries are gaining popularity as a method for interactive information visualization. Many implementations have been made on personal computers, and there is increasing interest in web-based designs. While Java and Flash strategies have been developed, we believe that a Dynamic HTML implementation could help promote more widespread use. We implemented double-box range sliders with choropleth maps and scattergrams, which are two popular visualizations, using HTML layers and tables. This paper describes our methods for slider control, visual presentation, and displaying/updating results for these visualizations. Visual design issues innovations and performance enhancements were necessary to create viable designs. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6491 %0 Report %D 2005 %T Dynamic Querying for Pattern Identification in Microarray and Genomic Data (2003) %A Hochheiser,Harry %A Baehrecke,Eric H %A Mount, Stephen M. %A Shneiderman, Ben %K Technical Report %X Data sets involving linear ordered sequences are a recurring theme in bioinformatics. Dynamic query tools that support exploration of these data sets can be useful for identifying patterns of interest. This paper describes the use of one such tool TimeSearcher - to interactively explore linear sequence data sets taken from two bioinformatics problems. Microarray time course data sets involve expression levels for large numbers of genes over multiple time points. TimeSearcher can be used to interactively search these data sets for genes with expression profiles of interest. The occurrence frequencies of short sequences of DNA in aligned exons can be used to identify sequences that play a role in the pre-mRNA splicing. TimeSearcher can be used to search these data sets for candidate splicing signals. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6503 %0 Journal Article %J Circuits and Systems for Video Technology, IEEE Transactions on %D 2005 %T Efficient bandwidth resource allocation for low-delay multiuser video streaming %A Su,Guan-Ming %A M. Wu %K algorithm; %K allocation %K allocation; %K approaches; %K average %K Bandwidth %K channel %K channels; %K coding; %K complexity; %K computational %K control; %K distortion; %K dynamic %K efficient %K fine %K fluctuation; %K granularity %K look-ahead %K low-delay %K MPEG-4 %K multiple %K multiuser %K QUALITY %K quality; %K rate-control %K resource %K scalability %K scenario; %K single-user %K sliding-window %K streaming; %K Telecommunication %K utilization; %K video %K visual %X This paper studies efficient bandwidth resource allocation for streaming multiple MPEG-4 fine granularity scalability (FGS) video programs to multiple users. We begin with a simple single-user scenario and propose a rate-control algorithm that has low delay and achieves an excellent tradeoff between the average visual distortion and the quality fluctuation. The proposed algorithm employs two weight factors for adjusting the tradeoff, and the optimal choice of these factors is derived. We then extend to the multiuser case and propose a dynamic resource allocation algorithm with low delay and low computational complexity. By exploring the variations in the scene complexity of video programs as well as dynamically and jointly distributing the available system resources among users, our proposed algorithm provides low fluctuation of quality for each user, and can support consistent or differentiated quality among all users to meet applications' needs. Experimental results show that compared to traditional look-ahead sliding-window approaches, our algorithm can achieve comparable visual quality and channel utilization at a much lower cost of delay, computation, and storage. %B Circuits and Systems for Video Technology, IEEE Transactions on %V 15 %P 1124 - 1137 %8 2005/09// %@ 1051-8215 %G eng %N 9 %R 10.1109/TCSVT.2005.852626 %0 Journal Article %J Computer Methods in Applied Mechanics and Engineering %D 2005 %T Efficient iterative algorithms for the stochastic finite element method with application to acoustic scattering %A Elman, Howard %A Ernst, Oliver G. %A O’Leary,Dianne P. %A Stewart, Michael %K 01 %K 28 %K 60 %K finite element %K multigrid %K Multiple right-hand side %K Scattering %K Stochastic %X In this study, we describe the algebraic computations required to implement the stochastic finite element method for solving problems in which uncertainty is restricted to right-hand side data coming from forcing functions or boundary conditions. We show that the solution can be represented in a compact outer product form which leads to efficiencies in both work and storage, and we demonstrate that block iterative methods for algebraic systems with multiple right-hand sides can be used to advantage to compute this solution. We also show how to generate a variety of statistical quantities from the computed solution. Finally, we examine the behavior of these statistical quantities in one setting derived from a model of acoustic scattering. %B Computer Methods in Applied Mechanics and Engineering %V 194 %P 1037 - 1055 %8 2005/03/18/ %@ 0045-7825 %G eng %U http://www.sciencedirect.com/science/article/pii/S0045782504003123 %N 9–11 %R 10.1016/j.cma.2004.06.028 %0 Conference Paper %B Proceedings of the twenty-fourth annual ACM symposium on Principles of distributed computing %D 2005 %T Efficient lookup on unstructured topologies %A Morselli,Ruggero %A Bhattacharjee, Bobby %A Srinivasan, Aravind %A Marsh,Michael A %K distributed algorithms %K lookup protocols %K peer-to-peer networks %K random walks %X We present LMS, a protocol for efficient lookup on unstructured networks. Our protocol uses a virtual namespace without imposing specific topologies. It is more efficient than existing lookup protocols for unstructured networks, and thus is an attractive alternative for applications in which the topology cannot be structured as a Distributed Hash Table (DHT).We present analytic bounds for the worst-case performance of our protocol. Through detailed simulations (with up to 100,000 nodes), we show that the actual performance on realistic topologies is significantly better. We also show in both simulations and a complete implementation (which includes over five hundred nodes) that our protocol is inherently robust against multiple node failures and can adapt its replication strategy to optimize searches according to a specific heuristic. Moreover, the simulation demonstrates the resilience of LMS to high node turnover rates, and that it can easily adapt to orders of magnitude changes in network size. The overhead incurred by LMS is small, and its performance approaches that of DHTs on networks of similar size. %B Proceedings of the twenty-fourth annual ACM symposium on Principles of distributed computing %S PODC '05 %I ACM %C New York, NY, USA %P 77 - 86 %8 2005/// %@ 1-58113-994-2 %G eng %U http://doi.acm.org/10.1145/1073814.1073828 %R 10.1145/1073814.1073828 %0 Conference Paper %B Proceedings of the 2005 national conference on Digital government research %D 2005 %T An efficient nearest neighbor algorithm for P2P settings %A Tanin,Egemen %A Nayar,Deepa %A Samet, Hanan %K geographic information systems %K nearest neighbor query %K peer-to-peer networks %K quadtree %K spatial data %X New Peer-to-Peer (P2P) applications like P2P job-employee seeker networks and P2P virtual cities, for application domains such as collaborative urban planning and forming virtual communities, are about to emerge. An important component in these applications is spatial data, i.e., data with locational components. Many requests initiated on spatial data involve finding the spatial objects that are nearest to a query location. In this paper, we propose an efficient algorithm that finds the spatial objects that are nearest to a given query location on a P2P network in the order of their minimum distance to the query point. The proposed algorithm makes use of a distributed spatial index that does not rely on the use of a central server. The algorithm is designed to be more efficient by utilizing the parallel nature of the P2P network. A demonstration of the proposed algorithm was implemented as a prototype P2P application that finds events and places of interest in a city. %B Proceedings of the 2005 national conference on Digital government research %S dg.o '05 %I Digital Government Society of North America %P 21 - 28 %8 2005/// %G eng %U http://dl.acm.org/citation.cfm?id=1065226.1065237 %0 Conference Paper %B Proceedings of the 13th annual ACM international workshop on Geographic information systems %D 2005 %T Efficient query processing on spatial networks %A Sankaranarayanan,Jagan %A Alborzi,Houman %A Samet, Hanan %K location-based services %K path coherence %K Query processing %K SILC framework %K Spatial databases %K spatial networks %X A framework for determining the shortest path and the distance between every pair of vertices on a spatial network is presented. The framework, termed SILC, uses path coherence between the shortest path and the spatial positions of vertices on the spatial network, thereby, resulting in an encoding that is compact in representation and fast in path and distance retrievals. Using this framework, a wide variety of spatial queries such as incremental nearest neighbor searches and spatial distance joins can be shown to work on datasets of locations residing on a spatial network of sufficiently large size. The suggested framework is suitable for both main memory and disk-resident datasets. %B Proceedings of the 13th annual ACM international workshop on Geographic information systems %S GIS '05 %I ACM %C New York, NY, USA %P 200 - 209 %8 2005/// %@ 1-59593-146-5 %G eng %U http://doi.acm.org/10.1145/1097064.1097093 %R 10.1145/1097064.1097093 %0 Journal Article %J UMD, CS Tech. Rep. CS-TR %D 2005 %T Efficient strategies for channel management in wireless lans %A Mishra,A. %A Brik,V. %A Banerjee,S. %A Srinivasan, Aravind %A Arbaugh, William A. %X We define efficient algorithms for channel management (channelassignment and load balancing among APs) in 802.11-based WLANs that lead to better usage of the wireless spectrum. These algorithms (called CFAssign) are based on a “conflict-free set coloring” for- mulation that jointly perform load balancing along with channel assignment. Such a formulation has a number of advantages. First, it explicitly captures interference effects at clients. Next, it intrinsi- cally exposes opportunities for better channel re-use. Finally, algo- rithms based on this formulation do not depend on specific physical RF models and hence can be applied efficiently to a wide-range of in-building as well as outdoor scenarios. We have performed extensive packet-level simulations and mea- surements on a deployed wireless testbed of 70 APs to validate the performance of our proposed algorithms. We show that in ad- dition to single network scenarios, CFAssign algorithms are well suited for channel assignment in scenarios where multiple wireless networks share the same physical space and contend for the same frequency spectrum. Our results over a wide range of scenarios indicate that CFAssign reduces the interference at clients by about 50-70% in comparison to current best-known techniques. %B UMD, CS Tech. Rep. CS-TR %V 4729 %8 2005/// %G eng %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International %D 2005 %T EMPS: an environment for memory performance studies %A Hollingsworth, Jeffrey K %A Snavely, A. %A Sbaraglia, S. %A Ekanadham, K. %K Application software %K cache storage %K Computational modeling %K Concurrent computing %K data gathering %K Delay %K digital simulation %K Economic forecasting %K EMP radiation effects %K high end computing system %K High performance computing %K memory architecture %K memory performance %K Moore's Law %K PARALLEL PROCESSING %K parallel program %K Predictive models %K simulation tool %K Writing %X This paper describes an overview of environment for memory performance studies (EMPS). EMPS is a framework to allow different data gathering and simulation tools to be composed together to predict the performance of parallel programs on a variety of current and future high end computing (HEC) systems. The framework seeks to combine the automated nature of direct execution simulation with the predictive capabilities of performance modeling. %B Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International %I IEEE %8 2005/04// %@ 0-7695-2312-9 %G eng %R 10.1109/IPDPS.2005.196 %0 Journal Article %J Proceedings of the American Society for Information Science and Technology %D 2005 %T Envisioning help resources for the future information ecology: Toward an enriched sense of help %A Haas,Stephanie W %A Brown,Laurie %A Denn,Sheila %A Locke,David %A Shneiderman, Ben %B Proceedings of the American Society for Information Science and Technology %V 42 %P n/a-n/a - n/a-n/a %8 2005/01/01/ %@ 1550-8390 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/meet.14504201123/abstract %N 1 %R 10.1002/meet.14504201123 %0 Journal Article %J SIAM Journal on Matrix Analysis and Applications %D 2005 %T Error Analysis of the Quasi-Gram–Schmidt Algorithm %A Stewart, G.W. %K Gram–Schmidt algorithm %K orthogonalization %K QR factorization %K rounding-error analysis %K sparse matrix %X Let the $n\,{\times}\,p$ $(n\geq p)$ matrix $X$ have the QR factorization $X = QR$, where $R$ is an upper triangular matrix of order $p$ and $Q$ is orthonormal. This widely used decomposition has the drawback that $Q$ is not generally sparse even when $X$ is. One cure is to discard $Q$, retaining only $X$ and $R$. Products like $a = Q\trp y = R\itp X\trp y$ can then be formed by computing $b = X\trp y$ and solving the system $R\trp a = b$. This approach can be used to modify the Gram--Schmidt algorithm for computing $Q$ and $R$ to compute $R$ without forming $Q$ or altering $X$. Unfortunately, this quasi-Gram--Schmidt algorithm can produce inaccurate results. In this paper it is shown that with reorthogonalization the inaccuracies are bounded under certain natural conditions. %B SIAM Journal on Matrix Analysis and Applications %V 27 %P 493 - 506 %8 2005/// %G eng %U http://link.aip.org/link/?SML/27/493/1 %N 2 %R 10.1137/040607794 %0 Conference Paper %B Proceedings of the 6th ACM SIGPLAN-SIGSOFT workshop on Program analysis for software tools and engineering %D 2005 %T Evaluating and tuning a static analysis to find null pointer bugs %A Hovemeyer,D. %A Spacco,J. %A Pugh, William %B Proceedings of the 6th ACM SIGPLAN-SIGSOFT workshop on Program analysis for software tools and engineering %P 13 - 19 %8 2005/// %G eng %0 Journal Article %J Innovations in Systems and Software Engineering %D 2005 %T An evolutionary testbed for software technology evaluation %A Lindvall,M. %A Rus,I. %A Shull, F. %A Zelkowitz, Marvin V %A Donzelli,P. %A Memon, Atif M. %A Basili, Victor R. %A Costa,P. %A Tvedt,R. %A Hochstein, L. %X Empirical evidence and technology evaluation are needed to close the gap between the state of the art and the state of the practice in software engineering. However, there are difficulties associated with evaluating technologies based on empirical evidence: insufficient specification of context variables, cost of experimentation, and risks associated with trying out new technologies. In this paper, we propose the idea of an evolutionary testbed for addressing these problems. We demonstrate the utility of the testbed in empirical studies involving two different research technologies applied to the testbed, as well as the results of these studies. The work is part of NASArsquos High Dependability Computing Project (HDCP), in which we are evaluating a wide range of new technologies for improving the dependability of NASA mission-critical systems. %B Innovations in Systems and Software Engineering %V 1 %P 3 - 11 %8 2005/// %G eng %N 1 %R 10.1007/s11334-005-0007-z %0 Journal Article %J Applications of evolutionary computing: EvoWorkshops 2005, EvoBIO, EvoCOMNET, EvoHOT, EvoIASP, EvoMUSART, and EvoSTOC %D 2005 %T EvoSTOC Contributions %A Merkle,D. %A Middendorf,M. %A Scheidler,A. %A Avigad,G. %A Moshaiov,A. %A Brauner,N. %A Parsopoulos,K.E. %A Vrahatis,M.N. %A Rand, William %A Riolo,R %B Applications of evolutionary computing: EvoWorkshops 2005, EvoBIO, EvoCOMNET, EvoHOT, EvoIASP, EvoMUSART, and EvoSTOC %8 2005/// %G eng %0 Journal Article %J Distributed Computing and Internet Technology %D 2005 %T Executable requirements specifications using triggered message sequence charts %A Sengupta,B. %A Cleaveland, Rance %B Distributed Computing and Internet Technology %P 482 - 493 %8 2005/// %G eng %0 Report %D 2005 %T Exploring Personal Media: A Spatial Interface Supporting User-Defined Semantic Regions (2004) %A Kang,Hyunmo %A Shneiderman, Ben %K Technical Report %X Graphical mechanisms for spatially organizing personal media data could enable users to fruitfully apply their mental models. This paper introduces Semantic Regions, an innovative way for users to construct mental models by drawing regions on 2D space and specifying the semantics for each region. Then users can apply personal ontologies to personal media data using the fling-and-flock metaphor. This allows photos (or other personal media) to be dragged to the display and automatically grouped according to time, geography, family trees, groups of friends, or other conceptual maps. The prototype implementation for Semantic Regions, MediaFinder, was validated with two usability tests for usage and construction of user-defined mental models. Applications and extensions are discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6513 %0 Report %D 2005 %T Extending the Utility of Treemaps with Flexible Hierarchy (2004) %A Chintalapani,Gouthami %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X Treemaps is a visualization technique for presenting hierarchical information on two dimensional displays. Prior implementations limit the visualization to pre-defined static hierarchies. Flexible hierarchy, a new capability of Treemap 4.0, enables users to define various hierarchies through dynamically selecting a series of data attributes so that they can discover patterns, clusters and outliers. This paper describes the design and implementation issues of flexible hierarchy. It then reports on a usability study which led to enhancements to the interface. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6514 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Extending User Understanding of Federal Statistics in Tables (2000) %A Marchionini,Gary %A Hert,Carol %A Liddy,Liz %A Shneiderman, Ben %K Technical Report %X This paper describes progress toward improving user interfaces for US Federal government statistics that are presented in tables. Based on studies of user behaviors and needs related to statistical tables, we describe interfaces to assist diverse users with a range of statistical literacy to explore, find, understand, and use US Federal government statistics. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6473 %0 Report %D 2005 %T Finding Governmental Statistical Data on the Web: Three Empirical Studies of the FedStats Topics Page (2003) %A Ceaparu,Irina %A Shneiderman, Ben %K Technical Report %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6509 %0 Book %D 2005 %T Finite Elements and Fast Iterative Solvers : with Applications in Incompressible Fluid Dynamics %A Elman, Howard %A Silvester, David J %A Wathen, Andrew J %K Computers / Computer Science %K Mathematics / Numerical Analysis %K Science / Mechanics / Fluids %K Science / Physics / General %K Technology & Engineering / Mechanical %X The subject of this book is the efficient solution of partial differential equations (PDEs) that arise when modelling incompressible fluid flow. The material is organized into four groups of two chapters each, covering the Poisson equation (chapters 1 & 2); the convection-diffucion equation (chapters 3 & 4); the Stokes equations (chapters 5 & 6); and the Navier-Stokes equations (chapters 7 & 8). These equations represent important models within the domain of computational fluid dynamics, but they also arise in many other settings. For each PDE model, there is a chapter concerned with finite element discretization. For each problem and associated solvers there is a description of how to compute along with theoretical analysis which guides the choice of approaches. Illustrative numerical results occur throughout the book, which have been computed with the freely downloadable IFISS software. All numerical results should be reproducible by readers who have access to MATLAB and there is considerable scope for experimentation in the 'computational laboratory' provided by the software. This book provides an excellent introduction to finite elements, iterative linear solvers and scientific computing aimed at graduates in engineering, numerical analysis, applied mathematics and interdisciplinary scientific computing. Including theoretical problems and practical exercises closely tied with freely downloadable MATLAB software, this book is an ideal teaching and learning resource. %I Oxford University Press %P 416 %8 2005 %@ 9780191523786 %G eng %0 Conference Paper %B Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems %D 2005 %T A framework for decomposing reputation in MAS into competence and integrity %A Smith,Michael J. %A desJardins, Marie %B Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems %S AAMAS '05 %I ACM %C New York, NY, USA %P 1253 - 254 %8 2005/// %@ 1-59593-093-0 %G eng %U http://doi.acm.org/10.1145/1082473.1082719 %R 10.1145/1082473.1082719 %0 Journal Article %J Inductive Logic Programming %D 2005 %T A framework for set-oriented computation in inductive logic programming and its application in generalizing inverse entailment %A Corrada Bravo, Hector %A Page,D. %A Ramakrishnan,R. %A Shavlik,J. %A Costa,V. S %B Inductive Logic Programming %P 69 - 86 %8 2005/// %G eng %0 Conference Paper %D 2005 %T Game-tree search with combinatorially large belief states %A Parker,A. %A Nau, Dana S. %A V.S. Subrahmanian %X In games such as kriegspiel chess (a chess variantwhere players have no direct knowledge of the op- ponent’s pieces’ locations) the belief state’s sizes dwarf those of other partial information games like bridge, scrabble, and poker–and there is no easy way to generate states satisfying the given observa- tions. We show that statistical sampling approaches can be developed to do well in such games. We show that it is not necessary for the random sample to consist only of game boards that satisfy each and every one of a player’s observations. In fact, we win 24% more often by beginning with such completely consistent boards and gradually switching (as the game progressed) to boards that are merely consistent with the latest observation. This surprising result is explained by noting that as the game progresses, a board that is consistent with the last move becomes more and more likely to be consistent with the entire set of observations, even if we have no idea what sequence of moves might have actually generated this board. %V 19 %P 254 - 254 %8 2005/// %G eng %U http://www.ijcai.org/papers/0878.pdf %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Generating and Querying Semantic Web Environments for Photo Libraries (2005) %A Axelrod,Adam %A Golbeck,Jennifer %A Shneiderman, Ben %K Technical Report %X Online photo libraries require a method to efficiently search a collection of photographs, and retrieve photos with similar attributes. Our motivation was to incorporate an existing collection of over 250 photographs of over 200 faculty members and events spanning 7 decades into a library called CS PhotoHistory that is available in hypertext and on the Semantic Web. In this paper, we identify challenges related to making this repository available on the Semantic Web, including issues of automation, modeling, and expressivity. Using CS PhotoHistory as a case study, we describe the process of creating an ontology and a querying interface for interacting with a digital photo library on the Semantic Web. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6538 %0 Conference Paper %B Proceedings of the second international workshop on Software engineering for high performance computing system applications %D 2005 %T Generating testable hypotheses from tacit knowledge for high productivity computing %A Asgari, Sima %A Hochstein, Lorin %A Basili, Victor R. %A Zelkowitz, Marvin V %A Hollingsworth, Jeffrey K %A Carver, Jeff %A Shull, Forrest %K folklore elicitation %K high productivity development time experimental studies %K hypothesis generation %K tacit knowledge solicitation %K testable hypotheses %X In this research, we are developing our understanding of how the high performance computing community develops effective parallel implementations of programs by collecting the folklore within the community. We use this folklore as the basis for a series of experiments, which we expect, will validate or negate these assumptions. %B Proceedings of the second international workshop on Software engineering for high performance computing system applications %S SE-HPCS '05 %I ACM %C New York, NY, USA %P 17 - 21 %8 2005/// %@ 1-59593-117-1 %G eng %U http://doi.acm.org/10.1145/1145319.1145325 %R 10.1145/1145319.1145325 %0 Journal Article %J Nature %D 2005 %T The genome of the protist parasite Entamoeba histolytica %A Loftus,B. %A Anderson,I. %A Davies,R. %A Alsmark,U. C.M %A Samuelson,J. %A Amedeo,P. %A Roncaglia,P. %A Berriman,M. %A Hirt,R. P %A Mann,B. J %A others %B Nature %V 433 %P 865 - 868 %8 2005 %G eng %N 7028 %0 Journal Article %J Science %D 2005 %T The Genome Sequence of Trypanosoma cruzi, Etiologic Agent of Chagas Disease %A El‐Sayed, Najib M. %A Myler,Peter J. %A Bartholomeu,Daniella C. %A Nilsson,Daniel %A Aggarwal,Gautam %A Tran,Anh-Nhi %A Ghedin,Elodie %A Worthey,Elizabeth A. %A Delcher,Arthur L. %A Blandin,Gaëlle %A Westenberger,Scott J. %A Caler,Elisabet %A Cerqueira,Gustavo C. %A Branche,Carole %A Haas,Brian %A Anupama,Atashi %A Arner,Erik %A Åslund,Lena %A Attipoe,Philip %A Bontempi,Esteban %A Bringaud,Frédéric %A Burton,Peter %A Cadag,Eithon %A Campbell,David A. %A Carrington,Mark %A Crabtree,Jonathan %A Darban,Hamid %A da Silveira,Jose Franco %A de Jong,Pieter %A Edwards,Kimberly %A Englund,Paul T. %A Fazelina,Gholam %A Feldblyum,Tamara %A Ferella,Marcela %A Frasch,Alberto Carlos %A Gull,Keith %A Horn,David %A Hou,Lihua %A Huang,Yiting %A Kindlund,Ellen %A Klingbeil,Michele %A Kluge,Sindy %A Koo,Hean %A Lacerda,Daniela %A Levin,Mariano J. %A Lorenzi,Hernan %A Louie,Tin %A Machado,Carlos Renato %A McCulloch,Richard %A McKenna,Alan %A Mizuno,Yumi %A Mottram,Jeremy C. %A Nelson,Siri %A Ochaya,Stephen %A Osoegawa,Kazutoyo %A Pai,Grace %A Parsons,Marilyn %A Pentony,Martin %A Pettersson,Ulf %A Pop, Mihai %A Ramirez,Jose Luis %A Rinta,Joel %A Robertson,Laura %A Salzberg,Steven L. %A Sanchez,Daniel O. %A Seyler,Amber %A Sharma,Reuben %A Shetty,Jyoti %A Simpson,Anjana J. %A Sisk,Ellen %A Tammi,Martti T. %A Tarleton,Rick %A Teixeira,Santuza %A Van Aken,Susan %A Vogt,Christy %A Ward,Pauline N. %A Wickstead,Bill %A Wortman,Jennifer %A White,Owen %A Fraser,Claire M. %A Stuart,Kenneth D. %A Andersson,Björn %X Whole-genome sequencing of the protozoan pathogen Trypanosoma cruzi revealed that the diploid genome contains a predicted 22,570 proteins encoded by genes, of which 12,570 represent allelic pairs. Over 50% of the genome consists of repeated sequences, such as retrotransposons and genes for large families of surface molecules, which include trans-sialidases, mucins, gp63s, and a large novel family (>1300 copies) of mucin-associated surface protein (MASP) genes. Analyses of the T. cruzi, T. brucei, and Leishmania major (Tritryp) genomes imply differences from other eukaryotes in DNA repair and initiation of replication and reflect their unusual mitochondrial DNA. Although the Tritryp lack several classes of signaling molecules, their kinomes contain a large and diverse set of protein kinases and phosphatases; their size and diversity imply previously unknown interactions and regulatory processes, which may be targets for intervention. %B Science %V 309 %P 409 - 415 %8 2005/07/15/ %G eng %U http://www.sciencemag.org/content/309/5733/409.abstract %N 5733 %R 10.1126/science.1112631 %0 Journal Article %J ScienceScience %D 2005 %T The Genome Sequence of Trypanosoma Cruzi, Etiologic Agent of Chagas Disease %A El‐Sayed, Najib M. %A Myler,Peter J. %A Bartholomeu,Daniella C. %A Nilsson,Daniel %A Aggarwal,Gautam %A Tran,Anh-Nhi %A Ghedin,Elodie %A Worthey,Elizabeth A. %A Delcher,Arthur L. %A Blandin,Gaëlle %A Westenberger,Scott J. %A Caler,Elisabet %A Cerqueira,Gustavo C. %A Branche,Carole %A Haas,Brian %A Anupama,Atashi %A Arner,Erik %A Åslund,Lena %A Attipoe,Philip %A Bontempi,Esteban %A Bringaud,Frédéric %A Burton,Peter %A Cadag,Eithon %A Campbell,David A. %A Carrington,Mark %A Crabtree,Jonathan %A Darban,Hamid %A da Silveira,Jose Franco %A de Jong,Pieter %A Edwards,Kimberly %A Englund,Paul T. %A Fazelina,Gholam %A Feldblyum,Tamara %A Ferella,Marcela %A Frasch,Alberto Carlos %A Gull,Keith %A Horn,David %A Hou,Lihua %A Huang,Yiting %A Kindlund,Ellen %A Klingbeil,Michele %A Kluge,Sindy %A Koo,Hean %A Lacerda,Daniela %A Levin,Mariano J. %A Lorenzi,Hernan %A Louie,Tin %A Machado,Carlos Renato %A McCulloch,Richard %A McKenna,Alan %A Mizuno,Yumi %A Mottram,Jeremy C. %A Nelson,Siri %A Ochaya,Stephen %A Osoegawa,Kazutoyo %A Pai,Grace %A Parsons,Marilyn %A Pentony,Martin %A Pettersson,Ulf %A Pop, Mihai %A Ramirez,Jose Luis %A Rinta,Joel %A Robertson,Laura %A Salzberg,Steven L. %A Sanchez,Daniel O. %A Seyler,Amber %A Sharma,Reuben %A Shetty,Jyoti %A Simpson,Anjana J. %A Sisk,Ellen %A Tammi,Martti T. %A Tarleton,Rick %A Teixeira,Santuza %A Van Aken,Susan %A Vogt,Christy %A Ward,Pauline N. %A Wickstead,Bill %A Wortman,Jennifer %A White,Owen %A Fraser,Claire M. %A Stuart,Kenneth D. %A Andersson,Björn %X Whole-genome sequencing of the protozoan pathogen Trypanosoma cruzi revealed that the diploid genome contains a predicted 22,570 proteins encoded by genes, of which 12,570 represent allelic pairs. Over 50% of the genome consists of repeated sequences, such as retrotransposons and genes for large families of surface molecules, which include trans-sialidases, mucins, gp63s, and a large novel family (>1300 copies) of mucin-associated surface protein (MASP) genes. Analyses of the T. cruzi, T. brucei, and Leishmania major (Tritryp) genomes imply differences from other eukaryotes in DNA repair and initiation of replication and reflect their unusual mitochondrial DNA. Although the Tritryp lack several classes of signaling molecules, their kinomes contain a large and diverse set of protein kinases and phosphatases; their size and diversity imply previously unknown interactions and regulatory processes, which may be targets for intervention. %B ScienceScience %V 309 %P 409 - 415 %8 2005/07/15/ %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/309/5733/409 %N 5733 %R 10.1126/science.1112631 %0 Journal Article %J Journal of VirologyJ. Virol. %D 2005 %T Genome-Wide Analysis of Chromosomal Features Repressing Human Immunodeficiency Virus Transcription %A Lewinski,M. K %A Bisgrove,D. %A Shinn,P. %A Chen,H. %A Hoffmann,C. %A Hannenhalli, Sridhar %A Verdin,E. %A Berry,C. C %A Ecker,J. R %A Bushman,F. D %X We have investigated regulatory sequences in noncoding human DNA that are associated with repression of an integrated human immunodeficiency virus type 1 (HIV-1) promoter. HIV-1 integration results in the formation of precise and homogeneous junctions between viral and host DNA, but integration takes place at many locations. Thus, the variation in HIV-1 gene expression at different integration sites reports the activity of regulatory sequences at nearby chromosomal positions. Negative regulation of HIV transcription is of particular interest because of its association with maintaining HIV in a latent state in cells from infected patients. To identify chromosomal regulators of HIV transcription, we infected Jurkat T cells with an HIV-based vector transducing green fluorescent protein (GFP) and separated cells into populations containing well-expressed (GFP-positive) or poorly expressed (GFP-negative) proviruses. We then determined the chromosomal locations of the two classes by sequencing 971 junctions between viral and cellular DNA. Possible effects of endogenous cellular transcription were characterized by transcriptional profiling. Low-level GFP expression correlated with integration in (i) gene deserts, (ii) centromeric heterochromatin, and (iii) very highly expressed cellular genes. These data provide a genome-wide picture of chromosomal features that repress transcription and suggest models for transcriptional latency in cells from HIV-infected patients. %B Journal of VirologyJ. Virol. %V 79 %P 6610 - 6619 %8 2005/06/01/ %@ 0022-538X, 1098-5514 %G eng %U http://jvi.asm.org/content/79/11/6610 %N 11 %R 10.1128/JVI.79.11.6610-6619.2005 %0 Journal Article %J Computer-Aided Design %D 2005 %T Geometric algorithms for computing cutter engagement functions in 2.5D milling operations %A Gupta, Satyandra K. %A Saini,Sunil K. %A Spranklin,Brent W. %A Yao,Zhiyang %K Cutter engagement %K Cutter path planning %K Feed rate adjustments %X Cutter engagement is a measure that describes what portion of the cutter is involved in machining at a given instant of time. During profile milling of complex geometries, cutter engagement varies significantly along the cutter path. Cutter engagement information helps in determining the efficiency of the cutter path and also helps in improving it by adjusting the feed rate. This paper describes geometric algorithms for computing piece-wise continuous closed-form cutter engagement functions for 2.5D milling operations. The results produced by our algorithm are compared with the results obtained by discrete simulations of the cutting process and appear to match very well. %B Computer-Aided Design %V 37 %P 1469 - 1480 %8 2005/12// %@ 0010-4485 %G eng %U http://www.sciencedirect.com/science/article/pii/S0010448505000801 %N 14 %R 10.1016/j.cad.2005.03.001 %0 Conference Paper %D 2005 %T Going far, logically %A Yaman,F. %A Nau, Dana S. %A V.S. Subrahmanian %X There are numerous applications where we need toensure that multiple moving objects are sufficiently far apart. Furthermore, in many moving object do- mains, there is positional indeterminacy — we are not 100% sure exactly when a given moving object will be at a given location. [Yaman et al., 2004] provided a logic of motion but did not provide algo- rithms to ensure that moving objects are kept suf- ficiently far apart. In this paper, we extend their logic to include a “far” predicate. We develop the CheckFar algorithm that checks if any given two objects will always be sufficiently far apart at dur- ing a time interval. We have run a set of experi- ments showing that our CheckFar algorithm scales very well. %V 19 %P 615 - 615 %8 2005/// %G eng %U http://www.cs.umd.edu/~nau/papers/yaman2005going.pdf %0 Conference Paper %B PROCEEDINGS OF THE NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE %D 2005 %T A graph theoretical foundation for integrating RDF ontologies %A Udrea,O. %A Deng,Y. %A Ruckhaus,E. %A V.S. Subrahmanian %X RDF ontologies are rapidly increasing in number. We studythe problem of integrating two RDF ontologies under a given set H of Horn clauses that specify semantic relationships be- tween terms in the ontology, as well as under a given set of negative constraints. We formally define the notion of a “wit- ness” to the integrability of two RDF ontologies under such constraints. A witness represents a way of integrating the ontologies together. We define a “minimal” witnesses and provide the polynomial CROW (Computing RDF Ontology Witness) algorithm to find a witness. We report on the per- formance of CROW both on DAML, SchemaWeb and Onto- Broker ontologies as well as on synthetically generated data. The experiments show that CROW works very well on real- life ontologies and scales to massive ontologies. %B PROCEEDINGS OF THE NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE %V 20 %P 1442 - 1442 %8 2005/// %G eng %0 Conference Paper %B Shape Modeling and Applications, 2005 International Conference %D 2005 %T The half-edge tree: a compact data structure for level-of-detail tetrahedral meshes %A Danovaro,E. %A De Floriani, Leila %A Magillo,P. %A Puppo,E. %A Sobrero,D. %A Sokolovsky,N. %K application; %K compact %K computational %K data %K detection; %K edge %K encoding; %K generation; %K geometry; %K half-edge %K iterative %K level-of-detail %K mesh %K meshes; %K methods; %K model; %K structure; %K structures; %K tetrahedral %K tree %K tree; %X We propose a new data structure for the compact encoding of a level-of detail (LOD) model of a three-dimensional scalar field based on unstructured tetrahedral meshes. Such data structure, called a half-edge tree (HET), is built through the iterative application of a half-edge collapse, i.e. by contracting an edge to one of its endpoints. We also show that selective refined meshes extracted from an HET contain on average about 34% and up to 75% less tetrahedra than those extracted from an LOD model built through a general edge collapse. %B Shape Modeling and Applications, 2005 International Conference %P 332 - 337 %8 2005/06// %G eng %R 10.1109/SMI.2005.47 %0 Report %D 2005 %T Headline Generation for Written and Broadcast News %A Zajic, David %A Dorr, Bonnie J %A Schwartz,Richard %K *INFORMATION RETRIEVAL %K *RADIO BROADCASTING %K AUTOMATIC %K AUTOMATIC SUMMARIZATION %K HEDGE TRIMMER %K INFORMATION SCIENCE %K RADIO COMMUNICATIONS %K STATISTICAL PROCESSES %K TEST AND EVALUATION %K WORDS(LANGUAGE) %X This technical report is an overview of work done on Headline Generation for written and broadcast news. The report covers HMM Hedge, a statistical approach based on the noisy channel model, Hedge Trimmer, a parse-and-trim approach using linguistically motivated trimming rules, and Topiary, a combination of Trimmer and Unsupervised Topic Discovery. Automatic evaluation of summaries using ROUGE and BLEU is described and used to evaluate the Headline Generation systems. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %8 2005/03// %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA454198 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Help! I'm Lost: User Frustration in Web Navigation (2003) %A Lazar,Jonathan %A Bessiere,Katie %A Ceaparu,Irina %A Robinson,John %A Shneiderman, Ben %K Technical Report %X Computers can be valuable tools, and networked resources via the Internet can be beneficial to many different populations and communities. Unfortunately, when people are unable to reach their task goals due to frustrating experiences, this can hinder the effectiveness of technology. This research summary provides information about the user frustration research that has been performed at the University of Maryland and Towson University. Causes of user frustration are discussed in this research summary, along with the surprising finding that nearly one-third to one-half of the time spent in front of the computer is wasted due to frustrating experiences. Furthermore, when interfaces are planned to be deceptive and confusing, this %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6508 %0 Conference Paper %B Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing %D 2005 %T The Hiero machine translation system: Extensions, evaluation, and analysis %A Chiang,D. %A Lopez,A. %A Madnani,N. %A Monz,C. %A Resnik, Philip %A Subotin,M. %B Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing %P 779 - 786 %8 2005/// %G eng %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International %D 2005 %T High Performance Communication between Parallel Programs %A Jae-Yong Lee %A Sussman, Alan %K Adaptive arrays %K Analytical models %K Chaotic communication %K Computational modeling %K Computer science %K Data analysis %K data distribution %K Educational institutions %K high performance communication %K image data analysis %K image resolution %K inter-program communication patterns %K InterComm %K Libraries %K Message passing %K parallel languages %K parallel libraries %K parallel programming %K parallel programs %K performance evaluation %K Wind %X We present algorithms for high performance communication between message-passing parallel programs, and evaluate the algorithms as implemented in InterComm. InterComm is a framework to couple parallel programs in the presence of complex data distributions within a coupled application. Multiple parallel libraries and languages may be used in the different programs of a single coupled application. The ability to couple such programs is required in many emerging application areas, such as complex simulations that model physical phenomena at multiple scales and resolutions, and image data analysis applications. We describe the new algorithms we have developed for computing inter-program communication patterns. We present experimental results showing the performance of various algorithmic tradeoffs, and also compare performance against an earlier system. %B Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International %I IEEE %P 177b- 177b - 177b- 177b %8 2005/04// %@ 0-7695-2312-9 %G eng %R 10.1109/IPDPS.2005.243 %0 Conference Paper %B Proceedings of the 2005 national conference on Digital government research %D 2005 %T Human-computer interaction themes in digital government: web site comprehension and statistics visualization %A Shneiderman, Ben %X Digital government applications often involve web sites to provide information for citizens and visitors from essential services such as passport application or motor vehicle registration to discretionary, but highly popular applications such as recreation and parks information. Another aspect of government web sites is the delivery of statistical reports with summary tables, aggregated data resources, and extensive raw data files. This review focuses on human-computer interaction themes to improve designs of web sites and statistics visualization, ??? specifically as they relate to digital government sites. It also addresses research methods that are appropriate for digital government interfaces. %B Proceedings of the 2005 national conference on Digital government research %S dg.o '05 %I Digital Government Society of North America %P 7 - 8 %8 2005/// %G eng %U http://dl.acm.org/citation.cfm?id=1065226.1065230 %0 Conference Paper %B CHI '05 extended abstracts on Human factors in computing systems %D 2005 %T "I hear the pattern": interactive sonification of geographical data patterns %A Zhao,Haixia %A Plaisant, Catherine %A Shneiderman, Ben %K auditory user interfaces %K information seeking %K sonification %K universal usability %K vision impairment %X Interactive sonification (non-speech sound) is a novel strategy to present the geographical distribution patterns of statistical data to vision impaired users. We discuss the design space with dimensions of interaction actions, data representation forms, input devices, navigation structures, and sound feedback encoding. Two interfaces were designed, one using a keyboard and another using a smooth surface touch tablet. A study with three blind users shows that they are able to perceive patterns of 5-category values on both familiar and unknown maps, and learn new map geography, in both interfaces. %B CHI '05 extended abstracts on Human factors in computing systems %S CHI EA '05 %I ACM %C New York, NY, USA %P 1905 - 1908 %8 2005/// %@ 1-59593-002-7 %G eng %U http://doi.acm.org/10.1145/1056808.1057052 %R 10.1145/1056808.1057052 %0 Book Section %B Handbook of Image and Video Processing (Second Edition)Handbook of Image and Video Processing (Second Edition) %D 2005 %T Image Sequence Stabilization, Mosaicking, and Superresolution %A Chellapa, Rama %A Srinivasan, S. %A Aggarwal,G. %A Veeraraghavan,A. %E Al Bovik %X A sequence of temporal images gathered from a single sensor adds a whole new dimension to two-dimensional (2D) image data. Availability of an image sequence permits the measurement of quantities such as subpixel intensities, camera motion and depth, and detection and tracking of moving objects. In turn, the processing of image sequences necessitates the development of sophisticated techniques to extract this information. With the recent availability of powerful yet inexpensive computers, data storage systems, and image acquisition devices, image sequence analysis has transitioned from an esoteric research domain to a practical area with significant commercial interest. %B Handbook of Image and Video Processing (Second Edition)Handbook of Image and Video Processing (Second Edition) %I Academic Press %C Burlington %P 309-VII - 309-VII %8 2005/// %@ 978-0-12-119792-6 %G eng %U http://www.sciencedirect.com/science/article/pii/B9780121197926500826 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Image-Based Highly Interactive Web Mapping for Geo-Referenced Data Publishing (2002) %A Zhao,Haixia %A Shneiderman, Ben %K Technical Report %X This paper describes an image-based technique that enables highly interactive Web choropleth maps for geo-referenced data publishing and visual exploration. Geographic knowledge is encoded into raster images and delivered to the client, instead of in vector formats. Differing from traditional raster-imagebased approaches that are static and allow very little user interaction, it allows varieties of sub-second fine-grained interface controls such as dynamic query, dynamic classification, geographic object data identification, user setting adjusting, as well as turning on/off layers, panning and zooming, with no or minimum server support. Compared to Web GIS approaches that are based on vector geographic data, this technique has the features of short initial download time, near-constant performance scalability for larger numbers of geographic objects, and download-map-segment-only-when-necessary which potentially reduces the overall data transfer over the network. As a result, it accommodates general public users with slow modem network connections and low-end machines, as well as users with fast T-1 connections and fast machines. The client-side (browser) is implemented as light-weight Java applets. YMap, an easy-to-use, user-task-oriented highly interactive mapping tool prototype for visual georeferenced data exploration is implemented using this technique. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6498 %0 Report %D 2005 %T Immediate Usability: A Case Study of Public Access Design for a Community Photo Library (2003) %A Kules,Bill %A Kang,Hyunmo %A Plaisant, Catherine %A Rose,Anne %A Shneiderman, Ben %K Technical Report %X This paper describes a novel instantiation of a digital photo library in a public access system. It demonstrates how designers can utilize characteristics of a target user community (social constraints, trust, and a lack of anonymity) to provide capabilities that would be impractical in other types of public access systems. It also presents a compact set of design principles and guidelines for ensuring the immediate usability of public access information systems. These principles and guidelines were derived from our experience developing PhotoFinder Kiosk, a community photo library. Attendees of a major HCI conference (CHI 2001 Conference on Human Factors in Computing Systems) successfully used the tool to browse and annotate collections of photographs spanning 20 years of HCI-related conferences, producing a richly annotated photo history of the field of human-computer interaction. Observations and log data were used to evaluate the tool and develop the guidelines. They provide specific guidance for practitioners, as well as a useful framework for additional research in public access interfaces. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6504 %0 Report %D 2005 %T Immediate Usability: Kiosk design principles from the CHI 2001 Photo Library (2001) %A Kules,Bill %A Kang,Hyunmo %A Plaisant, Catherine %A Rose,Anne %A Shneiderman, Ben %K Technical Report %X This paper describes a novel set of design principles and guidelines for ensuring the immediate usability of public access systems. These principles and guidelines were formulated while developing PhotoFinder Kiosk, a community photo library. Attendees of CHI 2001 successfully used the tool to browse and annotate collections of photographs spanning 20 years of CHI and related conferences, producing a richly annotated photo history of the field of human-computer interaction. We used observations and log data to evaluate the tool and refine the guidelines. They provide specific guidance for practitioners, as well as a useful framework for additional research in public access interfaces. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6488 %0 Report %D 2005 %T Improving Accessibility and Usability of Geo-referenced Statistical Data (2003) %A Zhao,Haixia %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X Several technology breakthroughs are needed to achieve the goals of universal accessibility and usability. These goals are especially challenging in the case of geo-referenced statistical data that many U.S. government agencies supply. We present technical and user-interface design challenges in accommodating users with low-end technology (slow network connection and low-end machine) and users who are blind or vision-impaired. Our solutions are presented and future work is discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6502 %0 Report %D 2005 %T Improving Web-based Civic Information Access: A Case Study of the 50 US States (2002) %A Ceaparu,Irina %A Shneiderman, Ben %K Technical Report %X An analysis of the home pages of all fifty U. S. states reveals great variety in key design features that influence efficacy. Some states had excessively large byte counts that would slow users connected by commonly-used 56K modems. Many web sites had low numbers of or poorly organized links that would make it hard for citizens to find what they were interested in. Features such as search boxes, privacy policies, online help, or contact information need to be added by several states. Our analysis concludes with ten recommendations and finds many further opportunities for individual states to improve their websites. However still greater benefits will come through collaboration among the states that would lead to consistency, appropriate tagging, and common tools. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6494 %0 Journal Article %J Advances in Multimedia Information Processing-PCM 2004 %D 2005 %T Indexing issues in supporting similarity searching %A Samet, Hanan %X Indexing issues that arise in the support of similarity searching are presented. This includes a discussion of the curse of dimensionality, as well as multidimensional indexing, distance-based indexing, dimension reduction, and embedding methods. %B Advances in Multimedia Information Processing-PCM 2004 %P 463 - 470 %8 2005/// %G eng %R 10.1007/978-3-540-30542-2_57 %0 Conference Paper %D 2005 %T Information and Knowledge Modeling for Computer Supported Micro Electro-Mechanical Systems Design and Development %A Zha,X. F. %A Sriram,R.D. %A Gupta,S.K. %X In this paper, we present a preliminary research effort towardsan effective computer support environment for the design and development of micro-electro-mechanical systems (MEMS). We first identify the characteristics of MEMS product design and development processes and examine the state-of-the-art of MEMS Computer-aided Design (CAD) and simulation systems. We then propose a function-(environment-effect)-behavior- (principle-state)-form (FEEBPSF) framework based on the NIST core product model and its extensions for modeling MEMS products, and apply the OESM (open embedded system model) developed to model information and knowledge for embedded MEMS design and development. Moreover, in order to tackle the knowledge-intensive tasks of design and development for MEMS products, we develop a general and flexible knowledge repository, called KR-MEMS, based on the FEEBPSF framework and the Unified Modeling Language (UML)/ Extensible Markup Language (XML) model, and integrate KR-MEMS into a web-based MEMS design support system. Throughout the paper, a micro fluidic dispensing device (a biomedical device for drug-delivery) is used as an example to illustrate the concepts, models, and knowledge bases necessary to support the MEMS product design and development. %8 2005/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.160.7822&rep=rep1&type=pdf %0 Journal Article %J Web Semantics: Science, Services and Agents on the World Wide Web %D 2005 %T Information gathering during planning for Web Service composition %A Kuter,Ugur %A Sirin,Evren %A Parsia,Bijan %A Nau, Dana S. %A Hendler,James %K HTN planning %K Information gathering %K Web Service composition %X Hierarchical task network (HTN) based planning techniques have been applied to the problem of composing Web Services, especially when described using the OWL-S service ontologies. Many of the existing Web Services are either exclusively information providing or crucially depend on information-providing services. Thus, many interesting service compositions involve collecting information either during execution or during the composition process itself. In this paper, we focus on the latter issue. In particular, we present ENQUIRER , an HTN-planning algorithm designed for planning domains in which the information about the initial state of the world may not be complete, but it is discoverable through plan-time information-gathering queries. We have shown that ENQUIRER is sound and complete, and derived several mathematical relationships among the amount of available information, the likelihood of the planner finding a plan, and the quality of the plan found. We have performed experimental tests that confirmed our theoretical results and that demonstrated how ENQUIRER can be used for Web Service composition. %B Web Semantics: Science, Services and Agents on the World Wide Web %V 3 %P 183 - 205 %8 2005/10// %@ 1570-8268 %G eng %U http://www.sciencedirect.com/science/article/pii/S1570826805000168 %N 2–3 %R 10.1016/j.websem.2005.07.001 %0 Conference Paper %B Integrated Formal Methods %D 2005 %T An integrated framework for scenarios and state machines %A Sengupta,B. %A Cleaveland, Rance %B Integrated Formal Methods %P 366 - 385 %8 2005/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Interactive Color Mosaic and Dendogram Displays for Signal/Noise Optimization in Microarray Data Analysis (2003) %A Seo,Jinwook %A Bakay,Marina %A Zhao,Po %A Chen,Yi-Wen %A Clarkson,Priscilla %A Shneiderman, Ben %A Hoffman,Eric P %K Technical Report %X Data analysis and visualization is strongly influenced by noise and noise filters. There are multiple sources of oisein microarray data analysis, but signal/noise ratios are rarely optimized, or even considered. Here, we report a noise analysis of a novel 13 million oligonucleotide dataset - 25 human U133A (~500,000 features) profiles of patient muscle biposies. We use our recently described interactive visualization tool, the Hierarchical Clustering Explorer (HCE) to systemically address the effect of different noise filters on resolution of arrays into orrectbiological groups (unsupervised clustering into three patient groups of known diagnosis). We varied probe set interpretation methods (MAS 5.0, RMA), resent callfilters, and clustering linkage methods, and investigated the results in HCE. HCE interactive features enabled us to quickly see the impact of these three variables. Dendrogram displays showed the clustering results systematically, and color mosaic displays provided a visual support for the results. We show that each of these three variables has a strong effect on unsupervised clustering. For this dataset, the strength of the biological variable was maximized, and noise minimized, using MAS 5.0, 10% present call filter, and Average Group Linkage. We propose a general method of using interactive tools to identify the optimal signal/noise balance or the optimal combination of these three variables to maximize the effect of the desired biological variable on data interpretation. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6506 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Interactive Exploration of Multidimensional Microarray Data: Scatterplot Ordering, Gene Ontology Browser, and Profile Search (2003) %A Seo,Jinwook %A Shneiderman, Ben %K Technical Report %X Multidimensional data sets are common in many research areas, including microarray experiment data sets. Genome researchers are using cluster analysis to find meaningful groups in microarray data. However, the high dimensionality of the data sets hinders users from finding interesting patterns, clusters, and outliers. Determining the biological significance of such features remains problematic due to the difficulties of integrating biological knowledge. In addition, it is not efficient to perform a cluster analysis over the whole data set in cases where researchers know the approximate temporal pattern of the gene expression that they are seeking.

To address these problems, we add three new features to the Hierarchical Clustering Explorer (HCE): (1) scatterplot ordering methods so that all 2D projections of a high dimensional data set can be ordered according to relevant criteria, (2) a gene ontology browser, coupled with clustering results so that known gene functions within a cluster can be easily studied, (3) a profile search so that genes with a certain temporal pattern can be easily identified. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6528 %0 Journal Article %J Proceedings of SPIE %D 2005 %T Interactive pattern search in time series %A Buono,Paolo %A Aris,Aleks %A Plaisant, Catherine %A Khella,Amir %A Shneiderman, Ben %X The need for pattern discovery in long time series data led researchers to develop algorithms for similarity search. Most of the literature about time series focuses on algorithms that index time series and bring the data into the main storage, thus providing fast information retrieval on large time series. This paper reviews the state of the art in visualizing time series, and focuses on techniques that enable users to visually and interactively query time series. Then, it presents TimeSearcher 2, a tool that enables users to explore multidimensional data using synchronized tables and graphs with overview+detail, filter the time series data to reduce the scope of the search, select an existing pattern to find similar occurrences, and interactively adjust similarity parameters to narrow the result set. This tool is an extension of previous work, TimeSearcher 1, which uses graphical timeboxes to interactively query time series data. %B Proceedings of SPIE %V 5669 %P 175 - 186 %8 2005/03/11/ %@ 0277786X %G eng %U http://spiedigitallibrary.org/proceedings/resource/2/psisdg/5669/1/175_1?isAuthorized=no %N 1 %R doi:10.1117/12.587537 %0 Report %D 2005 %T Interactive Pattern Search in Time Series (2004) %A Buono,Paolo %A Aris,Aleks %A Plaisant, Catherine %A Khella,Amir %A Shneiderman, Ben %K Technical Report %X The need for pattern discovery in long time series data led researchers to develop algorithms for similarity search. Most of the literature about time series focuses on algorithms that index time series and bring the data into the main storage, thus providing fast information retrieval on large time series. This paper reviews the state of the art in visualizing time series, and focuses on techniques that enable users to interactively query time series. Then it presents TimeSearcher 2, a tool that enables users to explore multidimensional data using coordinated tables and graphs with overview+detail, filter the time series data to reduce the scope of the search, select an existing pattern to find similar occurrences, and interactively adjust similarity parameters to narrow the result set. This tool is an extension of previous work, TimeSearcher 1, which uses graphical timeboxes to interactively query time series data. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6519 %0 Journal Article %J IEEE Multimedia %D 2005 %T Interactive sonification of choropleth maps %A Zhao,Haixia %A Smith,B. K %A Norman,K. %A Plaisant, Catherine %A Shneiderman, Ben %K audio signal processing %K audio user interfaces %K Auditory (non-speech) feedback %K auditory information %K cartography %K choropleth maps %K data collections %K decision making %K Evaluation %K Feedback %K georeferenced data %K Guidelines %K handicapped aids %K Hardware %K HUMANS %K information resources %K interaction style %K Interactive sonification %K interactive systems %K Navigation %K nonspeech audio %K problem solving %K Problem-solving %K sound %K universal usability %K US Government %K User interfaces %K vision impairments %K World Wide Web %X Auditory information is an important channel for the visually impaired. Effective sonification (the use of non-speech audio to convey information) promotes equal working opportunities for people with vision impairments by helping them explore data collections for problem solving and decision making. Interactive sonification systems can make georeferenced data accessible to people with vision impairments. The authors compare methods for using sound to encode georeferenced data patterns and for navigating maps. %B IEEE Multimedia %V 12 %P 26 - 35 %8 2005/06//April %@ 1070-986X %G eng %N 2 %R 10.1109/MMUL.2005.28 %0 Report %D 2005 %T Inventing Discovery Tools: Combining Information Visualization with Data Mining (2001) %A Shneiderman, Ben %K Technical Report %X The growing use of information visualization tools and data mining algorithms stems from two separate lines of research. Information visualization researchers believe in the importance of giving users an overview and insight into the data distributions, while data mining researchers believe that statistical algorithms and machine learning can be relied on to find the interesting patterns. This paper discusses two issues that influence design of discovery tools: statistical algorithms vs. visual data presentation, and hypothesis testing vs. exploratory data analysis. I claim that a combined approach could lead to novel discovery tools that preserve user control, enable more effective exploration, and promote responsibility. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6484 %0 Conference Paper %B Proceedings of the 7th international ACM SIGACCESS conference on Computers and accessibility %D 2005 %T iSonic: interactive sonification for non-visual data exploration %A Zhao,Haixia %A Plaisant, Catherine %A Shneiderman, Ben %K auditory user interfaces %K information seeking %K sonification %K universal usability %K vision impairment %X iSonic is an interactive sonification tool for vision impaired users to explore geo-referenced statistical data, such as population or crime rates by geographical regions. Users use a keyboard or a smooth surface touchpad to interact with coordinated map and table views of the data. The integrated use of musical sounds and speech allows users to grasp the overall data trends and to explore the data to get more details. Scenarios of use are described. %B Proceedings of the 7th international ACM SIGACCESS conference on Computers and accessibility %S Assets '05 %I ACM %C New York, NY, USA %P 194 - 195 %8 2005/// %@ 1-59593-159-7 %G eng %U http://doi.acm.org/10.1145/1090785.1090826 %R 10.1145/1090785.1090826 %0 Journal Article %J Johns Hopkins University, Tech. Rep %D 2005 %T Johns Hopkins summer workshop final report on parsing and spoken structural event detection %A Harper,MP %A Dorr, Bonnie J %A Hale,J. %A Roark,B. %A Shafran,I. %A Lease,M. %A Liu,Y. %A Snover,M. %A Yung,L. %A Krasnyanskaya,A. %X This report describes research conducted by the Parsing and Spoken Structural Event Detection(PaSSED) team as part of the 2005 Johns Hopkins Summer Workshop on Language Engineering. This project investigated the interaction between parsing and the detection of structural metadata in conversational speech, including sentence boundaries, edits (the reparandum portion of speech repairs), and fillers. In terms of parsing, we explored alternative methods of exploiting metadata information in parsing models and measured how varying accuracy in transcription and metadata information affects parsing accuracy. In the other direction, we similarly considered how syntactic and prosodic knowledge could be leveraged in metadata detection, measuring how this knowledge impacts metadata detection accuracy. As part of this work, we investigated metrics for evaluating parse accuracy in the presence of tran- scription and metadata detection errors, and we report on our experience using these metrics with several parsers and across varying experimental conditions. A range of methods for handling edits during pars- ing were evaluated in this research (excision, addition of markups to the input string, and grammar modification). We also developed a ToBI (a prosodic structure annotation scheme [SBP+92]) prosodic event classifier and describe its evaluation. Finally, we present methods for effective n-best sentence boundary candidate generation and reranking using syntactic, prosodic, and other features. These stud- ies are complemented by a second set of reranking investigations wherein we optimize sentence boundary detection explicitly to improve parse accuracy. The PaSSED project has: • investigated various techniques to enhance parsing of speech given metadata detection on conver- sational speech; • defined metrics for evaluating speech parsing accuracy, implemented them in the publically avail- able SParseval software package, and evaluated them under a wide variety of conditions; • recast SU detection as an n-best reranking problem with a relatively small n. Using this approach, we demonstrated significant improvements over a very strong baseline SU detection system. • reported on the interaction between parsing and metadata detection and their synergy; • fostered new collaborations and identified a number of interesting avenues for future work. %B Johns Hopkins University, Tech. Rep %8 2005/// %G eng %0 Conference Paper %B Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on %D 2005 %T Joint uplink and downlink optimization for video conferencing over wireless LAN %A Su,Guan-Ming %A Han,Zhu %A M. Wu %A Liu,K. J.R %K 802.11 %K conferencing %K conferencing; %K distortion; %K end-to-end %K expected %K IEEE %K joint %K LAN; %K maximal %K minimisation; %K modes; %K multiple %K networks; %K optimization; %K pairs; %K PHY %K PSNR; %K real-time %K teleconferencing; %K time; %K transmission %K uplink-downlink %K video %K wireless %X A real-time video conferencing framework is proposed for multiple conferencing pairs by jointly considering the uplink and downlink conditions within IEEE 802.11 networks. We formulate this system so as to minimize the maximal end-to-end expected distortion received by all users by selecting the PHY modes and transmission time. Compared with the strategy of individually optimizing uplink and downlink, the proposed framework outperforms by 3.67-8.65 dB for the average received PSNR among all users. %B Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on %V 2 %P ii/1101 - ii/1104 Vol. 2 - ii/1101 - ii/1104 Vol. 2 %8 2005/03// %G eng %R 10.1109/ICASSP.2005.1415601 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Knowledge Discovery in High Dimensional Data: Case Studies and a User Survey for an Information Visualization Tool %A Seo,Jinwook %A Shneiderman, Ben %K Information Visualization %X Knowledge discovery in high dimensional data is a challenging enterprise, but new visual analytic tools appear to offer users remarkable powers if they are ready to learn new concepts and interfaces. Our 3-year effort to develop versions of the Hierarchical Clustering Explorer (HCE) began with building an interactive tool for exploring clustering results. It expanded, based on user needs, to include other potent analytic and visualization tools for multivariate data, especially the rank-by-feature framework. Our own successes using HCE provided some testimonial evidence of its utility, but we felt it necessary to get beyond our subjective impressions. This paper presents an evaluation of the Hierarchical Clustering Explorer (HCE) using three case studies and an email user survey (n=57) to focus on skill acquisition with the novel concepts and interface for the rank-by-feature framework. Knowledgeable and motivated users in diverse fields provided multiple perspectives that refined our understanding of strengths and weaknesses. A user survey confirmed the benefits of HCE, but gave less guidance about improvements. Both evaluations suggested improved training methods. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6566 %0 Book Section %B From Integrated Publication and Information Systems to Information and Knowledge EnvironmentsFrom Integrated Publication and Information Systems to Information and Knowledge Environments %D 2005 %T A Knowledge Integration Framework for Information Visualization %A Seo,Jinwook %A Shneiderman, Ben %E Hemmje,Matthias %E Niederée,Claudia %E Risse,Thomas %X Users can better understand complex data sets by combining insights from multiple coordinated visual displays that include relevant domain knowledge. When dealing with multidimensional data and clustering results, the most familiar displays and comprehensible are 1- and 2-dimensional projections (histograms, and scatterplots). Other easily understood displays of domain knowledge are tabular and hierarchical information for the same or related data sets. The novel parallel coordinates view [6] powered by a direct-manipulation search, offers strong advantages, but requires some training for most users. We provide a review of related work in the area of information visualization, and introduce new tools and interaction examples on how to incorporate users’ domain knowledge for understanding clustering results. Our examples present hierarchical clustering of gene expression data, coordinated with a parallel coordinates view and with the gene annotation and gene ontology. %B From Integrated Publication and Information Systems to Information and Knowledge EnvironmentsFrom Integrated Publication and Information Systems to Information and Knowledge Environments %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3379 %P 207 - 220 %8 2005/// %@ 978-3-540-24551-3 %G eng %U http://dx.doi.org/10.1007/978-3-540-31842-2_21 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T A Knowledge Integration Framework for Information Visualization (2004) %A Seo,Jinwook %A Shneiderman, Ben %K Technical Report %X Users can better understand complex data sets by combining insights from multiple coordinated visual displays that include relevant domain knowl-edge. When dealing with multidimensional data and clustering results, the most familiar displays and comprehensible are 1- and 2-dimensional projections (his-tograms, and scatterplots). Other easily understood displays of domain knowl-edge are tabular and hierarchical information for the same or related data sets. The novel parallel coordinates view [6] powered by a direct-manipulation search, offers strong advantages, but requires some training for most users. We provide a review of related work in the area of information visualization, and introduce new tools and interaction examples on how to incorporate usersdo-main knowledge for understanding clustering results. Our examples present hi-erarchical clustering of gene expression data, coordinated with a parallel coor-dinates view and with the gene annotation and gene ontology. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6525 %0 Conference Paper %B Proceedings of the 14th ACM international conference on Information and knowledge management %D 2005 %T Leonardo's laptop: human needs and the new computing technologies %A Shneiderman, Ben %X The old computing was about what computers could do; the new computing is about what people can do.To accelerate the shift from the old to the new computing designers need to:reduce computer user frustration. Recent studies show 46% of time is lost to crashes, confusing instructions, navigation problems, etc. Public pressure for change could promote design improvements and increase reliability, thereby dramatically enhancing user experiences.promote universal usability. Interfaces must be tailorable to a wide range of hardware, software, and networks, and users. When broad services such as voting, healthcare, and education are envisioned, the challenge to designers is substantial.envision a future in which human needs more directly shape technology evolution. Four circles of human relationships and four human activities map out the human needs for mobility, ubiquity, creativity, and community. The World Wide Med and million-person communities will be accessible through desktop, palmtop and fingertip devices to support e-learning, e-business, e-healthcare, and e-government.Leonardo da Vinci could help as an inspirational muse for the new computing. His example could push designers to improve quality through scientific study and more elegant visual design. Leonardo's example can guide us to the new computing, which emphasizes empowerment, creativity, and collaboration. Information visualization and personal photo interfaces will be shown: PhotoMesa (www.cs.umd.edu/hcil/photomesa) and PhotoFinder (www.cs.umd.edu/hcil/photolib).For more: http://mitpress.mit.edu/leonardoslaptop and http://www.cs.umd.edu/hcil/newcomputing. %B Proceedings of the 14th ACM international conference on Information and knowledge management %S CIKM '05 %I ACM %C New York, NY, USA %P 1 - 1 %8 2005/// %@ 1-59593-140-6 %G eng %U http://doi.acm.org/10.1145/1099554.1099555 %R 10.1145/1099554.1099555 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T The limits of speech recognition: Understanding acoustic memory and appreciating prosody (2000) %A Shneiderman, Ben %K Technical Report %X Human-human relationships are rarely a good model for the design of effective user interfaces. Spoken language is effective for human-human interaction (HHI), but it often has severe limitations when applied to human-computer interaction (HCI). Speech is slow for presenting information, it is difficult to review or edit, and it interferes with other cognitive tasks. However speech has proven to be useful for store-and-forward messages, alerts in busy environments, and input-output for blind or motor-impaired users. Speech recognition for control is helpful for hands-busy, eyes-busy, mobilityrequired, or hostile environments and it shows promise for use in telephone-based services. Dictation input is increasingly accurate, but adoption outside the disabled users community has been slow compared to visual interfaces. Obvious physical problems include fatigue from speaking continuously and the disruption in an office filled with people speaking.

By understanding the cognitive processes surrounding human acoustic memory and processing, interface designers may be able to integrate speech more effectively and guide users more successfully. Then by appreciating the differences between HHI and HCI designers may be able to choose appropriate applications for human use of speech with computers. The key distinction may be the rich emotional content conveyed by prosody -- the pacing, intonation, and amplitude in spoken language. Prosody is potent for HHI, but may be disruptive for HCI. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6532 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Listening to Maps: User Evaluation of Interactive Sonifications of Geo-Referenced Data (2004) %A Zhao,Haixia %A Smith,Benjamin K %A Norman,Kent L %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X In this paper, we summarize the Auditory Information Seeking Principle (AISP) (gist, navigate, filter, and details-ondemand). To improve blind access to geo-referenced statistical data, we developed several interactive sonifications, adhering to the above AISP. Two user studies are presented. In the first user study with nine sighted subjects, a preliminary map design is compared with an enhanced table design. The study shows subjects can recognize geographic data distribution patterns on a real map with 51 geographic regions, in both designs. The map-based design was strongly preferred. The study also shows evidence that AISP conforms to people information seeking strategies. Based on the observations from the first user study, a second user study was conducted with forty-eight sighted subjects comparing four map designs. The effects of using sound to encode vertical geographic positions and two map navigation methods were compared. The result is presented and future work is discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6518 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Listening to Maps: User Evaluation of Interactive Sonifications of Geo-Referenced Data %A Zhao,Haixia %A Smith,Benjamin K %A Norman,Kent L %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X In this paper, we summarize the Auditory Information Seeking Principle (AISP) (gist, navigate, filter, and details-ondemand). To improve blind access to geo-referenced statistical data, we developed several interactive sonifications, adhering to the above AISP. Two user studies are presented. In the first user study with nine sighted subjects, a preliminary map design is compared with an enhanced table design. The study shows subjects can recognize geographic data distribution patterns on a real map with 51 geographic regions, in both designs. The map-based design was strongly preferred. The study also shows evidence that AISP conforms to people information seeking strategies. Based on the observations from the first user study, a second user study was conducted with forty-eight sighted subjects comparing four map designs. The effects of using sound to encode vertical geographic positions and two map navigation methods were compared. The result is presented and future work is discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6518 %0 Conference Paper %B Proceedings of the 27th international conference on Software engineering - ICSE '05 %D 2005 %T Main effects screening %A Yilmaz,Cemal %A Krishna,Arvind S. %A Memon, Atif M. %A Porter, Adam %A Schmidt,Douglas C. %A Gokhale,Aniruddha %A Natarajan,Balachandran %B Proceedings of the 27th international conference on Software engineering - ICSE '05 %C St. Louis, MO, USA %P 293 - 293 %8 2005/// %G eng %U http://dl.acm.org/citation.cfm?id=1062515 %R 10.1145/1062455.1062515 %0 Conference Paper %B Proceedings of the 27th international conference on Software engineering %D 2005 %T Main effects screening: a distributed continuous quality assurance process for monitoring performance degradation in evolving software systems %A Yilmaz,Cemal %A Krishna,Arvind S. %A Memon, Atif M. %A Porter, Adam %A Schmidt,Douglas C. %A Gokhale,Aniruddha %A Natarajan,Balachandran %K design of experiment theory %K distributed continuous quality assurance %K performance-oriented regression testing %X Developers of highly configurable performance-intensive software systems often use a type of in-house performance-oriented "regression testing" to ensure that their modifications have not adversely affected their software's performance across its large configuration space. Unfortunately, time and resource constraints often limit developers to in-house testing of a small number of configurations and unreliable extrapolation from these results to the entire configuration space, which allows many performance bottlenecks and sources of QoS degradation to escape detection until systems are fielded. To improve performance assessment of evolving systems across large configuration spaces, we have developed a distributed continuous quality assurance (DCQA) process called main effects screening that uses in-the-field resources to execute formally designed experiments to help reduce the configuration space, thereby allowing developers to perform more targeted in-house QA. We have evaluated this process via several feasibility studies on several large, widely-used performance-intensive software systems. Our results indicate that main effects screening can detect key sources of performance degradation in large-scale systems with significantly less effort than conventional techniques. %B Proceedings of the 27th international conference on Software engineering %S ICSE '05 %I ACM %C New York, NY, USA %P 293 - 302 %8 2005/// %@ 1-58113-963-2 %G eng %U http://doi.acm.org/10.1145/1062455.1062515 %R 10.1145/1062455.1062515 %0 Conference Paper %B Proceedings of the 2005 conference on Designing for User eXperience %D 2005 %T Maintaining concentration to achieve task completion %A Shneiderman, Ben %A Bederson, Benjamin B. %K avoid interruption %K calm %K flow %K focus %K goal attainment %K interface design %K maintain concentration %K situation awareness %K task completion %X When faced with a challenging goal, knowledge workers need to concentrate on their tasks so that they move forward toward completion. Since frustrations, distractions, and interruptions can interfere with their smooth progress, design strategies should enable users to maintain concentration. This paper promotes awareness of this issue, reviews related work, and suggests three initial strategies: Reduce short-term and working memory load, provide information abundant interfaces, and increase automaticity. %B Proceedings of the 2005 conference on Designing for User eXperience %S DUX '05 %I AIGA: American Institute of Graphic Arts %C New York, NY, USA %8 2005/// %@ 1-59593-250-X %G eng %U http://dl.acm.org/citation.cfm?id=1138235.1138246 %0 Conference Paper %B Proceedings of the 2005 conference on Designing for User eXperience %D 2005 %T Maintaining concentration to achieve task completion %A Shneiderman, Ben %A Bederson, Benjamin B. %B Proceedings of the 2005 conference on Designing for User eXperience %P 9 - 9 %8 2005/// %G eng %0 Conference Paper %B Data Engineering, International Conference on %D 2005 %T Maintaining Implicated Statistics in Constrained Environments %A Sismanis,Yannis %A Roussopoulos, Nick %X Aggregated information regarding implicated entities is critical for online applications like network management, traffic characterization or identifying patters of resource consumption. Recently there has been a flurry of research for online aggregation on streams (like quantiles, hot items, hierarchical heavy hitters) but surprizingly the problem of summarizing implicated information in stream data has received no attention. As an example, consider an IP-network and the implication source → destination. Flash crowds, — such as those that follow recent sport events (like the olympics) or seek information regarding catastrophic events — or denial of service attacks direct a large volume of traffic from a huge number of sources to a very small number of destinations. In this paper we present novel randomized algorithms for monitoring such implications with constraints in both memory and processing power for environments like network routers. Our experiments demonstrate several factors of improvements over straightforward approaches. %B Data Engineering, International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %P 730 - 741 %8 2005/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/ICDE.2005.84 %0 Report %D 2005 %T Making Computer and Internet Usability a Priority (2000) %A Kreitzberg,Charles B %A Shneiderman, Ben %K Technical Report %X As usability professionals, we are all too aware of the productivity losses, frustration, and lost business that results from poorly designed user interfaces. And we are uncomfortable with the risks created by poorly designed computer systems in life-critical applications such as air travel, medical care, and military applications. Yet despite the common sense of our approach, we still find it difficult to convince the technical and managerial communities that usability is a critical business parameter. While senior managers may support the concept of usability, project mangers and developers, coping with too tight schedules, often see it as a nicety that can be eliminated. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6527 %0 Conference Paper %B Computer Vision for Interactive and Intelligent Environment, 2005 %D 2005 %T Markerless Motion Capture using Multiple Cameras %A Sundaresan, A. %A Chellapa, Rama %X Motion capture has important applications in different areas such as biomechanics, computer animation, and human-computer interaction. Current motion capture methods use passive markers that are attached to different body parts of the subject and are therefore intrusive in nature. In applications such as pathological human movement analysis, these markers may introduce an unknown artifact in the motion, and are, in general, cumbersome. We present computer vision based methods for performing markerless human motion capture. We model the human body as a set of super-quadrics connected in an articulated structure and propose algorithms to estimate the parameters of the model from video sequences. We compute a volume data (voxel) representation from the images and combine bottom-up approach with top down approach guided by our knowledge of the model. We propose a tracking algorithm that uses this model to track human pose. The tracker uses an iterative framework akin to an Iterated Extended Kalman Filter to estimate articulated human motion using multiple cues that combine both spatial and temporal information in a novel manner. We provide preliminary results using data collected from 8-16 cameras. The emphasis of our work is on models and algorithms that are able to scale with respect to the requirement for accuracy. Our ultimate objective is to build an end-to-end system that can integrate the above mentioned components into a completely automated markerless motion capture system. %B Computer Vision for Interactive and Intelligent Environment, 2005 %P 15 - 26 %8 2005/11// %G eng %R 10.1109/CVIIE.2005.13 %0 Journal Article %J Concurrency and Computation: Practice and Experience %D 2005 %T MEAD: support for Real-Time Fault-Tolerant CORBA %A Narasimhan, P. %A Tudor Dumitras %A Paulos, A. M. %A Pertet, S. M. %A Reverte, C. F. %A Slember, J. G. %A Srivastava, D. %K CORBA %K Fault tolerance %K non-determinism %K predictability %K real-time %K recovery %K trade-offs %X The OMG's Real-Time CORBA (RT-CORBA) and Fault-Tolerant CORBA (FT-CORBA) specifications make it possible for today's CORBA implementations to exhibit either real-time or fault tolerance in isolation. While real-time requires a priori knowledge of the system's temporal operation, fault tolerance necessarily deals with faults that occur unexpectedly, and with possibly unpredictable fault recovery times. The MEAD (Middleware for Embedded Adaptive Dependability) system attempts to identify and to reconcile the conflicts between real-time and fault tolerance, in a resource-aware manner, for distributed CORBA applications. MEAD supports transparent yet tunable fault tolerance in real-time, proactive dependability, resource-aware system adaptation to crash, communication and timing faults with bounded fault detection and fault recovery. Copyright © 2005 John Wiley & Sons, Ltd. %B Concurrency and Computation: Practice and Experience %V 17 %P 1527 - 1545 %8 2005/// %@ 1532-0634 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/cpe.882/abstract %N 12 %0 Conference Paper %B Proceedings of the 5th ACM/IEEE-CS Joint Conference on Digital Libraries, 2005. JCDL '05 %D 2005 %T Meaningful presentations of photo libraries: rationale and applications of bi-level radial quantum layouts %A Kustanowitz,J. %A Shneiderman, Ben %K 1024 pixel %K 1200 pixel %K 1280 pixel %K 1310720 pixel %K 1600 pixel %K 1920000 pixel %K Application software %K bi-level radial quantum layouts %K Computer displays %K Computer science %K digital libraries %K Educational institutions %K Image retrieval %K Layout %K layout generation %K Lifting equipment %K linear strips %K Permission %K photo layouts %K photo library searching %K photo management %K Photography %K Quantum computing %K software libraries %K Strips %K two-dimensional grid %K User interfaces %K visual databases %K visual presentation %K zoomable three dimensional arrangements %X Searching photo libraries can be made more satisfying and successful if search results are presented in a way that allows users to gain an overview of the photo categories. Since photo layouts on computer displays are the primary way that users get an overview, we propose a novel approach to show more photos in meaningful groupings. Photo layouts can be linear strips, or zoomable three dimensional arrangements, but the most common form is the two-dimensional grid. This paper introduces a novel bi-level hierarchical layout with motivating examples. In a bilevel hierarchy, one region is designated for primary content - an image, text, or combination. Adjacent to that region, groups of photos are placed radially in an ordered fashion, such that the relationship of the single primary region to its many secondary regions is apparent. A compelling aspect is the interactive experience in which the layout is dynamically resized, allowing users to rapidly, incrementally, and reversibly alter the dimensions and content. It can accommodate hundreds of photos in dozens of regions, can be customized in a corner or center layout, and can scale from an element on a web page to a large poster size. On typical displays (1024 times 1280 or 1200 times 1600 pixels), bi-level radial quantum layouts can conveniently accommodate 2-20 regions with tens or hundreds of photos per region %B Proceedings of the 5th ACM/IEEE-CS Joint Conference on Digital Libraries, 2005. JCDL '05 %I IEEE %P 188 - 196 %8 2005/06/07/11 %@ 1-58113-876-8 %G eng %R 10.1145/1065385.1065431 %0 Conference Paper %B INFOCOM 2005. 24th Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings IEEE %D 2005 %T Measurement-based multipath multicast %A Guven,T. %A La,R.J. %A Shayman,M.A. %A Bhattacharjee, Bobby %K algorithm; %K allocation; %K application-layer %K approximation %K approximation; %K balance %K communication; %K Convergence %K convergence; %K intradomain %K load %K measurement-based %K methods; %K Multicast %K multipath %K network %K numerical %K of %K overlaying; %K perturbation %K processes; %K resource %K Routing %K routing; %K source; %K Stochastic %K techniques; %K Telecommunication %K theory; %K traffic; %X We propose a measurement-based routing algorithm to load balance intradomain traffic along multiple paths for multiple multicast sources. Multiple paths are established using application-layer overlaying. The proposed algorithm is able to converge under different network models, where each model reflects a different set of assumptions about the multicasting capabilities of the network. The algorithm is derived from simultaneous perturbation stochastic approximation and relies only on noisy estimates from measurements. Simulation results are presented to demonstrate the additional benefits obtained by incrementally increasing the multicasting capabilities. %B INFOCOM 2005. 24th Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings IEEE %V 4 %P 2803 - 2808 vol. 4 - 2803 - 2808 vol. 4 %8 2005/03// %G eng %R 10.1109/INFCOM.2005.1498566 %0 Report %D 2005 %T MediaFinder: An Interface for Dynamic Personal Media Management with Semantic Regions (2003) %A Kang,Hyunmo %A Shneiderman, Ben %K Technical Report %X Computer users deal with large amounts of personal media often face problems in managing and exploring it. This paper presents Semantic Regions, rectangular regions that enable users to specify their semantics or mental models, and the MediaFinder application, which uses Semantic Regions as the basis of a personal media management tool. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6499 %0 Patent %D 2005 %T Method and system for providing physical security in an area of interest %A Agrawala, Ashok K. %A Youssef,Moustafa Amin %A Shahamat,Layla %X A system for detecting the presence of an intruder in a protected area utilizes a received signal strength indicator (RSSI) value of signals broadcast from transmitting stations deployed in the protected area. The system includes monitoring points for receiving broadcast signals, measuring the RSSI values of the received signals, and transmitting the measured RSSI values to a security system server. The security system server analyzes the RSSI values, and initiates security measures when the physical security of the protected area is violated which is detected when the measured RSSI values deviate from a predetermined strength of the broadcast signals. The security system also has the ability to track objects in the protected area and keep track of their movement in real time and report such movement. The system may be based on a Wi-Fi infrastructure in the protected area. %V 10/916,545 %8 2005/03/10/ %G eng %U http://www.google.com/patents?id=VIaWAAAAEBAJ %0 Journal Article %J Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization %D 2005 %T A methodology for extrinsic evaluation of text summarization: Does ROUGE correlate %A Dorr, Bonnie J %A Monz,C. %A President,S. %A Schwartz,R. %A Zajic, David %X This paper demonstrates the usefulness of sum-maries in an extrinsic task of relevance judgment based on a new method for measuring agree- ment, Relevance-Prediction, which compares sub- jects’ judgments on summaries with their own judg- ments on full text documents. We demonstrate that, because this measure is more reliable than previ- ous gold-standard measures, we are able to make stronger statistical statements about the benefits of summarization. We found positive correlations be- tween ROUGE scores and two different summary types, where only weak or negative correlations were found using other agreement measures. How- ever, we show that ROUGE may be sensitive to the choice of summarization style. We discuss the im- portance of these results and the implications for fu- ture summarization evaluations. %B Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization %P 1 - 8 %8 2005/// %G eng %0 Journal Article %J Neural computation %D 2005 %T Mirror symmetric topographic maps can arise from activity-dependent synaptic changes %A Schulz,R. %A Reggia, James A. %B Neural computation %V 17 %P 1059 - 1083 %8 2005/// %G eng %N 5 %0 Conference Paper %B Proceedings of the 12th ACM conference on Computer and communications security %D 2005 %T Misbehaving TCP receivers can cause internet-wide congestion collapse %A Sherwood,Rob %A Bhattacharjee, Bobby %A Braud,Ryan %K Congestion control %K distributed denial of service %X An optimistic acknowledgment (opt-ack) is an acknowledgment sent by a misbehaving client for a data segment that it has not received. Whereas previous work has focused on opt-ack as a means to greedily improve end-to-end performance, we study opt-ack exclusively as a denial of service attack. Specifically, an attacker sends optimistic acknowledgments to many victims in parallel, thereby amplifying its effective bandwidth by a factor of 30 million (worst case). Thus, even a relatively modest attacker can totally saturate the paths from many victims back to the attacker. Worse, a distributed network of compromised machines ("zombies") attacking in parallel can exploit over-provisioning in the Internet to bring about wide-spread, sustained congestion collapse.We implement this attack both in simulation and in a wide-area network, and show it severity both in terms of number of packets and total traffic generated. We engineer and implement a novel solution that does not require client or network modifications allowing for practical deployment. Additionally, we demonstrate the solution's efficiency on a real network. %B Proceedings of the 12th ACM conference on Computer and communications security %S CCS '05 %I ACM %C New York, NY, USA %P 383 - 392 %8 2005/// %@ 1-59593-226-7 %G eng %U http://doi.acm.org/10.1145/1102120.1102170 %R 10.1145/1102120.1102170 %0 Conference Paper %B Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on %D 2005 %T Modeling image processing systems with homogeneous parameterized dataflow graphs %A Sen,Mainak %A Bhattacharyya, Shuvra S. %A Lv,Tiehan %A Wolf,W. %K data consumption %K data flow graphs %K data production %K dynamic dataflow graph %K graph edges %K homogeneous parameterized dataflow graphs %K IMAGE PROCESSING %K image processing system modeling %K iteration %K iterative methods %X We describe a new dataflow model called homogeneous parameterized dataflow (HPDF). This form of dynamic dataflow graph takes advantage of the fact that in a large number of image processing applications, data production and consumption rates, though dynamic, are equal across graph edges for any particular iteration, which leads to a homogeneous rate of actor execution, even though data production and consumption values are dynamic and vary across graph edges. We discuss existing dataflow models and formulate in detail the HPDF model. We develop examples of applications that are described naturally in terms of HPDF semantics and present experimental results that demonstrate the efficacy of the HPDF approach. %B Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on %V 5 %P v/133 - v/136 Vol. 5 - v/133 - v/136 Vol. 5 %8 2005/03// %G eng %R 10.1109/ICASSP.2005.1416258 %0 Conference Paper %B Proceedings of the 12th ACM conference on Computer and communications security %D 2005 %T Modeling insider attacks on group key-exchange protocols %A Katz, Jonathan %A Shin,Ji Sun %K group key exchange %K insider attacks %K universal composability %X Protocols for authenticated key exchange (AKE) allow parties within an insecure network to establish a common session key which can then be used to secure their future communication. It is fair to say that group AKE is currently less well understood than the case of two-party AKE; in particular, attacks by malicious insiders --- a concern specific to the group setting --- have so far been considered only in a relatively "ad-hoc" fashion. The main contribution of this work is to address this deficiency by providing a formal, comprehensive model and definition of security for group AKE which automatically encompasses insider attacks. We do so by defining an appropriate ideal functionality for group AKE within the universal composability (UC) framework. As a side benefit, any protocol secure with respect to our definition is secure even when run concurrently with other protocols, and the key generated by any such protocol may be used securely in any subsequent application.In addition to proposing this definition, we show that the resulting notion of security is strictly stronger than the one proposed by Bresson, et al. (termed "AKE-security"), and that our definition implies all previously-suggested notions of security against insider attacks. We also show a simple technique for converting any AKE-secure protocol into one secure with respect to our definition. %B Proceedings of the 12th ACM conference on Computer and communications security %S CCS '05 %I ACM %C New York, NY, USA %P 180 - 189 %8 2005/// %@ 1-59593-226-7 %G eng %U http://doi.acm.org/10.1145/1102120.1102146 %R 10.1145/1102120.1102146 %0 Conference Paper %D 2005 %T A motion closed world asumption %A Yaman,F. %A Nau, Dana S. %A V.S. Subrahmanian %X Yaman et. al. [Yaman et al., 2004] introduce “gotheories” to reason about moving objects. In this paper, we show that this logic often does not allow us to infer that an object is not present at a given place or region, even though common sense would dictate that this is a reasonable inference to make. We define a class of models of go-theories called coherent models. We use this concept to define a motion closed world assumption (MCWA) and de- velop a notion of MCWA-entailment. We show that checking if a go-theory has a coherent model is NP-complete. An in atom checks if a given ob- ject is present in a given region sometime in a given time interval. We provide sound and complete al- gorithms to check if a ground in literal (positive or negative in atom) can be inferred from a go- theory using the MCWA. In our experiments our algorithms answer such queries in less than 1 sec- ond when there are up to 1,000 go-atoms per object. %V 19 %P 621 - 621 %8 2005/// %G eng %U http://ijcai.org/papers/1227.pdf %0 Journal Article %J University of Marlyand %D 2005 %T Motivating annotation for personal digital photo libraries: Lowering barriers while raising incentives %A Kustanowitz,J. %A Shneiderman, Ben %X The value of personal digital photo libraries grows immensely when users invest effort to annotate their photos. Frameworksfor understanding annotation requirements could guide improved strategies that would motivate more users to invest the necessary effort. We propose one framework for annotation techniques along with the strengths and weaknesses of each one, and a second framework for target user groups and their motivations. Several applications are described that provide useful and information-rich representations, but which require good annotations, in the hope of providing incentives for high quality annotation. We describe how annotations make possible four novel presentations of personal photo collections: (1) Birthday Collage to show growth of a child over several years, (2) FamiliarFace to show family trees of photos, (3) Kaleidoscope to show photos of related people in an appealing tableau, and (4) TripPics to show photos from a sequential story such as a vacation trip. %B University of Marlyand %8 2005/// %G eng %0 Conference Paper %B Proceedings of the 13th annual ACM international workshop on Geographic information systems %D 2005 %T Multi-resolution out-of-core modeling of terrain and geological data %A Danovaro,Emanuele %A De Floriani, Leila %A Puppo,Enrico %A Samet, Hanan %K Multi-resolution %K out-of-core %K terrain models %X Multi-resolution is a useful tool for managing the complexity of huge terrain and geological data sets. Since encoding large data sets may easily exceed main memory capabilities, data structures and algorithms capable of efficiently working in external memory are needed. In our work, we aim at developing an out-of-core multi-resolution model dimension-independent, that can be used for both terrains, represented by Triangulated Irregular Networks(TINs), and 3D data, such as geological data, represented by tetrahedral meshes. We have based our approach on a general multi-resolution model, that we have proposed in our previous work, which supports the extraction of variable-resolution representations. As first step, we have developed, in a prototype simulation system, a large number of clustering techniques for the modifications in a multi-resolution model. Here, we describe such techniques, and analyze and evaluate them experimentally. The result of this investigation has led us to select a specific clustering approach as the basis for an efficient out-of-core data structure. %B Proceedings of the 13th annual ACM international workshop on Geographic information systems %S GIS '05 %I ACM %C New York, NY, USA %P 143 - 152 %8 2005/// %@ 1-59593-146-5 %G eng %U http://doi.acm.org/10.1145/1097064.1097085 %R 10.1145/1097064.1097085 %0 Conference Paper %B Proceedings of the 32nd ACM SIGPLAN-SIGACT symposium on Principles of programming languages %D 2005 %T Mutatis mutandis: safe and predictable dynamic software updating %A Stoyle,Gareth %A Hicks, Michael W. %A Bierman,Gavin %A Sewell,Peter %A Neamtiu,Iulian %K capability %K dynamic software updating %K proteus %K Type inference %K updateability analysis %X Dynamic software updates can be used to fix bugs or add features to a running program without downtime. Essential for some applications and convenient for others, low-level dynamic updating has been used for many years. Perhaps surprisingly, there is little high-level understanding or language support to help programmers write dynamic updates effectively.To bridge this gap, we present Proteus, a core calculus for dynamic software updating in C-like languages that is flexible, safe, and predictable. Proteus supports dynamic updates to functions (even active ones), to named types and to data, allowing on-line evolution to match source-code evolution as we have observed it in practice. We ensure updates are type-safe by checking for a property we call "con-freeness" for updated types t at the point of update. This means that non-updated code will not use t concretely beyond that point (concrete usages are via explicit coercions) and thus t's representation can safely change. We show how con-freeness can be enforced dynamically for a particular program state. We additionally define a novel and efficient static updateability analysis to establish con-freeness statically, and can thus automatically infer program points at which all future (well-formed) updates will be type-safe. We have implemented our analysis for C and tested it on several well-known programs. %B Proceedings of the 32nd ACM SIGPLAN-SIGACT symposium on Principles of programming languages %S POPL '05 %I ACM %C New York, NY, USA %P 183 - 194 %8 2005/// %@ 1-58113-830-X %G eng %U http://doi.acm.org/10.1145/1040305.1040321 %R 10.1145/1040305.1040321 %0 Journal Article %J EURASIP Journal on Applied Signal Processing %D 2005 %T Neuromimetic sound representation for percept detection and manipulation %A Zotkin,Dmitry N %A Chi,T. %A Shamma,S.A. %A Duraiswami, Ramani %X The acoustic wave received at the ears is processed by the human auditory system to separate different sounds along the intensity,pitch, and timbre dimensions. Conventional Fourier-based signal processing, while endowed with fast algorithms, is unable to easily represent a signal along these attributes. In this paper, we discuss the creation of maximally separable sounds in auditory user interfaces and use a recently proposed cortical sound representation, which performs a biomimetic decomposition of an acoustic signal, to represent and manipulate sound for this purpose. We briefly overview algorithms for obtaining, manipulating, and inverting a cortical representation of a sound and describe algorithms for manipulating signal pitch and timbre separately. The algorithms are also used to create sound of an instrument between a “guitar” and a “trumpet.” Excellent sound quality can be achieved if processing time is not a concern, and intelligible signals can be reconstructed in reasonable processing time (about ten seconds of computational time for a one-second signal sampled at 8 kHz). Work on bringing the algorithms into the real-time processing domain is ongoing. %B EURASIP Journal on Applied Signal Processing %V 9 %P 1350 - 1350 %8 2005/// %G eng %0 Journal Article %J AIAA paper %D 2005 %T New experimental data of STBLI at DNS/LES accessible Reynolds numbers %A Bookey,P. %A Wyckham,C. %A Smits,A. J. %A Martin, M.P %B AIAA paper %V 309 %P 2005 - 2005 %8 2005/// %G eng %0 Journal Article %J Theoretical Computer Science %D 2005 %T A new framework for addressing temporal range queries and some preliminary results %A Shi,Qingmin %A JaJa, Joseph F. %K algorithms %K Data structures %K Orthogonal range search %K temporal data %X Given a set of n objects, each characterized by d attributes specified at m fixed time instances, we are interested in the problem of designing space efficient indexing structures such that a class of temporal range search queries can be handled efficiently. When m = 1 , our problem reduces to the d-dimensional orthogonal search problem. We establish efficient data structures to handle several classes of the general problem. Our results include a linear size data structure that enables a query time of O ( log n log m + f ) for one-sided queries when d = 1 , where f is the number of objects satisfying the query. A similar result is shown for counting queries. We also show that the most general problem can be solved with a polylogarithmic query time using superlinear space data structures. %B Theoretical Computer Science %V 332 %P 109 - 121 %8 2005/02/28/ %@ 0304-3975 %G eng %U http://www.sciencedirect.com/science/article/pii/S0304397504007005 %N 1–3 %R 10.1016/j.tcs.2004.10.013 %0 Conference Paper %B Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on %D 2005 %T Non-negative lighting and specular object recognition %A Shirdhonkar,S. %A Jacobs, David W. %K and %K distribution %K eigenfunctions;image %K eigenvalue %K harmonic %K Lambertian %K lighting;nonnegative %K lighting;semidefinite %K matching;object %K object %K objects;Szego %K optimization;incident %K programming;specular %K recognition;optimisation; %K recognition;spherical %K representation;eigenvalues %K theorem;constrained %X Recognition of specular objects is particularly difficult because their appearance is much more sensitive to lighting changes than that of Lambertian objects. We consider an approach in which we use a 3D model to deduce the lighting that best matches the model to the image. In this case, an important constraint is that incident lighting should be non-negative everywhere. In this paper, we propose a new method to enforce this constraint and explore its usefulness in specular object recognition, using the spherical harmonic representation of lighting. The method follows from a novel extension of Szego's eigenvalue distribution theorem to spherical harmonics, and uses semidefinite programming to perform a constrained optimization. The new method is faster as well as more accurate than previous methods. Experiments on both synthetic and real data indicate that the constraint can improve recognition of specular objects by better separating the correct and incorrect models %B Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on %V 2 %P 1323 -1330 Vol. 2 - 1323 -1330 Vol. 2 %8 2005/10// %G eng %R 10.1109/ICCV.2005.168 %0 Journal Article %J Proc. of IEEE International Conference on Computer Vision %D 2005 %T Non-negative lighting and specular object recognition %A Jacobs, David W. %A Shirdhonkar,S. %B Proc. of IEEE International Conference on Computer Vision %V 2 %8 2005/// %G eng %0 Journal Article %J SIAM Journal on Computing %D 2005 %T Novel transformation techniques using q-heaps with applications to computational geometry %A Shi,Q. %A JaJa, Joseph F. %X Using the notions of Q-heaps and fusion trees developed by Fredman and Willard,we develop general transformation techniques to reduce a number of computational geometry prob- lems to their special versions in partially ranked spaces. In particular, we develop a fast fractional cascading technique, which uses linear space and enables sublogarithmic iterative search on catalog trees in the case when the degree of each node is bounded by O(log∈ n), for some constant ϵ > 0, where n is the total size of all the lists stored in the tree. We apply the fast fractional cascading tech- nique in combination with the other techniques to derive the first linear-space sublogarithmic time algorithms for the two fundamental geometric retrieval problems: orthogonal segment intersection and rectangular point enclosure. %B SIAM Journal on Computing %V 34 %P 1474 - 1492 %8 2005/// %G eng %N 6 %0 Journal Article %J Information Processing Letters %D 2005 %T Optimal and near-optimal algorithms for generalized intersection reporting on pointer machines %A Shi,Qingmin %A JaJa, Joseph F. %K algorithms %K computational geometry %K Generalized intersection %X We develop efficient algorithms for a number of generalized intersection reporting problems, including orthogonal and general segment intersection, 2D range searching, rectangular point enclosure, and rectangle intersection search. Our results for orthogonal and general segment intersection, 3-sided 2D range searching, and rectangular pointer enclosure problems match the lower bounds for their corresponding standard versions under the pointer machine model. Our results for the remaining problems improve upon the best known previous algorithms. %B Information Processing Letters %V 95 %P 382 - 388 %8 2005/08/16/ %@ 0020-0190 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020019005001183 %N 3 %R 10.1016/j.ipl.2005.04.008 %0 Conference Paper %B Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems %D 2005 %T Optimal status sets of heterogeneous agent programs %A Stroe,Bogdan %A V.S. Subrahmanian %A Dasgupta,Sudeshna %K heterogeneous agents %K heuristic search %K non-ground representations %K optimal status sets %X There are many situations where an agent can perform one of several sets of actions in responses to changes in its environment, and the agent chooses to perform the set of actions that optimizes some objective function. In past work, Eiter et. al. have proposed a rule based framework for programming agents on top of heterogeneous data sources, but they provide no solutions to the above problem. In this paper, we propose a semantics called optimal feasible status set semantics for agents which allows the agent to associate an objective function with feasible status sets and act according to the feasible status set that optimizes this objective function. We provide both an algorithm to compute exact optimal feasible status sets as well as the TierOpt and FastOpt algorithms to find (suboptimal) feasible status set much faster. We report on experiments on a suite of real agent applications showing that the heuristic algorithms works well in practice. %B Proceedings of the fourth international joint conference on Autonomous agents and multiagent systems %S AAMAS '05 %I ACM %C New York, NY, USA %P 709 - 715 %8 2005/// %@ 1-59593-093-0 %G eng %U http://doi.acm.org/10.1145/1082473.1082581 %R 10.1145/1082473.1082581 %0 Report %D 2005 %T Ordered and Quantum Treemaps: Making Effective Use of 2D Space to Display Hierarchies (2001) %A Bederson, Benjamin B. %A Shneiderman, Ben %A Wattenberg,Martin %K Technical Report %X Treemaps, a space- filling method of visualizing large hierarchical data sets, are receiving increasing attention. Several algorithms have been proposed to create more useful displays by controlling the aspect ratios of the rectangles that make up a treemap. While these algorithms do improve visibility of small items in a single layout, they introduce instability over time in the display of dynamically changing data, fail to preserve order of the underlying data, and create layouts that are difficult to visually search. In addition, continuous treemap algorithms are not suitable for displaying quantum-sized objects within them, such as images. This paper introduces several new treemap algorithms, which address these shortcomings. In addition, we show a new application of these treemaps, using them to present groups of images. The ordered treemap algorithms ensure that items near each other in the given order will be near each other in the treemap layout. Using experimental evidence from Monte Carlo trials, we show that compared to other layout algorithms ordered treemaps are more stable while maintaining relatively favorable aspect ratios of the constituent rectangles. A second test set uses stock market data. The quantum treemap algorithms modify the layout of the continuous treemap algorithms to generate rectangles that are integral multiples of an input object size. The quantum treemap algorithm has been applied to PhotoMesa, an application that supports browsing of large numbers of images. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6486 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Ordered Treemap Layouts (2001) %A Shneiderman, Ben %A Wattenberg,Martin %K Technical Report %X Treemaps, a space-filling method of visualizing large hierarchical data sets, are receiving increasing attention. Several algorithms have been proposed to create more useful displays by controlling the aspect ratios of the rectangles that make up a treemap. While these algorithms do improve visibility of small items in a single layout, they introduce instability over time in the display of dynamically changing data, and fail to preserve an ordering of the underlying data. This paper introduces the ordered treemap, which addresses these two shortcomings. The ordered treemap algorithm ensures that items near each other in the given order will be near each other in the treemap layout. Using experimental evidence from Monte Carlo trials, we show that compared to other layout algorithms ordered treemaps are more stable while maintaining relatively low aspect ratios of the constituent rectangles. A second test set uses stock market data. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6480 %0 Journal Article %J Yugoslav Journal of Operations Research %D 2005 %T A parametric visualization software for the assignment problem %A Charalampos Papamanthou %A Paparrizos, Konstantinos %A Samaras, Nikolaos %B Yugoslav Journal of Operations Research %V 15 %P 147 - 158 %8 2005/// %@ 0354-0243 %G eng %U http://www.doiserbia.nb.rs/Article.aspx?id=0354-02430501147P&AspxAutoDetectCookieSupport=1 %N 1 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T A Photo History of SIGCHI: Evolution of Design from Personal to Public (2002) %A Shneiderman, Ben %K Technical Report %X For 20 years I have been photographing personalities and events in the emerging discipline of humanomputer interaction. Until now, only a few of these photos were published in newsletters or were shown to visitors who sought them out. Now this photo history is going from a personal record to a public archive. This archive should be interesting for professional members of this community who want to reminisce, as well as for historians and journalists who want to understand what happened. Students and Web surfers may also want to look at the people who created better interfaces and more satisfying user experiences. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6529 %0 Journal Article %J ACM SIGOPS Operating Systems Review %D 2005 %T Pioneer: verifying code integrity and enforcing untampered code execution on legacy systems %A Seshadri, Arvind %A Luk, Mark %A Elaine Shi %A Perrig, Adrian %A van Doorn, Leendert %A Khosla, Pradeep %K dynamic root of trust %K rootkit detection %K self-check-summing code %K software-based code attestation %K verifiable code execution %X We propose a primitive, called Pioneer, as a first step towards verifiable code execution on untrusted legacy hosts. Pioneer does not require any hardware support such as secure co-processors or CPU-architecture extensions. We implement Pioneer on an Intel Pentium IV Xeon processor. Pioneer can be used as a basic building block to build security systems. We demonstrate this by building a kernel rootkit detector. %B ACM SIGOPS Operating Systems Review %V 39 %P 1 - 16 %8 2005 %@ 0163-5980 %G eng %U http://doi.acm.org/10.1145/1095809.1095812 %N 5 %0 Conference Paper %B 35th AIAA Fluid Dynamics Conference %D 2005 %T Preliminary Study of the Turbulence Structure in Supersonic Boundary Layers using DNS Data %A Taylor,E. M %A Martin, M.P %A Smits,A. J. %X Direct numerical simulation data are used to visualize coherent structures in turbulentboundary layers at Mach numbers from 0.3 to 7. Different criteria to identify the three- dimensional turbulence structure are selected. We find that using the discriminant of the velocity gradient tensor, the swirling strength and the λ2 criteria give nearly identical results, with λ2 identifying more structures very close to the wall. %B 35th AIAA Fluid Dynamics Conference %C Toronto, Canada %8 2005/// %G eng %0 Book Section %B On the Move to Meaningful Internet Systems 2005: CoopIS, DOA, and ODBASEOn the Move to Meaningful Internet Systems 2005: CoopIS, DOA, and ODBASE %D 2005 %T Probabilistic Ontologies and Relational Databases %A Udrea,Octavian %A Yu,Deng %A Hung,Edward %A Subrahmanian,V. %E Meersman,Robert %E Tari,Zahir %K Computer %K Science %X The relational algebra and calculus do not take the semantics of terms into account when answering queries. As a consequence, not all tuples that should be returned in response to a query are always returned, leading to low recall. In this paper, we propose the novel notion of a constrained probabilistic ontology (CPO). We developed the concept of a CPO-enhanced relation in which each attribute of a relation has an associated CPO. These CPOs describe relationships between terms occurring in the domain of that attribute. We show that the relational algebra can be extended to handle CPO-enhanced relations. This allows queries to yield sets of tuples, each of which has a probability of being correct. %B On the Move to Meaningful Internet Systems 2005: CoopIS, DOA, and ODBASEOn the Move to Meaningful Internet Systems 2005: CoopIS, DOA, and ODBASE %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3760 %P 1 - 17 %8 2005/// %@ 978-3-540-29736-9 %G eng %U http://dx.doi.org/10.1007/11575771_1 %0 Journal Article %J Molecular Microbiology %D 2005 %T Promoter architecture and response to a positive regulator of archaeal transcription %A Ouhammouch,Mohamed %A Langham,Geoffrey E %A Hausner,Winfried %A Simpson,Anjana J. %A El‐Sayed, Najib M. %A Geiduschek,E. Peter %X The archaeal transcription apparatus is chimeric: its core components (RNA polymerase and basal factors) closely resemble those of eukaryotic RNA polymerase II, but the putative archaeal transcriptional regulators are overwhelmingly of bacterial type. Particular interest attaches to how these bacterial-type effectors, especially activators, regulate a eukaryote-like transcription system. The hyperthermophilic archaeon Methanocaldococcus jannaschii encodes a potent transcriptional activator, Ptr2, related to the Lrp/AsnC family of bacterial regulators. Ptr2 activates rubredoxin 2 (rb2) transcription through a bipartite upstream activating site (UAS), and conveys its stimulatory effects on its cognate transcription machinery through direct recruitment of the TATA binding protein (TBP). A functional dissection of the highly constrained architecture of the rb2 promoter shows that a ‘one-site’ minimal UAS suffices for activation by Ptr2, and specifies the required placement of this site. The presence of such a simplified UAS upstream of the natural rubrerythrin (rbr) promoter also suffices for positive regulation by Ptr2 in vitro, and TBP recruitment remains the primary means of transcriptional activation at this promoter. %B Molecular Microbiology %V 56 %P 625 - 637 %8 2005/05/01/ %@ 1365-2958 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1365-2958.2005.04563.x/abstract %N 3 %R 10.1111/j.1365-2958.2005.04563.x %0 Report %D 2005 %T Promoting Universal Usability with Multi-Layer Interface Design (2003) %A Shneiderman, Ben %K Technical Report %X Increased interest in universal usability is causing some researchers to study advanced strategies for satisfying first-time as well as intermittent and expert users. This paper promotes the idea of multi-layer interface designs that enable first-time and novice users to begin with a limited set of features at layer 1. They can remain at layer 1, then move up to higher layers when needed or when they have time to learn further features. The arguments for and against multi-layer interfaces are presented with two example systems: a word processor with 8 layers and an interactive map with 3 layers. New research methods and directions are proposed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6510 %0 Conference Paper %B Cluster Computing and the Grid, IEEE International Symposium on %D 2005 %T Query planning for the grid: adapting to dynamic resource availability %A Zhang,K. %A Andrade,H. %A Raschid, Louiqa %A Sussman, Alan %X The availability of massive datasets, comprising sensor measurements or the results of scientific simulations, has had a significant impact on the methodology of scientific reasoning. Scientists require storage, bandwidth and computational capacity to query and analyze these datasets, to understand physical phenomena or to test hypotheses. This paper addresses the challenge of identifying and selecting resources to develop an evaluation plan for large scale data analysis queries when data processing capabilities and datasets are dispersed across nodes in one or more computing and storage clusters. We show that generating an optimal plan is hard and we propose heuristic techniques to find a good choice of resources. We also consider heuristics to cope with dynamic resource availability; in this situation we have stale information about reusable cached results (datasets) and the load on various nodes. %B Cluster Computing and the Grid, IEEE International Symposium on %I IEEE Computer Society %C Los Alamitos, CA, USA %V 2 %P 751 - 758 %8 2005/// %@ 0-7803-9074-1 %G eng %R http://doi.ieeecomputersociety.org/10.1109/CCGRID.2005.1558638 %0 Journal Article %J Information VisualizationInformation Visualization %D 2005 %T A Rank-by-Feature Framework for Interactive Exploration of Multidimensional Data %A Seo,Jinwook %A Shneiderman, Ben %K dynamic query %K exploratory data analysis %K feature detection/selection %K graphical displays %K Information Visualization %K Rank-by-feature framework %X Interactive exploration of multidimensional data sets is challenging because: (1) it is difficult to comprehend patterns in more than three dimensions, and (2) current systems often are a patchwork of graphical and statistical methods leaving many researchers uncertain about how to explore their data in an orderly manner. We offer a set of principles and a novel rank-by-feature framework that could enable users to better understand distributions in one (1D) or two dimensions (2D), and then discover relationships, clusters, gaps, outliers, and other features. Users of our framework can view graphical presentations (histograms, boxplots, and scatterplots), and then choose a feature detection criterion to rank 1D or 2D axis-parallel projections. By combining information visualization techniques (overview, coordination, and dynamic query) with summaries and statistical methods users can systematically examine the most important 1D and 2D axis-parallel projections. We summarize our Graphics, Ranking, and Interaction for Discovery (GRID) principles as: (1) study 1D, study 2D, then find features (2) ranking guides insight, statistics confirm. We implemented the rank-by-feature framework in the Hierarchical Clustering Explorer, but the same data exploration principles could enable users to organize their discovery process so as to produce more thorough analyses and extract deeper insights in any multidimensional data application, such as spreadsheets, statistical packages, or information visualization tools. %B Information VisualizationInformation Visualization %V 4 %P 96 - 113 %8 2005/06/20/ %@ 1473-8716, 1473-8724 %G eng %U http://ivi.sagepub.com/content/4/2/96 %N 2 %R 10.1057/palgrave.ivs.9500091 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T A Rank-by-Feature Framework for Interactive Exploration of Multidimensional Data (2004) %A Seo,Jinwook %A Shneiderman, Ben %K Technical Report %X Interactive exploration of multidimensional data sets is challenging because: (1) it is difficult to comprehend patterns in more than three dimensions, and (2) current systems often are a patchwork of graphical and statistical methods leaving many researchers uncertain about how to explore their data in an orderly manner. We offer a set of principles and a novel rank-by-feature framework that could enable users to better understand distributions in one (1D) or two dimensions (2D), and then discover relationships, clusters, gaps, outliers, and other features. Users of our framework can view graphical presentations (histograms, boxplots, and scatterplots), and then choose a feature detection criterion to rank 1D or 2D axis-parallel projections. By combining information visualization techniques (overview, coordination, and dynamic query) with summaries and statistical methods users can systematically examine the most important 1D and 2D axis-parallel projections. We summarize our Graphics, Ranking, and Interaction for Discovery (GRID) principles as: (1) 1D, 2D, then features (2) graphics, ranking, summaries, then statistics. We implemented the rank-by-feature framework in the Hierarchical Clustering Explorer, but the same data exploration principles could enable users to organize their discovery process so as to produce more thorough analyses and extract deeper insights in any multidimensional data application, such as spreadsheets, statistical packages, or information visualization tools. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6524 %0 Conference Paper %B Data Engineering, 2005. ICDE 2005. Proceedings. 21st International Conference on %D 2005 %T RDF aggregate queries and views %A Hung,E. %A Deng,Yu %A V.S. Subrahmanian %K aggregate %K databases; %K DBMS; %K description %K framework; %K languages; %K Maintenance %K methods; %K processing; %K queries; %K query %K RDF %K relational %K resource %K standard; %K standards; %K view %K Web %X Resource description framework (RDF) is a rapidly expanding Web standard. RDF databases attempt to track the massive amounts of Web data and services available. In this paper, we study the problem of aggregate queries. We develop an algorithm to compute answers to aggregate queries over RDF databases and algorithms to maintain views involving those aggregates. Though RDF data can be stored in a standard relational DBMS (and hence we can execute standard relational aggregate queries and view maintenance methods on them), we show experimentally that our algorithms that operate directly on the RDF representation exhibit significantly superior performance. %B Data Engineering, 2005. ICDE 2005. Proceedings. 21st International Conference on %P 717 - 728 %8 2005/04// %G eng %R 10.1109/ICDE.2005.121 %0 Report %D 2005 %T Recovering from Intrusions in the OSPF Data-plane %A Elaine Shi %A Lu, Yong %A Reid, Matt %X In this paper, we propose CONS-ROUTE, a data-plane intrusion recoverymechanism for securing the OSPF routing protocol. CONS-ROUTE allows routers to perform intrusion detection in a distributed manner. The intrusion detection outcome can be used globally to reevaluate routing decisions in a way that is resilient to the slandering attack, where a malicious router claims that a legitimate router is misbehaving. We evaluate CONS-ROUTE through simulation and compare it with several simple OSPF data plane resilience techniques. %B Selected Project Reports, Spring 2005 Advanced OS & Distributed Systems (15-712) %I School of Computer Science, Carnegie Mellon University %8 2005 %G eng %U http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.60.5274&rep=rep1&type=pdf#page=51 %R Technical Report %0 Journal Article %J Advances in Cryptology–EUROCRYPT 2005 %D 2005 %T Reducing complexity assumptions for statistically-hiding commitment %A Haitner,I. %A Horvitz,O. %A Katz, Jonathan %A Koo,C. Y %A Morselli,R. %A Shaltiel,R. %X Determining the minimal assumptions needed to construct various cryptographic building blocks has been a focal point of research in theoretical cryptography. Here, we revisit the following question: what are the minimal assumptions needed to construct statistically-hiding commitment schemes? Previously, it was known how to construct such schemes based on one-way permutations. We improve upon this by constructing statistically-hiding commitment schemes based on approximable-preimage-size one-way functions. These are one-way functions for which there is an efficient way to approximate the number of preimages of a given output. A special case (for which we show a somewhat simpler construction) is that of regular one-way functions where all outputs have the same number of preimages.We utilize two different approaches in constructing statistically-hiding commitment schemes. Our first approach proceeds by showing that the scheme of Naor et al. can be implemented using any one-way function having an output distribution which is “sufficiently similar” to uniform. We then construct one-way functions with this property from approximable-preimage-size one-way functions. Our second approach begins by constructing a commitment scheme which is statistically hiding against an honest-but-curious receiver. We then demonstrate a compiler which transforms any such commitment scheme into one which is statistically hiding even against a malicious receiver. This compiler and its analysis may be of independent interest. %B Advances in Cryptology–EUROCRYPT 2005 %P 614 - 614 %8 2005/// %G eng %R 10.1007/11426639_4 %0 Book Section %B Human-Computer Interaction - INTERACT 2005Human-Computer Interaction - INTERACT 2005 %D 2005 %T Representing Unevenly-Spaced Time Series Data for Visualization and Interactive Exploration %A Aris,Aleks %A Shneiderman, Ben %A Plaisant, Catherine %A Shmueli,Galit %A Jank,Wolfgang %E Costabile,Maria %E Paternò,Fabio %X Visualizing time series is useful to support discovery of relations and patterns in financial, genomic, medical and other applications. Often, measurements are equally spaced over time. We discuss the challenges of unevenly-spaced time series and present fourrepresentationmethods: sampled events, aggregated sampled events, event index and interleaved event index. We developed these methods while studying eBay auction data with TimeSearcher. We describe the advantages, disadvantages, choices for algorithms and parameters, and compare the different methods for different tasks. Interaction issues such as screen resolution, response time for dynamic queries, and learnability are governed by these decisions. %B Human-Computer Interaction - INTERACT 2005Human-Computer Interaction - INTERACT 2005 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3585 %P 835 - 846 %8 2005/// %@ 978-3-540-28943-2 %G eng %U http://dx.doi.org/10.1007/11555261_66 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Representing Unevenly-Spaced Time Series Data for Visualization and Interactive Exploration (2005) %A Aris,Aleks %A Shneiderman, Ben %A Plaisant, Catherine %A Shmueli,Galit %A Jank,Wolfgang %K Technical Report %X Visualizing time series data is useful to support discovery of relations and patterns in financial, genomic, medical and other applications. In most time series, measurements are equally spaced over time. This paper discusses the challenges for unevenly-spaced time series data and presents four methods to represent them: sampled events, aggregated sampled events, event index and interleaved event index. We developed these methods while studying eBay auction data with TimeSearcher. We describe the advantages, disadvantages, choices for algorithms and parameters, and compare the different methods. Since each method has its advantages, this paper provides guidance for choosing the right combination of methods, algorithms, and parameters to solve a given problem for unevenly-spaced time series. Interaction issues such as screen resolution, response time for dynamic queries, and meaning of the visual display are governed by these decisions. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6537 %0 Journal Article %J Information Security Practice and Experience %D 2005 %T Robust routing in malicious environment for ad hoc networks %A Yu,Z. %A Seng,C. Y %A Jiang,T. %A Wu,X. %A Arbaugh, William A. %B Information Security Practice and Experience %P 36 - 47 %8 2005/// %G eng %0 Conference Paper %B Proceedings of the 2005 national conference on Digital government research %D 2005 %T Scalable data collection infrastructure for digital government applications %A Samet, Hanan %A Tanin,Egemen %A Golubchik,Leana %X Highlights are presented of a project on the development of a scalable data collection infrastructure of digital government applications. The approach takes advantage of both the distributed nature of the data and the distributed ways of interacting with it. The approach is three-parted having theoretical, systems, and application components. On the theoretical and systems levels, work has continued on the development of the BISTRO system. On the applications level, work has continued on the investigation of the use of peer-to-peer networks in a number of domains in which the proper handling of spatial data is important. %B Proceedings of the 2005 national conference on Digital government research %S dg.o '05 %I Digital Government Society of North America %P 305 - 306 %8 2005/// %G eng %U http://dl.acm.org/citation.cfm?id=1065226.1065333 %0 Book Section %B Approximation, Randomization and Combinatorial Optimization. Algorithms and TechniquesApproximation, Randomization and Combinatorial Optimization. Algorithms and Techniques %D 2005 %T Scheduling on Unrelated Machines Under Tree-Like Precedence Constraints %A Kumar, V. %A Marathe,Madhav %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %E Chekuri,Chandra %E Jansen,Klaus %E Rolim,José %E Trevisan,Luca %X We present polylogarithmic approximations for the R | prec | C max and R | prec |∑ j w j C j problems, when the precedence constraints are “treelike” – i.e., when the undirected graph underlying the precedences is a forest. We also obtain improved bounds for the weighted completion time and flow time for the case of chains with restricted assignment – this generalizes the job shop problem to these objective functions. We use the same lower bound of “congestion+dilation”, as in other job shop scheduling approaches. The first step in our algorithm for the R | prec | C max problem with treelike precedences involves using the algorithm of Lenstra, Shmoys and Tardos to obtain a processor assignment with the congestion + dilation value within a constant factor of the optimal. We then show how to generalize the random delays technique of Leighton, Maggs and Rao to the case of trees. For the weighted completion time, we show a certain type of reduction to the makespan problem, which dovetails well with the lower bound we employ for the makespan problem. For the special case of chains, we show a dependent rounding technique which leads to improved bounds on the weighted completion time and new bicriteria bounds for the flow time. %B Approximation, Randomization and Combinatorial Optimization. Algorithms and TechniquesApproximation, Randomization and Combinatorial Optimization. Algorithms and Techniques %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3624 %P 609 - 609 %8 2005/// %@ 978-3-540-28239-6 %G eng %U http://dx.doi.org/10.1007/11538462_13 %0 Journal Article %J Advances in Cryptology–EUROCRYPT 2005 %D 2005 %T Secure remote authentication using biometric data %A Boyen,X. %A Dodis,Y. %A Katz, Jonathan %A Ostrovsky,R. %A Smith,A. %X Biometric data offer a potential source of high-entropy, secret information that can be used in cryptographic protocols provided two issues are addressed: (1) biometric data are not uniformly distributed; and (2) they are not exactly reproducible. Recent work, most notably that of Dodis, Reyzin, and Smith, has shown how these obstacles may be overcome by allowing some auxiliary public information to be reliably sent from a server to the human user. Subsequent work of Boyen has shown how to extend these techniques, in the random oracle model, to enable unidirectional authentication from the user to the server without the assumption of a reliable communication channel.We show two efficient techniques enabling the use of biometric data to achieve mutual authentication or authenticated key exchange over a completely insecure (i.e., adversarially controlled) channel. In addition to achieving stronger security guarantees than the work of Boyen, we improve upon his solution in a number of other respects: we tolerate a broader class of errors and, in one case, improve upon the parameters of his solution and give a proof of security in the standard model. %B Advances in Cryptology–EUROCRYPT 2005 %P 147 - 163 %8 2005/// %G eng %R 10.1007/11426639_9 %0 Journal Article %J Distributed Computing and Internet Technology %D 2005 %T Secure requirements elicitation through triggered message sequence charts %A Ray,A. %A Sengupta,B. %A Cleaveland, Rance %B Distributed Computing and Internet Technology %P 273 - 282 %8 2005/// %G eng %0 Conference Paper %B Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on %D 2005 %T Security of feature extraction in image hashing %A Swaminathan,A. %A Mao,Yinian %A M. Wu %K cryptography; %K differential %K digital %K entropy; %K extraction; %K feature %K functions; %K hash %K hashing; %K image %K metric; %K processing; %K randomness; %K robustness; %K Security %K signature; %K signatures; %X Security and robustness are two important requirements for image hash functions. We introduce "differential entropy" as a metric to quantify the amount of randomness in image hash functions and to study their security. We present a mathematical framework and derive expressions for the proposed security metric for various common image hashing schemes. Using the proposed security metric, we discuss the trade-offs between security and robustness in image hashing. %B Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on %V 2 %P ii/1041 - ii/1044 Vol. 2 - ii/1041 - ii/1044 Vol. 2 %8 2005/03// %G eng %R 10.1109/ICASSP.2005.1415586 %0 Conference Paper %B Proceedings of the 2005 international ACM SIGGROUP conference on Supporting group work %D 2005 %T Seeking the source: software source code as a social and technical artifact %A de Souza,Cleidson %A Jon Froehlich %A Dourish,Paul %K data mining %K social networks %K socio-technical systems %K software repositories %B Proceedings of the 2005 international ACM SIGGROUP conference on Supporting group work %S GROUP '05 %I ACM %C New York, NY, USA %P 197 - 206 %8 2005/// %@ 1-59593-223-2 %G eng %U http://doi.acm.org/10.1145/1099203.1099239 %R 10.1145/1099203.1099239 %0 Journal Article %J Proceedings of DUC2005 %D 2005 %T A sentence-trimming approach to multi-document summarization %A Zajic, David %A Dorr, Bonnie J %A Jimmy Lin %A Monz,C. %A Schwartz,R. %X We implemented an initial applicationof a sentence-trimming approach (Trim- mer) to the problem of multi-document summarization in the MSE2005 and DUC2005 tasks. Sentence trimming was incorporated into a feature-based summarization system, called Multi- Document Trimmer (MDT), by us- ing sentence trimming as both a pre- processing stage and a feature for sen- tence ranking. We demonstrate that we were able to port Trimmer easily to this new problem. Although the direct im- pact of sentence trimming was minimal compared to other features used in the system, the interaction of the other fea- tures resulted in trimmed sentences ac- counting for nearly half of the selected summary sentences. %B Proceedings of DUC2005 %8 2005/// %G eng %0 Journal Article %J Genome Biology %D 2005 %T Serendipitous discovery of Wolbachia genomes in multiple Drosophila species %A Salzberg,Steven L. %A Hotopp,Julie CD %A Delcher,Arthur L. %A Pop, Mihai %A Smith,Douglas R. %A Eisen,Michael B. %A Nelson,William C. %X The Trace Archive is a repository for the raw, unanalyzed data generated by large-scale genome sequencing projects. The existence of this data offers scientists the possibility of discovering additional genomic sequences beyond those originally sequenced. In particular, if the source DNA for a sequencing project came from a species that was colonized by another organism, then the project may yield substantial amounts of genomic DNA, including near-complete genomes, from the symbiotic or parasitic organism. %B Genome Biology %V 6 %P R23 - R23 %8 2005/02/22/ %@ 1465-6906 %G eng %U http://genomebiology.com/2005/6/3/R23 %N 3 %R 10.1186/gb-2005-6-3-r23 %0 Report %D 2005 %T Severity and Impact of Computer User Frustration: A Comparison of Student and Workplace Users (2002) %A Lazar,Jonathan %A Jones,Adam %A Hackley,Mary %A Shneiderman, Ben %K Technical Report %X User frustration with information and computing technology is a pervasive and persistent problem. When computers crash, network congestion causes delays, and poor user interfaces trigger confusion there are dramatic consequences for individuals, organizations, and society. These frustrations not only cause personal dissatisfaction and loss of self-efficacy, but may disrupt workplaces, slow learning, and reduce participation in local and national communities. Our study of 107 student computer users and 50 workplace computer users shows high levels of frustration and loss of 1/3 to 1/2 of time spent. This paper reports on the incident-specific and user-specific causes of frustration, and they raise frustration severity. It examines the frustration impacts on the daily interactions of the users. The time lost and time to fix problem, and importance of task, strongly correlate with frustration levels for both student and workplace users. Differences between students and workplace users are discussed in the paper. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6496 %0 Conference Paper %B 2005 IEEE Symposium on Visual Languages and Human-Centric Computing %D 2005 %T Show Me! Guidelines for producing recorded demonstrations %A Plaisant, Catherine %A Shneiderman, Ben %K animation %K auditory cues %K Computer aided instruction %K Computer animation %K Computer Graphics %K Computer interfaces %K computer literacy %K content guidelines %K documentation %K Government %K Graphical user interfaces %K Guidelines %K Laboratories %K narration %K recorded demonstrations %K screen capture animation %K technical guidelines %K usability %K User interfaces %K visual appeal %K visual cues %X Although recorded demonstrations (screen capture animations with narration) have become a popular form of instruction for user interfaces, little work has been done to describe guidelines for their design. Based on our experience in several projects, we offer a starting set of guidelines for the design of visually appealing and cognitively effective recorded demonstrations. Technical guidelines encourage users to keep file sizes small, strive for universal usability, and ensure user control etc. and provide tips to achieve those goals. Content guidelines include: create short demonstrations that focus on tasks, highlight each step with auditory and visual cues, synchronize narration and animation carefully, and create demonstrations with a clear beginning, middle, and end. %B 2005 IEEE Symposium on Visual Languages and Human-Centric Computing %I IEEE %P 171 - 178 %8 2005/09/20/24 %@ 0-7695-2443-5 %G eng %R 10.1109/VLHCC.2005.57 %0 Report %D 2005 %T Show Me! Guidelines for Producing Recorded Demonstrations (2005) %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X Although recorded demonstrations (screen capture animations with narration) have become a popular form of instruction for user interfaces, little work has been done to describe guidelines for their design. Based on our experience in several projects, we offer a starting set of guidelines for the design of recorded demonstrations. Technical guidelines encourage users to keep file sizes small, strive for universal usability, and ensure user control etc. and provide tips to achieve those goals. Content guidelines include: create short demonstrations that focus on tasks, highlight each step with auditory and visual cues, synchronize narration and animation carefully, and create demonstrations with a clear beginning, middle, and end. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6533 %0 Journal Article %J Concurrency and Computation: Practice and Experience %D 2005 %T A simulation and data analysis system for large-scale, data-driven oil reservoir simulation studies %A Kurc, T. %A Catalyurek,U. %A Zhang,X. %A Saltz, J. %A Martino,R. %A Wheeler,M. %A Peszyńska,M. %A Sussman, Alan %A Hansen,C. %A Sen,M. %A others %B Concurrency and Computation: Practice and Experience %V 17 %P 1441 - 1467 %8 2005/// %G eng %N 11 %0 Report %D 2005 %T Snap-Together Visualization: A User Interface for Coordinating Visualizations via Relational Schemata (2000) %A North,Chris %A Shneiderman, Ben %K Technical Report %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6470 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Social and Psychological Influences on Computer User Frustration (Newhagen book chapter, 2002) %A Bessiere,Katie %A Ceaparu,Irina %A Lazar,Jonathan %A Robinson,John %A Shneiderman, Ben %K Technical Report %X As computer usage has proliferated, so has user frustration. Even devoted and knowledgeable users encounter annoying delays, incomprehensible messages, incompatible files, and indecipherable menus. The frustration generated by these problems can be personally disturbing and socially disruptive. Psychological and social perspectives on frustration may clarify the relationships among variables such as personality types, cultural factors, goal attainment, workplace anger, and computer anxiety. These perspectives may also help designers, managers, and users understand the range of responses to frustration, which could lead to effective interventions such as redesign of software, improved training, better online help, user discipline, and even resetting of national research priorities. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6497 %0 Journal Article %J Software, IEEE %D 2005 %T Software process improvement in small organizations: a case study %A Dangle,K.C. %A Larsen,P. %A Shaw,M. %A Zelkowitz, Marvin V %K area; %K aspects; %K capability %K CMM %K development %K improvement; %K key %K management; %K Maturity %K model; %K organisational %K organization; %K process %K small %K software %X The Capability Maturity Model for software has become a major force in software development process improvement. We looked at the role of process improvement in the context of a small organization. Although the Capability Maturity Model integration is replacing the CMM, we focused on the older CMM. We had to look at more than which CMM key process areas apply to small businesses. Our overall goal was to institute good software development practices, we used the language of the CMM to gain the management's cooperation. Applying these practices is essential to managing growth, yet undertaking this effort without prior experience could impede a small company's innovative nature. This case study's purpose was to investigate the success factors of a software process improvement effort for a small software development organization. %B Software, IEEE %V 22 %P 68 - 75 %8 2005/12//nov %@ 0740-7459 %G eng %N 6 %R 10.1109/MS.2005.162 %0 Journal Article %J ACM SIGSOFT Software Engineering Notes %D 2005 %T Software repository mining with Marmoset: an automated programming project snapshot and testing system %A Spacco,J. %A Strecker,J. %A Hovemeyer,D. %A Pugh, William %B ACM SIGSOFT Software Engineering Notes %V 30 %P 1 - 5 %8 2005/// %G eng %N 4 %0 Journal Article %J Algorithms and Computation %D 2005 %T Space-efficient and fast algorithms for multidimensional dominance reporting and counting %A JaJa, Joseph F. %A Mortensen,C. %A Shi,Q. %X We present linear-space sub-logarithmic algorithms for handling the 3-dimensional dominance reporting and the 2-dimensional dominance counting problems. Under the RAM model as described in [M. L. Fredman and D. E. Willard. ldquoSurpassing the information theoretic bound with fusion treesrdquo, Journal of Computer and System Sciences, 47:424–436, 1993], our algorithms achieve O(log n/loglog n+f) query time for the 3-dimensional dominance reporting problem, where f is the output size, and O(log n/loglog n) query time for the 2-dimensional dominance counting problem. We extend these results to any constant dimension d ge 3, achieving O(n(log n/loglog n)d – 3) space and O((log n/loglog n)d – 2+ f) query time for the reporting case and O(n(log n/loglog n)d – 2) space and O((log n/loglog n)d – 1) query time for the counting case. %B Algorithms and Computation %P 1755 - 1756 %8 2005/// %G eng %R 10.1007/978-3-540-30551-4_49 %0 Conference Paper %B IEEE International Symposium on Cluster Computing and the Grid, 2005. CCGrid 2005 %D 2005 %T Spatial indexing of distributed multidimensional datasets %A Nam,B. %A Sussman, Alan %K centralized global index algorithm %K centralized index server %K Computer science %K database indexing %K distributed databases %K distributed multidimensional dataset %K Educational institutions %K File servers %K Indexing %K Large-scale systems %K Multidimensional systems %K Network servers %K replication protocol %K replication techniques %K scalability %K Sensor systems %K spatial data structures %K spatial indexing %K two-level hierarchical index algorithm %K wide area networks %X While declustering methods for distributed multidimensional indexing of large datasets have been researched widely in the past, replication techniques for multidimensional indexes have not been investigated deeply. In general, a centralized index server may become the performance bottleneck in a wide area network rather than the data servers, since the index is likely to be accessed more often than any of the datasets in the servers. In this paper, we present two different multidimensional indexing algorithms for a distributed environment - a centralized global index and a two-level hierarchical index. Our experimental results show that the centralized scheme does not scale well for either insertion or searching the index. In order to improve the scalability of the index server, we have employed a replication protocol for both the centralized and two-level index schemes that allows some inconsistency between replicas without affecting correctness. Our experiments show that the two-level hierarchical index scheme shows better scalability for both building and searching the index than the non-replicated centralized index, but replication can make the centralized index faster than the two-level hierarchical index for searching in some cases. %B IEEE International Symposium on Cluster Computing and the Grid, 2005. CCGrid 2005 %I IEEE %V 2 %P 743- 750 Vol. 2 - 743- 750 Vol. 2 %8 2005/05/09/12 %@ 0-7803-9074-1 %G eng %R 10.1109/CCGRID.2005.1558637 %0 Report %D 2005 %T Structural metadata and parsing speech %A Harper,M. %A Dorr, Bonnie J %A Hale,J. %A Roark,B. %A Shafran,I. %A Lease,M. %A Liu,Y. %A Snover,M. %A Yung,L. %A Krasnyanskaya,A. %A others %I JHU Language Engineering Workshop %8 2005/// %G eng %0 Report %D 2005 %T Stub Domain DDoS Detection %A Kommareddy,C. %A Levin,D. %A Bhattacharjee, Bobby %A La,R.J. %A Shayman,M. %I Technical Report of the Department of Computer Science, University of Maryland, College Park %8 2005/// %G eng %0 Conference Paper %D 2005 %T A system for performing content-based searches on a database of mechanical assemblies %A Deshmukh,A. S. %A Gupta,S.K. %A Karnik,M. V. %A Sriram,R.D. %X The increased use of 3D CAD systems by productdevelopment organizations has resulted in large databases of assemblies; this explosion of assembly data will continue in the future. Currently, there are no effective content-based techniques to search these databases. Ability to perform content-based searches on these databases is expected to help the designers in the following two ways. First, it can facilitate reuse of existing assembly designs, and thereby reducing the design time. Second, a lot of useful Design for Manufacturing and Assembly (DFMA) knowledge is embedded in existing assemblies. Therefore a capability to locate existing assemblies and examine them can be used as a learning tool by the designers to learn from the existing assembly designs and hence transfer the best DFMA practices to new designers. This paper describes a system for performing content-based searches on assembly databases. We identify templates for comprehensive search definitions and describe algorithms to perform content-based searches for mechanical assemblies. We also illustrate capabilities of our system through several examples. %C Orlando, Florida USA %8 2005/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.160.8372&rep=rep1&type=pdf %0 Journal Article %J Algorithms and Computation %D 2005 %T Techniques for indexing and querying temporal observations for a collection of objects %A Shi,Q. %A JaJa, Joseph F. %X We consider the problem of dynamically indexing temporal observations about a collection of objects, each observation consisting of a key identifying the object, a list of attribute values and a timestamp indicating the time at which these values were recorded. We make no assumptions about the rates at which these observations are collected, nor do we assume that the various objects have about the same number of observations. We develop indexing structures that are almost linear in the total number of observations available at any given time instance, and that support dynamic additions of new observations in polylogarithmic time. Moreover, these structures allow the quick handling of queries to identify objects whose attribute values fall within a certain range at every time instance of a specified time interval. Provably good bounds are established. %B Algorithms and Computation %P 822 - 834 %8 2005/// %G eng %R 10.1007/978-3-540-30551-4_70 %0 Conference Paper %B Information Visualization, 2005. INFOVIS 2005. IEEE Symposium on %D 2005 %T Temporal visualization of planning polygons for e~cient partitioning of geo-spatial data %A Shanbhag,P. %A Rheingans,P. %A desJardins, Marie %X Partitioning of geo-spatial data for efficient allocation of resources such as schools and emergency health care services is driven by a need to provide better and more effective services. Partitioning of spatial data is a complex process that depends on numerous factors such as population, costs incurred in deploying or utilizing resources and target capacity of a resource. Moreover, complex data such as population distributions are dynamic i.e. they may change over time. Simple animation may not effectively show temporal changes in spatial data. We propose the use of three temporal visualization techniques -wedges, rings and time slices - to display the nature of change in temporal data in a single view. Along with maximizing resource utilization and minimizing utilization costs, a partition should also ensure the long term effectiveness of the plan. We use multi-attribute visualization techniques to highlight the strengths and identify the weaknesses of a partition. Comparative visualization techniques allow multiple partitions to be viewed simultaneously. Users can make informed decisions about how to partition geo spatial data by using a combination of our techniques for multi-attribute visualization, temporal visualization and comparative visualization. %B Information Visualization, 2005. INFOVIS 2005. IEEE Symposium on %I IEEE %P 211 - 218 %8 2005/// %@ 0-7803-9464-X %G eng %U http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=1532149 %R 10.1109/INFVIS.2005.1532149 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Toward a Statistical Knowledge Network (2004) %A Marchionini,Gary %A Haas,Stephanie %A Plaisant, Catherine %A Shneiderman, Ben %A Hert,Carol %K Technical Report %X Statistics support planning and decision making and enormous efforts are made to collect data and produce statistics at all levels of governance. An important principle of democratic societies is that government statistics should be accessible to the broadest possible constituencies to empower better plans and decisions in all aspects of life. Given the potential of near-ubiquitous Internet access in homes and workplaces and efforts by government agencies to mount websites, physical access to large volumes of government is close to a fait accompli. What remains a significant challenge is enabling access to the right statistical information at the right time and in the right form. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6522 %0 Journal Article %J ICASSP apos %D 2005 %T Tracking Algorithm Using Background-Foreground Motion Models and Multiple Cues %A Shao, J. %A Zhou,S. K %A Chellapa, Rama %X We present a stochastic tracking algorithm for surveillance videoswhere targets are dim and of low resolution. Our tracker is mainly based on the particle filter algorithm. Two important novel fea- tures of the tracker include: A motion model consisting of both background and foreground motion parameters is built. Multiple cues are adaptively integrated in a system observation model when estimating the likelihood functions. Based on these features, the accuracy and robustness of the tracker has been improved, which is very important for surveillance problems. We present the results of applying the proposed algorithm to many videos. %B ICASSP apos %P 233 - 236 %8 2005/// %G eng %0 Conference Paper %B Image Processing, 2005. ICIP 2005. IEEE International Conference on %D 2005 %T Tracking objects in video using motion and appearance models %A Sankaranarayanan,A. C %A Chellapa, Rama %A Qinfen Zheng %K algorithm; %K analysis; %K appearance %K background %K estimation; %K image %K likelihood %K maximum %K model; %K models; %K MOTION %K object %K processing; %K signal %K target %K tracking %K tracking; %K video %K visual %X This paper proposes a visual tracking algorithm that combines motion and appearance in a statistical framework. It is assumed that image observations are generated simultaneously from a background model and a target appearance model. This is different from conventional appearance-based tracking, that does not use motion information. The proposed algorithm attempts to maximize the likelihood ratio of the tracked region, derived from appearance and background models. Incorporation of motion in appearance based tracking provides robust tracking, even when the target violates the appearance model. We show that the proposed algorithm performs well in tracking targets efficiently over long time intervals. %B Image Processing, 2005. ICIP 2005. IEEE International Conference on %V 2 %P II - 394-7 - II - 394-7 %8 2005/09// %G eng %R 10.1109/ICIP.2005.1530075 %0 Journal Article %J Environmental Microbiology %D 2005 %T Transcriptional profiling of the hyperthermophilic methanarchaeon Methanococcus jannaschii in response to lethal heat and non‐lethal cold shock %A Boonyaratanakornkit,Boonchai B %A Simpson,Anjana J. %A Whitehead,Timothy A %A Fraser,Claire M. %A El‐Sayed, Najib M. %A Clark,Douglas S %X Temperature shock of the hyperthermophilic methanarchaeon Methanococcus jannaschii from its optimal growth temperature of 85°C to 65°C and 95°C resulted in different transcriptional responses characteristic of both the direction of shock (heat or cold shock) and whether the shock was lethal. Specific outcomes of lethal heat shock to 95°C included upregulation of genes encoding chaperones, and downregulation of genes encoding subunits of the H+ transporting ATP synthase. A gene encoding an α subunit of a putative prefoldin was also upregulated, which may comprise a novel element in the protein processing pathway in M. jannaschii. Very different responses were observed upon cold shock to 65°C. These included upregulation of a gene encoding an RNA helicase and other genes involved in transcription and translation, and upregulation of genes coding for proteases and transport proteins. Also upregulated was a gene that codes for an 18 kDa FKBP-type PPIase, which may facilitate protein folding at low temperatures. Transcriptional profiling also revealed several hypothetical proteins that respond to temperature stress conditions. %B Environmental Microbiology %V 7 %P 789 - 797 %8 2005/06/01/ %@ 1462-2920 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1462-2920.2005.00751.x/abstract %N 6 %R 10.1111/j.1462-2920.2005.00751.x %0 Conference Paper %B IEEE Symposium on Information Visualization, 2005. INFOVIS 2005 %D 2005 %T Turning information visualization innovations into commercial products: lessons to guide the next success %A Shneiderman, Ben %A Rao,R. %A Andrews,K. %A Ahlberg,C. %A Brodbeck,D. %A Jewitt,T. %A Mackinlay,J. %K Books %K commercial development %K commercial product %K Computer interfaces %K Computer science %K data visualisation %K Data visualization %K Educational institutions %K exploratory data analysis %K information visualization innovation %K information visualization tool %K innovation management %K Laboratories %K Management training %K new technology emergence %K Technological innovation %K technology transfer %K Turning %K User interfaces %X As information visualization matures as an academic research field, commercial spinoffs are proliferating, but success stories are harder to find. This is the normal process of emergence for new technologies, but the panel organizers believe that there are certain strategies that facilitate success. To teach these lessons, we have invited several key figures who are seeking to commercialize information visualization tools. The panelists make short presentations, engage in a moderated discussion, and respond to audience questions. %B IEEE Symposium on Information Visualization, 2005. INFOVIS 2005 %I IEEE %P 241 - 244 %8 2005/10/23/25 %@ 0-7803-9464-X %G eng %R 10.1109/INFVIS.2005.1532153 %0 Journal Article %J Electronic Notes in Theoretical Computer Science %D 2005 %T A Type and Effect System for Flexible Abstract Interpretation of Java: (Extended Abstract) %A Skalka, Christian %A Smith, Scott %A David Van Horn %K language security %K object oriented languages %K Type analysis %X This paper describes a flexible type and effect inference system for Featherweight Java (FJ). The effect terms generated by static type and effect inference embody the abstract interpretation of pro- gram event sequences. Flexibility in the analysis is obtained by post-processing of inferred effects, allowing a modular adaptation to extensions of the language. Several example transformations are discussed, including how inferred effects can be transformed to reflect the impact of exceptions on FJ control flow. %B Electronic Notes in Theoretical Computer Science %V 131 %P 111 - 124 %8 2005/05/24/ %@ 1571-0661 %G eng %U http://www.sciencedirect.com/science/article/pii/S1571066105002628 %! Electronic Notes in Theoretical Computer Science %0 Journal Article %J Proceedings of the MSE2005 track of the Association for Computational Linguistics Workshop on Intrinsic and Extrinsic Evaluation Meatures for MT and/or Summarization, Ann Arbor, MI %D 2005 %T Umd/bbn at mse2005 %A Zajic, David %A Dorr, Bonnie J %A Jimmy Lin %A Schwartz,R. %X We implemented an initial applicationof a sentence-trimming approach (Trim- mer) to the problem of multi-document summarization in the MSE2005 task. Sentence trimming was incorporated into a feature-based summarization sys- tem, called Multi-Document Trimmer (MDT), by using sentence trimming as both a pre-processing stage and a fea- ture for sentence ranking. We demon- strate that we were able to port Trimmer easily to this new problem, although the impact of sentence trimming was min- imal compared to other features used in the system. The performance of our sys- tem in the official MSE2005 task was around the middle of the pack (16 out of 27). After some minor bug fixes and a simple correction (dateline removal) we obtained an improvement on a post-hoc run on the test data. %B Proceedings of the MSE2005 track of the Association for Computational Linguistics Workshop on Intrinsic and Extrinsic Evaluation Meatures for MT and/or Summarization, Ann Arbor, MI %8 2005/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Understanding Clusters in Multidimensional Spaces: Making Meaning by Combining Insights from Coordinated Views of Domain Knowledge (2004) %A Seo,Jinwook %A Shneiderman, Ben %K Technical Report %X Cluster analysis of multidimensional data is widely used in many research areas including financial, economical, sociological, and biological analyses. Finding natural subclasses in a data set not only reveals interesting patterns but also serves as a basis for further analyses. One of the troubles with cluster analysis is that evaluating how interesting a clustering result is to researchers is subjective, application-dependent, and even difficult to measure. This problem generally gets worse as dimensionality and the number of items grows. The remedy is to enable researchers to apply domain knowledge to facilitate insight about the significance of the clustering result. This article presents a way to better understand a clustering result by combining insights from two interactively coordinated visual displays of domain knowledge. The first is a parallel coordinates view powered by a direct-manipulation search. The second is a domain knowledge view containing a well-understood and meaningful tabular or hierarchical information for the same data set. Our examples depend on hierarchical clustering of gene expression data, coordinated with a parallel coordinates view and with the gene annotation and gene ontology. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6512 %0 Report %D 2005 %T Universal Usability as a Stimulus to Advanced Interface Design (2001) %A Shneiderman, Ben %A Hochheiser,Harry %K Technical Report %X The desire to make computing available to broader populations has historically been a motivation for research and innovation that led to new breakthroughs in usability. Menus, graphical user interfaces, and the World-Wide Web, are examples of innovative technological solutions that have arisen out of the challenge of bringing larger and more diverse groups of users into the world of computing. Universal usability is the latest such challenge: In order to build systems that are universally usable, designers must account for technology variety, user diversity, and gaps in user knowledge. These issues are particularly challenging and important in the context of increasing the usability of the World-Wide-Web. To raise awareness, we urge web designers to provide universal usability statements, to provide users with useful information about the usability of their sites. These statements can inform users and thereby reduce frustration and confusion. Further steps toward universal usability might be achieved through research aimed at developing tools that would encourage or promote usability. We close with five proposals for future research. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6485 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T User Frustration with Technology in the Workplace (2004) %A Lazar,Jonathan %A Jones,Adam %A Bessiere,Katie %A Ceaparu,Irina %A Shneiderman, Ben %K Technical Report %X When hard to use computers cause users to become frustrated, it can affect workplace productivity, user mood, and interactions with other co-workers. Previous research has examined the frustration that graduate students and their families face in using computers. To learn more about the causes and effects of user frustration with computers in the workplace, we collected modified time diaries from 50 workplace users, who spent an average of 5.1 hours on the computer. In this experiment, users reported wasting on average, 42-43% of their time on the computer due to frustrating experiences. The causes of the frustrating experiences, the time lost due to the frustrating experiences, and the effects of the frustrating experiences on the mood of the users are discussed in this paper. Implications for designers, managers, users, information technology staff, and policymakers are discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6515 %0 Journal Article %J Institute for Systems Research Technical Reports %D 2005 %T Using Categorical Information in Multidimensional Data Sets: Interactive Partition and Cluster Comparison %A Seo,Jinwook %A Shneiderman, Ben %K Societal Infrastructure Systems %X Multidimensional data sets often include categorical information. When most columns have categorical information, clustering the data set by similarity of categorical values can reveal interesting patterns in the data set. However, when the data set includes only a small number (one or two) of categorical columns, the categorical information is probably more useful as a way to partition the data set. For example, researchers might be interested in gene expression data for healthy vs. diseased patients or stock performance for common, preferred, or convertible shares. For these cases, we present a novel way to utilize the categorical information together with clustering algorithms. Instead of incorporating categorical information into the clustering process, we can partition the data set according to categorical information. Clustering is then performed with each subset to generate two or more clustering results, each of which is homogeneous (i.e. only includes the same categorical value for the categorical column). By comparing the partitioned clustering results, users can get meaningful insights into the data set: users can identify an interesting group of items that are differentially/similarly expressed in two different homogeneous partitions. The partition can be done in two different directions: (1) by rows if categorical information is available for each column (e.g. some columns are from disease samples and other columns are from healthy samples) or (2) by a column if a column contains categorical information (e.g. a column represents a categorical attribute such as colors or sex). We designed and implemented an interface to facilitate this interactive partition-based clustering results comparison. Coordination between clustering results displays and comparison results overview enables users to identify interesting clusters, and a simple grid display clearly reveals correspondence between two clusters. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6568 %0 Conference Paper %B Proceedings of the 2005 ACM/IEEE conference on Supercomputing %D 2005 %T Using Dynamic Tracing Sampling to Measure Long Running Programs %A Odom, Jeffrey %A Hollingsworth, Jeffrey K %A DeRose,Luiz %A Ekanadham, Kattamuri %A Sbaraglia, Simone %K data communications %K design %K experimentation %K measurement %K performance %K tracing %X Detailed cache simulation can be useful to both system developers and application writers to understand an application's performance. However, measuring long running programs can be extremely slow. In this paper we present a technique to use dynamic sampling of trace snippets throughout an application's execution. We demonstrate that our approach improves accuracy compared to sampling a few timesteps at the beginning of execution by judiciously choosing the frequency, as well as the points in the control flow, at which samples are collected. Our approach is validated using the SIGMA tracing and simulation framework for the IBM Power family of processors. %B Proceedings of the 2005 ACM/IEEE conference on Supercomputing %S SC '05 %I IEEE Computer Society %P 59– - 59– %8 2005/// %@ 1-59593-061-2 %G eng %U http://dx.doi.org/10.1109/SC.2005.77 %R http://dx.doi.org/10.1109/SC.2005.77 %0 Conference Paper %B Advanced Video and Signal Based Surveillance, 2005. AVSS 2005. IEEE Conference on %D 2005 %T VidMAP: video monitoring of activity with Prolog %A Shet,V. D %A Harwood,D. %A Davis, Larry S. %K activities %K algorithms; %K based %K Computer %K computerised %K engine; %K higher %K image %K level %K Logic %K monitoring; %K multicamera %K processing; %K programming; %K Prolog %K PROLOG; %K reasoning %K recognition; %K scenario; %K signal %K streaming; %K streams; %K Surveillance %K surveillance; %K system; %K video %K VISION %K vision; %K visual %X This paper describes the architecture of a visual surveillance system that combines real time computer vision algorithms with logic programming to represent and recognize activities involving interactions amongst people, packages and the environments through which they move. The low level computer vision algorithms log primitive events of interest as observed facts, while the higher level Prolog based reasoning engine uses these facts in conjunction with predefined rules to recognize various activities in the input video streams. The system is illustrated in action on a multi-camera surveillance scenario that includes both security and safety violations. %B Advanced Video and Signal Based Surveillance, 2005. AVSS 2005. IEEE Conference on %P 224 - 229 %8 2005/09// %G eng %R 10.1109/AVSS.2005.1577271 %0 Report %D 2005 %T Visual Specification of Queries for Finding Patterns in Time-Series Data (2001) %A Hochheiser,Harry %A Shneiderman, Ben %K Technical Report %X Widespread interest in discovering features and trends in time- series has generated a need for tools that support interactive exploration.This paper introduces timeboxes: a powerful graphical, directmanipulation metaphor for the specification of queries over time-series datasets. Our TimeFinder implementation of timeboxes supports interactive formulation and modification of queries, thus speeding the process of exploring time-series data sets and guiding data mining. TimeFinder includes windows for timebox queries, individual time-series, and details-on-demand. Other features include drag-and-drop support for query-by-example and graphical envelopes for displaying the extent of the entire data set and result set from a given query. Extensions involving increased expressive power and general temporal data sets are discussed. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6479 %0 Report %D 2005 %T Visualization Methods for Personal Photo Collections: Browsing and Searching in the PhotoFinder (2000) %A Kang,Hyunmo %A Shneiderman, Ben %K Technical Report %X Software tools for personal photo collection management are proliferating, but they usually have limited searching and browsing functions. We implemented the PhotoFinder prototype to enable non-technical users of personal photo collections to search and browse easily. PhotoFinder provides a set of visual Boolean query interfaces, coupled with dynamic query and query preview features. It gives users powerful search capabilities. Using a scatter plot thumbnail display and drag-and-drop interface, PhotoFinder is designed to be easy to use for searching and browsing photos. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6472 %0 Report %D 2005 %T Why Not Make Interfaces Better than 3D Reality? (2004) %A Shneiderman, Ben %K Technical Report %X Some designers dream about building interfaces that approach the richness of 3D reality. They believe that the closer an interface resembles the real world, the easier the usage. They strive for resolution that matches filmith rapid camera movement and lively animated objects. This is a dubious proposition since user studies show that disorienting navigation, complex user actions, and annoying occlusions can slow performance in the real world as well as 3D interfaces. %B Institute for Systems Research Technical Reports %8 2005/// %G eng %U http://drum.lib.umd.edu/handle/1903/6523 %0 Journal Article %J Proc. of Eighth International Symposium on the %D 2004 %T 3d modelling of human motion using kinematic chains and multiple cameras for tracking %A Sundaresan, A. %A RoyChowdhury, A. %A Chellapa, Rama %B Proc. of Eighth International Symposium on the %8 2004/// %G eng %0 Journal Article %J Journal of industrial microbiology & biotechnology %D 2004 %T Acinetobacter lipases: molecular biology, biochemical properties and biotechnological potential %A Snellman,E. A. %A Rita R Colwell %X Lipases (EC 3.1.1.3) have received increased attention recently, evidenced by the increasing amount of information about lipases in the current literature. The renewed interest in this enzyme class is due primarily to investigations of their role in pathogenesis and their increasing use in biotechnological applications [38]. Also, many microbial lipases are available as commercial products, the majority of which are used in detergents, cosmetic production, food flavoring, and organic synthesis. Lipases are valued biocatalysts because they act under mild conditions, are highly stable in organic solvents, show broad substrate specificity, and usually show high regio- and/or stereo-selectivity in catalysis. A number of lipolytic strains of Acinetobacter have been isolated from a variety of sources and their lipases possess many biochemical properties similar to those that have been developed for biotechnological applications. This review discusses the biology of lipase expression in Acinetobacter, with emphasis on those aspects relevant to potential biotechnology applications. %B Journal of industrial microbiology & biotechnology %V 31 %P 391 - 400 %8 2004/// %G eng %N 9 %R 10.1007/s10295-004-0167-0 %0 Journal Article %J International Journal of Image and Graphics %D 2004 %T Active deformable models using density estimation %A Abd-Almageed, Wael %A Smith,C.E. %B International Journal of Image and Graphics %V 4 %P 343 - 362 %8 2004/// %G eng %N 3 %0 Journal Article %J Proceedings of the AAAI 2004 Fall Symposium on Artificial Multi-agent Learning %D 2004 %T Adapting network structure for efficient team formation %A Gaston,M. %A Simmons,J. %A desJardins, Marie %B Proceedings of the AAAI 2004 Fall Symposium on Artificial Multi-agent Learning %8 2004/// %G eng %0 Conference Paper %B Distributed Computing Systems, 2004. Proceedings. 24th International Conference on %D 2004 %T Adaptive replication in peer-to-peer systems %A Gopalakrishnan,V. %A Silaghi,B. %A Bhattacharjee, Bobby %A Keleher,P. %K adaptive %K allocation; %K data %K databases; %K decentralized %K delivery %K distributed %K LAR %K low-latency %K peer-to-peer %K processing; %K protocol; %K replicated %K replication %K resource %K strategies; %K structured %K system-neutral %K system; %K systems; %X Peer-to-peer systems can be used to form a low-latency decentralized data delivery system. Structured peer-to-peer systems provide both low latency and excellent load balance with uniform query and data distributions. Under the more common skewed access distributions, however, individual nodes are easily overloaded, resulting in poor global performance and lost messages. This paper describes a lightweight, adaptive, and system-neutral replication protocol, called LAR, that maintains low access latencies and good load balance even under highly skewed demand. We apply LAR to Chord and show that it has lower overhead and better performance than existing replication strategies. %B Distributed Computing Systems, 2004. Proceedings. 24th International Conference on %P 360 - 369 %8 2004/// %G eng %R 10.1109/ICDCS.2004.1281601 %0 Journal Article %J Multimedia Tools and Applications %D 2004 %T An algebra for powerpoint sources %A Fayzullin,M. %A V.S. Subrahmanian %X There are now millions of PowerPoint documents available within corporate intranets and/or over the Internet. In this paper, we develop a formal model of PowerPoint databases. We propose a relational style algebra called pptA (PowerPoint Algebra) to query PowerPoint databases. The algebra contains some new operators (such as the APPLY operator that changes properties of objects, slides and presentations) as well as interesting twists on relational operators (e.g. join and cartesian product allow different entities being joined together to share attributes whose values may be merged). We prove a set of equivalence results within this algebra. We have implemented a version of pptA—the paper provides a cost model and experimental results on the conditions under which these equivalences are useful. %B Multimedia Tools and Applications %V 24 %P 273 - 301 %8 2004/// %G eng %N 3 %R 10.1023/B:MTAP.0000039422.87260.52 %0 Conference Paper %B Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on %D 2004 %T Appearance-based tracking and recognition using the 3D trilinear tensor %A Jie Shao %A Zhou,S. K %A Chellapa, Rama %K 3D %K adaptive %K affine-transformation %K airborne %K algorithm; %K appearance %K appearance-based %K based %K estimation; %K geometrical %K image %K mathematical %K novel %K object %K operator; %K operators; %K perspective %K prediction; %K processing; %K recognition; %K representation; %K signal %K structure %K synthesis; %K template %K tensor %K tensor; %K tensors; %K tracking; %K transformation; %K trilinear %K updating; %K video %K video-based %K video; %K view %X The paper presents an appearance-based adaptive algorithm for simultaneous tracking and recognition by generalizing the transformation model to 3D perspective transformation. A trilinear tensor operator is used to represent the 3D geometrical structure. The tensor is estimated by predicting the corresponding points using the existing affine-transformation based algorithm. The estimated tensor is used to synthesize novel views to update the appearance templates. Some experimental results using airborne video are presented. %B Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on %V 3 %P iii - 613-16 vol.3 - iii - 613-16 vol.3 %8 2004/05// %G eng %R 10.1109/ICASSP.2004.1326619 %0 Journal Article %J IEEE Software %D 2004 %T Applying Model-based Distributed Continuous Quality Assurance Processes to Enhance Persistent Software Attributes %A Krishna,A. S %A Yilmaz,C. %A Memon, Atif M. %A Porter, Adam %A Schmidt,D. C %A Gokhale,A. %A Natarajan,B. %X Time and resource constraints often force developers of highlyconfigurable quality of service (QoS)-intensive software sys- tems to guarantee their system’s persistent software attributes (PSAs) (e.g., functional correctness, portability, efficiency, and QoS) on very few platform configurations and to extrapolate from these configurations to the entire configuration space, which allows many sources of degradation to escape detec- tion until systems are fielded. This article illustrates how distributed continuous quality assurance (DCQA) processes help improve the assessment of these PSAs across large QoS- intensive software system configuration spaces. We also il- lustrate how model-based DCQA processes enable developers to run formally-designed screening experiments that isolate the most significant configuration options, such as different workload parameters, operating system, compiler flags, fea- ture sets, and/or run-time optimization controls. Our empir- ical results show that DCQA processes can be used monitor, safeguard, and enforce PSAs at an acceptable level of cost and effort. %B IEEE Software %V 21 %P 32 - 40 %8 2004/// %G eng %N 6 %0 Journal Article %J Journal of Algorithms %D 2004 %T Approximation algorithms for partial covering problems %A Gandhi,Rajiv %A Khuller, Samir %A Srinivasan, Aravind %K Approximation algorithms %K Partial covering %K Primal-dual methods %K Randomized rounding %K Set cover %K Vertex cover %X We study a generalization of covering problems called partial covering. Here we wish to cover only a desired number of elements, rather than covering all elements as in standard covering problems. For example, in k-partial set cover, we wish to choose a minimum number of sets to cover at least k elements. For k-partial set cover, if each element occurs in at most f sets, then we derive a primal-dual f-approximation algorithm (thus implying a 2-approximation for k-partial vertex cover) in polynomial time. Without making any assumption about the number of sets an element is in, for instances where each set has cardinality at most three, we obtain an approximation of 4/3. We also present better-than-2-approximation algorithms for k-partial vertex cover on bounded degree graphs, and for vertex cover on expanders of bounded average degree. We obtain a polynomial-time approximation scheme for k-partial vertex cover on planar graphs, and for covering k points in Rd by disks. %B Journal of Algorithms %V 53 %P 55 - 84 %8 2004/10// %@ 0196-6774 %G eng %U http://www.sciencedirect.com/science/article/pii/S0196677404000689 %N 1 %R 10.1016/j.jalgor.2004.04.002 %0 Journal Article %J Proceedings of the HLT-NAACL 2004 Document Understanding Workshop, Boston %D 2004 %T Bbn/umd at duc-2004: Topiary %A Zajic, David %A Dorr, Bonnie J %A Schwartz,R. %X This paper reports our results at DUC-2004 and describes our approach, im- plemented in a system called Topiary. We will show that the combination of linguistically motivated sentence com- pression with statistically selected topic terms performs better than either alone, according to some automatic summary evaluation measures. %B Proceedings of the HLT-NAACL 2004 Document Understanding Workshop, Boston %P 112 - 119 %8 2004/// %G eng %0 Conference Paper %B Proc. 3rd Workshop on Non-Slicon Computation, held in conjunction with the 31st International Symposium on Computer Architecture (ISCA 2004) %D 2004 %T Bending light for multi-chip virtual PRAMs? %A Vishkin, Uzi %A Smolyaninov,I. %X A new paradigm for an all-to-alloptical interconnection network is presented. An interesting modeling aspect is that (limited) bending of optical communication channels is allowed. In a computer system, the paradigm could provide part of an interconnection fabric between several tens (e.g., 64) of chips comprising parallel processing elements and the first level of the cache. An optical interconnection network raises an intriguing possibility: obtain both improved performance and significant cost reduction with respect to standard serial computer system models. %B Proc. 3rd Workshop on Non-Slicon Computation, held in conjunction with the 31st International Symposium on Computer Architecture (ISCA 2004) %P 19 - 23 %8 2004/// %G eng %0 Conference Paper %B Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval %D 2004 %T Building an information retrieval test collection for spontaneous conversational speech %A Oard, Douglas %A Soergel,Dagobert %A David Doermann %A Huang,Xiaoli %A Murray,G. Craig %A Wang,Jianqiang %A Ramabhadran,Bhuvana %A Franz,Martin %A Gustman,Samuel %A Mayfield,James %A Kharevych,Liliya %A Strassel,Stephanie %K assessment %K Automatic speech recognition %K oral history %K search-guided relevance %X Test collections model use cases in ways that facilitate evaluation of information retrieval systems. This paper describes the use of search-guided relevance assessment to create a test collection for retrieval of spontaneous conversational speech. Approximately 10,000 thematically coherent segments were manually identified in 625 hours of oral history interviews with 246 individuals. Automatic speech recognition results, manually prepared summaries, controlled vocabulary indexing, and name authority control are available for every segment. Those features were leveraged by a team of four relevance assessors to identify topically relevant segments for 28 topics developed from actual user requests. Search-guided assessment yielded sufficient inter-annotator agreement to support formative evaluation during system development. Baseline results for ranked retrieval are presented to illustrate use of the collection. %B Proceedings of the 27th annual international ACM SIGIR conference on Research and development in information retrieval %S SIGIR '04 %I ACM %C New York, NY, USA %P 41 - 48 %8 2004/// %@ 1-58113-881-4 %G eng %U http://doi.acm.org/10.1145/1008992.1009002 %R 10.1145/1008992.1009002 %0 Conference Paper %B Proceedings of the ACM SIGCOMM workshop on Future directions in network architecture %D 2004 %T The case for separating routing from routers %A Feamster, Nick %A Balakrishnan,Hari %A Rexford,Jennifer %A Shaikh,Aman %A van der Merwe,Jacobus %K BGP %K interdomain routing %K routing architecture %X Over the past decade, the complexity of the Internet's routing infrastructure has increased dramatically. This complexity and the problems it causes stem not just from various new demands made of the routing infrastructure, but also from fundamental limitations in the ability of today's distributed infrastructure to scalably cope with new requirements.The limitations in today's routing system arise in large part from the fully distributed path-selection computation that the IP routers in an autonomous system (AS) must perform. To overcome this weakness, interdomain routing should be separated from today's IP routers, which should simply forward packets (for the most part). Instead, a separate Routing Control Platform (RCP) should select routes on behalf of the IP routers in each AS and exchange reachability information with other domains.Our position is that an approach like RCP is a good way of coping with complexity while being responsive to new demands and can lead to a routing system that is substantially easier to manage than today. We present a design overview of RCP based on three architectural principles path computation based on a consistent view of network state, controlled interactions between routing protocol layers, and expressive specification of routing policies and discuss the architectural strengths and weaknesses of our proposal. %B Proceedings of the ACM SIGCOMM workshop on Future directions in network architecture %S FDNA '04 %I ACM %C New York, NY, USA %P 5 - 12 %8 2004/// %@ 1-58113-942-X %G eng %U http://doi.acm.org/10.1145/1016707.1016709 %R 10.1145/1016707.1016709 %0 Journal Article %J Briefings in Bioinformatics %D 2004 %T Comparative Genome Assembly %A Pop, Mihai %A Phillippy,Adam %A Delcher,Arthur L. %A Salzberg,Steven L. %K Assembly %K comparative genomics %K open source %K shotgun sequencing %X One of the most complex and computationally intensive tasks of genome sequence analysis is genome assembly. Even today, few centres have the resources, in both software and hardware, to assemble a genome from the thousands or millions of individual sequences generated in a whole-genome shotgun sequencing project. With the rapid growth in the number of sequenced genomes has come an increase in the number of organisms for which two or more closely related species have been sequenced. This has created the possibility of building a comparative genome assembly algorithm, which can assemble a newly sequenced genome by mapping it onto a reference genome.We describe here a novel algorithm for comparative genome assembly that can accurately assemble a typical bacterial genome in less than four minutes on a standard desktop computer. The software is available as part of the open-source AMOS project. %B Briefings in Bioinformatics %V 5 %P 237 - 248 %8 2004/09/01/ %@ 1467-5463, 1477-4054 %G eng %U http://bib.oxfordjournals.org/content/5/3/237 %N 3 %R 10.1093/bib/5.3.237 %0 Conference Paper %B Scientific and Statistical Database Management, International Conference on %D 2004 %T A Comparative Study of Spatial Indexing Techniques for Multidimensional Scientific Datasets %A Nam,Beomseok %A Sussman, Alan %X Scientific applications that query into very large multi-dimensional datasets are becoming more common. These datasets are growing in size every day, and are becoming truly enormous, making it infeasible to index individual data elements. We have instead been experimenting with chunking the datasets to index them, grouping data elements into small chunks of a fixed, but dataset-specific, size to take advantage of spatial locality. While spatial indexing structures based on R-trees perform reasonably well for the rectangular bounding boxes of such chunked datasets, other indexing structures based on KDB-trees, such as Hybrid trees, have been shown to perform very well for point data. In this paper, we investigate how all these indexing structures perform for multidimensional scientific datasets, and compare their features and performance with that of SH-trees, an extension of Hybrid trees, for indexing multi-dimensional rectangles. Our experimental results show that the algorithms for building and searching SH-trees outperform those for R-trees, R*-trees, and X-trees for both real application and synthetic datasets and queries. We show that the SH-tree algorithms perform well for both low and high dimensional data, and that they scale well to high dimensions both for building and searching the trees. %B Scientific and Statistical Database Management, International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %P 171 - 171 %8 2004/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/SSDM.2004.1311209 %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2004. IEEE %D 2004 %T Computation of Green's function for finite-size photonic crystals by boundary element method %A Seydou,F. %A Ramahi,O. %A Duraiswami, Ramani %A Seppdnen,T. %K 2D %K boundary %K boundary-elements %K circular %K crystals; %K cylinders; %K electromagnetic %K element %K equation %K equations; %K expansions; %K finite-size %K function %K function; %K Green %K Green's %K integral %K method; %K methods; %K multipole %K photonic %K propagation; %K wave %X We have derived a new integral equation method for computing the Green's function for 2D electromagnetic wave propagation in a finite size photonic crystal. The method is also used for deriving the multipole expansions for the case of circular cylinders. This derivation is different from the ones in the literature. Our numerical results show an excellent agreement of the two algorithms. In the future we would like to expand the method to the numerical solution of the three-dimensional electromagnetic problem. %B Antennas and Propagation Society International Symposium, 2004. IEEE %V 4 %P 4320 - 4323 Vol.4 - 4320 - 4323 Vol.4 %8 2004/06// %G eng %R 10.1109/APS.2004.1330307 %0 Journal Article %J Applied Mathematics and Computation %D 2004 %T Computational experience with exterior point algorithms for the transportation problem %A Charalampos Papamanthou %A Paparrizos, Konstantinos %A Samaras, Nikolaos %K Algorithm evaluation %K Experimental computational study %K Exterior point simplex algorithms %K Network simplex algorithm %K Transportation problem %X An experimental computational study to compare the classical primal simplex algorithm and the exterior point algorithms for the transportation problem (TP) is presented. Totally, four algorithms are compared on uniformly randomly generated test problems. The results are very encouraging for one of the competitive algorithms. In particular, a dual forest exterior point algorithm is on average up to 4.5 times faster than network simplex algorithm on TPs of size 300 × 300 and for all classes. This result leads into corresponding savings in computational time. From the computational performance we conclude that as the problem size increases, exterior point algorithm get relatively faster. %B Applied Mathematics and Computation %V 158 %P 459 - 475 %8 2004/11/05/ %@ 0096-3003 %G eng %U http://www.sciencedirect.com/science/article/pii/S0096300303010294 %N 2 %! Applied Mathematics and Computation %0 Conference Paper %B Proceedings of the twentieth annual symposium on Computational geometry %D 2004 %T A computational framework for incremental motion %A Mount, Dave %A Netanyahu,Nathan S. %A Piatko,Christine D. %A Silverman,Ruth %A Wu,Angela Y. %K competitive analysis %K incremental motion %K kinetic data structures %X We propose a generic computational framework for maintaining a discrete geometric structure defined by a collection of static and mobile objects. We assume that the mobile objects move incrementally, that is, in discrete time steps. We assume that the structure to be maintained is a function of the current locations of the mobile and static objects (independent of their prior motion). Unlike other models for kinetic computation, we place no restrictions on the motion nor on its predictability.In order to handle unrestricted incremental motion, our framework is based on the coordination of two computational entities. The first is the incremental motion algorithm. It is responsible for maintaining the structure and a set of certificates, or conditions, that prove the structure's correctness. The other entity, called the motion processor, is responsible for handling all the low-level aspects of motion, including computing and/or tracking the motion of the mobile objects, answering queries about their current positions and velocities, and validating that the object motions satisfy simple motion estimates, which are generated by the incremental motion algorithm. Computational efficiency is measured in terms of the number of interactions between these two entities. %B Proceedings of the twentieth annual symposium on Computational geometry %S SCG '04 %I ACM %C New York, NY, USA %P 200 - 209 %8 2004/// %@ 1-58113-885-7 %G eng %U http://doi.acm.org/10.1145/997817.997849 %R 10.1145/997817.997849 %0 Conference Paper %B Shape Modeling Applications, 2004. Proceedings %D 2004 %T Constant-time navigation in four-dimensional nested simplicial meshes %A Lee,M. %A De Floriani, Leila %A Samet, Hanan %K 4-dimensional %K adaptive %K algorithm; %K approximations; %K bit %K codes; %K computational %K constant-time %K continuous %K decomposition; %K face-adjacent %K faces; %K fields; %K finding %K following; %K four-dimensional %K generation; %K geometry; %K hierarchy; %K hypercube; %K location %K manipulation %K mesh %K meshes; %K multiresolution %K navigation; %K neighbor %K nested %K operations; %K pentatopes; %K pointer %K recursive %K representation; %K scalar %K simplexes; %K simplicial %K tetrahedral %X We consider a recursive decomposition of a four-dimensional hypercube into a hierarchy of nested 4-dimensional simplexes, that we call pentatopes. The paper presents an algorithm for finding the neighbors of a pentatope along its five tetrahedral faces in constant time. To this aim, we develop a labeling technique for nested pentatopes that enables their identification by using location codes. The constant-time behavior is achieved through bit manipulation operations, thus avoiding traversing the simplicial hierarchy via pointer following. We discuss an application of this representation to multi-resolution representations of four-dimensional scalar fields. Extracting adaptive continuous approximations of the scalar field from such a model requires generating conforming meshes, i.e., meshes in which the pentatopes match along their tetrahedral faces. Our neighbor finding algorithm enables computing face-adjacent pentatopes efficiently. %B Shape Modeling Applications, 2004. Proceedings %P 221 - 230 %8 2004/06// %G eng %R 10.1109/SMI.2004.1314509 %0 Conference Paper %B IEEE Infocom %D 2004 %T A cooperative bulk transfer protocol %A Sherwood,R. %A Braud,R. %A Bhattacharjee, Bobby %B IEEE Infocom %8 2004/// %G eng %0 Journal Article %J Algorithms–ESA 2004 %D 2004 %T Data migration on parallel disks %A Golubchik,L. %A Khuller, Samir %A Kim,Y. A %A Shargorodskaya,S. %A Wan,Y. C %X Our work is motivated by the problem of managing data on storage devices, typically a set of disks. Such storage servers are used as web servers or multimedia servers, for handling high demand for data. As the system is running, it needs to dynamically respond to changes in demand for different data items. There are known algorithms for mapping demand to a layout. When the demand changes, a new layout is computed. In this work we study the data migration problem, which arises when we need to quickly change one layout to another. This problem has been studied earlier when for each disk the new layout has been prescribed. However, to apply these algorithms effectively, we identify another problem that we refer to as the correspondence problem, whose solution has a significant impact on the solution for the data migration problem. We study algorithms for the data migration problem in more detail and identify variations of the basic algorithm that seem to improve performance in practice, even though some of the variations have poor worst case behavior. %B Algorithms–ESA 2004 %P 689 - 701 %8 2004/// %G eng %R 10.1007/978-3-540-30140-0_61 %0 Journal Article %J ACM Trans. Database Syst. %D 2004 %T Decoupling partitioning and grouping: Overcoming shortcomings of spatial indexing with bucketing %A Samet, Hanan %K BV-trees %K decoupling %K object hierarchies %K PK-trees %K R-trees %K space decomposition %K spatial indexing %X The principle of decoupling the partitioning and grouping processes that form the basis of most spatial indexing methods that use tree directories of buckets is explored. The decoupling is designed to overcome the following drawbacks of traditional solutions:(1) multiple postings in disjoint space decomposition methods that lead to balanced trees such as the hB-tree where a node split in the event of node overflow may be such that one of the children of the node that was split becomes a child of both of the nodes resulting from the split;(2) multiple coverage and nondisjointness of methods based on object hierarchies such as the R-tree which lead to nonunique search paths;(3) directory nodes with similarly-shaped hyper-rectangle bounding boxes with minimum occupancy in disjoint space decomposition methods such as those based on quadtrees and k-d trees that make use of regular decomposition.The first two drawbacks are shown to be overcome by the BV-tree where as a result of decoupling the partitioning and grouping processes, the union of the regions associated with the nodes at a given level of the directory does not necessarily contain all of the data points although all searches take the same amount of time. The BV-tree is not plagued by the third drawback. The third drawback is shown to be overcome by the PK-tree where the grouping process is based on ensuring that every node has at least k objects or blocks. The PK-tree is not plagued by the first two drawbacks as they are inapplicable to it. In both cases, the downside of decoupling the partitioning and grouping processes is that the resulting structure is not necessarily balanced, although, since the nodes have a relatively large fanout, the deviation from a balanced structure is relatively small. %B ACM Trans. Database Syst. %V 29 %P 789 - 830 %8 2004/12// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/1042046.1042052 %N 4 %R 10.1145/1042046.1042052 %0 Journal Article %J Interactions %D 2004 %T Designing for fun: how can we design user interfaces to be more fun? %A Shneiderman, Ben %B Interactions %V 11 %P 48 - 50 %8 2004/09// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/1015530.1015552 %N 5 %R 10.1145/1015530.1015552 %0 Journal Article %J Wireless Communications, IEEE %D 2004 %T Designing secure sensor networks %A Elaine Shi %A Perrig, A. %K Pervasive computing %K secure sensor network design %K telecommunication security %K Ubiquitous Computing %K wireless sensor network %K Wireless sensor networks %X Sensor networks are expected to play an essential role in the upcoming age of pervasive computing. Due to their constraints in computation, memory, and power resources, their susceptibility to physical capture, and use of wireless communications, security is a challenge in these networks. The scale of deployments of wireless sensor networks require careful decisions and trade-offs among various security measures. The authors discuss these issues and consider mechanisms to achieve secure communication in these networks. %B Wireless Communications, IEEE %V 11 %P 38 - 43 %8 2004 %@ 1536-1284 %G eng %N 6 %0 Book %D 2004 %T Designing the user interface: Strategies for Effective Human-Computer Interaction %A Plaisant, Catherine %A Schneiderman,B. %I Pearson Higher Education %8 2004/// %G eng %0 Journal Article %J International Journal of Human-Computer Interaction %D 2004 %T Determining Causes and Severity of End-User Frustration %A Ceaparu,Irina %A Lazar,Jonathan %A Bessiere,Katie %A Robinson,John %A Shneiderman, Ben %X Although computers are beneficial to individuals and society, frequently users encounter frustrating experiences when using computers. This study represents an attempt to measure, in 111 participants, the frequency, the cause, and the level of severity of frustrating experiences. The data show that frustrating experiences happen on a frequent basis. The applications in which the frustrating experiences happened most frequently were Web browsing, e-mail, and word processing. The most-cited causes of frustrating experiences were error messages, dropped network connections, long download times, and hard-to-find features. The time lost due to frustrating experiences ranged from 47% to 53% of time spent on a computer, depending on the location and study method. After extreme cases were discarded, the time lost was still above 38%. These disturbing results should be a basis for future study.Although computers are beneficial to individuals and society, frequently users encounter frustrating experiences when using computers. This study represents an attempt to measure, in 111 participants, the frequency, the cause, and the level of severity of frustrating experiences. The data show that frustrating experiences happen on a frequent basis. The applications in which the frustrating experiences happened most frequently were Web browsing, e-mail, and word processing. The most-cited causes of frustrating experiences were error messages, dropped network connections, long download times, and hard-to-find features. The time lost due to frustrating experiences ranged from 47% to 53% of time spent on a computer, depending on the location and study method. After extreme cases were discarded, the time lost was still above 38%. These disturbing results should be a basis for future study. %B International Journal of Human-Computer Interaction %V 17 %P 333 - 356 %8 2004/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1207/s15327590ijhc1703_3 %N 3 %R 10.1207/s15327590ijhc1703_3 %0 Journal Article %J Computer Systems: Architectures, Modeling, and Simulation %D 2004 %T DIF: An interchange format for dataflow-based design tools %A Hsu,C. J %A Keceli,F. %A Ko,M. Y %A Shahparnia,S. %A Bhattacharyya, Shuvra S. %B Computer Systems: Architectures, Modeling, and Simulation %P 3 - 32 %8 2004/// %G eng %0 Conference Paper %B Communications, 2004 IEEE International Conference on %D 2004 %T Distortion management of real-time MPEG-4 video over downlink multicode CDMA networks %A Su,Guan-Ming %A Han,Zhu %A Kwasinski,A. %A M. Wu %A Liu,K. J.R %A Farvardin,N. %K access; %K adaptation; %K allocation; %K CDMA %K channel %K code %K coding %K coding; %K combined %K communication; %K compression; %K control; %K data %K distortion %K division %K downlink %K links; %K management; %K MPEG-4 %K multicode %K multiple %K networks; %K power %K radio %K rate %K real-time %K resource %K source %K source-channel %K video %K video; %K visual %X In this paper, a protocol is designed to manage source rate/channel coding rate adaptation, code allocation, and power control to transmit real-time MPEG-4 FGS video over downlink multicode CDMA networks. We develop a fast adaptive scheme of distortion management to reduce the overall distortion received by all users subject, to the limited number of codes and maximal transmitted power. Compared with a modified greedy method in literature, our proposed algorithm can reduce the overall system's distortion by at least 45%. %B Communications, 2004 IEEE International Conference on %V 5 %P 3071 - 3075 Vol.5 - 3071 - 3075 Vol.5 %8 2004/06// %G eng %R 10.1109/ICC.2004.1313096 %0 Conference Paper %B 19th ACM OOPSLA Workshop on Component and Middleware Performance %D 2004 %T A distributed continuous quality assurance process to manage variability in performance-intensive software %A Krishna,A. %A Yilmaz,C. %A Memon, Atif M. %A Porter, Adam %A Schmidt,D. C %A Gokhale,A. %A Natarajan,B. %B 19th ACM OOPSLA Workshop on Component and Middleware Performance %8 2004/// %G eng %0 Journal Article %J Journal of Systems and Software %D 2004 %T Distributed prototyping from validated specifications %A Hansel,David %A Cleaveland, Rance %A Smolka,Scott A. %X We present vpl2cxx, a translator that automatically generates efficient, fully distributed C++ code from high-level system models specified in the mathematically well-founded VPL design language. As the Concurrency Workbench of the New Century (CWB-NC) verification tool includes a front-end for VPL, designers may use the full range of automatic verification and simulation checks provided by this tool on their VPL system designs before invoking the translator, thereby generating distributed prototypes from validated specifications. Besides being fully distributed, the code generated by vpl2cxx is highly readable and portable to a host of execution environments and real-time operating systems (RTOSes). This is achieved by encapsulating all generated code dealing with low-level interprocess communication issues in a library for synchronous communication, which in turn is built upon the adaptive communication environment (ACE) client-server network programming interface. Finally, example applications show that the performance of the generated code is very good, especially for prototyping purposes. We discuss two such examples, including the RETHER real-time Ethernet protocol for voice and video applications. %B Journal of Systems and Software %V 70 %P 275 - 298 %8 2004/03// %@ 0164-1212 %G eng %U http://www.sciencedirect.com/science/article/pii/S0164121203000748 %N 3 %R 10.1016/S0164-1212(03)00074-8 %0 Journal Article %J GeneticsGenetics %D 2004 %T The Drosophila U1-70K Protein Is Required for Viability, but Its Arginine-Rich Domain Is Dispensable %A Salz,Helen K. %A Mancebo,Ricardo S. Y. %A Nagengast,Alexis A. %A Speck,Olga %A Psotka,Mitchell %A Mount, Stephen M. %X The conserved spliceosomal U1-70K protein is thought to play a key role in RNA splicing by linking the U1 snRNP particle to regulatory RNA-binding proteins. Although these protein interactions are mediated by repeating units rich in arginines and serines (RS domains) in vitro, tests of this domain's importance in intact multicellular organisms have not been carried out. Here we report a comprehensive genetic analysis of U1-70K function in Drosophila. Consistent with the idea that U1-70K is an essential splicing factor, we find that loss of U1-70K function results in lethality during embryogenesis. Surprisingly, and contrary to the current view of U1-70K function, animals carrying a mutant U1-70K protein lacking the arginine-rich domain, which includes two embedded sets of RS dipeptide repeats, have no discernible mutant phenotype. Through double-mutant studies, however, we show that the U1-70K RS domain deletion no longer supports viability when combined with a viable mutation in another U1 snRNP component. Together our studies demonstrate that while the protein interactions mediated by the U1-70K RS domain are not essential for viability, they nevertheless contribute to an essential U1 snRNP function. %B GeneticsGenetics %V 168 %P 2059 - 2065 %8 2004/12/01/ %@ 0016-6731, 1943-2631 %G eng %U http://www.genetics.org/content/168/4/2059 %N 4 %R 10.1534/genetics.104.032532 %0 Conference Paper %B Global Telecommunications Conference, 2004. GLOBECOM '04. IEEE %D 2004 %T Dynamic distortion control for 3-D embedded wavelet video over multiuser OFDM networks %A Su,Guan-Ming %A Han,Zhu %A M. Wu %A Liu,K. J.R %K 3D %K 802.11a; %K channels; %K codec; %K codecs; %K communication; %K control; %K deviation; %K distortion %K diversity %K diversity; %K downlink %K dynamic %K embedded %K fairness; %K Frequency %K IEEE %K LAN; %K maximal %K minimax %K minimization; %K modulation; %K multimedia %K multiuser %K OFDM %K OFDM; %K PSNR %K rate %K reception; %K streaming; %K systems; %K techniques; %K theory; %K TIME %K transforms; %K video %K video; %K wavelet %K wireless %X In this paper, we propose a system to transmit multiple 3D embedded wavelet video programs over downlink multiuser OFDM. We consider the fairness among users and formulate the problem as minimizing the users' maximal distortion subject to power, rate, and subcarrier constraints. By exploring frequency, time, and multiuser diversity in OFDM and flexibility of the 3D embedded wavelet video codec, the proposed algorithm can achieve fair video qualities among all users. Compared to a scheme similar to the current multiuser OFDM standard (IEEE 802.11a), the proposed scheme outperforms it by 1-5 dB on the worst received PSNR among all users and has much smaller PSNR deviation. %B Global Telecommunications Conference, 2004. GLOBECOM '04. IEEE %V 2 %P 650 - 654 Vol.2 - 650 - 654 Vol.2 %8 2004/12/03/nov %G eng %R 10.1109/GLOCOM.2004.1378042 %0 Conference Paper %B INFOCOM 2004. Twenty-third AnnualJoint Conference of the IEEE Computer and Communications Societies %D 2004 %T Dynamic join-exit amortization and scheduling for time-efficient group key agreement %A Mao,Yinian %A Sun,Yan %A M. Wu %A Liu,K. J.R %K agreement; %K asymptotic %K average %K communication; %K complexity; %K computational %K cost; %K cryptography; %K dynamic %K group %K join-tree/exit-tree %K key %K logical %K Multicast %K network %K secure %K security; %K special %K Telecommunication %K TIME %K time-efficient %K topology; %K tree; %X We propose a time-efficient contributory key agreement framework for secure communications in dynamic groups. The proposed scheme employs a special join-tree/exit-tree topology in the logical key tree and effectively exploits the efficiency of amortized operations. We derive the optimal parameters and design an activation algorithm for the join and exit trees. We also show that the asymptotic average time cost per user join and leave event is theta;(log (log n)), where n is the group size. Our experiment results on both simulated user activities and the real MBone data have shown that the proposed scheme outperforms the existing tree-based schemes. %B INFOCOM 2004. Twenty-third AnnualJoint Conference of the IEEE Computer and Communications Societies %V 4 %P 2617 - 2627 vol.4 - 2617 - 2627 vol.4 %8 2004/03// %G eng %R 10.1109/INFCOM.2004.1354681 %0 Journal Article %J Information VisualizationInformation Visualization %D 2004 %T Dynamic Query Tools for Time Series Data Sets: Timebox Widgets for Interactive Exploration %A Hochheiser,Harry %A Shneiderman, Ben %K angular queries %K Bioinformatics %K dynamic query %K Graphical user interfaces %K temporal data %K time series %K timeboxes %K TimeSearcher %K visual query %X Timeboxes are rectangular widgets that can be used in direct-manipulation graphical user interfaces (GUIs) to specify query constraints on time series data sets. Timeboxes are used to specify simultaneously two sets of constraints: given a set of N time series profiles, a timebox covering time periods x1…x2 (x1 ≤ x2) and values y1…y2 (y1 ≤ y2) will retrieve only those n√N that have values y1 ≤ y2 during all times x1 ≤ x ≤ x2. TimeSearcher is an information visualization tool that combines timebox queries with overview displays, query-by-example facilities, and support for queries over multiple time-varying attributes. Query manipulation tools including pattern inversion and ‘leaders & laggards’ graphical bookmarks provide additional support for interactive exploration of data sets. Extensions to the basic timebox model that provide additional expressivity include variable time timeboxes, which can be used to express queries with variability in the time interval, and angular queries, which search for ranges of differentials, rather than absolute values. Analysis of the algorithmic requirements for providing dynamic query performance for timebox queries showed that a sequential search outperformed searches based on geometric indices. Design studies helped identify the strengths and weaknesses of the query tools. Extended case studies involving the analysis of two different types of data from molecular biology experiments provided valuable feedback and validated the utility of both the timebox model and the TimeSearcher tool. Timesearcher is available at http://www.cs.umd.edu/hcil/timesearcher %B Information VisualizationInformation Visualization %V 3 %P 1 - 18 %8 2004/03/20/ %@ 1473-8716, 1473-8724 %G eng %U http://ivi.sagepub.com/content/3/1/1 %N 1 %R 10.1057/palgrave.ivs.9500061 %0 Conference Paper %B Communications, 2004 IEEE International Conference on %D 2004 %T Efficient bandwidth resource allocation for low-delay multiuser MPEG-4 video transmission %A Su,Guan-Ming %A M. Wu %K algorithm; %K allocation %K allocation; %K approach; %K Bandwidth %K broadband %K channel %K channels; %K coding; %K delays; %K expert %K FGS; %K fine %K granularity %K group; %K look-ahead %K moving %K MPEG-4 %K multiuser %K networks; %K of %K picture %K program; %K QUALITY %K resource %K scalability; %K service; %K sliding-window %K transmission; %K utilization; %K video %X An efficient bandwidth resource allocation algorithm with low delay and low fluctuation of quality to transmit multiple MPEG-4 fine granularity scalability (FGS) video programs to multiple users is proposed in this paper. By exploring the variation in the scene complexity of each video program and jointly redistributing available system resources among users, our proposed algorithm provides low fluctuation of quality for each user and consistent quality among all users. Experimental results show that compared to a traditional look-ahead sliding-window approach, our scheme can achieve comparable perceptual quality and channel utilization at a much lower cost of delay, computation, and storage. %B Communications, 2004 IEEE International Conference on %V 3 %P 1308 - 1312 Vol.3 - 1308 - 1312 Vol.3 %8 2004/06// %G eng %R 10.1109/ICC.2004.1312724 %0 Journal Article %J Languages and Compilers for Parallel Computing %D 2004 %T Efficient Execution of Multi-Query Data Analysis Batches Using Compiler Optimization Strategies %A Andrade,H. %A Aryangat,S. %A Kurc, T. %A Saltz, J. %A Sussman, Alan %X This work investigates the leverage that can be obtained from compiler optimization techniques for efficient execution of multi-query workloads in data analysis applications. Our approach is to address multi-query optimization at the algorithmic level, by transforming a declarative specification of scientific data analysis queries into a high-level imperative program that can be made more efficient by applying compiler optimization techniques. These techniques – including loop fusion, common subexpression elimination and dead code elimination – are employed to allow data and computation reuse across queries. We describe a preliminary experimental analysis on a real remote sensing application that analyzes very large quantities of satellite data. The results show our techniques achieve sizable reductions in the amount of computation and I/O necessary for executing query batches and in average execution times for the individual queries in a given batch. %B Languages and Compilers for Parallel Computing %8 2004/// %G eng %R 10.1007/978-3-540-24644-2_33 %0 Journal Article %J Linear Algebra and its Applications %D 2004 %T An Elsner-like perturbation theorem for generalized eigenvalues %A Stewart, G.W. %K eigenvalue %K Elsner’s theorem %K Generalized eigenvalue %K Perturbation theory %X In 1985 Elsner established a general bound on the distance between an eigenvalue of a matrix and the closest eigenvalue of a perturbation of that matrix. In this note, we show that a similar result holds for the generalized eigenvalue problem. %B Linear Algebra and its Applications %V 390 %P 1 - 5 %8 2004/10/01/ %@ 0024-3795 %G eng %U http://www.sciencedirect.com/science/article/pii/S0024379504002459 %R 10.1016/j.laa.2004.05.006 %0 Conference Paper %B Proceedings of the ICSE 2004 Workshop on Architecting Dependable Systems, Edinburgh, UK %D 2004 %T Empirical-based estimation of the effect on software dependability of a technique for architecture conformance verification %A Asgari, S. %A Basili, Victor R. %A Costa,P. %A Donzelli,P. %A Hochstein, L. %A Lindvall,M. %A Rus,I. %A Shull, F. %A Tvedt,R. %A Zelkowitz, Marvin V %X The High Dependability Computing Program (HDCP)project is a NASA initiative for increasing dependability of software-based systems. It researches achieving high dependability by introducing new technologies. We focus on the evaluation of the effectiveness of technologies with respect to dependability. We employ empirical evaluation methods along with evaluation testbeds. In this paper, our technology evaluation approach is described. A testbed representative of air traffic control applications is used. An experiment to evaluate a technology to for identification of architectural violations is presented. %B Proceedings of the ICSE 2004 Workshop on Architecting Dependable Systems, Edinburgh, UK %8 2004/// %G eng %0 Journal Article %J IEEE Transactions on Multimedia %D 2004 %T End-to-end analysis of distributed video-on-demand systems %A Mundur, Padma %A Simon,R. %A Sood,A. K %K Admission control %K admission control strategies %K Analytical models %K Computer science %K distributed architecture %K distributed video-on-demand systems %K end-to-end analysis %K global request handling %K High-speed networks %K Network servers %K Next generation networking %K Performance analysis %K performance evaluation methodology %K Protocols %K request handling %K reservation protocols %K resource allocation %K Resource management %K signaling protocols %K telecommunication congestion control %K video on demand %K Video sharing %X The focus of the research presented in this paper is the end-to-end analysis of a distributed Video-on-Demand (VoD) system. We analyze the distributed architecture of a VoD system to design global request handling and admission control strategies and evaluate them using global metrics. The performance evaluation methodology developed in this paper helps in determining efficient ways of using all resources in the VoD architecture within the constraints of providing guaranteed high quality service to each request. For instance, our simulation results show that request handling policies based on limited redirection of blocked requests to other resources perform better than load sharing policies. We also show that request handling policies based on redirection have simpler connection establishment semantics than load sharing policies and, therefore, are easily incorporated into reservation or signaling protocols. %B IEEE Transactions on Multimedia %V 6 %P 129 - 141 %8 2004/02// %@ 1520-9210 %G eng %N 1 %R 10.1109/TMM.2003.819757 %0 Conference Paper %B Proceedings of the fifteenth annual ACM-SIAM symposium on Discrete algorithms %D 2004 %T End-to-end packet-scheduling in wireless ad-hoc networks %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Parthasarathy,Srinivasan %A Srinivasan, Aravind %X Packet-scheduling is a particular challenge in wireless networks due to interference from nearby transmissions. A distance-2 interference model serves as a useful abstraction here, and we study packet routing and scheduling under this model. The main focus of our work is the development of fully-distributed (decentralized) protocols. We present polylogarithmic/constant factor approximation algorithms for various families of disk graphs (which capture the geometric nature of wireless-signal propagation), as well as near-optimal approximation algorithms for general graphs. The packet-scheduling work by L eighton, Maggs and Rao (Combinatorica, 1994) and a basic distributed coloring procedure, originally due to Luby (J. Computer and System Sciences, 1993), underlie many of our algorithms. Experimental work of Finocchi, Panconesi, and Silvestri (SODA 2002) showed that a natural modification of Luby's algorithm leads to improved performance, and a rigorous explanation of this was left as an open question; we prove that the modified algorithm is provably better in the worst-case. Finally, using simulations, we study the impact of the routing strategy and the choice of parameters on the performance of our distributed algorithm for unit disk graphs. %B Proceedings of the fifteenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '04 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 1021 - 1030 %8 2004/// %@ 0-89871-558-X %G eng %U http://dl.acm.org/citation.cfm?id=982792.982945 %0 Report %D 2004 %T Error Analysis of the Quasi-Gram--Schmidt Algorithm %A Stewart, G.W. %K Technical Report %X Let the $n{\times}p$ $(n\geq p)$ matrix $X$ have the QR~factorization$X = QR$, where $R$ is an upper triangular matrix of order $p$ and $Q$ is orthonormal. This widely used decomposition has the drawback that $Q$ is not generally sparse even when $X$ is. One cure is to discard $Q$ retaining only $X$ and $R$. Products like $a = Q\trp y = R\itp X\trp y$ can then be formed by computing $b = X\trp y$ and solving the system $R\trp a = b$. This approach can be used to modify the Gram--Schmidt algorithm for computing $Q$ and $R$ to compute $R$ without forming $Q$ or altering $X$. Unfortunately, this quasi-Gram--Schmidt algorithm can produce inaccurate results. In this paper it is shown that with reorthogonalization the inaccuracies are bounded under certain natural conditions. (UMIACS-TR-2004-17) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2004-17 %8 2004/04/19/ %G eng %U http://drum.lib.umd.edu/handle/1903/1346 %0 Conference Paper %B Eighth International Conference on Information Visualisation, 2004. IV 2004. Proceedings %D 2004 %T Extending the utility of treemaps with flexible hierarchy %A Chintalapani,G. %A Plaisant, Catherine %A Shneiderman, Ben %K 2D displays %K Computer displays %K Computer science %K data visualisation %K Data visualization %K Educational institutions %K flexible hierarchy %K graphical user interface %K Graphical user interfaces %K hierarchical information %K Nominations and elections %K Switches %K Tree data structures %K Tree graphs %K Treemap 4.0 %K Two dimensional displays %K usability %K visualization technique %X Treemaps are a visualization technique for presenting hierarchical information on two-dimensional displays. Prior implementations limit the visualization to pre-defined static hierarchies. Flexible hierarchy, a new capability of Treemap 4.0, enables users to define various hierarchies through dynamically selecting a series of data attributes so that they can discover patterns, clusters and outliers. This work describes the design and implementation issues of flexible hierarchy. It then reports on a usability study, which led to enhancements to the interface. %B Eighth International Conference on Information Visualisation, 2004. IV 2004. Proceedings %I IEEE %P 335 - 344 %8 2004/07/14/16 %@ 0-7695-2177-0 %G eng %R 10.1109/IV.2004.1320166 %0 Report %D 2004 %T Extrinsic Evaluation of Automatic Metrics for Summarization %A Dorr, Bonnie J %A Monz,Christof %A Oard, Douglas %A President,Stacy %A Zajic, David %A Schwartz,Richard %K *ABSTRACTS %K *ACCURACY %K *DATA SUMMARIZATION %K *DOCUMENT SUMMARIES %K *DOCUMENTS %K *READING MACHINES %K *SOFTWARE METRICS %K *STATISTICAL ANALYSIS %K ABSTRACT SELECTION %K ANNOTATIONS %K Automation %K COMPUTER PROGRAMMING AND SOFTWARE %K COMPUTER SYSTEMS %K CORRELATION TECHNIQUES %K DATA PROCESSING %K DOCUMENT SUMMARIES %K EXTRINSIC TASKS %K Information retrieval %K INFORMATION SCIENCE %K PERFORMANCE(HUMAN) %K PRECISION %K RELEVANCY %K ROUGE-1 COMPUTER PROGRAM %K STATISTICS AND PROBABILITY %K TEST AND EVALUATION %X This paper describes extrinsic-task evaluation of summarization. We show that it is possible to save time using summaries for relevance assessment without adversely impacting the degree of accuracy that would be possible with full documents. In addition, we demonstrate that the extrinsic task we have selected exhibits a high degree of interannotator agreement, i.e., consistent relevance decisions across subjects. We also conducted a composite experiment that better reflects the actual document selection process and found that using a surrogate improves the processing speed over reading the entire document. Finally, we have found a small yet statistically significant correlation between some of the intrinsic measures and a user's performance in an extrinsic task. The overall conclusion we can draw at this point is that ROUGE-1 does correlate with precision and to a somewhat lesser degree with accuracy, but that it remains to be investigated how stable these correlations are and how differences in ROUGE-1 translate into significant differences in human performance in an extrinsic task. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %8 2004/07/20/ %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA448065 %0 Conference Paper %B Eighth International Conference on Information Visualisation, 2004. IV 2004. Proceedings %D 2004 %T Facilitating understanding of information visualizations: emerging principles and examples %A Shneiderman, Ben %K Computer science %K data visualisation %K Educational institutions %K Guidelines %K hierarchical visualization %K HUMANS %K Information Visualization %K Laboratories %K multidimensional visualization %K Portable computers %K Testing %K usability %K User interfaces %K Visualization %X Summary form only given. The enthusiasm for information visualization has generated a wide variety of interesting tools for multi-dimensional, hierarchical, and other kinds of visualizations. However, some designs are readily accepted as understandable and useful, while others are perceived as confusing and useless. Controlled studies have begun to sort of some of the issues, but the insights of designers and usability tests are contributing interesting cognitive hypotheses for researchers and practical guidelines for developers. This paper offers examples of what works and what doesn't with a preliminary set of principles that might have wide applicability. %B Eighth International Conference on Information Visualisation, 2004. IV 2004. Proceedings %I IEEE %8 2004/07/14/16 %@ 0-7695-2177-0 %G eng %R 10.1109/IV.2004.1320117 %0 Conference Paper %D 2004 %T Ferret: a host vulnerability checking tool %A Sharma,Anil %A Martin,J.R. %A Anand,N. %A Michel Cukier %A Sanders,W. H. %K Ferret software tool %K host vulnerability checking tool %K open-source software %K Perl %K plug-in module %K program verification %K security auditing tool %K security evaluation %K security of data %K software tools %X Evaluation of computing system security requires knowledge of the vulnerabilities present in the system and of potential attacks against the system. Vulnerabilities can be classified based on their location as application vulnerabilities, network vulnerabilities, or host vulnerabilities. We describe Ferret, a new software tool for checking host vulnerabilities. Ferret helps system administrators by quickly finding vulnerabilities that are present on a host. It is designed and implemented in a modular way: a different plug-in module is used for each vulnerability checked, and each possible output format is specified by a plug-in module. As a result, Ferret is extensible, and can easily be kept up-to-date through addition of checks for new vulnerabilities as they are discovered; the modular approach also makes it easy to provide specific configurations of Ferret tailored to specific operating systems or use environments. Ferret is a freely available open-source software implemented in Perl. %P 389 - 394 %8 2004/03// %G eng %R 10.1109/PRDC.2004.1276595 %0 Journal Article %J BMC Bioinformatics %D 2004 %T Few amino acid positions in ıt rpoB are associated with most of the rifampin resistance in ıt Mycobacterium tuberculosis %A Cummings, Michael P. %A Segal,M. R %X BACKGROUND: Mutations in rpoB, the gene encoding the beta subunit of DNA-dependent RNA polymerase, are associated with rifampin resistance in Mycobacterium tuberculosis. Several studies have been conducted where minimum inhibitory concentration (MIC, which is defined as the minimum concentration of the antibiotic in a given culture medium below which bacterial growth is not inhibited) of rifampin has been measured and partial DNA sequences have been determined for rpoB in different isolates of M. tuberculosis. However, no model has been constructed to predict rifampin resistance based on sequence information alone. Such a model might provide the basis for quantifying rifampin resistance status based exclusively on DNA sequence data and thus eliminate the requirements for time consuming culturing and antibiotic testing of clinical isolates. RESULTS: Sequence data for amino acid positions 511-533 of rpoB and associated MIC of rifampin for different isolates of M. tuberculosis were taken from studies examining rifampin resistance in clinical samples from New York City and throughout Japan. We used tree-based statistical methods and random forests to generate models of the relationships between rpoB amino acid sequence and rifampin resistance. The proportion of variance explained by a relatively simple tree-based cross-validated regression model involving two amino acid positions (526 and 531) is 0.679. The first partition in the data, based on position 531, results in groups that differ one hundredfold in mean MIC (1.596 micrograms/ml and 159.676 micrograms/ml). The subsequent partition based on position 526, the most variable in this region, results in a > 354-fold difference in MIC. When considered as a classification problem (susceptible or resistant), a cross-validated tree-based model correctly classified most (0.884) of the observations and was very similar to the regression model. Random forest analysis of the MIC data as a continuous variable, a regression problem, produced a model that explained 0.861 of the variance. The random forest analysis of the MIC data as discrete classes produced a model that correctly classified 0.942 of the observations with sensitivity of 0.958 and specificity of 0.885. CONCLUSIONS: Highly accurate regression and classification models of rifampin resistance can be made based on this short sequence region. Models may be better with improved (and consistent) measurements of MIC and more sequence data. %B BMC Bioinformatics %V 5 %P 137 - 137 %8 2004/09// %G eng %R 10.1186/1471-2105-5-137 %0 Journal Article %J Journal of the American Society for Information Science and Technology %D 2004 %T Finding governmental statistical data on the Web: A study of categorically organized links for the FedStats topics page %A Ceaparu,Irina %A Shneiderman, Ben %X More than 100 U.S. governmental agencies offer links through FedStats, a centralized Web site that facilitates access to statistical tables, reports, and agencies. This and similar large collections need appropriate interfaces to guide the general public to easily and successfully find information they seek. This paper summarizes the results of 3 empirical studies of alternate organization concepts of the FedStats Topics Web page. Each study had 15 participants. The evolution from 645 alphabetically organized links, to 549 categorically organized links, to 215 categorically organized links tied to portal pages produced a steady rise in successful task completion from 15.6 to 24.4 to 42.2%. User satisfaction also increased. We make recommendations based on these data and our observations of users. %B Journal of the American Society for Information Science and Technology %V 55 %P 1008 - 1015 %8 2004/09/01/ %@ 1532-2890 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/asi.20052/abstract %N 11 %R 10.1002/asi.20052 %0 Journal Article %J Journal of the American Society for Information Science and Technology %D 2004 %T Finding Governmental Statistical Data on the Web: Three Empirical Studies of the FedStats Topics Page %A Ceaparu,Irina %A Shneiderman, Ben %X Governmental agencies provide statistical data on their web sites. These large collections of data need appropriate interfaces that would guide the general public, as well as the researchers, to easily and successfully find information they seek. This paper summarizes the results of three empirical studies with 15 users in each group of the FedStats Topics web page. The evolution from 645 alphabetically organized links, to 549 categorically organized links, to 215 categorically organized links tied to portal pages produced a steady rise in successful task completion from 15 % to 28 % to 42%. User satisfaction also increased. We make recommendations based on these data and our observations of users. %B Journal of the American Society for Information Science and Technology %V 55 %P 1008 - 1015 %8 2004/// %G eng %0 Conference Paper %B Proceedings of the 5th IEEE/ACM International Workshop on Grid Computing %D 2004 %T Flexible Control of Data Transfers between Parallel Programs %A Wu,Joe Shang-Chieh %A Sussman, Alan %X Allowing loose coupling between complex e-Science applications has many advantages, such as being able to easily incorporate new applications and to flexibly specify how the applications are connected to transfer data between them. To facilitate efficient, flexible data transfers between applications, in this paper we describe methods for making decisions at runtime about when such transfers will occur, and for flexibly specifying when data transfers are desired. We also present preliminary experimental results that measure the overheads incurred by our approach. %B Proceedings of the 5th IEEE/ACM International Workshop on Grid Computing %S GRID '04 %I IEEE Computer Society %C Washington, DC, USA %P 226 - 234 %8 2004/// %@ 0-7695-2256-4 %G eng %U http://dx.doi.org/10.1109/GRID.2004.31 %R 10.1109/GRID.2004.31 %0 Journal Article %J Handbook on Pattern Recognition and Computer Vision %D 2004 %T Gait-based human identification from a monocular video sequence %A Kale, A. %A Sundaresan, A. %A Roy Chowdhury, A.K. %A Chellapa, Rama %X Human gait is a spatio-temporal phenomenon that characterizes the motion char-acteristics of an individual. It is possible to detect and measure gait even in low- resolution video. In this chapter, we discuss algorithms for identifying people by their gait from a monocular video sequence. Human identification using gait, sim- ilar to text-based speaker identification, involves different individuals performing the same task and a template-matching approach is suitable for such problems. In situations where the amount of training data is limited, we demonstrate the utility of a simple width feature for gait recognition. By virtue of their determin- istic nature, template matching methods have limited noise resilience. In order to deal with noise we introduce a systematic approach to gait recognition by building representations for the structural and dynamic components of gait using exemplars and hidden Markov models (HMMs). The above methods assume that an exact side-view of the subject is available in the probe sequence. For the case when the person walks at an arbitrary angle far away from the camera we present a view invariant gait recognition algorithm which is based on synthesizing a side view of a person from an arbitrary monocular view. %B Handbook on Pattern Recognition and Computer Vision %8 2004/// %G eng %0 Journal Article %J Parallel and Distributed Systems, IEEE Transactions on %D 2004 %T A global-state-triggered fault injector for distributed system evaluation %A Chandra,Ramesh %A Lefever,R.M. %A Joshi,K.R. %A Michel Cukier %A Sanders,W. H. %K distributed processing %K distributed system evaluation %K fault tolerant computing %K global-state-based fault injection mechanism %K Loki %K offline clock synchronization %K performance evaluation %K post-runtime analysis %K Synchronisation %K system recovery %K user-specified performance %X Validation of the dependability of distributed systems via fault injection is gaining importance because distributed systems are being increasingly used in environments with high dependability requirements. The fact that distributed systems can fail in subtle ways that depend on the state of multiple parts of the system suggests that a global-state-based fault injection mechanism should be used to validate them. However, global-state-based fault injection is challenging since it is very difficult in practice to maintain the global state of a distributed system at runtime with minimal intrusion into the system execution. We present Loki, a global-state-based fault injector, which has been designed with the goals of low intrusion, high precision, and high flexibility. Loki achieves these goals by utilizing the ideas of partial view of global state, optimistic synchronization, and offline analysis. In Loki, faults are injected based on a partial, view of the global state of the system, and a post-runtime analysis is performed to place events and injections into a single global timeline and to discard experiments with incorrect fault injections. Finally, the experiments with correct fault injections are used to estimate user-specified performance and dependability measures. A flexible measure language has been designed that facilitates the specification of a wide range of measures. %B Parallel and Distributed Systems, IEEE Transactions on %V 15 %P 593 - 605 %8 2004/07// %@ 1045-9219 %G eng %N 7 %R 10.1109/TPDS.2004.14 %0 Journal Article %J Journal of Computer and System Sciences %D 2004 %T Guest editors' foreword: special issue: 35th annual ACM symposium on theory of computing %A Khanna,S. %A Srinivasan, Aravind %B Journal of Computer and System Sciences %V 69 %P 305 - 305 %8 2004/// %G eng %N 3 %0 Report %D 2004 %T Headline evaluation experiment results %A Zajic, David %A Dorr, Bonnie J %A Schwartz,R. %A President,S. %X This technical report describes an experiment intending to show that different summarizationtechniques have an effect on human performance of an extrinsic task. The task is document selection in the context of information retrieval. We conclude that the task was too difficult and ambiguous for subjects to perform at the level required in order to make statistically significant claims about the relationship between summarization techniques and human performance. An alternate interpretation of the experimental results is described and plans for future experiments are discussed. %I University of Maryland, College Park, MD. UMIACS-TR-2004-18 %8 2004/// %G eng %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2004. Proceedings. 18th International %D 2004 %T Hierarchical routing with soft-state replicas in TerraDir %A Silaghi,B. %A Gopalakrishnan,Vijay %A Bhattacharjee, Bobby %A Kelcher,P. %K ad-hoc %K adaptive %K allocation; %K asymmetrical %K balancing; %K bottlenecks; %K consistency %K constraints; %K delivering; %K demand %K distribution; %K guarantees; %K hierarchical %K latency %K load %K low %K namespaces; %K peer-to-peer %K protocol; %K protocols; %K replica %K replicas; %K replication %K resource %K routing; %K soft-state %K systems; %K TerraDir; %K topological %X Summary form only given. Recent work on peer-to-peer systems has demonstrated the ability to deliver low latencies and good load balance when demand for data is relatively uniform. We describe an adaptive replication protocol that delivers low latencies, good load balance even when demand is heavily skewed. The protocol can withstand arbitrary and instantaneous changes in demand distribution. Our approach also addresses classical concerns related to topological constraints of asymmetrical namespaces, such as hierarchical bottlenecks in the context of hierarchical namespaces. The protocol replicates routing state in an ad-hoc manner based on profiled information, is lightweight, scalable, and requires no replica consistency guarantees. %B Parallel and Distributed Processing Symposium, 2004. Proceedings. 18th International %P 48 - 48 %8 2004/04// %G eng %R 10.1109/IPDPS.2004.1302967 %0 Journal Article %J Genome Research %D 2004 %T Hierarchical Scaffolding With Bambus %A Pop, Mihai %A Kosack,Daniel S. %A Salzberg,Steven L. %X The output of a genome assembler generally comprises a collection of contiguous DNA sequences (contigs) whose relative placement along the genome is not defined. A procedure called scaffolding is commonly used to order and orient these contigs using paired read information. This ordering of contigs is an essential step when finishing and analyzing the data from a whole-genome shotgun project. Most recent assemblers include a scaffolding module; however, users have little control over the scaffolding algorithm or the information produced. We thus developed a general-purpose scaffolder, called Bambus, which affords users significant flexibility in controlling the scaffolding parameters. Bambus was used recently to scaffold the low-coverage draft dog genome data. Most significantly, Bambus enables the use of linking data other than that inferred from mate-pair information. For example, the sequence of a completed genome can be used to guide the scaffolding of a related organism. We present several applications of Bambus: support for finishing, comparative genomics, analysis of the haplotype structure of genomes, and scaffolding of a mammalian genome at low coverage. Bambus is available as an open-source package from our Web site. %B Genome Research %V 14 %P 149 - 159 %8 2004/01/01/ %G eng %U http://genome.cshlp.org/content/14/1/149.abstract %N 1 %R 10.1101/gr.1536204 %0 Journal Article %J Web Semantics: Science, Services and Agents on the World Wide Web %D 2004 %T HTN planning for Web Service composition using SHOP2 %A Sirin,Evren %A Parsia,Bijan %A Wu,Dan %A Hendler,James %A Nau, Dana S. %K HTN planning %K OWL-S %K SHOP2 %K Web Service composition %K Web services %X Automated composition of Web Services can be achieved by using AI planning techniques. Hierarchical Task Network (HTN) planning is especially well-suited for this task. In this paper, we describe how HTN planning system SHOP2 can be used with OWL-S Web Service descriptions. We provide a sound and complete algorithm to translate OWL-S service descriptions to a SHOP2 domain. We prove the correctness of the algorithm by showing the correspondence to the situation calculus semantics of OWL-S. We implemented a system that plans over sets of OWL-S descriptions using SHOP2 and then executes the resulting plans over the Web. The system is also capable of executing information-providing Web Services during the planning process. We discuss the challenges and difficulties of using planning in the information-rich and human-oriented context of Web Services. %B Web Semantics: Science, Services and Agents on the World Wide Web %V 1 %P 377 - 396 %8 2004/10// %@ 1570-8268 %G eng %U http://www.sciencedirect.com/science/article/pii/S1570826804000113 %N 4 %R 10.1016/j.websem.2004.06.005 %0 Conference Paper %B Proceedings of the 2nd international conference on Mobile systems, applications, and services %D 2004 %T Human needs and mobile technologies: small, fast, and fun %A Shneiderman, Ben %X The central thesis of "Leonardo's Laptop" (MIT Press, 2002) is that designers who are sensitive to human needs are more likely to make the breakthroughs that yield new technologies successes. Therefore, a theory of mobile devices would focus on compact devices that support human relationships, provide salient information, and enable creative expression. The foundations are not only the megahertz of connectivity, but also the usability and universality of interfaces. Demonstrations include digital photo applications, personal info, healthcare, and e-commerce. %B Proceedings of the 2nd international conference on Mobile systems, applications, and services %S MobiSys '04 %I ACM %C New York, NY, USA %P 1 - 1 %8 2004/// %@ 1-58113-793-1 %G eng %U http://doi.acm.org/10.1145/990064.990065 %R 10.1145/990064.990065 %0 Journal Article %J Comparative Evaluation of Multilingual Information Access Systems %D 2004 %T iCLEF 2003 at Maryland: Translation selection and document selection %A Dorr, Bonnie J %A He,D. %A Luo,J. %A Oard, Douglas %A Schwartz,R. %A Wang,J. %A Zajic, David %B Comparative Evaluation of Multilingual Information Access Systems %P 231 - 265 %8 2004/// %G eng %0 Book Section %B Comparative Evaluation of Multilingual Information Access SystemsComparative Evaluation of Multilingual Information Access Systems %D 2004 %T iCLEF 2003 at Maryland: Translation Selection and Document Selection %A Dorr, Bonnie J %A He,Daqing %A Luo,Jun %A Oard, Douglas %A Schwartz,Richard %A Wang,Jianqiang %A Zajic, David %E Peters,Carol %E Gonzalo,Julio %E Braschler,Martin %E Kluck,Michael %X Maryland performed two sets of experiments for the 2003 Cross-Language Evaluation Forum’s interactive track, one focused on interactive selection of appropriate translations for query terms, the second focused on interactive selection of relevant documents. Translation selection was supported using possible synonyms discovered through back translation and two techniques for generating KeyWord In Context (KWIC) examples of usage. The results indicate that searchers typically achieved a similar search effectiveness using fewer query iterations when interactive translation selection was available. For document selection, a complete extract of the first 40 words of each news story was compared to a compressed extract generated using an automated parse-and-trim approach that approximates one way in which people can produce headlines. The results indicate that compressed “headlines” result in faster assessment, but with a 20% relative reduction in the F α = 0.8 search effectiveness measure. %B Comparative Evaluation of Multilingual Information Access SystemsComparative Evaluation of Multilingual Information Access Systems %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3237 %P 231 - 265 %8 2004/// %@ 978-3-540-24017-4 %G eng %U http://dx.doi.org/10.1007/978-3-540-30222-3_42 %0 Journal Article %J Image Processing, IEEE Transactions on %D 2004 %T Identification of humans using gait %A Kale, A. %A Sundaresan, A. %A Rajagopalan, AN %A Cuntoor, N.P. %A Roy-Chowdhury, A.K. %A Kruger, V. %A Chellapa, Rama %K Automated;Reproducibility of Results;Sensitivity and Specificity;Signal Processing %K binary silhouette;frame to exemplar distance;gait databases;gait recognition;hidden Markov model;human identification;image features;observation probability;observation vector;hidden Markov models;image recognition;image representation;image sequences;pro %K Biological;Models %K Computer-Assisted;Models %K Computer-Assisted;Subtraction Technique;Task Performance and Analysis;Video Recording; %K Statistical;Pattern Recognition %X We propose a view-based approach to recognize humans from their gait. Two different image features have been considered: the width of the outer contour of the binarized silhouette of the walking person and the entire binary silhouette itself. To obtain the observation vector from the image features, we employ two different methods. In the first method, referred to as the indirect approach, the high-dimensional image feature is transformed to a lower dimensional space by generating what we call the frame to exemplar (FED) distance. The FED vector captures both structural and dynamic traits of each individual. For compact and effective gait representation and recognition, the gait information in the FED vector sequences is captured in a hidden Markov model (HMM). In the second method, referred to as the direct approach, we work with the feature vector directly (as opposed to computing the FED) and train an HMM. We estimate the HMM parameters (specifically the observation probability B) based on the distance between the exemplars and the image features. In this way, we avoid learning high-dimensional probability density functions. The statistical nature of the HMM lends overall robustness to representation and recognition. The performance of the methods is illustrated using several databases. %B Image Processing, IEEE Transactions on %V 13 %P 1163 - 1173 %8 2004/09// %@ 1057-7149 %G eng %N 9 %R 10.1109/TIP.2004.832865 %0 Conference Paper %B Multimedia Signal Processing, 2004 IEEE 6th Workshop on %D 2004 %T Image hashing resilient to geometric and filtering operations %A Swaminathan,A. %A Mao,Yinian %A M. Wu %K compact %K cryptographic %K cryptography; %K discrete %K distortion; %K Filtering %K Fourier %K function; %K geometric %K hash %K image %K key %K key; %K operation; %K polar %K PROCESSING %K public %K representation; %K theory; %K transform; %K transforms; %X Image hash functions provide compact representations of images, which is useful for search and authentication applications. In this work, we have identified a general three step framework and proposed a new image hashing scheme that achieves a better overall performance than the existing approaches under various kinds of image processing distortions. By exploiting the properties of discrete polar Fourier transform and incorporating cryptographic keys, the proposed image hash is resilient to geometric and filtering operations, and is secure against guessing and forgery attacks. %B Multimedia Signal Processing, 2004 IEEE 6th Workshop on %P 355 - 358 %8 2004/10/01/sept %G eng %R 10.1109/MMSP.2004.1436566 %0 Journal Article %J Interacting with Computers %D 2004 %T Immediate usability: a case study of public access design for a community photo library %A Kules,Bill %A Kang,Hyunmo %A Plaisant, Catherine %A Rose,Anne %A Shneiderman, Ben %K Casual use %K Community photo library %K direct annotation %K direct manipulation %K Drag-and-drop %K Group annotation %K Immediate usability %K Photo collection %K Public access system %K Walk-up-and-use %K Zero-trial learning %X This paper describes a novel instantiation of a digital photo library in a public access system. It demonstrates how designers can utilize characteristics of a target user community (social constraints, trust, and a lack of anonymity) to provide capabilities, such as unrestricted annotation and uploading of photos, which would be impractical in other types of public access systems. It also presents a compact set of design principles and guidelines for ensuring the immediate usability of public access information systems. These principles and guidelines were derived from our experience developing PhotoFinder Kiosk, a community photo library. Attendees of a major HCI conference (CHI 2001 Conference on Human Factors in Computing Systems) successfully used the tool to browse and annotate collections of photographs spanning 20 years of HCI-related conferences, producing a richly annotated photo history of the field of human–computer interaction. Observations and usage log data were used to evaluate the tool and develop the guidelines. They provide specific guidance for practitioners, as well as a useful framework for additional research in public access interfaces. %B Interacting with Computers %V 16 %P 1171 - 1193 %8 2004/12// %@ 0953-5438 %G eng %U http://www.sciencedirect.com/science/article/pii/S0953543804000840 %N 6 %R 10.1016/j.intcom.2004.07.005 %0 Conference Paper %B Logic Programming and Non-Monotonic Reasoning, Proceedings of the Second International Workshop %D 2004 %T Implementing stable semantics by linear programming %A Bell,C. %A Nerode,A. %A Ng,R. T %A V.S. Subrahmanian %B Logic Programming and Non-Monotonic Reasoning, Proceedings of the Second International Workshop %V 7 %P 23 - 42 %8 2004/// %G eng %0 Conference Paper %B Proceedings of the 2004 annual national conference on Digital government research %D 2004 %T Importing abstract spatial data into the SAND database system %A Samet, Hanan %A Brabec,Frantisek %A Sankaranarayanan,Jagan %X The World Wide Web has opened ways to operate large applications by users with simple client platforms located half way around the world by connecting servers around the world together. We present a system that makes a centrally stored spatial database available to off-site users. Regardless of the specific platform available to a user, all they need to do is simply establish a link between their client and the server. Unlike well-known web-based services such as MapQuest [2] that rely on computational power of the server, our system distributes the workload among the client and the server in such a manner that the user will observe the system as being interactive, with minimal delay between the user action and appropriate response for most types of operations. %B Proceedings of the 2004 annual national conference on Digital government research %S dg.o '04 %I Digital Government Society of North America %P 62:1–62:2 - 62:1–62:2 %8 2004/// %G eng %U http://dl.acm.org/citation.cfm?id=1124191.1124253 %0 Journal Article %J Signal %D 2004 %T Improved Multiuser Detectors Employing Genetic Algorithms in a Space-Time Block Coded System, Yinggang Du and Kam Tai Chan %A Self-Similarity,G.R.U.I. %A BenAbdelkader,C. %A Cutler,R.G. %A Davis, Larry S. %A Local,F.R.U. %A Global Features,J.H. %A Yuen, P.C. %A Lai,JH %A Li,C. %B Signal %V 2004 %P 640 - 648 %8 2004/// %G eng %N 5 %0 Journal Article %J Annals of Mathematics and Artificial Intelligence %D 2004 %T Improving performance of heterogeneous agents %A \textbackslashÖzcan,F. %A V.S. Subrahmanian %A Dix,J. %B Annals of Mathematics and Artificial Intelligence %V 41 %P 339 - 395 %8 2004/// %G eng %N 2 %0 Conference Paper %B Proceedings of the 8th International Conference on Software Reuse %D 2004 %T Improving the Quality of Performance-intensive Software via Model-integrated Distributed Continuous Quality Assurance %A Krishna,A. S %A Schmidt,D. C %A Porter, Adam %A Memon, Atif M. %A Sevilla-Ruiz,D. %B Proceedings of the 8th International Conference on Software Reuse %8 2004/// %G eng %0 Conference Paper %B Proceedings of the 2004 annual national conference on Digital government research %D 2004 %T Indexing distributed complex data for complex queries %A Tanin,Egemen %A Harwood,Aaron %A Samet, Hanan %X Peer-to-peer networks are becoming a common form of online data exchange. Querying data, mostly files, using keywords on peer-to-peer networks is well-known. But users cannot perform many types of queries on complex data and on many of the attributes of the data on such networks other than mostly exact-match queries. We introduce a distributed hashing-based index for enabling more powerful accesses on complex data over peer-to-peer networks that we expect to be commonly deployed for digital government applications. Preliminary experiments show that our index scales well and we believe that it can be extended to obtain similar indices for many other data types for performing various complex queries, such as range queries. %B Proceedings of the 2004 annual national conference on Digital government research %S dg.o '04 %I Digital Government Society of North America %P 63:1–63:10 - 63:1–63:10 %8 2004/// %G eng %U http://dl.acm.org/citation.cfm?id=1124191.1124254 %0 Book Section %B The Semantic Web – ISWC 2004 %D 2004 %T Information Gathering During Planning for Web Service Composition %A Kuter,Ugur %A Sirin,Evren %A Nau, Dana S. %A Parsia,Bijan %A Hendler,James %E McIlraith,Sheila %E Plexousakis,Dimitris %E van Harmelen,Frank %K Computer science %X Hierarchical Task-Network (HTN) based planning techniques have been applied to the problem of composing Web Services, especially when described using the OWL − S service ontologies. Many of the existing Web Services are either exclusively information providing or crucially depend on information-providing services. Thus, many interesting service compositions involve collecting information either during execution or during the composition process itself. In this paper, we focus on the latter issue. In particular, we present ENQUIRER , an HTN-planning algorithm designed for planning domains in which the information about the initial state of the world may not be complete, but it is discoverable through plan-time information-gathering queries. We have shown that ENQUIRER is sound and complete, and derived several mathematical relationships among the amount of available information, the likelihood of the planner finding a plan, and the quality of the plan found. We have performed experimental tests that confirmed our theoretical results and that demonstrated how ENQUIRER can be used in Web Service composition. %B The Semantic Web – ISWC 2004 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3298 %P 335 - 349 %8 2004/// %@ 978-3-540-23798-3 %G eng %U http://www.springerlink.com/content/v829m5080fc0bpng/abstract/ %0 Journal Article %J information retrieval %D 2004 %T Interactive cross-language document selection %A Oard, Douglas %A Gonzalo,J. %A Sanderson,M. %A López-Ostenero,F. %A Wang,J. %B information retrieval %V 7 %P 205 - 228 %8 2004/// %G eng %N 1 %0 Journal Article %J BioinformaticsBioinformatics %D 2004 %T Interactively Optimizing Signal-to-Noise Ratios in Expression Profiling: Project-Specific Algorithm Selection and Detection P-Value Weighting in Affymetrix Microarrays %A Seo,Jinwook %A Bakay,Marina %A Chen,Yi-Wen %A Hilmer,Sara %A Shneiderman, Ben %A Hoffman,Eric P %X Motivation: The most commonly utilized microarrays for mRNA profiling (Affymetrix) include ‘probe sets’ of a series of perfect match and mismatch probes (typically 22 oligonucleotides per probe set). There are an increasing number of reported ‘probe set algorithms’ that differ in their interpretation of a probe set to derive a single normalized ‘signal’ representative of expression of each mRNA. These algorithms are known to differ in accuracy and sensitivity, and optimization has been done using a small set of standardized control microarray data. We hypothesized that different mRNA profiling projects have varying sources and degrees of confounding noise, and that these should alter the choice of a specific probe set algorithm. Also, we hypothesized that use of the Microarray Suite (MAS) 5.0 probe set detection p-value as a weighting function would improve the performance of all probe set algorithms.Results: We built an interactive visual analysis software tool (HCE2W) to test and define parameters in Affymetrix analyses that optimize the ratio of signal (desired biological variable) versus noise (confounding uncontrolled variables). Five probe set algorithms were studied with and without statistical weighting of probe sets using the MAS 5.0 probe set detection p-values. The signal-to-noise ratio optimization method was tested in two large novel microarray datasets with different levels of confounding noise, a 105 sample U133A human muscle biopsy dataset (11 groups: mutation-defined, extensive noise), and a 40 sample U74A inbred mouse lung dataset (8 groups: little noise). Performance was measured by the ability of the specific probe set algorithm, with and without detection p-value weighting, to cluster samples into the appropriate biological groups (unsupervised agglomerative clustering with F-measure values). Of the total random sampling analyses, 50% showed a highly statistically significant difference between probe set algorithms by ANOVA [F(4,10) > 14, p < 0.0001], with weighting by MAS 5.0 detection p-value showing significance in the mouse data by ANOVA [F(1,10) > 9, p < 0.013] and paired t-test [t(9) = −3.675, p = 0.005]. Probe set detection p-value weighting had the greatest positive effect on performance of dChip difference model, ProbeProfiler and RMA algorithms. Importantly, probe set algorithms did indeed perform differently depending on the specific project, most probably due to the degree of confounding noise. Our data indicate that significantly improved data analysis of mRNA profile projects can be achieved by optimizing the choice of probe set algorithm with the noise levels intrinsic to a project, with dChip difference model with MAS 5.0 detection p-value continuous weighting showing the best overall performance in both projects. Furthermore, both existing and newly developed probe set algorithms should incorporate a detection p-value weighting to improve performance. Availability: The Hierarchical Clustering Explorer 2.0 is available at http://www.cs.umd.edu/hcil/hce/. Murine arrays (40 samples) are publicly available at the PEPR resource (http://microarray.cnmcresearch.org/pgadatatable.asp; http://pepr.cnmcresearch.org; Chen et al., 2004). %B BioinformaticsBioinformatics %V 20 %P 2534 - 2544 %8 2004/11/01/ %@ 1367-4803, 1460-2059 %G eng %U http://bioinformatics.oxfordjournals.org/content/20/16/2534 %N 16 %R 10.1093/bioinformatics/bth280 %0 Book Section %B Machine Translation: From Real Users to ResearchMachine Translation: From Real Users to Research %D 2004 %T Interlingual Annotation for MT Development %A Reeder,Florence %A Dorr, Bonnie J %A Farwell,David %A Habash,Nizar %A Helmreich,Stephen %A Hovy,Eduard %A Levin,Lori %A Mitamura,Teruko %A Miller,Keith %A Rambow,Owen %A Siddharthan,Advaith %E Frederking,Robert %E Taylor,Kathryn %X MT systems that use only superficial representations, including the current generation of statistical MT systems, have been successful and useful. However, they will experience a plateau in quality, much like other “silver bullet” approaches to MT. We pursue work on the development of interlingual representations for use in symbolic or hybrid MT systems. In this paper, we describe the creation of an interlingua and the development of a corpus of semantically annotated text, to be validated in six languages and evaluated in several ways. We have established a distributed, well-functioning research methodology, designed a preliminary interlingua notation, created annotation manuals and tools, developed a test collection in six languages with associated English translations, annotated some 150 translations, and designed and applied various annotation metrics. We describe the data sets being annotated and the interlingual (IL) representation language which uses two ontologies and a systematic theta-role list. We present the annotation tools built and outline the annotation process. Following this, we describe our evaluation methodology and conclude with a summary of issues that have arisen. %B Machine Translation: From Real Users to ResearchMachine Translation: From Real Users to Research %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 3265 %P 236 - 245 %8 2004/// %@ 978-3-540-23300-8 %G eng %U http://dx.doi.org/10.1007/978-3-540-30194-3_26 %0 Book Section %B DUC 04 Conference ProceedingsDUC 04 Conference Proceedings %D 2004 %T Left-Brain/Right-Brain Multi-Document Summarization %A Conroy,John M. %A Schlesinger,Judith D. %A Goldstein,Jade %A O'Leary, Dianne P. %B DUC 04 Conference ProceedingsDUC 04 Conference Proceedings %I U.S. National Inst. of Standards and Technology %8 2004/// %G eng %U \tt http://duc.nist.gov/ %0 Conference Paper %B Proceedings of HLT-NAACL 2004: Short Papers %D 2004 %T A lexically-driven algorithm for disfluency detection %A Snover,Matthew %A Dorr, Bonnie J %A Schwartz,Richard %X This paper describes a transformation-based learning approach to disfluency detection in speech transcripts using primarily lexical features. Our method produces comparable results to two other systems that make heavy use of prosodic features, thus demonstrating that reasonable performance can be achieved without extensive prosodic cues. In addition, we show that it is possible to facilitate the identification of less frequently disfluent discourse markers by taking speaker style into account. %B Proceedings of HLT-NAACL 2004: Short Papers %S HLT-NAACL-Short '04 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 157 - 160 %8 2004/// %@ 1-932432-24-8 %G eng %U http://dl.acm.org/citation.cfm?id=1613984.1614024 %0 Conference Paper %D 2004 %T A logic of motion %A Yaman,F. %A Nau, Dana S. %A V.S. Subrahmanian %X There are numerous applications such as air traffic manage- ment, cellular phone location tracking, and vehicle protection systems where there is a critical need to reason about moving objects. In this paper, we propose a formal logic of motion (LOM for short). We provide a formal syntax for LOM, as well as a model theory for LOM. In addition, we develop al- gorithms to check consistency of LOM theories, as well as to answer certain kinds of queries posed to LOM theories. We have implemented these algorithms in a prototype LOM sys- tem - we describe experiments showing that such queries can be efficiently executed in practice. %P 85 - 94 %8 2004/// %G eng %U http://www.aaai.org/Papers/KR/2004/KR04-011.pdf %0 Report %D 2004 %T Maintaining RDF views %A Deng,Y. %A Hung,E. %A V.S. Subrahmanian %X Resource Description Framework (RDF) has been recommended bythe World Wide Web Consortium for integrating web data. RDF specifications are increasingly being stored in RDF databases. Many applications will maintain views over RDF databases – these views may describe resources of interest to the owner of the views. As data about the resources available on the web is subject to frequent change, there is a need to incrementally maintain such views. We present algorithms to maintain materialized RDF views when data is inserted into, deleted from, or modified in RDF database instances. We have developed a prototype of our algorithms. For each of these three cases, we compare the performance of our algorithm against the naive approach of storing RDF data in a relational database and using the standard view maintenance algorithm %I Tech. Rep. CS-TR-4612 (UMIACS-TR-2004-54) %8 2004/// %G eng %0 Book Section %B Proceedings 2004 VLDB ConferenceProceedings 2004 VLDB Conference %D 2004 %T Maintenance of Spatial Semijoin Queries on Moving Points %A Iwerks,Glenn S. %A Samet, Hanan %A Smith,Kenneth P. %X In this paper, we address the maintenance of spatial semijoin queries over continuously moving points, where points are modeled as linear functions of time. This is analogous to the maintenance of a materialized view except, as time advances, the query result may change independently of updates. As in a materialized view, we assume there is no prior knowledge of updates before they occur. We present a new approach, continuous fuzzy sets (CFS), to maintain continuous spatial semijoins efficiently. CFS is compared experimentally to a simple scaling of previous work. The result is significantly better performance of CFS compared to previous work by up to an order of magnitude in some cases. %B Proceedings 2004 VLDB ConferenceProceedings 2004 VLDB Conference %I Morgan Kaufmann %C St Louis %P 828 - 839 %8 2004/// %@ 978-0-12-088469-8 %G eng %U http://www.sciencedirect.com/science/article/pii/B9780120884698500735 %0 Journal Article %J Computer Speech & Language %D 2004 %T Mandarin-English Information (MEI): investigating translingual speech retrieval %A Meng,Helen M. %A Chen,Berlin %A Khudanpur,Sanjeev %A Levow,Gina-Anne %A Lo,Wai-Kit %A Oard, Douglas %A Schone,Patrick %A Tang,Karen %A Wang,Hsin-Min %A Wang,Jianqiang %K English–Chinese cross-language spoken document retrieval %K Multi-scale spoken document retrieval %X This paper describes the Mandarin–English Information (MEI) project, where we investigated the problem of cross-language spoken document retrieval (CL-SDR), and developed one of the first English–Chinese CL-SDR systems. Our system accepts an entire English news story (text) as query, and retrieves relevant Chinese broadcast news stories (audio) from the document collection. Hence, this is a cross-language and cross-media retrieval task. We applied a multi-scale approach to our problem, which unifies the use of phrases, words and subwords in retrieval. The English queries are translated into Chinese by means of a dictionary-based approach, where we have integrated phrase-based translation with word-by-word translation. Untranslatable named entities are transliterated by a novel subword translation technique. The multi-scale approach can be divided into three subtasks – multi-scale query formulation, multi-scale audio indexing (by speech recognition) and multi-scale retrieval. Experimental results demonstrate that the use of phrase-based translation and subword translation gave performance gains, and multi-scale retrieval outperforms word-based retrieval. %B Computer Speech & Language %V 18 %P 163 - 179 %8 2004/04// %@ 0885-2308 %G eng %U http://www.sciencedirect.com/science/article/pii/S0885230803000524 %N 2 %R 10.1016/j.csl.2003.09.003 %0 Journal Article %J Monitoring, Security and Rescue Techniques in Multiagent Systems %D 2004 %T MASS: Multiagent Security and Survivability %A V.S. Subrahmanian %X Multiagent applications are not likely to be very useful in thereal world if they are unable to ensure the security and survivability of agents in the multiagent system. In this talk, I will focus on survivabil- ity of MASs and present a set of architectures that can greatly enhance survivability of MASs. I will discuss the advantages and disadvantages of the architectures and discuss their impact on scalability of the MAS applications. The talk reflects joint work with S. Kraus, C. Tas and Y. Zhang. %B Monitoring, Security and Rescue Techniques in Multiagent Systems %P 112 - 112 %8 2004/// %G eng %0 Conference Paper %B INFOCOM 2004. Twenty-third AnnualJoint Conference of the IEEE Computer and Communications Societies %D 2004 %T Measurement based optimal multi-path routing %A Guven,T. %A Kommareddy,C. %A La,R.J. %A Shayman,M.A. %A Bhattacharjee, Bobby %K algorithm;network %K approximation;routing %K IP %K measurement;network %K monitoring;optimization;perturbation %K multipath %K network;measurement-based %K networks;Internet;optimisation;routing %K processes; %K protocol;IP %K protocols;stochastic %K Routing %K Stochastic %X We propose a new architecture for efficient network monitoring and measurements in a traditional IP network. This new architecture enables establishment of multiple paths (tunnels) between source-destination pairs without having to modify the underlying routing protocol(s). Based on the proposed architecture we propose a measurement-based multipath routing algorithm derived from simultaneous perturbation stochastic approximation. The proposed algorithm does not assume that the gradient of analytical cost function is known to the algorithm, but rather relies on noisy estimates from measurements. Using the analytical model presented in the paper we prove the convergence of the algorithm to the optimal solution. Simulation results are presented to demonstrate the advantages of the proposed algorithm under a variety of network scenarios. A comparative study with an existing optimal routing algorithm, MATE, is also provided %B INFOCOM 2004. Twenty-third AnnualJoint Conference of the IEEE Computer and Communications Societies %V 1 %P 187 - 196 %8 2004/03// %G eng %R 10.1109/INFCOM.2004.1354493 %0 Patent %D 2004 %T Method of designing signaling networks for internet telephony %A Aravamudan,Murali %A Kumaran,Krishnan %A Ramakrishnan,Kajamalai Gopalaswamy %A Srinivasan, Aravind %E Lucent Technologies Inc. %X A method is disclosed for designing a signaling network of call coordinators (CCs) for internet telephony. The new method can be used to design a CC network of arbitrary size that satisfies, with high probability, limitations on the maximum number of sockets per CC and on the maximum number of hops between an arbitrary pair of switches in the network. According to the disclosed method, the network of CCs is treated initially as a collection of isolated points, one point for each CC. Then, links are added between pairs of CCs, excluding pairs lying within the same switch. The links are added randomly, but with a particular probability p, which may be different for different pairs. Thus, whether a given link is added depends, in effect, upon the outcome of a loaded coin toss in which the probability of a positive outcome is p. %V : 09/517,658 %8 2004/02/03/ %G eng %U http://www.google.com/patents?id=VTsSAAAAEBAJ %N 6687363 %0 Conference Paper %B Proceedings of the 2004 conference on Interaction design and children: building a community %D 2004 %T Mixing ideas: a new technique for working with young children as design partners %A Guha,M.L. %A Druin, Allison %A Chipman,G. %A Fails,J. A %A Simms,S. %A Farber,A. %B Proceedings of the 2004 conference on Interaction design and children: building a community %P 35 - 42 %8 2004/// %G eng %0 Journal Article %J Lecture Notes in Computer Science %D 2004 %T Model-Based Approaches-Validating Quality of Service for Reusable Software Via Model-Integrated Distributed Continuous Quality Assurance %A Krishna,A. S %A Schmidt,D. C %A Memon, Atif M. %A Porter, Adam %A Sevilla,D. %B Lecture Notes in Computer Science %V 3107 %P 286 - 295 %8 2004/// %G eng %0 Journal Article %J Proceedings of the 2nd ICSE Workshop on Remote Analysis and Measurement of Software Systems (RAMSS), Edinburgh, Scotland, UK %D 2004 %T A Model-based Distributed Continuous Quality Assurance Process to Enhance the Quality of Service of Evolving Performance-intensive Software Systems %A Yilmaz,C. %A Krishna,A. S %A Memon, Atif M. %A Porter, Adam %A Schmidt,D. C %A Gokhale,A. %A Natarajan,B. %X Performance-intensive software, such as that found in high-perfo-rmance computing systems and distributed real-time and embedded systems, increasingly executes on a multitude of platforms and user contexts. To ensure that performance-intensive software meets its quality of service (QoS) requirements, it must often be fine-tuned to specific platforms/contexts by adjusting many (in some cases hun- dreds of) configuration options. Developers who write these types of systems must therefore try to ensure that their additions and mod- ifications work across this large configuration space. In practice, however, time and resource constraints often force developers to assess performance on very few configurations and to extrapolate from these to the entire configuration space, which allows many performance bottlenecks and sources of QoS degradation to escape detection until systems are fielded. To improve the assessment of performance across large config- uration spaces, we present a model-based approach to develop- ing and deploying a new distributed continuous quality assurance (DCQA) process. Our approach builds upon and extends the Skoll environment, which is developing and validating novel software QA processes and tools that leverage the extensive computing resources of worldwide user communities in a distributed, continuous man- ner to significantly and rapidly improve software quality. This pa- per describes how our new DCQA performance assessment process enables developers to run formally-designed screening experiments that isolate the most significant options. After that, exhaustive ex- periments (on the now much smaller configuration space) are con- ducted. We implemented this process using model-based software tools and executed it in the Skoll environment to demonstrate its ef- fectiveness via two experiments on widely used QoS-enabled mid- dleware. Our results show that model-based DCQA processes im- proves developer insight into the effect of system changes on per- formance at an acceptable cost. %B Proceedings of the 2nd ICSE Workshop on Remote Analysis and Measurement of Software Systems (RAMSS), Edinburgh, Scotland, UK %8 2004/// %G eng %0 Journal Article %J Nature %D 2004 %T Modelling disease outbreaks in realistic urban social networks %A Eubank,Stephen %A Guclu,Hasan %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Srinivasan, Aravind %A Toroczkai,Zolt|[aacute]|n %A Wang,Nan %X Most mathematical models for the spread of disease use differential equations based on uniform mixing assumptions1 or ad hoc models for the contact process2, 3, 4. Here we explore the use of dynamic bipartite graphs to model the physical contact patterns that result from movements of individuals between specific locations. The graphs are generated by large-scale individual-based urban traffic simulations built on actual census, land-use and population-mobility data. We find that the contact network among people is a strongly connected small-world-like5 graph with a well-defined scale for the degree distribution. However, the locations graph is scale-free6, which allows highly efficient outbreak detection by placing sensors in the hubs of the locations network. Within this large-scale simulation framework, we then analyse the relative merits of several proposed mitigation strategies for smallpox spread. Our results suggest that outbreaks can be contained by a strategy of targeted vaccination combined with early detection without resorting to mass vaccination of a population. %B Nature %V 429 %P 180 - 184 %8 2004/05/13/ %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v429/n6988/full/nature02541.html %N 6988 %R 10.1038/nature02541 %0 Report %D 2004 %T Motivating annotation for digital photographs: Lowering barriers while raising incentives %A Kustanowitz,J. %A Shneiderman, Ben %X Frameworks for understanding annotation requirements could guide improved strategies that would motivate more users to invest the necessary effort. We propose one framework for annotation techniques along with the st rengths and weaknesses of each one, and a second framework for target user groups and their motivations. Several applications are described that provide useful and information-rich representations, but which require good annotations, in the hope of providing incentives for high quality annotation. We describe how annotations make possible four novel presentations of photo collections: (1) Birthday Collage to show growth of a child over several years, (2) FamiliarFace to show family trees of photos, (3) Kaleidoscope to show photos of related people in an appeal tableau, and (4) TripPics to show photos from a sequential story such as a vacation trip. %B HCIL-2004-18 %8 2004/// %G eng %0 Journal Article %J Languages and Compilers for Parallel Computing %D 2004 %T Mpjava: High-performance message passing in java using java. nio %A Pugh, William %A Spacco,J. %B Languages and Compilers for Parallel Computing %P 323 - 339 %8 2004/// %G eng %0 Journal Article %J Indian Conference on Computer Vision, Graphics and Image Processing %D 2004 %T Multi-cue exemplar-based nonparametric model for gesture recognition %A Shet,V. D %A Prasad,V. S.N %A Elgammal,A. %A Yacoob,Yaser %A Davis, Larry S. %X This paper presents an approach for a multi-cue, view-based recognition of gestures. We describe an exemplar- based technique that combines two different forms of exem- plars - shape exemplars and motion exemplars - in a uni- fied probabilistic framework. Each gesture is represented as a sequence of learned body poses as well as a sequence of learned motion parameters. The shape exemplars are comprised of pose contours, and the motion exemplars are represented as affine motion parameters extracted using a robust estimation approach. The probabilistic framework learns by employing a nonparametric estimation technique to model the exemplar distributions. It imposes temporal constraints between different exemplars through a learned Hidden Markov Model (HMM) for each gesture. We use the proposed multi-cue approach to recognize a set of four- teen gestures and contrast it against a shape only, single- cue based system. %B Indian Conference on Computer Vision, Graphics and Image Processing %P 16 - 18 %8 2004/// %G eng %0 Conference Paper %B Cluster Computing and the Grid, 2004. CCGrid 2004. IEEE International Symposium on %D 2004 %T Multi-dimensional quorum sets for read-few write-many replica control protocols %A Silaghi,B. %A Keleher,P. %A Bhattacharjee, Bobby %K accesses; %K availability; %K cache %K caching; %K Communication %K complexity; %K CONTROL %K d-spaces; %K distributed %K efficiency; %K lightweight %K logical %K multi-dimensional %K processing; %K protocols; %K quorum %K read %K read-few %K reconfiguration; %K replica %K sets; %K storage; %K structures; %K update %K write-many %X We describe d-spaces, a replica control protocol defined in terms of quorum sets on multi-dimensional logical structures. Our work is motivated by asymmetrical access patterns, where the number of read accesses to data are dominant relative to update accesses, i.e. where the protocols should be read-few write-many. D-spaces are optimal with respect to quorum group sizes. The quality of the tradeoff between read efficiency and update availability is not matched by existing quorum protocols. We also propose a novel scheme for implementing d-spaces that combines caching and local information to provide a best-effort form of global views. This allows quorum reconfiguration to be lightweight without impacting access latencies, even when the rate of membership changes is very high. %B Cluster Computing and the Grid, 2004. CCGrid 2004. IEEE International Symposium on %P 355 - 362 %8 2004/04// %G eng %R 10.1109/CCGrid.2004.1336588 %0 Conference Paper %B Image Processing, 2004. ICIP '04. 2004 International Conference on %D 2004 %T Multiple view tracking of humans modelled by kinematic chains %A Sundaresan, A. %A Chellapa, Rama %A RoyChowdhury, R. %K 3D %K algorithm; %K analysis; %K body %K calibrated %K cameras; %K chain %K displacement; %K error %K estimation; %K human %K image %K iterative %K kinematic %K kinematics; %K methods; %K model; %K MOTION %K motion; %K multiple %K parameters; %K perspective %K Pixel %K processing; %K projection %K sequences; %K signal %K tracking; %K video %K view %X We use a kinematic chain to model human body motion. We estimate the kinematic chain motion parameters using pixel displacements calculated from video sequences obtained from multiple calibrated cameras to perform tracking. We derive a linear relation between the 2D motion of pixels in terms of the 3D motion parameters of various body parts using a perspective projection model for the cameras, a rigid body motion model for the base body and the kinematic chain model for the body parts. An error analysis of the estimator is provided, leading to an iterative algorithm for calculating the motion parameters from the pixel displacements. We provide experimental results to demonstrate the accuracy of our formulation. We also compare our iterative algorithm to the noniterative algorithm and discuss its robustness in the presence of noise. %B Image Processing, 2004. ICIP '04. 2004 International Conference on %V 2 %P 1009 - 1012 Vol.2 - 1009 - 1012 Vol.2 %8 2004/10// %G eng %R 10.1109/ICIP.2004.1419472 %0 Journal Article %J Eurographics 2004, Tutorials 2: Multi-resolution Modeling, Visualization and Streaming of Volume Meshes %D 2004 %T Multi-resolution modeling, visualization and streaming of volume meshes %A Cignoni,P. %A De Floriani, Leila %A Lindstrom,P. %A Pascucci,V. %A Rossignac,J. %A Silva,C. %B Eurographics 2004, Tutorials 2: Multi-resolution Modeling, Visualization and Streaming of Volume Meshes %8 2004/// %G eng %0 Journal Article %J Computer-Aided Design %D 2004 %T A multi-resolution topological representation for non-manifold meshes %A De Floriani, Leila %A Magillo,Paola %A Puppo,Enrico %A Sobrero,Davide %K Data structures %K Multi-resolution %K Non-manifold modeling %X We address the problem of representing and processing 3D objects, described through simplicial meshes, which consist of parts of mixed dimensions, and with a non-manifold topology, at different levels of detail. First, we describe a multi-resolution model, that we call a non-manifold multi-tessellation (NMT), and we consider the selective refinement query, which is at the heart of several analysis operations on multi-resolution meshes. Next, we focus on a specific instance of a NMT, generated by simplifying simplicial meshes based on vertex-pair contraction, and we describe a compact data structure for encoding such a model. We also propose a new data structure for two-dimensional simplicial meshes, capable of representing both connectivity and adjacency information with a small memory overhead, which is used to describe the mesh extracted from an NMT through selective refinement. Finally, we present algorithms to efficiently perform updates on such a data structure. %B Computer-Aided Design %V 36 %P 141 - 159 %8 2004/02// %@ 0010-4485 %G eng %U http://www.sciencedirect.com/science/article/pii/S0010448503000587 %N 2 %R 10.1016/S0010-4485(03)00058-7 %0 Conference Paper %D 2004 %T New directions in design for manufacturing %A Herrmann,J.W. %A Cooper,J. %A Gupta,S.K. %A Hayes,C. C. %A Ishii,K. %A Kazmer,D. %A Sandborn,P. A. %A Wood,W. H. %X This paper gives an overview of research that is expandingthe domain of design for manufacturing (DFM) into new and important areas. This paper covers DFM and concurrent engineering, DFM for conceptual design, DFM for embodiment design, DFM for detailed design, design for production, platform design for reducing time-to-market, design for system quality, design for life cycle costs, and design for environment. The paper concludes with some general guidelines that suggest how manufacturing firms can develop useful, effective DFM tools. %P 1 - 9 %8 2004/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.198.8706&rep=rep1&type=pdf %0 Conference Paper %B AAAI Spring symposium on Computational Approaches to Analyzing Weblogs %D 2004 %T Oasys: An opinion analysis system %A Cesarano,C. %A Dorr, Bonnie J %A Picariello, A. %A Reforgiato,D. %A Sagoff,A. %A V.S. Subrahmanian %X There are numerous applications in which we would like toassess what opinions are being expressed in text documents. For example, Martha Stewart’s company may have wished to assess the degree of harshness of news articles about her in the recent past. Likewise, a World Bank official may wish to as- sess the degree of criticism of a proposed dam in Bangladesh. The ability to gauge opinion on a given topic is therefore of critical interest. In this paper, we develop a suite of algo- rithms which take as input, a set D of documents as well as a topic t, and gauge the degree of opinion expressed about topic t in the set D of documents. Our algorithms can return both a number (larger the number, more positive the opinion) as well as a qualitative opinion (e.g. harsh, complimentary). We as- sess the accuracy of these algorithms via human experiments and show that the best of these algorithms can accurately re- flect human opinions. We have also conducted performance experiments showing that our algorithms are computationally fast. %B AAAI Spring symposium on Computational Approaches to Analyzing Weblogs %8 2004/// %G eng %0 Journal Article %J ACM Comput. Surv. %D 2004 %T Object-based and image-based object representations %A Samet, Hanan %K Access methods %K feature query %K geographic information systems (GIS) %K image space %K location query %K object space %K octrees %K Pyramids %K quadtrees %K R-trees %K space-filling curves %K Spatial databases %X An overview is presented of object-based and image-based representations of objects by their interiors. The representations are distinguished by the manner in which they can be used to answer two fundamental queries in database applications: (1) Feature query: given an object, determine its constituent cells (i.e., their locations in space). (2) Location query: given a cell (i.e., a location in space), determine the identity of the object (or objects) of which it is a member as well as the remaining constituent cells of the object (or objects). Regardless of the representation that is used, the generation of responses to the feature and location queries is facilitated by building an index (i.e., the result of a sort) either on the objects or on their locations in space, and implementing it using an access structure that correlates the objects with the locations. Assuming the presence of an access structure, implicit (i.e., image-based) representations are described that are good for finding the objects associated with a particular location or cell (i.e., the location query), while requiring that all cells be examined when determining the locations associated with a particular object (i.e., the feature query). In contrast, explicit (i.e., object-based) representations are good for the feature query, while requiring that all objects be examined when trying to respond to the location query. The goal is to be able to answer both types of queries with one representation and without possibly having to examine every cell. Representations are presented that achieve this goal by imposing containment hierarchies on either space (i.e., the cells in the space in which the objects are found), or objects. In the former case, space is aggregated into successively larger-sized chunks (i.e., blocks), while in the latter, objects are aggregated into successively larger groups (in terms of the number of objects that they contain). The former is applicable to image-based interior-based representations of which the space pyramid is an example. The latter is applicable to object-based interior-based representations of which the R-tree is an example. The actual mechanics of many of these representations are demonstrated in the VASCO JAVA applets found at http://www.cs.umd.edu/˜hjs/quadtree/index.html. %B ACM Comput. Surv. %V 36 %P 159 - 217 %8 2004/06// %@ 0360-0300 %G eng %U http://doi.acm.org/10.1145/1031120.1031123 %N 2 %R 10.1145/1031120.1031123 %0 Journal Article %J Environmental Microbiology %D 2004 %T Occurrence and distribution of Vibrio cholerae in the coastal environment of Peru %A Gil,Ana I. %A Louis,Valérie R. %A Rivera,Irma N. G. %A Lipp,Erin %A Huq,Anwar %A Lanata,Claudio F. %A Taylor,David N. %A Russek‐Cohen,Estelle %A Choopun,Nipa %A Sack,R. Bradley %A Rita R Colwell %X The occurrence and distribution of Vibrio cholerae in sea water and plankton along the coast of Peru were studied from October 1997 to June 2000, and included the 1997–98 El Niño event. Samples were collected at four sites in coastal waters off Peru at monthly intervals. Of 178 samples collected and tested, V. cholerae O1 was cultured from 10 (5.6%) samples, and V. cholerae O1 was detected by direct fluorescent antibody assay in 26 out of 159 samples tested (16.4%). Based on the number of cholera cases reported in Peru from 1997 to 2000, a significant correlation was observed between cholera incidence and elevated sea surface temperature (SST) along the coast of Peru (P < 0.001). From the results of this study, coastal sea water and zooplankton are concluded to be a reservoir for V. cholerae in Peru. The climate–cholera relationship observed for the 1997–98 El Niño year suggests that an early warning system for cholera risk can be established for Peru and neighbouring Latin American countries. %B Environmental Microbiology %V 6 %P 699 - 706 %8 2004/07/01/ %@ 1462-2920 %G eng %U http://onlinelibrary.wiley.com/doi/10.1111/j.1462-2920.2004.00601.x/abstract?userIsAuthenticated=false&deniedAccessCustomisedMessage= %N 7 %R 10.1111/j.1462-2920.2004.00601.x %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 2004 %T Optimal models of disjunctive logic programs: semantics, complexity, and computation %A Leone,N. %A Scarcello,F. %A V.S. Subrahmanian %K complexity; %K computational %K disjunctive %K function; %K knowledge %K LANGUAGE %K Logic %K minimal %K model %K nonmonotonic %K objective %K optimisation; %K OPTIMIZATION %K problems; %K program %K Programming %K programming; %K reasoning; %K representation; %K semantics; %K stable %K user-specified %X Almost all semantics for logic programs with negation identify a set, SEM(P), of models of program P, as the intended semantics of P, and any model M in this class is considered a possible meaning of P with regard to the semantics the user has in mind. Thus, for example, in the case of stable models [M. Gelfond et al., (1988)], choice models [D. Sacca et al., (1990)], answer sets [M. Gelfond et al., (1991)], etc., different possible models correspond to different ways of "completing" the incomplete information in the logic program. However, different end-users may have different ideas on which of these different models in SEM(P) is a reasonable one from their point of view. For instance, given SEM(P), user U1 may prefer model M1 isin;SEM(P) to model M2 isin;SEM(P) based on some evaluation criterion that she has. We develop a logic program semantics based on optimal models. This semantics does not add yet another semantics to the logic programming arena - it takes as input an existing semantics SEM(P) and a user-specified objective function Obj, and yields a new semantics Opt(P)_ sube; SEM(P) that realizes the objective function within the framework of preferred models identified already by SEM(P). Thus, the user who may or may not know anything about logic programming has considerable flexibility in making the system reflect her own objectives by building "on top" of existing semantics known to the system. In addition to the declarative semantics, we provide a complete complexity analysis and algorithms to compute optimal models under varied conditions when SEM(P) is the stable model semantics, the minimal models semantics, and the all-models semantics. %B Knowledge and Data Engineering, IEEE Transactions on %V 16 %P 487 - 503 %8 2004/04// %@ 1041-4347 %G eng %N 4 %R 10.1109/TKDE.2004.1269672 %0 Journal Article %J Canadian Journal of Microbiology %D 2004 %T Pandemic strains of O3:K6 Vibrio parahaemolyticus in the aquatic environment of Bangladesh %A Islam,M. S. %A Tasmin,Rizwana %A Khan,Sirajul I. s l a m %A Bakht,Habibul B. M. %A Mahmood,Zahid H. a y a t %A Rahman,M. Z. i a u r %A Bhuiyan,Nurul A. m i n %A Nishibuchi,Mitsuaki %A Nair,G. B. a l a k r i s h %A Sack,R. B. r a d l e y %A Huq,Anwar %A Rita R Colwell %A Sack,David A. %X A total of 1500 environmental strains of Vibrio parahaemolyticus, isolated from the aquatic environment of Bangladesh, were screened for the presence of a major V. parahaemolyticus virulence factor, the thermostable direct haemolysin (tdh) gene, by the colony blot hybridization method using a digoxigenin-labeled tdh gene probe. Of 1500 strains, 5 carried the tdh sequence, which was further confirmed by PCR using primers specific for the tdh gene. Examination by PCR confirmed that the 5 strains were V. parahamolyticus and lacked the thermostable direct haemolysin-related haemolysin (trh) gene, the alternative major virulence gene known to be absent in pandemic strains. All 5 strains gave positive Kanagawa phenomenon reaction with characteristic beta-haemolysis on Wagatsuma agar medium. Southern blot analysis of the HindIII-digested chromosomal DNA demonstrated, in all 5 strains, the presence of 2 tdh genes common to strains positive for Kanagawa phenomenon. However, the 5 strains were found to belong to 3 different serotypes (O3:K29, O4:K37, and O3:K6). The 2 with pandemic serotype O3:K6 gave positive results in group-specific PCR and ORF8 PCR assays, characteristics unique to the pandemic clone. Clonal variations among the 5 isolates were analyzed by comparing RAPD and ribotyping patterns. Results showed different patterns for the 3 serotypes, but the pattern was identical among the O3:K6 strains. This is the first report on the isolation of pandemic O3:K6 strains of V. parahaemolyticus from the aquatic environment of Bangladesh. %B Canadian Journal of Microbiology %V 50 %P 827 - 834 %8 2004/10// %G eng %U http://umd.library.ingentaconnect.com/content/nrc/cjm/2004/00000050/00000010/art00007 %N 10 %0 Report %D 2004 %T PAWN: Producer-Archive Workflow Network in support of digital preservation %A Smorul,M. %A JaJa, Joseph F. %A Wang, Y. %A McCall,F. %X We describe the design and the implementation of the PAWN (Producer – ArchiveWorkflow Network) environment to enable secure and distributed ingestion of digital objects into a persistent archive. PAWN was developed to capture the core elements required for long term preservation of digital objects as identified by previous research in the digital library and archiving communities. In fact, PAWN can be viewed as an implementation of the Ingest Process as defined by the Open Archival Information System (OAIS) Reference Model, and is currently being used to ingest significant collections into a pilot persistent archive developed through a collaboration between the San Diego Supercomputer Center, the University of Maryland, and the National Archives and Records Administration. We make use of METS (Metadata Encoding and Transmission Standards) to encapsulate content, structural, descriptive, and preservation metadata. The basic software components are based on open standards and web technologies, and hence are platform independent. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2004 %P 2006 - 2006 %8 2004/// %G eng %0 Book Section %B Foundations of Information and Knowledge Systems %D 2004 %T Plan Databases: Model and Algebra %A Yaman,Fusun %A Adali,Sibel %A Nau, Dana S. %A Sapino,Maria %A Subrahmanian,V. %E Seipel,Dietmar %E Turull-Torres,José %K Computer science %X Despite the fact that thousands of applications manipulate plans, there has been no work to date on managing large databases of plans. In this paper, we first propose a formal model of plan databases. We describe important notions of consistency and coherence for such databases. We then propose a set of operators similar to the relational algebra to query such databases of plans. %B Foundations of Information and Knowledge Systems %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2942 %P 302 - 319 %8 2004/// %@ 978-3-540-20965-2 %G eng %U http://www.springerlink.com/content/yqhlqtu4te18q2e1/abstract/ %0 Journal Article %J Software, IEEE %D 2004 %T Preserving distributed systems critical properties: a model-driven approach %A Yilmaz,C. %A Memon, Atif M. %A Porter, Adam %A Krishna,A. S %A Schmidt,D. C %A Gokhale,A. %A Natarajan,B. %K configuration management %K formal verification %K Middleware %K middleware suite %K model-driven approach %K persistent software attributes %K QoS requirements %K Quality assurance %K quality of service %K quality-of-service %K Skoll distributed computing resources %K software configuration %K Software maintenance %K Software quality %K software quality assurance process %K system dependability %X The need for creating predictability in distributed systems is most often specified in terms of quality-of-service (QoS) requirements, which help define the acceptable levels of dependability with which capabilities such as processing capacity, data throughput, or service availability reach users. For longer-term properties such as scalability, maintainability, adaptability, and system security, we can similarly use persistent software attributes (PSAs) to specify how and to what degree such properties must remain intact as a network expands and evolves over time. The Skoll distributed continuous software quality assurance process helps to identify viable system and software configurations for meeting stringent QOS and PSA requirements by coordinating the use of distributed computing resources. The authors tested their process using the large, rapidly evolving ACE+TAO middleware suite. %B Software, IEEE %V 21 %P 32 - 40 %8 2004/// %@ 0740-7459 %G eng %N 6 %R 10.1109/MS.2004.50 %0 Conference Paper %B Proceedings of the 2nd ACM international workshop on Multimedia databases %D 2004 %T The priority curve algorithm for video summarization %A Fayzullin,M. %A V.S. Subrahmanian %A Albanese, M. %A Picariello, A. %K probabilistic %K Summarization %K system %K video %X In this paper, we introduce the concept of a priority curve associated with a video. We then provide an algorithm that can use the priority curve to create a summary (of a desired length) of any video. The summary thus created exhibits nice continuity properties and also avoids repetition. We have implemented the priority curve algorithm (PCA) and compared it with other summarization algorithms in the literature. We show that PCA is faster than existing algorithms and also produces better quality summaries. The quality of summaries was evaluated by a group of 200 students in Naples, Italy, who watched soccer videos. We also briefly describe a soccer video summarization system we have built on using the PCA architecture and various (classical) image processing algorithms. %B Proceedings of the 2nd ACM international workshop on Multimedia databases %S MMDB '04 %I ACM %C New York, NY, USA %P 28 - 35 %8 2004/// %@ 1-58113-975-6 %G eng %U http://doi.acm.org/10.1145/1032604.1032611 %R 10.1145/1032604.1032611 %0 Journal Article %J IEEE Wireless Communications %D 2004 %T Proactive key distribution using neighbor graphs %A Mishra,A. %A Min Ho Shin %A Petroni,N. L. %A Clancy,T. C %A Arbaugh, William A. %K access points %K Authentication %K authentication time %K Base stations %K Communication system security %K Delay %K graph theory %K GSM %K IEEE 802.11 handoff %K Land mobile radio cellular systems %K Message authentication %K mobile radio %K Multiaccess communication %K neighbor graph %K Network topology %K Roaming %K telecommunication security %K Telephone sets %K user mobility %K Wi-Fi networks %K wireless data networks %K Wireless LAN %K Wireless networks %X User mobility in wireless data networks is increasing because of technological advances, and the desire for voice and multimedia applications. These applications, however, require that handoffs between base stations (or access points) be fast to maintain the quality of the connections. In this article we introduce a novel data structure, the neighbor graph, that dynamically captures the mobility topology of a wireless network. We show how neighbor graphs can be utilized to obtain a 99 percent reduction in the authentication time of an IEEE 802.11 handoff (full EAP-TLS) by proactively distributing necessary key material one hop ahead of the mobile user. We also present a reactive method for fast authentication that requires only firmware changes to access points and hence can easily be deployed on existing wireless networks. %B IEEE Wireless Communications %V 11 %P 26 - 36 %8 2004/02// %@ 1536-1284 %G eng %N 1 %R 10.1109/MWC.2004.1269714 %0 Conference Paper %B Proceedings of the 2004 annual national conference on Digital government research %D 2004 %T Project highlight: toward the statistical knowledge network %A Marchionini,Gary %A Haas,Stephanie %A Shneiderman, Ben %A Plaisant, Catherine %A Hert,Carol A. %X This project aims to help people find and understand government statistical information. To achieve this goal, we envision a statistical knowledge network that brings stakeholders from government at all levels together with citizens who provide or seek statistical information. The linchpin of this network is a series of human-computer interfaces that facilitate information seeking, understanding, and use. In turn, these interfaces depend on high-quality metadata and intra-agency cooperation. In this briefing, we summarize our accomplishments in the second year of the project. %B Proceedings of the 2004 annual national conference on Digital government research %S dg.o '04 %I Digital Government Society of North America %P 125:1–125:2 - 125:1–125:2 %8 2004/// %G eng %U http://dl.acm.org/citation.cfm?id=1124191.1124316 %0 Conference Paper %B IEEE Symposium on Information Visualization, 2004. INFOVIS 2004 %D 2004 %T A Rank-by-Feature Framework for Unsupervised Multidimensional Data Exploration Using Low Dimensional Projections %A Seo,J. %A Shneiderman, Ben %K axis-parallel projections %K boxplot %K color-coded lower-triangular matrix %K computational complexity %K computational geometry %K Computer displays %K Computer science %K Computer vision %K Data analysis %K data mining %K data visualisation %K Data visualization %K Displays %K dynamic query %K Educational institutions %K exploratory data analysis %K feature detection %K feature detection/selection %K Feature extraction %K feature selection %K graph theory %K graphical displays %K histogram %K Information Visualization %K interactive systems %K Laboratories %K Multidimensional systems %K Principal component analysis %K rank-by-feature prism %K scatterplot %K statistical analysis %K statistical graphics %K statistical graphs %K unsupervised multidimensional data exploration %K very large databases %X Exploratory analysis of multidimensional data sets is challenging because of the difficulty in comprehending more than three dimensions. Two fundamental statistical principles for the exploratory analysis are (1) to examine each dimension first and then find relationships among dimensions, and (2) to try graphical displays first and then find numerical summaries (D.S. Moore, (1999). We implement these principles in a novel conceptual framework called the rank-by-feature framework. In the framework, users can choose a ranking criterion interesting to them and sort 1D or 2D axis-parallel projections according to the criterion. We introduce the rank-by-feature prism that is a color-coded lower-triangular matrix that guides users to desired features. Statistical graphs (histogram, boxplot, and scatterplot) and information visualization techniques (overview, coordination, and dynamic query) are combined to help users effectively traverse 1D and 2D axis-parallel projections, and finally to help them interactively find interesting features %B IEEE Symposium on Information Visualization, 2004. INFOVIS 2004 %I IEEE %P 65 - 72 %8 2004/// %@ 0-7803-8779-3 %G eng %R 10.1109/INFVIS.2004.3 %0 Journal Article %J Proc. DARPA Rich Transcription Workshop %D 2004 %T RT-S: Surface rich transcription scoring, methodology, and initial results %A Snover,M. %A Schwartz,R. %A Dorr, Bonnie J %A Makhoul,J. %X In this paper we present a methodoly for the scoring ofpunctuation annotated texts, as well as a preliminary system to perform the task. We modify SCLITE’s scoring method- ology to support scoring of punctuation. Using this method- ology, we show that the error rate of an initial automatic sys- tem is comparable to annotator inconsistency. However, the use of multiple references allows us to differentiate between human inconsistencies and system errors. %B Proc. DARPA Rich Transcription Workshop %8 2004/// %G eng %0 Conference Paper %B Proceedings of the 14th international workshop on Network and operating systems support for digital audio and video %D 2004 %T Scalable resilient media streaming %A Banerjee,Suman %A Lee,Seungjoon %A Braud,Ryan %A Bhattacharjee, Bobby %A Srinivasan, Aravind %K media streaming %K Multicast %K overlay network %K Resilience %X We present a low-overhead media streaming system, called SRMS (Scalable Resilient Media Streaming) that can be used to scalably deliver streaming data to a large group of receivers. SRMS uses overlay multicast for data distribution. to a large group of users. SRMS leverages a probabilistic loss recovery technique to provide high data delivery guarantees even under large network losses and overlay node failures. The clients in the SRMS system are able to interoperate with existing media streaming servers that use RTP for data transport. One of the interesting features of SRMS is that it can simultaneously support clients with disparate access bandwidths. It enables the necessary bandwidth adaptations using standard Real-time Transport Protocol (RTP) mechanisms, e.g. RTP translators. We have implemented and evaluated the SRMS system in detail on an emulated network as well as on a wide-area testbed with up to 128 clients. Our results show that clients using SRMS achieve high (97%) data delivery ratios with low overheads (<5%) even for a very dynamic network (up to five membership changes per minute). %B Proceedings of the 14th international workshop on Network and operating systems support for digital audio and video %S NOSSDAV '04 %I ACM %C New York, NY, USA %P 4 - 9 %8 2004/// %@ 1-58113-801-6 %G eng %U http://doi.acm.org/10.1145/1005847.1005851 %R 10.1145/1005847.1005851 %0 Journal Article %J Visualization and Computer Graphics, IEEE Transactions on %D 2004 %T Selective refinement queries for volume visualization of unstructured tetrahedral meshes %A Cignoni,P. %A De Floriani, Leila %A Magillo,P. %A Puppo,E. %A Scopigno,R. %K Automated;Signal Processing %K Computer-Assisted;Imaging %K Computer-Assisted;User-Computer Interface; %K data structure;geometric modeling;interactive visualization;large data sets;multiresolution model;selective refinement queries;unstructured tetrahedral meshes;variable resolution queries;volume data visualization;data visualisation;mesh generation;query p %K Three-Dimensional;Online Systems;Pattern Recognition %X We address the problem of the efficient visualization of large irregular volume data sets by exploiting a multiresolution model based on tetrahedral meshes. Multiresolution models, also called Level-Of-Detail (LOD) models, allow encoding the whole data set at a virtually continuous range of different resolutions. We have identified a set of queries for extracting meshes at variable resolution from a multiresolution model, based on field values, domain location, or opacity of the transfer function. Such queries allow trading off between resolution and speed in visualization. We define a new compact data structure for encoding a multiresolution tetrahedral mesh built through edge collapses to support selective refinement efficiently and show that such a structure has a storage cost from 3 to 5.5 times lower than standard data structures used for tetrahedral meshes. The data structures and variable resolution queries have been implemented together with state-of-the art visualization techniques in a system for the interactive visualization of three-dimensional scalar fields defined on tetrahedral meshes. Experimental results show that selective refinement queries can support interactive visualization of large data sets. %B Visualization and Computer Graphics, IEEE Transactions on %V 10 %P 29 - 45 %8 2004/// %@ 1077-2626 %G eng %N 1 %R 10.1109/TVCG.2004.1260756 %0 Journal Article %J INFORMS Journal on Computing %D 2004 %T A semidefinite programming approach to side chain positioning with new rounding strategies %A Chazelle,B. %A Kingsford, Carl %A Singh,M. %B INFORMS Journal on Computing %V 16 %P 380 - 392 %8 2004/// %G eng %N 4 %0 Conference Paper %B Proceedings of the 12th annual ACM international workshop on Geographic information systems %D 2004 %T A serverless 3D world %A Tanin,Egemen %A Harwood,Aaron %A Samet, Hanan %A Nutanong,Sarana %A Truong,Minh Tri %K distributed hash tables %K peer-to-peer systems %K spatial data %K virtual-worlds %X Online multi-participant virtual-world systems have attracted significant interest from the Internet community but are hindered by their inability to efficiently support interactivity for a large number of participants. Current solutions divide a large virtual-world into a few mutually exclusive zones, with each zone controlled by a different server, and/or limit the number of participants per server or per virtual-world. Peer-to-Peer (P2P) systems are known to provide excellent scalability in a networked environment (one peer is introduced to the system by each participant), however current P2P applications can only provide file sharing and other forms of relatively simple data communications. In this paper, we present a generic 3D virtual-world application that runs on a P2P network with no central administration or server. Two issues are addressed by this paper to enable such a spatial application on a P2P network. First, we demonstrate how to index and query a 3D space on a dynamic distributed network. Second, we show how to build such a complex application from the ground level of a P2P routing algorithm. Our work leads to new directions for the development of online virtual-worlds that we believe can be used for many government, industry, and public domain applications. %B Proceedings of the 12th annual ACM international workshop on Geographic information systems %S GIS '04 %I ACM %C New York, NY, USA %P 157 - 165 %8 2004/// %@ 1-58113-979-9 %G eng %U http://doi.acm.org/10.1145/1032222.1032246 %R 10.1145/1032222.1032246 %0 Journal Article %J Proceedings of the 14th International Workshop on Network and Operating Systems Support for Digital Audio and Video: NOSSDAV 2003: June 16-18, 2004, Cork, Ireland %D 2004 %T SESSION 2: STREAMING %A Banerjee,S. %A Lee,S. %A Braud,R. %A Bhattacharjee, Bobby %A Srinivasan, Aravind %A Chu,Y. %A Zhang,H. %A Sinha,R. %A Papadopoulos,C. %A Boustead,P. %A others %B Proceedings of the 14th International Workshop on Network and Operating Systems Support for Digital Audio and Video: NOSSDAV 2003: June 16-18, 2004, Cork, Ireland %P 3 - 3 %8 2004/// %G eng %0 Conference Paper %B Image Processing, 2004. ICIP '04. 2004 International Conference on %D 2004 %T Simultaneous background and foreground modeling for tracking in surveillance video %A Shao, J. %A Zhou,S. K %A Chellapa, Rama %K algorithm; %K analysis; %K background-foreground %K displacement %K estimation; %K image %K information; %K INTENSITY %K modeling; %K MOTION %K processes; %K processing; %K resolution; %K sequences; %K signal %K Stochastic %K Surveillance %K surveillance; %K tracking %K tracking; %K video %X We present a stochastic tracking algorithm for surveillance video where targets are dim and at low resolution. The algorithm builds motion models for both background and foreground by integrating motion and intensity information. Some other merits of the algorithm include adaptive selection of feature points for scene description and defining proper cost functions for displacement estimation. The experimental results show tracking robustness and precision in a challenging video sequences. %B Image Processing, 2004. ICIP '04. 2004 International Conference on %V 2 %P 1053 - 1056 Vol.2 - 1053 - 1056 Vol.2 %8 2004/10// %G eng %R 10.1109/ICIP.2004.1419483 %0 Conference Paper %B Software Engineering, 2004. ICSE 2004. Proceedings. 26th International Conference on %D 2004 %T Skoll: distributed continuous quality assurance %A Memon, Atif M. %A Porter, Adam %A Yilmaz,C. %A Nagarajan,A. %A Schmidt,D. %A Natarajan,B. %K 1MLOC+ software package %K ACE+TAO %K around-the-clock QA process %K around-the-world QA process %K distributed continuous QA %K distributed continuous quality assurance %K distributed programming %K program verification %K Quality assurance %K Skoll %K software performance evaluation %K software profiling %K Software quality %K Software testing %X Quality assurance (QA) tasks, such as testing, profiling, and performance evaluation, have historically been done in-house on developer-generated workloads and regression suites. Since this approach is inadequate for many systems, tools and processes are being developed to improve software quality by increasing user participation in the QA process. A limitation of these approaches is that they focus on isolated mechanisms, not on the coordination and control policies and tools needed to make the global QA process efficient, effective, and scalable. To address these issues, we have initiated the Skoll project, which is developing and validating novel software QA processes and tools that leverage the extensive computing resources of worldwide user communities in a distributed, continuous manner to significantly and rapidly improve software quality. This paper provides several contributions to the study of distributed continuous QA. First, it illustrates the structure and functionality of a generic around-the-world, around-the-clock QA process and describes several sophisticated tools that support this process. Second, it describes several QA scenarios built using these tools and process. Finally, it presents a feasibility study applying these scenarios to a 1MLOC+ software package called ACE+TAO. While much work remains to be done, the study suggests that the Skoll process and tools effectively manage and control distributed, continuous QA processes. Using Skoll we rapidly identified problems that had taken the ACE+TAO developers substantially longer to find and several of which had previously not been found. Moreover, automatic analysis of QA task results often provided developers information that quickly led them to the root cause of the problems. %B Software Engineering, 2004. ICSE 2004. Proceedings. 26th International Conference on %P 459 - 468 %8 2004/05// %G eng %R 10.1109/ICSE.2004.1317468 %0 Conference Paper %B INFOCOM 2004. Twenty-third AnnualJoint Conference of the IEEE Computer and Communications Societies %D 2004 %T Slurpie: a cooperative bulk data transfer protocol %A Sherwood,R. %A Braud,R. %A Bhattacharjee, Bobby %K adaptive %K bulk %K client-server %K clients; %K computing; %K cooperative %K data %K data; %K downloading %K network; %K peer-to-peer %K protocol; %K protocols; %K Slurpie %K strategy; %K systems; %K transfer %K transport %X We present Slurpie: a peer-to-peer protocol for bulk data transfer. Slurpie is specifically designed to reduce client download times for large, popular files, and to reduce load on servers that serve these files. Slurpie employs a novel adaptive downloading strategy to increase client performance, and employs a randomized backoff strategy to precisely control load on the server. We describe a full implementation of the Slurpie protocol, and present results from both controlled local-area and wide-area testbeds. Our results show that Slurpie clients improve performance as the size of the network increases, and the server is completely insulated from large flash crowds entering the Slurpie network. %B INFOCOM 2004. Twenty-third AnnualJoint Conference of the IEEE Computer and Communications Societies %V 2 %P 941 - 951 vol.2 - 941 - 951 vol.2 %8 2004/03// %G eng %R 10.1109/INFCOM.2004.1356981 %0 Book Section %B Media access: Social and psychological dimensions of new technology useMedia access: Social and psychological dimensions of new technology use %D 2004 %T Social and psychological influences on computer user frustration %A Bessiere,K. %A Ceaparu,I. %A Lazar,J. %A Robinson,J. %A Shneiderman, Ben %B Media access: Social and psychological dimensions of new technology useMedia access: Social and psychological dimensions of new technology use %P 91 - 103 %8 2004/// %G eng %0 Journal Article %J International Journal of Computer Vision %D 2004 %T SoftPOSIT: Simultaneous pose and correspondence determination %A David,P. %A DeMenthon,D. %A Duraiswami, Ramani %A Samet, Hanan %X The problem of pose estimation arises in many areas of computer vision, including object recognition, object tracking, site inspection and updating, and autonomous navigation when scene models are available. We present a new algorithm, called SoftPOSIT, for determining the pose of a 3D object from a single 2D image when correspondences between object points and image points are not known. The algorithm combines the iterative softassign algorithm (Gold and Rangarajan, 1996; Gold et al., 1998) for computing correspondences and the iterative POSIT algorithm (DeMenthon and Davis, 1995) for computing object pose under a full-perspective camera model. Our algorithm, unlike most previous algorithms for pose determination, does not have to hypothesize small sets of matches and then verify the remaining image points. Instead, all possible matches are treated identically throughout the search for an optimal pose. The performance of the algorithm is extensively evaluated in Monte Carlo simulations on synthetic data under a variety of levels of clutter, occlusion, and image noise. These tests show that the algorithm performs well in a variety of difficult scenarios, and empirical evidence suggests that the algorithm has an asymptotic run-time complexity that is better than previous methods by a factor of the number of image points. The algorithm is being applied to a number of practical autonomous vehicle navigation problems including the registration of 3D architectural models of a city to images, and the docking of small robots onto larger robots. %B International Journal of Computer Vision %V 59 %P 259 - 284 %8 2004/// %G eng %N 3 %R 10.1023/B:VISI.0000025800.10423.1f %0 Journal Article %J Proceedings of ICAD %D 2004 %T Sonification of geo-referenced data for auditory information seeking: Design principle and pilot study %A Zhao,H. %A Plaisant, Catherine %A Shneiderman, Ben %A Duraiswami, Ramani %X We present an Auditory Information Seeking Principle (AISP)(gist, navigate, filter, and details-on-demand) modeled after the visual information seeking mantra [1]. We propose that data sonification designs should conform to this principle. We also present some design challenges imposed by human auditory perception characteristics. To improve blind access to geo- referenced statistical data, we developed two preliminary sonifications adhering to the above AISP, an enhanced table and a spatial choropleth map. Our pilot study shows people can recognize geographic data distribution patterns on a real map with 51 geographic regions, in both designs. The study also shows evidence that AISP conforms to people’s information seeking strategies. Future work is discussed, including the improvement of the choropleth map design. %B Proceedings of ICAD %8 2004/// %G eng %0 Journal Article %J Journal of Computer and System Sciences %D 2004 %T Special issue: 35th Annual ACM Symposium on Theory of Computing %A Khanna,Sanjeev %A Srinivasan, Aravind %B Journal of Computer and System Sciences %V 69 %P 305 - 305 %8 2004/11// %@ 0022-0000 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022000004000881 %N 3 %R 10.1016/j.jcss.2004.06.004 %0 Conference Paper %B Proceedings of the fifteenth annual ACM-SIAM symposium on Discrete algorithms %D 2004 %T Structural and algorithmic aspects of massive social networks %A Eubank,Stephen %A Kumar,V. S. Anil %A Marathe,Madhav V. %A Srinivasan, Aravind %A Wang,Nan %X We study the algorithmic and structural properties of very large, realistic social contact networks. We consider the social network for the city of Portland, Oregon, USA, developed as a part of the TRANSIMS/EpiSims project at the Los Alamos National Laboratory. The most expressive social contact network is a bipartite graph, with two types of nodes: people and locations; edges represent people visiting locations on a typical day. Three types of results are presented. (i) Our empirical results show that many basic characteristics of the dataset are well-modeled by a random graph approach suggested by Fan Chung Graham and Lincoln Lu (the CL-model), with a power-law degree distribution. (ii) We obtain fast approximation algorithms for computing basic structural properties such as clustering coefficients and shortest paths distribution. We also study the dominating set problem for such networks; this problem arose in connection with optimal sensor-placement for disease-detection. We present a fast approximation algorithm for computing near-optimal dominating sets. (iii) Given the close approximations provided by the CL-model to our original dataset and the large data-volume, we investigate fast methods for generating such random graphs. We present methods that can generate such a random network in near-linear time, and show that these variants asymptotically share many key features of the CL-model, and also match the Portland social network.The structural results have been used to study the impact of policy decisions for controlling large-scale epidemics in urban environments. %B Proceedings of the fifteenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '04 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 718 - 727 %8 2004/// %@ 0-89871-558-X %G eng %U http://dl.acm.org/citation.cfm?id=982792.982902 %0 Conference Paper %B First International Workshop On Software Engineering for High Performance Computing System Applications %D 2004 %T Studying code development for high performance computing: the HPCS program %A Carver, J. %A Asgari, S. %A Basili, Victor R. %A Hochstein, L. %A Hollingsworth, Jeffrey K %A Shull, F. %A Zelkowitz, Marvin V %B First International Workshop On Software Engineering for High Performance Computing System Applications %P 32 - 32 %8 2004/// %G eng %0 Conference Paper %D 2004 %T The sybil attack in sensor networks: analysis & defenses %A Newsome, James %A Elaine Shi %A Song, Dawn %A Perrig, Adrian %K Security %K sensor networks %K sybil attack %X Security is important for many sensor network applications. A particularly harmful attack against sensor and ad hoc networks is known as the Sybil attack [6], where a node illegitimately claims multiple identities. This paper systematically analyzes the threat posed by the Sybil attack to wireless sensor networks. We demonstrate that the attack can be exceedingly detrimental to many important functions of the sensor network such as routing, resource allocation, misbehavior detection, etc. We establish a classification of different types of the Sybil attack, which enables us to better understand the threats posed by each type, and better design countermeasures against each type. We then propose several novel techniques to defend against the Sybil attack, and analyze their effectiveness quantitatively. %S IPSN '04 %I ACM %P 259 - 268 %8 2004 %@ 1-58113-846-6 %G eng %U http://doi.acm.org/10.1145/984622.984660 %0 Conference Paper %B Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on %D 2004 %T Systematic exploitation of data parallelism in hardware synthesis of DSP applications %A Sen,M. %A Bhattacharyya, Shuvra S. %K area-performance trade-off curve %K automatic Verilog code generation %K circuit optimisation %K data flow graphs %K data parallelism %K dedicated hardware implementation synthesis %K design tool %K Digital signal processing %K DSP applications %K embedded systems %K hardware description languages %K hardware synthesis %K high level synthesis %K high level synthesis algorithm %K IMAGE PROCESSING %K PARALLEL PROCESSING %K power consumption %K Signal processing %K synchronous dataflow graph %K video processing %X We describe an approach that we have explored for low-power synthesis and optimization of image, video, and digital signal processing (DSP) applications. In particular, we consider the systematic exploitation of data parallelism across the operations of an application dataflow graph when synthesizing a dedicated hardware implementation. Data parallelism occurs commonly in DSP applications, and provides flexible opportunities to increase throughput or lower power consumption. Exploiting this parallelism in a dedicated hardware implementation comes at the expense of increased resource requirements, which must be balanced carefully when applying the technique in a design tool. We propose a high level synthesis algorithm to determine the data parallelism factor for each computation, and, based on the area and performance trade-off curve, design an efficient hardware representation of the dataflow graph. For performance estimation, our approach uses a cyclostatic dataflow intermediate representation of the hardware structure under synthesis. We then apply an automatic hardware generation framework to build the actual circuit. %B Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on %V 5 %P V - 229-32 vol.5 - V - 229-32 vol.5 %8 2004/05// %G eng %R 10.1109/ICASSP.2004.1327089 %0 Journal Article %J Neural Computation %D 2004 %T Temporally asymmetric learning supports sequence processing in multi-winner self-organizing maps %A Schulz,R. %A Reggia, James A. %B Neural Computation %V 16 %P 535 - 561 %8 2004/// %G eng %N 3 %0 Conference Paper %B Proceedings of the 18th annual international conference on Supercomputing %D 2004 %T Time and space optimization for processing groups of multi-dimensional scientific queries %A Aryangat,Suresh %A Andrade,Henrique %A Sussman, Alan %K multi-query optimization %K Scientific databases %X Data analysis applications in areas as diverse as remote sensing and telepathology require operating on and processing very large datasets. For such applications to execute efficiently, careful attention must be paid to the storage, retrieval, and manipulation of the datasets. This paper addresses the optimizations performed by a high performance database system that processes groups of data analysis requests for these applications, which we call queries. The system performs end-to-end processing of the requests, formulated as PostgreSQL declarative queries. The queries are converted into imperative descriptions, multiple imperative descriptions are merged into a single execution plan, the plan is optimized to decrease execution time via common compiler optimization techniques, and, finally, the plan is optimized to decrease memory consumption. The last two steps are experimentally shown to effectively reduc the amount of time required while conserving memory space as a group of queries is processed by the database. %B Proceedings of the 18th annual international conference on Supercomputing %S ICS '04 %I ACM %C New York, NY, USA %P 95 - 105 %8 2004/// %@ 1-58113-839-3 %G eng %U http://doi.acm.org/10.1145/1006209.1006224 %R 10.1145/1006209.1006224 %0 Journal Article %J Applied Computational Electromagnetics Society journal %D 2004 %T TM Electromagnetic Scattering from 2D Multilayered Dielectric Bodies-Numerical Solution %A Duraiswami, Ramani %A Seydou,F. %A Gumerov, Nail A. %X An integral equation approach is derived for an electromagnetic scattering from an M arbitrary multilayereddielectric domain. The integral equation is valid for the 2D and 3D Helmholtz equation. Here we show the numerical solution for the 2D case by using the Nyström method. For validating the method we develop a mode matching method for the case when the domains are multilayered circular cylinders and give numerical results for illustrating the algorithm. %B Applied Computational Electromagnetics Society journal %P 100 - 107 %8 2004/// %G eng %N 2 %0 Conference Paper %B Proceedings of the 2004 ACM SIGMOD international conference on Management of data %D 2004 %T TOSS: an extension of TAX with ontologies and similarity queries %A Hung,Edward %A Deng,Yu %A V.S. Subrahmanian %K ontologies %K semantic integration of heterogeneous data %K similarity enhancement %K XML databases %X TAX is perhaps the best known extension of the relational algebra to handle queries to XML databases. One problem with TAX (as with many existing relational DBMSs) is that the semantics of terms in a TAX DB are not taken into account when answering queries. Thus, even though TAX answers queries with 100% precision, the recall of TAX is relatively low. Our TOSS system improves the recall of TAX via the concept of a similarity enhanced ontology (SEO). Intuitively, an ontology is a set of graphs describing relationships (such as isa, partof, etc.) between terms in a DB. An SEO also evaluates how similarities between terms (e.g. "J. Ullman", "Jeff Ullman", and "Jeffrey Ullman") affect ontologies. Finally, we show how the algebra proposed in TAX can be extended to take SEOs into account. The result is a system that provides a much higher answer quality than TAX does alone (quality is defined as the square root of the product of precision and recall). We experimentally evaluate the TOSS system on the DBLP and SIGMOD bibliographic databases and show that TOSS has acceptable performance. %B Proceedings of the 2004 ACM SIGMOD international conference on Management of data %S SIGMOD '04 %I ACM %C New York, NY, USA %P 719 - 730 %8 2004/// %@ 1-58113-859-8 %G eng %U http://doi.acm.org/10.1145/1007568.1007649 %R 10.1145/1007568.1007649 %0 Conference Paper %B Proceedings of the 19th annual ACM SIGPLAN conference on Object-oriented programming, systems, languages, and applications %D 2004 %T Transparent proxies for java futures %A Pratikakis,Polyvios %A Spacco,Jaime %A Hicks, Michael W. %K future %K java %K proxy %K Type inference %K type qualifier %X A proxy object is a surrogate or placeholder that controls access to another target object. Proxies can be used to support distributed programming, lazy or parallel evaluation, access control, and other simple forms of behavioral reflection. However, wrapper proxies (like futures or suspensions> for yet-to-be-computed results) can require significant code changes to be used in statically-typed languages, while proxies more generally can inadvertently violate assumptions of transparency, resulting in subtle bugs. To solve these problems, we have designed and implemented a simple framework for proxy programming that employs a static analysis based on qualifier inference, but with additional novelties. Code for using wrapper proxies is automatically introduced via a classfile-to-classfile transformation, and potential violations of transparency are signaled to the programmer. We have formalized our analysis and proven it sound. Our framework has a variety of applications, including support for asynchronous method calls returning futures. Experimental results demonstrate the benefits of our framework: programmers are relieved of managing and/or checking proxy usage, analysis times are reasonably fast, overheads introduced by added dynamic checks are negligible, and performance improvements can be significant. For example, changing two lines in a simple RMI-based peer-to-peer application and then using our framework resulted in a large performance gain. %B Proceedings of the 19th annual ACM SIGPLAN conference on Object-oriented programming, systems, languages, and applications %S OOPSLA '04 %I ACM %C New York, NY, USA %P 206 - 223 %8 2004/// %@ 1-58113-831-8 %G eng %U http://doi.acm.org/10.1145/1028976.1028994 %R 10.1145/1028976.1028994 %0 Journal Article %J Science %D 2004 %T Ubistatins Inhibit Proteasome-Dependent Degradation by Binding the Ubiquitin Chain %A Verma,Rati %A Peters,Noel R. %A D'Onofrio,Mariapina %A Tochtrop,Gregory P. %A Sakamoto,Kathleen M. %A Varadan,Ranjani %A Zhang,Mingsheng %A Coffino,Philip %A Fushman, David %A Deshaies,Raymond J. %A King,Randall W. %X To identify previously unknown small molecules that inhibit cell cycle machinery, we performed a chemical genetic screen in Xenopus extracts. One class of inhibitors, termed ubistatins, blocked cell cycle progression by inhibiting cyclin B proteolysis and inhibited degradation of ubiquitinated Sic1 by purified proteasomes. Ubistatins blocked the binding of ubiquitinated substrates to the proteasome by targeting the ubiquitin-ubiquitin interface of Lys48-linked chains. The same interface is recognized by ubiquitin-chain receptors of the proteasome, indicating that ubistatins act by disrupting a critical protein-protein interaction in the ubiquitin-proteasome system. %B Science %V 306 %P 117 - 120 %8 2004/10/01/ %G eng %U http://www.sciencemag.org/cgi/content/abstract/sci;306/5693/117 %N 5693 %R 10.1126/science.1100946 %0 Journal Article %J Artificial Intelligence in Medicine %D 2004 %T Understanding tuberculosis epidemiology using structured statistical models %A Getoor, Lise %A Rhee,Jeanne T %A Koller,Daphne %A Small,Peter %K Bayesian networks %K epidemiology %K Probabilistic and statistical relational models %K Tuberculosis %X Molecular epidemiological studies can provide novel insights into the transmission of infectious diseases such as tuberculosis. Typically, risk factors for transmission are identified using traditional hypothesis-driven statistical methods such as logistic regression. However, limitations become apparent in these approaches as the scope of these studies expand to include additional epidemiological and bacterial genomic data. Here we examine the use of Bayesian models to analyze tuberculosis epidemiology. We begin by exploring the use of Bayesian networks (BNs) to identify the distribution of tuberculosis patient attributes (including demographic and clinical attributes). Using existing algorithms for constructing BNs from observational data, we learned a BN from data about tuberculosis patients collected in San Francisco from 1991 to 1999. We verified that the resulting probabilistic models did in fact capture known statistical relationships. Next, we examine the use of newly introduced methods for representing and automatically constructing probabilistic models in structured domains. We use statistical relational models (SRMs) to model distributions over relational domains. SRMs are ideally suited to richly structured epidemiological data. We use a data-driven method to construct a statistical relational model directly from data stored in a relational database. The resulting model reveals the relationships between variables in the data and describes their distribution. We applied this procedure to the data on tuberculosis patients in San Francisco from 1991 to 1999, their Mycobacterium tuberculosis strains, and data on contact investigations. The resulting statistical relational model corroborated previously reported findings and revealed several novel associations. These models illustrate the potential for this approach to reveal relationships within richly structured data that may not be apparent using conventional statistical approaches. We show that Bayesian methods, in particular statistical relational models, are an important tool for understanding infectious disease epidemiology. %B Artificial Intelligence in Medicine %V 30 %P 233 - 256 %8 2004/03// %@ 0933-3657 %G eng %U http://www.sciencedirect.com/science/article/pii/S0933365703001337 %N 3 %R 10.1016/j.artmed.2003.11.003 %0 Journal Article %J Media access: social and psychological dimensions of new technology use %D 2004 %T Universal Usability: Pushing Human-Computer Interaction Research to Empower Every Citizen %A Shneiderman, Ben %B Media access: social and psychological dimensions of new technology use %P 255 - 255 %8 2004/// %G eng %0 Report %D 2004 %T Use of Minimal Lexical Conceptual Structures for Single-Document Summarization %A Dorr, Bonnie J %A Habash,Nizar Y %A Monz,Christof %A Schwartz,Richard %K *COMPUTATIONAL LINGUISTICS %K AUTOMATIC HEADLINE GENERATION %K CYBERNETICS %K DOCUMENT SUMMARIZATION %K Information retrieval %K INFORMATION SCIENCE %K linguistics %K Machine translation %X This reports provides an overview of the findings and software that have evolved from the Use of Minimal Lexical Conceptual Structures for Single-Document Summarization project over the last six months. We present the major goals that have been achieved and discuss some of the open issues that we intend to address in the near future. This report also contains some details on the usage of some software that has been implemented during the project. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %8 2004/06// %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA458797 %0 Journal Article %J Sprache und Datenverarbeitung %D 2004 %T User interface design with speech technologies : A cognitive limitations review %A Shneiderman, Ben %K Communication homme-machine %K computational linguistics %K Human-computer communication %K interface %K Linguistique informatique %K Speech processing %K Traitement automatique de la parole %B Sprache und Datenverarbeitung %V 28 %P 101 - 109 %8 2004/// %@ 0343-5202 %G eng %U http://cat.inist.fr/?aModele=afficheN&cpsidt=16961708 %N 2 %0 Book Section %B Bacterial Artificial Chromosomes %D 2004 %T Using the TIGR Assembler in Shotgun Sequencing Projects %A Pop, Mihai %A Kosack,Dan %E Zhao,Shaying %E Stodolsky,Marvin %X The TIGR Assembler (TA) ( 1 ) is the sequence assembly program used in sequencing projects at The Institute for Genomic Research (TIGR). Development of the TA was based on the experience obtained in more than 20 sequencing projects completed at TIGR ( see http://www.tigr.org ). This extensive experience led to a sequence assembler that produces few misassemblies ( 2 , 3 ) and has been used successfully in whole-genome shotgun sequencing of prokaryotic and eukaryotic organisms, bacterial artificial chromosome-based sequencing of eukaryotic organisms, and expressed sequence tag assembly. %B Bacterial Artificial Chromosomes %S Methods in Molecular Biology %I Humana Press %V 255 %P 279 - 294 %8 2004/// %@ 978-1-59259-752-9 %G eng %U http://dx.doi.org/10.1385/1-59259-752-1:279 %0 Conference Paper %B Proceedings of the European Conference on Artificial Intelligence (ECAI) %D 2004 %T Utilizing volatile external information during planning %A Au,T. C %A Nau, Dana S. %A V.S. Subrahmanian %B Proceedings of the European Conference on Artificial Intelligence (ECAI) %P 647 - 651 %8 2004/// %G eng %0 Journal Article %J Software Reuse: Methods, Techniques, and Tools %D 2004 %T Validating quality of service for reusable software via model-integrated distributed continuous quality assurance %A Krishna,A. %A Schmidt,D. %A Memon, Atif M. %A Porter, Adam %A Sevilla,D. %X Quality assurance (QA) tasks, such as testing, profiling, and performance evaluation, have historically been done in-house on developer-generated workloads and regression suites. Performance-intensive systems software, such as that found in the scientific computing grid and distributed real-time and embedded (DRE) domains, increasingly run on heterogeneous combinations of OS, compiler, and hardware platforms. Such software has stringent quality of service (QoS) requirements and often provides a variety of configuration options to optimize QoS. As a result, QA performed solely in-house is inadequate since it is hard to manage software variability, i.e., ensuring software quality on all supported target platforms across all desired configuration options. This paper describes how the Skoll project is addressing these issues by developing advanced QA processes and tools that leverage the extensive computing resources of user communities in a distributed, continuous manner to improve key software quality attributes. %B Software Reuse: Methods, Techniques, and Tools %P 286 - 295 %8 2004/// %G eng %R 10.1007/978-3-540-27799-6_24 %0 Journal Article %J Microbiology and immunology %D 2004 %T Variation of toxigenic Vibrio cholerae O1 in the aquatic environment of Bangladesh and its correlation with the clinical strains %A Islam,M. S. %A Talukder,K. A. %A Khan,N. H. %A Mahmud,Z. H. %A Rahman,M. Z. %A Nair,G. B. %A Siddique,A. K. M. %A Yunus,M. %A Sack,D. A. %A Sack,R. B. %A Rita R Colwell %B Microbiology and immunology %V 48 %P 773 - 777 %8 2004/// %G eng %N 10 %0 Journal Article %J Lecture Notes in Computer Science %D 2004 %T Visual Content Mining in Multimedia Documents-Indexing Issues in Supporting Similarity Searching %A Samet, Hanan %B Lecture Notes in Computer Science %V 3332 %P 463 - 470 %8 2004/// %G eng %0 Journal Article %J BMC Bioinformatics %D 2004 %T Visualization and analysis of microarray and gene ontology data with treemaps %A Baehrecke,Eric H %A Dang,Niem %A Babaria,Ketan %A Shneiderman, Ben %X The increasing complexity of genomic data presents several challenges for biologists. Limited computer monitor views of data complexity and the dynamic nature of data in the midst of discovery increase the challenge of integrating experimental results with information resources. The use of Gene Ontology enables researchers to summarize results of quantitative analyses in this framework, but the limitations of typical browser presentation restrict data access. %B BMC Bioinformatics %V 5 %P 84 - 84 %8 2004/06/28/ %@ 1471-2105 %G eng %U http://www.biomedcentral.com/1471-2105/5/84 %N 1 %R 10.1186/1471-2105-5-84 %0 Journal Article %J Proceedings of the National Academy of Sciences of the United States of America %D 2004 %T Whole-genome shotgun assembly and comparison of human genome assemblies %A Istrail,S. %A Sutton,G. G %A Florea,L. %A Halpern,A. L %A Mobarry,C. M %A Lippert,R. %A Walenz,B. %A Shatkay,H. %A Dew,I. %A Miller,J. R %A others %B Proceedings of the National Academy of Sciences of the United States of America %V 101 %P 1916 - 1916 %8 2004/// %G eng %N 7 %0 Journal Article %J SIGKDD Explor. Newsl. %D 2004 %T A workshop report: mining for and from the Semantic Web at KDD 2004 %A Hotho,Andreas %A Sure,York %A Getoor, Lise %X The international workshop on Mining for and from the Semantic Web (MSW) at the KDD 2004 successfully brought together people from the communities Semantic Web and Knowledge Discovery. The goal of the workshop was to strengthen the communication and interaction between these communities as there is currently an agreement that both can benefit largely from each other. Overall the contributions and discussions showed that today mining for the semantic web is the primarily targeted research area. However, with a growing amount of available semantic web data also mining from the semantic web will be more prominent. %B SIGKDD Explor. Newsl. %V 6 %P 142 - 143 %8 2004/12// %@ 1931-0145 %G eng %U http://doi.acm.org/10.1145/1046456.1046482 %N 2 %R 10.1145/1046456.1046482 %0 Book %D 2003 %T 19th Biennial Conference on Mechanical Vibration and Noise %A Gupta,A. K. %A Fadel,G. M. %A Lewicki,D. G. %A Shabana,A. A. %A Royston,T. J. %A Gupta,S.K. %A Hayes,C. %A Herrmann,J.W. %A Dym,C. L. %A Schmidt,L. C. %I American Society of Mechanical Engineers %8 2003/// %G eng %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2003. IEEE %D 2003 %T A 2D profile reconstruction in a multilayered waveguide structure %A Seydou,F. %A Duraiswami, Ramani %A Seppanen,T. %K (mathematics); %K 2D %K dual %K duality %K electromagnetic %K equations; %K inhomogeneous %K Maxwell %K media; %K method; %K multilayered %K multilayers; %K profile %K reconstruction; %K scattering; %K space %K structure; %K theory; %K wave %K waveguide %K waveguides; %X We discuss the problem of finding a profile or its location for a 2D scattering of electromagnetic waves with fixed frequencies in a waveguide multilayered domain. We use the dual space method (DSM) of Colton and Monk. Our goal is to extend our previous work of TE and TM cases to the more complicated case of a waveguide. We emphasize on the frequency range in the reconstruction. %B Antennas and Propagation Society International Symposium, 2003. IEEE %V 1 %P 531 - 534 vol.1 - 531 - 534 vol.1 %8 2003/06// %G eng %R 10.1109/APS.2003.1217513 %0 Journal Article %J Journal of Infectious DiseasesJ Infect Dis. %D 2003 %T A 4-Year Study of the Epidemiology of Vibrio Cholerae in Four Rural Areas of Bangladesh %A Sack,R. Bradley %A Siddique,A. Kasem %A Longini,Ira M. %A Nizam,Azhar %A Yunus,Md %A M. Sirajul Islam %A Morris,Jr %A Ali,Afsar %A Huq,Anwar %A Nair,G. Balakrish %A Qadri,Firdausi %A Faruque,Shah M. %A Sack,David A. %A Rita R Colwell %X How Vibrio cholerae spreads around the world and what determines its seasonal peaks in endemic areas are not known. These features of cholera have been hypothesized to be primarily the result of environmental factors associated with aquatic habitats that can now be identified. Since 1997, fortnightly surveillance in 4 widely separated geographic locations in Bangladesh has been performed to identify patients with cholera and to collect environmental data. A total of 5670 patients (53% <5 years of age) have been studied; 14.3% had cholera (10.4% due to V. cholerae O1 El Tor, 3.8% due to O139). Both serogroups were found in all locations; outbreaks were seasonal and often occurred simultaneously. Water-use patterns showed that bathing and washing clothes in tube-well water was significantly protective in two of the sites. These data will be correlated with environmental factors, to develop a model for prediction of cholera outbreaks %B Journal of Infectious DiseasesJ Infect Dis. %V 187 %P 96 - 101 %8 2003/01/01/ %@ 0022-1899, 1537-6613 %G eng %U http://jid.oxfordjournals.org/content/187/1/96 %N 1 %R 10.1086/345865 %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T Accessing diverse geo-referenced data sources with the SAND spatial DBMS %A Sankaranarayanan,Jagan %A Tanin,Egemen %A Samet, Hanan %A Brabec,Frantisek %X The Internet has become the most frequently accessed medium for obtaining various types of data. In particular, government agencies, academic institutions, and private enterprises have published gigabytes of geo-referenced data on the Web. However, to obtain geo-referenced data from the Web successfully, systems must be designed to be capable of understanding the data sets published in different data formats. Also, even if the data sets are available in a simple known format, they often have poorly defined structures. With these issues in mind, we have developed an Internet-enabled data collection and conversion utility that interfaces with our prototype spatial database system, SAND. Using this utility, data can be retrieved from many different sources on the Web and converted into a format understandable by the SAND spatial database management system. Our collection and conversion utility is able to import the most popular data formats; namely, ESRI Shapefiles, Microsoft Excel files, HTML files, and GML files. Data in unstructured formats are verified for correct selection of the data types and handling of missing tuples before the insertion operation into the database. Moreover, our utility makes it possible to download any nonspatial data set and combine it internally with a relevant spatial data set. These features are accessible through a spreadsheet-like interface for online editing and structuring of data. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 4 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123206 %0 Journal Article %J Parallel and Distributed Systems, IEEE Transactions on %D 2003 %T An adaptive quality of service aware middleware for replicated services %A Krishnamurthy,Sudha %A Sanders,W. H. %A Michel Cukier %K distributed resource sharing %K Middleware %K online performance monitoring %K probabilistic modeling %K QoS specification %K quality of service %K replica consistency %K replicated services %K resource allocation %K time-sensitive client applications %K timeliness constraints %X A dependable middleware should be able to adaptively share the distributed resources it manages in order to meet diverse application requirements, even when the quality of service (QoS) is degraded due to uncertain variations in load and unanticipated failures. We have addressed this issue in the context of a dependable middleware that adaptively manages replicated servers to deliver a timely and consistent response to time-sensitive client applications. These applications have specific temporal and consistency requirements, and can tolerate a certain degree of relaxed consistency in exchange for better response time. We propose a flexible QoS model that allows clients to specify their timeliness and consistency constraints. We also propose an adaptive framework that dynamically selects replicas to service a client's request based on the prediction made by probabilistic models. These models use the feedback from online performance monitoring of the replicas to provide probabilistic guarantees for meeting a client's QoS specification. The experimental results we have obtained demonstrate the role of feedback and the efficacy of simple analytical models for adaptively sharing the available replicas among the users under different workload scenarios. %B Parallel and Distributed Systems, IEEE Transactions on %V 14 %P 1112 - 1125 %8 2003/11// %@ 1045-9219 %G eng %N 11 %R 10.1109/TPDS.2003.1247672 %0 Book Section %B Video RegistrationVideo Registration %D 2003 %T Airborne Video Registration for Activity Monitoring %A Shekhar,Chandra %A Chellapa, Rama %E Shah,Mubarak %E Kumar,Rakesh %E Shah,Mubarak %X The analysis of ground activities involving vehicles and humans is of interest in airborne surveillance applications. The high spatiotemporal resolution of a video camera makes it the sensor of choice for such applications. In the past, video sensors were not preferred, precisely because of the high volume of data generated, which had to be processed manually for the most part. Further, the moving platform gives rise to numerous false alarms. %B Video RegistrationVideo Registration %S The Kluwer International Series in Video Computing %I Springer US %V 5 %P 130 - 143 %8 2003/// %@ 978-1-4615-0459-7 %G eng %U http://dx.doi.org/10.1007/978-1-4615-0459-7_6 %0 Journal Article %J The Journal of the Acoustical Society of America %D 2003 %T Analysis of particular phononic structures using single integral equations %A Seydou,Fadoulourahmane %A Duraiswami, Ramani %A Gumerov, Nail A. %X A fast method for determining phononic (and photonic) bandgaps in composite materials is developed. It is known that in the propagation of waves in a 3D medium containing N scatterers arranged periodically, there exist refractive indices for which such structures have bandgaps, i.e., frequencies for which no waves can propagate inside. Our task is to find the frequencies that generate these prohibited waves. This requires the solution of an eigenvalue problem for the Helmholtz operator. To solve this problem we choose an alternate route which uses boundary integral equations. We derive a single integral equation on each of the interfaces between the outer region and the scatterers, considering a general transmission boundary condition, by using a hybrid method using layer potentials and Green’s formula. This approach reduces the number of unknowns considerably in comparison to other methods, but requires the treatment of large dense matrices and many matrix vector multiplications. To remedy this, we use the Fast Multipole Method. For solving the eigenvalue problem we discuss two methods: the Newton method and a method based on the Cauchy formula. Details of the numerical implementation, and results will be presented for different cases: sound hard, sound soft, impedance and transmission boundary conditions. [Work partially supported by NSF Award 0219681 is gratefully acknowledged.] %B The Journal of the Acoustical Society of America %V 113 %P 2284 - 2284 %8 2003/// %G eng %U http://link.aip.org/link/?JAS/113/2284/1 %N 4 %0 Journal Article %J ACM Computer Communications Review %D 2003 %T An Analysis of the Layer 2 Handoff costs in Wireless Local Area Networks %A Mishra,A. %A Shin,M. H. %A Arbaugh, William A. %B ACM Computer Communications Review %V 33 %8 2003/// %G eng %N 2 %0 Journal Article %J Journal of Computer and System Sciences %D 2003 %T On the approximability of clique and related maximization problems %A Srinivasan, Aravind %K Approximation algorithms %K Clique %K Inapproximability %K Independent set %K Packing integer programs %K random sampling %X We consider approximations of the form n1−o(1) for the Maximum Clique problem, where n is the number of vertices in the input graph and where the “o(1)” term goes to zero as n increases. We show that sufficiently strong negative results for such problems, which we call strong inapproximability results, have interesting consequences for exact computation. A simple sampling method underlies most of our results. %B Journal of Computer and System Sciences %V 67 %P 633 - 651 %8 2003/11// %@ 0022-0000 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022000003001107 %N 3 %R 10.1016/S0022-0000(03)00110-7 %0 Journal Article %J SIAM Journal on Computing %D 2003 %T Approximating the domatic number %A Feige,U. %A Halldórsson,M.M. %A Kortsarz,G. %A Srinivasan, Aravind %X A set of vertices in a graph is a dominating set if every vertex outside the set hasa neighbor in the set. The domatic number problem is that of partitioning the vertices of a graph into the maximum number of disjoint dominating sets. Let n denote the number of vertices, δ the minimum degree, and ∆ the maximum degree. We show that every graph has a domatic partition with (1 − o(1))(δ + 1)/ ln n dominating sets and, moreover, that such a domatic partition can be found in polynomial-time. This implies a (1 + o(1)) ln n-approximation algorithm for domatic number, since the domatic number is always at most δ+1. We also show this to be essentially best possible. Namely, extending the approximation hardness of set cover by combining multiprover protocols with zero-knowledge techniques, we show that for every ϵ > 0, a (1 − ϵ) ln n-approximation implies that NP ⊆ DTIME(nO(log log n)). This makes domatic number the first natural maximization problem (known to the authors) that is provably approximable to within polylogarithmic factors but no better. We also show that every graph has a domatic partition with (1 − o(1))(δ + 1)/ ln ∆ dominating sets, where the “o(1)” term goes to zero as ∆ increases. This can be turned into an efficient algorithm that produces a domatic partition of Ω(δ/ ln ∆) sets. %B SIAM Journal on Computing %V 32 %P 172 - 195 %8 2003/// %G eng %N 1 %0 Book Section %B Approximation, Randomization, and Combinatorial Optimization: Algorithms and Techniques %D 2003 %T Approximation Algorithms for Channel Allocation Problems in Broadcast Networks %A Gandhi,Rajiv %A Khuller, Samir %A Srinivasan, Aravind %A Wang,Nan %E Arora,Sanjeev %E Jansen,Klaus %E Rolim,José %E Sahai,Amit %X We study two packing problems that arise in the area of dissemination-based information systems; a second theme is the study of distributed approximation algorithms. The problems considered have the property that the space occupied by a collection of objects together could be significantly less than the sum of the sizes of the individual objects. In the Channel Allocation Problem , there are users who request subsets of items. There are a fixed number of channels that can carry an arbitrary amount of information. Each user must get all of the requested items from one channel, i.e., all the data items of each request must be broadcast on some channel. The load on any channel is the number of items that are broadcast on that channel; the objective is to minimize the maximum load on any channel. We present approximation algorithms for this problem and also show that the problem is MAX-SNP hard. The second problem is the Edge Partitioning Problem addressed by Goldschmidt, Hochbaum, Levin, and Olinick ( Networks, 41:13-23, 2003 ). Each channel here can deliver information to at most k users, and we aim to minimize the total load on all channels. We present an O ( n⅓ )–approximation algorithm and also show that the algorithm can be made fully distributed with the same approximation guarantee; we also generalize to the case of hypergraphs. %B Approximation, Randomization, and Combinatorial Optimization: Algorithms and Techniques %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2764 %P 821 - 826 %8 2003/// %@ 978-3-540-40770-6 %G eng %U http://dx.doi.org/10.1007/978-3-540-45198-3_5 %0 Journal Article %J Computers, IEEE Transactions on %D 2003 %T AQuA: an adaptive architecture that provides dependable distributed objects %A Ren,Yansong %A Bakken,D. E. %A Courtney,T. %A Michel Cukier %A Karr,D. A. %A Rubel,P. %A Sabnis,C. %A Sanders,W. H. %A Schantz,R.E. %A Seri,M. %K active replication pass-first scheme %K adaptive architecture %K adaptive fault tolerance %K AQuA %K CORBA %K data consistency %K data integrity %K dependable distributed objects %K distributed object management %K performance measurements %K quality of service %K replicated dependability manager %K replication schemes %K software fault tolerance %K system resources %X Building dependable distributed systems from commercial off-the-shelf components is of growing practical importance. For both cost and production reasons, there is interest in approaches and architectures that facilitate building such systems. The AQuA architecture is one such approach; its goal is to provide adaptive fault tolerance to CORBA applications by replicating objects. The AQuA architecture allows application programmers to request desired levels of dependability during applications' runtimes. It provides fault tolerance mechanisms to ensure that a CORBA client can always obtain reliable services, even if the CORBA server object that provides the desired services suffers from crash failures and value faults. AQuA includes a replicated dependability manager that provides dependability management by configuring the system in response to applications' requests and changes in system resources due to faults. It uses Maestro/Ensemble to provide group communication services. It contains a gateway to intercept standard CORBA IIOP messages to allow any standard CORBA application to use AQuA. It provides different types of replication schemes to forward messages reliably to the remote replicated objects. All of the replication schemes ensure strong, data consistency among replicas. This paper describes the AQuA architecture and presents, in detail, the active replication pass-first scheme. In addition, the interface to the dependability manager and the design of the dependability manager replication are also described. Finally, we describe performance measurements that were conducted for the active replication pass-first scheme, and we present results from our study of fault detection, recovery, and blocking times. %B Computers, IEEE Transactions on %V 52 %P 31 - 50 %8 2003/01// %@ 0018-9340 %G eng %N 1 %R 10.1109/TC.2003.1159752 %0 Journal Article %J IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS SPEECH AND SIGNAL PROCESSING %D 2003 %T AUDIO-P2. 1: PITCH AND TIMBRE MANIPULATIONS USING CORTICAL REPRESENTATION OF SOUND %A Zotkin,Dmitry N %A Shamma,S.A. %A Ru,P. %A Duraiswami, Ramani %A Davis, Larry S. %B IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS SPEECH AND SIGNAL PROCESSING %V 5 %8 2003/// %G eng %0 Journal Article %J Proc. International Conference on Auditory Display (ICAD’03) %D 2003 %T Augmented intelligibility in simultaneous multi-talker environments %A Mesgarani,N. %A Shamma,S. %A Grant,K.W. %A Duraiswami, Ramani %X Speech intelligibility can be severely compromised inenvironments where there are several competing speakers. It may be possible, however, to improve speech intelligibility in such environments by manipulating certain acoustic parameters known to facilitate the segregation of competing signals. The first step in designing such a feature-rich audio display is to understand the significant elements of human auditory perception that affect information transmission capacity. We review a series of experiments to examine the impact of different audio-display design parameters on overall intelligibility of simultaneous speech messages and show how using these features may improve intelligibility. %B Proc. International Conference on Auditory Display (ICAD’03) %P 71 - 74 %8 2003/// %G eng %0 Conference Paper %B Proceedings of the 16th annual ACM symposium on User interface software and technology %D 2003 %T Automatic thumbnail cropping and its effectiveness %A Suh,B. %A Ling,H. %A Bederson, Benjamin B. %A Jacobs, David W. %B Proceedings of the 16th annual ACM symposium on User interface software and technology %P 95 - 104 %8 2003/// %G eng %0 Conference Paper %B Proceedings of the 16th annual ACM symposium on User interface software and technology %D 2003 %T Automatic thumbnail cropping and its effectiveness %A Suh,Bongwon %A Ling,Haibin %A Bederson, Benjamin B. %A Jacobs, David W. %K Face detection %K image cropping %K saliency map %K thumbnail %K usability study %K visual search %K zoomable user interfaces %X Thumbnail images provide users of image retrieval and browsing systems with a method for quickly scanning large numbers of images. Recognizing the objects in an image is important in many retrieval tasks, but thumbnails generated by shrinking the original image often render objects illegible. We study the ability of computer vision systems to detect key components of images so that automated cropping, prior to shrinking, can render objects more recognizable. We evaluate automatic cropping techniques 1) based on a general method that detects salient portions of images, and 2) based on automatic face detection. Our user study shows that these methods result in small thumbnails that are substantially more recognizable and easier to find in the context of visual search. %B Proceedings of the 16th annual ACM symposium on User interface software and technology %S UIST '03 %I ACM %C New York, NY, USA %P 95 - 104 %8 2003/// %@ 1-58113-636-6 %G eng %U http://doi.acm.org/10.1145/964696.964707 %R 10.1145/964696.964707 %0 Book Section %B The Semantic Web - ISWC 2003 %D 2003 %T Automating DAML-S Web Services Composition Using SHOP2 %A Wu,Dan %A Parsia,Bijan %A Sirin,Evren %A Hendler,James %A Nau, Dana S. %E Fensel,Dieter %E Sycara,Katia %E Mylopoulos,John %K Computer science %X The DAML-S Process Model is designed to support the application of AI planning techniques to the automated composition of Web services. SHOP2 is an Hierarchical Task Network (HTN) planner well-suited for working with the Process Model. We have proven the correspondence between the semantics of SHOP2 and the situation calculus semantics of the Process Model. We have also implemented a system which soundly and completely plans over sets of DAML-S descriptions using a SHOP2 planner, and then executes the resulting plans over the Web. We discuss the challenges and difficulties of using SHOP2 in the information-rich and human-oriented context of Web services. %B The Semantic Web - ISWC 2003 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2870 %P 195 - 210 %8 2003/// %@ 978-3-540-20362-9 %G eng %U http://www.springerlink.com/content/rm5ejwlmbw0mdv97/abstract/ %0 Journal Article %J Commun. ACM %D 2003 %T Bistro: a scalable and secure data transfer service for digital government applications %A Golubchik,Leana %A Cheng,William C. %A Chou,Cheng-Fu %A Khuller, Samir %A Samet, Hanan %A Wan,C. Justin %X Government at all levels is a major collector and provider of data.Our project focuses on the collection of data over wide-area networks (WANs) and addresses the scalability issues that arise in the context of Internet-based massive data collection applications. Furthermore, security, due to the need for privacy and integrity of the data, is a central issue for data collection applications that use a public infrastructure such as the Internet. Numerous digital government applications require data collection over WANs [5].One compelling example of such an application is the Internal Revenue Service's electronic submission of income tax forms. Other digital government applications include collecting census data, federal statistics, and surveys; gathering and tallying of electronic votes; collecting crime data for the U.S. Justice department; collecting data from sensors for disaster response applications; collecting data from geological surveys; collecting electronic filings of patents, permits, and securities (for SEC) applications; grant proposals and contract bids submissions; and so on. All these applications have scalability and security needs in common.The poor performance that may be experienced by current digital government users, given the existing state of technology (as in Figure 1a), is largely due to how (independent) data transfers using TCP/IP work over the Internet. TCP/IP is good at equally sharing bandwidth between data streams, which in large-scale applications can lead to poor performance for individual clients (as they receive only a very small share of this bandwidth). Given that TCP/IP is here to stay for the foreseeable future, what is needed is a scalable yet cost- effective solution that can be easily deployed over the existing Internet technology.We are designing and developing a system called Bistro, which addresses the scalability needs of digital government data collection applications while allowing them to share the same infrastructure and resources efficiently, cost-effectively, and securely [1]. Bistro's basic approach is to introduce intermediate hosts---bistros---which allow replacement of a traditionally "synchronized client push" approach with a "nonsynchronized combination of client-push and server-pull" approach (as depicted in Figure 1b). This in turn allows spreading of the workload on the destination server and the network over time, with subsequent elimination of hot spots as well as significant improvements in performance for both clients and servers. Our ongoing research [2, 4] indicates that orders of magnitude of improvement can be achieved with the Bistro architecture and the corresponding data collection algorithms it affords.Bistro's design allows for a gradual deployment and experimentation over the Internet (by simply downloading Bistro server software and installing it on public servers). Bistro's security protocol and trust structure [3] are designed such that only encrypted data travels through (not necessarily trusted) bistros. This means a government agency does not need to trust bistros installed by other agencies or commercial institutions. At the same time, these (untrusted) bistros can significantly improve the agency's data collection performance. Each application (within each agency) can have its own scalability, security, fault tolerance, and other data collection needs, and these applications and agencies can still share available resources, if so desired, across all Bistro servers.We believe an appropriately designed single infrastructure such as Bistro can address all digital government wide-area data collection needs in a scalable, secure, and cost-effective manner. (For more information, see bourbon.usc.edu/iml/bistro/. %B Commun. ACM %V 46 %P 50 - 51 %8 2003/01// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/602421.602448 %N 1 %R 10.1145/602421.602448 %0 Journal Article %J Communications of the ACM %D 2003 %T Bistro: a scalable and secure data transfer service for digital government applications %A Golubchik,Leana %A Cheng,William C. %A Chou,Cheng-Fu %A Khuller, Samir %A Samet, Hanan %A Wan,C. Justin %X Government at all levels is a major collector and provider of data.Our project focuses on the collection of data over wide-area networks (WANs) and addresses the scalability issues that arise in the context of Internet-based massive data collection applications. Furthermore, security, due to the need for privacy and integrity of the data, is a central issue for data collection applications that use a public infrastructure such as the Internet. Numerous digital government applications require data collection over WANs [5].One compelling example of such an application is the Internal Revenue Service's electronic submission of income tax forms. Other digital government applications include collecting census data, federal statistics, and surveys; gathering and tallying of electronic votes; collecting crime data for the U.S. Justice department; collecting data from sensors for disaster response applications; collecting data from geological surveys; collecting electronic filings of patents, permits, and securities (for SEC) applications; grant proposals and contract bids submissions; and so on. All these applications have scalability and security needs in common.The poor performance that may be experienced by current digital government users, given the existing state of technology (as in Figure 1a), is largely due to how (independent) data transfers using TCP/IP work over the Internet. TCP/IP is good at equally sharing bandwidth between data streams, which in large-scale applications can lead to poor performance for individual clients (as they receive only a very small share of this bandwidth). Given that TCP/IP is here to stay for the foreseeable future, what is needed is a scalable yet cost- effective solution that can be easily deployed over the existing Internet technology.We are designing and developing a system called Bistro, which addresses the scalability needs of digital government data collection applications while allowing them to share the same infrastructure and resources efficiently, cost-effectively, and securely [1]. Bistro's basic approach is to introduce intermediate hosts---bistros---which allow replacement of a traditionally "synchronized client push" approach with a "nonsynchronized combination of client-push and server-pull" approach (as depicted in Figure 1b). This in turn allows spreading of the workload on the destination server and the network over time, with subsequent elimination of hot spots as well as significant improvements in performance for both clients and servers. Our ongoing research [2, 4] indicates that orders of magnitude of improvement can be achieved with the Bistro architecture and the corresponding data collection algorithms it affords.Bistro's design allows for a gradual deployment and experimentation over the Internet (by simply downloading Bistro server software and installing it on public servers). Bistro's security protocol and trust structure [3] are designed such that only encrypted data travels through (not necessarily trusted) bistros. This means a government agency does not need to trust bistros installed by other agencies or commercial institutions. At the same time, these (untrusted) bistros can significantly improve the agency's data collection performance. Each application (within each agency) can have its own scalability, security, fault tolerance, and other data collection needs, and these applications and agencies can still share available resources, if so desired, across all Bistro servers.We believe an appropriately designed single infrastructure such as Bistro can address all digital government wide-area data collection needs in a scalable, secure, and cost-effective manner. (For more information, see bourbon.usc.edu/iml/bistro/. %B Communications of the ACM %V 46 %P 50 - 51 %8 2003/01// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/602421.602448 %N 1 %R 10.1145/602421.602448 %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2003. IEEE %D 2003 %T A boundary element method for electromagnetic scattering by multiple cylinders %A Seydou,F. %A Duraiswami, Ramani %A Seppanen,T. %K algorithm; %K boundary %K boundary-elements %K cylinders; %K electromagnetic %K element %K equations; %K Fredholm %K integral %K method; %K methods; %K multiple %K Nystrom %K scattering; %K two-dimensional %K type %K wave %X An integral equation approach is derived for an electromagnetic scattering from M multiple cylinders. The problem is two-dimensional and the integral equation is solved using the Nystrom method. We give numerical examples that illustrate the algorithm. %B Antennas and Propagation Society International Symposium, 2003. IEEE %V 3 %P 516 - 519 vol.3 - 516 - 519 vol.3 %8 2003/06// %G eng %R 10.1109/APS.2003.1219899 %0 Journal Article %J The craft of information visualization: readings and reflections %D 2003 %T Bringing Treasures to the Surface: Previews and Overviews in a Prototype for the Library of Congress National Digital Library %A Marchionini,A.S. %A Nation,D. %A Karasik,S. %A Cronnell,T. %A Shneiderman, Ben %A Spotfire,V.D.M.U. %A Bederson, Benjamin B. %A Data,T.S. %A Hochheiser,H. %A Shneiderman, Ben %A others %B The craft of information visualization: readings and reflections %P 388 - 388 %8 2003/// %G eng %0 Book Section %B The Craft of Information VisualizationThe Craft of Information Visualization %D 2003 %T Broadening Access to Large Online Databases by Generalizing Query Previews %A Tanin,Egemen %A Plaisant, Catherine %A Shneiderman, Ben %E Bederson, Benjamin B. %E Shneiderman, Ben %X Companies, government agencies, and other types of organizations are making their large databases available to the world over the Internet. Current database front-ends do not give users information about the distribution of data. This leads many users to waste time and network resources posing queries that have either zero-hit or mega-hit result sets. Query previews form a novel visual approach for browsing large databases. Query previews supply data distribution information about the database that is being searched and give continuous feedback about the size of the result set for the query as it is being formed. On the other hand, query previews use only a few pre-selected attributes of the database. The distribution information is displayed only on these attributes. Unfortunately, many databases are formed of numerous relations and attributes. This paper introduces a generalization of query previews. We allow users to browse all of the relations and attributes of a database using a hierarchical browser. Any of the attributes can be used to display the distribution information, making query previews applicable to many public online databases. %B The Craft of Information VisualizationThe Craft of Information Visualization %I Morgan Kaufmann %C San Francisco %P 31 - 37 %8 2003/// %@ 978-1-55860-915-0 %G eng %U http://www.sciencedirect.com/science/article/pii/B978155860915050007X %0 Report %D 2003 %T Building an Old-Fashioned Sparse Solver %A Stewart, G.W. %K Technical Report %X A sparse matrix is a matrix with very few nonzero elements. Manyapplications in diverse fields give rise to linear systems of the form $Ax = b$, where $A$ is sparse. The problem in solving these systems is to take advantage of the preponderance of zero elements to reduce both memory use and comutation time. The purpose of this paper is to introduce students (and perhaps their teachers) to sparse matrix technology. It is impossible to treat all the techniques developed since the subject started in the 1960's. Instead, this paper constructs a sparse solver for positive definite systems that would have been state of the art around 1980, emphasizing equally theory and computational practice. It is hoped that a mastery of this material will allow the reader to study the subject independently. (UMIACS-TR-2003-95) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-95 %8 2003/09/25/ %G eng %U http://drum.lib.umd.edu/handle/1903/1312 %0 Conference Paper %B Proceedings of the 2003 ACM SIGMOD international conference on Management of data %D 2003 %T Cache-and-query for wide area sensor databases %A Deshpande, Amol %A Nath,Suman %A Gibbons,Phillip B. %A Seshan,Srinivasan %X Webcams, microphones, pressure gauges and other sensors provide exciting new opportunities for querying and monitoring the physical world. In this paper we focus on querying wide area sensor databases, containing (XML) data derived from sensors spread over tens to thousands of miles. We present the first scalable system for executing XPATH queries on such databases. The system maintains the logical view of the data as a single XML document, while physically the data is fragmented across any number of host nodes. For scalability, sensor data is stored close to the sensors, but can be cached elsewhere as dictated by the queries. Our design enables self starting distributed queries that jump directly to the lowest common ancestor of the query result, dramatically reducing query response times. We present a novel query-evaluate gather technique (using XSLT) for detecting (1) which data in a local database fragment is part of the query result, and (2) how to gather the missing parts. We define partitioning and cache invariants that ensure that even partial matches on cached data are exploited and that correct answers are returned, despite our dynamic query-driven caching. Experimental results demonstrate that our techniques dramatically increase query throughputs and decrease query response times in wide area sensor databases. %B Proceedings of the 2003 ACM SIGMOD international conference on Management of data %S SIGMOD '03 %I ACM %C New York, NY, USA %P 503 - 514 %8 2003/// %@ 1-58113-634-X %G eng %U http://doi.acm.org/10.1145/872757.872818 %R 10.1145/872757.872818 %0 Conference Paper %B Proceedings of the 29th international conference on Very large data bases-Volume 29 %D 2003 %T Continuous k-nearest neighbor queries for continuously moving points with updates %A Iwerks,G. S %A Samet, Hanan %A Smith,K. %B Proceedings of the 29th international conference on Very large data bases-Volume 29 %P 512 - 523 %8 2003/// %G eng %0 Conference Paper %B INFOCOM 2003. Twenty-Second Annual Joint Conference of the IEEE Computer and Communications. IEEE Societies %D 2003 %T Cooperative peer groups in NICE %A Lee,S. %A Sherwood,R. %A Bhattacharjee, Bobby %K algorithm; %K cooperative %K decentralized %K distributed %K identification; %K individual %K inference %K inference; %K information %K Internet; %K malicious %K manner; %K mechanisms; %K networks; %K NICE %K node; %K noncooperative %K peer-to-peer %K platform; %K reputation %K retrieval; %K storage; %K system; %K trust %K users %K users; %X A distributed scheme for trust inference in peer-to-peer networks is presented. Our work is in context of the NICE system, which is a platform for implementing cooperative applications over the Internet. We describe a technique for efficiently storing user reputation information in a completely decentralized manner, and show how this information can be used to efficiently identify noncooperative users in NICE. We present a simulation based study of our algorithms, in which we show our scheme scales to thousands of users using modest amounts of storage, processing, and bandwidth at any individual node. Lastly, we show that our scheme is robust and can form cooperative groups in systems where the vast majority of users are malicious. %B INFOCOM 2003. Twenty-Second Annual Joint Conference of the IEEE Computer and Communications. IEEE Societies %V 2 %P 1272 - 1282 vol.2 - 1272 - 1282 vol.2 %8 2003/// %G eng %R 10.1109/INFCOM.2003.1208963 %0 Journal Article %J Cognitive Systems Research %D 2003 %T Cost minimization during simulated evolution of paired neural networks leads to asymmetries and specialization %A Shkuro,Y. %A Reggia, James A. %B Cognitive Systems Research %V 4 %P 365 - 383 %8 2003/// %G eng %N 4 %0 Book Section %B FST TCS 2003: Foundations of Software Technology and Theoretical Computer ScienceFST TCS 2003: Foundations of Software Technology and Theoretical Computer Science %D 2003 %T On the Covering Steiner Problem %A Gupta,Anupam %A Srinivasan, Aravind %E Pandya,Paritosh %E Radhakrishnan,Jaikumar %X The Covering Steiner problem is a common generalization of the k -MST and Group Steiner problems. An instance of the Covering Steiner problem consists of an undirected graph with edge-costs, and some subsets of vertices called groups , with each group being equipped with a non-negative integer value (called its requirement ); the problem is to find a minimum-cost tree which spans at least the required number of vertices from every group. When all requirements are equal to 1, this is the Group Steiner problem. While many covering problems (e.g., the covering integer programs such as set cover) become easier to approximate as the requirements increase, the Covering Steiner problem remains at least as hard to approximate as the Group Steiner problem; in fact, the best guarantees previously known for the Covering Steiner problem were worse than those for Group Steiner as the requirements became large. In this work, we present an improved approximation algorithm whose guarantee equals the best known guarantee for the Group Steiner problem. %B FST TCS 2003: Foundations of Software Technology and Theoretical Computer ScienceFST TCS 2003: Foundations of Software Technology and Theoretical Computer Science %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2914 %P 244 - 251 %8 2003/// %@ 978-3-540-20680-4 %G eng %U http://dx.doi.org/10.1007/978-3-540-24597-1_21 %0 Conference Paper %B Proceedings of the 1st ACM international workshop on Multimedia databases %D 2003 %T The CPR model for summarizing video %A Fayzullin,M. %A V.S. Subrahmanian %A Picariello, A. %A Sapino,M. L %K multimedia %K Summarization %K video %X Most past work on video summarization has been based on selecting key frames from videos. We propose a model of video summarization based on three important parameters: Priority (of frames), Continuity (of the summary), and non-Repetition (of the summary). In short, a summary must include high priority frames, must be continuous and non-repetitive. An optimal summary is one that maximizes an objective function based on these three parameters. We develop formal definitions of all these concepts and provide algorithms to find optimal summaries. We briefly report on the performance of these algorithms. %B Proceedings of the 1st ACM international workshop on Multimedia databases %S MMDB '03 %I ACM %C New York, NY, USA %P 2 - 9 %8 2003/// %@ 1-58113-726-5 %G eng %U http://doi.acm.org/10.1145/951676.951679 %R 10.1145/951676.951679 %0 Book %D 2003 %T The craft of information visualization: readings and reflections %A Bederson, Benjamin B. %A Shneiderman, Ben %K Computers / Computer Graphics %K Computers / Computer Science %K Computers / Digital Media / Desktop Publishing %K Computers / General %K Computers / Social Aspects / Human-Computer Interaction %K Computers / User Interfaces %K Human-computer interaction %K Information Visualization %K Science / General %K User-centered system design %X Since the beginning of the computer age, researchers from many disciplines have sought to facilitate people's use of computers and to provide ways for scientists to make sense of the immense quantities of data coming out of them. One gainful result of these efforts has been the field of information visualization, whose technology is increasingly applied in scientific research, digital libraries, data mining, financial data analysis, market studies, manufacturing production control, and data discovery.This book collects 38 of the key papers on information visualization from a leading and prominent research lab, the University of Maryland's Human-Computer Interaction Lab (HCIL). Celebrating HCIL's 20th anniversary, this book presents a coherent body of work from a respected community that has had many success stories with its research and commercial spin-offs. Each chapter contains an introduction specifically written for this volume by two leading HCI researchers, to describe the connections among those papers and reveal HCIL's individual approach to developing innovations.*Presents key ideas, novel interfaces, and major applications of information visualization tools, embedded in inspirational prototypes.*Techniques can be widely applied in scientific research, digital libraries, data mining, financial data analysis, business market studies, manufacturing production control, drug discovery, and genomic studies.*Provides an "insider" view to the scientific process and evolution of innovation, as told by the researchers themselves.*This work comes from the prominent and high profile University of Maryland's Human Computer Interaction Lab %I Morgan Kaufmann %8 2003/// %@ 9781558609150 %G eng %0 Journal Article %J Magnetics, IEEE Transactions on %D 2003 %T Critical fields and pulse durations for precessional switching of thin magnetic films %A Bertotti,G. %A Mayergoyz, Issak D %A Serpico,C. %K anisotropy; %K axis; %K critical %K durations; %K field; %K fields; %K film; %K films; %K hard %K in-plane %K integrals; %K magnetic %K magnetisation %K magnetization %K magnetization; %K magnetized %K motion; %K precessional %K pulse %K pulsed %K rectangular %K reversal; %K switching; %K thin %K uniformly %X The precessional switching process of a uniformly magnetized thin film with in-plane anisotropy subject to pulsed magnetic fields applied in the film plane is analyzed. Critical fields required to achieve switching are studied for the case when the applied field is constant during the pulse duration and forms arbitrary angles with the hard axis. By using two integrals of magnetization motion, the explicit expressions for the critical field are derived. The formulas for durations of rectangular magnetic field pulses that guarantee the precessional switching of magnetization are presented as well. %B Magnetics, IEEE Transactions on %V 39 %P 2504 - 2506 %8 2003/09// %@ 0018-9464 %G eng %N 5 %R 10.1109/TMAG.2003.816454 %0 Journal Article %J ACM Transactions on Asian Language Information Processing (TALIP) %D 2003 %T Cross-language headline generation for Hindi %A Dorr, Bonnie J %A Zajic, David %A Schwartz,Richard %X This paper presents new approaches to headline generation for English newspaper texts, with an eye toward the production of document surrogates for document selection in cross-language information retrieval. This task is difficult because the user must make decisions about relevance based on (often poor) translations of retrieved documents. To facilitate the decision-making process we need translations that can be assessed rapidly and accurately; our approach is to provide an English headline for the non-English document. We describe two approaches to headline generation and their application to the recent DARPA TIDES-2003 Surprise Language Exercise for Hindi. For comparison, we also implemented an alternative method for surrogate generation: a system that produces topic lists for (Hindi) articles. We present the results of a series of experiments comparing each of these approaches. We demonstrate in both automatic and human evaluations that our linguistically motivated approach outperforms two other surrogate-generation methods: a statistical system and a topic discovery system. %B ACM Transactions on Asian Language Information Processing (TALIP) %V 2 %P 270 - 289 %8 2003/09// %@ 1530-0226 %G eng %U http://doi.acm.org/10.1145/979872.979878 %N 3 %R 10.1145/979872.979878 %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T Data exploration with paired hierarchical visualizations: initial designs of PairTrees %A Kules,Bill %A Shneiderman, Ben %A Plaisant, Catherine %X Paired hierarchical visualizations (PairTrees) integrate treemaps, node-link diagrams, choropleth maps and other information visualization techniques to support exploration of hierarchical data sets at multiple levels of abstraction. This paper describes several novel applications of PairTrees in the econometric and health statistics domains, as well as some challenges and trade-offs inherent in the technique. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 6 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123233 %0 Book Section %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %D 2003 %T Database Discovery with Dynamic Queries %A Fredrikson,A. %A North,C. %A Plaisant, Catherine %A Shneiderman, Ben %A Tanin,E. %A Pkisant,C. %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %I Morgan Kaufmann Publishers Inc. %P 1 - 1 %8 2003/// %@ 978-1-55860-915-0 %G eng %0 Journal Article %J Parallel Computing %D 2003 %T Data-parallel polygonization %A Hoel,Erik G. %A Samet, Hanan %K Data-parallel algorithms %K Hierarchical spatial data structures %K Lines %K Polygonization %X Data-parallel algorithms are presented for polygonizing a collection of line segments represented by a data-parallel bucket PMR quadtree, a data-parallel R-tree, and a data-parallel R+-tree. Such an operation is useful in a geographic information system (GIS). A sample performance comparison of the three data-parallel structures for this operation is also given. %B Parallel Computing %V 29 %P 1381 - 1401 %8 2003/10// %@ 0167-8191 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167819103001091 %N 10 %R 10.1016/j.parco.2003.05.001 %0 Journal Article %J Journal of Intelligent Information Systems %D 2003 %T A Deductive Database Approach to AI Planning %A Brogi,A. %A V.S. Subrahmanian %A Zaniolo,C. %X In this paper, we show that the classical A.I. planning problem can be modelled using simple database constructs with logic-based semantics. The approach is similar to that used to model updates and nondeterminism in active database rules. We begin by showing that planning problems can be automatically converted to Datalog1S programs with nondeterministic choice constructs, for which we provide a formal semantics using the concept of stable models. The resulting programs are characterized by a syntactic structure (XY-stratification) that makes them amenable to efficient implementation using compilation and fixpoint computation techniques developed for deductive database systems. We first develop the approach for sequential plans, and then we illustrate its flexibility and expressiveness by formalizing a model for parallel plans, where several actions can be executed simultaneously. The characterization of parallel plans as partially ordered plans allows us to develop (parallel) versions of partially ordered plans that can often be executed faster than the original partially ordered plans. %B Journal of Intelligent Information Systems %V 20 %P 215 - 253 %8 2003/// %G eng %N 3 %R 10.1023/A:1022808724136 %0 Conference Paper %B Image Analysis and Processing, 2003.Proceedings. 12th International Conference on %D 2003 %T Depth-first k-nearest neighbor finding using the MaxNearestDist estimator %A Samet, Hanan %K branch-and-bound %K data %K depth-first %K distance; %K DNA %K documents; %K estimation; %K estimator; %K finding; %K images; %K k-nearest %K matching; %K maximum %K MaxNearestDist %K mining; %K neighbor %K parameter %K pattern %K possible %K process; %K processing; %K query %K search %K searching; %K sequences; %K series; %K similarity %K text %K TIME %K tree %K video; %X Similarity searching is an important task when trying to find patterns in applications which involve mining different types of data such as images, video, time series, text documents, DNA sequences, etc. Similarity searching often reduces to finding the k nearest neighbors to a query object. A description is given of how to use an estimate of the maximum possible distance at which a nearest neighbor can be found to prune the search process in a depth-first branch-and-bound k-nearest neighbor finding algorithm. Using the MaxNearestDist estimator (Larsen, S. and Kanal, L.N., 1986) in the depth-first k-nearest neighbor algorithm provides a middle ground between a pure depth-first and a best-first k-nearest neighbor algorithm. %B Image Analysis and Processing, 2003.Proceedings. 12th International Conference on %P 486 - 491 %8 2003/09// %G eng %R 10.1109/ICIAP.2003.1234097 %0 Journal Article %J The craft of information visualization: readings and reflections %D 2003 %T Designing a digital library for young children: An intergenerational partnership %A Druin, Allison %A Bederson, Benjamin B. %A Hourcade,J. P %A Sherman,L. %A Revelle,G. %A Platner,M. %A Weng,S. %B The craft of information visualization: readings and reflections %V 178 %8 2003/// %G eng %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T Designing a metadata-driven visual information browser for federal statistics %A Kules,Bill %A Shneiderman, Ben %X When looking for federal statistics, finding the right table, chart or report can be a daunting task for anyone not thoroughly familiar with the federal statistical system. Search tools help, but differing terminologies within the statistical agencies and a lack of familiarity of terms by information seekers limit their effectiveness. The FedStats Browser is a design for visually browsing federal agency statistical products and publications, using techniques that allow users to reformulate queries and iteratively refine results via simple, reversible actions with immediate feedback. This paper also discusses the characteristics of metadata needed for such a browser and the challenges inherent in acquiring that metadata. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 6 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123235 %0 Journal Article %J Communications of the ACM %D 2003 %T Digital government %A Marchionini,G. %A Samet, Hanan %A Brandt,L. %X Information technologies are being applied vigorously by governmental units at national,regional, and local levels around the world. The application of IT to government service is often termed “e-government” and the larger concept of government that depends upon IT to achieve basic missions is termed “digital government.” This distinction is, of course, lexically arbitrary, but serves to distinguish R&D specifically aimed at creating techniques for applying IT to government operations. Such R&D efforts also consider the long-term impact of these applications on citizens and government itself. %B Communications of the ACM %V 46 %P 25 - 27 %8 2003/// %G eng %N 1 %0 Journal Article %J Communications of the ACM-Association for Computing Machinery-CACM %D 2003 %T Digital Government-Bistro: A Scalable and Secure Data Transfer Service for Digital Government Applications %A Golubchik,L. %A Cheng,W. C %A Chou,C. F %A Khuller, Samir %A Samet, Hanan %A Wan,Y. C.J %B Communications of the ACM-Association for Computing Machinery-CACM %V 46 %P 50 - 51 %8 2003/// %G eng %N 1 %0 Conference Paper %B Workshop on Remote Analysis and Measurement of Software Systems (RAMSS) %D 2003 %T Distributed continuous quality assurance: The Skoll project %A Yilmaz,C. %A Porter, Adam %A Schmidt,D. C %B Workshop on Remote Analysis and Measurement of Software Systems (RAMSS) %8 2003/// %G eng %0 Journal Article %J Computational Science—ICCS 2003 %D 2003 %T Driving scientific applications by data in distributed environments %A Saltz, J. %A Catalyurek,U. %A Kurc, T. %A Gray,M. %A Hastings,S. %A Langella,S. %A Narayanan,S. %A Martino,R. %A Bryant,S. %A Peszynska,M. %A others %B Computational Science—ICCS 2003 %P 713 - 713 %8 2003/// %G eng %0 Journal Article %J Proc. Human-Computer Interaction International %D 2003 %T Dynamic query choropleth maps for information seeking and decision making %A Norman,K. %A Zhao,H. %A Shneiderman, Ben %A Golub,E. %X Information retrieval and visualization can be combined in dynamic query systems that allow users unparalleled access to information for decision making. In this paper, we report on the development and evaluation of a dynamic query system (YMap) that displays information on a chloropleth map using double thumb sliders to select ranges of query variables. The YMap prototype is a Java-Applet that supports panning and zooming. Several usability studies were conducted on early prototypes that resulted in the current version. Applications of YMap for decision making tasks are discussed. %B Proc. Human-Computer Interaction International %8 2003/// %G eng %0 Journal Article %J International journal of Web engineering and technology %D 2003 %T Dynamic query visualisations on world wide web clients: A DHTML solution for maps and scattergrams %A Golub,E. %A Shneiderman, Ben %X Dynamic queries are gaining popularity as a method for interactive information visualisation. Many implementations have been made on personal computers, and there is increasing interest in Web-based designs. While Java and Flash strategies have been developed, we believe that a Dynamic HTML implementation could help promote more widespread use. We implemented double-box range sliders with choropleth maps and scattergrams, which are two popular visualisations, using HTML layers and tables. This paper describes our methods for slider control, visual presentation, and displaying/updating results for these visualisations. Visual design issues innovations and performance enhancements were necessary to create viable designs. %B International journal of Web engineering and technology %V 1 %P 63 - 78 %8 2003/// %G eng %N 1 %0 Conference Paper %B 2003 International Conference on Multimedia and Expo, 2003. ICME '03. Proceedings %D 2003 %T Dynamic querying for pattern identification in microarray and genomic data %A Hochheiser,H. %A Baehrecke,E. H %A Mount, Stephen M. %A Shneiderman, Ben %K Bioinformatics %K data sets %K Displays %K dynamic querying %K expression profiles %K Frequency %K Gene expression %K genes %K Genetics %K genomic data %K Genomics %K linear ordered sequences %K macromolecules %K medical signal processing %K Mice %K Microarray %K pattern identification %K pattern recognition %K premRNA splicing %K Query processing %K sequences %K Signal processing %K splicing %K TimeSearcher %X Data sets involving linear ordered sequences are a recurring theme in bioinformatics. Dynamic query tools that support exploration of these data sets can be useful for identifying patterns of interest. This paper describes the use of one such tool - timesearcher - to interactively explore linear sequence data sets taken from two bioinformatics problems. Microarray time course data sets involve expression levels for large numbers of genes over multiple time points. Timesearcher can be used to interactively search these data sets for genes with expression profiles of interest. The occurrence frequencies of short sequences of DNA in aligned exons can be used to identify sequences that play a role in the pre-mRNA splicing. Timesearcher can be used to search these data sets for candidate splicing signals. %B 2003 International Conference on Multimedia and Expo, 2003. ICME '03. Proceedings %I IEEE %V 3 %P III- 453-6 vol.3 - III- 453-6 vol.3 %8 2003/07/06/9 %@ 0-7803-7965-9 %G eng %R 10.1109/ICME.2003.1221346 %0 Conference Paper %B Proceedings of the eighth ACM SIGPLAN international conference on Functional programming %D 2003 %T Dynamic rebinding for marshalling and update, with destruct-time ? %A Bierman,Gavin %A Hicks, Michael W. %A Sewell,Peter %A Stoyle,Gareth %A Wansbrough,Keith %K distributed programming %K dynamic binding %K dynamic update %K lambda calculus %K marshalling %K programming languages %K serialisation %X Most programming languages adopt static binding, but for distributed programming an exclusive reliance on static binding is too restrictive: dynamic binding is required in various guises, for example when a marshalled value is received from the network, containing identifiers that must be rebound to local resources. Typically it is provided only by ad-hoc mechanisms that lack clean semantics.In this paper we adopt a foundational approach, developing core dynamic rebinding mechanisms as extensions to simply-typed call-by-value ? -calculus. To do so we must first explore refinements of the call-by-value reduction strategy that delay instantiation, to ensure computations make use of the most recent versions of rebound definitions. We introduce redex-time and destruct-time strategies. The latter forms the basis for a ?marsh calculus that supports dynamic rebinding of marshalled values, while remaining as far as possible statically-typed. We sketch an extension of ? marsh with concurrency and communication, giving examples showing how wrappers for encapsulating untrusted code can be expressed. Finally, we show that a high-level semantics for dynamic updating can also be based on the destruct-time strategy, defining a ?marsh calculus with simple primitives to provide type-safe updating of running code. We thereby establish primitives and a common semantic foundation for a variety of real-world dynamic rebinding requirements. %B Proceedings of the eighth ACM SIGPLAN international conference on Functional programming %S ICFP '03 %I ACM %C New York, NY, USA %P 99 - 110 %8 2003/// %@ 1-58113-756-7 %G eng %U http://doi.acm.org/10.1145/944705.944715 %R 10.1145/944705.944715 %0 Journal Article %J Veterinary Parasitology %D 2003 %T Effect of treatment on the dynamics of circulating hypodermin C in cattle naturally infested with Hypoderma lineatum (Diptera: Oestridae) %A Douglas D Colwell %A Panadero-Fontan,Rosario %A López-Sandez,Ceferino %A Parra-Fernandez,Francisco %A Paz-Silva,A %A Sánchez-Andrade,Rita %A Dı́ez-Baños,Pablo %K Capture ELISA %K Cattle-arthropoda %K Control methods-arthropoda %K Eprinomectin %K Fenthion %K Hypoderma spp. %X An antigen capture ELISA, using a murine monoclonal antibody recognising recombinant hypodermin C (rHyC), was used to evaluate the influence of early treatment with eprinomectin (Eprinex®) or fenthion (Spotton®) on the kinetics of circulating hypodermin C in calves naturally infested with Hypoderma lineatum. No viable larvae were collected from treated animals, whereas a variable number of warbles were found in control animals. Treatment provoked a decrease in circulating HyC levels that was significant 9 days post-treatment (p.t.). Circulating antigen levels in the treated cattle remained detectable for approximately 99 days p.t. In contrast, control animals had no detectable antigen at 64 days p.t., 42 days earlier than in the treated animals. These results suggest that larvae were either gradually killed, resulting in slow release of antigen or they were encapsulated, leading to the slow liberation of antigen. Kinetics of circulating HyC did not differ among the two insecticide treatments. Antibodies persisted, in all groups, throughout the 120-day study. These results suggest that the antigen capture ELISA will be useful as a technique for detecting successful treatment of cattle grub infestations and for the detection of new infestations in previously infested cattle. %B Veterinary Parasitology %V 113 %P 263 - 272 %8 2003/05/01/ %@ 0304-4017 %G eng %U http://www.sciencedirect.com/science/article/pii/S0304401703000840 %N 3–4 %R 10.1016/S0304-4017(03)00084-0 %0 Conference Paper %B International Workshop on the Web and Databases (WebDB) %D 2003 %T Efficient dissemination of aggregate data over the wireless web %A Sharaf,M. A %A Sismanis,Y. %A Labrinidis,A. %A Chrysanthis,P. %A Roussopoulos, Nick %B International Workshop on the Web and Databases (WebDB) %P 93 - 98 %8 2003/// %G eng %0 Journal Article %J Peer-to-Peer Systems II %D 2003 %T Efficient peer-to-peer searches using result-caching %A Bhattacharjee, Bobby %A Chawathe,S. %A Gopalakrishnan,V. %A Keleher,P. %A Silaghi,B. %X Existing peer-to-peer systems implement a single function well: data lookup. There is now a wealth of research describing how to reliably disseminate, and to later retrieve, data in a scalable and load-balanced manner. %B Peer-to-Peer Systems II %P 225 - 236 %8 2003/// %G eng %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2003. IEEE %D 2003 %T Electromagnetic scattering from a multilayered cylindrical waveguide %A Seydou,F. %A Duraiswami, Ramani %A Seppanen,T. %K circular %K core; %K cylinder; %K cylindrical %K dielectric %K dielectric-loaded %K electromagnetic %K EM %K equations; %K matching; %K Maxwell %K mode %K multilayered %K scattering; %K theory; %K wave %K waveguide %K waveguide; %K waveguides; %X This paper is devoted to electromagnetic scattering from an N multilayered circular cylinder. We consider waveguides in the z direction; that is, we look for the solution of Maxwell equations along the z direction. We assume a dielectric core and derive a mode matching approach for solving the problem. A numerical result is presented that illustrates the algorithm. %B Antennas and Propagation Society International Symposium, 2003. IEEE %V 3 %P 332 - 335 vol.3 - 332 - 335 vol.3 %8 2003/06// %G eng %R 10.1109/APS.2003.1219855 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 2003 %T Electronic voting system usability issues %A Bederson, Benjamin B. %A Lee,B. %A Sherman,R.M. %A Herrnson,P.S. %A Niemi,R.G. %B Proceedings of the SIGCHI conference on Human factors in computing systems %P 145 - 152 %8 2003/// %G eng %0 Journal Article %J Proceedings of the National Academy of SciencesPNAS %D 2003 %T Emergence and Evolution of Vibrio Cholerae O139 %A Faruque,Shah M. %A Sack,David A. %A Sack,R. Bradley %A Rita R Colwell %A Takeda,Yoshifumi %A Nair,G. Balakrish %X The emergence of Vibrio cholerae O139 Bengal during 1992–1993 was associated with large epidemics of cholera in India and Bangladesh and, initially, with a total displacement of the existing V. cholerae O1 strains. However, the O1 strains reemerged in 1994 and initiated a series of disappearance and reemergence of either of the two serogroups that was associated with temporal genetic and phenotypic changes sustained by the strains. Since the initial emergence of the O139 vibrios, new variants of the pathogen derived from multiple progenitors have been isolated and characterized. The clinical and epidemiological characteristics of these strains have been studied. Rapid genetic reassortment in O139 strains appears to be a response to the changing epidemiology of V. cholerae O1 and also a strategy for persistence in competition with strains of the O1 serogroup. The emergence of V. cholerae O139 has provided a unique opportunity to witness genetic changes in V. cholerae that may be associated with displacement of an existing serogroup by a newly emerging one and, thus, provide new insights into the epidemiology of cholera. The genetic changes and natural selection involving both environmental and host factors are likely to influence profoundly the genetics, epidemiology, and evolution of toxigenic V. cholerae, not only in the Ganges Delta region of India and Bangladesh, but also in other areas of endemic and epidemic cholera. %B Proceedings of the National Academy of SciencesPNAS %V 100 %P 1304 - 1309 %8 2003/02/04/ %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/100/3/1304 %N 3 %R 10.1073/pnas.0337468100 %0 Book Section %B Encyclopedia of Computer ScienceEncyclopedia of Computer Science %D 2003 %T Errors %A Ralston,Anthony %A Shneiderman, Ben %X The indignant customer who receives an incorrect bill from a department store probably does not care what the source of the error was or even that, almost certainly, the fault was not a computer's but rather that of its data entry personnel or programmers. Neither is the astronaut who is descending toward the surface of the Moon very concerned about the precise source of the error that caused the onboard computer to fail. But an understanding of the sources of errors in computers is important to anyone who wishes to use or comprehend digital computers. %B Encyclopedia of Computer ScienceEncyclopedia of Computer Science %I John Wiley and Sons Ltd. %C Chichester, UK %P 677 - 680 %8 2003/// %@ 0-470-86412-5 %G eng %U http://dl.acm.org/citation.cfm?id=1074100.1074382 %0 Conference Paper %B Image and Signal Processing and Analysis, 2003. ISPA 2003. Proceedings of the 3rd International Symposium on %D 2003 %T Exemplar-based tracking and recognition of arm gestures %A Elgammal,A. %A Shet,V. %A Yacoob,Yaser %A Davis, Larry S. %K arm %K constrains; %K correspondence-free %K edge %K exemplar-based %K framework; %K gesture %K hidden %K HMM; %K image %K logic; %K Markov %K MATCHING %K matching; %K model; %K models; %K probabilistic %K recognition; %K scheme; %K segmentation; %K temporal %K tracking; %K weighted %X This paper presents a probabilistic exemplar-based framework for recognizing gestures. The approach is based on representing each gesture as a sequence of learned body poses. The gestures are recognized through a probabilistic framework for matching these body poses and for imposing temporal constrains between different poses. Matching individual poses to image data is performed using a probabilistic formulation for edge matching to obtain a likelihood measurement for each individual pose. The paper introduces a correspondence-free weighted matching scheme for edge templates that emphasize discriminating features in the matching. The weighting does not require establishing correspondences between the different pose models. The probabilistic framework also imposes temporal constrains between different pose through a learned hidden Markov model (HMM) of each gesture. %B Image and Signal Processing and Analysis, 2003. ISPA 2003. Proceedings of the 3rd International Symposium on %V 2 %P 656 - 661 Vol.2 - 656 - 661 Vol.2 %8 2003/09// %G eng %R 10.1109/ISPA.2003.1296358 %0 Conference Paper %D 2003 %T An experimental evaluation of correlated network partitions in the Coda distributed file system %A Lefever,R.M. %A Michel Cukier %A Sanders,W. H. %K client-server systems %K Coda distributed file system %K correlated network partitions %K distributed file systems %K experimental evaluation %K fault model %K Loki fault injector %K multiple correlated failures %K network failure %K Network topology %K performance evaluation %K replicated data %K replicated databases %K software fault tolerance %K system performance evaluation %X Experimental evaluation is an important way to assess distributed systems, and fault injection is the dominant technique in this area for the evaluation of a system's dependability. For distributed systems, network failure is an important fault model. Physical network failures often have far-reaching effects, giving rise to multiple correlated failures as seen by higher-level protocols. This paper presents an experimental evaluation, using the Loki fault injector, which provides insight into the impact that correlated network partitions have on the Coda distributed file system. In this evaluation, Loki created a network partition between two Coda file servers, during which updates were made at each server to the same replicated data volume. Upon repair of the partition, a client requested directory resolution to converge the diverging replicas. At various stages of the resolution, Loki invoked a second correlated network partition, thus allowing us to evaluate its impact on the system's correctness, performance, and availability. %P 273 - 282 %8 2003/10// %G eng %R 10.1109/RELDIS.2003.1238077 %0 Conference Paper %B Parallel and Distributed Processing Symposium, International %D 2003 %T Exploiting Functional Decomposition for Efficient Parallel Processing of Multiple Data Analysis Queries %A Andrade,Henrique %A Kurc,Tahsin %A Sussman, Alan %A Saltz,Joel %X Reuse is a powerful method for increasing system performance. In this paper, we examine functional decomposition for improving data and computation reuse and, therefore, overall query execution performance in the context of data analysis applications. Additionally, we look at the performance effects of using various projection primitives that make it possible to transform intermediate results generated by a query so that they can be reused by a new query. A satellite data analysis application is used to experimentally show the performance benefits achieved using functional decomposition and projection primitives. %B Parallel and Distributed Processing Symposium, International %I IEEE Computer Society %C Los Alamitos, CA, USA %P 81a - 81a %8 2003/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/IPDPS.2003.1213184 %0 Report %D 2003 %T Fast Algorithms for 3-D Dominance Reporting and Counting %A Shi,Qingmin %A JaJa, Joseph F. %K Technical Report %X We present in this paper fast algorithms for the 3-D dominance reportingand counting problems, and generalize the results to the d-dimensional case. Our 3-D dominance reporting algorithm achieves $O(\log n/\log\log n +f)$ query time using $O(n\log^{\epsilon}n)$ space, where $f$ is the number of points satisfying the query and $\epsilon>0$ is an arbitrary small constant. For the 3-D dominance counting problem (which is equivalent to the 3-D range counting problem), our algorithm runs in $O((\log n/\log\log n)^2)$ time using $O(nlog^{1+\epsilon}n/\log\log n)$ space. Also UMIACS-TR-2003-06 %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-06 %8 2003/02/05/ %G eng %U http://drum.lib.umd.edu/handle/1903/1253 %0 Journal Article %J Algorithms and Data Structures %D 2003 %T Fast algorithms for a class of temporal range queries %A Shi,Q. %A JaJa, Joseph F. %X Given a set of n objects, each characterized by d attributes specified at m fixed time instances, we are interested in the problem of designing efficient indexing structures such that the following type of queries can be handled efficiently: given d value ranges and a time interval, report or count all the objects whose attributes fall within the corresponding d value ranges at each time instance lying in the specified time interval. We establish efficient data structures to handle several classes of the general problem. Our results include a linear size data structure that enables a query time of O(log n log m + f) for one-sided queries when d=1, where f is the output size. We also show that the most general problem can be solved with polylogarithmic query time using nonlinear space data structures. %B Algorithms and Data Structures %P 91 - 102 %8 2003/// %G eng %R 10.1007/978-3-540-45078-8_9 %0 Report %D 2003 %T Fast Fractional Cascading and Its Applications %A Shi,Qingmin %A JaJa, Joseph F. %K Technical Report %X Using the notions of Q-heaps and fusion trees developed by Fredman andWillard, we develop a faster version of the fractional cascading technique while maintaining the linear space structure. The new version enables sublogarithmic iterative search in the case when we have a search tree and the degree of each node is bounded by $O(\log^{\epsilon}n)$, for some constant $\epsilon >0$, where $n$ is the total size of all the lists stored in the tree. The fast fractional cascading technique is used in combination with other techniques to derive sublogarithmic time algorithms for the geometric retrieval problems: orthogonal segment intersection and rectangular point enclosure. The new algorithms use $O(n)$ space and achieve a query time of $O(\log n/\log\log n + f)$, where $f$ is the number of objects satisfying the query. All our algorithms assume the version of the RAM model used by Fredman and Willard. (UMIACS-TR-2003-71) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-71 %8 2003/08/01/ %G eng %U http://drum.lib.umd.edu/handle/1903/1296 %0 Conference Paper %B Mobile Computing Systems and Applications, 2003. Proceedings. Fifth IEEE Workshop on %D 2003 %T Footloose: a case for physical eventual consistency and selective conflict resolution %A Paluska %A Saff %A Yeh %A Chen %X Users are increasingly inundated with small devices with communication and storage capabilities. Unfortunately, the user is still responsible for reconciling all of the devices whenever a change is made. We present Footloose, a user-centered data store that can share data and reconcile conflicts across diverse devices. Footloose is an optimistic system based on physical eventual consistency: consistency based on the movement of devices, and selective conflict resolution, which allows conflicts to flow through devices that cannot resolve the conflict to devices that can. Using these techniques, Footloose can present consistent views of data on the devices closest to the user without user interaction. %B Mobile Computing Systems and Applications, 2003. Proceedings. Fifth IEEE Workshop on %I IEEE %P 170 - 179 %8 2003/// %@ 0-7695-1995-4 %G eng %U http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=1240778 %R 10.1109/MCSA.2003.1240778 %0 Journal Article %J IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS E SERIES D %D 2003 %T Formal verification of an intrusion-tolerant group membership protocol %A Ramasamy,H. V. %A Michel Cukier %A Sanders,W. H. %X The traditional approach for establishing thecorrectness of group communication protocols is through rigor- ous arguments. While this is a valid approach, the likelihood of subtle errors in the design and implementation of such complex distributed protocols is not negligible. The use of formal verifi- cation methods has been widely advocated to instill confidence in the correctness of protocols. In this paper, we describe how we used the SPIN model checker to formally verify a group mem- bership protocol that is part of an intrusion-tolerant group com- munication system. We describe how we successfully tackled the state-space explosion problem by determining the right abstrac- tion level for formally specifying the protocol. The verification exercise not only formally showed that the protocol satisfies its correctness claims, but also provided information that will help us make the protocol more efficient without violating correctness. key words: intrusion tolerance, group communication systems, validation, formal methods %B IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS E SERIES D %V 86 %P 2612 - 2622 %8 2003/// %G eng %U https://www.perform.csl.illinois.edu/Papers/USAN_papers/03RAM01.pdf %N 12 %0 Journal Article %J Proceedings of the Second International Workshop on Unanticipated Software Evolution (USE) %D 2003 %T Formalizing dynamic software updating %A Bierman,G. %A Hicks, Michael W. %A Sewell,P. %A Stoyle,G. %X Dynamic software updating (DSU) enables running programs to beupdated with new code and data without interrupting their execution. A number of DSU systems have been designed, but there is still little rig- orous understanding of how to use DSU technology so that updates are safe. As a first step in this direction, we introduce a small update calculus with a precise mathematical semantics. The calculus is formulated as an extension of a typed lambda calculus, and supports updating technology similar to that of the programming language Erlang [2]. Our goal is to provide a simple yet expressive foundation for reasoning about dynam- ically updateable software. In this paper, we present the details of the calculus, give some examples of its expressive power, and discuss how it might be used or extended to guarantee safety properties. %B Proceedings of the Second International Workshop on Unanticipated Software Evolution (USE) %8 2003/// %G eng %0 Report %D 2003 %T A Fortran 95 Matrix Wrapper %A Stewart, G.W. %K Technical Report %X {\Matran} is an wrapper written in Fortran~95 that implements matrixoperations and computes matrix decompositions using {\lapack} and the {\blas}. This document describes a preliminary release of {\matran}, which treats only real matrices. Its purpose is to get outside comments and suggestions before the package jells. Consequently, this documentation is slanted toward the experienced programmer familiar with both matrix computations and Fortran~90/95. User oriented documentation will accompany the final release. (UMIACS-TR-2003-89) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-89 %8 2003/09/02/ %G eng %U http://drum.lib.umd.edu/handle/1903/1308 %0 Journal Article %J Nature %D 2003 %T The genome sequence of Bacillus anthracis Ames and comparison to closely related bacteria %A Read,Timothy D. %A Peterson,Scott N. %A Tourasse,Nicolas %A Baillie,Les W. %A Paulsen,Ian T. %A Nelson,Karen E. %A Tettelin,Herv|[eacute]| %A Fouts,Derrick E. %A Eisen,Jonathan A. %A Gill,Steven R. %A Holtzapple,Erik K. %A |[Oslash]|kstad,Ole Andreas %A Helgason,Erlendur %A Rilstone,Jennifer %A Wu,Martin %A Kolonay,James F. %A Beanan,Maureen J. %A Dodson,Robert J. %A Brinkac,Lauren M. %A Gwinn,Michelle %A DeBoy,Robert T. %A Madpu,Ramana %A Daugherty,Sean C. %A Durkin,A. Scott %A Haft,Daniel H. %A Nelson,William C. %A Peterson,Jeremy D. %A Pop, Mihai %A Khouri,Hoda M. %A Radune,Diana %A Benton,Jonathan L. %A Mahamoud,Yasmin %A Jiang,Lingxia %A Hance,Ioana R. %A Weidman,Janice F. %A Berry,Kristi J. %A Plaut,Roger D. %A Wolf,Alex M. %A Watkins,Kisha L. %A Nierman,William C. %A Hazen,Alyson %A Cline,Robin %A Redmond,Caroline %A Thwaite,Joanne E. %A White,Owen %A Salzberg,Steven L. %A Thomason,Brendan %A Friedlander,Arthur M. %A Koehler,Theresa M. %A Hanna,Philip C. %A Kolst|[oslash]|,Anne-Brit %A Fraser,Claire M. %X Bacillus anthracis is an endospore-forming bacterium that causes inhalational anthrax1. Key virulence genes are found on plasmids (extra-chromosomal, circular, double-stranded DNA molecules) pXO1 (ref. 2) and pXO2 (ref. 3). To identify additional genes that might contribute to virulence, we analysed the complete sequence of the chromosome of B. anthracis Ames (about 5.23 megabases). We found several chromosomally encoded proteins that may contribute to pathogenicity—including haemolysins, phospholipases and iron acquisition functions—and identified numerous surface proteins that might be important targets for vaccines and drugs. Almost all these putative chromosomal virulence and surface proteins have homologues in Bacillus cereus, highlighting the similarity of B. anthracis to near-neighbours that are not associated with anthrax4. By performing a comparative genome hybridization of 19 B. cereus and Bacillus thuringiensis strains against a B. anthracis DNA microarray, we confirmed the general similarity of chromosomal genes among this group of close relatives. However, we found that the gene sequences of pXO1 and pXO2 were more variable between strains, suggesting plasmid mobility in the group. The complete sequence of B. anthracis is a step towards a better understanding of anthrax pathogenesis. %B Nature %V 423 %P 81 - 86 %8 2003/05/01/ %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v423/n6935/full/nature01586.html %N 6935 %R 10.1038/nature01586 %0 Journal Article %J Magnetics, IEEE Transactions on %D 2003 %T Geometrical analysis of precessional switching and relaxation in uniformly magnetized bodies %A Bertotti,G. %A Mayergoyz, Issak D %A Serpico,C. %A d'Aquino,M. %K analysis; %K attraction; %K basins %K bodies; %K dynamical %K dynamics; %K entanglement; %K equation; %K film; %K films; %K geometrical %K integrals %K Landau-Lifshitz-Gilbert %K magnetic %K magnetisation %K magnetization %K magnetized %K motion; %K of %K particle; %K particles; %K phase %K portrait; %K precessional %K relaxation; %K reversal; %K riddled %K separatrices; %K single-domain %K spin %K switching; %K system; %K thin %K uniformly %X Precessional switching and relaxation in a single-domain particle or film is studied by using the Landau-Lifshitz-Gilbert equation. The analysis of the switching process is based on the explicit knowledge of two integrals of motion for the magnetization dynamics in the conservative case. The knowledge of these integrals of motions enables one to carry out the geometrical analysis of the system and give the complete phase portrait. The relaxation process which occurs after the magnetization is reversed is analyzed by using geometrical methods and it is showed that the dynamical system exhibits entanglement of separatrices and riddled basins of attraction. %B Magnetics, IEEE Transactions on %V 39 %P 2501 - 2503 %8 2003/09// %@ 0018-9464 %G eng %N 5 %R 10.1109/TMAG.2003.816453 %0 Journal Article %J The craft of information visualization: readings and reflections %D 2003 %T HCIL Technical Report Listing (1993-2002) %A Ceaparu,I. %A Druin, Allison %A Guimbretiere,F. %A Manipulation,D. %A Shneiderman, Ben %A Westerlurid,B. %A Keogh,E. %A Hochheiser,H. %A Shrieidemian,B. %A Shneidemiari,B. %A others %B The craft of information visualization: readings and reflections %V 27 %P 393 - 393 %8 2003/// %G eng %N 29 %0 Conference Paper %B Proceedings of the HLT-NAACL 03 on Text summarization workshop - Volume 5 %D 2003 %T Hedge Trimmer: a parse-and-trim approach to headline generation %A Dorr, Bonnie J %A Zajic, David %A Schwartz,Richard %X This paper presents Hedge Trimmer, a HEaDline GEneration system that creates a headline for a newspaper story using linguistically-motivated heuristics to guide the choice of a potential headline. We present feasibility tests used to establish the validity of an approach that constructs a headline by selecting words in order from a story. In addition, we describe experimental results that demonstrate the effectiveness of our linguistically-motivated approach over a HMM-based model, using both human evaluation and automatic metrics for comparing the two approaches. %B Proceedings of the HLT-NAACL 03 on Text summarization workshop - Volume 5 %S HLT-NAACL-DUC '03 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 1 - 8 %8 2003/// %G eng %U http://dx.doi.org/10.3115/1119467.1119468 %R 10.3115/1119467.1119468 %0 Journal Article %J IT & Society %D 2003 %T Help! I’m lost: User frustration in web navigation %A Lazar,J. %A Bessiere,K. %A Ceaparu,I. %A Robinson,J. %A Shneiderman, Ben %X Computers can be valuable tools, and networked resources via the Internet can be beneficial to many different populations and communities. Unfortunately, when people are unable to reach their task goals due to frustrating experiences, this can hinder the effectiveness of technology. This research summary provides information about the user frustration research that has been performed at the University of Maryland and Towson University. Causes of user frustration are discussed in this research summary, along with the surprising finding that nearly one-third to one-half of the time spent in front of the computer is wasted due to frustrating experiences. Furthermore, when interfaces are planned to be deceptive and confusing, this can lead to increased frustration. Implications for designers and users are discussed. %B IT & Society %V 1 %P 18 - 26 %8 2003/// %G eng %N 3 %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T Helping users get started with visual interfaces: multi-layered interfaces, integrated initial guidance and video demonstrations %A Kang,Hyunmo %A Plaisant, Catherine %A Shneiderman, Ben %X We are investigating new ways to help users learn to use public access interactive tools, in particular for the visual exploration of government statistics. Our work led to a series of interfaces using multi-layered design, a new help method called Integrated Initial Guidance, and video demonstrations. Multi-layer designs structure an interface so that a simpler interface is available for users to get started and more complex features are accessed as users move through the more advanced layers. Integrated Initial Guidance provides help within the working interface, right at the start of the application. Using the metaphor of "sticky notes" overlaid on top of the functional interface locates the main widgets, demonstrates their manipulation, and explains the resulting actions using preset activations of the interface. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 1 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123257 %0 Conference Paper %B Image Processing, 2003. ICIP 2003. Proceedings. 2003 International Conference on %D 2003 %T A hidden Markov model based framework for recognition of humans from gait sequences %A Sundaresan,Aravind %A RoyChowdhury,Amit %A Chellapa, Rama %K analysis; %K background-subtracted %K binarized %K discrete %K distance %K feature %K Gait %K hidden %K human %K image %K image; %K Markov %K metrics; %K model; %K models; %K postures; %K recognition; %K sequences; %K vector; %X In this paper we propose a generic framework based on hidden Markov models (HMMs) for recognition of individuals from their gait. The HMM framework is suitable, because the gait of an individual can be visualized as his adopting postures from a set, in a sequence which has an underlying structured probabilistic nature. The postures that the individual adopts can be regarded as the states of the HMM and are typical to that individual and provide a means of discrimination. The framework assumes that, during gait, the individual transitions between N discrete postures or states but it is not dependent on the particular feature vector used to represent the gait information contained in the postures. The framework, thus, provides flexibility in the selection of the feature vector. The statistical nature of the HMM lends robustness to the model. In this paper we use the binarized background-subtracted image as the feature vector and use different distance metrics, such as those based on the L1 and L2 norms of the vector difference, and the normalized inner product of the vectors, to measure the similarity between feature vectors. The results we obtain are better than the baseline recognition rates reported before. %B Image Processing, 2003. ICIP 2003. Proceedings. 2003 International Conference on %V 2 %P II - 93-6 vol.3 - II - 93-6 vol.3 %8 2003/09// %G eng %R 10.1109/ICIP.2003.1246624 %0 Conference Paper %B Proceedings of the 6th ACM international workshop on Data warehousing and OLAP - DOLAP '03 %D 2003 %T Hierarchical dwarfs for the rollup cube %A Sismanis,Yannis %A Deligiannakis,Antonios %A Kotidis,Yannis %A Roussopoulos, Nick %B Proceedings of the 6th ACM international workshop on Data warehousing and OLAP - DOLAP '03 %C New Orleans, Louisiana, USA %P 17 - 17 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=956064 %R 10.1145/956060.956064 %0 Conference Paper %B Proceedings of the 17th annual international conference on Supercomputing %D 2003 %T A high performance multi-perspective vision studio %A Borovikov,Eugene %A Sussman, Alan %K database %K distributed system %K high-performance %K IMAGE PROCESSING %K multi-perspective %K VISION %K volumetric reconstruction %X We describe a multi-perspective vision studio as a flexible high performance framework for solving complex image processing and machine vision problems on multi-view image sequences. The studio abstracts multi-view image data from image sequence acquisition facilities, stores and catalogs sequences in a high performance distributed database, allows customization of back-end processing services, and can serve custom client applications, thus helping make multi-view video sequence processing efficient and generic. To illustrate our approach, we describe two multi-perspective studio applications, and discuss performance and scalability results. %B Proceedings of the 17th annual international conference on Supercomputing %S ICS '03 %I ACM %C New York, NY, USA %P 348 - 357 %8 2003/// %@ 1-58113-733-8 %G eng %U http://doi.acm.org/10.1145/782814.782862 %R 10.1145/782814.782862 %0 Journal Article %J Technical Reports from UMIACS %D 2003 %T Image-based highly interactive Web mapping for geo-referenced data publishing %A Zhao,Haixia %A Shneiderman, Ben %K Technical Report %X This paper describes an image-based technique that enables highly interactiveWeb choropleth maps for geo-referenced data publishing and visual exploration. Geographic knowledge is encoded into raster images and delivered to the client, instead of in vector formats. Differing from traditional raster-image-based approaches that are static and allow very little user interaction, it allows varieties of sub-second fine-grained interface controls such as dynamic query, dynamic classification, geographic object data identification, user setting adjusting, as well as turning on/off layers, panning and zooming, with no or minimum server support. Compared to Web GIS approaches that are based on vector geographic data, this technique has the features of short initial download time, near-constant performance scalability for larger numbers of geographic objects, and download-map-segment-only-when-necessary which potentially reduces the overall data transfer over the network. As a result, it accommodates general public users with slow modem network connections and low-end machines, as well as users with fast T-1 connections and fast machines. The client-side (browser) is implemented as light-weight Java applets. YMap, an easy-to-use, user-task-oriented highly interactive mapping tool prototype for visual geo-referenced data exploration is implemented using this technique. (UMIACS-TR-2003-02) (HCIL-TR-2002-26) %B Technical Reports from UMIACS %8 2003/01/21/ %G eng %U http://drum.lib.umd.edu/handle/1903/1248 %0 Journal Article %J Technical Reports from UMIACS %D 2003 %T Immediate Usability: Kiosk design principles from the CHI 2001 Photo Library %A Kules,Bill %A Kang,Hyunmo %A Plaisant, Catherine %A Rose,Anne %A Shneiderman, Ben %K Technical Report %X This paper describes a novel set of design principles and guidelines forensuring the immediate usability of public access systems. These principles and guidelines were formulated while developing PhotoFinder Kiosk, a community photo library. Attendees of CHI 2001 successfully used the tool to browse and annotate collections of photographs spanning 20 years of CHI and related conferences, producing a richly annotated photo history of the field of human-computer interaction. We used observations and log data to evaluate the tool and refine the guidelines. They provide specific guidance for practitioners, as well as a useful framework for additional research in public access interfaces. Keywords Photo collection, community photo library, group annotation, public access system, direct annotation, direct manipulation, drag-and-drop, immediate usability, zero-trial learning, walk-up-and-use, casual use. (UMIACS-TR-2001-71) (HCIL-TR-2001-23) %B Technical Reports from UMIACS %8 2003/01/21/ %G eng %U http://drum.lib.umd.edu/handle/1903/1155 %0 Journal Article %J Theory of Computing %D 2003 %T An Improved Approximation Ratio for the Covering Steiner Problem %A Gupta,Anupam %A Srinivasan, Aravind %X In the Covering Steiner problem, we are given an undirected graph with edge-costs, and some subsets of vertices called groups, with each group being equipped with a non-negative integer value (called its requirement); the problem is to find a minimum-cost tree which spans at least the required number of vertices from every group. The Covering Steiner problem is a common generalization of the k-MST and the Group Steiner problems; indeed, when all the vertices of the graph lie in one group with a requirement of k, we get the k-MST problem, and when there are multiple groups with unit requirements, we obtain the Group Steiner problem. While many covering problems (e.g., the covering integer programs such as set cover) become easier to approximate as the requirements increase, the Covering Steiner problem remains at least as hard to approximate as the Group Steiner problem; in fact, the best guarantees previously known for the Covering Steiner problem were worse than those for Group Steiner as the requirements became large. In this work, we present an improved approximation algorithm whose guarantee equals the best known guarantee for the Group Steiner problem. %B Theory of Computing %V 2 %P 53 - 64 %8 2003/// %G eng %U http://repository.cmu.edu/compsci/850 %0 Journal Article %J Pattern Recognition Letters %D 2003 %T Improved search heuristics for the sa-tree %A Hjaltason,Gı́sli R. %A Samet, Hanan %K distance-based indexing %K Metric spaces %K Nearest neighbor algorithms %X The sa-tree is an interesting metric space indexing structure that is inspired by the Voronoi diagram. In essence, the sa-tree records a portion of the Delaunay graph of the data set, a graph whose vertices are the Voronoi cells, with edges between adjacent cells. An improvement is presented on the original search strategy for the sa-tree. This consists of details on the intuition behind the improvement as well as the original search strategy and a proof of their correctness. Furthermore, it is shown how to adapt an incremental nearest neighbor algorithm to the sa-tree, which allows computing nearest neighbor in a progressive manner. Unlike other adaptations, the resulting algorithm does not take the unnecessary steps to ensure that keys of “node” elements are monotonically non-decreasing. %B Pattern Recognition Letters %V 24 %P 2785 - 2795 %8 2003/11// %@ 0167-8655 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167865503001223 %N 15 %R 10.1016/S0167-8655(03)00122-3 %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T Improving access to large volumes of online data %A Tanin,Egemen %A Samet, Hanan %X The Internet has recently become the medium of interaction with large volumes of data. Enterprises in the public and private sectors made their databases available over the Internet. Working with such large volumes of online data is a challenging task. For efficient access to large data from remote locations we introduced APPOINT (an Approach for Peer-to-Peer Offloading the INTernet). In APPOINT, active clients of a client-server architecture act on the server's behalf and communicate with each other to transfer large volumes of online data more efficiently. In essence, a server is enabled to appoint alternatives and create a scalable collaborative virtual mirror from the active clients. Multiple parameters such as availability of clients and bandwidth information on clients are considered to decide on how to best forward a download request. APPOINT is built as an add-on to existing client-server systems. A library of functions, with a simple application programming interface (API) that can be used within an existing client-server system to improve the service, is developed. Our experimental findings show that APPOINT can dramatically improve the performance of existing client-server based database systems. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 6 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123260 %0 Conference Paper %B 3rd IEEE/ACM International Symposium on Cluster Computing and the Grid, 2003. Proceedings. CCGrid 2003 %D 2003 %T Improving access to multi-dimensional self-describing scientific datasets %A Nam,B. %A Sussman, Alan %K Application software %K application-specific semantic metadata %K Bandwidth %K Computer science %K database indexing %K disk I/O bandwidth %K distributed databases %K Educational institutions %K Indexing %K indexing structures %K Libraries %K meta data %K Middleware %K multidimensional arrays %K multidimensional datasets %K Multidimensional systems %K NASA %K NASA remote sensing data %K Navigation %K query formulation %K self-describing scientific data file formats %K structural metadata %K very large databases %X Applications that query into very large multidimensional datasets are becoming more common. Many self-describing scientific data file formats have also emerged, which have structural metadata to help navigate the multi-dimensional arrays that are stored in the files. The files may also contain application-specific semantic metadata. In this paper, we discuss efficient methods for performing searches for subsets of multi-dimensional data objects, using semantic information to build multidimensional indexes, and group data items into properly sized chunks to maximize disk I/O bandwidth. This work is the first step in the design and implementation of a generic indexing library that will work with various high-dimension scientific data file formats containing semantic information about the stored data. To validate the approach, we have implemented indexing structures for NASA remote sensing data stored in the HDF format with a specific schema (HDF-EOS), and show the performance improvements that are gained from indexing the datasets, compared to using the existing HDF library for accessing the data. %B 3rd IEEE/ACM International Symposium on Cluster Computing and the Grid, 2003. Proceedings. CCGrid 2003 %I IEEE %P 172 - 179 %8 2003/05/12/15 %@ 0-7695-1919-9 %G eng %R 10.1109/CCGRID.2003.1199366 %0 Report %D 2003 %T Improving Accessibility and Usability of Geo-referenced Statistical Data %A Zhao,Haixia %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X Several technology breakthroughs are needed to achieve the goals ofuniversal accessibility and usability. These goals are especially challenging in the case of geo-referenced statistical data that many U.S. government agencies supply. We present technical and user-interface design challenges in accommodating users with low-end technology (slow network connection and low-end machine) and users who are blind or vision-impaired. Our solutions are presented and future work is discussed. (UMIACS-TR-2003-37) (HCIL-2003-11) %B Technical Reports from UMIACS %8 2003/06/04/ %G eng %U http://drum.lib.umd.edu/handle/1903/1277 %0 Journal Article %J Nucleic Acids ResearchNucl. Acids Res. %D 2003 %T Improving the Arabidopsis genome annotation using maximal transcript alignment assemblies %A Haas,Brian J. %A Delcher,Arthur L. %A Mount, Stephen M. %A Wortman,Jennifer R. %A Jr,Roger K. Smith %A Hannick,Linda I. %A Maiti,Rama %A Ronning,Catherine M. %A Rusch,Douglas B %A Town,Christopher D. %A Salzberg,Steven L. %A White,Owen %X The spliced alignment of expressed sequence data to genomic sequence has proven a key tool in the comprehensive annotation of genes in eukaryotic genomes. A novel algorithm was developed to assemble clusters of overlapping transcript alignments (ESTs and full‐length cDNAs) into maximal alignment assemblies, thereby comprehensively incorporating all available transcript data and capturing subtle splicing variations. Complete and partial gene structures identified by this method were used to improve The Institute for Genomic Research Arabidopsis genome annotation (TIGR release v.4.0). The alignment assemblies permitted the automated modeling of several novel genes and >1000 alternative splicing variations as well as updates (including UTR annotations) to nearly half of the ∼27 000 annotated protein coding genes. The algorithm of the Program to Assemble Spliced Alignments (PASA) tool is described, as well as the results of automated updates to Arabidopsis gene annotations. %B Nucleic Acids ResearchNucl. Acids Res. %V 31 %P 5654 - 5666 %8 2003/10/01/ %@ 0305-1048, 1362-4962 %G eng %U http://nar.oxfordjournals.org/content/31/19/5654 %N 19 %R 10.1093/nar/gkg770 %0 Journal Article %J Comptes Rendus Biologies %D 2003 %T In vivo filtering of in vitro expression data reveals MyoD targets %A Zhao,Po %A Seo,Jinwook %A Wang,Zuyi %A Wang,Yue %A Shneiderman, Ben %A Hoffman,Eric P %K clustering %K expression profiling %K Microarray %K microréseaux %K muscle regeneration %K MyoD %K profil d'expression %K régénération musculaire %K regroupement %X A published set of downstream targets of MyoD defined in a well-controlled in vitro experiment was filtered for relevance to muscle regeneration using a 27-time-point in vivo murine regeneration series. Using interactive hierarchical and Bayes soft clustering, only a minority of the targets defined in vitro can be confirmed in vivo (∼50% of induced transcripts, and none of repressed transcripts). This approach provided strong support that 18 targets including of MyoD are biologically relevant during myoblast differentiation. To cite this article: P. Zhao et al., C. R. Biologies 326 (2003). %B Comptes Rendus Biologies %V 326 %P 1049 - 1065 %8 2003/10// %@ 1631-0691 %G eng %U http://www.sciencedirect.com/science/article/pii/S1631069103002324 %N 10–11 %R 10.1016/j.crvi.2003.09.035 %0 Journal Article %J ACM Trans. Database Syst. %D 2003 %T Index-driven similarity search in metric spaces (Survey Article) %A Hjaltason,Gisli R. %A Samet, Hanan %K distance-based indexing %K Hiearchical metric data structures %K nearest neighbor queries %K range queries %K Ranking %K similarity searching %X Similarity search is a very important operation in multimedia databases and other database applications involving complex objects, and involves finding objects in a data set S similar to a query object q, based on some similarity measure. In this article, we focus on methods for similarity search that make the general assumption that similarity is represented with a distance metric d. Existing methods for handling similarity search in this setting typically fall into one of two classes. The first directly indexes the objects based on distances (distance-based indexing), while the second is based on mapping to a vector space (mapping-based approach). The main part of this article is dedicated to a survey of distance-based indexing methods, but we also briefly outline how search occurs in mapping-based methods. We also present a general framework for performing search based on distances, and present algorithms for common types of queries that operate on an arbitrary "search hierarchy." These algorithms can be applied on each of the methods presented, provided a suitable search hierarchy is defined. %B ACM Trans. Database Syst. %V 28 %P 517 - 580 %8 2003/12// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/958942.958948 %N 4 %R 10.1145/958942.958948 %0 Book Section %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %D 2003 %T Innovating the Interaction %A Bederson, Benjamin B. %A Plaisant, Catherine %A Mushlin,R. %A Snyder,A. %A Li,J. %A Heller,D. %A Shneiderman, Ben %A Hochheiser,H. %A Fekete,J. D %A Czenvinski,M. %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %I Morgan Kaufmann %C San Francisco %P 295 - 295 %8 2003/// %@ 978-1-55860-915-0 %G eng %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2003. IEEE %D 2003 %T Integral equation solution of electromagnetic scattering from a multilayered cylindrical waveguide %A Seydou,F. %A Duraiswami, Ramani %A Seppanen,T. %K approximation %K circular %K core; %K cylinder; %K cylindrical %K dielectric %K dielectric-loaded %K electromagnetic %K EM %K equations; %K integral %K Maxwell %K method; %K multilayered %K numerical %K Nystrom %K scattering; %K theory; %K wave %K waveguide %K waveguides; %X This paper is devoted to the electromagnetic scattering from an N multilayered cylinder. We consider waveguides in the z direction, that is: we look for the solution of Maxwell equations along the z direction. We assume a dielectric core and discuss the problem for the case of general domains. We use an integral equation approach to solve the problem and the Nystrom method for the numerical approximation. %B Antennas and Propagation Society International Symposium, 2003. IEEE %V 3 %P 524 - 527 vol.3 - 524 - 527 vol.3 %8 2003/06// %G eng %R 10.1109/APS.2003.1219901 %0 Conference Paper %B Proceedings of the fourteenth annual ACM-SIAM symposium on Discrete algorithms %D 2003 %T Integrality ratio for group Steiner trees and directed steiner trees %A Halperin,Eran %A Kortsarz,Guy %A Krauthgamer,Robert %A Srinivasan, Aravind %A Wang,Nan %X We present an Ω(log2k) lower bound on the integrality ratio of the flow-based relaxation for the Group Steiner Tree problem, where k denotes the number of groups; this holds even for input graphs that are Hierarchically Well-Separated Trees, introduced by Bartal [Symp. Foundations of Computer Science, pp. 184--193, 1996], in which case this lower bound is tight. This relaxation appears to be the only one that have been studied for the problem, as well as for its generalization, the Directed Steiner Tree problem. For the latter problem, our results imply an Ω(log2n/(log logn)2) integrality ratio, where n is the number of vertices in the graph. For both problems, this is the first known lower bound on the integrality ratio that is superlogarithmic in the input size. We also show algorithmically that the integrality ratio for Group Steiner Tree is much better for certain families of instances, which helps pinpoint the types of instances that appear to be most difficult to approximate. %B Proceedings of the fourteenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '03 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 275 - 284 %8 2003/// %@ 0-89871-538-5 %G eng %U http://dl.acm.org/citation.cfm?id=644108.644155 %0 Conference Paper %B Proceedings of the 9th international conference on Tools and algorithms for the construction and analysis of systems %D 2003 %T The integrated CWB-NC/PIOAtool for functional verification and performance analysis of concurrent systems %A Zhang,D. %A Cleaveland, Rance %A Stark,E. W %B Proceedings of the 9th international conference on Tools and algorithms for the construction and analysis of systems %P 431 - 436 %8 2003/// %G eng %0 Conference Paper %B Multimedia and Expo, 2003. ICME'03. Proceedings. 2003 International Conference on %D 2003 %T Interactive color mosaic and dendrogram displays for signal/noise optimization in microarray data analysis %A Seo,J. %A Bakay,M. %A Zhao,P. %A Chen,Y.W. %A Clarkson,P. %A Shneiderman, Ben %A Hoffman,E.P. %X Data analysis and visualization is strongly influenced by noise and noise filters. There are multiple sources of oisein microarray data analysis, but signal/noise ratios are rarely optimized, or even considered. Here, we report a noise analysis of a novel 13 million oligonucleotide dataset - 25 human U133A (~500,000 features) profiles of patient muscle biposies. We use our recently described interactive visualization tool, the Hierarchical Clustering Explorer (HCE) to systemically address the effect of different noise filters on resolution of arrays into orrectbiological groups (unsupervised clustering into three patient groups of known diagnosis). We varied probe set interpretation methods (MAS 5.0, RMA), resent callfilters, and clustering linkage methods, and investigated the results in HCE. HCE interactive features enabled us to quickly see the impact of these three variables. Dendrogram displays showed the clustering results systematically, and color mosaic displays provided a visual support for the results. We show that each of these three variables has a strong effect on unsupervised clustering. For this dataset, the strength of the biological variable was maximized, and noise minimized, using MAS 5.0, 10% present call filter, and Average Group Linkage. We propose a general method of using interactive tools to identify the optimal signal/noise balance or the optimal combination of these three variables to maximize the effect of the desired biological variable on data interpretation. %B Multimedia and Expo, 2003. ICME'03. Proceedings. 2003 International Conference on %V 3 %P III–461 - III–461 %8 2003/// %G eng %0 Journal Article %J IEEE Transactions on Visualization and Computer Graphics %D 2003 %T Interactive Visualization of Large Tetrahedral Meshes through Selective Refinement %A Cignoni,P. %A De Floriani, Leila %A Magillo,P. %A Puppo,E. %A Scopigno,R. %A di Genova,U. %X In this paper, we address the problem of the efficient visualization of very largeirregular volume datasets. To this aim, we exploit a multiresolution representation based on a domain decomposition into tetrahedral cells. A new compact data struc- ture is described which encodes the whole dataset at a virtually continuous range of different resolutions, with a storage cost six times lower than a standard data struc- ture for tetrahedral meshes. Such structure supports on-line selective refinement to focus resolution on areas that the user considers more critical, based on either field values, or domain location, or opacity of the transfer function. Selective refinement is used to trade-off between resolution and speed in visualization, according to user needs and hardware constraints. These features have been implemented in a new system, called TAn2 (Tetrahedra Analyzer), for the interactive visualization of three- dimensional scalar fields defined on very large tetrahedral meshes. Multiresolution representation and selective refinement make the system fully scalable with respect to the size of the dataset and to hardware requirements. %B IEEE Transactions on Visualization and Computer Graphics %8 2003/// %G eng %0 Conference Paper %B Proceedings of the 11th ACM international symposium on Advances in geographic information systems %D 2003 %T The internet spatial spreadsheet: enabling remote visualization of dynamic spatial data and ongoing query results over a network %A Iwerks,Glenn S. %A Samet, Hanan %K client server %K GIS %K Spatial databases %K Visualization %X Moving object databases store and process data for objects that change location frequently. Materialized views maintained over time must be updated to reflect changes due to the motion of objects in their environment. To visualize view query results, displays must be updated to reflect the change. In this paper we present the Internet Spatial Spreadsheet (ISS) as a means to organize, query, and visualize changing spatial data in a network environment such as the Internet.The goal of the ISS is to keep client visualizations of query results up to date with the server state. This is accomplished by pushing the minimal set of spatial data needed for rendering query results on the client. Incremental changes to query results are subsequently transmitted to the client as the database is updated to keep the visualization current. Additional constraints in the network environment such as firewall limitations are also considered. %B Proceedings of the 11th ACM international symposium on Advances in geographic information systems %S GIS '03 %I ACM %C New York, NY, USA %P 154 - 160 %8 2003/// %@ 1-58113-730-3 %G eng %U http://doi.acm.org/10.1145/956676.956697 %R 10.1145/956676.956697 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2003-70 %D 2003 %T Intradomain Overlays: Architecture and Applications %A Kommareddy,Christopher %A Guven,Tuna %A Bhattacharjee, Bobby %A La,Richard %A Shayman,Mark %K Technical Report %X We introduce an architecture for ``Intradomain Overlays'', where a subset of routers within a domain is augmented with a dedicated host. These strategically placed hosts form an overlay network, and we describe a number of applications for such overlays. These applications include efficient network monitoring, policy- and load-based packet re-routing, and network resource accounting. In this paper, we elaborate on the network monitoring application and describe a distributed protocol for monitoring routers within an AS which has been augmented with a few overlay nodes. The routers and other infrastructure are unaware of the overlay nodes, and the monitoring of individual elements is conducted using plain SNMP. We describe techniques for efficiently synthesizing and transporting the monitored SNMP data, and present results using trace data collected from an AS with 400+ routers. Our results show that the overlay-based monitoring reduces overheads by 2--4 orders of magnitude, and thus enables much finer grained monitoring and traffic engineering than is otherwise possible. (UMIACS-TR-2003-70) %B Technical Reports from UMIACS, UMIACS-TR-2003-70 %8 2003/08/01/ %G eng %U http://drum.lib.umd.edu/handle/1903/1295 %0 Journal Article %J IT&SOCIETY %D 2003 %T INTRODUCTION TO VOLUME 1, ISSUE 3: WEB NAVIGATION %A Shneiderman, Ben %A Lazar,J. %A IVORY,M. %B IT&SOCIETY %V 1 %8 2003/// %G eng %N 3 %0 Journal Article %J Intel Research, UC Berkeley, Carnegie Mellon University %D 2003 %T IRIS: Internet-scale Resource-Intensive Sensor Services %A Deshpande, Amol %A Nath,S. %A Gibbons,P.B. %A Seshan,S. %X The proliferation and affordability of smart sensors such aswebcams, microphones etc., has created opportunities for exciting new classes of distributed services. A key stum- bling block to mining these rich information sources is the lack of a common, scalable networked infrastructure for col- lecting, filtering, and combining the video feeds, extracting the useful information, and enabling distributed queries. In this demo, we demonstrate the design and an early prototype of such an infrastructure, called IRIS (Internet- scale Resource-Intensive Sensor services). IRIS is a poten- tially global network of smart sensor nodes, with webcams or other sensors, and organizing nodes that provide the means to query recent and historical sensor-based data. IRIS ex- ploits the fact that high-volume sensor feeds are typically attached to devices with significant computing power and storage, and running a standard operating system. Aggres- sive filtering, smart query routing, and semantic caching are used to dramatically reduce network bandwidth utilization and improve query response times, as we will demonstrate. The service that we demonstrate here is that of a park- ing space finder. This service utilizes webcams that monitor parking spaces to answer queries such as the availability of parking spaces near a user’s destination. %B Intel Research, UC Berkeley, Carnegie Mellon University %8 2003/// %G eng %0 Conference Paper %B Proceedings of the 29th international conference on Very large data bases - Volume 29 %D 2003 %T IrisNet: an architecture for internet-scale sensing services %A Nath,Suman %A Deshpande, Amol %A Ke,Yan %A Gibbons,Phillip B. %A Karp,Brad %A Seshan,Srinivasan %X We demonstrate the design and an early prototype of IrisNet (Internet-scale Resource-Intensive Sensor Network services), a common, scalable networked infrastructure for deploying wide area sensing services. IrisNet is a potentially global network of smart sensing nodes, with webcams or other monitoring devices, and organizing nodes that provide the means to query recent and historical sensor-based data. IrisNet exploits the fact that high-volume sensor feeds are typically attached to devices with significant computing power and storage, and running a standard operating system. It uses aggressive filtering, smart query routing, and semantic caching to dramatically reduce network bandwidth utilization and improve query response times, as we demonstrate. Our demo will present two services built on Iris-Net, from two very different application domains. The first one, a parking space finder, utilizes webcams that monitor parking spaces to answer queries such as the availability of parking spaces near a user's destination. The second one, a distributed infrastructure monitor, uses measurement tools installed in individual nodes of a large distributed infrastructure to answer queries such as average network bandwidth usage of a set of nodes. %B Proceedings of the 29th international conference on Very large data bases - Volume 29 %S VLDB '03 %I VLDB Endowment %P 1137 - 1140 %8 2003/// %@ 0-12-722442-4 %G eng %U http://dl.acm.org/citation.cfm?id=1315451.1315568 %0 Journal Article %J ACM Trans. Database Syst. %D 2003 %T Iterative spatial join %A Jacox,Edwin H. %A Samet, Hanan %K external memory algorithms %K plane-sweep %K Spatial databases %K Spatial join %X The key issue in performing spatial joins is finding the pairs of intersecting rectangles. For unindexed data sets, this is usually resolved by partitioning the data and then performing a plane sweep on the individual partitions. The resulting join can be viewed as a two-step process where the partition corresponds to a hash-based join while the plane-sweep corresponds to a sort-merge join. In this article, we look at extending the idea of the sort-merge join for one-dimensional data to multiple dimensions and introduce the Iterative Spatial Join. As with the sort-merge join, the Iterative Spatial Join is best suited to cases where the data is already sorted. However, as we show in the experiments, the Iterative Spatial Join performs well when internal memory is limited, compared to the partitioning methods. This suggests that the Iterative Spatial Join would be useful for very large data sets or in situations where internal memory is a shared resource and is therefore limited, such as with today's database engines which share internal memory amongst several queries. Furthermore, the performance of the Iterative Spatial Join is predictable and has no parameters which need to be tuned, unlike other algorithms. The Iterative Spatial Join is based on a plane sweep algorithm, which requires the entire data set to fit in internal memory. When internal memory overflows, the Iterative Spatial Join simply makes additional passes on the data, thereby exhibiting only a gradual performance degradation. To demonstrate the use and efficacy of the Iterative Spatial Join, we first examine and analyze current approaches to performing spatial joins, and then give a detailed analysis of the Iterative Spatial Join as well as present the results of extensive testing of the algorithm, including a comparison with partitioning-based spatial join methods. These tests show that the Iterative Spatial Join overcomes the performance limitations of the other algorithms for data sets of all sizes as well as differing amounts of internal memory. %B ACM Trans. Database Syst. %V 28 %P 230 - 256 %8 2003/09// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/937598.937600 %N 3 %R 10.1145/937598.937600 %0 Conference Paper %B IEEE International Conference on Systems, Man and Cybernetics, 2003 %D 2003 %T Kernel snakes: non-parametric active contour models %A Abd-Almageed, Wael %A Smith,C.E. %A Ramadan,S. %K Active contours %K Artificial intelligence %K Bayes methods %K Bayesian decision theory %K Bayesian methods %K decision theory %K Deformable models %K Image edge detection %K Image segmentation %K Intelligent robots %K Kernel %K kernel snakes %K Laboratories %K multicolored target tracking %K nonparametric active contour models %K nonparametric generalized formulation %K nonparametric model %K nonparametric statistics %K nonparametric techniques %K real time performance %K Robot vision systems %K statistical pressure snakes %K target tracking %X In this paper, a new non-parametric generalized formulation to statistical pressure snakes is presented. We discuss the shortcomings of the traditional pressure snakes. We then introduce a new generic pressure model that alleviates these shortcomings, based on the Bayesian decision theory. Non-parametric techniques are used to obtain the statistical models that drive the snake. We discuss the advantages of using the proposed non-parametric model compared to other parametric techniques. Multi-colored-target tracking is used to demonstrate the performance of the proposed approach. Experimental results show enhanced, real-time performance. %B IEEE International Conference on Systems, Man and Cybernetics, 2003 %I IEEE %V 1 %P 240- 244 vol.1 - 240- 244 vol.1 %8 2003/10// %@ 0-7803-7952-7 %G eng %R 10.1109/ICSMC.2003.1243822 %0 Conference Paper %B Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on %D 2003 %T Learning dynamics for exemplar-based gesture recognition %A Elgammal,A. %A Shet,V. %A Yacoob,Yaser %A Davis, Larry S. %K arbitrary %K body %K by %K Computer %K constraint; %K detection; %K discrete %K distribution %K dynamics; %K edge %K estimation; %K example; %K exemplar %K exemplar-based %K extraction; %K feature %K framework; %K gesture %K gesture; %K hidden %K HMM; %K human %K image %K learning %K Markov %K matching; %K model; %K models; %K motion; %K nonparametric %K pose %K probabilistic %K recognition; %K sequence; %K space; %K state; %K statistics; %K system %K temporal %K tool; %K view-based %K vision; %X This paper addresses the problem of capturing the dynamics for exemplar-based recognition systems. Traditional HMM provides a probabilistic tool to capture system dynamics and in exemplar paradigm, HMM states are typically coupled with the exemplars. Alternatively, we propose a non-parametric HMM approach that uses a discrete HMM with arbitrary states (decoupled from exemplars) to capture the dynamics over a large exemplar space where a nonparametric estimation approach is used to model the exemplar distribution. This reduces the need for lengthy and non-optimal training of the HMM observation model. We used the proposed approach for view-based recognition of gestures. The approach is based on representing each gesture as a sequence of learned body poses (exemplars). The gestures are recognized through a probabilistic framework for matching these body poses and for imposing temporal constraints between different poses using the proposed non-parametric HMM. %B Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on %V 1 %P I-571 - I-578 vol.1 - I-571 - I-578 vol.1 %8 2003/06// %G eng %R 10.1109/CVPR.2003.1211405 %0 Book %D 2003 %T Leonardo's Laptop: Human Needs and the New Computing Technologies %A Shneiderman, Ben %K Computers / Data Processing %K Computers / Data Processing / General %K Computers / Data Processing / Storage & Retrieval %K Computers / Social Aspects / Human-Computer Interaction %K Computers / System Administration / Storage & Retrieval %K Electronic data processing %K Human-computer interaction %K Technological forecasting %K Technology & Engineering / Social Aspects %X 2003 IEEE-USAB Award for Distinguished Literary Contributions Furthering Public Understanding of the Profession. and Selected as a Finalist in the category of Computer/Internet in the 2002 Independent Publisher Book Awards (IPPYs) presented by Independent Publisher MagazineBen Shneiderman's book dramatically raises computer users' expectations of what they should get from technology. He opens their eyes to new possibilities and invites them to think freshly about future technology. He challenges developers to build products that better support human needs and that are usable at any bandwidth. Shneiderman proposes Leonardo da Vinci as an inspirational muse for the "new computing." He wonders how Leonardo would use a laptop and what applications he would create.Shneiderman shifts the focus from what computers can do to what users can do. A key transformation is to what he calls "universal usability," enabling participation by young and old, novice and expert, able and disabled. This transformation would empower those yearning for literacy or coping with their limitations. Shneiderman proposes new computing applications in education, medicine, business, and government. He envisions a World Wide Med that delivers secure patient histories in local languages at any emergency room and thriving million-person communities for e-commerce and e-government. Raising larger questions about human relationships and society, he explores the computer's potential to support creativity, consensus-seeking, and conflict resolution. Each chapter ends with a Skeptic's Corner that challenges assumptions about trust, privacy, and digital divides. %I MIT Press %8 2003/09/01/ %@ 9780262692991 %G eng %0 Book Section %B The Craft of Information VisualizationThe Craft of Information Visualization %D 2003 %T LifeLines: Using Visualization to Enhance Navigation and Analysis of Patient Records %A Plaisant, Catherine %A Mushlin,Richard %A Snyder,Aaron %A Li,Jia %A Heller,Dan %A Shneiderman, Ben %E Bederson, Benjamin B. %E Shneiderman, Ben %B The Craft of Information VisualizationThe Craft of Information Visualization %I Morgan Kaufmann %C San Francisco %P 308 - 312 %8 2003/// %@ 978-1-55860-915-0 %G eng %U http://www.sciencedirect.com/science/article/pii/B978155860915050038X %0 Journal Article %J EURASIP J. Appl. Signal Process. %D 2003 %T Logic foundry: rapid prototyping for FPGA-based DSP systems %A Spivey,Gary %A Bhattacharyya, Shuvra S. %A Nakajima,Kazuo %K CAD tools %K Design methodology %K DSP %K FPGA %K integration %K rapid prototyping %X We introduce the Logic Foundry, a system for the rapid creation and integration of FPGA-based digital signal processing systems. Recognizing that some of the greatest challenges in creating FPGA-based systems occur in the integration of the various components, we have proposed a system that targets the following four areas of integration: design flow integration, component integration, platform integration, and software integration. Using the Logic Foundry, a system can be easily specified, and then automatically constructed and integrated with system level software. %B EURASIP J. Appl. Signal Process. %V 2003 %P 565 - 579 %8 2003/01// %@ 1110-8657 %G eng %U http://dx.doi.org/10.1155/S1110865703301039 %R 10.1155/S1110865703301039 %0 Journal Article %J IEEE Intelligent Systems (special issue on Natural Language Processing) %D 2003 %T Machine and Human Performance for Single- and Multi-Document Summarization %A Schlesinger,Judith D. %A Conroy,John M. %A Okurowski,Mary Ellen %A O'Leary, Dianne P. %B IEEE Intelligent Systems (special issue on Natural Language Processing) %V 18 %P 46 - 54 %8 2003/// %G eng %U http://csdl.computer.org/comp/mags/ex/2003/01/x1toc.htmhttp://csdl.computer.org/comp/mags/ex/2003/01/x1toc.htm %N 1 %0 Journal Article %J ACM Transactions on Asian Language Information Processing (TALIP) %D 2003 %T Making miracles: Interactive translingual search for cebuano and hindi %A He,D. %A Oard, Douglas %A Wang,J. %A Luo,J. %A Demner-Fushman,D. %A Darwish,K. %A Resnik, Philip %A Khudanpur,S. %A Nossal,M. %A Subotin,M. %A others %B ACM Transactions on Asian Language Information Processing (TALIP) %V 2 %P 219 - 244 %8 2003/// %G eng %N 3 %0 Journal Article %J Applied Mechanics Reviews %D 2003 %T Matrix Algorithms, Volume II: Eigensystems %A Stewart, G.W. %A Author %A Mahajan,A. %A Reviewer %K algorithm theory %K Eigenvalues and eigenfunctions %K linear algebra %K matrix algebra %K reviews %B Applied Mechanics Reviews %V 56 %P B2-B2 - B2-B2 %8 2003/// %G eng %U http://link.aip.org/link/?AMR/56/B2/1 %N 1 %R 10.1115/1.1523352 %0 Conference Paper %B 5th IASTEDInternational Conference on Signal and Image Processing %D 2003 %T Measuring Structural Similarity of Document Pages for Searching Document Image Databases %A Shin,C. %A David Doermann %A Rosenfeld, A. %X Current document management and database systems provide text search and retrieval capabilities, but generally lack the ability to utilize the documents’ logical and physical structures. This paper describes a general system for document image retrieval that is able to make use of document structure. It discusses the use of structural similarity for retrieval; it defines a measure of structural similarity between document images based on content area overlap, and also compares similarity ratings based on this measure with human relevance judgments. %B 5th IASTEDInternational Conference on Signal and Image Processing %P 320 - 325 %8 2003/08// %G eng %0 Conference Paper %B CHI '03 extended abstracts on Human factors in computing systems %D 2003 %T MediaFinder: an interface for dynamic personal media management with semantic regions %A Kang,Hyunmo %A Shneiderman, Ben %K dynamic queries %K Fling-and-flock %K Personal media management %K spatial information organization %K User interfaces %X Computer users deal with large amounts of personal media often face problems in managing and exploring it. This paper presents Semantic Regions, rectangular regions that enable users to specify their semantics or mental models, and the MediaFinder application, which uses Semantic Regions as the basis of a personal media management tool. %B CHI '03 extended abstracts on Human factors in computing systems %S CHI EA '03 %I ACM %C New York, NY, USA %P 764 - 765 %8 2003/// %@ 1-58113-637-4 %G eng %U http://doi.acm.org/10.1145/765891.765977 %R 10.1145/765891.765977 %0 Journal Article %J SIGPLAN Fortran Forum %D 2003 %T Memory leaks in derived types revisited %A Stewart, G.W. %X In a note in the Fortran Forum, Markus describes a technique for avoiding memory leaks with derived types. In this note, we show by a simple example that this technique does not work when the object in question is a parameter in nested subprogram invocations. A fix is proposed and illustrated with code from MATRAN, a Fortran 95 package for performing matrix manipulations. %B SIGPLAN Fortran Forum %V 22 %P 25 - 27 %8 2003/12// %@ 1061-7264 %G eng %U http://doi.acm.org/10.1145/962180.962183 %N 3 %R 10.1145/962180.962183 %0 Journal Article %J SIGMOBILE Mob. Comput. Commun. Rev. %D 2003 %T MobiCom poster: mining a world of smart sensors %A Nath,Suman %A Deshpande, Amol %A Gibbons,Phillip B. %A Seshan,Srinivasan %B SIGMOBILE Mob. Comput. Commun. Rev. %V 7 %P 34 - 36 %8 2003/01// %@ 1559-1662 %G eng %U http://doi.acm.org/10.1145/881978.881987 %N 1 %R 10.1145/881978.881987 %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T New approaches to help users get started with visual interfaces: multi-layered interfaces and integrated initial guidance %A Kang,Hyunmo %A Plaisant, Catherine %A Shneiderman, Ben %X We are investigating new ways to help users learn to use public access interactive tools, in particular for the visual exploration of government statistics. Our work led to a series of interfaces using multi-layered design and a new help method called Integrated Initial Guidance. Multi-layer designs structure an interface so that a simpler interface is available for users to get started and more complex features are accessed as users move through the more advanced layers. Integrated Initial Guidance provides help within the working interface, right at the start of the application. Using the metaphor of "sticky notes" overlaid on top of the functional interface locates the main widgets, demonstrates their manipulation, and explains the resulting actions using preset animation of the interface. Usability testing with 12 participants led to refined designs and guidelines for the design of Integrated Initial Guidance interfaces. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 6 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123269 %0 Conference Paper %B IEEE International Conference on Systems, Man and Cybernetics, 2003 %D 2003 %T Non-parametric expectation maximization: a learning automata approach %A Abd-Almageed, Wael %A El-Osery,A. %A Smith,C.E. %K Automatic speech recognition %K bivariate simulation data %K Computational modeling %K Computer vision %K Density functional theory %K expectation maximization technique %K learning automata %K mixture parameters estimation %K nonparametric expectation maximization %K nonparametric statistics %K optimisation %K parameter estimation %K Parzen Window %K PDF %K probability %K probability density function %K Speech processing %K Speech recognition %K stochastic learning automata %K Stochastic processes %K stochastic technique %X The famous expectation maximization technique suffers two major drawbacks. First, the number of components has to be specified apriori. Also, the expectation maximization is sensitive to initialization. In this paper, we present a new stochastic technique for estimating the mixture parameters. Parzen Window is used to estimate a discrete estimate of the PDF of the given data. Stochastic learning automata is then used to select the mixture parameters that minimize the distance between the discrete estimate of the PDF and the estimate of the expectation maximization. The validity of the proposed approach is verified using bivariate simulation data. %B IEEE International Conference on Systems, Man and Cybernetics, 2003 %I IEEE %V 3 %P 2996- 3001 vol.3 - 2996- 3001 vol.3 %8 2003/10// %@ 0-7803-7952-7 %G eng %R 10.1109/ICSMC.2003.1244347 %0 Report %D 2003 %T An O(n)-Space O(log n/log log n + f)-Query Time Algorithm for 3-D Dominance Reporting %A Shi,Qingmin %A JaJa, Joseph F. %K Technical Report %X We present a linear-space algorithm for handling the {\emthree-dimensional dominance reporting problem}: given a set $S$ of $n$ three-dimensional points, design a data structure for $S$ so that the points in $S$ which dominate a given query point can be reported quickly. Under the variation of the RAM model introduced by Fredman and Willard~\cite{Fredman94}, our algorithm achieves $O(\log n/\log\log n+f)$ query time, where $f$ is the number of points reported. Extensions to higher dimensions are also reported. (UMIACS-TR-2003-77) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-77 %8 2003/08/01/ %G eng %U http://drum.lib.umd.edu/handle/1903/1301 %0 Journal Article %J Advances in Informatics %D 2003 %T The Opsis project: materialized views for data warehouses and the web %A Roussopoulos, Nick %A Kotidis,Y. %A Labrinidis,A. %A Sismanis,Y. %B Advances in Informatics %P 64 - 81 %8 2003/// %G eng %0 Report %D 2003 %T Optimizing Heavily Loaded Agents %A V.S. Subrahmanian %X We develop algorithms to help scale software agents built on top of heterogeneous, legacy codebases. The algorithms apply to large data sets, to large volumes of workloads on agents, as well as algorithms for computationally intensive functions. %I Department of Computer Science, University of Maryland, College Park %8 2003/06// %G eng %0 Journal Article %J Infection and ImmunityInfect. Immun. %D 2003 %T Pathogenic Potential of Environmental Vibrio Cholerae Strains Carrying Genetic Variants of the Toxin-Coregulated Pilus Pathogenicity Island %A Faruque,Shah M. %A Kamruzzaman,M. %A Meraj,Ismail M. %A Chowdhury,Nityananda %A Nair,G. Balakrish %A Sack,R. Bradley %A Rita R Colwell %A Sack,David A. %X The major virulence factors of toxigenic Vibrio cholerae are cholera toxin (CT), which is encoded by a lysogenic bacteriophage (CTXΦ), and toxin-coregulated pilus (TCP), an essential colonization factor which is also the receptor for CTXΦ. The genes for the biosynthesis of TCP are part of a larger genetic element known as the TCP pathogenicity island. To assess their pathogenic potential, we analyzed environmental strains of V. cholerae carrying genetic variants of the TCP pathogenicity island for colonization of infant mice, susceptibility to CTXΦ, and diarrheagenicity in adult rabbits. Analysis of 14 environmental strains, including 3 strains carrying a new allele of the tcpA gene, 9 strains carrying a new allele of the toxT gene, and 2 strains carrying conventional tcpA and toxT genes, showed that all strains colonized infant mice with various efficiencies in competition with a control El Tor biotype strain of V. cholerae O1. Five of the 14 strains were susceptible to CTXΦ, and these transductants produced CT and caused diarrhea in adult rabbits. These results suggested that the new alleles of the tcpA and toxT genes found in environmental strains of V. cholerae encode biologically active gene products. Detection of functional homologs of the TCP island genes in environmental strains may have implications for understanding the origin and evolution of virulence genes of V. cholerae. %B Infection and ImmunityInfect. Immun. %V 71 %P 1020 - 1025 %8 2003/02/01/ %@ 0019-9567, 1098-5522 %G eng %U http://iai.asm.org/content/71/2/1020 %N 2 %R 10.1128/IAI.71.2.1020-1025.2003 %0 Book Section %B DUC 03 Conference ProceedingsDUC 03 Conference Proceedings %D 2003 %T Performance of a Three-Stage System for Multi-Document Summarization %A Dunlavy,Daniel M. %A Conroy,John M. %A Schlesinger,Judith D. %A Goodman,Sarah A. %A Okurowski,Mary Ellen %A O'Leary, Dianne P. %A Halteren,Hans van %B DUC 03 Conference ProceedingsDUC 03 Conference Proceedings %I U.S. National Inst. of Standards and Technology %8 2003/// %G eng %U http://duc.nist.gov/http://duc.nist.gov/ %0 Journal Article %J Environmental Microbiology %D 2003 %T Persistence of adhesive properties in Vibrio cholerae after long‐term exposure to sea water %A Pruzzo,Carla %A Tarsi,Renato %A Del Mar Lleò,Maria %A Signoretto,Caterina %A Zampini,Massimiliano %A Pane,Luigi %A Rita R Colwell %A Canepari,Pietro %X The effect of exposure to artificial sea water (ASW) on the ability of classical Vibrio cholerae O1 cells to interact with chitin-containing substrates and human intestinal cells was studied. Incubation of vibrios in ASW at 5°C and 18°C resulted in two kinds of cell responses: the viable but non-culturable (VBNC) state (i.e. <0.1 colony forming unit ml−1) at 5°C, and starvation (i.e. maintenance of culturability of the population) at 18°C. The latter remained rod shaped and, after 40 days’ incubation, presented a 47–58% reduction in the number of cells attached to chitin, a 48–53% reduction in the number of bacteria adhering to copepods, and a 48–54% reduction in the number of bacteria adhering to human cultured intestinal cells, compared to control cells not suspended in ASW. Bacteria suspended in ASW at 5°C became coccoid and, after 40 days, showed 34–42% fewer cells attached to chitin, 52–55% fewer adhering to copep-ods, and 45–48% fewer cells adhering to intestinal cell monolayers, compared to controls. Sarkosyl-insoluble membrane proteins that bind chitin particles were isolated and analysed by SDS-PAGE. After 40 days incubation in ASW at both 5°C and 18°C vibrios expressed chitin-binding ligands similar to bacteria harvested in the stationary growth phase. It is concluded that as vibrios do not lose adhesive properties after long-term exposure to ASW, it is important to include methods for VBNC bacteria when testing environmental and clinical samples for purposes of public health safety. %B Environmental Microbiology %V 5 %P 850 - 858 %8 2003/10/01/ %@ 1462-2920 %G eng %U http://onlinelibrary.wiley.com/doi/10.1046/j.1462-2920.2003.00498.x/abstract?userIsAuthenticated=false&deniedAccessCustomisedMessage= %N 10 %R 10.1046/j.1462-2920.2003.00498.x %0 Conference Paper %B Multimedia and Expo, IEEE International Conference on %D 2003 %T Pitch and timbre manipulations using cortical representation of sound %A Zotkin,Dmitry N %A Shamma,S.A. %A Ru,P. %A Duraiswami, Ramani %A Davis, Larry S. %X The sound receiver at the ears is processed by humans using signal processing that separate the signal along intensity, pitch and timbre dimensions. Conventional Fourier-based signal processing, while endowed with fast algorithms, is unable to easily represent signal along these attributes. In this paper we use a cortical representation to represent the manipulate sound. We briefly overview algorithms for obtaining, manipulating and inverting cortical representation of sound and describe algorithms for manipulating signal pitch and timbre separately. The algorithms are first used to create sound of an instrument between a guitar and a trumpet. Applications to creating maximally separable sounds in auditory user interfaces are discussed. %B Multimedia and Expo, IEEE International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %V 3 %P 381 - 384 %8 2003/// %@ 0-7803-7965-9 %G eng %R http://doi.ieeecomputersociety.org/10.1109/ICME.2003.1221328 %0 Journal Article %J International Conference on Automated Planning and Scheduling (ICAPS) 2003 Workshop on planning for web services %D 2003 %T A planner for composing services described in DAML-S %A Sheshagiri,M. %A desJardins, Marie %A Finin,T. %B International Conference on Automated Planning and Scheduling (ICAPS) 2003 Workshop on planning for web services %8 2003/// %G eng %0 Journal Article %J Numerische Mathematik %D 2003 %T On the powers of a matrix with perturbations %A Stewart, G.W. %X Let A be a matrix of order n. The properties of the powers A k of A have been extensively studied in the literature. This paper concerns the perturbed powers Pk=(A+Ek)(A+Ek−1)(A+E1) where the E k are perturbation matrices. We will treat three problems concerning the asymptotic behavior of the perturbed powers. First, determine conditions under which Pk0 . Second, determine the limiting structure of P k . Third, investigate the convergence of the power method with error: that is, given u 1 , determine the behavior of u k =ngr k P k u 1 , where ngr k is a suitable scaling factor. %B Numerische Mathematik %V 96 %P 363 - 376 %8 2003/// %G eng %N 2 %R 10.1007/s00211-003-0470-0 %0 Conference Paper %B INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE %D 2003 %T Probabilistically survivable mass %A Kraus,S. %A V.S. Subrahmanian %A Tas,N. C %X Multiagent systems (MAS) can "go down" for alarge number of reasons, ranging from system mal- functions and power failures to malicious attacks. The placement of agents on nodes is called a de- ployment of the MAS. We develop a probabilis- tic model of survivability of a deployed MAS and provide two algorithms to compute the probability of survival of a deployed MAS. Our probabilistic model docs not make independence assumptions though such assumptions can be added if so de- sired. An optimal deployment of a MAS is one that maximizes its survival probability. We provide a mathematical answerto this question, an algorithm that computes an exact solution to this problem, as well as several algorithms that quickly compute approximate solutions to the problem. We have implemented our algorithms - our implementation demonstrates that computing deployments can be done scalably. %B INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE %V 18 %P 789 - 795 %8 2003/// %G eng %0 Journal Article %J CONCUR 2003-Concurrency Theory %D 2003 %T A process-algebraic language for probabilistic I/O automata %A Stark,E. W %A Cleaveland, Rance %A Smolka,S. A %B CONCUR 2003-Concurrency Theory %P 193 - 207 %8 2003/// %G eng %0 Journal Article %J Pattern Analysis and Machine Intelligence, IEEE Transactions on %D 2003 %T Properties of embedding methods for similarity searching in metric spaces %A Hjaltason,G. R %A Samet, Hanan %K complex %K contractiveness; %K data %K databases; %K decomposition; %K dimension %K distance %K distortion; %K DNA %K documents; %K EMBEDDING %K embeddings; %K Euclidean %K evaluations; %K FastMap; %K images; %K Lipschitz %K methods; %K metric %K MetricMap; %K multimedia %K processing; %K query %K reduction %K search; %K searching; %K sequences; %K similarity %K singular %K spaces; %K SparseMap; %K types; %K value %X Complex data types-such as images, documents, DNA sequences, etc.-are becoming increasingly important in modern database applications. A typical query in many of these applications seeks to find objects that are similar to some target object, where (dis)similarity is defined by some distance function. Often, the cost of evaluating the distance between two objects is very high. Thus, the number of distance evaluations should be kept at a minimum, while (ideally) maintaining the quality of the result. One way to approach this goal is to embed the data objects in a vector space so that the distances of the embedded objects approximates the actual distances. Thus, queries can be performed (for the most part) on the embedded objects. We are especially interested in examining the issue of whether or not the embedding methods will ensure that no relevant objects are left out. Particular attention is paid to the SparseMap, FastMap, and MetricMap embedding methods. SparseMap is a variant of Lipschitz embeddings, while FastMap and MetricMap are inspired by dimension reduction methods for Euclidean spaces. We show that, in general, none of these embedding methods guarantee that queries on the embedded objects have no false dismissals, while also demonstrating the limited cases in which the guarantee does hold. Moreover, we describe a variant of SparseMap that allows queries with no false dismissals. In addition, we show that with FastMap and MetricMap, the distances of the embedded objects can be much greater than the actual distances. This makes it impossible (or at least impractical) to modify FastMap and MetricMap to guarantee no false dismissals. %B Pattern Analysis and Machine Intelligence, IEEE Transactions on %V 25 %P 530 - 549 %8 2003/05// %@ 0162-8828 %G eng %N 5 %R 10.1109/TPAMI.2003.1195989 %0 Conference Paper %B Data Engineering, 2003. Proceedings. 19th International Conference on %D 2003 %T PXML: a probabilistic semistructured data model and algebra %A Hung,E. %A Getoor, Lise %A V.S. Subrahmanian %K algebra; %K data %K databases; %K instances; %K model; %K models; %K probabilistic %K processing; %K PXML; %K query %K relational %K semistructured %K structures; %K tree %K XML; %X Despite the recent proliferation of work on semistructured data models, there has been little work to date on supporting uncertainty in these models. We propose a model for probabilistic semistructured data (PSD). The advantage of our approach is that it supports a flexible representation that allows the specification of a wide class of distributions over semistructured instances. We provide two semantics for the model and show that the semantics are probabilistically coherent. Next, we develop an extension of the relational algebra to handle probabilistic semistructured data and describe efficient algorithms for answering queries that use this algebra. Finally, we present experimental results showing the efficiency of our algorithms. %B Data Engineering, 2003. Proceedings. 19th International Conference on %P 467 - 478 %8 2003/03// %G eng %R 10.1109/ICDE.2003.1260814 %0 Journal Article %J Topology %D 2003 %T Quantifying the causes of path inflation %A Spring, Neil %A Mahajan,R. %A Anderson,T. %B Topology %V 100 %P 2 - 1 %8 2003/// %G eng %0 Conference Paper %B Proceedings of the fourteenth annual ACM-SIAM symposium on Discrete algorithms %D 2003 %T Random MAX SAT, random MAX CUT, and their phase transitions %A Coppersmith,D. %A Gamarnik,D. %A Hajiaghayi, Mohammad T. %A Sorkin,G. B %B Proceedings of the fourteenth annual ACM-SIAM symposium on Discrete algorithms %P 364 - 373 %8 2003/// %G eng %0 Report %D 2003 %T Recovery of a Digital Image Collection Through the SDSC/UMD/NARA Prototype Persistent Archive %A Smorul,Mike %A JaJa, Joseph F. %A McCall,Fritz %A Brown,Susan Fitch %A Moore,Reagan %A Marciano,Richard %A Chen,Sheau-Yen %A Lopez,Rick %A Chadduck,Robert %K Technical Report %X The San Diego Supercomputer Center (SDSC), the University of Maryland, and theNational Archives and Records Administration (NARA) are collaborating on building a pilot persistent archive using and extending data grid and digital library technologies. The current prototype consists of node servers at SDSC, University of Maryland, and NARA, connected through the Storage Request Broker (SRB) data grid middleware, and currently holds several terabytes of NARA selected collections. In particular, a historically important image collection that was on the verge of becoming inaccessible was fully restored and ingested into our pilot system. In this report, we describe the methodology behind our approach to fully restore this image collection and the process used to ingest it into the prototype persistent archive. (UMIACS-TR-2003-105) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-105 %8 2003/11/25/ %G eng %U http://drum.lib.umd.edu/handle/1903/1321 %0 Journal Article %J Proceedings of the National Academy of SciencesPNAS %D 2003 %T Reduction of Cholera in Bangladeshi Villages by Simple Filtration %A Rita R Colwell %A Huq,Anwar %A M. Sirajul Islam %A K. M. A. Aziz %A Yunus,M. %A N. Huda Khan %A A. Mahmud %A Sack,R. Bradley %A Nair,G. B. %A J. Chakraborty %A Sack,David A. %A E. Russek-Cohen %X Based on results of ecological studies demonstrating that Vibrio cholerae, the etiological agent of epidemic cholera, is commensal to zooplankton, notably copepods, a simple filtration procedure was developed whereby zooplankton, most phytoplankton, and particulates >20 μm were removed from water before use. Effective deployment of this filtration procedure, from September 1999 through July 2002 in 65 villages of rural Bangladesh, of which the total population for the entire study comprised ≈133,000 individuals, yielded a 48% reduction in cholera (P < 0.005) compared with the control. %B Proceedings of the National Academy of SciencesPNAS %V 100 %P 1051 - 1055 %8 2003/02/04/ %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/100/3/1051 %N 3 %R 10.1073/pnas.0237386100 %0 Journal Article %J ACM SIGSOFT Software Engineering Notes %D 2003 %T Regression testing of GUIs %A Memon, Atif M. %A Soffa,Mary Lou %K call-tree %K classification of events %K GUI call-graph %K GUI control-flow graph %K GUI testing %K regression testing %K repairing test cases %X Although graphical user interfaces (GUIs) constitute a large part of the software being developed today and are typically created using rapid prototyping, there are no effective regression testing techniques for GUIs. The needs of GUI regression testing differ from those of traditional software. When the structure of a GUI is modified, test cases from the original GUI are either reusable or unusable on the modified GUI. Since GUI test case generation is expensive, our goal is to make the unusable test cases usable. The idea of reusing these unusable (a.k.a. obsolete) test cases has not been explored before. In this paper, we show that for GUIs, the unusability of a large number of test cases is a serious problem. We present a novel GUI regression testing technique that first automatically determines the usable and unusable test cases from a test suite after a GUI modification. It then determines which of the unusable test cases can be repaired so they can execute on the modified GUI. The last step is to repair the test cases. Our technique is integrated into a GUI testing framework that, given a test case, automatically executes it on the GUI. We implemented our regression testing technique and demonstrate for two case studies that our approach is effective in that many of the test cases can be repaired, and is practical in terms of its time performance. %B ACM SIGSOFT Software Engineering Notes %V 28 %P 118 - 127 %8 2003/09// %@ 0163-5948 %G eng %U http://doi.acm.org/10.1145/949952.940088 %N 5 %R 10.1145/949952.940088 %0 Journal Article %J SERIES ON SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING %D 2003 %T Replicated studies: building a body of knowledge about software reading techniques %A Shull, F. %A Carver, J. %A Travassos,G.H. %A Maldonado,J.C. %A Conradi,R. %A Basili, Victor R. %B SERIES ON SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING %V 12 %P 39 - 84 %8 2003/// %G eng %0 Journal Article %J SIGMETRICS Perform. Eval. Rev. %D 2003 %T Resilient multicast using overlays %A Banerjee,Suman %A Lee,Seungjoon %A Bhattacharjee, Bobby %A Srinivasan, Aravind %K overlay multicast %K randomized forwarding %K Resilience %X We introduce PRM (Probabilistic Resilient Multicast): a multicast data recovery scheme that improves data delivery ratios while maintaining low end-to-end latencies. PRM has both a proactive and a reactive component; in this paper we describe how PRM can be used to improve the performance of application-layer multicast protocols, especially when there are high packet losses and host failures. Further, using analytic techniques, we show that PRM can guarantee arbitrarily high data delivery ratios and low latency bounds. As a detailed case study, we show how PRM can be applied to the NICE application-layer multicast protocol. We present detailed simulations of the PRM-enhanced NICE protocol for 10,000 node Internet-like topologies. Simulations show that PRM achieves a high delivery ratio (> 97%) with a low latency bound (600 ms) for environments with high end-to-end network losses (1-5%) and high topology change rates (5 changes per second) while incurring very low overheads (< 5%). %B SIGMETRICS Perform. Eval. Rev. %V 31 %P 102 - 113 %8 2003/06// %@ 0163-5999 %G eng %U http://doi.acm.org/10.1145/885651.781041 %N 1 %R 10.1145/885651.781041 %0 Conference Paper %B Proceedings of the IASTED NCI %D 2003 %T RGL study in a hybrid real-time system %A Hennacy,K. %A Swamy,N. %A Perlis, Don %B Proceedings of the IASTED NCI %8 2003/// %G eng %0 Conference Paper %B Proceedings of the 22nd international conference on Theory and applications of cryptographic techniques %D 2003 %T Round efficiency of multi-party computation with a dishonest majority %A Katz, Jonathan %A Ostrovsky,Rafail %A Smith,Adam %X We consider the round complexity of multi-party computation in the presence of a static adversary who controls a majority of the parties. Here, n players wish to securely compute some functionality and up to n - 1 of these players may be arbitrarily malicious. Previous protocols for this setting (when a broadcast channel is available) require O(n) rounds. We present two protocols with improved round complexity: The first assumes only the existence of trapdoor permutations and dense cryptosystems, and achieves round complexity O(log n) based on a proof scheduling technique of Chor and Rabin [13]; the second requires a stronger hardness assumption (along with the non-black-box techniques of Barak [2]) and achieves O(1) round complexity. %B Proceedings of the 22nd international conference on Theory and applications of cryptographic techniques %S EUROCRYPT'03 %I Springer-Verlag %C Berlin, Heidelberg %P 578 - 595 %8 2003/// %@ 3-540-14039-5 %G eng %U http://dl.acm.org/citation.cfm?id=1766171.1766222 %0 Conference Paper %B Proceedings of the 4th conference on USENIX Symposium on Internet Technologies and Systems-Volume 4 %D 2003 %T Scriptroute: A public Internet measurement facility %A Spring, Neil %A Wetherall,D. %A Anderson,T. %B Proceedings of the 4th conference on USENIX Symposium on Internet Technologies and Systems-Volume 4 %P 17 - 17 %8 2003/// %G eng %0 Journal Article %J Proceedings of the American Society for Information Science and Technology %D 2003 %T Searching large collections of recorded speech: A preliminary study %A Kim,J. %A Oard, Douglas %A Soergel,D. %B Proceedings of the American Society for Information Science and Technology %V 40 %P 330 - 339 %8 2003/// %G eng %N 1 %0 Conference Paper %B Mobile Data Management %D 2003 %T SEB-tree: An approach to index continuously moving objects %A Song,Z. %A Roussopoulos, Nick %B Mobile Data Management %P 340 - 344 %8 2003/// %G eng %0 Journal Article %J Annals of Mathematics and Artificial Intelligence %D 2003 %T Secure agents %A Bonatti,P. A %A Kraus,S. %A V.S. Subrahmanian %X With the rapid proliferation of software agents, there comes an increased need for agents to ensure that they do not provide data and/or services to unauthorized users. We first develop an abstract definition of what it means for an agent to preserve data/action security. Most often, this requires an agent to have knowledge that is impossible to acquire – hence, we then develop approximate security checks that take into account, the fact that an agent usually has incomplete/approximate beliefs about other agents. We develop two types of security checks – static ones that can be checked prior to deploying the agent, and dynamic ones that are executed at run time. We prove that a number of these problems are undecidable, but under certain conditions, they are decidable and (our definition of) security can be guaranteed. Finally, we propose a language within which the developer of an agent can specify her security needs, and present provably correct algorithms for static/dynamic security verification. %B Annals of Mathematics and Artificial Intelligence %V 37 %P 169 - 235 %8 2003/// %G eng %N 1 %R 10.1023/A:1020233522878 %0 Journal Article %J IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews %D 2003 %T A secure PLAN %A Hicks, Michael W. %A Keromytis,A. D %A Smith,J. M %K active networks %K active-network firewall %K Authentication %K Authorization %K Contracts %K cryptography %K Environmental management %K Extraterrestrial measurements %K functionally restricted packet language %K general-purpose service routines %K Internet %K latency overhead %K namespace-based security %K packet switching %K PLANet %K Planets %K privilege level %K programmable networks %K Safety %K safety risks %K secure PLAN %K security of data %K security risks %K trust management %K two-level architecture %K virtual private network %K Virtual private networks %K Web and internet services %X Active networks, being programmable, promise greater flexibility than current networks. Programmability, however, may introduce safety and security risks. This correspondence describes the design and implementation of a security architecture for the active network PLANet. Security is obtained with a two-level architecture that combines a functionally restricted packet language, PLAN, with an environment of general-purpose service routines governed by trust management. In particular, a technique is used which expands or contracts a packet's service environment based on its level of privilege, termed namespace-based security. The design and implementation of an active-network firewall and virtual private network is used as an application of the security architecture. Measurements of the system show that the addition of the firewall imposes an approximately 34% latency overhead and as little as a 6.7% space overhead to incoming packets. %B IEEE Transactions on Systems, Man, and Cybernetics, Part C: Applications and Reviews %V 33 %P 413 - 426 %8 2003/08// %@ 1094-6977 %G eng %N 3 %R 10.1109/TSMCC.2003.817347 %0 Book Section %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %D 2003 %T Seeing the World Through Image Libraries %A North,C. %A Shneiderman, Ben %A Plaisant, Catherine %A Kang,H. %A Bederson, Benjamin B. %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %I Morgan Kaufmann Publishers Inc. %P 47 - 47 %8 2003/// %@ 978-1-55860-915-0 %G eng %0 Journal Article %J Nucleic acids research %D 2003 %T The sequence and analysis of Trypanosoma brucei chromosome II %A El‐Sayed, Najib M. %A Ghedin,E. %A Song,J. %A MacLeod,A. %A Bringaud,F. %A Larkin,C. %A Wanless,D. %A Peterson,J. %A Hou,L. %A Taylor,S. %A others %B Nucleic acids research %V 31 %P 4856 - 4856 %8 2003/// %G eng %N 16 %0 Journal Article %J DevelopmentDevelopment %D 2003 %T Sex-lethal splicing autoregulation in vivo: interactions between SEX-LETHAL, the U1 snRNP and U2AF underlie male exon skipping %A Nagengast,Alexis A. %A Stitzinger,Shane M. %A Tseng,Chin-Hsiu %A Mount, Stephen M. %A Salz,Helen K. %K Drosophila %K SNF %K Splicing regulation %K SXL %K U1 snRNP %K U2AF %X Alternative splicing of the Sex-lethal pre-mRNA has long served as a model example of a regulated splicing event, yet the mechanism by which the female-specific SEX-LETHAL RNA-binding protein prevents inclusion of the translation-terminating male exon is not understood. Thus far, the only general splicing factor for which there is in vivo evidence for a regulatory role in the pathway leading to male-exon skipping is sans-fille (snf), a protein component of the spliceosomal U1 and U2 snRNPs. Its role, however, has remained enigmatic because of questions about whether SNF acts as part of an intact snRNP or a free protein. We provide evidence that SEX-LETHAL interacts with SANS-FILLE in the context of the U1 snRNP, through the characterization of a point mutation that interferes with both assembly into the U1 snRNP and complex formation with SEX-LETHAL. Moreover, we find that SEX-LETHAL associates with other integral U1 snRNP components, and we provide genetic evidence to support the biological relevance of these physical interactions. Similar genetic and biochemical approaches also link SEX-LETHAL with the heterodimeric splicing factor, U2AF. These studies point specifically to a mechanism by which SEX-LETHAL represses splicing by interacting with these key splicing factors at both ends of the regulated male exon. Moreover, because U2AF and the U1 snRNP are only associated transiently with the pre-mRNA during the course of spliceosome assembly, our studies are difficult to reconcile with the current model that proposes that the SEX-LETHAL blocks splicing at the second catalytic step, and instead argue that the SEX-LETHAL protein acts after splice site recognition, but before catalysis begins. %B DevelopmentDevelopment %V 130 %P 463 - 471 %8 2003/02/01/ %@ 0950-1991, 1477-9129 %G eng %U http://dev.biologists.org/content/130/3/463 %N 3 %R 10.1242/dev.00274 %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2003. IEEE %D 2003 %T A simplified Newton method for the inverse orthotropic problem %A Seydou,F. %A Seppanen,T. %A Duraiswami, Ramani %K algebra; %K electromagnetic %K inhomogeneous %K inverse %K matrix %K matrix; %K media; %K medium; %K method; %K Newton %K orthotropic %K problem; %K problems; %K scattering; %K wave %X We consider one of the most challenging inverse problems in electromagnetic scattering for inhomogeneous medium. We try to reconstruct the elements of a matrix from scattering data in an orthotropic medium. To solve the inverse problem, we implement a simplified Newton method and present some numerical results. %B Antennas and Propagation Society International Symposium, 2003. IEEE %V 1 %P 535 - 538 vol.1 - 535 - 538 vol.1 %8 2003/06// %G eng %R 10.1109/APS.2003.1217514 %0 Conference Paper %B Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on %D 2003 %T Simultaneous pose and correspondence determination using line features %A David,P. %A DeMenthon,D. %A Duraiswami, Ramani %A Samet, Hanan %K algorithm; %K algorithms; %K annealing; %K clutter; %K cluttered %K Computer %K correspondence %K detection; %K determination; %K deterministic %K environment; %K extraction; %K feature %K feature; %K image %K image; %K imagery; %K images; %K joint %K line %K local %K man-made %K MATCHING %K matching; %K measurement; %K model-to-image %K noise; %K occlusion; %K optimum; %K perspective %K point %K pose %K position %K problem; %K processing; %K real %K realistic %K registration %K simulated %K softassign; %K SoftPOSIT %K stereo %K synthetic %K vision; %X We present a new robust line matching algorithm for solving the model-to-image registration problem. Given a model consisting of 3D lines and a cluttered perspective image of this model, the algorithm simultaneously estimates the pose of the model and the correspondences of model lines to image lines. The algorithm combines softassign for determining correspondences and POSIT for determining pose. Integrating these algorithms into a deterministic annealing procedure allows the correspondence and pose to evolve from initially uncertain values to a joint local optimum. This research extends to line features the SoftPOSIT algorithm proposed recently for point features. Lines detected in images are typically more stable than points and are less likely to be produced by clutter and noise, especially in man-made environments. Experiments on synthetic and real imagery with high levels of clutter, occlusion, and noise demonstrate the robustness of the algorithm. %B Computer Vision and Pattern Recognition, 2003. Proceedings. 2003 IEEE Computer Society Conference on %V 2 %P II-424 - II-431 vol.2 - II-424 - II-431 vol.2 %8 2003/06// %G eng %R 10.1109/CVPR.2003.1211499 %0 Report %D 2003 %T Space-Efficient and Fast Algorithms for Multidimensional Dominance Reporting and Range Counting %A Shi,Qingmin %A JaJa, Joseph F. %A Mortensen,Christian %K Technical Report %X We present linear-space sublogarithmic algorithms for handling the {\emthree-dimensional dominance reporting problem} and the {\em two-dimensional range counting problem}. Under the RAM model as described in~[M.~L. Fredman and D.~E. Willard. ``Surpassing the information theoretic bound with fusion trees'', {\em Journal of Computer and System Sciences}, 47:424--436, 1993], our algorithms achieve $O(\log n/\log\log n+f)$ query time for 3-D dominance reporting, where $f$ is the number of points reported, and $O(\log n/\log\log n)$ query time for 2-D range counting case. We extend these results to any constant dimension $d$ achieving $O(n(\log n/\log\log n)^{d-3})$-space and $O((\log n/\log\log )^{d-2}+f)$-query time for the reporting case and $O(n(\log n/\log\log n)^{d-2})$-space and $O((\log n/\log\log n)^{d-1})$ query time for the counting case. (UMIACS-TR-2003-101) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2003-101 %8 2003/11/25/ %G eng %U http://drum.lib.umd.edu/handle/1903/1318 %0 Conference Paper %B Proceedings of the 2003 conference on Interaction design and children %D 2003 %T Starting an intergenerational technology design team: a case study %A Knudtzon,Kendra %A Druin, Allison %A Kaplan,Nancy %A Summers,Kathryn %A Chisik,Yoram %A Kulkarni,Rahul %A Moulthrop,Stuart %A Weeks,Holly %A Bederson, Benjamin B. %K children %K cooperative inquiry %K design methods %K educational applications %K intergenerational design teams %K participatory design %X This paper presents a case study of the first three months of a new intergenerational design team with children ages 10--13. It discusses the research and design methods used for working with children of this age group, the challenges and opportunities of starting a new team, and the lessons learned. %B Proceedings of the 2003 conference on Interaction design and children %S IDC '03 %I ACM %C Preston, England %P 51 - 58 %8 2003/// %@ 1-58113-732-X %G eng %R 10.1145/953536.953545 %0 Journal Article %J Commun. ACM %D 2003 %T Supporting statistical electronic table usage by citizens %A Hert,Carol A. %A Liddy,Elizabeth D. %A Shneiderman, Ben %A Marchionini,Gary %X Over 70 agencies at the federal level are charged with collecting data and producing and disseminating statistics. These statistics are used to inform government policy, shape health care initiatives, provide information on the state of the economy, and others. They also have significant impact on the lives of citizens who use the statistics, for example, to determine job opportunities, changes in social security benefits, and quality of life in particular areas. Our digital government project developed several specific technologies to support the location, manipulation, and understanding of a quintessential format for statistical information---the table. %B Commun. ACM %V 46 %P 52 - 54 %8 2003/01// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/602421.602449 %N 1 %R 10.1145/602421.602449 %0 Journal Article %J Journal of manufacturing systems %D 2003 %T A system for generating process and material selection advice during embodiment design of mechanical components %A Gupta,S.K. %A Chen,Y. %A Feng,S. %A Sriram,R. %X This paper describes a systematic approach to material and process selection during theembodiment design of mechanical components and a system for generating process and material selection advice. Quite often during the embodiment design stage, design requirements are not precisely defined. Therefore, the system described in this paper accounts for imprecision in design requirements during generation and evaluation of alternative process sequences and material options. To reduce the computational effort, the system uses a depth-first branch-and- bound search algorithm. This aids in exploring promising process sequences and material options that can be used to meet the given set of design requirements. Various process sequences and material options are evaluated by using a commercial cost estimation tool. %B Journal of manufacturing systems %V 22 %P 28 - 45 %8 2003/// %G eng %U ftp://ftp.eng.umd.edu/:/home/glue/s/k/skgupta/backup/pub/Publication/JMS03_Gupta.pdf %N 1 %0 Journal Article %J IEEE Data Engineering Bulletin %D 2003 %T TelegraphCQ: An architectural status report %A Krishnamurthy, S. %A Chandrasekaran,S. %A Cooper,O. %A Deshpande, Amol %A Franklin,M.J. %A Hellerstein,J. M %A Hong,W. %A Madden,S. %A Reiss,F. %A Shah,M. A %B IEEE Data Engineering Bulletin %V 26 %P 11 - 18 %8 2003/// %G eng %N 1 %0 Conference Paper %B Proceedings of the 2003 ACM SIGMOD international conference on Management of data %D 2003 %T TelegraphCQ: continuous dataflow processing %A Chandrasekaran,Sirish %A Cooper,Owen %A Deshpande, Amol %A Franklin,Michael J. %A Hellerstein,Joseph M. %A Wei Hong %A Krishnamurthy,Sailesh %A Madden,Samuel R. %A Reiss,Fred %A Shah,Mehul A. %B Proceedings of the 2003 ACM SIGMOD international conference on Management of data %S SIGMOD '03 %I ACM %C New York, NY, USA %P 668 - 668 %8 2003/// %@ 1-58113-634-X %G eng %U http://doi.acm.org/10.1145/872757.872857 %R 10.1145/872757.872857 %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 2003 %T Temporal probabilistic object bases %A Biazzo,V. %A Giugno,R. %A Lukasiewicz,T. %A V.S. Subrahmanian %K algebra; %K algebraic %K bases; %K constraints; %K data %K database %K database; %K databases; %K distribution %K explicit %K functions; %K handling; %K implicit %K instances; %K integrity; %K intervals; %K management; %K model; %K models; %K object %K object-oriented %K operations; %K probabilistic %K probability %K probability; %K relational %K temporal %K theory; %K Uncertainty %K uncertainty; %X There are numerous applications where we have to deal with temporal uncertainty associated with objects. The ability to automatically store and manipulate time, probabilities, and objects is important. We propose a data model and algebra for temporal probabilistic object bases (TPOBs), which allows us to specify the probability with which an event occurs at a given time point. In explicit TPOB-instances, the sets of time points along with their probability intervals are explicitly enumerated. In implicit TPOB-instances, sets of time points are expressed by constraints and their probability intervals by probability distribution functions. Thus, implicit object base instances are succinct representations of explicit ones; they allow for an efficient implementation of algebraic operations, while their explicit counterparts make defining algebraic operations easy. We extend the relational algebra to both explicit and implicit instances and prove that the operations on implicit instances correctly implement their counterpart on explicit instances. %B Knowledge and Data Engineering, IEEE Transactions on %V 15 %P 921 - 939 %8 2003/08//july %@ 1041-4347 %G eng %N 4 %R 10.1109/TKDE.2003.1209009 %0 Journal Article %J The craft of information visualization: readings and reflections %D 2003 %T Theories for Understanding Information Visualization %A Shneiderman, Ben %B The craft of information visualization: readings and reflections %P 349 - 349 %8 2003/// %G eng %0 Conference Paper %B Antennas and Propagation Society International Symposium, 2003. IEEE %D 2003 %T Three dimensional acoustic scattering from an M multilayered domain via an integral equation approach %A Seydou,F. %A Duraiswami, Ramani %A Seppanen,T. %K acoustic %K collocation %K dimensional %K domain; %K equation; %K equations; %K fast %K inhomogeneous %K integral %K matrix %K media; %K method; %K methods; %K multilayered %K multilayers; %K multiple %K multiplication; %K multiplications; %K scattering; %K three %K vector %K wave %X An integral equation approach is derived for three dimensional acoustic scattering from an M multilayered domain. The integral equation is solved using the collocation method. To validate the method, we compare the exact and approximate results. The numerous matrix vector multiplications of our method may be thought a disadvantage. This problem can be overcome by using fast multiple methods where these operations are done very quickly. %B Antennas and Propagation Society International Symposium, 2003. IEEE %V 1 %P 669 - 672 vol.1 - 669 - 672 vol.1 %8 2003/06// %G eng %R 10.1109/APS.2003.1217547 %0 Conference Paper %B Proceedings of the 2003 annual national conference on Digital government research %D 2003 %T Toward a statistical knowledge network %A Marchionini,Gary %A Haas,Stephanie %A Plaisant, Catherine %A Shneiderman, Ben %A Hert,Carol %X This paper describes an architecture for a statistical knowledge network that depends on user interfaces as the glue among the components. These interfaces aim to serve non-expert users with diverse needs and statistical and computing experiences. Such interfaces are crucially dependent on different indexing schemes and good metadata. %B Proceedings of the 2003 annual national conference on Digital government research %S dg.o '03 %I Digital Government Society of North America %P 1 - 6 %8 2003/// %G eng %U http://dl.acm.org/citation.cfm?id=1123196.1123312 %0 Conference Paper %B 2nd International Workshop on Scenarios and State Machines: Models, Algorithms and Tools %D 2003 %T Towards formal but flexible scenarios %A Sengupta,B. %A Cleaveland, Rance %B 2nd International Workshop on Scenarios and State Machines: Models, Algorithms and Tools %8 2003/// %G eng %0 Conference Paper %B Computer Aided Verification %D 2003 %T TRIM: A tool for triggered message sequence charts %A Sengupta,B. %A Cleaveland, Rance %B Computer Aided Verification %P 106 - 109 %8 2003/// %G eng %0 Journal Article %J The Journal of the Acoustical Society of America %D 2003 %T Two methods for solving a 3D acoustic inverse scattering problem %A Seydou,Fadoulourahmane %A Gumerov, Nail A. %A Duraiswami, Ramani %X We consider the problem of finding the refractive index of a buried object by using far‐field measurements in an inhomogeneous medium. We describe two methods for solving the inverse problem. Both methods are implemented in two steps in order to better deal with the ill‐posedness of the problem. In the first method an integral equation of the first kind is derived for the far‐field operator which is solved via least‐squares and Tikhonov regularization. We then use the solution of the integral equation to derive an over‐posed boundary value problem, i.e., the Helmholtz equation in a bounded domain with Cauchy data on the boundary. The index that satisfies this over‐posed problem most closely is obtained via the Levenberg–Marquardt algorithm. The second method is an iterative method and is based on the Lippmann–Schwinger equation. It is implemented via the Newton method. The first step consists, as for the other method. Here we use a Fourier integral approach and regularization via discretization. The second step is to obtain the index by iterating the Lippmann–Schwinger equation starting with the Born approximation. %B The Journal of the Acoustical Society of America %V 113 %P 2191 - 2191 %8 2003/// %G eng %U http://link.aip.org/link/?JAS/113/2191/5 %N 4 %0 Journal Article %J Technical Reports from UMIACS %D 2003 %T Understanding Computer User Frustration: Measuring and Modeling the Disruption from Poor Designs %A Bessiere,Katie %A Ceaparu,Irina %A Lazar,Jonathan %A Robinson,John %A Shneiderman, Ben %K Technical Report %X User frustration in the use of information and computing technology is apervasive and persistent problem. When computers crash, network congestion causes delays, and poor user interfaces trigger confusion there are dramatic consequences for individuals, organizations, and society. These frustrations not only cause personal dissatisfaction and loss of self-efficacy, but may disrupt workplaces, slow learning, and reduce participation in local and national communities. We propose a Computing Frustration Model with incident specific and individual variables to guide research. Our study of 108 computer users shows high levels of frustration and loss of 1/3 to 1/2 of time spent. The importance of the users' goals and the severity of the disruption were correlated with frustration. Those who had strong self-efficacy, indicated by a willingness to stick to solving the problem, reported lower levels of frustration. Implications for users, software developers, managers, and policymakers are discussed. Keywords: user frustration, user interface design, training, helpdesk, computer experience, computer anxiety (UMIACS-TR-2002-89) (HCIL-TR-2002-18) %B Technical Reports from UMIACS %8 2003/01/21/ %G eng %U http://drum.lib.umd.edu/handle/1903/1233 %0 Journal Article %J Technical Reports from UMIACS %D 2003 %T Understanding Hierarchical Clustering Results by Interactive Exploration of Dendrograms: A Case Study with Genomic Microarray Data %A Seo,Jinwook %A Shneiderman, Ben %K Technical Report %X Abstract: Hierarchical clustering is widely used to find patterns inmulti-dimensional datasets, especially for genomic microarray data. Finding groups of genes with similar expression patterns can lead to better understanding of the functions of genes. Early software tools produced only printed results, while newer ones enabled some online exploration. We describe four general techniques that could be used in interactive explorations of clustering algorithms: (1) overview of the entire dataset, coupled with a detail view so that high-level patterns and hot spots can be easily found and examined, (2) dynamic query controls so that users can restrict the number of clusters they view at a time and show those clusters more clearly, (3) coordinated displays: the overview mosaic has a bi-directional link to 2-dimensional scattergrams, (4) cluster comparisons to allow researchers to see how different clustering algorithms group the genes. (UMIACS-TR-2002-50) (HCIL-TR-2002-10) %B Technical Reports from UMIACS %8 2003/01/21/ %G eng %U http://drum.lib.umd.edu/handle/1903/1203 %0 Conference Paper %B Global Telecommunications Conference, 2003. GLOBECOM '03. IEEE %D 2003 %T On the use of flow migration for handling short-term overloads %A Kuo,Kuo-Tung %A Phuvoravan,S. %A Bhattacharjee, Bobby %A Jun La,R. %A Shayman,M. %A Chang,Hyeong Soo %K computing; %K CONGESTION %K congestion; %K CONTROL %K control; %K dynamic %K end-to-end %K engineering %K fast-timescale %K flow %K Internet %K IP %K label %K long-term %K mapping; %K migration; %K MPLS %K multiprotocol %K network %K network; %K networks; %K of %K optimal %K overloads; %K protocol; %K QoS; %K QUALITY %K quality; %K routers; %K routing; %K service; %K set-up %K short-term %K software %K software; %K static %K switching; %K Telecommunication %K telephony; %K time; %K transient %K voice-over-IP; %X In this work, we investigate flow migration as a mechanism to sustain QoS to network users during short-term overloads in the context of an MPLS IP network. We experiment with three different control techniques: static long-term optimal mapping of flows to LSPs; on-line locally optimal mapping of flows to LSPs at flow set-up time; and dynamic flow migration in response to transient congestion. These techniques are applicable over different timescales, have different run-time overheads, and require different levels of monitoring and control software inside the network. We present results both from detailed simulations and a complete implementation using software IP routers. We use voice-over-IP as our test application, and show that if end-to-end quality is to be maintained during short unpredictable bursts of high load, then a fast-timescale control such as migration is required. %B Global Telecommunications Conference, 2003. GLOBECOM '03. IEEE %V 6 %P 3108 - 3112 vol.6 - 3108 - 3112 vol.6 %8 2003/12// %G eng %R 10.1109/GLOCOM.2003.1258807 %0 Journal Article %J Commun. ACM %D 2003 %T Use of the SAND spatial browser for digital government applications %A Samet, Hanan %A Alborzi,Houman %A Brabec,František %A Esperança,Claudio %A Hjaltason,Gísli R. %A Morgan,Frank %A Tanin,Egemen %K GIS %X Numerous federal agencies produce official statistics made accessible to ordinary citizens for searching and data retrieval. This is frequently done via the Internet through a Web browser interface. If this data is presented in textual format, it can often be searched and retrieved by such attributes as topic, responsible agency, keywords, or press release. However, if the data is of spatial nature, for example, in the form of a map, then using text-based queries is often too cumbersome for the intended audience. We describe the use of the SAND Spatial Browser to provide more power to users of these databases by enabling them to define and explore the specific spatial region of interest graphically. The SAND Spatial Browser allows users to form either purely spatial or mixed spatial/nonspatial queries intuitively, which can present information to users that might have been missed if only a textual interface was available. %B Commun. ACM %V 46 %P 61 - 64 %8 2003/01// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/602421.602453 %N 1 %R 10.1145/602421.602453 %0 Journal Article %J Software Engineering, IEEE Transactions on %D 2003 %T User interface evaluation and empirically-based evolution of a prototype experience management tool %A Seaman,C.B %A Mendonca,M.G. %A Basili, Victor R. %A Kim,Y. M %K data %K development; %K empirical %K EMS; %K engineering; %K evaluation; %K experience %K experience; %K Factory; %K interface %K interfaces; %K knowledge %K management %K management; %K model; %K models; %K organization %K performance %K prototype %K Q-Labs %K re-engineering; %K reusability; %K reuse; %K software %K system; %K systems %K tool; %K user %X Experience management refers to the capture, structuring, analysis, synthesis, and reuse of an organization's experience in the form of documents, plans, templates, processes, data, etc. The problem of managing experience effectively is not unique to software development, but the field of software engineering has had a high-level approach to this problem for some time. The Experience Factory is an organizational infrastructure whose goal is to produce, store, and reuse experiences gained in a software development organization. This paper describes The Q-Labs Experience Management System (Q-Labs EMS), which is based on the Experience Factory concept and was developed for use in a multinational software engineering consultancy. A critical aspect of the Q-Labs EMS project is its emphasis on empirical evaluation as a major driver of its development and evolution. The initial prototype requirements were grounded in the organizational needs and vision of Q-Labs, as were the goals and evaluation criteria later used to evaluate the prototype. However, the Q-Labs EMS architecture, data model, and user interface were designed to evolve, based on evolving user needs. This paper describes this approach, including the evaluation that was conducted of the initial prototype and its implications for the further development of systems to support software experience management. %B Software Engineering, IEEE Transactions on %V 29 %P 838 - 850 %8 2003/09// %@ 0098-5589 %G eng %N 9 %R 10.1109/TSE.2003.1232288 %0 Conference Paper %B ACM SIGOPS Operating Systems Review %D 2003 %T User-level internet path diagnosis %A Mahajan,R. %A Spring, Neil %A Wetherall,D. %A Anderson,T. %B ACM SIGOPS Operating Systems Review %V 37 %P 106 - 119 %8 2003/// %G eng %0 Conference Paper %B SPE Annual Technical Conference and Exhibition %D 2003 %T Using Visualization Tools to Gain Insight Into Your Data %A Plaisant, Catherine %A Chintalapani,G. %A Lukehart,C. %A Schiro,D. %A Ryan,J. %X When the term ‘visualization’ is used in the oil and gas industry, it is usually referring to the viewing of complex geologic structures in three-dimensional space. This paper illustrates insights gained by applying interactive visual environments to petroleum industry data that has traditionally been presented in spreadsheets, line graphs, and bar charts. Two information visualization tools, Treemap and SpaceTree, will be shown applied to a variety of oilfield related data. %B SPE Annual Technical Conference and Exhibition %I Society of Petroleum Engineers %8 2003/10// %@ 9781555631529 %G eng %U http://www.onepetro.org/mslib/servlet/onepetropreview?id=00084439 %R 10.2118/84439-MS %0 Conference Paper %B Proceedings of the nineteenth annual symposium on Computational geometry %D 2003 %T VASCO: visualizing and animating spatial constructs and operations %A Brabec,Frantivsek %A Samet, Hanan %A Yilmaz,Cemal %K k-d trees %K Nearest neighbor algorithms %K quadtrees %K R-trees %K Visualization %X A video is used to demonstrate a set of spatial index JAVA applets that enable users on the worldwide web to experiment with a number of variants of the quadtree spatial data structure for different spatial data types, and, most importantly, enable them to see in an animated manner how a number of basic search operations are executed for them. The spatial data types are points, line segments, rectangles, and regions. The search operations are the window query (i.e., a spatial range query) and a nearest neighbor query that enables ranking spatial objects in the order of their distance from a given query object. The representations and algorithms are visualized and animated in a consistent manner using the same primitives and colors so that the differences between the effects of the representations can be easily understood. The video demonstrates the PR quadtree, PM1 quadtree, and R-tree. The applets can be found at: www.cs.umd.edu/~hjs/quadtree/. %B Proceedings of the nineteenth annual symposium on Computational geometry %S SCG '03 %I ACM %C New York, NY, USA %P 374 - 375 %8 2003/// %@ 1-58113-663-3 %G eng %U http://doi.acm.org/10.1145/777792.777849 %R 10.1145/777792.777849 %0 Journal Article %J IEEE Transactions on Information Technology in Biomedicine %D 2003 %T The virtual microscope %A Catalyurek,U. %A Beynon,M. D %A Chang,Chialin %A Kurc, T. %A Sussman, Alan %A Saltz, J. %K biomedical optical imaging %K client software %K client/server architecture %K Computer architecture %K Computer Graphics %K computer platforms %K Computer simulation %K Concurrent computing %K configured data server %K data server software %K database management systems %K database software %K Database systems %K digital slide images %K digital telepathology %K diseases %K emulation %K Environment %K Equipment Design %K Equipment Failure Analysis %K high power light microscope %K Image databases %K Image Enhancement %K Image Interpretation, Computer-Assisted %K Image retrieval %K Information retrieval %K Information Storage and Retrieval %K java %K local disks %K microscope image data %K Microscopy %K multiple clients %K optical microscopy %K PACS %K software %K Software design %K software system %K Software systems %K Systems Integration %K Telepathology %K User-Computer Interface %K virtual microscope design %K Virtual reality %K Workstations %X We present the design and implementation of the virtual microscope, a software system employing a client/server architecture to provide a realistic emulation of a high power light microscope. The system provides a form of completely digital telepathology, allowing simultaneous access to archived digital slide images by multiple clients. The main problem the system targets is storing and processing the extremely large quantities of data required to represent a collection of slides. The virtual microscope client software runs on the end user's PC or workstation, while database software for storing, retrieving and processing the microscope image data runs on a parallel computer or on a set of workstations at one or more potentially remote sites. We have designed and implemented two versions of the data server software. One implementation is a customization of a database system framework that is optimized for a tightly coupled parallel machine with attached local disks. The second implementation is component-based, and has been designed to accommodate access to and processing of data in a distributed, heterogeneous environment. We also have developed caching client software, implemented in Java, to achieve good response time and portability across different computer platforms. The performance results presented show that the Virtual Microscope systems scales well, so that many clients can be adequately serviced by an appropriately configured data server. %B IEEE Transactions on Information Technology in Biomedicine %V 7 %P 230 - 248 %8 2003/12// %@ 1089-7771 %G eng %N 4 %R 10.1109/TITB.2004.823952 %0 Book Section %B The Craft of Information VisualizationThe Craft of Information Visualization %D 2003 %T Visual Information Management for Network Configuration %A Kumar,Harsha %A Plaisant, Catherine %A Teittinen,Marko %A Shneiderman, Ben %E Bederson, Benjamin B. %E Shneiderman, Ben %X Current network management systems rely heavily on forms in their user interfaces. The interfaces reflect the intricacies of the network hardware components but provide little support for guiding users through tasks. There is a scarcity of useful graphical visualizations and decision-support tools. %B The Craft of Information VisualizationThe Craft of Information Visualization %I Morgan Kaufmann %C San Francisco %P 239 - 256 %8 2003/// %@ 978-1-55860-915-0 %G eng %U http://www.sciencedirect.com/science/article/pii/B9781558609150500329 %0 Journal Article %J IEEE Computer Graphics and Applications %D 2003 %T Visualization Viewpoints %A Shneiderman, Ben %B IEEE Computer Graphics and Applications %V 23 %P 12 - 15 %8 2003/// %G eng %N 6 %0 Conference Paper %B Human-Computer Interaction, Interact '03 %D 2003 %T What makes a good answer? The role of context in question answering %A Jimmy Lin %A Quan,D. %A Sinha,V. %A Bakshi,K. %A Huynh,D. %A Katz,B. %A Karger,D. R %B Human-Computer Interaction, Interact '03 %P 25 - 25 %8 2003/// %G eng %0 Journal Article %J Theoretical Computer Science %D 2003 %T When does a random Robin Hood win? %A Gasarch,William %A Golub,Evan %A Srinivasan, Aravind %K Games %K Randomized strategy %K STRATEGY %X A certain two-person infinite game (between “Robin Hood” and the “Sheriff”) has been studied in the context of set theory. In certain cases, it is known that for any deterministic strategy of Robin Hood's, if the Sheriff knows Robin Hood's strategy, he can adapt a winning counter-strategy. We show that in these cases, Robin Hood wins with “probability one” if he adopts a natural random strategy. We then characterize when this random strategy has the almost-surely winning property. We also explore the case of a random Sheriff versus a deterministic Robin Hood. %B Theoretical Computer Science %V 304 %P 477 - 484 %8 2003/07/28/ %@ 0304-3975 %G eng %U http://www.sciencedirect.com/science/article/pii/S0304397503002895 %N 1–3 %R 10.1016/S0304-3975(03)00289-5 %0 Conference Paper %B Proceedings of the 14th IEEE Visualization 2003 (VIS'03) %D 2003 %T Which comes first, usability or utility? %A Grinstein,Georges %A Kobsa,Alfred %A Plaisant, Catherine %A Shneiderman, Ben %A Stasko,John T %B Proceedings of the 14th IEEE Visualization 2003 (VIS'03) %S VIS '03 %I IEEE Computer Society %C Washington, DC, USA %P 112– - 112– %8 2003/// %@ 0-7695-2030-8 %G eng %U http://dx.doi.org/10.1109/VISUAL.2003.1250426 %R 10.1109/VISUAL.2003.1250426 %0 Journal Article %J IEEE Computer Graphics and Applications %D 2003 %T Why not make interfaces better than 3D reality? %A Shneiderman, Ben %K 3D interfaces %K 3D reality %K Atmosphere %K Avatars %K Cities and towns %K Collaboration %K Computer displays %K Computer Graphics %K error prevention %K history keeping %K information exploration %K Information Visualization %K movements %K Navigation %K occlusion %K overviews %K programming-by-demonstration %K rapid situation awareness %K Testing %K usability %K user action feedback %K User interfaces %K Virtual reality %K Visualization %X Many constrained interfaces are designed to be simpler than the real world by restricting movement, limiting interface actions, and keeping interface objects in a plane. However, the strong utility of pure 3D interfaces for medical, architectural, product design, and scientific visualization means that interface design for pure 3D remains an important challenge. An intriguing possibility is that enhanced 3D interfaces might offer simpler navigation, more compelling functionality, safer movements, and less occlusion, than 3D reality, especially for information exploration and visualization tasks. Such features can enable superhuman capabilities such as faster-than-light teleportation, flying through objects, and X-ray vision. Enhanced 3D interfaces might have supernatural tools such as magic wands for instantly shrinking, enlarging, duplicating, or sending objects and enchanted environments that provide error prevention, history keeping, and programming-by-demonstration. Playful game designers and creative application developers have already pushed the technology further than those who seek merely to mimic reality. Advanced designs are marked by their support of rapid situation awareness through effective overviews, reduced numbers of actions to accomplish tasks; and prompt, meaningful feedback for user actions. This article reviews these clever enhanced 3D-design features and encourages approaches that facilitate user tasks rather than mimic reality. %B IEEE Computer Graphics and Applications %V 23 %P 12 - 15 %8 2003/12//Nov %@ 0272-1716 %G eng %N 6 %R 10.1109/MCG.2003.1242376 %0 Book Section %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %D 2003 %T The World's Information in Digital Libraries %A Plaisant, Catherine %A Marchimini,G. %A Bruns,T. %A Komlodi,A. %A Campbell,L. %A Rose,A. %A Ding,G.M. %A Beale Jr,J. %A Nolet,V. %A Shneiderman, Ben %A others %B The craft of information visualization: readings and reflectionsThe craft of information visualization: readings and reflections %I Morgan Kaufmann Publishers Inc. %P 149 - 149 %8 2003/// %@ 978-1-55860-915-0 %G eng %0 Journal Article %J AI Magazine %D 2002 %T AAAI 2002 Workshops %A Blake,Brian %A Haigh,Karen %A Hexmoor,Henry %A Falcone,Rino %A Soh,Leen-Kiat %A Baral,Chitta %A McIlraith,Sheila %A Gmytrasiewicz,Piotr %A Parsons,Simon %A Malaka,Rainer %A Krueger,Antonio %A Bouquet,Paolo %A Smart,Bill %A Kurumantani,Koichi %A Pease,Adam %A Brenner,Michael %A desJardins, Marie %A Junker,Ulrich %A Delgrande,Jim %A Doyle,Jon %A Rossi,Francesca %A Schaub,Torsten %A Gomes,Carla %A Walsh,Toby %A Guo,Haipeng %A Horvitz,Eric J %A Ide,Nancy %A Welty,Chris %A Anger,Frank D %A Guegen,Hans W %A Ligozat,Gerald %B AI Magazine %V 23 %P 113 - 113 %8 2002/12/15/ %@ 0738-4602 %G eng %U http://www.aaai.org/ojs/index.php/aimagazine/article/viewArticle/1678 %N 4 %R 10.1609/aimag.v23i4.1678 %0 Journal Article %J Communications of the ACM %D 2002 %T ACM's computing professionals face new challenges %A Shneiderman, Ben %X Technologists can make a difference on so many critical fronts. %B Communications of the ACM %V 45 %P 31 - 34 %8 2002/02// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/503124.503142 %N 2 %R 10.1145/503124.503142 %0 Conference Paper %B Proceedings of the 2002 ACM/IEEE conference on Supercomputing %D 2002 %T Active Proxy-G: Optimizing the query execution process in the Grid %A Andrade,H. %A Kurc, T. %A Sussman, Alan %A Saltz, J. %B Proceedings of the 2002 ACM/IEEE conference on Supercomputing %P 1 - 15 %8 2002/// %G eng %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2001-45 %D 2002 %T AD (Attacker Defender) Game %A Kochut,Andrzej %A Agrawala, Ashok K. %A Larsen,Ronald L %A A. Udaya Shankar %K Technical Report %X Information Dynamics is a framework for agent-based systems that givesa central position to the role of information, time, and the value of information. We illustrate system design in the Information Dynamics Framework by developing an intelligence game called AD involving attackers, defenders and targets operating in some space of locations. The goal of the attackers is to destroy all targets. Target destruction takes place when the number of attackers in the target's neighborhood exceeds the number of defenders in this neighborhood by a value WINNING_DIFFERENCE. The goal of defenders is to prevent attackers from achieving their goal. (Also UMIACS-TR-2001-45) %B Technical Reports from UMIACS, UMIACS-TR-2001-45 %8 2002/01/31/ %G eng %U http://drum.lib.umd.edu/handle/1903/1142 %0 Conference Paper %D 2002 %T An adaptive framework for tunable consistency and timeliness using replication %A Krishnamurthy, S. %A Sanders,W. H. %A Michel Cukier %K adaptive framework %K concurrent multiple client service %K consistency requirements %K dependability requirements %K distributed object management %K flexible QoS model %K lazy update propagation %K performance history %K probabilistic approach %K Protocols %K quality of service %K replication %K temporal requirements %K timely consistent response %K tunable consistency %K tunable timeliness %X One well-known challenge in using replication to service multiple clients concurrently is that of delivering a timely and consistent response to the clients. In this paper, we address this problem in the context of client applications that have specific temporal and consistency requirements. These applications can tolerate a certain degree of relaxed consistency, in exchange for better response time. We propose a flexible QoS model that allows these clients to specify their temporal and consistency constraints. In order to select replicas to serve these clients, we need to control of the inconsistency of the replicas, so that we have a large enough pool of replicas with the appropriate state to meet a client's timeliness, consistency, and dependability requirements. We describe an adaptive framework that uses lazy update propagation to control the replica inconsistency and employs a probabilistic approach to select replicas dynamically to service a client, based on its QoS specification. The probabilistic approach predicts the ability of a replica to meet a client's QoS specification by using the performance history collected by monitoring the replicas at runtime. We conclude with experimental results based on our implementation. %P 17 - 26 %8 2002/// %G eng %R 10.1109/DSN.2002.1028882 %0 Report %D 2002 %T Addendum to "A Krylov--Schur Algorithm for Large Eigenproblems" %A Stewart, G.W. %K Technical Report %X In this addendum to an earlier paper by the author, it is shown how tocompute a Krylov decomposition corresponding to an arbitrary Rayleigh-Quotient. This decomposition can be used to restart an Arnoldi process, with a selection of the Ritz vectors corresponding to the Rayleigh quotient. (Also UMIACS-TR-2001-90) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2001-90 %8 2002/01/31/ %G eng %U http://drum.lib.umd.edu/handle/1903/1171 %0 Conference Paper %B Proceedings of the first international joint conference on Autonomous agents and multiagent systems: part 2 %D 2002 %T Agents dealing with time and uncertainty %A Dix,Jürgen %A Kraus,Sarit %A V.S. Subrahmanian %K formalisms and logics %K logic programming %K probabilistic/uncertain reasoning %K temporal reasoning %K theories of agency %X Situated agents in the real world need to handle the fact that events occur frequently, as well as the fact that the agent typically has uncertain knowledge about what is true in the world. The ability to reason about both time and uncertainty is therefore very important. In this paper, we develop a formal theory of agents that can reason about both time and uncertainty. The theory extends the notion of agents described in [10, 21] and proposes the notion of temporal probabilistic (or TP) agents. A formal semantics for TP-agents is proposed - this semantics is described via structures called feasible TP-status interpretations (FTPSI's). TP-agents continuously evaluate changes (in the state of the environment they are situated in) and compute appropriate FTPSI's. For a class of TP-agents called positive TP-agents, we develop a provably sound and complete procedure to compute FTPSI's. %B Proceedings of the first international joint conference on Autonomous agents and multiagent systems: part 2 %S AAMAS '02 %I ACM %C New York, NY, USA %P 912 - 919 %8 2002/// %@ 1-58113-480-0 %G eng %U http://doi.acm.org/10.1145/544862.544953 %R 10.1145/544862.544953 %0 Book %D 2002 %T Algorithm Engineering and Experiments: 4th International Workshop, Alenex 2002, San Francisco, Ca, Usa, January 4-5. 2002 : Revised Papers %A Mount, Dave %A Stein,Clifford %K Architecture / General %K Artificial intelligence %K Computer algorithms %K Computer algorithms/ Congresses %K COMPUTERS %K Computers / Computer Graphics %K Computers / Computer Science %K Computers / Data Modeling & Design %K Computers / Data Processing %K Computers / General %K Computers / Programming / Algorithms %K Electronic books %K Mathematics / Discrete Mathematics %K Medical / General %X This book constitutes the thoroughly refereed post-proceedings of the 4th International Workshop on Algorithm Engineering and Experiments, ALENEX 2002, held in San Francisico, CA, USA in January 2002.The 15 revised full papers presented were carefully reviewed and selected from 34 submissions. Among the topics addressed are hewistics for algorithms, combinatorial optimization, searching, graph computation, network optimization, scheduling, computational geometry, sorting, and clustering algorithms. %I Springer %8 2002/09/17/ %@ 9783540439776 %G eng %0 Journal Article %J ACM Comput. Surv. %D 2002 %T Algorithmic issues in modeling motion %A Agarwal,Pankaj K. %A Guibas,Leonidas J. %A Edelsbrunner,Herbert %A Erickson,Jeff %A Isard,Michael %A Har-Peled,Sariel %A Hershberger,John %A Jensen,Christian %A Kavraki,Lydia %A Koehl,Patrice %A Lin,Ming %A Manocha,Dinesh %A Metaxas,Dimitris %A Mirtich,Brian %A Mount, Dave %A Muthukrishnan,S. %A Pai,Dinesh %A Sacks,Elisha %A Snoeyink,Jack %A Suri,Subhash %A Wolefson,Ouri %K computational geometry %K Computer vision %K mobile networks %K modeling %K molecular biology %K motion modeling %K physical simulation %K robotoics %K spatio-temporal databases %X This article is a survey of research areas in which motion plays a pivotal role. The aim of the article is to review current approaches to modeling motion together with related data structures and algorithms, and to summarize the challenges that lie ahead in producing a more unified theory of motion representation that would be useful across several disciplines. %B ACM Comput. Surv. %V 34 %P 550 - 572 %8 2002/12// %@ 0360-0300 %G eng %U http://doi.acm.org/10.1145/592642.592647 %N 4 %R 10.1145/592642.592647 %0 Journal Article %J Multimedia Tools and Applications %D 2002 %T The Analysis and Performance of Multi-Rate Service in Distributed Video-on-Demand Systems %A Mundur, Padma %A Simon,R. %A Sood,A. %B Multimedia Tools and Applications %V 17 %P 51 - 75 %8 2002/// %G eng %N 1 %0 Conference Paper %B Proceedings of the 2002 annual national conference on Digital government research %D 2002 %T APPOINT: an Approach for Peer-to-Peer Offloading the INTernet %A Tanin,Egemen %A Samet, Hanan %X The Internet has recently become the medium of interaction with large volumes of online data. Enterprises in the public and private sectors made their data archives available over the Internet. Working with such large volumes of online data is a challenging task. APPOINT (an Approach for Peer-to-Peer Offloading the INTernet), a centralized peer-to-peer approach that helps users of the Internet transfer large volumes of online data efficiently, is introduced with this paper. In APPOINT, active clients of a client-server architecture act on the server's behalf and communicate with each other to decrease network latency, improve service bandwidth, and resolve server congestions. In essence, a server is enabled to appoint alternatives when needed and create a scalable collaborative virtual mirror from the active clients. Multiple parameters such as locality and availability to decide on how to best forward a download request are considered. The problems that can arise in symmetric upload requests are also addressed. APPOINT is an ongoing project and it is being built as an add-on to existing client-server systems. A library of functions, with a simple application programming interface (API) that can be used within an existing client-server system to improve the service, is currently being developed. %B Proceedings of the 2002 annual national conference on Digital government research %S dg.o '02 %I Digital Government Society of North America %P 1 - 7 %8 2002/// %G eng %U http://dl.acm.org/citation.cfm?id=1123098.1123102 %0 Journal Article %J Random Structures & Algorithms %D 2002 %T Approximation algorithms for the covering Steiner problem %A Konjevod,Goran %A Ravi,R. %A Srinivasan, Aravind %X The covering Steiner problem is a generalization of both the k-MST and the group Steiner problems: given an edge-weighted graph, with subsets of vertices called the groups, and a nonnegative integer value (called the requirement) for each group, the problem is to find a minimum-weight tree spanning at least the required number of vertices of every group. When all requirements are equal to 1, this becomes the group Steiner problem, while if there is only one group which contains all vertices of the graph the problem reduces to k-MST with k equal to the requirement of this unique group. We discuss two different (but equivalent) linear relaxations of the problem for the case when the given graph is a tree and construct polylogarithmic approximation algorithms based on randomized LP rounding of these relaxations. By using a probabilistic approximation of general metrics by tree metrics due to Bartal, our algorithms also solve the covering Steiner problem on general graphs with a further polylogarithmic worsening in the approximation ratio. © 2002 Wiley Periodicals, Inc. Random Struct. Alg., 20:465–483, 2002 %B Random Structures & Algorithms %V 20 %P 465 - 482 %8 2002/05/01/ %@ 1098-2418 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/rsa.10038/abstract?systemMessage=Wiley+Online+Library+will+be+disrupted+17+March+from+10-14+GMT+%2806-10+EDT%29+for+essential+maintenance %N 3 %R 10.1002/rsa.10038 %0 Journal Article %J Peer-to-Peer Systems %D 2002 %T Are virtualized overlay networks too much of a good thing? %A Keleher,P. %A Bhattacharjee, Bobby %A Silaghi,B. %X The majority of recent high-profile work in peer-to-peer networks has approached the problem of location by abstracting over object lookup services. Namespace virtualization in the overlay layer provides load balance and provable bounds on latency at low costs.We contend that namespace virtualization comes at a significant cost for applications that naturally describe their data sets in a hierarchical manner. Opportunities for enhancing browsing, prefetching and efficient attribute-based searches are lost. A hierarchy exposes relationships between items near to each other in the topology; virtualization of the namespace discards this information even if present at client, higherlevel protocols. We advocate encoding application hierarchies directly into the structure of the overlay network, and revisit this argument through a newly proposed distributed directory service. %B Peer-to-Peer Systems %P 225 - 231 %8 2002/// %G eng %0 Book Section %B Flexible Query Answering SystemsFlexible Query Answering Systems %D 2002 %T An Augmented Visual Query Mechanism for Finding Patterns in Time Series Data %A Keogh,Eamonn %A Hochheiser,Harry %A Shneiderman, Ben %E Carbonell,Jaime %E Siekmann,Jörg %E Andreasen,Troels %E Christiansen,Henning %E Motro,Amihai %E Legind Larsen,Henrik %X Relatively few query tools exist for data exploration and pattern identification in time series data sets. In previous work we introduced Timeboxes . Timeboxes are rectangular, direct-manipulation queries for studying time-series datasets. We demonstrated how Timeboxes can be used to support interactive exploration via dynamic queries, along with overviews of query results and drag-and-drop support for query-by-example. In this paper, we extend our work by introducing Variable Time Timeboxes (VTT). VTTs are a natural generalization of Timeboxes, which permit the specification of queries that allow a degree of uncertainty in the time axis. We carefully motivate the need for these more expressive queries, and demonstrate the utility of our approach on several data sets. %B Flexible Query Answering SystemsFlexible Query Answering Systems %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2522 %P 240 - 250 %8 2002/// %@ 978-3-540-00074-7 %G eng %U http://dx.doi.org/10.1007/3-540-36109-X_19 %0 Journal Article %J Proceedings of the ACL-2002 Workshop on Text Summarization %D 2002 %T Automatic headline generation for newspaper stories %A Zajic, David %A Dorr, Bonnie J %A Schwartz,R. %X In this paper we propose a novelapplication of Hidden Markov Models to automatic generation of informative headlines for English texts. We propose four decoding parameters to make the headlines appear more like Headlinese, the language of informative newspaper headlines. We also allow for morphological variation in words between headline and story English. Informal and formal evaluations indicate that our approach produces informative headlines, mimicking a Headlinese style generated by humans. %B Proceedings of the ACL-2002 Workshop on Text Summarization %8 2002/// %G eng %0 Conference Paper %B Proceedings of the The 5th International Conference on Text, Speech and Dialogue %D 2002 %T Automatic transcription of Czech language oral history in the MALACH project: resources and initial experiments %A Oard, Douglas %A Demner-Fushman,D. %A Hajic,J. %A Ramabhadran,B. %A Gustman,S. %A Byrne,WJ %A Soergel,D. %A Dorr, Bonnie J %A Resnik, Philip %A Picheny,M. %B Proceedings of the The 5th International Conference on Text, Speech and Dialogue %P 57 - 64 %8 2002/// %G eng %0 Journal Article %J International Journal for Numerical Methods in Fluids %D 2002 %T Block preconditioners for the discrete incompressible Navier–Stokes equations %A Elman, Howard %A Silvester, David J %A Wathen, Andrew J %K Iterative algorithms %K Navier–Stokes equations %K preconditioning %X We examine the convergence characteristics of iterative methods based on a new preconditioning operator for solving the linear systems arising from discretization and linearization of the steady-state Navier–Stokes equations. For steady-state problems, we show that the preconditioned problem has an eigenvalue distribution consisting of a tightly clustered set together with a small number of outliers. These characteristics are directly correlated with the convergence properties of iterative solvers, with convergence rates independent of mesh size and only mildly dependent on viscosity. For evolutionary problems, we show that implicit treatment of the time derivatives leads to systems for which convergence is essentially independent of viscosity. Copyright © 2002 John Wiley & Sons, Ltd. %B International Journal for Numerical Methods in Fluids %V 40 %P 333 - 344 %8 2002/09/30/ %@ 1097-0363 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/fld.311/abstract %N 3‐4 %R 10.1002/fld.311 %0 Journal Article %J Interactions-New York %D 2002 %T Book Preview-Leonardo's Laptop: Human Needs and the New Computing Technologies %A Shneiderman, Ben %B Interactions-New York %V 9 %P 43 - 43 %8 2002/// %G eng %N 4 %0 Conference Paper %B Proceedings of the 2nd Workshop on Caching, Coherence, and Consistency, held in conjunction with the 16th ACM International Conference on Supercomputing %D 2002 %T On cache replacement policies for servicing mixed data intensive query workloads %A Andrade,H. %A Kurc, T. %A Sussman, Alan %A Borovikov,E. %A Saltz, J. %X When data analysis applications are employed in a multi-client environment, a data server must service multiple si- multaneous queries, each of which may employ complex user-defined data structures and operations on the data. It is then necessary to harness inter- and intra-query com- monalities and system resources to improve the performance of the data server. We have developed a framework and customizable middleware to enable reuse of intermediate and final results among queries, through an in-memory ac- tive semantic cache and user-defined transformation func- tions. Since resources such as processing power and mem- ory space are limited on the machine hosting the server, ef- fective scheduling of incoming queries and efficient cache replacement policies are challenging issues that must be ad- dressed. We have worked on the scheduling problem in ear- lier work, and in this paper we describe and evaluate several cache replacement policies. We present experimental eval- uation of the policies on a shared-memory parallel system using two applications from different application domains. %B Proceedings of the 2nd Workshop on Caching, Coherence, and Consistency, held in conjunction with the 16th ACM International Conference on Supercomputing %8 2002/// %G eng %0 Conference Paper %B CHI '02 extended abstracts on Human factors in computing systems %D 2002 %T CHI@20: fighting our way from marginality to power %A Shneiderman, Ben %A Card,Stuart %A Norman,Donald A. %A Tremaine,Marilyn %A Waldrop,M. Mitchell %K education %K future vision %K Human-computer interaction %K practice %K professional %K THEORY %K usability engineering %X The Special Interest Group on Computer Human Interaction (SIGCHI) has had a successful history of 20 years of growth in its numbers and influence. To help guide the continued evolution of the academic discipline and professional community, we invite several senior members to offer their visions for what the field of CHI actually accomplished over the past several decades, and what do we still need to accomplish? What do we need to do differently/ better/smarter? What haven't we tried because the technology, the money or the will wasn't there in the past, but perhaps is now.The CHI field is more than just technology. We understand that our work can have a profound effect on individuals, families, neighborhoods, corporations, and countries. We know that we can influence education, commerce, healthcare, and government. How can we contribute to bridging the digital divides in developed and developing countries? What agendas can we offer for the academic, research, industrial, and civic spheres for the next 20 years? How can we be more ambitious? How can we truly serve human needs. %B CHI '02 extended abstracts on Human factors in computing systems %S CHI EA '02 %I ACM %C New York, NY, USA %P 688 - 691 %8 2002/// %@ 1-58113-454-1 %G eng %U http://doi.acm.org/10.1145/506443.506548 %R 10.1145/506443.506548 %0 Conference Paper %B IEEE INFOCOM 2002. Twenty-First Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %D 2002 %T Clustering and server selection using passive monitoring %A Andrews,M. %A Shepherd,B. %A Srinivasan, Aravind %A Winkler,P. %A Zane,F. %K client assignment %K client-server systems %K clustering %K content servers %K Delay %K distributed system %K Educational institutions %K Internet %K IP addresses %K Monitoring %K network conditions %K Network servers %K Network topology %K optimal content server %K passive monitoring %K server selection %K Space technology %K TCPIP %K Transport protocols %K Web pages %K Web server %K Webmapper %X We consider the problem of client assignment in a distributed system of content servers. We present a system called Webmapper for clustering IP addresses and assigning each cluster to an optimal content server. The system is passive in that the only information it uses comes from monitoring the TCP connections between the clients and the servers. It is also flexible in that it makes no a priori assumptions about network topology and server placement and it can react quickly to changing network conditions. We present experimental results to evaluate the performance of Webmapper. %B IEEE INFOCOM 2002. Twenty-First Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %I IEEE %V 3 %P 1717- 1725 vol.3 - 1717- 1725 vol.3 %8 2002/// %@ 0-7803-7476-2 %G eng %R 10.1109/INFCOM.2002.1019425 %0 Book Section %B Chip TechnologyChip Technology %D 2002 %T Combinatorial Algorithms for Design of DNA Arrays %A Hannenhalli, Sridhar %A Hubbell,Earl %A Lipshutz,Robert %A Pevzner,Pavel %E Hoheisel,Jörg %E Brazma,A. %E Büssow,K. %E Cantor,C. %E Christians,F. %E Chui,G. %E Diaz,R. %E Drmanac,R. %E Drmanac,S. %E Eickhoff,H. %E Fellenberg,K. %E Hannenhalli, Sridhar %E Hoheisel,J. %E Hou,A. %E Hubbell,E. %E Jin,H. %E Jin,P. %E Jurinke,C. %E Konthur,Z. %E Köster,H. %E Kwon,S. %E Lacy,S. %E Lehrach,H. %E Lipshutz,R. %E Little,D. %E Lueking,A. %E McGall,G. %E Moeur,B. %E Nordhoff,E. %E Nyarsik,L. %E Pevzner,P. %E Robinson,A. %E Sarkans,U. %E Shafto,J. %E Sohail,M. %E Southern,E. %E Swanson,D. %E Ukrainczyk,T. %E van den Boom,D. %E Vilo,J. %E Vingron,M. %E Walter,G. %E Xu,C. %X Optimal design of DNA arrays requires the development of algorithms with two-fold goals: reducing the effects caused by unintended illumination ( border length minimization problem ) and reducing the complexity of masks ( mask decomposition problem ). We describe algorithms that reduce the number of rectangles in mask decomposition by 20–30% as compared to a standard array design under the assumption that the arrangement of oligonucleotides on the array is fixed. This algorithm produces provably optimal solution for all studied real instances of array design. We also address the difficult problem of finding an arrangement which minimizes the border length and come up with a new idea of threading that significantly reduces the border length as compared to standard designs. %B Chip TechnologyChip Technology %S Advances in Biochemical Engineering/Biotechnology %I Springer Berlin / Heidelberg %V 77 %P 1 - 19 %8 2002/// %@ 978-3-540-43215-9 %G eng %U http://dx.doi.org/10.1007/3-540-45713-5_1 %0 Journal Article %J Science %D 2002 %T Comparative Genome Sequencing for Discovery of Novel Polymorphisms in Bacillus Anthracis %A Read,Timothy D. %A Salzberg,Steven L. %A Pop, Mihai %A Shumway,Martin %A Umayam,Lowell %A Jiang,Lingxia %A Holtzapple,Erik %A Busch,Joseph D %A Smith,Kimothy L %A Schupp,James M %A Solomon,Daniel %A Keim,Paul %A Fraser,Claire M. %X Comparison of the whole-genome sequence ofBacillus anthracis isolated from a victim of a recent bioterrorist anthrax attack with a reference reveals 60 new markers that include single nucleotide polymorphisms (SNPs), inserted or deleted sequences, and tandem repeats. Genome comparison detected four high-quality SNPs between the two sequenced B. anthracischromosomes and seven differences among different preparations of the reference genome. These markers have been tested on a collection of anthrax isolates and were found to divide these samples into distinct families. These results demonstrate that genome-based analysis of microbial pathogens will provide a powerful new tool for investigation of infectious disease outbreaks. %B Science %V 296 %P 2028 - 2033 %8 2002/06/14/ %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/296/5575/2028 %N 5575 %R 10.1126/science.1071837 %0 Conference Paper %B Application-Specific Systems, Architectures and Processors, 2002. Proceedings. The IEEE International Conference on %D 2002 %T A component architecture for FPGA-based, DSP system design %A Spivey,G. %A Bhattacharyya, Shuvra S. %A Nakajima,K. %B Application-Specific Systems, Architectures and Processors, 2002. Proceedings. The IEEE International Conference on %P 41 - 51 %8 2002/// %G eng %0 Journal Article %J Kowalski, Part I %D 2002 %T Computational Logic: Logic Programming and Beyond, Essays in Honour of Robert A %A Eiter,T. %A Mascardi,V. %A V.S. Subrahmanian %B Kowalski, Part I %P 586 - 625 %8 2002/// %G eng %0 Conference Paper %D 2002 %T A configurable CORBA gateway for providing adaptable system properties %A Seri,M. %A Courtney,T. %A Michel Cukier %A Gupta,V. %A Krishnmamurthy,S. %A Lyons,J. %A Ramasamy,H. %A Ren,J. %A Sanders,W. H. %8 2002/// %G eng %U https://www.perform.csl.illinois.edu/Papers/USAN_papers/02SER01.pdf %0 Conference Paper %B Pattern Recognition, 2002. Proceedings. 16th International Conference on %D 2002 %T Content-based image retrieval using Fourier descriptors on a logo database %A Folkers,A. %A Samet, Hanan %K abstraction; %K analysis; %K constraints; %K content-based %K contour %K database %K database; %K databases; %K descriptors; %K detection; %K edge %K Fourier %K image %K logos; %K pictorial %K processing; %K query %K retrieval; %K SHAPE %K spatial %K specification; %K theory; %K visual %X A system that enables the pictorial specification of queries in an image database is described. The queries are comprised of rectangle, polygon, ellipse, and B-spline shapes. The queries specify which shapes should appear in the target image as well as spatial constraints on the distance between them and their relative position. The retrieval process makes use of an abstraction of the contour of the shape which is invariant against translation, scale, rotation, and starting point, that is based on the use of Fourier descriptors. These abstractions are used in a system to locate logos in an image database. The utility of this approach is illustrated using some sample queries. %B Pattern Recognition, 2002. Proceedings. 16th International Conference on %V 3 %P 521 - 524 vol.3 - 521 - 524 vol.3 %8 2002/// %G eng %R 10.1109/ICPR.2002.1047991 %0 Conference Paper %B IEEE/RSJ International Conference on Intelligent Robots and Systems, 2002 %D 2002 %T Contour migration: solving object ambiguity with shape-space visual guidance %A Abd-Almageed, Wael %A Smith,C.E. %K Artificial intelligence %K camera motion %K CAMERAS %K Computer vision %K contour migration %K Databases %K edge detection %K Intelligent robots %K Laboratories %K Machine vision %K object ambiguity %K Object recognition %K pattern matching %K Robot vision systems %K servomechanisms %K SHAPE %K shape matching %K shape-space visual guidance %K silhouette matching %K visual servoing %X A fundamental problem in computer vision is the issue of shape ambiguity. Simply stated, a silhouette cannot uniquely identify an object or an object's classification since many unique objects can present identical occluding contours. This problem has no solution in the general case for a monocular vision system. This paper presents a method for disambiguating objects during silhouette matching using a visual servoing system. This method identifies the camera motion(s) that gives disambiguating views of the objects. These motions are identified through a new technique called contour migration. The occluding contour's shape is used to identify objects or object classes that are potential matches for that shape. A contour migration is then determined that disambiguates the possible matches by purposive viewpoint adjustment. The technique is demonstrated using an example set of objects. %B IEEE/RSJ International Conference on Intelligent Robots and Systems, 2002 %I IEEE %V 1 %P 330- 335 vol.1 - 330- 335 vol.1 %8 2002/// %@ 0-7803-7398-7 %G eng %R 10.1109/IRDS.2002.1041410 %0 Journal Article %J Journal of Systems and Software %D 2002 %T COTS-based software development: Processes and open issues %A Morisio,M %A Seaman,C.B %A Basili, Victor R. %A Parra,A.T %A Kraft,S.E %A Condon,S.E %K Commercial off-the-shelf %K Component-based %K COTS %K empirical study %K Software development process %X The work described in this paper is an investigation of the COTS-based software development within a particular NASA environment, with an emphasis on the processes used. Fifteen projects using a COTS-based approach were studied and their actual process was documented. This process is evaluated to identify essential differences in comparison to traditional software development. The main differences, and the activities for which projects require more guidance, are requirements definition and COTS selection, high level design, integration and testing. Starting from these empirical observations, a new process and set of guidelines for COTS-based development are developed and briefly presented. %B Journal of Systems and Software %V 61 %P 189 - 199 %8 2002/04/01/ %@ 0164-1212 %G eng %U http://www.sciencedirect.com/science/article/pii/S0164121201001479 %N 3 %R 10.1016/S0164-1212(01)00147-9 %0 Journal Article %J Commun. ACM %D 2002 %T Creativity support tools %A Shneiderman, Ben %X Establishing a framework of activities for creative work. %B Commun. ACM %V 45 %P 116 - 120 %8 2002/10// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/570907.570945 %N 10 %R 10.1145/570907.570945 %0 Conference Paper %B Proceedings of the 4th conference on Creativity & cognition %D 2002 %T Creativity support tools: a tutorial overview %A Shneiderman, Ben %K consultation %K creativity %K design environments %K empowerment %K innovation %K knowledge work %K models of the user creativity %K software integration %K software tools %X The tutorial begins with a model of creative processes and refines it into design principles. Participants will learn of eight ways to improve software that supports human creative processes: - searching and browsing digital libraries - visualizing data and processes, - consulting with peers and mentors, -thinking by free associations, -exploring solutions: what- if tools, -composing artifacts and performances, -reviewing and replaying session histories, and -disseminating results.These practical suggestions can be integrated into existing software applications, built into web services, or serve as the framework for novel tools. Software examples will be shown and brainstorming sessions will use projects from participants. %B Proceedings of the 4th conference on Creativity & cognition %S C&C '02 %I ACM %C New York, NY, USA %P 1 - 2 %8 2002/// %@ 1-58113-465-7 %G eng %U http://doi.acm.org/10.1145/581710.581711 %R 10.1145/581710.581711 %0 Conference Paper %B The 43rd Annual IEEE Symposium on Foundations of Computer Science, 2002. Proceedings %D 2002 %T Dependent rounding in bipartite graphs %A Gandhi,R. %A Khuller, Samir %A Parthasarathy,S. %A Srinivasan, Aravind %K Application software %K Approximation algorithms %K bipartite graph %K bipartite graphs %K broadcast channels %K broadcast scheduling %K Broadcasting %K capacitated vertex cover %K Character generation %K computational complexity %K Computer science %K Delay %K edge-sets %K Educational institutions %K fair scheduling %K fractional vectors %K graph theory %K per-user fairness properties %K pipage rounding technique %K Processor scheduling %K Random variables %K random-graph models %K randomized rounding approach %K rounding method %K scheduling %K Scheduling algorithm %K telecommunication computing %K unrelated parallel machines %X We combine the pipage rounding technique of Ageev & Sviridenko with a recent rounding method developed by Srinivasan (2001), to develop a new randomized rounding approach for fractional vectors defined on the edge-sets of bipartite graphs. We show various ways of combining this technique with other ideas, leading to the following applications: richer random-graph models for graphs with a given degree-sequence; improved approximation algorithms for: (i) throughput-maximization in broadcast scheduling, (ii) delay-minimization in broadcast scheduling, and (iii) capacitated vertex cover; fair scheduling of jobs on unrelated parallel machines. A useful feature of our method is that it lets us prove certain (probabilistic) per-user fairness properties. %B The 43rd Annual IEEE Symposium on Foundations of Computer Science, 2002. Proceedings %I IEEE %P 323 - 332 %8 2002/// %@ 0-7695-1822-2 %G eng %R 10.1109/SFCS.2002.1181955 %0 Journal Article %J Computational Logic: Logic Programming and Beyond %D 2002 %T Disjunctive logic programming: A survey and assessment %A Minker, Jack %A Seipel,D. %X We describe the fields of disjunctive logic programming and disjunctive deductive databases from the time of their inception to the current time. Contributions with respect to semantics, implementations and applications are surveyed.In the last decade many semantics have been proposed out of which we highlight what we believe to be the most influential ones and compare them. Basic ideas have been borrowed from the semantics of normal logic programs such as stable model semantics and well-founded semantics, which have been generalized in various ways to obtain semantics of disjunctive logic programs. We discuss disjunctive systems such as DLV and Smodels, and related non-disjunctive systems such as XSB and DeReS, that have been implemented. We also describe applications of disjunctive logic programming: reasoning about declarative specifications, reasoning about actions, diagnosis (e.g. in medicine or biology), and in data integration that have resource predicates defined by multiple rules. We discuss the future needs to make the field practical: e.g. integrating concepts from databases (such as aggregation), optimization methods, and object orientation. %B Computational Logic: Logic Programming and Beyond %P 171 - 197 %8 2002/// %G eng %R 10.1007/3-540-45628-7_18 %0 Conference Paper %B Proceedings of the 2002 ACM SIGMOD international conference on Management of data - SIGMOD '02 %D 2002 %T Dwarf %A Sismanis,Yannis %A Deligiannakis,Antonios %A Roussopoulos, Nick %A Kotidis,Yannis %B Proceedings of the 2002 ACM SIGMOD international conference on Management of data - SIGMOD '02 %C Madison, Wisconsin %P 464 - 464 %8 2002/// %G eng %U http://dl.acm.org/citation.cfm?id=564745 %R 10.1145/564691.564745 %0 Conference Paper %B IEEE 2002 Symposia on Human Centric Computing Languages and Environments, 2002. Proceedings %D 2002 %T Dynamic layout management in a multimedia bulletin board %A Kang,Hyunmo %A Shneiderman, Ben %A Wolff,G. J %K Asynchronous communication %K asynchronous communication system %K Collaboration %K data mining %K dynamic layout management %K dynamic layout templates %K groupware %K Human computer interaction %K image %K Information services %K Laboratories %K moving picture %K Multimedia Bulletin Board system %K multimedia computing %K Multimedia databases %K multimedia objects %K multimedia systems %K office document %K Prototypes %K sound %K text %K user interface %K User interfaces %K user-controlled layout strategy %K voice %K Web %K Web pages %X This paper proposes a novel user interface to manage the dynamic layout of multimedia objects in the Multimedia Bulletin Board (MBB) system. The MBB has been designed and implemented as a prototype of an asynchronous communication system that enables rich communication and collaboration among users of multimedia objects such as text, image, moving picture, sound, voice, Web, office document, and other files. The layout properties of the multimedia objects on a board (e.g. x-y position, size, z-order, explicit and implicit links, etc.) show important and useful information on the user dynamics occurring within a board. However, a fixed layout created and edited by multiple users may prevent users from recognizing and identifying other information. This paper resolves this problem with a novel user-controlled layout strategy made visible with dynamic layout templates (DLT). Users can reorganize the objects to extract meaningful information related to time, source, geographic location, or topic. %B IEEE 2002 Symposia on Human Centric Computing Languages and Environments, 2002. Proceedings %I IEEE %P 51 - 53 %8 2002/// %@ 0-7695-1644-0 %G eng %R 10.1109/HCC.2002.1046344 %0 Conference Paper %B CHI '02 extended abstracts on Human factors in computing systems %D 2002 %T A dynamic query interface for finding patterns in time series data %A Hochheiser,Harry %A Shneiderman, Ben %K dynamic queries %K Information Visualization %K time series %X Identification of patterns in time series data sets is a task that arises in a wide variety of application domains. This demonstration presents the timebox model of rectangular regions that specify constraints for dynamic queries over time series data sets, and the TimeSearcher application, which uses timeboxes as the basis of an interactive query tool. %B CHI '02 extended abstracts on Human factors in computing systems %S CHI EA '02 %I ACM %C New York, NY, USA %P 522 - 523 %8 2002/// %@ 1-58113-454-1 %G eng %U http://doi.acm.org/10.1145/506443.506460 %R 10.1145/506443.506460 %0 Journal Article %J Cognitive, Affective, & Behavioral Neuroscience %D 2002 %T Effects of callosal lesions in a model of letter perception %A Shevtsova,N. %A Reggia, James A. %B Cognitive, Affective, & Behavioral Neuroscience %V 2 %P 37 - 37 %8 2002/// %G eng %N 1 %0 Journal Article %J Pattern Analysis and Machine Intelligence, IEEE Transactions on %D 2002 %T An efficient k-means clustering algorithm: analysis and implementation %A Kanungo,T. %A Mount, Dave %A Netanyahu,N. S %A Piatko,C. D %A Silverman,R. %A Wu,A. Y %K algorithm;color %K algorithm;image %K algorithm;kd-tree;mean %K analysis;filtering %K clustering %K clustering; %K compression;data %K distance;covariance %K Lloyd %K matrices;filtering %K quantization;data %K segmentation;k-means %K squared %K structure;data-sensitive %K theory;pattern %X In k-means clustering, we are given a set of n data points in d-dimensional space Rd and an integer k and the problem is to determine a set of k points in Rd, called centers, so as to minimize the mean squared distance from each data point to its nearest center. A popular heuristic for k-means clustering is Lloyd's (1982) algorithm. We present a simple and efficient implementation of Lloyd's k-means clustering algorithm, which we call the filtering algorithm. This algorithm is easy to implement, requiring a kd-tree as the only major data structure. We establish the practical efficiency of the filtering algorithm in two ways. First, we present a data-sensitive analysis of the algorithm's running time, which shows that the algorithm runs faster as the separation between clusters increases. Second, we present a number of empirical studies both on synthetically generated data and on real data sets from applications in color quantization, data compression, and image segmentation %B Pattern Analysis and Machine Intelligence, IEEE Transactions on %V 24 %P 881 - 892 %8 2002/07// %@ 0162-8828 %G eng %N 7 %R 10.1109/TPAMI.2002.1017616 %0 Conference Paper %B Parallel and Distributed Processing Symposium, International %D 2002 %T Efficient Manipulation of Large Datasets on Heterogeneous Storage Systems %A Beynon,Michael D. %A Sussman, Alan %A Kurc,Tahsin %A Catalyurek,Umit %A Saltz,Joel %K component-based frameworks %K data-intensive computing %K load balancing %X In this paper we are concerned with the efficient use of a collection of disk-based storage systems and computing platforms in a heterogeneous setting for retrieving and processing large scientific datasets. We demonstrate, in the context of a data-intensive visualization application, how heterogeneity affects performance and show a set of optimization techniques that can be used to improve performance in a component-based framework. In particular, we examine the application of parallelism via transparent copies of application components in the pipelined processing of data. %B Parallel and Distributed Processing Symposium, International %I IEEE Computer Society %C Los Alamitos, CA, USA %V 2 %P 0084 - 0084 %8 2002/// %@ 0-7695-1573-8 %G eng %R http://doi.ieeecomputersociety.org/10.1109/IPDPS.2002.1015655 %0 Conference Paper %B Scientific and Statistical Database Management, 2002. Proceedings. 14th International Conference on %D 2002 %T Efficient techniques for range search queries on earth science data %A Shi,Qingmin %A JaJa, Joseph F. %K based %K computing; %K content %K data %K data; %K databases; %K Earth %K factors; %K large %K mining %K mining; %K natural %K processing; %K queries; %K query %K range %K raster %K retrieval; %K scale %K Science %K sciences %K search %K spatial %K structures; %K tasks; %K temporal %K tree %K tree-of-regions; %K visual %X We consider the problem of organizing large scale earth science raster data to efficiently handle queries for identifying regions whose parameters fall within certain range values specified by the queries. This problem seems to be critical to enabling basic data mining tasks such as determining associations between physical phenomena and spatial factors, detecting changes and trends, and content based retrieval. We assume that the input is too large to fit in internal memory and hence focus on data structures and algorithms that minimize the I/O bounds. A new data structure, called a tree-of-regions (ToR), is introduced and involves a combination of an R-tree and efficient representation of regions. It is shown that such a data structure enables the handling of range queries in an optimal I/O time, under certain reasonable assumptions. We also show that updates to the ToR can be handled efficiently. Experimental results for a variety of multi-valued earth science data illustrate the fast execution times of a wide range of queries, as predicted by our theoretical analysis. %B Scientific and Statistical Database Management, 2002. Proceedings. 14th International Conference on %P 142 - 151 %8 2002/// %G eng %R 10.1109/SSDM.2002.1029714 %0 Journal Article %J Extreme Programming and Agile Methods—XP/Agile Universe 2002 %D 2002 %T Empirical findings in agile methods %A Lindvall,M. %A Basili, Victor R. %A Boehm,B. %A Costa,P. %A Dangle,K. %A Shull, F. %A Tesoriero,R. %A Williams,L. %A Zelkowitz, Marvin V %X In recent years, the use of, interest in, and controversy about Agile methodologies have realized dramatic growth. Anecdotal evidence is rising regarding the effectiveness of agile methodologies in certain environments and for specified projects. However, collection and analysis of empirical evidence of this effectiveness and classification of appropriate environments for Agile projects has not been conducted. Researchers from four institutions organized an eWorkshop to synchronously and virtually discuss and gather experiences and knowledge from eighteen Agile experts spread across the globe. These experts characterized Agile Methods and communicated experiences using these methods on small to very large teams. They discussed the importance of staffing Agile teams with highly skilled developers. They shared common success factors and identified warning signs of problems in Agile projects. These and other findings and heuristics gathered through this valuable exchange can be useful to researchers and to practitioners as they establish an experience base for better decision making. %B Extreme Programming and Agile Methods—XP/Agile Universe 2002 %P 81 - 92 %8 2002/// %G eng %R 10.1007/3-540-45672-4_19 %0 Book Section %B Computational Logic: Logic Programming and BeyondComputational Logic: Logic Programming and Beyond %D 2002 %T Error-Tolerant Agents %A Eiter,Thomas %A Mascardi,Viviana %A Subrahmanian,V. %E Kakas,Antonis %E Sadri,Fariba %K Computer %K Science %X The use of agents in today’s Internet world is expanding rapidly. Yet, agent developers proceed largely under the optimistic assumption that agents will be error-free. Errors may arise in agents for numerous reasons — agents may share a workspace with other agents or humans and updates made by these other entities may cause an agent to face a situation that it was not explicitly programmed to deal with. Likewise, errors in coding agents may lead to inconsistent situations where it is unclear how the agent should act. In this paper, we define an agent execution model that allows agents to continue acting “reasonably” even when some errors of the above types occur. More importantly, in our framework, agents take “repair” actions automatically when confronted with such situations, but while taking such repair actions, they can often continue to engage in work and/or interactions with other agents that are unaffected by repairs. %B Computational Logic: Logic Programming and BeyondComputational Logic: Logic Programming and Beyond %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2407 %P 83 - 104 %8 2002/// %@ 978-3-540-43959-2 %G eng %U http://dx.doi.org/10.1007/3-540-45628-7_22 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2002-22 %D 2002 %T Evaluation of the SoftPOSIT Model-to-Image Registration Algorithm %A David,P. %A DeMenthon,D. %A Duraiswami, Ramani %A Samet, Hanan %X The problem of pose estimation arises in many areas of computer vision, including object recognition,object tracking, site inspection and updating, and autonomous navigation when scene models are avail- able. We present a new algorithm, called SoftPOSIT, for determining the pose of a 3D object from a single 2D image when correspondences between model points and image points are not known. The algorithm combines Gold’s iterative softassign algorithm [20, 21] for computing correspondences and DeMenthon’s iterative POSIT algorithm [14] for computing object pose under a full-perspective camera model. Our algorithm, unlike most previous algorithms for pose determination, does not have to hypoth- esize small sets of matches and then verify the remaining image points. Instead, all possible matches are treated identically throughout the search for an optimal pose. The performance of the algorithm is extensively evaluated in Monte Carlo simulations on synthetic data under a variety of levels of clutter, occlusion, and image noise. These tests show that the algorithm performs well in a variety of difficult sce- narios, and empirical evidence suggests that the algorithm has an asymptotic run-time complexity that is better than previous methods by a factor of the number of image points. The algorithm is being applied to a number of practical autonomous vehicle navigation problems including the registration of 3D archi- tectural models of a city to images, and the docking of small robots onto larger robots. %B Technical Reports from UMIACS, UMIACS-TR-2002-22 %8 2002/// %G eng %0 Book Section %B Active NetworksActive Networks %D 2002 %T Evolution in Action: Using Active Networking to Evolve Network Support for Mobility %A Seong-Kyu,Song %A Shannon,Stephen %A Hicks, Michael W. %A Nettles,Scott %E Sterbenz,James %E Takada,Osamu %E Tschudin,Christian %E Plattner,Bernhard %X A key early objective of Active Networking (AN) was to support on-the-fly network evolution. Although AN has been used relatively extensively to build application-customized protocols and even whole networking systems, demonstrations of evolution have been limited. This paper examines three AN mechanisms and how they enable evolution: active packets and plug-in extensions, well-known to the AN community, and update extensions, which are novel to AN. We devote our presentation to a series of demonstrations of how each type of evolution can be applied to the problem of adding support for mobility to a network. This represents the most large-scale demonstration of AN evolution to date. These demonstrations show what previous AN research has not: that AN technology can, in fact, support very significant changes to the network, even while the network is operational. %B Active NetworksActive Networks %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2546 %P 146 - 161 %8 2002/// %@ 978-3-540-00223-9 %G eng %U http://dx.doi.org/10.1007/3-540-36199-5_12 %0 Conference Paper %B SC Conference %D 2002 %T Executing Multiple Pipelined Data Analysis Operations in the Grid %A Spencer,Matthew %A Ferreira,Renato %A Beynon,Michael %A Kurc,Tahsin %A Catalyurek,Umit %A Sussman, Alan %A Saltz,Joel %X Processing of data in many data analysis applications can be represented as an acyclic, coarse grain data flow, from data sources to the client. This paper is concerned with scheduling of multiple data analysis operations, each of which is represented as a pipelined chain of processing on data. We define the scheduling problem for effectively placing components onto Grid resources, and propose two scheduling algorithms. Experimental results are presented using a visualization application. %B SC Conference %I IEEE Computer Society %C Los Alamitos, CA, USA %P 54 - 54 %8 2002/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/SC.2002.10015 %0 Journal Article %J Behaviour & Information Technology %D 2002 %T Expandable indexes vs. sequential menus for searching hierarchies on the World Wide Web %A Zaphiris,Panayiotis %A Shneiderman, Ben %A Norman,Kent L %X An experiment is reported that compared expandable indexes providing full menu context with sequential menus providing only partial context. Menu depth was varied using hierarchies of two, three and four levels deep in an asymmetric structure of 457 root level items. Menus were presented on the World Wide Web within a browser. Participants searched for specific targets. Results suggest that reducing the depth of hierarchies improves performance in terms of speed and search efficiency. Surprisingly, expandable indexes resulted in poorer performance with deeper hierarchies than did sequential menus.An experiment is reported that compared expandable indexes providing full menu context with sequential menus providing only partial context. Menu depth was varied using hierarchies of two, three and four levels deep in an asymmetric structure of 457 root level items. Menus were presented on the World Wide Web within a browser. Participants searched for specific targets. Results suggest that reducing the depth of hierarchies improves performance in terms of speed and search efficiency. Surprisingly, expandable indexes resulted in poorer performance with deeper hierarchies than did sequential menus. %B Behaviour & Information Technology %V 21 %P 201 - 207 %8 2002/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/0144929021000009045 %N 3 %R 10.1080/0144929021000009045 %0 Journal Article %J IEEE Software %D 2002 %T The Experience factory organization %A Basili, Victor R. %A Seaman,C. %B IEEE Software %V 19 %P 30 - 31 %8 2002/// %G eng %N 3 %0 Journal Article %J Software, IEEE %D 2002 %T Experience in implementing a learning software organization %A Schneider,K. %A von Hunnius,J.-P. %A Basili, Victor R. %K ACQUISITION %K center;software %K centres;project %K DaimlerChrysler;explicit %K development %K experience %K improvement; %K improvement;automobile %K industry;computer %K knowledge %K management;software %K organization;previous %K process %K processes;software %K projects;software %K reuse;learning %K software %X In an effort to improve software development and acquisition processes and explicitly reuse knowledge from previous software projects, DaimlerChrysler created a software experience center (SEC). The authors report on challenges the company faced in creating the SEC %B Software, IEEE %V 19 %P 46 - 49 %8 2002/06//may %@ 0740-7459 %G eng %N 3 %R 10.1109/MS.2002.1003453 %0 Conference Paper %B Applications of Computer Vision, 2002. (WACV 2002). Proceedings. Sixth IEEE Workshop on %D 2002 %T An experimental evaluation of linear and kernel-based methods for face recognition %A Gupta, H. %A Agrawala, Ashok K. %A Pruthi, T. %A Shekhar, C. %A Chellapa, Rama %K analysis; %K classification; %K component %K discriminant %K Face %K image %K Kernel %K linear %K Machine; %K nearest %K neighbor; %K principal %K recognition; %K Support %K vector %X In this paper we present the results of a comparative study of linear and kernel-based methods for face recognition. The methods used for dimensionality reduction are Principal Component Analysis (PCA), Kernel Principal Component Analysis (KPCA), Linear Discriminant Analysis (LDA) and Kernel Discriminant Analysis (KDA). The methods used for classification are Nearest Neighbor (NN) and Support Vector Machine (SVM). In addition, these classification methods are applied on raw images to gauge the performance of these dimensionality reduction techniques. All experiments have been performed on images from UMIST Face Database. %B Applications of Computer Vision, 2002. (WACV 2002). Proceedings. Sixth IEEE Workshop on %P 13 - 18 %8 2002/// %G eng %R 10.1109/ACV.2002.1182137 %0 Book Section %B Dependable Computing EDCC-4 %D 2002 %T Experimental Evaluation of the Unavailability Induced by a Group Membership Protocol %A Joshi,Kaustubh %A Michel Cukier %A Sanders,William %E Bondavalli,Andrea %E Thevenod-Fosse,Pascale %K Computer science %X Group communication is an important paradigm for building highly available distributed systems. However, group membership operations often require the system to block message traffic, causing system services to become unavailable. This makes it important to quantify the unavailability induced by membership operations. This paper experimentally evaluates the blocking behavior of the group membership protocol of the Ensemble group communication system using a novel global-state-based fault injection technique. In doing so, we demonstrate how a layered distributed protocol such as the Ensemble group membership protocol can be modeled in terms of a state machine abstraction, and show how the resulting global state space can be used to specify fault triggers and define important measures on the system. Using this approach, we evaluate the cost associated with important states of the protocol under varying workload and group size. We also evaluate the sensitivity of the protocol to the occurrence of a second correlated crash failure during its operation. %B Dependable Computing EDCC-4 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2485 %P 644 - 648 %8 2002/// %@ 978-3-540-00012-9 %G eng %U http://www.springerlink.com/content/mnkd4upqr14w2r7e/abstract/ %0 Journal Article %J TRANSACTIONS-AMERICAN SOCIETY OF MECHANICAL ENGINEERS JOURNAL OF MECHANICAL DESIGN %D 2002 %T Extracting articulation models from CAD models of parts with curved surfaces %A Sinha,R. %A Gupta,S.K. %A Paredis,C. J. J. %A Khosla,P. K. %X Degrees of freedom in an assembly are realized by creating mating features that permit relativemotion between parts. In complex assemblies, interactions between individual degrees of freedom may result in a behavior different from the intended behavior. In addition, current methods perform assembly reasoning by approximating curved surfaces as piecewise linear surfaces. Therefore, it is important to be able to: reason about assemblies using exact representations of curved surfaces; verify global motion behavior of parts in the assembly; and create motion simulations of the assembly by examination of the geometry. In this paper, we present a linear algebraic constraint method to automatically construct the space of allowed instantaneous motions of an assembly from the geometry of its constituent parts. Our work builds on previous work on linear contact mechanics and on our previous work on curved surface contact mechanics. We enumerate the conditions under which general curved surfaces can be represented using a finite number of constraints linear in the instantaneous velocities. We compose such constraints to build a space of allowed instantaneous velocities for the assembly. The space is then described as a set-theoretic sum of contact-preserving and contact-breaking motion sub-spaces. Analysis of each subspace provides feedback to the designer, which we demonstrate through the use of an example assembly – a 4-part arm. Finally, the results of the analysis of a 4-bar linkage are compared to those from mechanism theory. %B TRANSACTIONS-AMERICAN SOCIETY OF MECHANICAL ENGINEERS JOURNAL OF MECHANICAL DESIGN %V 124 %P 106 - 114 %8 2002/// %G eng %U http://www-2.cs.cmu.edu/afs/cs.cmu.edu/user/cjp/www/pubs/MechDes02.pdf %N 1 %0 Conference Paper %B Proc. of the IIIrd Indian Conference on Computer Vision, Graphics and Image Processing %D 2002 %T Eye tracking using active deformable models %A Ramadan,S. %A Abd-Almageed, Wael %A Smith,C.E. %B Proc. of the IIIrd Indian Conference on Computer Vision, Graphics and Image Processing %8 2002/// %G eng %0 Conference Paper %D 2002 %T Formal specification and verification of a group membership protocol for an intrusion-tolerant group communication system %A Ramasamy,H. V. %A Michel Cukier %A Sanders,W. H. %K computer network reliability %K distributed processing %K distributed systems %K fault tolerant computing %K formal specification %K formal verification %K group membership protocol %K intrusion-tolerant group communication system %K PROMELA %K Protocols %X We describe a group membership protocol that is part of an intrusion-tolerant group communication system, and present an effort to use formal tools to model and validate our protocol. We describe in detail the most difficult part of the validation exercise, which was the determination of the right level of abstraction of the protocol for formally specifying the protocol. The validation exercise not only formally showed that the protocol satisfies its correctness claims, but also provided information that will help us make the protocol more efficient without violating correctness. %P 9 - 18 %8 2002/12// %G eng %R 10.1109/PRDC.2002.1185613 %0 Journal Article %J Science of Computer Programming %D 2002 %T Generic tools for verifying concurrent systems %A Cleaveland, Rance %A Sims,Steven T. %K Model checking %K Operational semantics %K Process algebra %K Specification %K verification %K Verification tools %X Despite the enormous strides made in automatic verification technology over the past decade and a half, tools such as model checkers remain relatively underused in the development of software. One reason for this is that the bewildering array of specification and verification formalisms complicates the development and adoption by users of relevant tool support. This paper proposes a remedy to this state of affairs in the case of finite-state concurrent systems by describing an approach to developing customizable yet efficient verification tools. %B Science of Computer Programming %V 42 %P 39 - 47 %8 2002/01// %@ 0167-6423 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167642301000338 %N 1 %R 10.1016/S0167-6423(01)00033-8 %0 Journal Article %J Nature %D 2002 %T Genome sequence and comparative analysis of the model rodent malaria parasite Plasmodium yoelii yoelii %A Carlton,Jane M. %A Angiuoli,Samuel V %A Suh,Bernard B. %A Kooij,Taco W. %A Pertea,Mihaela %A Silva,Joana C. %A Ermolaeva,Maria D. %A Allen,Jonathan E %A Jeremy D Selengut %A Koo,Hean L. %A Peterson,Jeremy D. %A Pop, Mihai %A Kosack,Daniel S. %A Shumway,Martin F. %A Bidwell,Shelby L. %A Shallom,Shamira J. %A Aken,Susan E. van %A Riedmuller,Steven B. %A Feldblyum,Tamara V. %A Cho,Jennifer K. %A Quackenbush,John %A Sedegah,Martha %A Shoaibi,Azadeh %A Cummings,Leda M. %A Florens,Laurence %A Yates,John R. %A Raine,J. Dale %A Sinden,Robert E. %A Harris,Michael A. %A Cunningham,Deirdre A. %A Preiser,Peter R. %A Bergman,Lawrence W. %A Vaidya,Akhil B. %A Lin,Leo H. van %A Janse,Chris J. %A Waters,Andrew P. %A Smith,Hamilton O. %A White,Owen R. %A Salzberg,Steven L. %A Venter,J. Craig %A Fraser,Claire M. %A Hoffman,Stephen L. %A Gardner,Malcolm J. %A Carucci,Daniel J. %X Species of malaria parasite that infect rodents have long been used as models for malaria disease research. Here we report the whole-genome shotgun sequence of one species, Plasmodium yoelii yoelii, and comparative studies with the genome of the human malaria parasite Plasmodium falciparum clone 3D7. A synteny map of 2,212 P. y. yoelii contiguous DNA sequences (contigs) aligned to 14 P. falciparum chromosomes reveals marked conservation of gene synteny within the body of each chromosome. Of about 5,300 P. falciparum genes, more than 3,300 P. y. yoelii orthologues of predominantly metabolic function were identified. Over 800 copies of a variant antigen gene located in subtelomeric regions were found. This is the first genome sequence of a model eukaryotic parasite, and it provides insight into the use of such systems in the modelling of Plasmodium biology and disease. %B Nature %V 419 %P 512 - 519 %8 2002/10/03/ %@ 0028-0836 %G eng %U http://www.nature.com/nature/journal/v419/n6906/full/nature01099.html %N 6906 %R 10.1038/nature01099 %0 Magazine Article %D 2002 %T Genome Sequence Assembly: Algorithms and Issues %A Pop, Mihai %A Salzberg,Steven L. %A Shumway,Martin %X Ultimately, genome sequencing seeks to provide an organism's complete DNA sequence. Automation of DNA sequencing allowed scientists to decode entire genomes and gave birth to genomics, the analytic and comparative study of genomes. Although genomes can include billions of nucleotides, the chemical reactions researchers use to decode the DNA are accurate for only about 600 to 700 nucleotides at a time.The DNA reads that sequencing produces must then be assembled into a complete picture of the genome. Errors and certain DNA characteristics complicate assembly. Resolving these problemsentails an additional and costly finishing phase that involves extensive human intervention. Assembly programs can dramatically reduce this cost by taking into account additional informationobtained during finishing. Algorithms that can assemble millions of DNA fragments into gene sequences underlie the current revolution in biotechnology, helping researchers build the growingdatabase of complete genomes. %B Computer %V 35 %P 47 - 54 %8 2002/// %@ 0018-9162 %G eng %N 7 %0 Journal Article %J Proceedings of the National Academy of Sciences %D 2002 %T Genomic profiles of clinical and environmental isolates of Vibrio cholerae O1 in cholera-endemic areas of Bangladesh %A Zo,Y. G. %A Rivera,I. N. G. %A E. Russek-Cohen %A Islam,M. S. %A Siddique,A. K. %A Yunus,M. %A Sack,R. B. %A Huq,A. %A Rita R Colwell %X Diversity, relatedness, and ecological interactions of toxigenic Vibrio cholerae O1 populations in two distinctive habitats, the human intestine and the aquatic environment, were analyzed. Twenty environmental isolates and 42 clinical isolates were selected for study by matching serotype, geographic location of isolation in Bangladesh, and season of isolation. Genetic profiling was done by enterobacterial repetitive intergenic consensus sequence–PCR, optimized for profiling by using the fully sequenced V. cholerae El Tor N16961 genome. Five significant clonal clusters of haplotypes were found from 57 electrophoretic types. Isolates from different areas or habitats intermingled in two of the five significant clusters. Frequencies of haplotypes differed significantly only between the environmental populations (exact test; P < 0.05). Analysis of molecular variance yielded a population genetic structure reflecting the differentiating effects of geographic area, habitat, and sampling time. Although a parameter confounding the latter differences explained 9% of the total molecular variance in the entire population (P < 0.01), the net effect of habitat and time could not be separated because of the small number of environmental isolates included in the study. Five subpopulations from a single area were determined, and from these we were able to estimate a relative differentiating effect of habitat, which was small compared with the effect of temporal change. In conclusion, the resulting population structure supports the hypothesis that spatial and temporal fluctuations in the composition of toxigenic V. cholerae populations in the aquatic environment can cause shifts in the dynamics of the disease. %B Proceedings of the National Academy of Sciences %V 99 %P 12409 - 12409 %8 2002/// %@ 0027-8424, 1091-6490 %G eng %U http://www.pnas.org/content/99/19/12409 %N 19 %R 10.1073/pnas.192426499 %0 Conference Paper %B CHI '02 extended abstracts on Human factors in computing systems %D 2002 %T Getting real about speech: overdue or overhyped? %A James,Frankie %A Lai,Jennifer %A Suhm,Bernhard %A Balentine,Bruce %A Makhoul,John %A Nass,Clifford %A Shneiderman, Ben %K cognitive load %K social responses %K speech interfaces %X Speech has recently made headway towards becoming a more mainstream interface modality. For example, there is an increasing number of call center applications, especially in the airline and banking industries. However, speech still has many properties that cause its use to be problematic, such as its inappropriateness in both very quiet and very noisy environments, and the tendency of speech to increase cognitive load. Concerns about such problems are valid; however, they do not explain why the use of speech is so controversial in the HCI community. This panel would like to address the issues underlying the controversy around speech, by discussing the current state of the art, the reasons it is so difficult to build a good speech interface, and how HCI research can contribute to the development of speech interfaces. %B CHI '02 extended abstracts on Human factors in computing systems %S CHI EA '02 %I ACM %C New York, NY, USA %P 708 - 709 %8 2002/// %@ 1-58113-454-1 %G eng %U http://doi.acm.org/10.1145/506443.506557 %R 10.1145/506443.506557 %0 Conference Paper %B Workshop on Automatic Summarization %D 2002 %T GLEANS: A generator of logical extracts and abstracts for nice summaries %A Daumé, Hal %A Echihabi,A. %A Marcu,D. %A Munteanu,D. %A Soricut,R. %B Workshop on Automatic Summarization %P 9 - 14 %8 2002/// %G eng %0 Conference Paper %B Frontiers in Handwriting Recognition, 2002. Proceedings. Eighth International Workshop on %D 2002 %T Hidden loop recovery for handwriting recognition %A David Doermann %A Intrator,N. %A Rivin,E. %A Steinherz,T. %K analysis; %K character %K contour %K cursive %K detection; %K distance %K edge %K ellipses; %K form %K handwritten %K hidden %K loop %K measurements; %K mutual %K partitioning; %K recognition; %K recovery; %K SHAPE %K shape; %K sophisticated %K strokes; %K symmetric %K truncated %K word %X One significant challenge in the recognition of off-line handwriting is in the interpretation of loop structures. Although this information is readily available in online representation, close proximity of strokes often merges their centers making them difficult to identify. In this paper a novel approach to the recovery of hidden loops in off-line scanned document images is presented. The proposed algorithm seeks blobs that resemble truncated ellipses. We use a sophisticated form analysis method based on mutual distance measurements between the two sides of a symmetric shape. The experimental results are compared with the ground truth of the online representations of each off-line word image. More than 86% percent of the meaningful loops are handled correctly. %B Frontiers in Handwriting Recognition, 2002. Proceedings. Eighth International Workshop on %P 375 - 380 %8 2002/// %G eng %R 10.1109/IWFHR.2002.1030939 %0 Conference Paper %B Automation Congress, 2002 Proceedings of the 5th Biannual World %D 2002 %T Hidden Markov models for silhouette classification %A Abd-Almageed, Wael %A Smith,C. %K Computer vision %K Feature extraction %K Fourier transforms %K hidden Markov models %K HMM %K image classification %K Neural networks %K object classification %K Object recognition %K parameter estimation %K pattern recognition %K Probability distribution %K Shape measurement %K silhouette classification %K Wavelet transforms %X In this paper, a new technique for object classification from silhouettes is presented. Hidden Markov models are used as a classification mechanism. Through a set of experiments, we show the validity of our approach and show its invariance under severe rotation conditions. Also, a comparison with other techniques that use hidden Markov models for object classification from silhouettes is presented. %B Automation Congress, 2002 Proceedings of the 5th Biannual World %I IEEE %V 13 %P 395 - 402 %8 2002/// %@ 1-889335-18-5 %G eng %R 10.1109/WAC.2002.1049575 %0 Conference Paper %B Proceedings of the Participatory Design Conference %D 2002 %T How young can our design partners be %A Farber,A. %A Druin, Allison %A Chipman,G. %A Julian,D. %A Somashekhar,S. %B Proceedings of the Participatory Design Conference %P 272 - 277 %8 2002/// %G eng %0 Journal Article %J IEEE Distributed Systems Online %D 2002 %T IEEE Multimedia: Visions and views: Meeting human needs with new digital imaging technologies %A Shneiderman, Ben %B IEEE Distributed Systems Online %V 3 %8 2002/// %G eng %N 11 %0 Journal Article %J Information Security %D 2002 %T Implementation of chosen-ciphertext attacks against PGP and GnuPG %A Jallad,K. %A Katz, Jonathan %A Schneier,B. %X We recently noted [6] that PGP and other e-mail encryption protocols are, in theory, highly vulnerable to chosen-ciphertext attacks in which the recipient of the e-mail acts as an unwitting “decryption oracle”. We argued further that such attacks are quite feasible and therefore represent a serious concern. Here, we investigate these claims in more detail by attempting to implement the suggested attacks. On one hand, we are able to successfully implement the described attacks against PGP and GnuPG (two widely-used software packages) in a number of different settings. On the other hand, we show that the attacks largely fail when data is compressed before encryption.Interestingly, the attacks are unsuccessful for largely fortuitous reasons; resistance to these attacks does not seem due to any conscious effort made to prevent them. Based on our work, we discuss those instances in which chosen-ciphertext attacks do indeed represent an important threat and hence must be taken into account in order to maintain confidentiality. We also recommend changes in the OpenPGP standard [3] to reduce the effectiveness of our attacks in these settings. %B Information Security %P 90 - 101 %8 2002/// %G eng %R 10.1007/3-540-45811-5_7 %0 Book Section %B Approximation Algorithms for Combinatorial OptimizationApproximation Algorithms for Combinatorial Optimization %D 2002 %T Improved Approximation Algorithms for the Partial Vertex Cover Problem %A Halperin,Eran %A Srinivasan, Aravind %E Jansen,Klaus %E Leonardi,Stefano %E Vazirani,Vijay %X The partial vertex cover problem is a generalization of the vertex cover problem:given an undirected graph G = ( V,E ) and an integer k , we wish to choose a minimum number of vertices such that at least k edges are covered. Just as for vertex cover, 2-approximation algorithms are known for this problem, and it is of interest to see if we can do better than this.The current-best approximation ratio for partial vertex cover, when parameterized by the maximum degree d of G , is (2 − Θ (1/ d )).We improve on this by presenting a $$ łeft( 2 - \Theta łeft( \tfracłn łn d łn d \right) \right) $$ -approximation algorithm for partial vertex cover using semidefinite programming, matching the current-best bound for vertex cover. Our algorithmuses a new rounding technique, which involves a delicate probabilistic analysis. %B Approximation Algorithms for Combinatorial OptimizationApproximation Algorithms for Combinatorial Optimization %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2462 %P 161 - 174 %8 2002/// %@ 978-3-540-44186-1 %G eng %U http://dx.doi.org/10.1007/3-540-45753-4_15 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2002-96 %D 2002 %T Improving Performance of Agents by Activity Partitioning %A Ozcan,F. %A V.S. Subrahmanian %X There is growing interest in software agents that provide a variety of services to humans, other agents, and third party software applications. Some of these agents are engaged in hundreds of activities at any given time point. In such cases, agents may try to examine a set A of activities and leverage commonalities between them in order to reduce their load. We call this activity merging. Unfortunately, in most application domains, activity merging turns out to be NP-complete. Thus, for each application domain, there is an integer k (which varies from domain to domain) such that activity merging can merge up to k activities while satisfying the application's performance expectations. In this paper, we consider the problem of what to do when the set of activities exceeds k. Our approach partitions A into disjoint sets A1 union A2 union ... union An such that each Ai contains at most k activities in it (thus the activities in each Ai can be merged using a merging algorithm). When creating such partitions, we would like to ensure that the activities inside each Ai share a lot of commonality, so that merging yields a lot of savings. In this paper, we propose two optimal algorithms (based on the A* algorithm and the branch and bound paradigm), as well as numerous greedy algorithms to solve the problem. We have implemented these algorithms and conducted detailed experiments. The results point out which algorithms are most appropriate for scaling agent performance. %B Technical Reports from UMIACS, UMIACS-TR-2002-96 %8 2002/// %G eng %0 Conference Paper %B 2002 International Symposium on Technology and Society, 2002. (ISTAS'02) %D 2002 %T Improving Web-based civic information access: a case study of the 50 US states %A Ceaparu,I. %A Shneiderman, Ben %K Computer aided software engineering %K Computer science %K contact information %K Educational institutions %K government data processing %K Guidelines %K home page design features %K information resources %K Laboratories %K Modems %K Navigation %K online help %K privacy %K privacy policies %K search boxes %K Tagging %K Uniform resource locators %K US states %K USA %K User interfaces %K Web sites %K Web-based civic information access %X An analysis of the home pages of all fifty US states reveals great variety in key design features that influence efficacy. Some states had excessively large byte counts that would slow users connected by commonly-used 56 K modems. Many web sites had low numbers of or poorly organized links that would make it hard for citizens to find what they were interested in. Features such as search boxes, privacy policies, online help, or contact information need to be added by several states. Our analysis concludes with ten recommendations and finds many further opportunities for individual states to improve their Websites. However still greater benefits will come through collaboration among the states that would lead to consistency, appropriate tagging, and common tools. %B 2002 International Symposium on Technology and Society, 2002. (ISTAS'02) %I IEEE %P 275 - 282 %8 2002/// %@ 0-7803-7284-0 %G eng %R 10.1109/ISTAS.2002.1013826 %0 Journal Article %J Current microbiology %D 2002 %T In vitro adhesion to human cells by viable but nonculturable Enterococcus faecalis %A Pruzzo,C. %A Tarsi,R. %A Lleò,M. M. %A Signoretto,C. %A Zampini,M. %A Rita R Colwell %A Canepari,P. %X The ability of viable but nonculturable (VBNC) Enterococcus faecalis to adhere to Caco-2 and Girardi heart cultured cells and to urinary tract epithelial cells (ECs) was studied. Enterococci were harvested during the vegetative growth phase (early exponential and stationary), in the VBNC state, and after recovery of the ability to divide. VBNC bacteria maintained their adherence capability but the efficiency of attachment was reduced by about 50 to 70%, depending on the target cell employed. The decrease was transient, since enterococci that regained their culturability showed adherence values similar to those observed for actively growing cells. Analysis of the invasive properties of E. faecalis revealed that the VBNC state caused a decrease in the number of bacteria that entered the cultured HEK cells as a result of the reduction in the number of adhering bacteria. These results highlight the importance of studies of the VBNC phenomenon, with respect to both microbial survival in the environment and the impact on human health. %B Current microbiology %V 45 %P 105 - 110 %8 2002/// %G eng %N 2 %R 10.1007/s00284-001-0089-2 %0 Journal Article %J E-Service: new directions in theory and practice %D 2002 %T In Web we trust: establishing strategic trust among online customers %A Ceaparu,I. %A Demner,D. %A Hung,E. %A Zhao,H. %A Shneiderman, Ben %X Electronic commerce (e-commerce) provides an important chance for established large companies to gain more customers as well as for new small companies to have a good start and rapid growth. However, establishing trust between customers and companies through web interface, a key component of successful e-service, is not as easy as through human-buyer-human-seller interaction. In the past few years, a number of experiments have been conducted in order to determine the factors that influence customers' trust in online businesses. The goal of our experiment was to establish which features that appear on commercial websites are trust-inducing. Our study focused on three independent variables: customer service (limited and extensive), testimonial (self and external) and security features representation (graphics and text). We designed the homepages of eight web sites with all combinations of treatments of the three independent variables. Each of the 52 subjects reviewed all the eight homepages and gave them a relative rank of trustworthiness. After the experiment subjects answered additional e-trust related questions. The experiment results and the survey support the belief that specific customer service commitments, third party testimonials and graphic security representations are important in establishing strategic trust among on-line customers. %B E-Service: new directions in theory and practice %P 90 - 107 %8 2002/// %G eng %0 Conference Paper %B Proceedings of the 2nd ACM SIGCOMM Workshop on Internet measurment %D 2002 %T Inferring link weights using end-to-end measurements %A Mahajan,Ratul %A Spring, Neil %A Wetherall,David %A Anderson,Tom %X We describe a novel constraint-based approach to approximate ISP link weights using only end-to-end measurements. Common routing protocols such as OSPF and IS-IS choose least-cost paths using link weights, so inferred weights provide a simple, concise, and useful model of intradomain routing. Our approach extends router-level ISP maps, which include only connectivity, with link weights that are consistent with routing. Our inferred weights agree well with observed routing: while our inferred weights fully characterize the set of shortest paths between 84--99% of the router-pairs, alternative models based on hop count and latency do so for only 47--81% of the pairs. %B Proceedings of the 2nd ACM SIGCOMM Workshop on Internet measurment %S IMW '02 %I ACM %C New York, NY, USA %P 231 - 236 %8 2002/// %@ 1-58113-603-X %G eng %U http://doi.acm.org/10.1145/637201.637237 %R 10.1145/637201.637237 %0 Journal Article %J Pattern Recognition Letters %D 2002 %T Integration of local and global shape analysis for logo classification %A Neumann, Jan %A Samet, Hanan %A Soffer,Aya %K Image databases %K Logos %K shape recognition %K shape representation %K Symbol recognition %X A comparison is made of global and local methods for the shape analysis of logos in an image database. The qualities of the methods are judged by using the shape signatures to define a similarity metric on the logos. As representatives for the two classes of methods, we use the negative shape method which is based on local shape information and a wavelet-based method which makes use of global information. We apply both methods to images with different kinds of degradations and examine how a particular degradation highlights the strengths and shortcomings of each method. Finally, we use these results to develop a new adaptive weighting scheme which is based on the relative performances of the two methods. This scheme gives rise to a new method that is much more robust with respect to all degradations examined and works by automatically predicting if the negative shape or wavelet method is performing better. %B Pattern Recognition Letters %V 23 %P 1449 - 1457 %8 2002/10// %@ 0167-8655 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167865502001058 %N 12 %R 10.1016/S0167-8655(02)00105-8 %0 Conference Paper %B CHI '02 extended abstracts on Human factors in computing systems %D 2002 %T Interacting with identification technology: can it make us more secure? %A Scholtz,Jean %A Johnson,Jeff %A Shneiderman, Ben %A Hope-Tindall,Peter %A Gosling,Marcus %A Phillips,Jonathon %A Wexelblat,Alan %K Biometrics %K civil liberties %K face recognition %K national id card %K privacy %K Security %B CHI '02 extended abstracts on Human factors in computing systems %S CHI EA '02 %I ACM %C New York, NY, USA %P 564 - 565 %8 2002/// %@ 1-58113-454-1 %G eng %U http://doi.acm.org/10.1145/506443.506484 %R 10.1145/506443.506484 %0 Journal Article %J THIRD INTERNATIONAL NASA WORKSHOP ON PLANNING AND SCHEDULING FOR SPACE %D 2002 %T Interactive, Incremental Scheduling for Virtual Telescopes in Education %A Rathod,Priyang %A desJardins, Marie %A Sansare,Suryakant %X The Telescopes in Education (TIE) project, which began in 1992, provides remote access for students to control large observatory telescopes in real time. TIE began with a single telescope, and with manual evaluation and scheduling of student requests. With the sucess of TIE, 20 or more additional telescopes are expected to come on line. With proportionally more student requests, we anticipate that managing the requests and telescope resources will rapidly become too complex and time-consuming to handle manually. To respond to this problem, the Virtual Telescopes in Education (VTIE) project was begun last year. VTIE will provide networked capabilities for automated proposal preparation and evaluation, scheduling, and data archival. This paper describes the interactive scheduling capability that is under development for VTIE. Accepted observation requests will be incrementally and interactively scheduled by a constraint-based scheduler, and rescheduled as needed in response to dynamically changing weather and telescope conditions. The scheduling system incorporates a novel method for cost-sensitive constraint satisfaction, which will enable modeling of the costs of data gathering and user interaction during schedule repair. %B THIRD INTERNATIONAL NASA WORKSHOP ON PLANNING AND SCHEDULING FOR SPACE %P 27--29 - 27--29 %8 2002/// %G eng %U http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.134.4105 %0 Journal Article %J Computer %D 2002 %T Interactively exploring hierarchical clustering results [gene identification] %A Seo,Jinwook %A Shneiderman, Ben %K algorithmic methods %K arrays %K Bioinformatics %K biological data sets %K biology computing %K Data analysis %K data mining %K data visualisation %K Data visualization %K DNA %K Fluorescence %K gene functions %K gene identification %K gene profiles %K Genetics %K Genomics %K Hierarchical Clustering Explorer %K hierarchical systems %K interactive exploration %K interactive information visualization tool %K interactive systems %K Large screen displays %K meaningful cluster identification %K metrics %K microarray data analysis %K pattern clustering %K pattern extraction %K Process control %K Sensor arrays %K sequenced genomes %K Tiles %X To date, work in microarrays, sequenced genomes and bioinformatics has focused largely on algorithmic methods for processing and manipulating vast biological data sets. Future improvements will likely provide users with guidance in selecting the most appropriate algorithms and metrics for identifying meaningful clusters-interesting patterns in large data sets, such as groups of genes with similar profiles. Hierarchical clustering has been shown to be effective in microarray data analysis for identifying genes with similar profiles and thus possibly with similar functions. Users also need an efficient visualization tool, however, to facilitate pattern extraction from microarray data sets. The Hierarchical Clustering Explorer integrates four interactive features to provide information visualization techniques that allow users to control the processes and interact with the results. Thus, hybrid approaches that combine powerful algorithms with interactive visualization tools will join the strengths of fast processors with the detailed understanding of domain experts %B Computer %V 35 %P 80 - 86 %8 2002/07// %@ 0018-9162 %G eng %N 7 %R 10.1109/MC.2002.1016905 %0 Journal Article %J Intel Corporation, Pittsburgh IRPTR-02 %D 2002 %T Irisnet: An architecture for compute-intensive wide-area sensor network services %A Nath,S. %A Deshpande, Amol %A Ke,Y. %A Gibbons,P. %A Karp,B. %A Seshan,S. %X Previous work on sensor networks has targeted adhoc wireless networks of closely-colocated, resource- constrained scalar sensor motes. Such work has over- looked richer sensor types such as webcams and mi- crophones, which are typically attached to Internet- connected machines with significant computing power and storage. In this paper, we describe IrisNet (Internet- scale Resource-Intensive Sensor Network services), an architecture for wide-area sensor networks based on these more capable sensing nodes. IrisNet provides a common, scalable software infrastructure that enables the flexible creation of sensor-based Internet services. It dramatically reduces network bandwidth utilization through the use of senselets, binary code fragments that perform intensive data filtering at the sensing nodes, leveraging the available processing power and mem- ory. IrisNet employs a two-tier hierarchy of sensing nodes and query processing nodes. Key features of IrisNet include flexible data partitioning, efficient and protected sharing of sensor nodes, low-latency queries, partial match caching, query-specified freshness toler- ances, and monitoring and logging support. This pa- per reports on experiments with a working IrisNet pro- totype running a parking space finder service, demon- strating the effectiveness of IrisNet’s features in achiev- ing scalability and reducing query response times. %B Intel Corporation, Pittsburgh IRPTR-02 %V 10 %8 2002/// %G eng %0 Conference Paper %B Engineering of Complex Computer Systems, 2002. Proceedings. Eighth IEEE International Conference on %D 2002 %T A light-weight process for capturing and evolving defect reduction experience %A Basili, Victor R. %A Lindvall,M. %A Shull, F. %K bases;experience %K CeBASE;NSF %K defect %K development;software %K experience;software %K funded %K knowledge;software %K management;lightweight %K management;software %K project;decision %K projects;software %K reduction %K reusability;software %K reuse;software %K support;eWorkshops;experience %K tools; %K tools;knowledge %X Selecting technologies for developing software is a crucial activity in software projects. Defect reduction is an example of an area in which software developers have to decide what technologies to use. CeBASE is a NSF funded project that has the role of improving software development by providing decision support on the selection of techniques and tools. The decision support is based on empirical data organized in experience bases and refined into high-level models. Empirical data is collected through various activities, for example through eWorkshops in which experts discuss important issues, and formalized using the lightweight knowledge dust to knowledge pearl process. %B Engineering of Complex Computer Systems, 2002. Proceedings. Eighth IEEE International Conference on %P 129 - 132 %8 2002/12// %G eng %R 10.1109/ICECCS.2002.1181505 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2002-97 %D 2002 %T A Linear Iterative Approach for Hierarchical Shortest Path Finding %A Samet, Hanan %A Filho,Gutemberg Guerra %X We present a hierarchical approach that subdivides a network with $n$ vertices into $r$ regions with the same number $m$ of vertices ($n = r m$) and creates higher levels merging a constant number $c$ of adjacent regions. We propose linear iterative algorithms to find a shortest path and to expand this path into the lowest level. Since our approach is non-recursive, the complexity constants are small and the algorithms are more efficient in practice than other recursive optimal approaches. A hybrid shortest path algorithm to perform intra-regional queries in the lowest level is introduced. This strategy uses a subsequence of vertices that belong to the shortest path while actually computing the whole shortest path. The hybrid algorithm requires $O(m)$ time and space assuming an uniform distribution of vertices. This represents a further improvement concerning space, since a path view approach requires $O(m^{1.5})$ space in the lowest level. For higher $k$-levels, a path view approach spends $O(1)$ time and requires $O(c^k m)$ space. %B Technical Reports from UMIACS, UMIACS-TR-2002-97 %8 2002/// %G eng %0 Conference Paper %B Proceedings of the eighteenth annual symposium on Computational geometry %D 2002 %T A local search approximation algorithm for k-means clustering %A Kanungo,Tapas %A Mount, Dave %A Netanyahu,Nathan S. %A Piatko,Christine D. %A Silverman,Ruth %A Wu,Angela Y. %K Approximation algorithms %K clustering %K computational geometry %K k-means %K local search %X In k-means clustering we are given a set of n data points in d-dimensional space Rd and an integer k, and the problem is to determine a set of k points in ÓC;d, called centers, to minimize the mean squared distance from each data point to its nearest center. No exact polynomial-time algorithms are known for this problem. Although asymptotically efficient approximation algorithms exist, these algorithms are not practical due to the extremely high constant factors involved. There are many heuristics that are used in practice, but we know of no bounds on their performance.We consider the question of whether there exists a simple and practical approximation algorithm for k-means clustering. We present a local improvement heuristic based on swapping centers in and out. We prove that this yields a (9+&egr;)-approximation algorithm. We show that the approximation factor is almost tight, by giving an example for which the algorithm achieves an approximation factor of (9-&egr;). To establish the practical value of the heuristic, we present an empirical study that shows that, when combined with Lloyd's algorithm, this heuristic performs quite well in practice. %B Proceedings of the eighteenth annual symposium on Computational geometry %S SCG '02 %I ACM %C New York, NY, USA %P 10 - 18 %8 2002/// %@ 1-58113-504-1 %G eng %U http://doi.acm.org/10.1145/513400.513402 %R 10.1145/513400.513402 %0 Conference Paper %B ACM SIGCOMM Computer Communication Review %D 2002 %T Measuring ISP topologies with Rocketfuel %A Spring, Neil %A Mahajan,R. %A Wetherall,D. %B ACM SIGCOMM Computer Communication Review %V 32 %P 133 - 145 %8 2002/// %G eng %0 Journal Article %J Technical Reports of the Computer Science Department %D 2002 %T Media Conversion from Visual to Audio: Voice Browsers %A Youssef,Adel %A Shneiderman, Ben %K Technical Report %X There is a large amount of information on the World Wide Web that is atthe fingertips of anyone with access to the internet. However, so far this information has primarily been used by people who connect to the web via traditional computer. This is about to change. Recent advances in wireless communication, speech recognition, and speech synthesis technologies have made it possible to access this information from any place, and at any time. In this paper, we discuss voice browsers as compared to current web browsers. Some of the primary techniques of universal accessible design are listed with their relation to voice browsers and some ideas are offered to help authors implementing these considerations when designing web pages. The new voice markup language is briefly discussed. %B Technical Reports of the Computer Science Department %8 2002/12/19/ %G eng %U http://drum.lib.umd.edu/handle/1903/534 %0 Journal Article %J IEEE Multimedia %D 2002 %T Meeting human needs with new digital imaging technologies %A Shneiderman, Ben %K animation %K animations %K Computer applications %K digital images %K digital imaging technologies %K Helium %K History %K HUMANS %K Image databases %K IMAGE PROCESSING %K Photography %K photos %K software libraries %K Videos %K visual media %K Web pages %X It's a visual world. Images capture many people's thrills, emotions, and concerns. Art can shock or inspire, and family photos are some of our greatest treasures. It's not surprising that visual information is a vital component of the new computing. Most people depend on visual input to understand the world around them and as a basis for further creative activities. The popularity of visual media such as photos, short videos, and animations attests to their mainstream acceptance. Digital photos and image-laden Web pages have already become major computing applications, but users still want higher-resolution images and faster downloads %B IEEE Multimedia %V 9 %P 8 - 14 %8 2002/12//Oct %@ 1070-986X %G eng %N 4 %R 10.1109/MMUL.2002.1041942 %0 Conference Paper %B 16th International Conference on Pattern Recognition, 2002. Proceedings %D 2002 %T Mixture models for dynamic statistical pressure snakes %A Abd-Almageed, Wael %A Smith,C.E. %K active contour models %K Active contours %K Artificial intelligence %K Bayes methods %K Bayesian methods %K Bayesian theory %K complex colored object %K Computer vision %K decision making %K decision making mechanism %K dynamic statistical pressure snakes %K Equations %K expectation maximization algorithm %K Gaussian distribution %K image colour analysis %K Image edge detection %K Image segmentation %K Intelligent robots %K mixture models %K mixture of Gaussians %K mixture pressure model %K Robot vision systems %K robust pressure model %K Robustness %K segmentation results %K statistical analysis %K statistical modeling %X This paper introduces a new approach to statistical pressure snakes. It uses statistical modeling for both object and background to obtain a more robust pressure model. The Expectation Maximization (EM) algorithm is used to model the data into a Mixture of Gaussians (MoG). Bayesian theory is then employed as a decision making mechanism. Experimental results using the traditional pressure model and the new mixture pressure model demonstrate the effectiveness of the new models. %B 16th International Conference on Pattern Recognition, 2002. Proceedings %I IEEE %V 2 %P 721- 724 vol.2 - 721- 724 vol.2 %8 2002/// %@ 0-7695-1695-X %G eng %R 10.1109/ICPR.2002.1048404 %0 Conference Paper %B 2nd IEEE/ACM International Symposium on Cluster Computing and the Grid, 2002 %D 2002 %T Multiple Query Optimization for Data Analysis Applications on Clusters of SMPs %A Andrade,H. %A Kurc, T. %A Sussman, Alan %A Saltz, J. %K Aggregates %K Application software %K Bandwidth %K Data analysis %K Data structures %K Delay %K Query processing %K scheduling %K Subcontracting %K Switched-mode power supply %X This paper is concerned with the efficient execution of multiple query workloads on a cluster of SMPs. We target applications that access and manipulate large scientific datasets. Queries in these applications involve user-defined processing operations and distributed data structures to hold intermediate and final results. Our goal is to implement system components to leverage previously computed query results and to effectively utilize processing power and aggregated I/O bandwidth on SMP nodes so that both single queries and multi-query batches can be efficiently executed. %B 2nd IEEE/ACM International Symposium on Cluster Computing and the Grid, 2002 %I IEEE %P 154 - 154 %8 2002/05/21/24 %@ 0-7695-1582-7 %G eng %R 10.1109/CCGRID.2002.1017123 %0 Conference Paper %B Shape Modeling International, 2002. Proceedings %D 2002 %T Multiresolution tetrahedral meshes: an analysis and a comparison %A Danovaro,E. %A De Floriani, Leila %A Lee,M. %A Samet, Hanan %K (computer %K bisection;vertex %K cells;tetrahedron %K data %K generation;rendering %K geometry;data %K graphics);solid %K meshes;rendering;tetrahedral %K modelling; %K refinement;multiresolution %K representations;multiresolution %K sets;mesh %K split;computational %K structures;data %K tetrahedral %K visualisation;mesh %K visualization;experiment;large-size %K volume %X We deal with the problem of analyzing and visualizing large-size volume data sets. To this aim, we consider multiresolution representations based on a decomposition of the field domain into tetrahedral cells. We compare two types of multiresolution representations that differ on the rule applied to refine an initial coarse mesh: one is based on tetrahedron bisection, and one based on vertex split. The two representations can be viewed as instances of a common multiresolution model, that we call a multiresolution mesh. Encoding data structures for the two representations are briefly described. An experimental comparison on structured volume data sets is presented %B Shape Modeling International, 2002. Proceedings %P 83 - 91 %8 2002/// %G eng %R 10.1109/SMI.2002.1003532 %0 Journal Article %J Proc. of the 1st Intl. Symp. on 3D Data Processing, Visualization, and Transmission %D 2002 %T Octree approximation and compression methods %A Samet, Hanan %A Kochut, A. %X Techniques are presented to progressively approximateand compress in a lossless manner two-colored (i.e. bi- nary) 3D objects (as well as objects of arbitrary dimen- sionality). The objects are represented by a region octree implemented using a pointerless representation based on locational codes. Approximation is achieved through the use of a forest. This method labels the internal nodes of the octree as GB or GW, depending on the number of chil- dren being of type GB or GW. In addition, all BLACK nodes are labeled GB, while all WHITE nodes are labeled GW. A number of different image approximation methods are dis- cussed that make use of a forest. The advantage of these methods is that they are progressive which means that as more of the object is transmitted, the better is the approx- imation. This makes these methods attractive for use on the worldwide web. Progressive transmission has the draw- back that there is an overhead in requiring extra storage. A progressive forest-based approximation and transmission method is presented where the total amount of data that is transmitted is not larger than MIN(B,W), where B and W are the number of BLACK and WHITE blocks, respectively, in the region octree of the set of objects. %B Proc. of the 1st Intl. Symp. on 3D Data Processing, Visualization, and Transmission %8 2002/// %G eng %0 Journal Article %J Future Generation Computer Systems %D 2002 %T Optimizing execution of component-based applications using group instances %A Beynon,Michael D. %A Kurc,Tahsin %A Sussman, Alan %A Saltz,Joel %K DataCutter %K Grid %K Wide-area network %X Recent research on programming models for developing applications on the Grid has proposed component-based models as a viable approach, in which an application is composed of multiple interacting computational objects. We have been developing a framework, called filter-stream programming, for building data-intensive applications that query, analyze and manipulate very large datasets in a distributed environment. In this model, the processing structure of an application is represented as a set of processing units, referred to as filters. In this paper, we develop the problem of scheduling instances of a filter group. A filter group is a set of filters collectively performing a computation for an application. In particular, we seek the answer to the following question: should a new instance be created, or an existing one reused? We experimentally investigate the effects on performance of instantiating multiple filter groups under varying application characteristics. %B Future Generation Computer Systems %V 18 %P 435 - 448 %8 2002/03// %@ 0167-739X %G eng %U http://www.sciencedirect.com/science/article/pii/S0167739X0100070X %N 4 %R 10.1016/S0167-739X(01)00070-X %0 Journal Article %J ACM Transactions on Graphics (TOG) %D 2002 %T Ordered and quantum treemaps: Making effective use of 2D space to display hierarchies %A Bederson, Benjamin B. %A Shneiderman, Ben %A Wattenberg,Martin %K hierarchies %K Human-computer interaction %K image browsers %K Information Visualization %K jazz %K ordered treemaps %K treemaps %K TREES %K zoomable user interfaces (ZUIs). %X Treemaps, a space-filling method for visualizing large hierarchical data sets, are receiving increasing attention. Several algorithms have been previously proposed to create more useful displays by controlling the aspect ratios of the rectangles that make up a treemap. While these algorithms do improve visibility of small items in a single layout, they introduce instability over time in the display of dynamically changing data, fail to preserve order of the underlying data, and create layouts that are difficult to visually search. In addition, continuous treemap algorithms are not suitable for displaying fixed-sized objects within them, such as images.This paper introduces a new "strip" treemap algorithm which addresses these shortcomings, and analyzes other "pivot" algorithms we recently developed showing the trade-offs between them. These ordered treemap algorithms ensure that items near each other in the given order will be near each other in the treemap layout. Using experimental evidence from Monte Carlo trials and from actual stock market data, we show that, compared to other layout algorithms, ordered treemaps are more stable, while maintaining relatively favorable aspect ratios of the constituent rectangles. A user study with 20 participants clarifies the human performance benefits of the new algorithms. Finally, we present quantum treemap algorithms, which modify the layout of the continuous treemap algorithms to generate rectangles that are integral multiples of an input object size. The quantum treemap algorithm has been applied to PhotoMesa, an application that supports browsing of large numbers of images. %B ACM Transactions on Graphics (TOG) %V 21 %P 833 - 854 %8 2002/10// %@ 0730-0301 %G eng %U http://doi.acm.org/10.1145/571647.571649 %N 4 %R 10.1145/571647.571649 %0 Conference Paper %B Proceedings of the Working Conference on Advanced Visual Interfaces %D 2002 %T OZONE: A zoomable interface for navigating ontology information %A Suh,B. %A Bederson, Benjamin B. %B Proceedings of the Working Conference on Advanced Visual Interfaces %P 139 - 143 %8 2002/// %G eng %0 Conference Paper %B 2002 IEEE Symposium on Security and Privacy, 2002. Proceedings %D 2002 %T P5 : a protocol for scalable anonymous communication %A Sherwood,R. %A Bhattacharjee, Bobby %A Srinivasan, Aravind %K Broadcasting %K communication efficiency %K Computer science %K cryptography %K data privacy %K Educational institutions %K Internet %K large anonymous groups %K P5 protocol %K packet-level simulations %K Particle measurements %K Peer to peer computing %K peer-to-peer personal privacy protocol %K privacy %K Protocols %K receiver anonymity %K scalable anonymous communication %K security of data %K sender anonymity %K sender-receiver anonymity %K Size measurement %K telecommunication security %X We present a protocol for anonymous communication over the Internet. Our protocol, called P5 (peer-to-peer personal privacy protocol) provides sender-, receiver-, and sender-receiver anonymity. P5 is designed to be implemented over current Internet protocols, and does not require any special infrastructure support. A novel feature of P5 is that it allows individual participants to trade-off degree of anonymity for communication efficiency, and hence can be used to scalably implement large anonymous groups. We present a description of P5, an analysis of its anonymity and communication efficiency, and evaluate its performance using detailed packet-level simulations. %B 2002 IEEE Symposium on Security and Privacy, 2002. Proceedings %I IEEE %P 58 - 70 %8 2002/// %@ 0-7695-1543-6 %G eng %R 10.1109/SECPRI.2002.1004362 %0 Report %D 2002 %T A Parallel Block Multi-level Preconditioner for the 3D Incompressible Navier--Stokes Equations %A Elman, Howard %A Howle, V. E %A Shadid,John %A Tuminaro,Ray %K Technical Report %X The development of robust and efficient algorithms for both steady-statesimulations and fully-implicit time integration of the Navier--Stokes equations is an active research topic. To be effective, the linear subproblems generated by these methods require solution techniques that exhibit robust and rapid convergence. In particular, they should be insensitive to parameters in the problem such as mesh size, time step, and Reynolds number. In this context, we explore a parallel preconditioner based on a block factorization of the coefficient matrix generated in an Oseen nonlinear iteration for the primitive variable formulation of the system. The key to this preconditioner is the approximation of a certain Schur complement operator by a technique first proposed by Kay, Loghin, and Wathen [25] and Silvester, Elman, Kay, and Wathen [45]. The resulting operator entails subsidiary computations (solutions of pressure Poisson and convection--diffusion subproblems) that are similar to those required for decoupled solution methods; however, in this case these solutions are applied as preconditioners to the coupled Oseen system. One important aspect of this approach is that the convection--diffusion and Poisson subproblems are significantly easier to solve than the entire coupled system, and a solver can be built using tools developed for the subproblems. In this paper, we apply smoothed aggregation algebraic multigrid to both subproblems. Previous work has focused on demonstrating the optimality of these preconditioners with respect to mesh size on serial, two-dimensional, steady-state computations employing geometric multi-grid methods; we focus on extending these methods to large-scale, parallel, three-dimensional, transient and steady-state simulations employing algebraic multigrid (AMG) methods. Our results display nearly optimal convergence rates for steady-state solutions as well as for transient solutions over a wide range of CFL numbers on the two-dimensional and three-dimensional lid-driven cavity problem. Also UMIACS-TR-2002-95 %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2002-95 %8 2002/10/25/ %G eng %U http://drum.lib.umd.edu//handle/1903/1239 %0 Journal Article %J 4th International Workshop on Algorithm Engineering and Experiments %D 2002 %T Partitioning planar graphs with costs and weights %A Mount, Dave %A Stein,C. %X A graph separator is a set of vertices or edges whose removal divides an input graph into components of bounded size. This paper describes new algorithms for computing separators in planar graphs as well as techniques that can be used to speed up their implementation and improve the partition quality. In particular, we consider planar graphs with costs and weights on the vertices, where weights are used to estimate the sizes of the components and costs are used to estimate the size of the separator. We show that one can find a small separator that divides the graph into components of bounded size. We describe implementations of the partitioning algorithms and discuss results of our experiments. %B 4th International Workshop on Algorithm Engineering and Experiments %V 2409 %P 98 - 110 %8 2002/// %G eng %0 Conference Paper %D 2002 %T Passive replication schemes in AQuA %A Ren,Yansong %A Rubel,P. %A Seri,M. %A Michel Cukier %A Sanders,W. H. %A Courtney,T. %K AQuA architecture %K distributed object management %K Fault tolerance %K group members %K large-scale distributed object-oriented systems %K management structure %K multidimensional quality of service %K passive replication scheme %K performance measurements %K reusable technologies %K scalability %K software fault tolerance %K software performance evaluation %K software reusability %K software solutions %X Building large-scale distributed object-oriented systems that provide multidimensional quality of service (QoS) in terms of fault tolerance, scalability, and performance is challenging. In order to meet this challenge, we need an architecture that can ensure that applications' requirements can be met while providing reusable technologies and software solutions. This paper describes techniques, based on the AQuA architecture, that enhance the applications' dependability and scalability by introducing two types of group members and a novel passive replication scheme. In addition, we describe how to make the management structure itself dependable by using the passive replication scheme. Finally, we provide performance measurements for the passive replication scheme. %P 125 - 130 %8 2002/12// %G eng %R 10.1109/PRDC.2002.1185628 %0 Journal Article %J Numerische Mathematik %D 2002 %T Performance and analysis of saddle point preconditioners for the discrete steady-state Navier-Stokes equations %A Elman, Howard %A Silvester, D. J %A Wathen, A. J %B Numerische Mathematik %V 90 %P 665 - 688 %8 2002/// %G eng %N 4 %0 Conference Paper %D 2002 %T Performance evaluation of a probabilistic replica selection algorithm %A Krishnamurthy, S. %A Sanders,W. H. %A Michel Cukier %K client-server systems %K Dependability %K distributed object management %K dynamic selection algorithm %K Middleware %K probabilistic model %K probabilistic model-based replica selection algorithm %K probability %K quality of service %K real-time systems %K replica failures %K round-robin selection scheme %K static scheme %K time-sensitive distributed applications %K timeliness %K timing failures %K transient overload %X When executing time-sensitive distributed applications, a middleware that provides dependability and timeliness is faced with the important problem of preventing timing failures both under normal conditions and when the quality of service is degraded due to replica failures and transient overload on the server. To address this problem, we have designed a probabilistic model-based replica selection algorithm that allows a middleware to choose a set of replicas to service a client based on their ability to meet a client's timeliness requirements. This selection is done based on the prediction made by a probabilistic model that uses the performance history of replicas as inputs. In this paper, we describe the experiments we have conducted to evaluate the ability of this dynamic selection algorithm to meet a client's timing requirements, and compare it with that of a static and round-robin selection scheme under different scenarios %P 119 - 127 %8 2002/// %G eng %R 10.1109/WORDS.2002.1000044 %0 Conference Paper %D 2002 %T Performance evaluation of a QoS-aware framework for providing tunable consistency and timeliness %A Krishnamurthy, S. %A Sanders,W. H. %A Michel Cukier %K client-server systems %K Computer networks %K CORBA-based middleware %K distributed applications %K distributed object management %K Network servers %K QoS %K quality of service %K replica consistency %K replicated services %K server replicas %K timeliness %X Strong replica consistency models ensure that the data delivered by a replica always includes the latest updates, although this may result in poor response times. On the other hand, weak replica consistency models provide quicker access to information, but do not usually provide guarantees about the degree of staleness in the data they deliver. In order to support emerging distributed applications that are characterized by high concurrency demands, an increasing shift towards dynamic content, and timely delivery, we need quality of service models that allow us to explore the intermediate space between these two extreme approaches to replica consistency. Further, for better support of time-sensitive applications that can tolerate relaxed consistency in exchange for better responsiveness, we need to understand how the desired level of consistency affects the timeliness of a response. The QoS model we have developed to realize these objectives considers both timeliness and consistency, and treats consistency along two dimensions: order and staleness. We evaluate experimentally the framework we have developed to study the timeliness/consistency tradeoffs for replicated services and present experimental results that compare these tradeoffs in the context of sequential and FIFO ordering. %P 214 - 223 %8 2002/// %G eng %R 10.1109/IWQoS.2002.1006589 %0 Conference Paper %B Proceedings of the 6th Workshop on Languages, Compilers, and Run-Time Systems for Scalable Computers %D 2002 %T Persistent caching in a multiple query optimization framework %A Andrade,H. %A Kurc, T. %A Catalyurek,U. %A Sussman, Alan %A Saltz, J. %X This paper focuses on persistent caching in multi-client en-vironments, which aims to improve the performance of a data server by caching on disk query results that can be expensive to produce. We present and evaluate extensions to an existing multi-query optimization framework, called MQO, to incorporate persistent caching capabilities. %B Proceedings of the 6th Workshop on Languages, Compilers, and Run-Time Systems for Scalable Computers %8 2002/// %G eng %0 Journal Article %J Magnetics, IEEE Transactions on %D 2002 %T Perturbation technique for LLG dynamics in uniformly magnetized bodies subject to RF fields %A Bertotti,G. %A Mayergoyz, Issak D %A Serpico,C. %K analytical %K anisotropy %K anisotropy; %K applied %K bodies; %K circularly %K component; %K constant-in-time %K constant; %K damping %K demagnetisation; %K demagnetizing %K differential %K dynamics; %K effective %K elliptically %K equation; %K equations; %K exactly %K factors; %K field; %K film; %K films; %K Frequency %K gyromagnetic %K harmonic; %K higher %K Landau-Lifshitz-Gilbert %K large %K linear %K magnetic %K magnetisation; %K magnetization %K magnetization; %K magnetized %K modes; %K MOTION %K order %K partial %K particle; %K particles; %K perturbation %K polarized %K radio %K ratio; %K RF %K saturation %K solution; %K solvable %K system; %K technique; %K techniques; %K thin %K time-harmonic %K uniaxial %K uniformly %X The problem of magnetization dynamics of a uniformly magnetized uniaxial particle or film, under elliptically polarized applied field, is considered. In the special case of circularly polarized applied field and particles (films) with a symmetry axis, pure time-harmonic magnetization modes exist that can be computed analytically. Deviations from these highly symmetric conditions are treated as perturbation of the symmetric case. The perturbation technique leads to the exactly solvable system of linear differential equations for the perturbations which enables one to compute higher order magnetization harmonic. The analytical solutions are obtained and then compared with numerical results. %B Magnetics, IEEE Transactions on %V 38 %P 2403 - 2405 %8 2002/09// %@ 0018-9464 %G eng %N 5 %R 10.1109/TMAG.2002.803596 %0 Journal Article %J interactions %D 2002 %T A photo history of SIGCHI: evolution of design from personal to public %A Shneiderman, Ben %A Kang,Hyunmo %A Kules,Bill %A Plaisant, Catherine %A Rose,Anne %A Rucheir,Richesh %X For 20 years I have been photographing personalities and events in the emerging discipline of human--computer interaction. Until now, only a few of these photos were published in newsletters or were shown to visitors who sought them out. Now this photo history is going from a personal record to a public archive. This archive should be interesting for professional members of this community who want to reminisce, as well as for historians and journalists who want to understand what happened. Students and Web surfers may also want to look at the people who created better interfaces and more satisfying user experiences. %B interactions %V 9 %P 17 - 23 %8 2002/05// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/506671.506682 %N 3 %R 10.1145/506671.506682 %0 Conference Paper %B Proceedings: SIGCHI %D 2002 %T Physical Programming: Designing Tools for Children to Create Physical Interactive %A Montemayor,J. %A Druin, Allison %A Farber,A. %A Simms,S. %A Churaman,W. %B Proceedings: SIGCHI %8 2002/// %G eng %0 Journal Article %J Artificial life %D 2002 %T Predicting nearest agent distances in artificial worlds %A Schulz,R. %A Reggia, James A. %B Artificial life %V 8 %P 247 - 264 %8 2002/// %G eng %N 3 %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 2002 %T Presentation planning for distributed VoD systems %A Hwang,Eenjun %A Prabhakaran,B. %A V.S. Subrahmanian %K Computer %K computing; %K databases; %K demand; %K distributed %K local %K multimedia %K network; %K on %K optimal %K plan; %K plans; %K presentation %K presentation; %K server; %K servers; %K video %K video-on-demand; %K VoD; %X A distributed video-on-demand (VoD) system is one where a collection of video data is located at dispersed sites across a computer network. In a single site environment, a local video server retrieves video data from its local storage device. However, in distributed VoD systems, when a customer requests a movie from the local server, the server may need to interact with other servers located across the network. In this paper, we present different types of presentation plans that a local server can construct in order to satisfy a customer request. Informally speaking, a presentation plan is a temporally synchronized sequence of steps that the local server must perform in order to present the requested movie to the customer. This involves obtaining commitments from other video servers, obtaining commitments from the network service provider, as well as making commitments of local resources, while keeping within the limitations of available bandwidth, available buffer, and customer data consumption rates. Furthermore, in order to evaluate the quality of a presentation plan, we introduce two measures of optimality for presentation plans: minimizing wait time for a customer and minimizing access bandwidth which, informally speaking, specifies how much network/disk bandwidth is used. We develop algorithms to compute three different optimal presentation plans that work at a block level, or at a segment level, or with a hybrid mix of the two, and compare their performance through simulation experiments. We have also mathematically proven effects of increased buffer or bandwidth and data replications for presentation plans which had previously been verified experimentally in the literature. %B Knowledge and Data Engineering, IEEE Transactions on %V 14 %P 1059 - 1077 %8 2002/10//sep %@ 1041-4347 %G eng %N 5 %R 10.1109/TKDE.2002.1033774 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2002-30 %D 2002 %T A Probabilistic Clustering-Based Indoor Location Determination System %A Youssef,Moustafa A %A Agrawala, Ashok K. %A A. Udaya Shankar %A Noh,Sam H %K Technical Report %X We present an indoor location determination system based on signalstrength probability distributions for tackling the noisy wireless channel and clustering to reduce computation requirements. We provide two implementation techniques, namely, Joint Clustering and Incremental Triangulation and describe their tradeoffs in terms of location determination accuracy and computation requirement. Both techniques have been incorporated in two implemented context-aware systems: User Positioning System and the Rover System, both running on Compaq iPAQ Pocket PC's with Familiar distribution of Linux for PDA's. The results obtained show that both techniques give the user location with over 90% accuracy to within 7 feet with very low computation requirements, hence enabling a set of context-aware applications. Also UMIACS-TR-2002-30 %B Technical Reports from UMIACS, UMIACS-TR-2002-30 %8 2002/04/04/ %G eng %U http://drum.lib.umd.edu/handle/1903/1192 %0 Book Section %B Computer Vision — ECCV 2002Computer Vision — ECCV 2002 %D 2002 %T Probabilistic Human Recognition from Video %A Zhou,Shaohua %A Chellapa, Rama %E Heyden,Anders %E Sparr,Gunnar %E Nielsen,Mads %E Johansen,Peter %X This paper presents a method for incorporating temporal information in a video sequence for the task of human recognition. A time series state space model, parameterized by a tracking state vector and a recognizing identity variable , is proposed to simultaneously characterize the kinematics and identity. Two sequential importance sampling (SIS) methods, a brute-force version and an efficient version, are developed to provide numerical solutions to the model. The joint distribution of both state vector and identity variable is estimated at each time instant and then propagated to the next time instant. Marginalization over the state vector yields a robust estimate of the posterior distribution of the identity variable. Due to the propagation of identity and kinematics, a degeneracy in posterior probability of the identity variable is achieved to give improved recognition. This evolving behavior is characterized using changes in entropy . The effectiveness of this approach is illustrated using experimental results on low resolution face data and upper body data. %B Computer Vision — ECCV 2002Computer Vision — ECCV 2002 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2352 %P 173 - 183 %8 2002/// %@ 978-3-540-43746-8 %G eng %U http://dx.doi.org/10.1007/3-540-47977-5_45 %0 Conference Paper %D 2002 %T Probabilistic validation of intrusion tolerance %A Sanders,W. H. %A Michel Cukier %A Webber,F. %A Pal,P. %A Watro,R. %P 78 - 79 %8 2002/// %G eng %U https://www.perform.csl.illinois.edu/Papers/USAN_papers/02SAN02.pdf %0 Journal Article %J Parallel Computing %D 2002 %T Processing large-scale multi-dimensional data in parallel and distributed environments %A Beynon,Michael %A Chang,Chialin %A Catalyurek,Umit %A Kurc,Tahsin %A Sussman, Alan %A Andrade,Henrique %A Ferreira,Renato %A Saltz,Joel %K Data-intensive applications %K Distributed computing %K Multi-dimensional datasets %K PARALLEL PROCESSING %K Runtime systems %X Analysis of data is an important step in understanding and solving a scientific problem. Analysis involves extracting the data of interest from all the available raw data in a dataset and processing it into a data product. However, in many areas of science and engineering, a scientist's ability to analyze information is increasingly becoming hindered by dataset sizes. The vast amount of data in scientific datasets makes it a difficult task to efficiently access the data of interest, and manage potentially heterogeneous system resources to process the data. Subsetting and aggregation are common operations executed in a wide range of data-intensive applications. We argue that common runtime and programming support can be developed for applications that query and manipulate large datasets. This paper presents a compendium of frameworks and methods we have developed to support efficient execution of subsetting and aggregation operations in applications that query and manipulate large, multi-dimensional datasets in parallel and distributed computing environments. %B Parallel Computing %V 28 %P 827 - 859 %8 2002/05// %@ 0167-8191 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167819102000972 %N 5 %R 10.1016/S0167-8191(02)00097-2 %0 Journal Article %J ACM SIGCAPH Computers and the Physically Handicapped %D 2002 %T Promoting universal usability with multi-layer interface design %A Shneiderman, Ben %K first-time user %K Graphical user interfaces %K multi-layer interface %K novice user %K online help %K universal usability %X Increased interest in universal usability is causing some researchers to study advanced strategies for satisfying first-time as well as intermittent and expert users. This paper promotes the idea of multi-layer interface designs that enable first-time and novice users to begin with a limited set of features at layer 1. They can remain at layer 1, then move up to higher layers when needed or when they have time to learn further features. The arguments for and against multi-layer interfaces are presented with two example systems: a word processor with 8 layers and an interactive map with 3 layers. New research methods and directions are proposed. %B ACM SIGCAPH Computers and the Physically Handicapped %P 1 - 8 %8 2002/06// %@ 0163-5727 %G eng %U http://doi.acm.org/10.1145/960201.957206 %N 73-74 %R 10.1145/960201.957206 %0 Journal Article %J European Journal of Biochemistry %D 2002 %T Purification and properties of the extracellular lipase, LipA, of Acinetobacter sp. RAG‐1 %A Snellman,Erick A. %A Sullivan,Elise R. %A Rita R Colwell %K Acinetobacter sp. RAG‐1 %K LipA %K lipase %K protein purification %K zymogram %X An extracellular lipase, LipA, extracted from Acinetobacter sp. RAG-1 grown on hexadecane was purified and properties of the enzyme investigated. The enzyme is released into the growth medium during the transition to stationary phase. The lipase was harvested from cells grown to stationary phase, and purified with 22% yield and > 10-fold purification. The protein demonstrates little affinity for anion exchange resins, with contaminating proteins removed by passing crude supernatants over a Mono Q column. The lipase was bound to a butyl Sepharose column and eluted in a Triton X-100 gradient. The molecular mass (33 kDa) was determined employing SDS/PAGE. LipA was found to be stable at pH 5.8–9.0, with optimal activity at 9.0. The lipase remained active at temperatures up to 70 °C, with maximal activity observed at 55 °C. LipA is active against a wide range of fatty acid esters of p-nitrophenyl, but preferentially attacks medium length acyl chains (C6, C8). The enzyme demonstrates hydrolytic activity in emulsions of both medium and long chain triglycerides, as demonstrated by zymogram analysis. RAG-1 lipase is stabilized by Ca2+, with no loss in activity observed in preparations containing the cation, compared to a 70% loss over 30 h without Ca2+. The lipase is strongly inhibited by EDTA, Hg2+, and Cu2+, but shows no loss in activity after incubation with other metals or inhibitors examined in this study. The protein retains more than 75% of its initial activity after exposure to organic solvents, but is rapidly deactivated by pyridine. RAG-1 lipase offers potential for use as a biocatalyst. %B European Journal of Biochemistry %V 269 %P 5771 - 5779 %8 2002/12/01/ %@ 1432-1033 %G eng %U http://onlinelibrary.wiley.com/doi/10.1046/j.1432-1033.2002.03235.x/full %N 23 %R 10.1046/j.1432-1033.2002.03235.x %0 Conference Paper %D 2002 %T Quantifying the cost of providing intrusion tolerance in group communication systems %A Ramasamy,H. V. %A Pandey,P. %A Lyons,J. %A Michel Cukier %A Sanders,W. H. %K consistent group membership %K crash-fault-tolerant group communication system %K cryptography %K finite state machines %K groupware %K intrusion-tolerant microprotocols %K malicious intrusions %K multicast communication %K reliable ordered multicast properties %K security of data %K Transport protocols %X Group communication systems that provide consistent group membership and reliable, ordered multicast properties in the presence of faults resulting from malicious intrusions have not been analyzed extensively to quantify the cost of tolerating these intrusions. This paper attempts to quantify this cost by presenting results from an experimental evaluation of three new intrusion-tolerant microprotocols that have been added to an existing crash-fault-tolerant group communication system. The results are analyzed to identify the parts that contribute the most overhead during provision of intrusion tolerance at the group communication system level. %P 229 - 238 %8 2002/// %G eng %R 10.1109/DSN.2002.1028904 %0 Conference Paper %B Proceedings of the SPIE ITCOM %D 2002 %T Query routing in the TerraDir distributed directory %A Silaghi,B. %A Bhattacharjee, Bobby %A Keleher,P. %X We present the design and evaluation of the query-routing protocol of the TerraDir distributed directory.TerraDir is a wide-area distributed directory designed for hierarchical namespaces, and provides a lookup service for mapping keys to objects. We introduce distributed lookup and caching algorithms that leverage the underlying data hierarchy. Our algorithms provide efficient lookups while avoiding the load imbalances often associated with hierarchical systems. The TerraDir load balancing scheme also incorporates a node replication algorithm that provides configurable failure resilience with provably low overheads. %B Proceedings of the SPIE ITCOM %V 4868 %P 299 - 309 %8 2002/// %G eng %0 Book %D 2002 %T RECOMB '02: Proceedings of the sixth annual international conference on Computational biology %E Myers,Gene %E Hannenhalli, Sridhar %E Sankoff,David %E Istrail,Sorin %E Pevzner,Pavel %E Waterman,Michael %X The papers in this volume were presented at the Sixth Annual International Conference on Computational Biology RECOMB 2002, held in Washington, D.C. on April 18-21, 2002. The RECOMB series was started in 1997 by Sorin Istrail, Pavel Pevzner and Michael Waterman. RECOMB '99 took place in Lyon, France, RECOMB 2000 was held in Tokyo, Japan, and RECOMB 2001 was held in Montreal, Quebec, Canada.This year's call for papers gave rise to 118 submissions, out of which the program committee selected 35 papers to be presented at the conference and included in the proceedings. Each submission was refereed by at least three members of the program committee. After the completion of the referees' reports, an extensive web-based discussion took place.RECOMB 2002 had 8 invited speakers: Ruben Abagyan (The Scripps Research Institute), Ali H. Brivanlou (Rockefeller University), Evan Eichler (Case Western Reserve University), Harold "Skip" Garner (University of Texas Southwestern Medical Center at Dallas), David Ho (Rockefeller University), Gerry Rubin (Howard Hughes Medical Institute), J. Craig Venter (Celera) and Marc Vidal (Dana-Farber Cancer Institute). The Stanislaw Ulam Memorial Lecture was given by J. Craig Venter. The Distinguished Biology Lecture was given by David Ho. The Distinguished New Technologies Lecture was given by Harold Garner.Complete final versions of many of the papers presented in the conference will appear in a special issue of the Journal of Computational Biology, which is closely affiliated with the conference. %I ACM %C New York, NY, USA %8 2002/// %@ 1-58113-498-3 %G eng %0 Conference Paper %B Proceedings of the 10th ACM international symposium on Advances in geographic information systems %D 2002 %T Remote access to large spatial databases %A Tanin,Egemen %A Brabec,František %A Samet, Hanan %K client/server %K GIS %K Internet %K peer-to-peer %X Enterprises in the public and private sectors have been making their large spatial data archives available over the Internet. However, interactive work with such large volumes of online spatial data is a challenging task. We propose two efficient approaches to remote access to large spatial data. First, we introduce a client-server architecture where the work is distributed between the server and the individual clients for spatial query evaluation, data visualization, and data management. We enable the minimization of the requirements for system resources on the client side while maximizing system responsiveness as well as the number of connections one server can handle concurrently. Second, for prolonged periods of access to large online data, we introduce APPOINT (an Approach for Peer-to-Peer Offloading the INTernet). This is a centralized peer-to-peer approach that helps Internet users transfer large volumes of online data efficiently. In APPOINT, active clients of the client-server architecture act on the server's behalf and communicate with each other to decrease network latency, improve service bandwidth, and resolve server congestions. %B Proceedings of the 10th ACM international symposium on Advances in geographic information systems %S GIS '02 %I ACM %C New York, NY, USA %P 5 - 10 %8 2002/// %@ 1-58113-591-2 %G eng %U http://doi.acm.org/10.1145/585147.585150 %R 10.1145/585147.585150 %0 Conference Paper %B Proceedings of the 2002 annual national conference on Digital government research %D 2002 %T Remote thin-client access to spatial database systems %A Samet, Hanan %A Brabec,František %X Numerous federal agencies produce official statistics that are made accessible to ordinary citizens for searching and data retrieval. This is often done via the Internet through a web browser interface. If this data is presented in textual format, it can often be searched and retrieved by such attributes as topic, responsible agency, keywords, or press release. However, if the data is of spatial nature, e.g., in the form of a map, then using text-based queries is often too cumbersome for the intended audience. We propose to use the capabilities of the SAND Spatial Browser to provide more power to users of these databases. Using the SAND Spatial Browser allows users to define the spatial region of interest with greater specificity, instead of forcing them to retrieve data just for a particular location or a region with a predefined boundary. They can also make use of ranking which is the ability to retrieve data in the order of distance from other instances of the data or aggregates of data that are user-defined. Work is distributed between the SAND server and the individual clients for query evaluation, data visualization and data management. This enables the minimization of the necessary requirements for system resources on the client side while maximizing the number of connections one server can handle concurrently. Concrete experience with interfacing the SAND system with FedStats data is also discussed. %B Proceedings of the 2002 annual national conference on Digital government research %S dg.o '02 %I Digital Government Society of North America %P 1 - 8 %8 2002/// %G eng %U http://dl.acm.org/citation.cfm?id=1123098.1123170 %0 Conference Paper %B Parallel and Distributed Processing Symposium., Proceedings International, IPDPS 2002, Abstracts and CD-ROM %D 2002 %T Scheduling multiple data visualization query workloads on a shared memory machine %A Andrade,H. %A Kurc, T. %A Sussman, Alan %A Saltz, J. %K Atomic force microscopy %K Biomedical informatics %K Computer science %K Data analysis %K data visualisation %K Data visualization %K datasets %K deductive databases %K digitized microscopy image browsing %K directed graph %K directed graphs %K dynamic query scheduling model %K Educational institutions %K high workloads %K image database %K limited resources %K multiple data visualization query workloads %K multiple query optimization %K performance %K priority queue %K Processor scheduling %K Query processing %K query ranking %K Relational databases %K scheduling %K shared memory machine %K shared memory systems %K Virtual Microscope %K visual databases %X Query scheduling plays an important role when systems are faced with limited resources and high workloads. It becomes even more relevant for servers applying multiple query optimization techniques to batches of queries, in which portions of datasets as well as intermediate results are maintained in memory to speed up query evaluation. We present a dynamic query scheduling model based on a priority queue implementation using a directed graph and a strategy for ranking queries. We examine the relative performance of several ranking strategies on a shared-memory machine using two different versions of an application, called the Virtual Microscope, for browsing digitized microscopy images %B Parallel and Distributed Processing Symposium., Proceedings International, IPDPS 2002, Abstracts and CD-ROM %I IEEE %P 11 - 18 %8 2002/// %@ 0-7695-1573-8 %G eng %R 10.1109/IPDPS.2002.1015482 %0 Conference Paper %B DARPA Active NEtworks Conference and Exposition, 2002. Proceedings %D 2002 %T A secure PLAN (extended version) %A Hicks, Michael W. %A Keromytis,A. D %A Smith,J. M %K active internetwork %K active networks %K active-network firewall %K Authentication %K authorisation %K Authorization %K Cities and towns %K Computer networks %K Computer science %K cryptography %K functionally restricted packet language %K general-purpose service routines %K Information security %K internetworking %K IP networks %K latency overhead %K namespace-based security %K PLAN %K PLANet %K Planets %K programmability %K Safety %K security architecture %K telecommunication security %K trust management %K two-level architecture %K Web and internet services %X Active networks promise greater flexibility than current networks, but threaten safety and security by virtue of their programmability. We describe the design and implementation of a security architecture for the active network PLANet (Hicks et al., 1999). Security is obtained with a two-level architecture that combines a functionally restricted packet language, PLAN (Hicks et al., 1998), with an environment of general-purpose service routines governed by trust management (Blaze et al., 1996). In particular, we employ a technique which expands or contracts a packet's service environment based on its level of privilege, termed namespace-based security. As an application of our security architecture, we present the design and implementation of an active-network firewall. We find that the addition of the firewall imposes an approximately 34% latency overhead and as little as a 6.7% space overhead to incoming packets %B DARPA Active NEtworks Conference and Exposition, 2002. Proceedings %I IEEE %P 224 - 237 %8 2002/// %@ 0-7695-1564-9 %G eng %R 10.1109/DANCE.2002.1003496 %0 Conference Paper %B SC Conference %D 2002 %T SIGMA: A Simulator Infrastructure to Guide Memory Analysis %A DeRose,Luiz %A Ekanadham, K. %A Hollingsworth, Jeffrey K %A Sbaraglia, Simone %X In this paper we present SIGMA (Simulation Infrastructure to Guide Memory Analysis), a new data collection framework and family of cache analysis tools. The SIGMA environment provides detailed cache information by gathering memory reference data using software-based instrumentation. This infrastructure can facilitate quick probing into the factors that influence the performance of an application by highlighting bottleneck scenarios including: excessive cache/TLB misses and inefficient data layouts. The tool can also assist in perturbation analysis to determine performance variations caused by changes to architecture or program. Our validation tests using the SPEC Swim benchmark show that most of the performance metrics obtained with SIGMA are within 1% of the metrics obtained with hardware performance counters, with the advantage that SIGMA provides performance data on a data structure level, as specified by the programmer. %B SC Conference %I IEEE Computer Society %C Los Alamitos, CA, USA %P 1 - 1 %8 2002/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/SC.2002.10055 %0 Journal Article %J The VLDB Journal %D 2002 %T Speeding up construction of PMR quadtree-based spatial indexes %A Hjaltason,Gisli R. %A Samet, Hanan %K Bulk-loading %K I/O %K spatial indexing %X Spatial indexes, such as those based on the quadtree, are important in spatial databases for efficient execution of queries involving spatial constraints, especially when the queries involve spatial joins. In this paper we present a number of techniques for speeding up the construction of quadtree-based spatial indexes, specifically the PMR quadtree, which can index arbitrary spatial data. We assume a quadtree implementation using the “linear quadtree”, a disk-resident representation that stores objects contained in the leaf nodes of the quadtree in a linear index (e.g., a B-tree) ordered based on a space-filling curve. We present two complementary techniques: an improved insertion algorithm and a bulk-loading method. The bulk-loading method can be extended to handle bulk-insertions into an existing PMR quadtree. We make some analytical observations about the I/O cost and CPU cost of our PMR quadtree bulk-loading algorithm, and conduct an extensive empirical study of the techniques presented in the paper. Our techniques are found to yield significant speedup compared to traditional quadtree building methods, even when the size of a main memory buffer is very small compared to the size of the resulting quadtrees. %B The VLDB Journal %V 11 %P 109 - 137 %8 2002/10// %@ 1066-8888 %G eng %U http://dx.doi.org/10.1007/s00778-002-0067-8 %N 2 %R 10.1007/s00778-002-0067-8 %0 Conference Paper %B Proceedings of the 2nd ACM/IEEE-CS joint conference on Digital libraries %D 2002 %T Supporting access to large digital oral history archives %A Gustman,Samuel %A Soergel,Dagobert %A Oard, Douglas %A Byrne,William %A Picheny,Michael %A Ramabhadran,Bhuvana %A Greenberg,Douglas %K cataloging %K oral history %K research agenda %X This paper describes our experience with the creation, indexing, and provision of access to a very large archive of videotaped oral histories - 116,000 hours of digitized interviews in 32 languages from 52,000 survivors, liberators, rescuers, and witnesses of the Nazi Holocaust. It goes on to identify a set of critical research issues that must be addressed if we are to provide full and detailed access to collections of this size: issues in user requirement studies, automatic speech recognition, automatic classification, segmentation, summarization, retrieval, and user interfaces. The paper ends by inviting others to discuss use of these materials in their own research. %B Proceedings of the 2nd ACM/IEEE-CS joint conference on Digital libraries %S JCDL '02 %I ACM %C New York, NY, USA %P 18 - 27 %8 2002/// %@ 1-58113-513-0 %G eng %U http://doi.acm.org/10.1145/544220.544224 %R 10.1145/544220.544224 %0 Patent %D 2002 %T System and method for entering text in a virtual environment %A Evans,Francine %A Skiena,Steven %A Varshney, Amitabh %E The Research Foundation of the State University of New York %X A system and method for entering text in a virtual environment by sensory gloves. The user enters a key that represents one or more letters by simulating a press of a keyboard in the gloves. The user calibrates the gloves by entering text, during which time the system establishes threshold values that represent simulated presses for each finger. After the initial calibration of the sensory gloves, the user enters text with simulated finger presses. The system distinguishes which movements are intended as simulated finger presses by examining the relative motions of fingers and maintaining dynamic thresholds. Errors are alleviated by providing feedback to the user, such as beeps and a visual display of the fingers and the current text. Because keys may represent more than one character, the system determines the intended text by probabilistic analysis and the Viterbi algorithm. %V : 09/364,433 %8 2002/06/18/ %G eng %U http://www.google.com/patents?id=w0QJAAAAEBAJ %N 6407679 %0 Conference Paper %B In Proceedings of the 3rd International NASA Workshop on Planning and Scheduling for %D 2002 %T Three-dimensional visualization of hierarchical task network plans %A Kundu,K. %A Sessions,C. %A desJardins, Marie %A Rheingans,P. %X Artificial Intelligence (AI) planning techniques, particularly hierarchical task network (HTN) planners, can produce very complex plans, with many sequential and parallel actions decomposed into multiple levels of abstraction. %B In Proceedings of the 3rd International NASA Workshop on Planning and Scheduling for %8 2002/// %G eng %0 Journal Article %J ACM SIGSOFT Software Engineering Notes %D 2002 %T Triggered message sequence charts %A Sengupta,Bikram %A Cleaveland, Rance %B ACM SIGSOFT Software Engineering Notes %V 27 %P 167 - 167 %8 2002/11/01/ %@ 01635948 %G eng %U http://dl.acm.org/citation.cfm?id=605492 %R 10.1145/605466.605492 %0 Journal Article %J Experimental Parasitology %D 2002 %T Trypanosoma cruzi: RNA structure and post-transcriptional control of tubulin gene expression %A Bartholomeu,Daniella C. %A Silva,Rosiane A. %A Galvão,Lucia M. C. %A El‐Sayed, Najib M. %A Donelson,John E. %A Teixeira,Santuza M.R. %X Changes in tubulin expression are among the biochemical and morphological adaptations that occur during the life cycle of Trypanosomatids. To investigate the mechanism responsible for the differential accumulation of tubulin mRNAs in Trypanosoma cruzi, we determine the sequences of [alpha]- and [beta]-tubulin transcripts and analyzed their expression during the life cycle of the parasite. Two [beta]-tubulin mRNAs of 1.9 and 2.3 kb were found to differ mainly by an additional 369 nucleotides at the end of the 3' untranslated region (UTR). Although their transcription rates are similar in epimastigotes and amastigotes, [alpha]- and [beta]-tubulin transcripts are 3- to 6-fold more abundant in epimastigotes than in trypomastigotes and amastigotes. Accordingly, the half-lives of [alpha]- and [beta]-tubulin mRNAs are significantly higher in epimastigotes than in amastigotes. Transient transfection experiments indicated that positive regulatory elements occur in the 3' UTR plus downstream intergenic region of the [alpha]-tubulin gene and that both positive and negative elements occur in the equivalent regions of the [beta]-tubulin gene.Index Descriptions and Abbreviations: Kinetoplastida; Trypanosoma cruzi; tubulin; gene regulation; PCR, polymerase chain reaction; UTR, untranslated region; IR, intergenic region; SL, spliced leader; BAC, bacterial artificial chromosome. %B Experimental Parasitology %V 102 %P 123 - 133 %8 2002/11// %@ 0014-4894 %G eng %U http://www.sciencedirect.com/science/article/pii/S0014489403000341 %N 3-4 %R 16/S0014-4894(03)00034-1 %0 Journal Article %J interactions %D 2002 %T Understanding human reactivites and relationships: an excerpt from Leonardo's laptop %A Shneiderman, Ben %X "These notes reveal the intimate tie in Leonardo's thinking between… phenomena in general and the need to put such information to practical use."
--- A. Richard Turner, Inventing Leonardo, 1994, p 184 %B interactions %V 9 %P 40 - 53 %8 2002/09// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/566981.566982 %N 5 %R 10.1145/566981.566982 %0 Book Section %B DUC 02 Conference ProceedingsDUC 02 Conference Proceedings %D 2002 %T Understanding Machine Performance in the Context of Human Performance for Multi-document Summarization %A Schlesinger,J.D. %A Okurowski,M. E. %A Conroy,J.M. %A O'Leary, Dianne P. %A Taylor,A. %A Hobbs,J. %A Wilson,H. T. %X We present a discussion of our summarization algorithm -- both single and multi-document -- along with a discussion of the evaluation experiments we have undertaken, what we have learned from them, and how we intend to utilize this information. %B DUC 02 Conference ProceedingsDUC 02 Conference Proceedings %I U.S. National Inst. of Standards and Technology %8 2002/// %G eng %U http://duc.nist.gov/http://duc.nist.gov/ %0 Journal Article %J Information Systems %D 2002 %T Using Hilbert curve in image storing and retrieving %A Song,Zhexuan %A Roussopoulos, Nick %K Hilbert order %K Row-wise order %K Subset query %X In this paper, we propose a method to accelerate the speed of subset query on uncompressed images. First, we change the method to store images: the pixels of images are stored on the disk in the Hilbert order instead of row-wise order that is used in traditional methods. After studying the properties of the Hilbert curve, we give a new algorithm which greatly reduces the number of data segments in subset query range. Although, we have to retrieve more data than necessary, because the speed of sequential readings is much faster than the speed of random access readings, it takes about 10% less elapsed time in our algorithm than in the traditional algorithms to execute the subset queries. In some systems, the saving is as much as 90%. %B Information Systems %V 27 %P 523 - 536 %8 2002/12// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306437902000194 %N 8 %R 16/S0306-4379(02)00019-4 %0 Journal Article %J Journal of Digital Information %D 2002 %T Virtual Telescopes in Education %A Hoban,S. %A desJardins, Marie %A Farrell,N. %A Rathod,P. %A Sachs,J. %A Sansare,S. %A Yesha,Y. %A Keating,J. %A Busschots,B. %A Means,J. %A others %X Virtual Telescopes in Education is providing the services required to operate a virtual observatory comprising distributed telescopes, including an interactive, constraint-based scheduling service, data and resource archive, proposal preparation and review environment, and a VTIE Journal. A major goal of VTIE is to elicit from learners questions about the nature of celestial objects and the physical processes that give rise to the spectacular imagery that catches their imaginations. Generation of constrained science questions will assist learners in the science process. To achieve interoperability with other NSDL resources, our approach follows the Open Archives Initiative and the W3C Semantic Web activity. %B Journal of Digital Information %V 2 %8 2002/// %G eng %N 4 %0 Journal Article %J University of Maryland, Computer Science Dept. Tech Report, CS-TR-4365 %D 2002 %T Visual queries for finding patterns in time series data %A Hochheiser,H. %A Shneiderman, Ben %X Few tools exist for data exploration and pattern iden-tification in time series data sets. Timeboxes are rectan- gular, direct-manipulation queries for studying time-series datasets. Timeboxes are the primary query tool in our Time- Searcher application, which supports interactive explo- ration via dynamic queries, along with overviews of query results and drag-and-drop support for query-by-example. This paper describes the TimeSearcher application and pos- sible extensions to the timebox query model, along with a discussion of the use of TimeSearcher for exploring a time series data set involving gene expression profiles. %B University of Maryland, Computer Science Dept. Tech Report, CS-TR-4365 %8 2002/// %G eng %0 Journal Article %J Journal of Science Education and Technology %D 2002 %T A visual search tool for early elementary science students %A Revelle,G. %A Druin, Allison %A Platner,M. %A Bederson, Benjamin B. %A Hourcade,J. P %A Sherman,L. %B Journal of Science Education and Technology %V 11 %P 49 - 57 %8 2002/// %G eng %N 1 %0 Journal Article %J Journal of Algorithms %D 2002 %T Wavelength rerouting in optical networks, or the Venetian Routing problem %A Caprara,Alberto %A Italiano,Giuseppe F. %A Mohan,G. %A Panconesi,Alessandro %A Srinivasan, Aravind %K Approximation algorithms %K Label Cover %K Optical networks %K Shortest paths %K Wavelength rerouting %K Wavelength-division multiplexing %X Wavelength rerouting has been suggested as a viable and cost-effective method to improve the blocking performance of wavelength-routed wavelength-division multiplexing (WDM) networks. This method leads to the following combinatorial optimization problem, dubbed Venetian Routing. Given a directed multigraph G along with two vertices s and t and a collection of pairwise arc-disjoint paths, we wish to find an st-path which arc-intersects the smallest possible number of the given paths. In this paper we prove the computational hardness of this problem even in various special cases, and present several approximation algorithms for its solution. In particular we show a non-trivial connection between Venetian Routing and Label Cover. %B Journal of Algorithms %V 45 %P 93 - 125 %8 2002/11// %@ 0196-6774 %G eng %U http://www.sciencedirect.com/science/article/pii/S0196677402002146 %N 2 %R 10.1016/S0196-6774(02)00214-6 %0 Conference Paper %B Software Metrics, 2002. Proceedings. Eighth IEEE Symposium on %D 2002 %T What we have learned about fighting defects %A Shull, F. %A Basili, Victor R. %A Boehm,B. %A Brown,A. W %A Costa,P. %A Lindvall,M. %A Port,D. %A Rus,I. %A Tesoriero,R. %A Zelkowitz, Marvin V %K based %K Center %K Computer %K defect %K development; %K education; %K electronic %K Empiric %K engineering %K engineering; %K eWorkshops; %K for %K heuristics; %K reduction; %K Science %K software %K workshops; %X The Center for Empirically Based Software Engineering helps improve software development by providing guidelines for selecting development techniques, recommending areas for further research, and supporting software engineering education. A central activity toward achieving this goal has been the running of "e- Workshops" that capture expert knowledge with a minimum of overhead effort to formulate heuristics on a particular topic. The resulting heuristics are a useful summary of the current state of knowledge in an area based on expert opinion. This paper discusses the results to date of a series of e-Workshops on software defect reduction. The original discussion items are presented along with an encapsulated summary of the expert discussion. The reformulated heuristics can be useful both to researchers (for pointing out gaps in the current state of the knowledge requiring further investigation) and to practitioners (for benchmarking or setting expectations about development practices). %B Software Metrics, 2002. Proceedings. Eighth IEEE Symposium on %P 249 - 258 %8 2002/// %G eng %R 10.1109/METRIC.2002.1011343 %0 Journal Article %J IEEE Communications Magazine %D 2002 %T Your 802.11 network has no clothes %A Mishra,A. %A Shin,M. %A Arbaugh, William A. %B IEEE Communications Magazine %V 9 %P 44 - 51 %8 2002/// %G eng %0 Journal Article %J Parallel and Distributed Systems, IEEE Transactions on %D 2001 %T An adaptive algorithm for tolerating value faults and crash failures %A Ren,Yansong %A Michel Cukier %A Sanders,W. H. %K active replication communication %K adaptive algorithm %K adaptive fault tolerance %K adaptive majority voting algorithm %K AQuA architecture %K client-server systems %K CORBA %K crash failures %K data consistency %K data integrity %K Dependability %K distributed object management %K fault tolerant computing %K objects replication %K value faults %X The AQuA architecture provides adaptive fault tolerance to CORBA applications by replicating objects and providing a high-level method that an application can use to specify its desired level of dependability. This paper presents the algorithms that AQUA uses, when an application's dependability requirements can change at runtime, to tolerate both value faults in applications and crash failures simultaneously. In particular, we provide an active replication communication scheme that maintains data consistency among replicas, detects crash failures, collates the messages generated by replicated objects, and delivers the result of each vote. We also present an adaptive majority voting algorithm that enables the correct ongoing vote while both the number of replicas and the majority size dynamically change. Together, these two algorithms form the basis of the mechanism for tolerating and recovering from value faults and crash failures in AQuA %B Parallel and Distributed Systems, IEEE Transactions on %V 12 %P 173 - 192 %8 2001/02// %@ 1045-9219 %G eng %N 2 %R 10.1109/71.910872 %0 Report %D 2001 %T Adjusting the Rayleigh Quotient in Semiorthogonal Lanczos Methods %A Stewart, G.W. %K Technical Report %X In a semiorthogonal Lanczos algorithm, the orthogonality of theLanczos vectors is allowed to deteriorate to roughly the square root of the rounding unit, after which the current vectors are reorthogonalized. A theorem of Simon \cite{simo:84} shows that the Rayleigh quotient\,---\,i.e., the tridiagonal matrix produced by the Lanczos recursion\,---\,contains fully accurate approximations to the Ritz values in spite of the lack of orthogonality. Unfortunately, the same lack of orthogonality can cause the Ritz vectors to fail to converge. It also makes the classical estimate for the residual norm misleadingly small. In this note we show how to adjust the Rayleigh quotient to overcome this problem. (Cross-referenced as UMIACS-TR-2001-31) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2001-31 %8 2001/05/10/ %G eng %U http://drum.lib.umd.edu/handle/1903/1132 %0 Journal Article %J Image Processing, IEEE Transactions on %D 2001 %T Approximating large convolutions in digital images %A Mount, Dave %A Kanungo,T. %A Netanyahu,N. S %A Piatko,C. %A Silverman,R. %A Wu,A. Y %K 2D %K algorithm;binary %K approximation;mathematical %K convolution %K convolution;binary %K convolutions;Bresenham's %K convolutions;convex %K convolutions;geometric %K images;discrete %K kernel;convex %K kernel;digital %K line-drawing %K morphology; %K morphology;approximation %K object;image %K polygonal %K processing;large %K processing;mathematical %K theory;convolution;image %K two-dimensional %X Computing discrete two-dimensional (2-D) convolutions is an important problem in image processing. In mathematical morphology, an important variant is that of computing binary convolutions, where the kernel of the convolution is a 0-1 valued function. This operation can be quite costly, especially when large kernels are involved. We present an algorithm for computing convolutions of this form, where the kernel of the binary convolution is derived from a convex polygon. Because the kernel is a geometric object, we allow the algorithm some flexibility in how it elects to digitize the convex kernel at each placement, as long as the digitization satisfies certain reasonable requirements. We say that such a convolution is valid. Given this flexibility we show that it is possible to compute binary convolutions more efficiently than would normally be possible for large kernels. Our main result is an algorithm which, given an m times;n image and a k-sided convex polygonal kernel K, computes a valid convolution in O(kmn) time. Unlike standard algorithms for computing correlations and convolutions, the running time is independent of the area or perimeter of K, and our techniques do not rely on computing fast Fourier transforms. Our algorithm is based on a novel use of Bresenham's (1965) line-drawing algorithm and prefix-sums to update the convolution incrementally as the kernel is moved from one position to another across the image %B Image Processing, IEEE Transactions on %V 10 %P 1826 - 1835 %8 2001/12// %@ 1057-7149 %G eng %N 12 %R 10.1109/83.974567 %0 Journal Article %J Automata, Languages and Programming %D 2001 %T Approximation algorithms for partial covering problems %A Gandhi,R. %A Khuller, Samir %A Srinivasan, Aravind %X We study the generalization of covering problems to partial covering. Here we wish to cover only a desired number of elements, rather than covering all elements as in standard covering problems. For example, in k-set cover, we wish to choose a minimum number of sets to cover at least k elements. For k-set cover, if each element occurs in at most f sets, then we derive a primal-dual f-approximation algorithm (thus implying a 2-approximation for k-vertex cover) in polynomial time. In addition to its simplicity, this algorithm has the advantage of being parallelizable. For instances where each set has cardinality at most three, we obtain an approximation of 4/3. We also present better-than-2-approximation algorithms for k-vertex cover on bounded degree graphs, and for vertex cover on expanders of bounded average degree. We obtain a polynomial-time approximation scheme for k-vertex cover on planar graphs, and for covering points in R d by disks. %B Automata, Languages and Programming %P 225 - 236 %8 2001/// %G eng %R 10.1007/3-540-48224-5_19 %0 Conference Paper %B 39th AIAA Aerospace Sciences Meeting and Exhibit %D 2001 %T Assessment of the k-epsilon turbulence model for compressible flows using direct simulation data %A Sinha,K. %A Candler,G. V %A Martin, M.P %X We evaluate the k-epsilon turbulence model using direct numerical simulation (DNS) data of a Mach 4 boundary layer. We find that the low Reynolds number damping functions for the Reynolds stress must be corrected by the density ratio to match the DNS data. We present the budget of the k equation and assess the modeling of the various source terms. The models for all the source terms, except for the production and dilatational dissipation terms, are found to be adequate. Finally, we present the solenoidal dissipation rate equation and compute its budget using the boundary layer data. We compare this equation with the dissipation rate equation in an incompressible flow to show the equivalence between the two equations. This is the basis for modeling the solenoidal dissipation equation. However, an additional term in the equation due to variation of fluid viscosity needs to be modeled. %B 39th AIAA Aerospace Sciences Meeting and Exhibit %C Reno, NV %8 2001/// %G eng %0 Conference Paper %B Automated Software Engineering, International Conference on %D 2001 %T Automated Validation of Software Models %A Sims,Steve %A Cleaveland, Rance %A Butts,Ken %A Ranville,Scott %X This paper describes the application of an automated verification tool to a software model developed at Ford. Ford already has in place an advanced model-based software development framework that employs the Matlab?, Simulink?, and Stateflow? modeling tools. During this project we applied the invariant checker Salsa to a Simulink?/ Stateflow? model of automotive software to check for nondeterminism, missing cases, dead code, and redundant code. During the analysis, a number of anomalies were detected that had not been found during manual review. We argue that the detection and correction of these problems demonstrates a cost-effective application of formal verification that elevates our level of confidence in the model. %B Automated Software Engineering, International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %P 91 - 91 %8 2001/// %@ 0-7695-1426-X %G eng %R http://doi.ieeecomputersociety.org/10.1109/ASE.2001.989794 %0 Report %D 2001 %T Backward Error Bounds for Approximate Krylov Subspaces %A Stewart, G.W. %K Technical Report %X Let $A$ be a matrix of order $n$ and let $\clu\subset\comp^{n}$ be asubspace of dimension $k$. In this note we determine a matrix $E$ of minimal norm such that $\clu$ is a Krylov subspace of $A+E$. (Cross-referenced as UMIACS-TR-2001-32) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2001-32 %8 2001/05/10/ %G eng %U http://drum.lib.umd.edu/handle/1903/1133 %0 Book %D 2001 %T Becoming a Scholar in the Digital Age %A Duderstadt,James %A Arms,William %A Messina,Paul %A Ellisman,Mark %A Atkins,Daniel %A Fox,Edward %A Shneiderman, Ben %A Nissenbaum,Helen %A Lederberg,Joshua %K bibtex-import %K digital-library %K dlbook %I National Research Council %8 2001/// %G eng %0 Journal Article %J SIAM Journal on Discrete Mathematics %D 2001 %T Better Approximation Guarantees for Job-Shop Scheduling %A Goldberg,Leslie Ann %A Paterson,Mike %A Srinivasan, Aravind %A Sweedyk,Elizabeth %X Job-shop scheduling is a classical NP-hard problem. Shmoys, Stein, and Wein presented the first polynomial-time approximation algorithm for this problem that has a good (polylogarithmic) approximation guarantee. We improve the approximation guarantee of their work and present further improvements for some important NP-hard special cases of this problem (e.g., in the preemptive case where machines can suspend work on operations and later resume). We also present NC algorithms with improved approximation guarantees for some NP-hard special cases. %B SIAM Journal on Discrete Mathematics %V 14 %P 67 - 67 %8 2001/// %@ 08954801 %G eng %U http://link.aip.org/link/SJDMEC/v14/i1/p67/s1&Agg=doi %N 1 %R 10.1137/S0895480199326104 %0 Journal Article %J Product Focused Software Process Improvement %D 2001 %T Building an Experience Base for Software Engineering: A report on the first CeBASE eWorkshop %A Basili, Victor R. %A Tesoriero,R. %A Costa,P. %A Lindvall,M. %A Rus,I. %A Shull, F. %A Zelkowitz, Marvin V %X New information is obtained by research and disseminated by papers in conferences and journals. The synthesis of knowledge depends upon social discourse among the experts in a given domain to discuss the impact of this new information. Meetings among such experts are, however, expensive and time consuming. In this paper we discuss the organization of CeBASE, a center whose goal is the collection and dissemination of empirically-based software engineering knowledge, and the concept of the online workshop or eWorkshop as a way to use the Internet to minimize the needs of face-to-face meetings. We discuss the design of our eWorkshop and give the results of one eWorkshop that discussed the impact of defect reduction strategies.We want to thank Barry Boehm, Scott Henninger, Rayford Vaughn, Winsor Brown, Dan Port and Michael Frey as well as all participants for their contribution to the success of the eWorkshop. We also want to thank students at USC and UMD for their contribution in testing the system and Jennifer Dix for proof reading this paper. %B Product Focused Software Process Improvement %P 110 - 125 %8 2001/// %G eng %R 10.1007/3-540-44813-6_13 %0 Conference Paper %B CHI '01 extended abstracts on Human factors in computing systems %D 2001 %T Component-based, user-constructed, multiple-view visualization %A North,Chris %A Shneiderman, Ben %K Information Visualization %K multiple views %K software components %K Specification %K tight coupling %K user interface %X A major hindrance to the usage of information visualization in common tasks is that typically a new visualization interface must be custom programmed to suit each task. This video demonstrates a system and user interface that attempts to solve this problem by enabling end users to construct their own multiple-view visualization interfaces that are appropriate for their tasks and data. Users accomplish this by snapping together component visualizations and specifying tight couplings between them. Then they can use their newly constructed visualization interfaces to perform their tasks. %B CHI '01 extended abstracts on Human factors in computing systems %S CHI EA '01 %I ACM %C New York, NY, USA %P 201 - 202 %8 2001/// %@ 1-58113-340-5 %G eng %U http://doi.acm.org/10.1145/634067.634188 %R 10.1145/634067.634188 %0 Journal Article %J Artificial Life %D 2001 %T Conditions enabling the emergence of inter-agent signalling in an artificial world %A Reggia, James A. %A Schulz,R. %A Wilkinson,G. S %A Uriagereka,J. %B Artificial Life %V 7 %P 3 - 32 %8 2001/// %G eng %N 1 %0 Conference Paper %B Proceedings of the 8th European software engineering conference held jointly with 9th ACM SIGSOFT international symposium on Foundations of software engineering %D 2001 %T Coverage criteria for GUI testing %A Memon, Atif M. %A Soffa,Mary Lou %A Pollack,Martha E. %K component testing %K event-based coverage %K event-flow graph %K GUI test coverage %K GUI testing %K integration tree %X A widespread recognition of the usefulness of graphical user interfaces (GUIs) has established their importance as critical components of today's software. GUIs have characteristics different from traditional software, and conventional testing techniques do not directly apply to GUIs. This paper's focus is on coverage critieria for GUIs, important rules that provide an objective measure of test quality. We present new coverage criteria to help determine whether a GUI has been adequately tested. These coverage criteria use events and event sequences to specify a measure of test adequacy. Since the total number of permutations of event sequences in any non-trivial GUI is extremely large, the GUI's hierarchical structure is exploited to identify the important event sequences to be tested. A GUI is decomposed into GUI components, each of which is used as a basic unit of testing. A representation of a GUI component, called an event-flow graph, identifies the interaction of events within a component and intra-component criteria are used to evaluate the adequacy of tests on these events. The hierarchical relationship among components is represented by an integration tree, and inter-component coverage criteria are used to evaluate the adequacy of test sequences that cross components. Algorithms are given to construct event-flow graphs and an integration tree for a given GUI, and to evaluate the coverage of a given test suite with respect to the new coverage criteria. A case study illustrates the usefulness of the coverage report to guide further testing and an important correlation between event-based coverage of a GUI and statement coverage of its software's underlying code. %B Proceedings of the 8th European software engineering conference held jointly with 9th ACM SIGSOFT international symposium on Foundations of software engineering %S ESEC/FSE-9 %I ACM %C New York, NY, USA %P 256 - 267 %8 2001/// %@ 1-58113-390-1 %G eng %U http://doi.acm.org/10.1145/503209.503244 %R 10.1145/503209.503244 %0 Journal Article %J Interactions %D 2001 %T Design: CUU: bridging the digital divide with universal usability %A Shneiderman, Ben %B Interactions %V 8 %P 11 - 15 %8 2001/03// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/361897.361905 %N 2 %R 10.1145/361897.361905 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2001-4299 %D 2001 %T The Design of TerraDir %A Bhattacharjee, Bobby %A Keleher,Pete %A Silaghi,Bujor %K Technical Report %X We present the design and initial evaluation of TerraDir: an approach forimplementing customizable, distributed, peer-to-peer directories over which a broad range of wide-area resource discovery applications can be implemented. TerraDir's structure is two-tiered, consisting of a base protocol for providing arbitrary application-layer connectivity, and dynamic view materializations that efficiently realize different user-specified views of application-layer resources. TerraDir is specifically designed for data that can be arranged in a rooted hierarchy, and provides very efficient and robust lookups over such data while still being able to allow efficient searching. (Also referenced as UMIACS-TR-2001-4299) %B Technical Reports from UMIACS, UMIACS-TR-2001-4299 %8 2001/12/17/ %G eng %U http://drum.lib.umd.edu/handle/1903/1161 %0 Conference Paper %B Proceedings of the 10th conference on USENIX Security Symposium-Volume 10 %D 2001 %T Detecting format string vulnerabilities with type qualifiers %A Shankar,U. %A Talwar,K. %A Foster, Jeffrey S. %A Wagner,D. %B Proceedings of the 10th conference on USENIX Security Symposium-Volume 10 %P 16 - 16 %8 2001/// %G eng %0 Book %D 2001 %T The Digital University: Building a Learning Community %A Shneiderman, Ben %K elearning %K koelpu %X University education continues to be revolutionized by the use of Web-based teaching and learning systems. Following on from "The Digital University: Reinventing the Academy", this book provides a fully up-to-date and practical guide to using and implementing this important technology. Looking specifically at asynchronous collaboration, it covers:- policies- management of collaboration- distance learning- support for authoring- course design- educational metadata schemaand will be an essential buy for managers, lecturers, administrators, department heads and researchers.It includes a foreword by Ben Shneiderman, Director of the HCI Laboratory at the University of Maryland, USA.} %I Springer %8 2001/12/18/ %@ 1852334789 %G eng %U http://www.amazon.de/exec/obidos/redirect?tag=citeulike01-21&path=ASIN/1852334789 %0 Journal Article %J Unpublished manuscript %D 2001 %T The discrepancy of permutation families %A Spencer,JH %A Srinivasan, Aravind %A Tetali,P. %X In this note, we show that the discrepancy of any family of l permutations of[n] = {1,2,...,n} is O( √ llog n), improving on the O(llog n) bound due to Bohus (Random Structures & Algorithms, 1:215–220, 1990). In the case where l ≥ n, we show that the discrepancy is Θ(min{√nlog(2l/n),n}). %B Unpublished manuscript %8 2001/// %G eng %0 Journal Article %J Parallel Computing %D 2001 %T Distributed processing of very large datasets with DataCutter %A Beynon,Michael D. %A Kurc,Tahsin %A Catalyurek,Umit %A Chang,Chialin %A Sussman, Alan %A Saltz,Joel %K Component architectures %K Data analysis %K Distributed computing %K Multi-dimensional datasets %K Runtime systems %X We describe a framework, called DataCutter, that is designed to provide support for subsetting and processing of datasets in a distributed and heterogeneous environment. We illustrate the use of DataCutter with several data-intensive applications from diverse fields, and present experimental results. %B Parallel Computing %V 27 %P 1457 - 1478 %8 2001/10// %@ 0167-8191 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167819101000990 %N 11 %R 10.1016/S0167-8191(01)00099-0 %0 Conference Paper %B 42nd IEEE Symposium on Foundations of Computer Science, 2001. Proceedings %D 2001 %T Distributions on level-sets with applications to approximation algorithms %A Srinivasan, Aravind %K Approximation algorithms %K approximation guarantee %K approximation theory %K bounded-degree graphs %K computational geometry %K Concrete %K fixed-weight vectors %K group Steiner tree problem %K level-sets %K linear-time algorithm %K low-congestion multi-path routing %K marginal distributions %K maximum coverage versions %K negative correlation properties %K partial vertex cover problems %K trees (mathematics) %X We consider a family of distributions on fixed-weight vectors in {0, 1}t; these distributions enjoy certain negative correlation properties and also satisfy pre-specified conditions on their marginal distributions. We show the existence of such families, and present a linear-time algorithm to sample from them. This yields improved approximation algorithms for the following problems: (a) low-congestion multi-path routing; (b) maximum coverage versions of set cover; (c) partial vertex cover problems for bounded-degree graphs; and (d) the Group Steiner Tree problem. For (a) and (b), the improvement is in the approximation ratio; for (c), we show how to speedup existing approximation algorithms while preserving the best-known approximation ratio; we also improve the approximation ratio for certain families of instances of unbounded degree. For (d), we derive an approximation algorithm whose approximation guarantee is at least as good as what is known; our algorithm is shown to have a better approximation guarantee for the worst known input families for existing algorithms. %B 42nd IEEE Symposium on Foundations of Computer Science, 2001. Proceedings %I IEEE %P 588 - 597 %8 2001/10/08/11 %@ 0-7695-1116-3 %G eng %R 10.1109/SFCS.2001.959935 %0 Conference Paper %B Proceedings of the twelfth annual ACM-SIAM symposium on Discrete algorithms %D 2001 %T Domatic partitions and the Lovász local lemma %A Srinivasan, Aravind %B Proceedings of the twelfth annual ACM-SIAM symposium on Discrete algorithms %S SODA '01 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 922 - 923 %8 2001/// %@ 0-89871-490-7 %G eng %U http://dl.acm.org/citation.cfm?id=365411.365810 %0 Conference Paper %B Proceedings of the 10th conference on USENIX Security Symposium - Volume 10 %D 2001 %T Dos and don'ts of client authentication on the web %A Fu,Kevin %A Sit,Emil %A Smith,Kendra %A Feamster, Nick %X Client authentication has been a continuous source of problems on the Web. Although many well-studied techniques exist for authentication, Web sites continue to use extremely weak authentication schemes, especially in non-enterprise environments such as store fronts. These weaknesses often result from careless use of authenticators within Web cookies. Of the twenty-seven sites we investigated, we weakened the client authentication on two systems, gained unauthorized access on eight, and extracted the secret key used to mint authenticators from one. We provide a description of the limitations, requirements, and security models specific to Web client authentication. This includes the introduction of the interrogative adversary, a surprisingly powerful adversary that can adaptively query a Web site. We propose a set of hints for designing a secure client authentication scheme. Using these hints, we present the design and analysis of a simple authentication scheme secure against forgeries by the interrogative adversary. In conjunction with SSL, our scheme is secure against forgeries by the active adversary. %B Proceedings of the 10th conference on USENIX Security Symposium - Volume 10 %S SSYM'01 %I USENIX Association %C Berkeley, CA, USA %P 19 - 19 %8 2001/// %G eng %U http://dl.acm.org/citation.cfm?id=1267612.1267631 %0 Book Section %B Discovery ScienceDiscovery Science %D 2001 %T Dynamic Aggregation to Support Pattern Discovery: A Case Study with Web Logs %A Tang,Lida %A Shneiderman, Ben %E Jantke,Klaus %E Shinohara,Ayumi %X Rapid growth of digital data collections is overwhelming the capabilities of humans to comprehend them without aid. The extraction of useful data from large raw data sets is something that humans do poorly. Aggregation is a technique that extracts important aspect from groups of data thus reducing the amount that the user has to deal with at one time, thereby enabling them to discover patterns, outliers, gaps, and clusters. Previous mechanisms for interactive exploration with aggregated data were either too complex to use or too limited in scope. This paper proposes a new technique for dynamic aggregation that can combine with dynamic queries to support most of the tasks involved in data manipulation. %B Discovery ScienceDiscovery Science %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2226 %P 464 - 469 %8 2001/// %@ 978-3-540-42956-2 %G eng %U http://dx.doi.org/10.1007/3-540-45650-3_42 %0 Conference Paper %B Fifth International Conference on Information Visualisation, 2001. Proceedings %D 2001 %T Dynamic queries and brushing on choropleth maps %A Dang,G. %A North,C. %A Shneiderman, Ben %K brushing %K business %K cartography %K choropleth maps %K color coding %K colour graphics %K complex data sets %K Computational Intelligence Society %K Computer science %K data visualisation %K Data visualization %K demographic data %K Demography %K Dynamaps %K dynamic queries %K economic data %K Educational institutions %K geographic information system %K geographic information systems %K Joining processes %K map representations %K map-based information visualization %K Query processing %K Scattering %K scatterplot view %K tabular representations %K user interface %K User interfaces %K World Wide Web %X Users who must combine demographic, economic or other data in a geographic context are often hampered by the integration of tabular and map representations. Static, paper-based solutions limit the amount of data that can be placed on a single map or table. By providing an effective user interface, we believe that researchers, journalists, teachers, and students can explore complex data sets more rapidly and effectively. This paper presents Dynamaps, a generalized map-based information visualization tool for dynamic queries and brushing on choropleth maps. Users can use color coding to show a variable on each geographic region, and then filter out areas that do not meet the desired criteria. In addition, a scatterplot view and a details-on-demand window support overviews and specific fact-finding %B Fifth International Conference on Information Visualisation, 2001. Proceedings %I IEEE %P 757 - 764 %8 2001/// %@ 0-7695-1195-3 %G eng %R 10.1109/IV.2001.942141 %0 Conference Paper %D 2001 %T A dynamic replica selection algorithm for tolerating timing faults %A Krishnamurthy, S. %A Sanders,W. H. %A Michel Cukier %K AQuA %K client %K client-server systems %K CORBA-based middleware %K distributed object management %K distributed services %K dynamic replica selection algorithm %K fault tolerant computing %K local area network %K Local area networks %K quality of service %K replica failures %K response time %K server replication %K time-critical applications %K timing failures %K timing fault tolerance %X Server replication is commonly used to improve the fault tolerance and response time of distributed services. An important problem when executing time-critical applications in a replicated environment is that of preventing timing failures by dynamically selecting the replicas that can satisfy a client's timing requirement, even when the quality of service is degraded due to replica failures and excess load on the server. We describe the approach we have used to solve this problem in AQuA, a CORBA-based middleware that transparently replicates objects across a local area network. The approach we use estimates a replica's response time distribution based on performance measurements regularly broadcast by the replica. An online model uses these measurements to predict the probability with which a replica can prevent a timing failure for a client. A selection algorithm then uses this prediction to choose a subset of replicas that can together meet the client's timing constraints with at least the probability requested by the client. We conclude with experimental results based on our implementation. %P 107 - 116 %8 2001/07// %G eng %R 10.1109/DSN.2001.941397 %0 Conference Paper %B IEEE Global Telecommunications Conference, 2001. GLOBECOM '01 %D 2001 %T Efficient algorithms for location and sizing problems in network design %A Kumaran,K. %A Srinivasan, Aravind %A Qiong Wang %A Lanning,S. %A Ramakrishnan,K. G %K Algorithm design and analysis %K Broadband communication %K broadband communication networks %K broadband networks %K capacity modularity %K Communication networks %K computer network reliability %K distributed network elements %K Economies of scale %K greedy randomized adaptive search %K homing %K integer programming %K Intelligent networks %K Internet telephony %K large-scale location problems %K Large-scale systems %K Linear programming %K mixed-integer programs %K near-optimal solutions %K network design %K Optical switches %K randomised algorithms %K reliability %K search problems %K sizing %K Technological innovation %K telecommunication network planning %K Web caching %X Large-scale location, sizing and homing problems of distributed network elements, have received much attention recently due to the massive deployment of broadband communication networks for services like Internet telephony and Web caching. Key considerations in designing these networks include modularity of capacity, economies of scale in cost, and reliability. We formulate a general class of such network design problems as Mixed-Integer Programs. These problems are computationally intractable in general; under various asymptotic conditions, we show how to compute near-optimal solutions. To deal with arbitrary instances, we develop new algorithms based on linear programming, as well as greedy randomized adaptive search. These algorithms achieved near-optimal solutions with reasonable computation time for our experiments %B IEEE Global Telecommunications Conference, 2001. GLOBECOM '01 %I IEEE %V 4 %P 2586-2590 vol.4 - 2586-2590 vol.4 %8 2001/// %@ 0-7803-7206-9 %G eng %R 10.1109/GLOCOM.2001.966243 %0 Book Section %B Advances in Cryptology — EUROCRYPT 2001Advances in Cryptology — EUROCRYPT 2001 %D 2001 %T Efficient and Non-interactive Non-malleable Commitment %A Di Crescenzo,Giovanni %A Katz, Jonathan %A Ostrovsky,Rafail %A Smith,Adam %E Pfitzmann,Birgit %K Computer science %X We present new constructions of non-malleable commitment schemes, in the public parameter model (where a trusted party makes parameters available to all parties), based on the discrete logarithm or RSA assumptions. The main features of our schemes are: they achieve near-optimal communication for arbitrarily-large messages and are non-interactive . Previous schemes either required (several rounds of) interaction or focused on achieving non-malleable commitment based on general assumptions and were thus efficient only when committing to a single bit. Although our main constructions are for the case of perfectly-hiding commitment, we also present a communication-efficient, non-interactive commitment scheme (based on general assumptions) that is perfectly binding. %B Advances in Cryptology — EUROCRYPT 2001Advances in Cryptology — EUROCRYPT 2001 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2045 %P 40 - 59 %8 2001/// %@ 978-3-540-42070-5 %G eng %U http://www.springerlink.com/content/nhbj60a9da101w0r/abstract/ %0 Conference Paper %B Proceedings of the 2001 Workshop on Parallel and Distributed Computing in Imaging Processing, Video Processing, and Multimedia %D 2001 %T An efficient system for multi-perspective imaging and volumetric shape analysis %A Borovikov,E. %A Sussman, Alan %A Davis, Larry S. %B Proceedings of the 2001 Workshop on Parallel and Distributed Computing in Imaging Processing, Video Processing, and Multimedia %8 2001/// %G eng %0 Journal Article %J Numerische Mathematik %D 2001 %T On the eigensystems of graded matrices %A Stewart, G.W. %B Numerische Mathematik %V 90 %P 349 - 370 %8 2001/// %G eng %N 2 %R 10.1007/s002110100290 %0 Journal Article %J Proceedings of the IFIP Conference on Human-Computer Interaction (Interact2001) %D 2001 %T Enabling Commuters to Find the Best Route: An Interface for Analyzing Driving History Logs %A Konishi,M. %A Plaisant, Catherine %A Shneiderman, Ben %X This paper describes a prototype interface design for an automobile driving history log. It allows drivers to choose the best route among several alternatives for their common trips. Recorded data includes time to complete the travel, fuel consumption, and number of stops. %B Proceedings of the IFIP Conference on Human-Computer Interaction (Interact2001) %P 799 - 800 %8 2001/// %G eng %0 Journal Article %J Handbook of Process Algebra %D 2001 %T Equivalence and preorder checking for finite-state systems %A Cleaveland, Rance %A Sokolsky,O. %B Handbook of Process Algebra %P 391 - 424 %8 2001/// %G eng %0 Journal Article %J SURFACE, the Syracuse University Research Facility And Collaborative Environment %D 2001 %T E-tables: Non-specialist use and understanding of statistical data %A Marchionini,G. %A Hert,C. %A Shneiderman, Ben %A Liddy,L. %X This paper provides a progress report on a project that aims to understand how people think about and understand statistical data presented in tables. The focus is on people who are not statistical specialists. To this end, studies of various user populations were conducted, underlying data enrichment and explanation was investigated, and interactive user interface prototypes were developed and tested. %B SURFACE, the Syracuse University Research Facility And Collaborative Environment %P 33 - 33 %8 2001/// %G eng %0 Conference Paper %B Software Engineering Workshop, 2001. Proceedings. 26th Annual NASA Goddard %D 2001 %T An experience management system for a software engineering research organization %A Basili, Victor R. %A Costa,P. %A Lindvall,M. %A Mendonca,M. %A Seaman,C. %A Tesoriero,R. %A Zelkowitz, Marvin V %K and %K approach;knowledge %K base;experience %K Center %K control;knowledge %K data %K development %K dust-to-pearls %K engineering %K Engineering;business %K engineering;software %K experimental %K for %K Fraunhofer %K houses; %K knowledge;employee %K knowledge;experience %K level %K maintenance;knowledge %K management %K management;knowledge %K management;software %K mechanisms;software %K organization;software %K organizational %K organizations;administrative %K organizations;knowledge %K organizations;low-barrier %K ownership;knowledge-intensive %K processing;personnel;research %K research %K software %K system;human-intensive %X Most businesses rely on the fact that their employees possess relevant knowledge and that they can apply it to the task at hand. The problem is that this knowledge is not owned by the organization. It is owned and controlled by its employees. Maintaining an appropriate level of knowledge in the organization is a very important issue. It is, however, not an easy task for most organizations, and it is particularly problematic for software organizations, which are human- and knowledge-intensive. Knowledge management is a relatively new area that has attempted to address these problems. This paper introduces an approach called the "knowledge dust-to-pearls" approach. This approach addresses some of the issues with knowledge management by providing low-barrier mechanisms to "jump start" the experience base. This approach allows the experience base to become more useful more quickly than traditional approaches. This paper describes the approach and gives an example of its use at the Fraunhofer Center for Experimental Software Engineering, Maryland, USA %B Software Engineering Workshop, 2001. Proceedings. 26th Annual NASA Goddard %P 29 - 35 %8 2001/// %G eng %R 10.1109/SEW.2001.992652 %0 Book Section %B Algorithm EngineeringAlgorithm Engineering %D 2001 %T Experimental Analysis of Algorithms for Bilateral-Contract Clearing Mechanisms Arising in Deregulated Power Industry %A Barrett,Chris %A Cook,Doug %A Hicks,Gregory %A Faber,Vance %A Marathe,Achla %A Marathe,Madhav %A Srinivasan, Aravind %A Sussmann,Yoram %A Thornquist,Heidi %E Brodal,Gerth %E Frigioni,Daniele %E Marchetti-Spaccamela,Alberto %X We consider the bilateral contract satisfaction problem arising from electrical power networks due to the proposed deregulation of the electric utility industry in the USA. Given a network and a (multi)set of pairs of vertices (contracts) with associated demands, the goal is to find the maximum number of simultaneously satisfiable contracts. We study how four different algorithms perform in fairly realistic settings; we use an approximate electrical power network from Colorado. Our experiments show that three heuristics outperform a theoretically better algorithm. We also test the algorithms on four types of scenarios that are likely to occur in a deregulated marketplace. Our results show that the networks that are adequate in a regulated marketplace might be inadequate for satisfying all the bilateral contracts in a deregulated industry. %B Algorithm EngineeringAlgorithm Engineering %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2141 %P 172 - 184 %8 2001/// %@ 978-3-540-42500-7 %G eng %U http://dx.doi.org/10.1007/3-540-44688-5_14 %0 Journal Article %J University of Maryland Computer Science Technical Report (June 2001) %D 2001 %T Exploration of Large Online Data Tables Using Generalized Query Previews %A Tanin,E. %A Shneiderman, Ben %X Companies, government agencies, and other organizations are making their data available to the world over the Internet. These organizations often store their data as large online relational database tables. Users query these databases with front-ends that mostly use menus or form fillin interfaces, but these interfaces rarely give users information about the contents and distribution of the data. This situation leads users to waste time and network resources posing queries that have zero-hit or mega-hit results. Generalized query previews form a user interface architecture for efficient browsing of large online databases by supplying data distribution information to the users. The data distribution information provides an overview of the data. Generalized query previews give continuous feedback about the size of the results as the query is being formed. This provides a preview of the results. This paper presents the generalized query previews user interface architecture and our experimental findings. Our user study shows that for exploratory querying tasks, generalized query previews speed user performance and reduce network load. Categories and Subject Descriptors: H.5.2 [Information Interfaces and Presentation]: User Interfaces - Graphical user interfaces (GUI), H.3.5 [Information Storage and Retrieval]: Online Information Services - Web-based services. General Terms: Design, Human Factors, Experimentation. Keywords: Visual Data Mining, Information Visualization, Graphical User Interfaces, Relational Databases, Database Querying, Online Databases, Dynamic Queries, Previews and Overviews. %B University of Maryland Computer Science Technical Report (June 2001) %8 2001/// %G eng %0 Conference Paper %B Network Protocols, 2001. Ninth International Conference on %D 2001 %T Finding close friends on the Internet %A Kommareddy,C. %A Shankar,N. %A Bhattacharjee, Bobby %K application-layer; %K application-peers; %K Beaconing; %K beacons; %K close %K distance %K Expanding %K friends; %K Internet-like %K Internet; %K IP; %K measurement %K nearby %K network %K peer-location %K points; %K protocols; %K Ring %K searches; %K service; %K solutions; %K testbed; %K topologies; %K topology; %K transport %K Triangulation; %K unicast-only %K wide-area %X We consider the problem of finding nearby application-peers (close friends) over the Internet. We focus on unicast-only solutions and introduce a new scheme -Beaconing-for finding peers that are near. Our scheme uses distance measurement points (called beacons) and can be implemented entirely in the application-layer without investing in large infrastructure changes. We present an extensive evaluation of Beaconing and compare it to existing schemes including Expanding Ring searches and Triangulation. Our experiments show that 3-8 beacons are sufficient to provide efficient peer-location service on 10 000 node Internet-like topologies. Further, our results are 2-5 times more accurate than existing techniques. We also present results from an implementation of Beaconing over a non-trivial wide-area testbed. In our experiments, Beaconing is able to efficiently (< 3 K Bytes and < 50 packets on average), quickly (< 1 second on average), and accurately (< 20 ms error on average) find nearby peers on the Internet. %B Network Protocols, 2001. Ninth International Conference on %P 301 - 309 %8 2001/11// %G eng %R 10.1109/ICNP.2001.992910 %0 Conference Paper %B Proceedings of the thirteenth annual ACM symposium on Parallel algorithms and architectures %D 2001 %T Finding large independent sets of hypergraphs in parallel %A Shachnai,Hadas %A Srinivasan, Aravind %K hypergraphs %K independent sets %K Parallel algorithms %K randomized algorithms %X A basic problem in hypergraphs is that of finding a large independent set-one of guaranteed size-in a given hypergraph. Understanding the parallel complexity of this and related independent set problems on hypergraphs is a fundamental open issue in parallel computation. Caro and Tuza (J. Graph Theory, Vol. 15, pp. 99-107, 1991) have shown a certain lower bound &agr;k(H) on the size of a maximum independent set in a given k-uniform hypergraph H, and have also presented an efficien sequential algorithm to find an independent set of size &agr;k (H). They also show that &agr;k (H) is the size of the maximum independent set for various hypergraph families. Here, we develop the first RNC algorithm to find an independent set of size &agr;k(H), and also derandomize it for various special cases. We also present lower bounds on independent set size and corresponding RNC algorithms for non-uniform hypergraphs. %B Proceedings of the thirteenth annual ACM symposium on Parallel algorithms and architectures %S SPAA '01 %I ACM %C New York, NY, USA %P 163 - 168 %8 2001/// %@ 1-58113-409-6 %G eng %U http://doi.acm.org/10.1145/378580.378622 %R 10.1145/378580.378622 %0 Journal Article %J ACM Trans. Database Syst. %D 2001 %T Flexible support for multiple access control policies %A Jajodia,Sushil %A Samarati,Pierangela %A Sapino,Maria Luisa %A V.S. Subrahmanian %K access control policy %K Authorization %K logic programming %X Although several access control policies can be devised for controlling access to information, all existing authorization models, and the corresponding enforcement mechanisms, are based on a specific policy (usually the closed policy). As a consequence, although different policy choices are possible in theory, in practice only a specific policy can actually be applied within a given system. In this paper, we present a unified framework that can enforce multiple access control policies within a single system. The framework is based on a language through which users can specify security policies to be enforced on specific accesses. The language allows the specification of both positive and negative authorizations and incorporates notions of authorization derivation, conflict resolution, and decision strategies. Different strategies may be applied to different users, groups, objects, or roles, based on the needs of the security policy. The overall result is a flexible and powerful, yet simple, framework that can easily capture many of the traditional access control policies as well as protection requirements that exist in real-world applications, but are seldom supported by existing systems. The major advantage of our approach is that it can be used to specify different access control policies that can all coexist in the same system and be enforced by the same security server. %B ACM Trans. Database Syst. %V 26 %P 214 - 260 %8 2001/06// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/383891.383894 %N 2 %R 10.1145/383891.383894 %0 Conference Paper %B Mobile Data Management %D 2001 %T Hashing moving objects %A Song,Z. %A Roussopoulos, Nick %B Mobile Data Management %P 161 - 172 %8 2001/// %G eng %0 Journal Article %J Software Engineering, IEEE Transactions on %D 2001 %T Hierarchical GUI test case generation using automated planning %A Memon, Atif M. %A Pollack,M. E %A Soffa,M. L %K Artificial intelligence %K automated planning %K automatic test case generation %K Automatic testing %K correctness testing %K goal state %K Graphical user interfaces %K hierarchical GUI test case generation %K initial state %K Microsoft WordPad %K operators %K plan-generation system %K planning (artificial intelligence) %K Planning Assisted Tester for Graphical User Interface Systems %K program testing %K software %X The widespread use of GUIs for interacting with software is leading to the construction of more and more complex GUIs. With the growing complexity come challenges in testing the correctness of a GUI and its underlying software. We present a new technique to automatically generate test cases for GUIs that exploits planning, a well-developed and used technique in artificial intelligence. Given a set of operators, an initial state, and a goal state, a planner produces a sequence of the operators that will transform the initial state to the goal state. Our test case generation technique enables efficient application of planning by first creating a hierarchical model of a GUI based on its structure. The GUI model consists of hierarchical planning operators representing the possible events in the GUI. The test designer defines the preconditions and effects of the hierarchical operators, which are input into a plan-generation system. The test designer also creates scenarios that represent typical initial and goal states for a GUI user. The planner then generates plans representing sequences of GUI interactions that a user might employ to reach the goal state from the initial state. We implemented our test case generation system, called Planning Assisted Tester for Graphical User Interface Systems (PATHS) and experimentally evaluated its practicality and effectiveness. We describe a prototype implementation of PATHS and report on the results of controlled experiments to generate test cases for Microsoft's WordPad %B Software Engineering, IEEE Transactions on %V 27 %P 144 - 155 %8 2001/02// %@ 0098-5589 %G eng %N 2 %R 10.1109/32.908959 %0 Conference Paper %B Proceedings of the Eleventh SIAM Conference on Parallel Processing for Scientific Computing %D 2001 %T A hypergraph-based workload partitioning strategy for parallel data aggregation %A Chang,C. %A Kurc, T. %A Sussman, Alan %A Catalyurek,U. %A Saltz, J. %B Proceedings of the Eleventh SIAM Conference on Parallel Processing for Scientific Computing %8 2001/// %G eng %0 Journal Article %J Journal of Computer and System Sciences %D 2001 %T Improved Bounds on the Sample Complexity of Learning %A Li,Yi %A Long,Philip M. %A Srinivasan, Aravind %K agnostic learning %K empirical process theory %K machine learning %K PAC learning %K sample complexity %X We present a new general upper bound on the number of examples required to estimate all of the expectations of a set of random variables uniformly well. The quality of the estimates is measured using a variant of the relative error proposed by Haussler and Pollard. We also show that our bound is within a constant factor of the best possible. Our upper bound implies improved bounds on the sample complexity of learning according to Haussler's decision theoretic model. %B Journal of Computer and System Sciences %V 62 %P 516 - 527 %8 2001/05// %@ 0022-0000 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022000000917410 %N 3 %R 10.1006/jcss.2000.1741 %0 Conference Paper %B Information Visualisation, International Conference on %D 2001 %T Information Visualization: The Path from Innovation to Adoption %A Shneiderman, Ben %B Information Visualisation, International Conference on %I IEEE Computer Society %C Los Alamitos, CA, USA %P 0003 - 0003 %8 2001/// %@ 0-7695-1195-3 %G eng %R http://doi.ieeecomputersociety.org/10.1109/IV.2001.10004 %0 Journal Article %J Assembly Automation %D 2001 %T Intelligent assembly modeling and simulation %A Gupta,S.K. %A Paredis,C. J. J. %A Sinha,R. %X Because of the intense competition in the current global economy, a company must conceive, design, and manufacture new products quickly and inexpensively. The design cycle can be shortened through simulation. Rapid technical advances in many different areas of scientific computing provide the enabling technologies for creating a comprehensive simulation and visualization environment for assembly design and planning. An intelligent environment has been built in which simple simulation tools can be composed into complex simulations for detecting potential assembly problems. The goal in this research is to develop high fidelity assembly simulation and visualization tools that can detect assembly related problems without going through physical mock-ups. In addition, these tools can be used to create easy-to-visualize instructions for performing assembly and service operations. %B Assembly Automation %V 21 %P 215 - 235 %8 2001/// %G eng %U http://www.ingentaconnect.com/content/mcb/033/2001/00000021/00000003/art00004 %N 3 %0 Book Section %B Discovery ScienceDiscovery Science %D 2001 %T Interactive Exploration of Time Series Data %A Hochheiser,Harry %A Shneiderman, Ben %E Jantke,Klaus %E Shinohara,Ayumi %X Widespread interest in discovering features and trends in time- series has generated a need for tools that support interactive exploration.This paper introduces timeboxes: a powerful direct-manipulation metaphor for the specification of queries over time series datasets. Our TimeSearcher implementation of timeboxes supports interactive formulation and modification of queries, thus speeding the process of exploring time series data sets and guiding data mining. %B Discovery ScienceDiscovery Science %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2226 %P 441 - 446 %8 2001/// %@ 978-3-540-42956-2 %G eng %U http://dx.doi.org/10.1007/3-540-45650-3_38 %0 Conference Paper %D 2001 %T Intrusion tolerance approaches in ITUA %A Michel Cukier %A Lyons,J. %A Pandey,P. %A Ramasamy,H. V. %A Sanders,W. H. %A Pal,P. %A Webber,F. %A Schantz,R. %A Loyall,J. %A Watro,R. %V 64 %8 2001/// %G eng %U http://www.dist-systems.bbn.com/papers/2001/ICDSN/01CUK01.pdf %0 Book Section %B Discovery Science %D 2001 %T Inventing Discovery Tools: Combining Information Visualization with Data Mining %A Shneiderman, Ben %E Jantke,Klaus %E Shinohara,Ayumi %X The growing use of information visualization tools and data mining algorithms stems from two separate lines of research. Information visualization researchers believe in the importance of giving users an overview and insight into the data distributions, while data mining researchers believe that statistical algorithms and machine learning can be relied on to find the interesting patterns. This paper discusses two issues that influence design of discovery tools: statistical algorithms vs. visual data presentation, and hypothesis testing vs. exploratory data analysis. I claim that a combined approach could lead to novel discovery tools that preserve user control, enable more effective exploration, and promote responsibility. %B Discovery Science %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 2226 %P 17 - 28 %8 2001/// %@ 978-3-540-42956-2 %G eng %U http://dx.doi.org/10.1007/3-540-45650-3_4 %0 Journal Article %J Advances in Spatial and Temporal Databases %D 2001 %T K-nearest neighbor search for moving query point %A Song,Z. %A Roussopoulos, Nick %B Advances in Spatial and Temporal Databases %P 79 - 96 %8 2001/// %G eng %0 Conference Paper %B Proceedings of the 1st Workshop on Open Source Software Engineering %D 2001 %T Leveraging open-source communities to improve the quality & performance of open-source software %A Schmidt,D. C %A Porter, Adam %B Proceedings of the 1st Workshop on Open Source Software Engineering %8 2001/// %G eng %0 Conference Paper %B Proceedings of the first international conference on Human language technology research %D 2001 %T Mandarin-English Information (MEI): investigating translingual speech retrieval %A Meng,Helen %A Chen,Berlin %A Khudanpur,Sanjeev %A Levow,Gina-Anne %A Lo,Wai-Kit %A Oard, Douglas %A Schone,Patrick %A Tang,Karen %A Wang,Hsin-Min %A Wang,Jianqiang %K cross-language %K English-Chinese %K spoken document retrieval %X This paper describes the Mandarin-English Information (MEI) project, where we investigated the problem of cross-language spoken document retrieval (CL-SDR), and developed one of the first English-Chinese CL-SDR systems. Our system accepts an entire English news story (text) as query, and retrieves relevant Chinese broadcast news stories (audio) from the document collection. Hence this is a cross-language and cross-media retrieval task. We applied a multi-scale approach to our problem, which unifies the use of phrases, words and subwords in retrieval. The English queries are translated into Chinese by means of a dictionary-based approach, where we have integrated phrase-based translation with word-by-word translation. Untranslatable named entities are transliterated by a novel subword translation technique. The multi-scale approach can be divided into three subtasks -- multi-scale query formulation, multi-scale audio indexing (by speech recognition) and multi-scale retrieval. Experimental results demonstrate that the use of phrase-based translation and subword translation gave performance gains, and multi-scale retrieval outperforms word-based retrieval. %B Proceedings of the first international conference on Human language technology research %S HLT '01 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 1 - 7 %8 2001/// %G eng %U http://dx.doi.org/10.3115/1072133.1072202 %R 10.3115/1072133.1072202 %0 Book %D 2001 %T Matrix Algorithms: Eigensystems %A Stewart, G.W. %K eigenvalues %K Mathematics / Algebra / Linear %K Mathematics / Applied %K Mathematics / General %K Mathematics / Mathematical Analysis %K Mathematics / Matrices %K Mathematics / Numerical Analysis %K Matrices %K Medical / General %X This is the second volume in a projected five-volume survey of numerical linear algebra and matrix algorithms. It treats the numerical solution of dense and large-scale eigenvalue problems with an emphasis on algorithms and the theoretical background required to understand them. The notes and reference sections contain pointers to other methods along with historical comments. The book is divided into two parts: dense eigenproblems and large eigenproblems. The first part gives a full treatment of the widely used QR algorithm, which is then applied to the solution of generalized eigenproblems and the computation of the singular value decomposition. The second part treats Krylov sequence methods such as the Lanczos and Arnoldi algorithms and presents a new treatment of the Jacobi-Davidson method. These volumes are not intended to be encyclopedic, but provide the reader with the theoretical and practical background to read the research literature and implement or modify new algorithms. %I SIAM %8 2001/08/30/ %@ 9780898715033 %G eng %0 Journal Article %J DIMACS Mini-Workshop on Quality of Service Issues in the Internet %D 2001 %T Near-optimal design of MP S tunnels with shared recovery %A Fleischer,L. %A Meyerson,A. %A Saniee,I. %A Shepherd,FB %A Srinivasan, Aravind %X We describe the optimization problem associ-ated with the concurrent routing of demands with guar- anteed shared recovery in case of network failures. This problem arises in routing with protection in meshes and is known to be hard. We describe the problem in the context of the efficient design of restorable MP S tunnels in optical networks. The underlying design gives rise to a stochastic optimization problem that is equivalent to a (very) large- scale linear programming (LP) problem that explicitly in- corporates the network failure scenarios. The feasible re- gion for this LP is given by combined packing and cover- ing constraints for concurrent and optimal multicommodity flows. We develop a novel -approximation procedure for this problem and demonstrate its performance for a variety of real network sizes. An attraction of our approach is that its main computation consists of routing flow along a pair of short paths and these paths are easily found. Commer- cial general-purpose LP solvers are typically unable to solve these problems once they become large enough, while our approach scales for large networks. We conclude that the proposed scheme provides guaranteed approximation to the design of restorable MP S tunnels with shared protection within realistic network settings. %B DIMACS Mini-Workshop on Quality of Service Issues in the Internet %8 2001/// %G eng %0 Conference Paper %B Proceedings of the twelfth annual ACM-SIAM symposium on Discrete algorithms %D 2001 %T New approaches to covering and packing problems %A Srinivasan, Aravind %X Covering and packing integer programs model a large family of combinatorial optimization problems. The current-best approximation algorithms for these are an instance of the basic probabilistic method: showing that a certain randomized approach produces a good approximation with positive probability. This approach seems inherently sequential; by employing the method of alteration we present the first RNC and NC approximation algorithms that match the best sequential guarantees. Extending our approach, we get the first RNC and NC approximation algorithms for certain multi-criteria versions of these problems. We also present the first NC algorithms for two packing and covering problems that are not subsumed by the above result: finding large independent sets in graphs, and rounding fractional Group Steiner solutions on trees. %B Proceedings of the twelfth annual ACM-SIAM symposium on Discrete algorithms %S SODA '01 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 567 - 576 %8 2001/// %@ 0-89871-490-7 %G eng %U http://dl.acm.org/citation.cfm?id=365411.365535 %0 Journal Article %J IEEE Transactions on Information Theory %D 2001 %T The one-inclusion graph algorithm is near-optimal for the prediction model of learning %A Li,Yi %A Long,P. M %A Srinivasan, Aravind %K Computer science %K concept class %K general-purpose algorithm %K graph theory %K learning prediction model %K learning systems %K near-optimal algorithm %K Neural networks %K one-inclusion graph algorithm %K optimisation %K Prediction algorithms %K prediction theory %K Predictive models %K probability %K Probability distribution %K Upper bound %K Vapnik-Chervonenkis dimension %K Virtual colonoscopy %X Haussler, Littlestone and Warmuth (1994) described a general-purpose algorithm for learning according to the prediction model, and proved an upper bound on the probability that their algorithm makes a mistake in terms of the number of examples seen and the Vapnik-Chervonenkis (VC) dimension of the concept class being learned. We show that their bound is within a factor of 1+o(1) of the best possible such bound for any algorithm %B IEEE Transactions on Information Theory %V 47 %P 1257 - 1261 %8 2001/03// %@ 0018-9448 %G eng %N 3 %R 10.1109/18.915700 %0 Conference Paper %B KI 2001: advances in artificial intelligence: Joint German/Austrian Conference on AI, Vienna, Austria, September 19-21, 2001: proceedings %D 2001 %T Optimal Agent Selection %A Ozcan,F. %A V.S. Subrahmanian %A Golubchik,L. %X The Internet contains a vast array of sources that provide identical or similar services. When an agent needs to solve a problem, it may split the problem into “subproblems” and find an agent to solve each of the subproblems. Later, it may combine the results of these subproblems to solve the original problem. In this case, the agent is faced with the task of determining to which agents to assign the subproblems.We call this the agent selection problem (ASP for short). Solving ASP is complex because it must take into account several different parameters. For instance, different agents might take different amounts of time to process a request. Different agents might provide varying “qualities” of answers. Network latencies associated with different agents might vary. In this paper, we first formalize the agent selection problem and show that it is NP-hard. We then propose a generic cost function that is general enough to take into account the costs of (i) network and server loads, (ii) source computations, and (iii) internal mediator costs. We then develop exact and heuristic based algorithms to solve the agent selection problem. %B KI 2001: advances in artificial intelligence: Joint German/Austrian Conference on AI, Vienna, Austria, September 19-21, 2001: proceedings %P 2 - 2 %8 2001/// %G eng %0 Conference Paper %B IEEE Symposium on Information Visualization, 2001. INFOVIS 2001 %D 2001 %T Ordered treemap layouts %A Shneiderman, Ben %A Wattenberg,M. %K Clustering algorithms %K Computer science %K Data visualization %K Displays %K Electronic switching systems %K Filling %K Monte Carlo methods %K Read only memory %K Testing %B IEEE Symposium on Information Visualization, 2001. INFOVIS 2001 %I IEEE %P 73 - 78 %8 2001/// %@ 0-7695-7342-5 %G eng %R 10.1109/INFVIS.2001.963283 %0 Conference Paper %B INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE %D 2001 %T Partitioning activities for agents %A Ozcan,F. %A V.S. Subrahmanian %X There are now numerous agent applications thattrack interests of thousands of users in situations where changes occur continuously. [Shim et al., 1994] suggested that such agents can be made effi- cient by merging commonalities in their activities. However, past algorithms cannot merge more than 10 or 20 concurrent activities. We develop tech- niques so that a large number of concurrent activ- ities (typically over 1000) can be partitioned into components (groups of activities) of small size (e.g. 10 to 50) so that each component’s activities can be merged using previously developed algorithms (e.g. [Shim et al., 1994]). We first formalize the prob- lem and show that finding optimal partitions is NP- hard. We then develop three algorithms - Greedy, A -based and BAB (branch and bound). A -based and BAB are both guaranteed to compute optimal solutions. Greedy on the other hand uses heuris- tics and typically finds suboptimal solutions. We implemented all three algorithms. We experimen- tally show that the greedy algorithm finds partitions whose costs are at most 14% worse than that found by A -based and BAB — however, Greedy is able to handle over thousand concurrent requests very fast while the other two methods are much slower and able to handle only 10-20 requests. Hence, Greedy appears to be the best. %B INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE %V 17 %P 1218 - 1228 %8 2001/// %G eng %0 Conference Paper %B Active Middleware Services, 2001. Third Annual International Workshop on %D 2001 %T Performance optimization for data intensive grid applications %A Beynon,M. D %A Sussman, Alan %A Catalyurek,U. %A Kurc, T. %A Saltz, J. %B Active Middleware Services, 2001. Third Annual International Workshop on %P 97 - 105 %8 2001/// %G eng %0 Journal Article %J Journal of Communications and Networks %D 2001 %T The price of safety in an active network %A Alexander,D. S %A Menage,P. B %A Keromytis,A. D %A Arbaugh, William A. %A Anagnostakis,K. G %A Smith,J. M %B Journal of Communications and Networks %V 3 %P 4 - 18 %8 2001/// %G eng %N 1 %0 Journal Article %J IJCAI workshop on text learning: beyond supervision %D 2001 %T Probabilistic models of text and link structure for hypertext classification %A Getoor, Lise %A Segal,E. %A Taskar,B. %A Koller,D. %X Most text classification methods treat each document as anindependent instance. However, in many text domains, doc- uments are linked and the topics of linked documents are cor- related. For example, web pages of related topics are often connected by hyperlinks and scientific papers from related fields are commonly linked by citations. We propose a unified probabilistic model for both the textual content and the link structure of a document collection. Our model is based on the recently introduced framework of Probabilistic Relational Models (PRMs), which allows us to capture cor- relations between linked documents. We show how to learn these models from data and use them efficiently for classifi- cation. Since exact methods for classification in these large models are intractable, we utilize belief propagation, an ap- proximate inference algorithm. Belief propagation automat- ically induces a very natural behavior, where our knowledge about one document helps us classify related ones, which in turn help us classify others. We present preliminary empiri- cal results on a dataset of university web pages. %B IJCAI workshop on text learning: beyond supervision %P 24 - 29 %8 2001/// %G eng %0 Journal Article %J ACM Trans. Database Syst. %D 2001 %T Probabilistic object bases %A Eiter,Thomas %A Lu,James J. %A Lukasiewicz,Thomas %A V.S. Subrahmanian %K consistency %K object-oriented database %K probabilistic object algebra %K probabilistic object base %K probability %K query language %K Query optimization %X Although there are many applications where an object-oriented data model is a good way of representing and querying data, current object database systems are unable to handle objects whose attributes are uncertain. In this article, we extend previous work by Kornatzky and Shimony to develop an algebra to handle object bases with uncertainty. We propose concepts of consistency for such object bases, together with an NP-completeness result, and classes of probabilistic object bases for which consistency is polynomially checkable. In addition, as certain operations involve conjunctions and disjunctions of events, and as the probability of conjunctive and disjunctive events depends both on the probabilities of the primitive events involved as well as on what is known (if anything) about the relationship between the events, we show how all our algebraic operations may be performed under arbitrary probabilistic conjunction and disjunction strategies. We also develop a host of equivalence results in our algebra, which may be used as rewrite rules for query optimization. Last but not least, we have developed a prototype probabilistic object base server on top of ObjectStore. We describe experiments to assess the efficiency of different possible rewrite rules. %B ACM Trans. Database Syst. %V 26 %P 264 - 312 %8 2001/09// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/502030.502031 %N 3 %R 10.1145/502030.502031 %0 Journal Article %J ACM Transaction on Database Systems %D 2001 %T Probabilistic temporal databases %A Alex,D. %A Robert,R. %A V.S. Subrahmanian %B ACM Transaction on Database Systems %V 26 %P 41 - 95 %8 2001/// %G eng %N 1 %0 Journal Article %J ACM Trans. Database Syst. %D 2001 %T Probabilistic temporal databases, I: algebra %A Dekhtyar,Alex %A Ross,Robert %A V.S. Subrahmanian %X Dyreson and Snodgrass have drawn attention to the fact that, in many temporal database applications, there is often uncertainty about the start time of events, the end time of events, and the duration of events. When the granularity of time is small (e.g., milliseconds), a statement such as “Packet p was shipped sometime during the first 5 days of January, 1998” leads to a massive amount of uncertainty (5×24×60×60×1000) possibilities. As noted in Zaniolo et al. [1997], past attempts to deal with uncertainty in databases have been restricted to relatively small amounts of uncertainty in attributes. Dyreson and Snodgrass have taken an important first step towards solving this problem.In this article, we first introduce the syntax of Temporal-Probabilistic (TP) relations and then show how they can be converted to an explicit, significantly more space-consuming form, called Annotated Relations. We then present a theoretical annotated temporal algebra (TATA). Being explicit, TATA is convenient for specifying how the algebraic operations should behave, but is impractical to use because annotated relations are overwhelmingly large. Next, we present a temporal probabilistic algebra (TPA). We show that our definition of the TP-algebra provides a correct implementation of TATA despite the fact that it operates on implicit, succinct TP-relations instead of overwhemingly large annotated relations. Finally, we report on timings for an implementation of the TP-Algebra built on top of ODBC. %B ACM Trans. Database Syst. %V 26 %P 41 - 95 %8 2001/03// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/383734.383736 %N 1 %R 10.1145/383734.383736 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-2001-79 %D 2001 %T Probabilistic Temporal Databases, II: Calculus and Query Processing %A Dekhtyar,A. %A Ozcan,F. %A Ross,R. %A V.S. Subrahmanian %X There is a vast class of applications in which we know that a certain event occurred, but do not know exactly when it occurred. However, as studied by Dyreson and Snodgrass \cite{ds98}, there are many natural scenarios where probability distributions exist and quantify this uncertainty. Dekhtyar et. al. extended Dyreson and Snodgrass's work and defined an extension of the relational algebra to handle such data. The first contribution of this paper is a declarative temporal probabilistic (TP for short) calculus which we show is equivalent in expressive power to the temporal probabilistic algebra of Dekhtyar et. al. Our second major contribution is a set of equivalence and containment results for the TP-algebra. Our third contribution is the development of cost models that may be used to estimate the cost of TP-algebra operations. Our fourth contribution is an experimental evaluation of the accuracy of our cost models and the use of the equivalence results as rewrite rules for optimizing queries by using an implementation of TP-databases on top of ODBC. %B Technical Reports from UMIACS, UMIACS-TR-2001-79 %8 2001/// %G eng %0 Conference Paper %B Thirteenth International Conference on Software Engineering and Knowledge Engineering %D 2001 %T A prototype experience management system for a software consulting organization %A Mendonça Neto,M. G. %A Seaman,C. %A Basili, Victor R. %A Kim,Y. M %X The Experience Management System (EMS) is aimed atsupporting the capture and reuse of software-related experience, based on the Experience Factory concept. It is being developed for use in a multinational software engineering consultancy, Q-Labs. Currently, a prototype EMS exists and has been evaluated. This paper focuses on the EMS architecture, underlying data model, implementation, and user interface. %B Thirteenth International Conference on Software Engineering and Knowledge Engineering %P 29 - 36 %8 2001/// %G eng %0 Conference Paper %B E-commerce Security and Privacy %D 2001 %T Provisional authorizations %A Jajodia,S. %A Kudo,M. %A V.S. Subrahmanian %X Past generations of access control systems, when faced with an access request, have issued a “yes” (resp. “no”) answer to the access request resulting in access being granted (resp. denied). In this chapter, we ar­gue that for the world’s rapidly proliferating business to business (B2B) applications and auctions, “yes/no” responses are just not enough. We propose the notion of a “provisional authorization” which intuitively says “You may perform the desired access provided you cause condition C to be satisfied.” For instance, a user accessing an online brokerage may receive some information if he fills out his name/address, but not otherwise. While a variety of such provisional authorization mecha­nisms exist on the web, they are all hardcoded on an application by application basis. We show that given (almost) any logic L, we may define a provisional authorization specification language pASLL. pASLL is based on the declarative, polynomially evaluable authorization spec­ification language ASL proposed by Jajodia et al [JSS97]. We define programs in pASLL, and specify how given any access request, we must find a “weakest” precondition under which the access can be granted (in the worst case, if this weakest precondition is “false” this amounts to a denial). We develop a model theoretic semantics for pASLL and show how it can be applied to online sealed-bid auction servers and online contracting. %B E-commerce Security and Privacy %V 2 %P 133 - 159 %8 2001/// %G eng %R 10.1007/978-1-4615-1467-1_8 %0 Journal Article %J Proceedings of the 10th USENIX Security Symposium %D 2001 %T Reading between the lines: Lessons from the SDMI challenge %A Craver,S.A. %A Wu,M. %A Liu,B. %A Stubblefield,A. %A Swartzlander,B. %A Wallach,D.S. %A Dean,D. %A Felten,E.W. %X The Secure Digital Music Initiative is a consortium ofparties interested in preventing piracy of digital music, and to this end they are developing architectures for con- tent protection on untrusted platforms. SDMI recently held a challenge to test the strength of four watermark- ing technologies, and two other security technologies. No documentation explained the implementations of the technologies, and neither watermark embedding nor de- tecting software was directly accessible to challenge par- ticipants. We nevertheless accepted the challenge, and explored the inner workings of the technologies. We re- port on our results here. %B Proceedings of the 10th USENIX Security Symposium %P 13 - 17 %8 2001/// %G eng %0 Journal Article %J Biometrics %D 2001 %T Relating amino acid sequence to phenotype: analysis of peptide-binding data %A Segal,M. R %A Cummings, Michael P. %A Hubbard,A. E %X We illustrate data analytic concerns that arise in the context of relating genotype, as represented by amino acid sequence, to phenotypes (outcomes). The present application examines whether peptides that bind to a particular major histocompatibility complex (MHC) class I molecule have characteristic amino acid sequences. However, the concerns identified and addressed are considerably more general. It is recognized that simple rules for predicting binding based solely on preferences for specific amino acids in certain (anchor) positions of the peptide's amino acid sequence are generally inadequate and that binding is potentially influenced by all sequence positions as well as between-position interactions. The desire to elucidate these more complex prediction rules has spawned various modeling attempts, the shortcomings of which provide motivation for the methods adopted here. Because of (i) this need to model between-position interactions, (ii) amino acids constituting a highly (20) multilevel unordered categorical covariate, and (iii) there frequently being numerous such covariates (i.e., positions) comprising the sequence, standard regression/classification techniques are problematic due to the proliferation of indicator variables required for encoding the sequence position covariates and attendant interactions. These difficulties have led to analyses based on (continuous) properties (e.g., molecular weights) of the amino acids. However, there is potential information loss in such an approach if the properties used are incomplete and/or do not capture the mechanism underlying association with the phenotype. Here we demonstrate that handling unordered categorical covariates with numerous levels and accompanying interactions can be done effectively using classification trees and recently devised bump-hunting methods. We further tackle the question of whether observed associations are attributable to amino acid properties as well as addressing the assessment and implications of between-position covariation. %B Biometrics %V 57 %P 632 - 642 %8 2001/06// %G eng %N 2 %0 Journal Article %J Journal of Molecular Biology %D 2001 %T Rescuing a destabilized protein fold through backbone cyclization %A Camarero,Julio A %A Fushman, David %A Sato,Satoshi %A Giriat,Izabela %A Cowburn,David %A Raleigh,Daniel P %A Muir,Tom W %K circular protein %K ligation %K SH3 domain %X We describe the physicochemical characterization of various circular and linear forms of the ∼60 residue N-terminal Src homology 3 (SH3) domain from the murine c-Crk adapter protein. Structural, dynamic, thermodynamic, kinetic and biochemical studies reveal that backbone circularization does not prevent the adoption of the natural folded structure in any of the circular proteins. Both the folding and unfolding rate of the protein increased slightly upon circularization. Circularization did not lead to a significant thermodynamic stabilization of the full-length protein, suggesting that destabilizing enthalpic effects (e.g. strain) negate the expected favorable entropic contribution to overall stability. In contrast, we find circularization results in a dramatic stabilization of a truncated version of the SH3 domain lacking a key glutamate residue. The ability to rescue the destabilized mutant indicates that circularization may be a useful tool in protein engineering programs geared towards generating minimized proteins. %B Journal of Molecular Biology %V 308 %P 1045 - 1062 %8 2001/05/18/ %@ 0022-2836 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022283601946315 %N 5 %R 10.1006/jmbi.2001.4631 %0 Journal Article %J International Journal of Speech Technology %D 2001 %T The role of a natural language conversational interface in online sales: A case study %A Chai,J. %A Jimmy Lin %A Zadrozny,W. %A Ye,Y. %A Stys-Budzikowska,M. %A Horvath,V. %A Kambhatla,N. %A Wolf,C. %X This paper describes the evaluation of a natural language dialog-based navigation system (HappyAssistant) that helps users access e-commerce sites to find relevant information about products and services. The prototype system leverages technologies in natural language processing and human-computer interaction to create a faster and more intuitive way of interacting with websites, especially for less experienced users. The result of a comparative study shows that users prefer the natural language-enabled navigation two to one over the menu driven navigation. In addition, the study confirmed the efficiency of using natural language dialog in terms of the number of clicks and the amount of time required to obtain the relevant information. In the case study, as compared to the menu driven system, the average number of clicks used in the natural language system was reduced by 63.2% and the average time was reduced by 33.3%. %B International Journal of Speech Technology %V 4 %P 285 - 295 %8 2001/// %G eng %N 3 %R 10.1023/A:1011316909641 %0 Patent %D 2001 %T Rotation, scale, and translation resilient public watermarking for images using a log-polar fourier transform %A Bloom,Jeffrey A. %A Cox,Ingemar J. %A Miller,Matthew L. %A M. Wu %A Lin,Ching-Yung %A Lui,Yui Man %E Signafy, Inc. %X A method for detecting a watermark signal in digital image data. The detecting method includes the steps of: computing a log-polar Fourier transform of the image data to obtain a log-polar Fourier spectrum; projecting the log-polar Fourier spectrum down to a lower dimensional space to obtain an extracted signal; comparing the extracted signal to a target watermark signal; and declaring the presence or absence of the target watermark signal in the image data based on the comparison. Also provided is a method for inserting a watermark signal in digital image data to obtain a watermarked image. The inserting method includes the steps of: computing a log-polar Fourier transform of the image data to obtain a log-polar Fourier spectrum; projecting the log-polar Fourier spectrum down to a lower dimensional space to obtain an extracted signal; modifying the extracted signal such that it is similar to a target watermark; performing a one-to-many mapping of the modified signal back to... %8 2001/08/28/ %G eng %U http://www.google.com/patents?id=v4AIAAAAEBAJ %N 6282300 %0 Journal Article %J Proceedings of the AMIA SymposiumProc AMIA Symp %D 2001 %T Secure Health Information Sharing System: SHARE %A M. Wu %A Mui,Lik %A Mohtashemi,Mojdeh %A Szolovits,Peter %B Proceedings of the AMIA SymposiumProc AMIA Symp %P 1060 - 1060 %8 2001/// %@ 1531-605X %G eng %0 Journal Article %J Data Warehousing and Knowledge Discovery %D 2001 %T Shared index scans for data warehouses %A Kotidis1⋆,Y. %A Sismanis,Y. %A Roussopoulos, Nick %B Data Warehousing and Knowledge Discovery %P 307 - 316 %8 2001/// %G eng %0 Journal Article %J Computer Vision and Image Understanding %D 2001 %T The Statistics of Optical Flow %A Fermüller, Cornelia %A Shulman,David %A Aloimonos, J. %X When processing image sequences some representation of image motion must be derived as a first stage. The most often used representation is the optical flow field, which is a set of velocity measurements of image patterns. It is well known that it is very difficult to estimate accurate optical flow at locations in an image which correspond to scene discontinuities. What is less well known, however, is that even at the locations corresponding to smooth scene surfaces, the optical flow field often cannot be estimated accurately.Noise in the data causes many optical flow estimation techniques to give biased flow estimates. Very often there is consistent bias: the estimate tends to be an underestimate in length and to be in a direction closer to the majority of the gradients in the patch. This paper studies all three major categories of flow estimation methods—gradient-based, energy-based, and correlation methods, and it analyzes different ways of compounding one-dimensional motion estimates (image gradients, spatiotemporal frequency triplets, local correlation estimates) into two-dimensional velocity estimates, including linear and nonlinear methods. Correcting for the bias would require knowledge of the noise parameters. In many situations, however, these are difficult to estimate accurately, as they change with the dynamic imagery in unpredictable and complex ways. Thus, the bias really is a problem inherent to optical flow estimation. We argue that the bias is also integral to the human visual system. It is the cause of the illusory perception of motion in the Ouchi pattern and also explains various psychophysical studies of the perception of moving plaids. %B Computer Vision and Image Understanding %V 82 %P 1 - 32 %8 2001/04// %@ 1077-3142 %G eng %U http://www.sciencedirect.com/science/article/pii/S1077314200909007 %N 1 %R 10.1006/cviu.2000.0900 %0 Conference Paper %B Proceedings of the Second International Workshop on Evaluating Word Sense Disambiguation Systems (SENSEVAL-2) %D 2001 %T Supervised sense tagging using support vector machines %A Cabezas,C. %A Resnik, Philip %A Stevens,J. %B Proceedings of the Second International Workshop on Evaluating Word Sense Disambiguation Systems (SENSEVAL-2) %P 59 - 62 %8 2001/// %G eng %0 Journal Article %J Human-Centered Computing, Online Communities, and Virtual Environments %D 2001 %T Supporting creativity with advanced information-abundant user interfaces %A Shneiderman, Ben %B Human-Centered Computing, Online Communities, and Virtual Environments %P 469 - 480 %8 2001/// %G eng %0 Journal Article %J Artificial Intelligence %D 2001 %T Temporal agent programs %A Dix,Jürgen %A Kraus,Sarit %A V.S. Subrahmanian %K agents %K Heterogenous systems %K Information integration %K logic programming %K temporal reasoning %X The “agent program” framework introduced by Eiter, Subrahmanian and Pick [Artificial Intelligence 108 (1–2) (1999) 179], supports developing agents on top of arbitrary legacy code. Such agents are continuously engaged in an “event occurs→think→act→event occurs…” cycle. However, this framework has two major limitations: (1)all actions are assumed to have no duration, and (2) all actions are taken now, but cannot be scheduled for the future. In this paper, we present the concept of a “temporal agent program” (tap for short) and show that using taps, it is possible to build agents on top of legacy code that can reason about the past and about the future, and that can make temporal commitments for the future now. We develop a formal semantics for such agents, extending the concept of a status set proposed by Eiter et al., and develop algorithms to compute the status sets associated with temporal agent programs. Last, but not least, we show how taps support the decision making of collaborative agents. %B Artificial Intelligence %V 127 %P 87 - 135 %8 2001/03// %@ 0004-3702 %G eng %U http://www.sciencedirect.com/science/article/pii/S0004370201000522 %N 1 %R 10.1016/S0004-3702(01)00052-2 %0 Conference Paper %B Deformable avatars: IFIP TC5/WG5. 10 DEFORM'2000 Workshop, November 29-30, 2000, Geneva, Switzerland and AVATARS'2000 Workshop, November 30-December 1, 2000, Lausanne, Switzerland %D 2001 %T Towards the ultimate motion capture technology %A Stuart, B. %A Baker, P. %A Aloimonos, J. %B Deformable avatars: IFIP TC5/WG5. 10 DEFORM'2000 Workshop, November 29-30, 2000, Geneva, Switzerland and AVATARS'2000 Workshop, November 30-December 1, 2000, Lausanne, Switzerland %V 68 %P 143 - 143 %8 2001/// %G eng %0 Book Section %B Frontiers of human-centred computing, online communities and virtual environmentsFrontiers of human-centred computing, online communities and virtual environments %D 2001 %T Universal usability: a research agenda for human—computer interaction to empower every citizen %A Shneiderman, Ben %B Frontiers of human-centred computing, online communities and virtual environmentsFrontiers of human-centred computing, online communities and virtual environments %I Springer-Verlag %C London, UK %P 179 - 189 %8 2001/// %@ 1-85233-238-7 %G eng %0 Journal Article %J Behaviour & Information Technology %D 2001 %T Universal usability as a stimulus to advanced interface design %A Shneiderman, Ben %A Hochheiser,Harry %X The desire to make computing available to broader populations has historically been a motivation for research and innovation that led to new breakthroughs in usability. Menus, graphical user interfaces and the World Wide Web are examples of innovative technological solutions that have arisen out of the challenge of bringing larger and more diverse groups of users into the world of computing. Universal usability is the latest such challenge: In order to build systems that are universally usable, designers must account for technology variety, user diversity and gaps in user knowledge. These issues are particularly challenging and important in the context of increasing the usability of the World Wide Web. To raise awareness, web designers are urged to provide universal usability statements that offer users information about the usability of their sites. These statements can inform users and thereby reduce frustration and confusion. Further steps toward universal usability might be achieved through research aimed at developing tools that would encourage or promote usability. The paper closes with five proposals for future research.The desire to make computing available to broader populations has historically been a motivation for research and innovation that led to new breakthroughs in usability. Menus, graphical user interfaces and the World Wide Web are examples of innovative technological solutions that have arisen out of the challenge of bringing larger and more diverse groups of users into the world of computing. Universal usability is the latest such challenge: In order to build systems that are universally usable, designers must account for technology variety, user diversity and gaps in user knowledge. These issues are particularly challenging and important in the context of increasing the usability of the World Wide Web. To raise awareness, web designers are urged to provide universal usability statements that offer users information about the usability of their sites. These statements can inform users and thereby reduce frustration and confusion. Further steps toward universal usability might be achieved through research aimed at developing tools that would encourage or promote usability. The paper closes with five proposals for future research. %B Behaviour & Information Technology %V 20 %P 367 - 376 %8 2001/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449290110083602 %N 5 %R 10.1080/01449290110083602 %0 Journal Article %J interactions %D 2001 %T Universal usability statements: Marking the trail for all users %A Hochheiser,Harry %A Shneiderman, Ben %B interactions %V 8 %P 16 - 18 %8 2001/03// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/361897.361913 %N 2 %R 10.1145/361897.361913 %0 Journal Article %J Fifth International Conference on Information Visualisation: 25-27 July 2001 London, England: Proceedings %D 2001 %T University of Maryland, USA %A Shneiderman, Ben %B Fifth International Conference on Information Visualisation: 25-27 July 2001 London, England: Proceedings %P 3 - 3 %8 2001/// %G eng %0 Conference Paper %B PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON VERY LARGE DATA BASES %D 2001 %T Update propagation strategies for improving the quality of data on the web %A Labrinidis,A. %A Roussopoulos, Nick %A SCIENCE,MARYLAND UNIV COLLEGE PARK DEPT OF COMPUTER %B PROCEEDINGS OF THE INTERNATIONAL CONFERENCE ON VERY LARGE DATA BASES %P 391 - 400 %8 2001/// %G eng %0 Journal Article %J Journal of the American Society for Information Science and Technology %D 2001 %T Using interactive visualizations of WWW log data to characterize access patterns and inform site design %A Hochheiser,Harry %A Shneiderman, Ben %X HTTP server log files provide Web site operators with substantial detail regarding the visitors to their sites. Interest in interpreting this data has spawned an active market for software packages that summarize and analyze this data, providing histograms, pie graphs, and other charts summarizing usage patterns. Although useful, these summaries obscure useful information and restrict users to passive interpretation of static displays. Interactive visualizations can be used to provide users with greater abilities to interpret and explore Web log data. By combining two-dimensional displays of thousands of individual access requests, color, and size coding for additional attributes, and facilities for zooming and filtering, these visualizations provide capabilities for examining data that exceed those of traditional Web log analysis tools. We introduce a series of interactive visualizations that can be used to explore server data across various dimensions. Possible uses of these visualizations are discussed, and difficulties of data collection, presentation, and interpretation are explored. %B Journal of the American Society for Information Science and Technology %V 52 %P 331 - 343 %8 2001/01/01/ %@ 1532-2890 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/1532-2890(2000)9999:9999%3C::AID-ASI1066%3E3.0.CO;2-Y/abstract %N 4 %R 10.1002/1532-2890(2000)9999:9999<::AID-ASI1066>3.0.CO;2-Y %0 Journal Article %J Proceedings of discovery science %D 2001 %T Visual specification of queries for finding patterns in time-series data %A Hochheiser,H. %A Shneiderman, Ben %X Widespread interest in discovering features and trends in time- series has generated a need for tools that support interactive exploration.This paper introduces timeboxes: a powerful graphical, direct-manipulation metaphor for the specification of queries over time-series datasets. Our TimeFinder implementation of timeboxes supports interactive formulation and modification of queries, thus speeding the process of exploring time-series data sets and guiding data mining. TimeFinder includes windows for timebox queries, individual time-series, and details-on-demand. Other features include drag-and-drop support for� query-by-example and graphical envelopes for displaying the extent of the entire data set and result set from a given query. Extensions involving increased expressive power and general temporal data sets are discussed. %B Proceedings of discovery science %P 441 - 446 %8 2001/// %G eng %0 Journal Article %J IEEE Computer Graphics and Applications %D 2001 %T Visualization of large data sets with the Active Data Repository %A Kurc, T. %A Catalyurek,U. %A Chang,Chialin %A Sussman, Alan %A Saltz, J. %K active data repository %K ADR runtime system %K Algorithm design and analysis %K application-specific processing %K C++ %K data mining %K data retrieval %K data visualisation %K Data visualization %K distributed-memory parallel machines %K Indexing %K Information retrieval %K isosurface rendering %K Isosurfaces %K large data sets %K large-scale multidimensional data %K Memory management %K modular services %K out-of-core data sets %K parallel machine %K Parallel machines %K Partitioning algorithms %K ray-casting-based volume rendering %K Rendering (computer graphics) %K Runtime %K software libraries %K storage management %X We implement ray-casting-based volume rendering and isosurface rendering methods using the Active Data Repository (ADR) for visualizing out-of-core data sets. We have developed the ADR object-oriented framework to provide support for applications that employ range queries with user-defined mapping and aggregation operations on large-scale multidimensional data. ADR targets distributed-memory parallel machines with one or more disks attached to each node. It is designed as a set of modular services implemented in C++, which can be customized for application-specific processing. The ADR runtime system supports common operations such as memory management, data retrieval, and scheduling of processing across a parallel machine %B IEEE Computer Graphics and Applications %V 21 %P 24 - 33 %8 2001/08//Jul %@ 0272-1716 %G eng %N 4 %R 10.1109/38.933521 %0 Journal Article %J IEEE Data Engineering Bulletin %D 2000 %T Adaptive query processing: Technology in evolution %A Hellerstein,J. M %A Franklin,M.J. %A Chandrasekaran,S. %A Deshpande, Amol %A Hildrum,K. %A Madden,S. %A Raman,V. %A Shah,M. A %X As query engines are scaled and federated, they must cope with highly unpredictable and changeableenvironments. In the Telegraph project, we are attempting to architect and implement a continuously adaptive query engine suitable for global-area systems, massive parallelism, and sensor networks. To set the stage for our research, we present a survey of prior work on adaptive query processing, focusing on three characterizations of adaptivity: the frequency of adaptivity, the effects of adaptivity, and the extent of adaptivity. Given this survey, we sketch directions for research in the Telegraph project. %B IEEE Data Engineering Bulletin %V 23 %P 7 - 18 %8 2000/// %G eng %N 2 %0 Journal Article %J Technical Reports (CIS) %D 2000 %T Agents in Network Management %A Ertugay,Osman %A Hicks, Michael W. %A Smith,Jonathan %A Kornblum,Jessica %X The ubiquity and complexity of modern networks require automated management and control. With increases in scale, automated solutions based on simple data access models such as SNMP will give way to more distributed and algorithmic techniques. This article outlines present and near-term solutions based on the ideas of active networks and mobile agents, which permit sophisticated programmable control and management of ultra large scale networks. %B Technical Reports (CIS) %8 2000/02/26/ %G eng %U http://repository.upenn.edu/cis_reports/138 %0 Conference Paper %B Proceedings of the sixteenth annual symposium on Computational geometry %D 2000 %T The analysis of a simple k-means clustering algorithm %A Kanungo,Tapas %A Mount, Dave %A Netanyahu,Nathan S. %A Piatko,Christine %A Silverman,Ruth %A Wu,Angela Y. %B Proceedings of the sixteenth annual symposium on Computational geometry %S SCG '00 %I ACM %C New York, NY, USA %P 100 - 109 %8 2000/// %@ 1-58113-224-7 %G eng %U http://doi.acm.org/10.1145/336154.336189 %R 10.1145/336154.336189 %0 Book Section %B Algebraic Frames for the Perception-Action CycleAlgebraic Frames for the Perception-Action Cycle %D 2000 %T Analyzing Action Representations %A Aloimonos, J. %A Fermüller, Cornelia %E Sommer,Gerald %E Zeevi,Yehoshua %X We argue that actions represent the basic seed of intelligence underlying perception of the environment, and the representations encoding actions should be the starting point upon which further studies of cognition are built. In this paper we make a first effort in characterizing these action representations. In particular, from the study of simple actions related to 3D rigid motion interpretation, we deduce a number of principles for the possible computations responsible for the interpretation of space-time geometry. Using these principles, we then discuss possible avenues on how to proceed in analyzing the representations of more complex human actions. %B Algebraic Frames for the Perception-Action CycleAlgebraic Frames for the Perception-Action Cycle %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1888 %P 1 - 21 %8 2000/// %@ 978-3-540-41013-3 %G eng %U http://dx.doi.org/10.1007/10722492_1 %0 Journal Article %J Information Processing Letters %D 2000 %T Approximating low-congestion routing and column-restricted packing problems %A Baveja,Alok %A Srinivasan, Aravind %K algorithms %K Approximation algorithms %K integer programming %K Packing %K Routing %X We contribute to a body of research asserting that the fractional and integral optima of column-sparse integer programs are “nearby”. This yields improved approximation algorithms for some generalizations of the knapsack problem, with applications to low-congestion routing in networks, file replication in distributed databases, and other packing problems. %B Information Processing Letters %V 74 %P 19 - 25 %8 2000/04// %@ 0020-0190 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020019000000338 %N 1–2 %R 10.1016/S0020-0190(00)00033-8 %0 Journal Article %J Mathematics of Operations ResearchMathematics of Operations Research %D 2000 %T Approximation Algorithms for Disjoint Paths and Related Routing and Packing Problems %A Baveja,Alok %A Srinivasan, Aravind %K Approximation algorithms %K Disjoint paths %K integer programming %K Linear programming %K multicommodity flow %K Packing %K randomized algorithms %K rounding %K Routing %K unsplittable flow %X Given a network and a set of connection requests on it, we consider the maximum edge-disjoint paths and related generalizations and routing problems that arise in assigning paths for these requests. We present improved approximation algorithms and/or integrality gaps for all problems considered; the central theme of this work is the underlying multicommodity flow relaxation. Applications of these techniques to approximating families of packing integer programs are also presented. %B Mathematics of Operations ResearchMathematics of Operations Research %V 25 %P 255 - 280 %8 2000/05/01/ %@ 0364-765X, 1526-5471 %G eng %U http://mor.journal.informs.org/content/25/2/255 %N 2 %R 10.1287/moor.25.2.255.12228 %0 Report %D 2000 %T An Arnoldi--Schur Algorithm for Large Eigenproblems %A Stewart, G.W. %K Technical Report %X Sorensen's iteratively restarted Arnoldi algorithm is one of the mostsuccessful and flexible methods for finding a few eigenpairs of a large matrix. However, the need to preserve structure of the Arnoldi decomposition, on which the algorithm is based, restricts the range of transformations that can be performed on it. In consequence, it is difficult to deflate converged Ritz vectors from the decomposition. Moreover, the potential forward instability of the implicit QR algorithm can cause unwanted Ritz vectors to persist in the computation. In this paper we introduce a generalized Arnoldi decomposition that solves both problems in a natural and efficient manner. (Also cross-referenced as UMIACS-TR-2000-21) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2000-21 %8 2000/04/25/ %G eng %U http://drum.lib.umd.edu/handle/1903/1066 %0 Conference Paper %B 38th AIAA Aerospace Sciences Meeting and Exhibit %D 2000 %T Assessment of Temperature Fluctuation Models for RANS Simulations of Hypersonic Reacting Flows %A Sinha,K. %A Martin, M.P %A Candler,G. V %X Turbulence-chemistry interactions play an important role in various fluid flow problems. In a hypersonic boundary layer, these interactions can alter the surface heating rate of the vehicle by modifying the temperature profile across the layer. The effect of these interactions is accounted for in a Reynolds-averaged Navier Stokes (RANS) simulation using a probability density function (PDF) of the temperature distribution. In this paper, two presumed PDF models are considered, namely, the Martin and Candler model and the Gaffney et al. model, and their accuracy is assessed by comparing to direct numerical simulation (DNS) of homogeneous isotropic turbulence. TheMartin and Candler model reproduces the turbulence-chemistry interactions in the flow correctly whereas the Gaffney model overpredicts the temperature fluctuations. %B 38th AIAA Aerospace Sciences Meeting and Exhibit %C Reno, NV %8 2000/// %G eng %0 Journal Article %J SIGSOFT Softw. Eng. Notes %D 2000 %T Automated test oracles for GUIs %A Memon, Atif M. %A Pollack,Martha E. %A Soffa,Mary Lou %K automated oracles %K GUI test oracles %K GUI testing %X Graphical User Interfaces (GUIs) are critical components of today's software. Because GUIs have different characteristics than traditional software, conventional testing techniques do not apply to GUI software. In previous work, we presented an approach to generate GUI test cases, which take the form of sequences of actions. In this paper we develop a test oracle technique to determine if a GUI behaves as expected for a given test case. Our oracle uses a formal model of a GUI, expressed as sets of objects, object properties, and actions. Given the formal model and a test case, our oracle automatically derives the expected state for every action in the test case. We represent the actual state of an executing GUI in terms of objects and their properties derived from the GUI's execution. Using the actual state acquired from an execution monitor, our oracle automatically compares the expected and actual states after each action to verify the correctness of the GUI for the test case. We implemented the oracle as a component in our GUI testing system, called Planning Assisted Tester for grapHical user interface Systems (PATHS), which is based on AI planning. We experimentally evaluated the practicality and effectiveness of our oracle technique and report on the results of experiments to test and verify the behavior of our version of the Microsoft WordPad's GUI. %B SIGSOFT Softw. Eng. Notes %V 25 %P 30 - 39 %8 2000/11// %@ 0163-5948 %G eng %U http://doi.acm.org/10.1145/357474.355050 %N 6 %R 10.1145/357474.355050 %0 Journal Article %J SIAM Journal on Discrete Mathematics %D 2000 %T The capacitated k-center problem %A Khuller, Samir %A Sussmann,Y. J %B SIAM Journal on Discrete Mathematics %V 13 %P 403 - 403 %8 2000/// %G eng %0 Journal Article %J Ninth USENIX Security Symposium %D 2000 %T A chosen ciphertext attack against several e-mail encryption protocols %A Katz, Jonathan %A Schneier,B. %B Ninth USENIX Security Symposium %8 2000/// %G eng %0 Journal Article %J Computational Geometry %D 2000 %T Chromatic nearest neighbor searching: A query sensitive approach %A Mount, Dave %A Netanyahu,Nathan S. %A Silverman,Ruth %A Wu,Angela Y. %K BBD trees %K Branch-and-bound search %K Chromatic nearest neighbors %K Classification algorithms %K Multidimensional searching %K pattern recognition %K Query sensitive analysis %X The nearest neighbor problem is that of preprocessing a set P of n data points in Rd so that, given any query point q, the closest point in P to q can be determined efficiently. In the chromatic nearest neighbor problem, each point of P is assigned a color, and the problem is to determine the color of the nearest point to the query point. More generally, given k⩾1, the problem is to determine the color occurring most frequently among the k nearest neighbors. The chromatic version of the nearest neighbor problem is used in many applications in pattern recognition and learning. In this paper we present a simple algorithm for solving the chromatic k nearest neighbor problem. We provide a query sensitive analysis, which shows that if the color classes form spatially well separated clusters (as often happens in practice), then queries can be answered quite efficiently. We also allow the user to specify an error bound ε⩾0, and consider the same problem in the context of approximate nearest neighbor searching. We present empirical evidence that for well clustered data sets, this approach leads to significant improvements in efficiency. %B Computational Geometry %V 17 %P 97 - 119 %8 2000/12// %@ 0925-7721 %G eng %U http://www.sciencedirect.com/science/article/pii/S0925772100000213 %N 3–4 %R 10.1016/S0925-7721(00)00021-3 %0 Journal Article %J NONCONVEX OPTIMIZATION AND ITS APPLICATIONS %D 2000 %T Combinatorial Problems Arising in Deregulated Electrical Power Industry: Survey and Future Directions %A Cook,D. %A Hicks,G. %A Faber,V. %A Marathe,M. V %A Srinivasan, Aravind %A Sussmann,Y. J %A Thornquist,H. %B NONCONVEX OPTIMIZATION AND ITS APPLICATIONS %V 42 %P 138 - 162 %8 2000/// %G eng %0 Conference Paper %B Proceedings of the eleventh annual ACM-SIAM symposium on Discrete algorithms %D 2000 %T Communication complexity of document exchange %A Cormode,Graham %A Paterson,Mike %A Sahinalp,Süleyman Cenk %A Vishkin, Uzi %B Proceedings of the eleventh annual ACM-SIAM symposium on Discrete algorithms %S SODA '00 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 197 - 206 %8 2000/// %@ 0-89871-453-2 %G eng %U http://dl.acm.org/citation.cfm?id=338219.338252 %0 Conference Paper %B Proceedings of the fourth international ACM conference on Assistive technologies %D 2000 %T A comparison of voice controlled and mouse controlled web browsing %A Christian,Kevin %A Kules,Bill %A Shneiderman, Ben %A Youssef,Adel %K human-computer %K interaction %K User interfaces %K voice browsers %K voice recognition %K web browsing %B Proceedings of the fourth international ACM conference on Assistive technologies %S Assets '00 %I ACM %C New York, NY, USA %P 72 - 79 %8 2000/// %@ 1-58113-313-8 %G eng %U http://doi.acm.org/10.1145/354324.354345 %R 10.1145/354324.354345 %0 Journal Article %J J. ACM %D 2000 %T Contention resolution with constant expected delay %A Goldberg,Leslie Ann %A Mackenzie,Philip D. %A Paterson,Mike %A Srinivasan, Aravind %K contention resolution %K Ethernet %K Markov chains %K multiple-access channel %X We study contention resolution in a multiple-access channel such as the Ethernet channel. In the model that we consider, n users generate messages for the channel according to a probability distribution. Raghavan and Upfal have given a protocol in which the expected delay (time to get serviced) of every message is O(log n) when messages are generated according to a Bernoulli distribution with generation rate up to about 1/10. Our main results are the following protocols: (a) one in which the expected average message delay is O(1) when messages are generated according to a Bernoulli distribution with a generation rate smaller than 1/e, and (b) one in which the expected delay of any message is O(1) for an analogous model in which users are synchronized (i.e., they agree about the time), there are potentially an infinite number of users, and messages are generated according to a Poisson distribution with generation rate up to 1/e. (Each message constitutes a new user.)To achieve (a), we first show how to simulate (b) using n synchronized users, and then show how to build the synchronization into the protocol. %B J. ACM %V 47 %P 1048 - 1096 %8 2000/11// %@ 0004-5411 %G eng %U http://doi.acm.org/10.1145/355541.355567 %N 6 %R 10.1145/355541.355567 %0 Journal Article %J Workshop on New Paradigms in Information Visualization and Manipulation (NPIVM 2000), ACM Press %D 2000 %T Coordinating overviews and detail views of WWW log data %A Hochheiser,H. %A Shneiderman, Ben %X Web server log analysis tools provide site operators with useful information regarding the visitors to their sites.� Unfortunately, the utility of these tools is often limited by the use of aggregate summaries that hide the information associated with individual requests, and by the absence of contextual data that might help users interpret those summaries. Building upon earlier work in the use of starfield visualizations to display web site requests as individual data points [8], this paper describes the use of multiple-coordinated visualizations of web log data at varying granularities, and alongside additional related displays of appropriate contextual information. %B Workshop on New Paradigms in Information Visualization and Manipulation (NPIVM 2000), ACM Press %8 2000/// %G eng %0 Conference Paper %B Network models for control and processing %D 2000 %T Cortical inhibition as explained by the competitive distribution hypothesis %A Reggia, James A. %A Sutton III,G. G. %A Lynne,C. %A D'Autrechy,S. C %A Armentrout,S. L %B Network models for control and processing %P 31 - 62 %8 2000/// %G eng %0 Journal Article %J ACM Trans. Comput.-Hum. Interact. %D 2000 %T Creating creativity: user interfaces for supporting innovation %A Shneiderman, Ben %K creativity support tools %K direct manipulation %K Graphical user interfaces %K Human-computer interaction %K Information Visualization %X A challenge for human-computer interaction researchers and user interf ace designers is to construct information technologies that support creativity. This ambitious goal can be attained by building on an adequate understanding of creative processes. This article offers a four-phase framework for creativity that might assist designers in providing effective tools for users: (1)Collect: learn from provious works stored in libraries, the Web, etc.; (2) Relate: consult with peers and mentors at early, middle, and late stages, (3)Create: explore, compose, evaluate possible solutions; and (4) Donate: disseminate the results and contribute to the libraries. Within this integrated framework, this article proposes eight activities that require human-computer interaction research and advanced user interface design. A scenario about an architect illustrates the process of creative work within such an environment. %B ACM Trans. Comput.-Hum. Interact. %V 7 %P 114 - 138 %8 2000/03// %@ 1073-0516 %G eng %U http://doi.acm.org/10.1145/344949.345077 %N 1 %R 10.1145/344949.345077 %0 Journal Article %J Ubiquity %D 2000 %T Credit for computer crashes? %A Gehl,J. %A Shneiderman, Ben %B Ubiquity %V 1 %P 2 - 9 %8 2000/// %G eng %N 31 %0 Conference Paper %B NASA conference publication %D 2000 %T DataCutter: Middleware for filtering very large scientific datasets on archival storage systems %A Beynon, M. %A Ferreira,R. %A Kurc, T. %A Sussman, Alan %A Saltz, J. %B NASA conference publication %P 119 - 134 %8 2000/// %G eng %0 Journal Article %J Computing in Science Engineering %D 2000 %T The decompositional approach to matrix computation %A Stewart, G.W. %K Algebra %K approach;history;matrix %K Cholesky %K computation;matrix %K computations;pivoted %K computing;matrix %K decomposition; %K decomposition;bibliographies;history;mathematics %K decomposition;decompositional %K decomposition;numerical %K decomposition;QR %K decomposition;Schur %K decomposition;singular %K decomposition;spectral %K linear %K LU %K matrix %K revolutionized %K value %X The introduction of matrix decomposition into numerical linear algebra revolutionized matrix computations. The article outlines the decompositional approach, comments on its history, and surveys the six most widely used decompositions: Cholesky decomposition; pivoted LU decomposition; QR decomposition; spectral decomposition; Schur decomposition; and singular value decomposition %B Computing in Science Engineering %V 2 %P 50 - 59 %8 2000/02//jan %@ 1521-9615 %G eng %N 1 %R 10.1109/5992.814658 %0 Conference Paper %B Heterogeneous Computing Workshop %D 2000 %T Design of a Framework for Data-Intensive Wide-Area Applications %A Beynon,Michael D. %A Kurc,Tahsin %A Sussman, Alan %A Saltz,Joel %K application decomposition %K data analysis and exploration %K Data-intensive applications %X Applications that use collections of very large, distributed datasets have become an increasingly important part of science and engineering. With high performance wide-area networks become more pervasive, there is interest in making collective use of distributed computational and data resources. Recent work has converged to the notion of the Grid, which attempts to uniformly present a heterogeneous collection of distributed resources. Current Grid research covers many areas from low-level infrastructure issues to high-level application concerns. However, providing support for efficient exploration and processing of very large scientific datasets stored in distributed archival storage systems remains a challenging research issue.We have initiated an effort that focuses on developing efficient data-intensive applications in a Grid environment. In this paper, we present a framework, called filter-stream programming that represents the processing units of a data-intensive application as a set of filters, which are designed to be efficient in their use of memory and scratch space. We describe a prototype infrastructure that supports execution of applications using the proposed framework. We present the implementation of two applications using the filter-stream programming framework, and discuss experimental results demonstrating the effects of heterogeneous resources on application performance. %B Heterogeneous Computing Workshop %I IEEE Computer Society %C Los Alamitos, CA, USA %P 116 - 116 %8 2000/// %G eng %R http://doi.ieeecomputersociety.org/10.1109/HCW.2000.843737 %0 Conference Paper %B Proceedings of the 3rd conference on Designing interactive systems: processes, practices, methods, and techniques %D 2000 %T Designing StoryRooms: interactive storytelling spaces for children %A Alborzi,H. %A Druin, Allison %A Montemayor,J. %A Platner,M. %A Porteous,J. %A Sherman,L. %A Boltman,A. %A Taxén,G. %A Best,J. %A Hammer,J. %A others %B Proceedings of the 3rd conference on Designing interactive systems: processes, practices, methods, and techniques %P 95 - 104 %8 2000/// %G eng %0 Journal Article %J Communications of the ACM %D 2000 %T Designing trust into online experiences %A Shneiderman, Ben %B Communications of the ACM %V 43 %P 57 - 59 %8 2000/12// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/355112.355124 %N 12 %R 10.1145/355112.355124 %0 Conference Paper %B IEEE International Conference on Information Visualization, 2000. Proceedings %D 2000 %T Direct annotation: a drag-and-drop strategy for labeling photos %A Shneiderman, Ben %A Kang,H. %K Biomedical imaging %K Cities and towns %K Computer errors %K Computer science %K database indexing %K database schema %K Databases %K digital libraries %K direct annotation %K direct manipulation %K drag-and-drop strategy %K Educational institutions %K graphical user interface %K Graphical user interfaces %K History %K hobby computing %K image searching %K label placement %K Labeling %K Laboratories %K Libraries %K personal information systems %K personal names database %K PhotoFinder prototype %K photograph labelling %K photographic libraries %K Photography %K scrolling list %K user interface design %K visual databases %X Annotating photographs is such a time-consuming, tedious and error-prone data entry task that it discourages most owners of personal photo libraries. By allowing the user to drag labels, such as personal names, from a scrolling list and drop them onto a photo, we believe we can make the task faster, easier and more appealing. Since the names are entered in a database, searching for all photos of a friend or family member is dramatically simplified. We describe the user interface design and the database schema to support direct annotation, as implemented in our PhotoFinder prototype %B IEEE International Conference on Information Visualization, 2000. Proceedings %I IEEE %P 88 - 95 %8 2000/// %@ 0-7695-0743-3 %G eng %R 10.1109/IV.2000.859742 %0 Conference Paper %B 11th International Workshop on Database and Expert Systems Applications, 2000. Proceedings %D 2000 %T Domain name based visualization of Web histories in a zoomable user interface %A Gandhi,R. %A Kumar,G. %A Bederson, Benjamin B. %A Shneiderman, Ben %K Computer science %K data visualisation %K domain name %K domain name based visualization %K Domain Tree Browser %K Educational institutions %K History %K hypermedia %K hypertext links %K hypertext systems %K information resources %K Navigation %K online front-ends %K thumbnails %K Tree graphs %K tree structured visual navigation history %K Uniform resource locators %K URLs %K User interfaces %K Visualization %K Web browser companion %K Web histories %K Web pages %K World Wide Web %K zoomable user interface %K zoomable window %X Users of hypertext systems like the World Wide Web (WWW) often find themselves following hypertext links deeper and deeper, only to find themselves “lost” and unable to find their way back to the previously visited pages. We have implemented a Web browser companion called Domain Tree Browser (DTB) that builds a tree structured visual navigation history while browsing the Web. The Domain Tree Browser organizes the URLs visited based on the domain name of each URL and shows thumbnails of each page in a zoomable window %B 11th International Workshop on Database and Expert Systems Applications, 2000. Proceedings %I IEEE %P 591 - 598 %8 2000/// %@ 0-7695-0680-1 %G eng %R 10.1109/DEXA.2000.875085 %0 Conference Paper %B Formal Techniques in Real-Time and Fault-Tolerant Systems %D 2000 %T A dual interpretation of “standard constraints” in parametric scheduling %A Subramani,K. %A Agrawala, Ashok K. %X Parametric scheduling in real-time systems, in the presence of linear relative constraints between the start and execution times of tasks, is a well-studied problem. Prior research established the existence of polynomial time algorithms for the case when the constraints are restricted to be standard and the execution time vectors belong to an axis-parallel hyper-rectangle. In this paper we present a polynomial time algorithm for the case when the execution time vectors belong to arbitrary convex domains. Our insights into the problem occur primarily as a result of studying the dual polytope of the constraint system. %B Formal Techniques in Real-Time and Fault-Tolerant Systems %P 205 - 223 %8 2000/// %G eng %0 Conference Paper %D 2000 %T Dynamic node management and measure estimation in a state-driven fault injector %A Chandra,R. %A Michel Cukier %A Lefever,R. %A Sanders,W. %A others %X Validation of distributed systems using fault injectionis difficult because of their inherent complexity, lack of a global clock, and lack of an easily accessible notion of a global state. To address these challenges, the Loki fault injector injects faults based on a partial view of the global state of a distributed system, and performs a post-runtime analysis using an off-line clock synchronization algorithm to determine whether the faults were properly injected. In this paper, we first describe an enhanced runtime architecture for the Loki fault injector and then present a new method for obtaining measures in Loki. The enhanced runtime allows dynamic entry and exit of nodes in the system. It also offers more efficient multicast of notification messages and more efficient communication between state machines on the same host, and is more scalable than the previous runtime. We then detail a new and flexible method for obtaining a wide range of performance and dependability measures in Loki. %P 248 - 261 %8 2000/// %G eng %U https://www.perform.csl.illinois.edu/Papers/USAN_papers/00CHA01.pdf %0 Journal Article %J International Journal of High Performance Computing Applications %D 2000 %T Efficient performance prediction for large-scale, data-intensive applications %A Kurc, T. %A Uysal, M. %A Eom, H. %A Hollingsworth, Jeffrey K %A Saltz, J. %A Sussman, Alan %B International Journal of High Performance Computing Applications %V 14 %P 216 - 216 %8 2000/// %G eng %N 3 %0 Journal Article %J Computer Aided Design %D 2000 %T Efficiently computing and updating triangle strips for real-time rendering %A El-Sana,J. %A Evans,F. %A Kalaiah,A. %A Varshney, Amitabh %A Skiena,S. %A Azanli,E. %X Triangle strips are a widely used hardware-supported data-structure to compactly represent and efficiently render polygonal meshes. In thispaper we survey the efficient generation of triangle strips as well as their variants. We present efficient algorithms for partitioning polygonal meshes into triangle strips. Triangle strips have traditionally used a buffer size of two vertices. In this paper we also study the impact of larger buffer sizes and various queuing disciplines on the effectiveness of triangle strips. View-dependent simplification has emerged as a powerful tool for graphics acceleration in visualization of complex environments. However, in a view-dependent framework the triangle mesh connectivity changes at every frame making it difficult to use triangle strips. In this paper we present a novel data-structure, Skip Strip, that efficiently maintains triangle strips during such view-dependent changes. A Skip Strip stores the vertex hierarchy nodes in a skip-list-like manner with path compression. We anticipate that Skip Strips will provide a road-map to combine rendering acceleration techniques for static datasets, typical of retained-mode graphics applications, with those for dynamic datasets found in immediate-mode applications. %B Computer Aided Design %V 32 %P 753 - 772 %8 2000/// %G eng %N 13 %0 Book %D 2000 %T Embedded multiprocessors: Scheduling and synchronization %A Sriram,S. %A Bhattacharyya, Shuvra S. %I CRC %V 3 %8 2000/// %G eng %0 Journal Article %J ACM SIGSOFT Software Engineering Notes %D 2000 %T The experimental validation and packaging of software technologies %A Shull, Forrest %A Basili, Victor R. %A Zelkowitz, Marvin V %B ACM SIGSOFT Software Engineering Notes %V 25 %P 82 - 83 %8 2000/01// %@ 0163-5948 %G eng %U http://doi.acm.org/10.1145/340855.341021 %N 1 %R 10.1145/340855.341021 %0 Conference Paper %B Proceedings of the first international symposium on Databases in parallel and distributed systems %D 2000 %T Exploiting concurrency in a DBMS implementation for production systems %A Raschid, Louiqa %A Sellis,T. %A Lin,C. C %B Proceedings of the first international symposium on Databases in parallel and distributed systems %P 34 - 45 %8 2000/// %G eng %0 Conference Paper %B Proceedings of the 2000 annual national conference on Digital government research %D 2000 %T Extending understanding of federal statistics in tables %A Marchionini,Gary %A Hert,Carol %A Liddy,Liz %A Shneiderman, Ben %K data exploration %K dynamic queries %K Statistics %K tablular data %K User interfaces %X This paper describes progress toward improving user interfaces for US Federal government statistics that are presented in tables. Based on studies of user behaviors and needs related to statistical tables, we describe interfaces to assist diverse users with a range of statistical literacy to explore, find, understand, and use US Federal government statistics. %B Proceedings of the 2000 annual national conference on Digital government research %S dg.o '00 %I Digital Government Society of North America %P 1 - 7 %8 2000/// %G eng %U http://dl.acm.org/citation.cfm?id=1123075.1123079 %0 Journal Article %J Behaviour & Information Technology %D 2000 %T Facilitating data exploration with query previews: A study of user performance and preference %A Tanin,Egemen %A Lotem,Amnon %A Haddadin,Ihab %A Shneiderman, Ben %A Plaisant, Catherine %A Slaughter,Laura %X Networked and local data exploration systems that use command languages, menus, or form fill-in interfaces rarely give users an indication of the distribution of data. This often leads users to waste time, posing queries that have zero-hit or mega-hit results. Query previews are a novel visual approach for browsing databases. Query previews supply users with data distribution information for selected attributes of the database, and give continuous feedback about the size of the result set as the query is being formed. Subsequent refinements might be necessary to narrow the search. As there is a risk that query previews are an additional step, leading to a more complex and slow search process, a within-subjects empirical study was ran with 12 subjects who used interfaces with and without query previews and with minimized network delays. Even with 12 subjects and minimized network delays statistically significant differences were found, showing that query previews could speed up performance 1.6 to 2.1 times and lead to higher user satisfaction.Networked and local data exploration systems that use command languages, menus, or form fill-in interfaces rarely give users an indication of the distribution of data. This often leads users to waste time, posing queries that have zero-hit or mega-hit results. Query previews are a novel visual approach for browsing databases. Query previews supply users with data distribution information for selected attributes of the database, and give continuous feedback about the size of the result set as the query is being formed. Subsequent refinements might be necessary to narrow the search. As there is a risk that query previews are an additional step, leading to a more complex and slow search process, a within-subjects empirical study was ran with 12 subjects who used interfaces with and without query previews and with minimized network delays. Even with 12 subjects and minimized network delays statistically significant differences were found, showing that query previews could speed up performance 1.6 to 2.1 times and lead to higher user satisfaction. %B Behaviour & Information Technology %V 19 %P 393 - 403 %8 2000/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/014492900750052651 %N 6 %R 10.1080/014492900750052651 %0 Journal Article %J Theoretical Computer Science %D 2000 %T Fault tolerant K-center problems %A Khuller, Samir %A Pless,Robert %A Sussmann,Yoram J. %K Approximation algorithms %K Facility location %K Fault-tolerance %K K-center %X The basic K-center problem is a fundamental facility location problem, where we are asked to locate K facilities in a graph, and to assign vertices to facilities, so as to minimize the maximum distance from a vertex to the facility to which it is assigned. This problem is known to be NP-hard, and several optimal approximation algorithms that achieve an approximation factor of 2 have been developed for it. We focus our attention on a generalization of this problem, where each vertex is required to have a set of α (α⩽K) centers close to it. In particular, we study two different versions of this problem. In the first version, each vertex is required to have at least α centers close to it. In the second version, each vertex that does not have a center placed on it is required to have at least α centers close to it. For both these versions we are able to provide polynomial time approximation algorithms that achieve constant approximation factors for any α. For the first version we give an algorithm that achieves an approximation factor of 3 for any α, and achieves an approximation factor of 2 for α<4. For the second version, we provide algorithms with approximation factors of 2 for any α. The best possible approximation factor for even the basic K-center problem is 2, assuming P≠NP. In addition, we give a polynomial time approximation algorithm for a generalization of the K-supplier problem where a subset of at most K supplier nodes must be selected as centers so that every demand node has at least α centers close to it. For this version our approximation factor is 3. The best possible approximation factor for even the basic K-supplier problem is 3, assuming P≠NP. %B Theoretical Computer Science %V 242 %P 237 - 245 %8 2000/07/06/ %@ 0304-3975 %G eng %U http://www.sciencedirect.com/science/article/pii/S0304397598002229 %N 1–2 %R 10.1016/S0304-3975(98)00222-9 %0 Journal Article %J Coordination Languages and Models %D 2000 %T GCCS: A graphical coordination language for system specification %A Cleaveland, Rance %A Du,X. %A Smolka,S. %B Coordination Languages and Models %P 207 - 212 %8 2000/// %G eng %0 Report %D 2000 %T A Generalization of Saad's Theorem on Rayleigh-Ritz Approximations %A Stewart, G.W. %K Technical Report %X Let $(\lambda,x)$ be an eigenpair of the Hermitian matrix $A$ of order$n$ and let $(\mu,u)$ be a Ritz pair from a subspace $\clk$ of $\comp^{2}$. Saad has given a simple inequality bounding $\sin\angle(x,u)$ in terms of $\sin\angle(x,\clk)$. In this note we show that this inequality can be extended to an equally simple inequality for eigenspaces of non-Hermitian matrices. (Also cross-referenced as UMIACS-TR-99-78) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-99-78 %8 2000/02/04/ %G eng %U http://drum.lib.umd.edu/handle/1903/1050 %0 Journal Article %J Science %D 2000 %T The genome sequence of Drosophila melanogaster %A Adams,M.D. %A Celniker,S.E. %A Holt,R.A. %A Evans,C.A. %A Gocayne,J.D. %A Amanatides,P.G. %A Scherer,S.E. %A Li,P.W. %A Hoskins,R.A. %A Galle,R.F. %A others %X The fly Drosophila melanogaster is one of the most intensively studied organisms in biology and serves as a model system for the investigation of many developmental and cellular processes common to higher eukaryotes, including humans. We have determined the nucleotide sequence of nearly all of the ∼120-megabase euchromatic portion of theDrosophila genome using a whole-genome shotgun sequencing strategy supported by extensive clone-based sequence and a high-quality bacterial artificial chromosome physical map. Efforts are under way to close the remaining gaps; however, the sequence is of sufficient accuracy and contiguity to be declared substantially complete and to support an initial analysis of genome structure and preliminary gene annotation and interpretation. The genome encodes ∼13,600 genes, somewhat fewer than the smaller Caenorhabditis elegansgenome, but with comparable functional diversity. %B Science %V 287 %P 2185 - 2185 %8 2000/// %@ 0036-8075, 1095-9203 %G eng %U http://www.sciencemag.org/content/287/5461/2185 %N 5461 %R 10.1126/science.287.5461.2185 %0 Conference Paper %B ICPR %D 2000 %T Hidden Markov Models for Images %A DeMenthon,D. %A Stuckelberg,M. Vuilleumier %A David Doermann %X In this paper we investigate how speech recognition techniques can be extended to image processing. We describe a method for learning statistical models of images using a second order hidden Markov mesh model. First, an image can be segmented in a way that best matches its statistical model by an approach related to the dynamic programming used for segmenting Markov chains. Second, given an image segmentation a statistical model (3D state transition matrix and observation distributions within states) can be estimated. These two steps are repeated until convergence to provide both a segmentation and a statistical model of the image. We also describe a semi-Markov modeling technique in which the distributions of widths and heights of segmented regions are modeled explicitly by gamma distributions in a way related to explicit duration modeling in HMMs. Finally, we propose a statistical distance measure between images based on the similarity of their statistical models for classication and retrieval tasks. %B ICPR %P 147 - 150 %8 2000/// %G eng %0 Journal Article %J Journal of Computer and System Sciences %D 2000 %T Improved Algorithms via Approximations of Probability Distributions %A Chari,Suresh %A Rohatgi,Pankaj %A Srinivasan, Aravind %K derandomization %K discrepancy %K explicit constructions %K graph coloring %K Parallel algorithms %K small sample spaces %X We present two techniques for constructing sample spaces that approximate probability distributions. The first is a simple method for constructing the small-bias probability spaces introduced by Naor and Naor. We show how to efficiently combine this construction with the method of conditional probabilities to yield improved parallel algorithms for problems such as set discrepancy, finding large cuts in graphs, and finding large acyclic subgraphs. The second is a construction of small probability spaces approximating general independent distributions which are of smaller size than the constructions of Even, Goldreich, Luby, Nisan, and Veličković. %B Journal of Computer and System Sciences %V 61 %P 81 - 107 %8 2000/08// %@ 0022-0000 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022000099916951 %N 1 %R 10.1006/jcss.1999.1695 %0 Journal Article %J IMA journal of numerical analysis %D 2000 %T Iterative methods for stabilized discrete convection-diffusion problems %A Shih, Y. T %A Elman, Howard %B IMA journal of numerical analysis %V 20 %P 333 - 333 %8 2000/// %G eng %N 3 %0 Journal Article %J SIAM Journal on Matrix Analysis and Applications %D 2000 %T Iterative regularization and MINRES %A Kilmer,M. %A Stewart, G.W. %X In this paper we present three theorems which give insight into the regularizingproperties of MINRES. While our theory does not completely characterize the regularizing behavior of the algorithm, it provides a partial explanation of the observed behavior of the method. Unlike traditional attempts to explain the regularizing properties of Krylov subspace methods, our approach focuses on convergence properties of the residual rather than on convergence analysis of the harmonic Ritz values. The import of our analysis is illustrated by two examples. In particular, our theoret- ical and numerical results support the following important observation: in some circumstances the dimension of the optimal Krylov subspace can be much smaller than the number of the components of the truncated spectral solution that must be computed to attain comparable accuracy. %B SIAM Journal on Matrix Analysis and Applications %V 21 %P 613 - 628 %8 2000/// %G eng %N 2 %0 Journal Article %J PE & RS- Photogrammetric Engineering and Remote Sensing %D 2000 %T Kronos: A software system for the processing and retrieval of large-scale AVHRR data sets %A Zhang,Z. %A JaJa, Joseph F. %A Bader, D.A. %A Kalluri, SNV %A Song,H. %A El Saleous,N. %A Vermote,E. %A Townshend,J.R.G. %X Raw remotely sensed satellite data have to be processed andmapped into a standard projection in order to produce a multi- temporal data set which can then be used for regional or global Earth science studies. However, traditional methods of processing remotely sensed satellite data have inherent limitations because they are based on a fixed processing chain. Different users may need the data in different forms with possibly different processing steps; hence, additional transformations may have to be applied to the processed data, resulting in potentially significant errors. In this paper, we describe a software system, Kronos, for the generation of custom-tailored products from the Advanced Very High Resolution Radiometer (AVHRR) sensor. It allows the generation of a rich set of products that can be easily specified through a simple interface by scientists wishing to carry out Earth system modeling or analysis. Kronos is based on a flexible methodology and consists of four major components: ingest and preprocessing, indexing and storage, a search and processing engine, and a Java interface. After geo-location and calibration, every pixel is indexed and stored using a combination of data structures. Following the users' queries, data are selectively retrieved and secondary processing such as atmospheric correction, compositing, and projection are performed as specified. The processing is divided into two stages, the first of which involves the geo-location and calibration of the remotely sensed data and, hence, results in no loss of information. The second stage involves the retrieval of the appropriate data subsets and the application of the secondary processing specified by the user. This scheme allows the indexing and the storage of data from different sensors without any loss of information and, therefore, allows assimilation of data from multiple sensors. User specified processing can be applied later as needed. %B PE & RS- Photogrammetric Engineering and Remote Sensing %V 66 %P 1073 - 1082 %8 2000/// %G eng %N 9 %0 Journal Article %J Communications of the ACM %D 2000 %T The limits of speech recognition %A Shneiderman, Ben %B Communications of the ACM %V 43 %P 63 - 65 %8 2000/09// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/348941.348990 %N 9 %R 10.1145/348941.348990 %0 Conference Paper %B Multimedia and Expo, 2000. ICME 2000. 2000 IEEE International Conference on %D 2000 %T Live multimedia adaptation through wireless hybrid networks %A Koivisto,A. %A Pietkainen,P. %A Sauvola,J. %A David Doermann %K adaptation;mobile %K adaptation;value-adding %K analysis;wireless %K annotation;content-recognition;end-to-end %K communication;multimedia %K communication;real-time %K hybrid %K information;video %K LAN; %K multimedia %K networks;mobile %K Princess;automated %K processing;mobile %K service;live %K service;mobile %K streams;real-time %K Surveillance %K system;terminal %K systems;surveillance;wireless %K terminals;multiple %K video %X We present new techniques to intelligently adapt and combine multimedia presentations in real-time, mobile service environments. We present the techniques necessary to perform mobile multimedia processing for multiple video streams with ldquo;value-adding rdquo; information. The adaptation uses video analysis, content-recognition and automated annotation to provide scalable and interactive presentations over hybrid networks to mobile terminals. As a concrete case, we present a mobile surveillance service called Princess, which has several video sources and supports control information, terminal adaptation and end-to-end service %B Multimedia and Expo, 2000. ICME 2000. 2000 IEEE International Conference on %V 3 %P 1697 -1700 vol.3 - 1697 -1700 vol.3 %8 2000/// %G eng %R 10.1109/ICME.2000.871098 %0 Conference Paper %D 2000 %T Loki: a state-driven fault injector for distributed systems %A Chandra,R. %A Lefever,R.M. %A Michel Cukier %A Sanders,W. H. %K campaign execution %K computer testing %K distributed applications failure %K distributed processing %K distributed system validation %K Fault diagnosis %K fault event placement %K fault specification %K global system state %K global timeline %K Graphical user interfaces %K Loki %K off-line clock synchronization %K post-runtime analysis %K specified measures estimation %K state machine specification %K state-driven fault injection %K Synchronisation %X Distributed applications can fail in subtle ways that depend on the state of multiple parts of a system. This complicates the validation of such systems via fault injection, since it suggests that faults should be injected based on the global state of the system. In Loki, fault injection is performed based on a partial view of the global state of a distributed system, i.e. faults injected in one node of the system can depend on the state of other nodes. Once faults are injected, a post-runtime analysis, using off-line clock synchronization, is used to place events and injections on a single global timeline and to determine whether the intended faults were properly injected. Finally, experiments containing successful fault injections are used to estimate the specified measures. In addition to briefly reviewing the concepts behind Loki and its organization, we detail Loki's user interface. In particular, we describe the graphical user interfaces for specifying state machines and faults, for executing a campaign and for verifying whether the faults were properly injected %P 237 - 242 %8 2000/// %G eng %R 10.1109/ICDSN.2000.857544 %0 Journal Article %J Information Processing Letters %D 2000 %T Low discrepancy sets yield approximate min-wise independent permutation families %A Saks,Michael %A Srinivasan, Aravind %A Zhou,Shiyu %A Zuckerman,David %K combinatorial problems %K Document filtering %K explicit constructions %K Information retrieval %K Min-wise independent permutations %K Pseudorandom permutations %X Motivated by a problem of filtering near-duplicate Web documents, Broder, Charikar, Frieze and Mitzenmacher defined the following notion of ϵ-approximate min-wise independent permutation families. A multiset F of permutations of {0,1,…,n−1} is such a family if for all K⫅{0,1,…,n−1} and any x∈K, a permutation π chosen uniformly at random from F satisfies |Pr[min{π(K)}=π(x)]−1|K||≤ϵ|K|. We show connections of such families with low discrepancy sets for geometric rectangles, and give explicit constructions of such families F of size nO(logn) for ϵ=1/nΘ(1), improving upon the previously best-known bound of Indyk. We also present polynomial-size constructions when the min-wise condition is required only for |K|≤2O(log2/3n), with ϵ≥2−O(log2/3n). %B Information Processing Letters %V 73 %P 29 - 32 %8 2000/01/31/ %@ 0020-0190 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020019099001635 %N 1–2 %R 10.1016/S0020-0190(99)00163-5 %0 Journal Article %J Bulletin of the EATCS %D 2000 %T Low-discrepancy sets for high-dimensional rectangles: a survey %A Srinivasan, Aravind %X A sub-area of discrepancy theory that has received much attention in computer science re-cently, is that of explicit constructions of low-discrepancy point sets for various types of rectangle families in high dimension. This research has led to interesting applications in error-control cod- ing, distributed protocols, Web document filtering, derandomization, and other areas. We give a short survey of this area here. %B Bulletin of the EATCS %V 70 %P 67 - 76 %8 2000/// %G eng %0 Report %D 2000 %T Making Computer and Internet Usability a Priority %A Kreitzberg,Charles B %A Shneiderman, Ben %K Technical Report %X A management oriented essay on how to promote usability inorganizations. It begins with crafting the message (Good usability is good business). Then it deals with creating a cultural shift in information technology and software engineering departments, so that the impact of poor design and the benefits of improvements are widely understood. It closes with an eight-step call to action. Keywords: usability, technology fluency, user-centered design, software engineering (Also cross-referenced as UMIACS-TR-2000-70) %B Technical Reports from UMIACS %8 2000/10/04/ %G eng %U http://drum.lib.umd.edu/handle/1903/1106 %0 Conference Paper %B Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, Texas %D 2000 %T MOCHA: a database middleware system featuring automatic deployment of application-specific functionality %A Rodrıguez-Martinez,M. %A Roussopoulos, Nick %A McGann,J. M %A Keyley,S. %A Katz,V. %A Song,Z. %A JaJa, Joseph F. %B Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, Texas %8 2000/// %G eng %0 Conference Paper %B IEEE Workshop on Omnidirectional Vision, 2000. Proceedings %D 2000 %T Multi-camera networks: eyes from eyes %A Fermüller, Cornelia %A Aloimonos, J. %A Baker, P. %A Pless, R. %A Neumann, J. %A Stuart, B. %K Biosensors %K CAMERAS %K Computer vision %K Eyes %K Image sequences %K intelligent systems %K Layout %K Machine vision %K Robot vision systems %K Robustness %K Spatiotemporal phenomena %K video cameras %K Virtual reality %X Autonomous or semi-autonomous intelligent systems, in order to function appropriately, need to create models of their environment, i.e., models of space time. These are descriptions of objects and scenes and descriptions of changes of space over time, that is, events and actions. Despite the large amount of research on this problem, as a community we are still far from developing robust descriptions of a system's spatiotemporal environment using video input (image sequences). Undoubtedly, some progress has been made regarding the understanding of estimating the structure of visual space, but it has not led to solutions to specific applications. There is, however, an alternative approach which is in line with today's “zeitgeist.” The vision of artificial systems can be enhanced by providing them with new eyes. If conventional video cameras are put together in various configurations, new sensors can be constructed that have much more power and the way they “see” the world makes it much easier to solve problems of vision. This research is motivated by examining the wide variety of eye design in the biological world and obtaining inspiration for an ensemble of computational studies that relate how a system sees to what that system does (i.e. relating perception to action). This, coupled with the geometry of multiple views that has flourished in terms of theoretical results in the past few years, points to new ways of constructing powerful imaging devices which suit particular tasks in robotics, visualization, video processing, virtual reality and various computer vision applications, better than conventional cameras. This paper presents a number of new sensors that we built using common video cameras and shows their superiority with regard to developing models of space and motion %B IEEE Workshop on Omnidirectional Vision, 2000. Proceedings %I IEEE %P 11 - 18 %8 2000/// %@ 0-7695-0704-2 %G eng %R 10.1109/OMNVIS.2000.853797 %0 Journal Article %J Linear Algebra and its Applications %D 2000 %T A new relative perturbation theorem for singular subspaces %A Li,Ren-Cang %A Stewart, G.W. %K Multiplicative perturbation %K Relative gap %K Relative perturbation theory %K Singular subspace %X This note addresses the sensitivity of singular subspaces of a matrix under relative perturbations. It employs a new technique of separating a multiplicative perturbation D into two components: one is the distance of a scalar multiple of D to the nearest unitary matrix Q and the other is the distance of Q to the identity. Consequently, the new bounds reflect the intrinsic differences in how left and right multiplicative perturbations affect left and right singular subspaces. %B Linear Algebra and its Applications %V 313 %P 41 - 51 %8 2000/07/01/ %@ 0024-3795 %G eng %U http://www.sciencedirect.com/science/article/pii/S0024379500000744 %N 1–3 %R 10.1016/S0024-3795(00)00074-4 %0 Conference Paper %B IEEE INFOCOM 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %D 2000 %T Optimal design of signaling networks for Internet telephony %A Srinivasan, Aravind %A Ramakrishnan,K. G %A Kumaran,K. %A Aravamudan,M. %A Naqvi,S. %K bandwidth allocation %K Computational efficiency %K Computer networks %K Cost function %K Demand forecasting %K demand forecasts %K graph theory %K graphical design tool %K Internet telephony %K Linear programming %K Load forecasting %K Load management %K Network topology %K optimal design %K optimal load balancing %K optimisation %K performance %K quadratic assignment problem %K random graphs %K randomised algorithms %K randomized heuristics %K Signal design %K signaling networks %K Switches %K telecommunication signalling %K topology design %X We present an approach for efficient design of a signaling network for a network of software switches supporting Internet telephony. While one may take an integer programming approach to solve this problem, it quickly becomes intractable even for modest-sized networks. Instead, our topology design uses random graphs that we show to be nearly optimal in cost, highly connected, and computationally efficient even for large networks. We then formulate a quadratic assignment problem (QAP) to map the abstract topology into the physical network to achieve optimal load balancing for given demand forecasts, which we solve using randomized heuristics. Numerical results on several example networks illustrate the performance and computational efficiency of our method. A graphical design tool has been developed based on our algorithms %B IEEE INFOCOM 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %I IEEE %V 2 %P 707-716 vol.2 - 707-716 vol.2 %8 2000/// %@ 0-7803-5880-5 %G eng %R 10.1109/INFCOM.2000.832245 %0 Conference Paper %B Parallel and Distributed Processing Symposium, 2000. IPDPS 2000. Proceedings. 14th International %D 2000 %T Optimizing retrieval and processing of multi-dimensional scientific datasets %A Chang,Chialin %A Kurc, T. %A Sussman, Alan %A Saltz, J. %K active data repository %K Area measurement %K Computer science %K Data analysis %K distributed memory parallel machines %K Educational institutions %K Information retrieval %K Information Storage and Retrieval %K infrastructure %K Microscopy %K Microwave integrated circuits %K multi-dimensional scientific datasets retrieval %K PARALLEL PROCESSING %K Pathology %K range queries %K regular d-dimensional array %K Satellites %K Tomography %X We have developed the Active Data Repository (ADR), an infrastructure that integrates storage, retrieval, and processing of large multi-dimensional scientific datasets on distributed memory parallel machines with multiple disks attached to each node. In earlier work, we proposed three strategies for processing range queries within the ADR framework. Our experimental results show that the relative performance of the strategies changes under varying application characteristics and machine configurations. In this work we investigate approaches to guide and automate the selection of the best strategy for a given application and machine configuration. We describe analytical models to predict the relative performance of the strategies where input data elements are uniformly distributed in the attribute space of the output dataset, restricting the output dataset to be a regular d-dimensional array %B Parallel and Distributed Processing Symposium, 2000. IPDPS 2000. Proceedings. 14th International %I IEEE %P 405 - 410 %8 2000/// %@ 0-7695-0574-0 %G eng %R 10.1109/IPDPS.2000.846013 %0 Journal Article %J International Journal of Human-Computer Interaction %D 2000 %T Performance Benefits of Simultaneous Over Sequential Menus as Task Complexity Increases %A Hochheiser,Harry %A Shneiderman, Ben %X To date, experimental comparisons of menu layouts have concentrated on variants of hierarchical structures of sequentially presented menus. Simultaneous menus-layouts that present multiple active menus on a screen at the same time-are an alternative arrangement that may be useful in many Web design situations. This article describes an experiment involving a between-subject comparison of simultaneous menus and their traditional sequential counterparts. A total of 20 experienced Web users used either simultaneous or sequential menus in a standard Web browser to answer questions based on U.S. Census data. Our results suggest that appropriate use of simultaneous menus can lead to improved task performance speeds without harming subjective satisfaction measures. For novice users performing simple tasks, the simplicity of sequential menus appears to be helpful, but experienced users performing complex tasks may benefit from simultaneous menus. Design improvements can amplify the benefits of simultaneous menu layouts.To date, experimental comparisons of menu layouts have concentrated on variants of hierarchical structures of sequentially presented menus. Simultaneous menus-layouts that present multiple active menus on a screen at the same time-are an alternative arrangement that may be useful in many Web design situations. This article describes an experiment involving a between-subject comparison of simultaneous menus and their traditional sequential counterparts. A total of 20 experienced Web users used either simultaneous or sequential menus in a standard Web browser to answer questions based on U.S. Census data. Our results suggest that appropriate use of simultaneous menus can lead to improved task performance speeds without harming subjective satisfaction measures. For novice users performing simple tasks, the simplicity of sequential menus appears to be helpful, but experienced users performing complex tasks may benefit from simultaneous menus. Design improvements can amplify the benefits of simultaneous menu layouts. %B International Journal of Human-Computer Interaction %V 12 %P 173 - 192 %8 2000/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1207/S15327590IJHC1202_2 %N 2 %R 10.1207/S15327590IJHC1202_2 %0 Journal Article %J J Mol Evol %D 2000 %T Phylogenetic relationships of Acanthocephala based on analysis of 18S ribosomal RNA gene sequences %A García-Varela,M %A Pérez-Ponce de León,G. %A de la Torre,P %A Cummings, Michael P. %A Sarma,SS %A Laclette,J. P %X Acanthocephala (thorny-headed worms) is a phylum of endoparasites of vertebrates and arthropods, included among the most phylogenetically basal tripoblastic pseudocoelomates. The phylum is divided into three classes: Archiacanthocephala, Palaeacanthocephala, and Eoacanthocephala. These classes are distinguished by morphological characters such as location of lacunar canals, persistence of ligament sacs in females, number and type of cement glands in males, number and size of proboscis hooks, host taxonomy, and ecology. To understand better the phylogenetic relationships within Acanthocephala, and between Acanthocephala and Rotifera, we sequenced the nearly complete 18S rRNA genes of nine species from the three classes of Acanthocephala and four species of Rotifera from the classes Bdelloidea and Monogononta. Phylogenetic relationships were inferred by maximum-likelihood analyses of these new sequences and others previously determined. The analyses showed that Acanthocephala is the sister group to a clade including Eoacanthocephala and Palaeacanthocephala. Archiacanthocephala exhibited a slower rate of evolution at the nucleotide level, as evidenced by shorter branch lengths for the group. We found statistically significant support for the monophyly of Rotifera, represented in our analysis by species from the clade Eurotatoria, which includes the classes Bdelloidea and Monogononta. Eurotatoria also appears as the sister group to Acanthocephala. %B J Mol Evol %V 50 %P 532 - 540 %8 2000/06// %G eng %N 6 %0 Journal Article %J Proceedings of The 13th International Software/Internet Quality Week %D 2000 %T A planning-based approach to GUI testing %A Memon, Atif M. %A Pollack,M. E %A Soffa,M. L %B Proceedings of The 13th International Software/Internet Quality Week %8 2000/// %G eng %0 Journal Article %J The Journal of Cell BiologyJ Cell Biol %D 2000 %T Pre-Messenger RNA Processing Factors in the Drosophila Genome %A Mount, Stephen M. %A Salz,Helen K. %X In eukaryotes, messenger RNAs are generated by a process that includes coordinated splicing and 3′ end formation. Factors essential for the splicing of mRNA precursors (pre-mRNA) in eukaryotes have been identified primarily through the study of nuclear extracts derived from mammalian cells and Saccharomyces cerevisiae genetics. Here, we identify homologues of most known pre-mRNA processing factors in the recently completed sequence of the Drosophila genome. The set of proteins required for RNA processing shows remarkably little variation among eukaryotic species, and individual proteins are highly conserved. In general, proteins involved in the mechanics of RNA processing are even more conserved than proteins involved in the interpretation of RNA processing signals. The genome does not appear to contain a gene for the U11 RNA, or for a protein unique to the U11 snRNP, which raises the possibility that the U12-dependent spliceosome functions without U11 in Drosophila. %B The Journal of Cell BiologyJ Cell Biol %V 150 %P F37-F44 - F37-F44 %8 2000/07/24/ %@ 0021-9525, 1540-8140 %G eng %U http://jcb.rupress.org/content/150/2/F37 %N 2 %R 10.1083/jcb.150.2.F37 %0 Journal Article %J Journal of the American Society for Information Science %D 2000 %T Previews and overviews in digital libraries: Designing surrogates to support visual information seeking %A Greene,Stephan %A Marchionini,Gary %A Plaisant, Catherine %A Shneiderman, Ben %X To aid designers of digital library interfaces, we present a framework for the design of information representations in terms of previews and overviews. Previews and overviews are graphic or textual representations of information abstracted from primary information objects. Previews act as surrogates for one or a few objects and overviews represent collections of objects. A design framework is elaborated in terms of the following three dimensions: (1) what information objects are available to users, (2) how information objects are related and displayed, and (3) how users can manipulate information objects. When utilized properly, previews and overviews allow users to rapidly discriminate objects of interest from those not of interest, and to more fully understand the scope and nature of digital libraries. This article presents a definition of previews and overviews in context, provides design guidelines, and describes four example applications. %B Journal of the American Society for Information Science %V 51 %P 380 - 393 %8 2000/01/01/ %@ 1097-4571 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/(SICI)1097-4571(2000)51:4%3C380::AID-ASI7%3E3.0.CO;2-5/abstract;jsessionid=E15C609DE95671E0E91A862B8AFD1CC6.d03t01 %N 4 %R 10.1002/(SICI)1097-4571(2000)51:4<380::AID-ASI7>3.0.CO;2-5 %0 Journal Article %J ACM SIGCOMM Computer Communication Review %D 2000 %T A protocol-independent technique for eliminating redundant network traffic %A Spring, Neil %A Wetherall,D. %B ACM SIGCOMM Computer Communication Review %V 30 %P 87 - 95 %8 2000/// %G eng %N 4 %0 Journal Article %J International Journal of Computational Geometry and Applications %D 2000 %T Quantile approximation for robust statistical estimation and k-enclosing problems %A Mount, Dave %A Netanyahu,N. S %A Piatko,C. D %A Silverman,R. %A Wu,A. Y %X Given a set P of n points in Rd, a fundamental problem in computational geometryis concerned with finding the smallest shape of some type that encloses all the points of P. Well-known instances of this problem include finding the smallest enclosing box, minimum volume ball, and minimum volume annulus. In this paper we consider the following variant: Given a set of n points in Rd, find the smallest shape in question that contains at least k points or a certain quantile of the data. This type of problem is known as a k-enclosing problem. We present a simple algorithmic framework for computing quantile approximations for the minimum strip, ellipsoid, and annulus containing a given quantile of the points. The algorithms run in O(n log n) time. %B International Journal of Computational Geometry and Applications %V 10 %P 593 - 608 %8 2000/// %G eng %N 6 %0 Conference Paper %B IEEE INFOCOM 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %D 2000 %T Receiver based management of low bandwidth access links %A Spring, Neil %A Chesire,M. %A Berryman,M. %A Sahasranaman,V. %A Anderson,T. %A Bershad,B. %K Bandwidth %K buffer storage %K bulk-transfer applications %K complex Web page %K congestion control policy %K Delay %K dynamically loadable Linux kernel module %K information resources %K interactive network %K Internet %K Kernel %K link utilization %K Linux %K low-bandwidth access links %K mixed traffic load %K packet latency %K queue length %K queueing theory %K receive socket buffer sizes %K receiver-based management %K response time %K short flow prioritizing %K Size control %K Sockets %K subscriber loops %K TCP flow control %K telecommunication congestion control %K telecommunication network management %K Telecommunication traffic %K Testing %K Throughput %K Transport protocols %K Unix %K Web pages %X In this paper, we describe a receiver-based congestion control policy that leverages TCP flow control mechanisms to prioritize mixed traffic loads across access links. We manage queueing at the access link to: (1) improve the response time of interactive network applications; (2) reduce congestion-related packet losses; while (3) maintaining high throughput for bulk-transfer applications. Our policy controls queue length by manipulating receive socket buffer sizes. We have implemented this solution in a dynamically loadable Linux kernel module, and tested it over low-bandwidth links. Our approach yields a 7-fold improvement in packet latency over an unmodified system while maintaining 94% link utilization. In the common case, congestion-related packet losses at the access link can be eliminated. Finally, by prioritizing short flows, we show that our system reduces the time to download a complex Web page during a large background transfer by a factor of two %B IEEE INFOCOM 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %I IEEE %V 1 %P 245-254 vol.1 - 245-254 vol.1 %8 2000/// %@ 0-7803-5880-5 %G eng %R 10.1109/INFCOM.2000.832194 %0 Journal Article %J ACM SIGCHI Bulletin %D 2000 %T Report on closing the digital divide: meeting of the Deparment of Commerce, Washington, DC December 9, 1999 %A Shneiderman, Ben %B ACM SIGCHI Bulletin %V 32 %P 43 - 44 %8 2000/04// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/360405.360445 %N 2 %R 10.1145/360405.360445 %0 Journal Article %J Interactions %D 2000 %T Research Alerts %E Shneiderman, Ben %B Interactions %V 7 %P 9 - 17 %8 2000/11// %@ 1072-5520 %G eng %U http://doi.acm.org/10.1145/352580.352583 %N 6 %R 10.1145/352580.352583 %0 Journal Article %J Circuits and Systems I: Fundamental Theory and Applications, IEEE Transactions on %D 2000 %T Resynchronization for multiprocessor DSP systems %A Bhattacharyya, Shuvra S. %A Sriram,S. %A Lee,E. A %B Circuits and Systems I: Fundamental Theory and Applications, IEEE Transactions on %V 47 %P 1597 - 1609 %8 2000/// %G eng %N 11 %0 Journal Article %J Multimedia Systems %D 2000 %T Retrieval scheduling for collaborative multimedia presentations %A Bai,Ping %A Prabhakaran,B. %A Srinivasan, Aravind %X The single-system approach is no longer sufficient to handle the load on popular Internet servers, especially for those offering extensive multimedia content. Such services have to be replicated to enhance their availability, performance, and reliability. In a highly replicated and available environment, server selection is an important issue. In this paper, we propose an application-layer broker (ALB) for this purpose. ALB employs a content-based, client-centric approach to negotiate with the servers and to identify the best server for the requested objects. ALB aims to maximize client buffer utilization in order to efficiently handle dynamic user interactions such as skip, reverse presentation, go back in time. We also present details of a collaborative multimedia presentation platform that we have developed based on ALB. %B Multimedia Systems %V 8 %P 146 - 155 %8 2000/// %@ 0942-4962 %G eng %U http://dx.doi.org/10.1007/s005300050157 %N 2 %0 Book Section %B Active NetworksActive Networks %D 2000 %T Scalable Resource Control in Active Networks %A Anagnostakis,Kostas %A Hicks, Michael W. %A Ioannidis,Sotiris %A Keromytis,Angelos %A Smith,Jonathan %E Yasuda,Hiroshi %X The increased complexity of the service model relative to store-and-forward routers has made resource management one of the paramount concerns in active networking research and engineering. In this paper,we address two major challenges in scaling resource management-to-many-node active networks. The first is the use of market mechanisms and trading amongst nodes and programs with varying degrees of competition and cooperation to provide a scalable approach to managing active network resources. The second is the use of a trust-management architecture to ensure that the participants in the resource management marketplace have a policy-driven “rule of law” in which marketplace decisions can be made and relied upon. We have used lottery scheduling and the Keynote trust-management system for our implementation, for which we provide some initial performance indications. %B Active NetworksActive Networks %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1942 %P 343 - 357 %8 2000/// %@ 978-3-540-41179-6 %G eng %U http://dx.doi.org/10.1007/3-540-40057-5_25 %0 Journal Article %J IEEE Communications Magazine %D 2000 %T Secure quality of service handling: SQoSH %A Alexander,D. S %A Arbaugh, William A. %A Keromytis,A. D %A Muir,S. %A Smith,J. M %K Acceleration %K Access control %K active networks %K ALIEN active loader %K Clocks %K Computer network management %K cryptographic credentials %K cryptography %K customized networking services %K Data security %K Data structures %K denial-of-service attacks %K interfaces %K Kernel %K loaded modules %K network resources %K network traffic %K open signaling %K packet switching %K Piglet lightweight device kernel %K programmable network element %K programmable network infrastructures %K Programming profession %K Proposals %K quality of service %K remote invocation %K resource control %K restricted control of quality of service %K SANE %K scheduling %K scheduling discipline %K secure active network environment architecture %K secure quality of service handling %K security infrastructure %K security risks %K SQoSH %K SwitchWare architecture %K telecommunication security %K tuning knobs %K virtual clock %X Proposals for programmable network infrastructures, such as active networks and open signaling, provide programmers with access to network resources and data structures. The motivation for providing these interfaces is accelerated introduction of new services, but exposure of the interfaces introduces many new security risks. We describe some of the security issues raised by active networks. We then describe our secure active network environment (SANE) architecture. SANE was designed as a security infrastructure for active networks, and was implemented in the SwitchWare architecture. SANE restricts the actions that loaded modules can perform by restricting the resources that can be named; this is further extended to remote invocation by means of cryptographic credentials. SANE can be extended to support restricted control of quality of service in a programmable network element. The Piglet lightweight device kernel provides a “virtual clock” type of scheduling discipline for network traffic, and exports several tuning knobs with which the clock can be adjusted. The ALIEN active loader provides safe access to these knobs to modules that operate on the network element. Thus, the proposed SQoSH architecture is able to provide safe, secure access to network resources, while allowing these resources to be managed by end users needing customized networking services. A desirable consequence of SQoSH's integration of access control and resource control is that a large class of denial-of-service attacks, unaddressed solely with access control and cryptographic protocols, can now be prevented %B IEEE Communications Magazine %V 38 %P 106 - 112 %8 2000/04// %@ 0163-6804 %G eng %N 4 %R 10.1109/35.833566 %0 Conference Paper %B CHI '00 extended abstracts on Human factors in computing systems %D 2000 %T Simulation based learning environments and the use of learning histories %A Rose,A. %A Salter,R. %A Keswani,S. %A Kositsyna,N. %A Plaisant, Catherine %A Rubloff,G. %A Shneiderman, Ben %K education %K engineering %K History %K learning %K simulation %X We have developed an application framework for constructing simulation-based learning environments using dynamic simulations and visualizations to represent realistic time-dependent behavior. The development environment is described and many examples are given. In particular we will focus on the learning historian which provides users and learners with a manipulatable recording of their actions which facilitates the exchange of annotated history records among peers and mentors. %B CHI '00 extended abstracts on Human factors in computing systems %S CHI EA '00 %I ACM %C New York, NY, USA %P 2 - 3 %8 2000/// %@ 1-58113-248-4 %G eng %U http://doi.acm.org/10.1145/633292.633294 %R 10.1145/633292.633294 %0 Report %D 2000 %T A SIMULATION ENVIRONMENT FOR EVOLVING MULTIAGENT COMMUNICATION %A Reggia, James A. %A Schultz,Reiner %A Uriagereka,Juan %A Wilkinson,Jerry %K Technical Report %X A simulation environment has been created to support study of emergentcommunication. Multiple agents exist in a two-dimensional world where they must find food and avoid predators. While non-communicating agents may survive, the world is configured so that survival and fitness can be enhanced through the use of inter-agent communication. The goal with this version of the simulator is to determine conditions under which simple communication (signaling) emerges and persists during an evolutionary process. (Also cross-referenced as UMIACS-TR-2000-64) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-2000-64 %8 2000/09/15/ %G eng %U http://drum.lib.umd.edu//handle/1903/1101 %0 Conference Paper %B Proceedings of the working conference on Advanced visual interfaces %D 2000 %T Snap-together visualization: a user interface for coordinating visualizations via relational schemata %A North,Chris %A Shneiderman, Ben %K Coordination %K Information Visualization %K multiple views %K relational database %K tight coupling %K user interface %K user study %X Multiple coordinated visualizations enable users to rapidly explore complex information. However, users often need unforeseen combinations of coordinated visualizations that are appropriate for their data. Snap-Together Visualization enables data users to rapidly and dynamically mix and match visualizations and coordinations to construct custom exploration interfaces without programming. Snap's conceptual model is based on the relational database model. Users load relations into visualizations then coordinate them based on the relational joins between them. Users can create different types of coordinations such as: brushing, drill down, overview and detail view, and synchronized scrolling. Visualization developers can make their independent visualizations snap-able with a simple API.Evaluation of Snap revealed benefits, cognitive issues, and usability concerns. Data savvy users were very capable and thrilled to rapidly construct powerful coordinated visualizations. A snapped overview and detail-view coordination improved user performance by 30-80%, depending on task. %B Proceedings of the working conference on Advanced visual interfaces %S AVI '00 %I ACM %C New York, NY, USA %P 128 - 135 %8 2000/// %@ 1-58113-252-2 %G eng %U http://doi.acm.org/10.1145/345513.345282 %R 10.1145/345513.345282 %0 Journal Article %J International Journal of Human-Computer Studies %D 2000 %T Snap-together visualization: can users construct and operate coordinated visualizations? %A North,Chris %A Shneiderman, Ben %X Multiple coordinated visualizations enable users to rapidly explore complex information. However, users often need unforeseen combinations of coordinated visualizations. Snap-together visualization (Snap) enables users to rapidly and dynamically construct coordinated–visualization interfaces, customized for their data, without programming. Users U001load data into desired visualizations, then construct coordinations between them for brushing and linking, overview and detail view, drill down, etc. Snap formalizes a conceptual model of visualization coordination based on the relational data model. Visualization developers can easily Snap-enable their independent visualizations using a simple API.Empirical evaluation reveals benefits, cognitive issues and usability concerns with coordination concepts and Snap. Two user studies explore coordination construction and operation. Data-savvy users successfully, enthusiastically and rapidly constructed powerful coordinated–visualization interfaces of their own. Operating an overview-and-detail coordination reliably improved user performance by 30–80% over detail-only and uncoordinated interfaces for most tasks. %B International Journal of Human-Computer Studies %V 53 %P 715 - 739 %8 2000/11// %@ 1071-5819 %G eng %U http://www.sciencedirect.com/science/article/pii/S107158190090418X %N 5 %R 10.1006/ijhc.2000.0418 %0 Journal Article %J Int'l Journal of Human-Computer Studies special issue on Empirical Studies of Information Visualization %D 2000 %T Snap-together visualization: Evaluating coordination usage and construction %A North,C. %A Shneiderman, Ben %X Multiple coordinated visualizations enable users to rapidly explore complex information. However, users often need unforeseen combinations of coordinated visualizations. Snap-Together Visualization is a conceptual model, based on the relational model, and system to enable users to quickly coordinate otherwise-independent visualization tools. Users construct customized browsing environments with coordinations for selecting, navigating, and loading data, without programming.Evaluation revealed benefits, cognitive issues, and usability concerns with coordination concepts and the Snap system. Two user studies explore the value of coordination usage and the learnability of coordination construction. The overview and detail-view coordination improved user performance by 30-80%, depending on task. Data savvy users were very capable and thrilled to rapidly construct powerful coordinated visualizations. %B Int'l Journal of Human-Computer Studies special issue on Empirical Studies of Information Visualization %V 53 %P 715 - 739 %8 2000/// %G eng %0 Conference Paper %B Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval %D 2000 %T Structured translation for cross-language information retrieval %A Sperer,Ruth %A Oard, Douglas %X The paper introduces a query translation model that reflects the structure of the cross-language information retrieval task. The model is based on a structured bilingual dictionary in which the translations of each term are clustered into groups with distinct meanings. Query translation is modeled as a two-stage process, with the system first determining the intended meaning of a query term and then selecting translations appropriate to that meaning that might appear in the document collection. An implementation of structured translation based on automatic dictionary clustering is described and evaluated by using Chinese queries to retrieve English documents. Structured translation achieved an average precision that was statistically indistinguishable from Pirkola's technique for very short queries, but Pirkola's technique outperformed structured translation on long queries. The paper concludes with some observations on future work to improve retrieval effectiveness and on other potential uses of structured translation in interactive cross-language retrieval applications. %B Proceedings of the 23rd annual international ACM SIGIR conference on Research and development in information retrieval %S SIGIR '00 %I ACM %C New York, NY, USA %P 120 - 127 %8 2000/// %@ 1-58113-226-3 %G eng %U http://doi.acm.org/10.1145/345508.345562 %R 10.1145/345508.345562 %0 Conference Paper %B Proceedings of the 33rd Annual Hawaii International Conference on System Sciences, 2000 %D 2000 %T Supporting creativity with powerful composition tools for artifacts and performances %A Shneiderman, Ben %K artifacts %K Artificial intelligence %K composition tools %K Computer science %K Context modeling %K creativity %K creativity support %K Educational institutions %K Information retrieval %K Laboratories %K Manufacturing %K music composition %K performances %K Software design %K software libraries %K software tools %K US Department of Transportation %X Modern software such as word processors, slide preparation/presentation tools, or music composition packages are designed to produce artifacts or performances. Now, some designers are expanding their goals to include creativity support into their software. This essay builds on the genex framework for creativity, which has four phases and eight activities. It focuses on the composition activity by considering tools to support initiation, revision, and evaluation. Some existing tools provide partial support that suggests ways in which to develop more ambitious low, middle, and high level tools. The goal is to enable more people to be more creative more of the time. %B Proceedings of the 33rd Annual Hawaii International Conference on System Sciences, 2000 %I IEEE %8 2000/01/04/7 %@ 0-7695-0493-0 %G eng %R 10.1109/HICSS.2000.926896 %0 Journal Article %J SIGCHI BULLETIN %D 2000 %T A Trip Report on Creativity and Cognition 1999: An Insider's Report on a Far Out Conference %A Shneiderman, Ben %A Hewett,T. %B SIGCHI BULLETIN %V 32 %P 43 - 46 %8 2000/// %G eng %N 1 %0 Journal Article %J Communications of the ACM %D 2000 %T Universal usability %A Shneiderman, Ben %B Communications of the ACM %V 43 %P 84 - 91 %8 2000/05// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/332833.332843 %N 5 %R 10.1145/332833.332843 %0 Conference Paper %B Proceedings of the 2000 ACM workshops on Multimedia - MULTIMEDIA '00 %D 2000 %T Using Hilbert curve in image storing and retrieving %A Song,Zhexuan %A Roussopoulos, Nick %B Proceedings of the 2000 ACM workshops on Multimedia - MULTIMEDIA '00 %C Los Angeles, California, United States %P 167 - 170 %8 2000/// %G eng %U http://dl.acm.org/citation.cfm?id=357921 %R 10.1145/357744.357921 %0 Conference Paper %B Proceedings of the thirty-second annual ACM symposium on Theory of computing %D 2000 %T The value of strong inapproximability results for clique %A Srinivasan, Aravind %K P vs. NP %K Approximation algorithms %K Clique %K inapproximabilit %K Independent set %K Packing integer programs %K random sampling %B Proceedings of the thirty-second annual ACM symposium on Theory of computing %S STOC '00 %I ACM %C New York, NY, USA %P 144 - 152 %8 2000/// %@ 1-58113-184-4 %G eng %U http://doi.acm.org/10.1145/335305.335322 %R 10.1145/335305.335322 %0 Conference Paper %B 2000 IEEE International Conference on Multimedia and Expo, 2000. ICME 2000 %D 2000 %T Visualization methods for personal photo collections: browsing and searching in the PhotoFinder %A Kang,Hyunmo %A Shneiderman, Ben %K Browsing %K Computer science %K digital photo library %K Displays %K drag-and-drop interface %K dynamic query %K Educational institutions %K Filters %K Histograms %K Image retrieval %K personal computing %K personal photo collections %K photo collection management %K PhotoFinder %K Prototypes %K query preview %K scatter plot thumbnail display %K Scattering %K searching %K software libraries %K software tools %K User interfaces %K visual Boolean query interfaces %K visual databases %K Visualization %K visualization methods %X Software tools for personal photo collection management are proliferating, but they usually have limited searching and browsing functions. We implemented the PhotoFinder prototype to enable non-technical users of personal photo collections to search and browse easily. PhotoFinder provides a set of visual Boolean query interfaces, coupled with dynamic query and query preview features. It gives users powerful search capabilities. Using a scatter plot thumbnail display and drag-and-drop interface, PhotoFinder is designed to be easy to use for searching and browsing photos %B 2000 IEEE International Conference on Multimedia and Expo, 2000. ICME 2000 %I IEEE %V 3 %P 1539-1542 vol.3 - 1539-1542 vol.3 %8 2000/// %@ 0-7803-6536-4 %G eng %R 10.1109/ICME.2000.871061 %0 Conference Paper %B Proceedings of the fifth ACM conference on Digital libraries %D 2000 %T Visualizing digital library search results with categorical and hierarchical axes %A Shneiderman, Ben %A Feldman,David %A Rose,Anne %A Grau,Xavier Ferré %K categorical axes %K digital libraries %K Graphical user interfaces %K hierarchy %K hieraxes %K Information Visualization %X Digital library search results are usually shown as a textual list, with 10-20 items per page. Viewing several thousand search results at once on a two-dimensional display with continuous variables is a promising alternative. Since these displays can overwhelm some users, we created a simplified two-dimensional display that uses categorical and hierarchical axes, called hieraxes. Users appreciate the meaningful and limited number of terms on each hieraxis. At each grid point of the display we show a cluster of color-coded dots or a bar chart. Users see the entire result set and can then click on labels to move down a level in the hierarchy. Handling broad hierarchies and arranging for imposed hierarchies led to additional design innovations. We applied hieraxes to a digital video library of science topics used by middle school teachers, a legal information system, and a technical library using the ACM Computing Classification System. Feedback from usability testing with 32 subjects revealed strengths and weaknesses. %B Proceedings of the fifth ACM conference on Digital libraries %S DL '00 %I ACM %C New York, NY, USA %P 57 - 66 %8 2000/// %@ 1-58113-231-X %G eng %U http://doi.acm.org/10.1145/336597.336637 %R 10.1145/336597.336637 %0 Book Section %B Approximation Algorithms for Combinatorial Optimization %D 2000 %T Wavelength Rerouting in Optical Networks, or the Venetian Routing Problem %A Caprara,Alberto %A Italiano,Giuseppe %A Mohan,G. %A Panconesi,Alessandro %A Srinivasan, Aravind %E Jansen,Klaus %E Khuller, Samir %X Wavelength rerouting has been suggested as a viable and cost-effective method to improve the blocking performance of wavelength-routed Wavelength-Division Multiplexing (WDM) networks. This method leads to the following combinatorial optimization problem, dubbed Venetian Routing. Given a directed multigraph G along with two vertices s and t and a collection of pairwise arc-disjoint paths, we wish to find an st -path which arc-intersects the smallest possible number of such paths. In this paper we prove the computational hardness oft his problem even in various special cases, and present several approximation algorithms for its solution. In particular we show a non-trivial connection between Venetian Routing and Label Cover. %B Approximation Algorithms for Combinatorial Optimization %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1913 %P 71 - 84 %8 2000/// %@ 978-3-540-67996-7 %G eng %U http://dx.doi.org/10.1007/3-540-44436-X_9 %0 Conference Paper %B Geoscience and Remote Sensing Symposium, 2000. Proceedings. IGARSS 2000. IEEE 2000 International %D 2000 %T Web based progressive transmission for browsing remotely sensed imagery %A Mareboyana,M. %A Srivastava,S. %A JaJa, Joseph F. %K based %K decomposition;geophysical %K image;model-based %K interest;remote %K interest;vector %K mapping;user %K mapping;vector %K measurement %K of %K processing;geophysical %K processing;image %K Progressive %K quantisation;wavelet %K quantization;wavelet %K refinement;region %K regions %K representation;land %K representation;remote %K sensing;scalar;terrain %K sensing;terrain %K signal %K specified %K specified;user %K surface;large %K technique;image %K techniques;image %K transforms; %K transmission;browsing;geophysical %K VQ;progressive %K Web %X This paper describes an image representation technique that entails progressive refinement of user specified regions of interest (ROI) of large images. Progressive refinement to original quality can be accomplished in theory. However, due to heavy burden on storage resources for the authors' applications, they restrict the refinement to about 25% of the original data resolution. A wavelet decomposition of the data combined with scalar and vector quantization (VQ) of the high frequency components and JPEG/DCT compression of low frequency component is used as representation framework. Their software will reconstruct the region selected by the user from its wavelet decomposition such that it fills up the preview window with the appropriate subimages at the desired resolution, including full resolution stored for preview. Further refinement from the first preview can be obtained progressively by transmitting high frequency coefficients from low resolution to high resolution which are compressed by variant of vector quantization called model-based VQ (MVQ). The user will have an option for progressive build up of the ROIs until full resolution stored or terminate the transmission at any time during the progressive refinement %B Geoscience and Remote Sensing Symposium, 2000. Proceedings. IGARSS 2000. IEEE 2000 International %V 2 %P 591 -593 vol.2 - 591 -593 vol.2 %8 2000/// %G eng %R 10.1109/IGARSS.2000.861640 %0 Journal Article %J Readings in information visualization %D 1999 %T 1D, 2D, 3D %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %P 57 - 61 %8 1999/// %G eng %0 Journal Article %J ACM SIGMOD Record %D 1999 %T The active MultiSync controller of the cubetree storage organization %A Roussopoulos, Nick %A Kotidis,Y. %A Sismanis,Y. %B ACM SIGMOD Record %V 28 %P 582 - 583 %8 1999/// %G eng %N 2 %0 Conference Paper %B Proceedings of the eighth international conference on World Wide Web %D 1999 %T Agents vs. direct manipulation: what's best to disentangle the Web? %A Sánchez,J.A. %A Etzioni,O. %A Kautz,H. %A Lieberman,H. %A Shneiderman, Ben %B Proceedings of the eighth international conference on World Wide Web %8 1999/// %G eng %0 Journal Article %J Computer Networks: The International Journal of Computer and Telecommunications Networking %D 1999 %T Agents vs. direct manipulation: what's best to disentangle the Web? %A Sánchez,J. Alfredo %A Etzioni,Oren %A Kautz,Henry %A Lieberman,Henry %A Shneiderman, Ben %B Computer Networks: The International Journal of Computer and Telecommunications Networking %V 31 %P 1765 - 1767 %8 1999/05// %@ 1389-1286 %G eng %U http://dl.acm.org/citation.cfm?id=313009.313142 %N 11-16 %0 Journal Article %J Readings in information visualization %D 1999 %T Alternate geometry %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %P 381 - 381 %8 1999/// %G eng %0 Report %D 1999 %T An analysis of the Rayleigh-Ritz method for approximating eigenspaces %A Jia,Zhongxiao %A Stewart, G.W. %X Abstract. This paper concerns the Rayleigh–Ritz method for computing an approximation to an eigenspace X of a general matrix A from a subspace W that contains an approximation to X. The method produces a pair (N, ˜ X) that purports to approximate a pair (L, X), where X is a basis for X and AX = XL. In this paper we consider the convergence of (N, ˜ X) as the sine ɛ of the angle between X and W approaches zero. It is shown that under a natural hypothesis — called the uniform separation condition — the Ritz pairs (N, ˜ X) converge to the eigenpair (L, X). When one is concerned with eigenvalues and eigenvectors, one can compute certain refined Ritz vectors whose convergence is guaranteed, even when the uniform separation condition is not satisfied. An attractive feature of the analysis is that it does not assume that A has distinct eigenvalues or is diagonalizable. 1. %I Math. Comp %8 1999/// %G eng %0 Conference Paper %B Proceedings of the seventh ACM international conference on Multimedia (Part 2) %D 1999 %T Application-layer broker for scalable Internet services with resource reservation %A Bai,Ping %A Prabhakaran,B. %A Srinivasan, Aravind %B Proceedings of the seventh ACM international conference on Multimedia (Part 2) %S MULTIMEDIA '99 %I ACM %C New York, NY, USA %P 103 - 106 %8 1999/// %@ 1-58113-239-5 %G eng %U http://doi.acm.org/10.1145/319878.319906 %R 10.1145/319878.319906 %0 Journal Article %J Readings in information visualization %D 1999 %T Applications and implications %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %P 625 - 635 %8 1999/// %G eng %0 Journal Article %J Lectures on Approximation and Randomized Algorithms (M. Karonski and HJ Promel, editors), Series in Advanced Topics in Mathematics, Polish Scientific Publishers PWN %D 1999 %T Approximation algorithms via randomized rounding: a survey %A Srinivasan, Aravind %X Approximation algorithms provide a natural way to approach computationally hardproblems. There are currently many known paradigms in this area, including greedy al- gorithms, primal-dual methods, methods based on mathematical programming (linear and semide nite programming in particular), local improvement, and \low distortion" embeddings of general metric spaces into special families of metric spaces. Random- ization is a useful ingredient in many of these approaches, and particularly so in the form of randomized rounding of a suitable relaxation of a given problem. We survey this technique here, with a focus on correlation inequalities and their applications. %B Lectures on Approximation and Randomized Algorithms (M. Karonski and HJ Promel, editors), Series in Advanced Topics in Mathematics, Polish Scientific Publishers PWN %P 9 - 71 %8 1999/// %G eng %0 Journal Article %J Readings in information visualization %D 1999 %T Bifocal lens %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %P 331 - 332 %8 1999/// %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Browsing hierarchical data with multi-level dynamic queries %A Kumar,H. P %A Plaisant, Catherine %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %P 295 - 305 %8 1999/// %@ 9781558605336 %G eng %0 Conference Paper %D 1999 %T Building dependable distributed applications using AQUA %A Ren,J. %A Michel Cukier %A Rubel,P. %A Sanders,W. H. %A Bakken,D. E. %A Karr,D. A. %K ad hoc methods %K application programmer %K AQuA %K complexities %K computational complexity %K dependable distributed applications %K distributed CORBA application %K distributed object management %K Fault tolerance %K fault tolerant computing %K proteus %X Building dependable distributed systems using ad hoc methods is a challenging task. Without proper support, an application programmer must face the daunting requirement of having to provide fault tolerance at the application level, in addition to dealing with the complexities of the distributed application itself. This approach requires a deep knowledge of fault tolerance on the part of the application designer, and has a high implementation cost. What is needed is a systematic approach to providing dependability to distributed applications. Proteus, part of the AQuA architecture, fills this need and provides facilities to make a standard distributed CORBA application dependable, with minimal changes to an application. Furthermore, it permits applications to specify, either directly or via the Quality Objects (QuO) infrastructure, the level of dependability they expect of a remote object, and will attempt to configure the system to achieve the requested dependability level. Our previous papers have focused on the architecture and implementation of Proteus. This paper describes how to construct dependable applications using the AQuA architecture, by describing the interface that a programmer is presented with and the graphical monitoring facilities that it provides %P 189 - 196 %8 1999/// %G eng %R 10.1109/HASE.1999.809494 %0 Conference Paper %B SDIUT99 %D 1999 %T Classification of Document Page Images %A Shin,C. %A David Doermann %B SDIUT99 %P 166 - 175 %8 1999/// %G eng %0 Journal Article %J Molecular and Cellular Biochemistry %D 1999 %T A comparative study of the backbone dynamics of two closely related lipid binding proteins: Bovine heart fatty acid binding protein and porcine ileal lipid binding protein %A Lücke,Christian %A Fushman, David %A Ludwig,Christian %A Hamilton,James A. %A Sacchettini,James C. %A Rüterjans,Heinz %K Biomedical and Life Sciences %X The backbone dynamics of bovine heart fatty acid binding protein (H-FABP) and porcine ileal lipid binding protein (ILBP) were studied by 15N NMR relaxation (T1 and T2) and steady state heteronuclear 15N1H NOE measurements. The microdynamic parameters characterizing the backbone mobility were determined using the ‘model-free’ approach. For H-FABP, the non-terminal backbone amide groups display a rather compact protein structure of low flexibility. In contrast, for ILBP an increased number of backbone amide groups display unusually high internal mobility. Furthermore, the data indicate a higher degree of conformational exchange processes in the µsec-msec time range for ILBP compared to H-FABP. These results suggest significant differences in the conformational stability for these two structurally highly homologous members of the fatty acid binding protein family. %B Molecular and Cellular Biochemistry %V 192 %P 109 - 121 %8 1999/// %@ 0300-8177 %G eng %U http://www.springerlink.com/content/w2j255781546j880/abstract/ %N 1 %R 10.1023/A:1006834708786 %0 Report %D 1999 %T Comparing causal-link and propositional planners: Tradeoffs between plan length and domain size %A Memon, Atif M. %A Pollack,M. %A Soffa,M. L %X Recent studies have shown that propositional planners, which de-rive from Graphplan and SATPLAN, can generate significantly longer plans than causal-link planners. We present experimental evidence demonstrating that while this may be true, propositional planners also have important limitations relative to the causal-link planners: specifically, they can generate plans only for smaller domains, where the size of a domain is defined by the number of distinguishable objects it contains. Our experiments were conducted in the domain of code optimization, in which the states of the world represent states of the computer program code and the planning operators are the optimiza- tion operators. This domain is well-suited to studying the trade-offs between plan length and domain size, because it is straightforward to manipulate both these factors. On the basis of our experiments, we conclude that causal-link and propositional planners have complementary strengths. %I Technical Report 99-06, University of Pittsburgh, Pittsburgh %8 1999/// %G eng %0 Conference Paper %B Proceedings of the tenth annual ACM-SIAM symposium on Discrete algorithms %D 1999 %T Computing nearest neighbors for moving points and applications to clustering %A Kanungo,Tapas %A Mount, Dave %A Netanyahu,Nathan S. %A Piatko,Christine %A Silverman,Ruth %A Wu,Angela Y. %B Proceedings of the tenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '99 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 931 - 932 %8 1999/// %@ 0-89871-434-6 %G eng %U http://dl.acm.org/citation.cfm?id=314500.315095 %0 Book %D 1999 %T Content-Based Access to Multimedia Information: From Technology Trends to Sate of the Art %A Perry,B. %A Chang,S-K. %A Dinsmore,J. %A David Doermann %A Rosenfeld, A. %A Stevens,S. %I Kluwer %8 1999/// %G eng %0 Conference Paper %B Proceedings of the Ninth SIAM Conference on Parallel Processing for Scientific Computing %D 1999 %T Coupling multiple simulations via a high performance customizable database system %A Kurc,T. M %A Sussman, Alan %A Saltz, J. %B Proceedings of the Ninth SIAM Conference on Parallel Processing for Scientific Computing %8 1999/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 1999 %T Creating Creativity for Everyone: User Interfaces for Supporting Innovation %A Shneiderman, Ben %K Technical Report %X A challenge for human-computer interaction researchers and user interface designers is to construct information technologies that support creativity.This ambitious goal can be attained by building on an adequate understand-ing of creative processes. This paper offers the four-phase genex frame-work for generating excellence: - Collect: learn from previous works stored in digital libraries- Relate: consult with peers and mentors at early, middle and late stages- Create: explore, compose, and evaluate possible solutions - Donate: disseminate the results and contribute to the digital libraries Within this integrated framework, this paper proposes eight activities that require human-computer interaction research and advanced user inter-face design. A scenario about an architect illustrates the process of creative work within a genex environment. %B Institute for Systems Research Technical Reports %8 1999/// %G eng %U http://drum.lib.umd.edu/handle/1903/6010 %0 Journal Article %J Interactions %D 1999 %T Departments-Book Preview-Readings in Information Visualization: Using Vision to Think %A Card,S.K. %A Mackinlay,J. %A Shneiderman, Ben %B Interactions %V 6 %P 51 - 51 %8 1999/// %G eng %N 2 %0 Conference Paper %B Proceedings of the 1999 conference on Computer support for collaborative learning %D 1999 %T The design of history mechanisms and their use in collaborative educational simulations %A Plaisant, Catherine %A Rose,Anne %A Rubloff,Gary %A Salter,Richard %A Shneiderman, Ben %K hci %K scaffolding %K simulation %X Reviewing past events has been useful in many domains. Videotapes and flight data recorders provide invaluable technological help to sports coaches or aviation engineers. Similarly, providing learners with a readable recording of their actions may help them monitor their behavior, reflect on their progress, and experiment with revisions of their experiences. It may also facilitate active collaboration among dispersed learning communities. Learning histories can help students and professionals make more effective use of digital library searching, word processing tasks, computer-assisted design tools, electronic performance support systems, and web navigation.This paper describes the design space and discusses the challenges of implementing learning histories. It presents guidelines for creating effective implementations, and the design tradeoffs between sparse and dense history records. The paper also presents a first implementation of learning histories for a simulation-based engineering learning environment called SimPLE (Simulated Processes in a Learning Environment) for the case of a semiconductor fabrication module, and reports on early user evaluation of learning histories implemented within SimPLE. %B Proceedings of the 1999 conference on Computer support for collaborative learning %S CSCL '99 %I International Society of the Learning Sciences %8 1999/// %G eng %U http://dl.acm.org/citation.cfm?id=1150240.1150284 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: the CHI is the limit %D 1999 %T Designing PETS: a personal electronic teller of stories %A Druin, Allison %A Montemayor,Jamie %A Hendler,Jim %A McAlister,Britt %A Boltman,Angela %A Fiterman,Eric %A Plaisant,Aurelie %A Kruskal,Alex %A Olsen,Hanne %A Revett,Isabella %A Schwenn,Thomas Plaisant %A Sumida,Lauren %A Wagner,Rebecca %K children %K cooperative inquiry %K design techniques %K educational applications %K intergenerational design team %K PETS %K ROBOTICS %X We have begun the development of a new robotic pet that cansupport children in the storytelling process. Children can build their own pet by snapping together the modular animal parts of the PETS robot. After their pet is built, children can tell stories using the My Pets software. These stories can then be acted out by their robotic pet. This video paper describes the motivation for this research and the design process of our intergenerational design team in building the first PETS prototypes. We will discuss our progress to date and our focus for the future. %B Proceedings of the SIGCHI conference on Human factors in computing systems: the CHI is the limit %S CHI '99 %I ACM %C New York, NY, USA %P 326 - 329 %8 1999/// %@ 0-201-48559-1 %G eng %U http://doi.acm.org/10.1145/302979.303103 %R 10.1145/302979.303103 %0 Conference Paper %B Geoscience and Remote Sensing Symposium, 1999. IGARSS '99 Proceedings. IEEE 1999 International %D 1999 %T Developing the next generation of Earth science data systems: the Global Land Cover Facility %A Lindsay,F.E. %A Townshend,J.R.G. %A JaJa, Joseph F. %A Humphries,J. %A Plaisant, Catherine %A Shneiderman, Ben %K Computer architecture %K data archiving %K data distribution system %K Data systems %K Distributed computing %K Earth science data products %K Earth science data system %K ESIP %K geographic information system %K geographic information systems %K Geography %K geophysical measurement technique %K geophysical signal processing %K geophysical techniques %K Geoscience %K GIS %K GLCF %K Global Land Cover Facility %K High performance computing %K Indexing %K information service %K Information services %K Institute for Advanced Computer Studies %K land cover %K NASA %K next generation %K PACS %K Remote sensing %K terrain mapping %K UMIACS %K University of Maryland %K User interfaces %K web-based interface %X A recent initiative by NASA has resulted in the formation of a federation of Earth science data partners. These Earth Science Information Partners (ESIPs) have been tasked with creating novel Earth science data products and services as well as distributing new and existing data sets to the Earth science community and the general public. The University of Maryland established its ESIP activities with the creation of the Global Land Cover Facility (GLCF). This joint effort of the Institute for Advanced Computer Studies (UMIACS) and the Department of Geography has developed an operational data archiving and distribution system aimed at advancing current land cover research efforts. The success of the GLCF is tied closely to assessing user needs as well. As the timely delivery of data products to the research community. This paper discusses the development and implementation of a web-based interface that allows users to query the authors' data holdings and perform user requested processing tasks on demand. The GLCF takes advantage of a scaleable, high performance computing architecture for the manipulation of very large remote sensing data sets and the rapid spatial indexing of multiple format data types. The user interface has been developed with the cooperation of the Human-Computer Interaction Laboratory (HCIL) and demonstrates advances in spatial and temporal querying tools as well as the ability to overlay multiple raster and vector data sets. Their work provides one perspective concerning how critical earth science data may be handled in the near future by a coalition of distributed data centers %B Geoscience and Remote Sensing Symposium, 1999. IGARSS '99 Proceedings. IEEE 1999 International %I IEEE %V 1 %P 616-618 vol.1 - 616-618 vol.1 %8 1999/// %@ 0-7803-5207-6 %G eng %R 10.1109/IGARSS.1999.773583 %0 Book Section %B Materialized views %D 1999 %T The dynamic homefinder: evaluating dynamic queries in a real-estate information exploration system %A Williamson,C. %A Shneiderman, Ben %B Materialized views %I MIT Press %C Cambridge, MA, USA %P 125 - 139 %8 1999/// %@ 0-262-57122-6 %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Dynamic queries %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %P 235 - 235 %8 1999/// %G eng %0 Book %D 1999 %T Efficient preconditioning of the linearized Navier-Stokes equations %A Silvester, D. %A Elman, Howard %A Kay, D. %A Wathen, A. %I Citeseer %8 1999/// %G eng %0 Journal Article %J International Journal on Digital Libraries %D 1999 %T The end of zero-hit queries: query previews for NASA’s Global Change Master Directory %A Greene,Stephan %A Tanin,Egemen %A Plaisant, Catherine %A Shneiderman, Ben %A Olsen,Lola %A Major,Gene %A Johns,Steve %X The Human-Computer Interaction Laboratory (HCIL) of the University of Maryland and NASA have collaborated over three years to refine and apply user interface research concepts developed at HCIL in order to improve the usability of NASA data services. The research focused on dynamic query user interfaces, visualization, and overview + preview designs. An operational prototype, using query previews, was implemented with NASA’s Global Change Master Directory (GCMD), a directory service for earth science datasets. Users can see the histogram of the data distribution over several attributes and choose among attribute values. A result bar shows the cardinality of the result set, thereby preventing users from submitting queries that would have zero hits. Our experience confirmed the importance of metadata accuracy and completeness. The query preview interfaces make visible the problems or gaps in the metadata that are undetectable with classic form fill-in interfaces. This could be seen as a problem, but we think that it will have a long-term beneficial effect on the quality of the metadata as data providers will be compelled to produce more complete and accurate metadata. The adaptation of the research prototype to the NASA data required revised data structures and algorithms. %B International Journal on Digital Libraries %V 2 %P 79 - 90 %8 1999/// %@ 1432-5012 %G eng %U http://dx.doi.org/10.1007/s007990050039 %N 2 %0 Conference Paper %B Proceedings of the CNDS Conference (Society for Computer Simulation) %D 1999 %T End-to-end request handling in distributed Video-on-Demand systems %A Mundur, Padma %A Simon,R. %A Sood,A. %B Proceedings of the CNDS Conference (Society for Computer Simulation) %P 151 - 157 %8 1999/// %G eng %0 Journal Article %J Lecture notes in computer science %D 1999 %T On the evolution of reactive components: A process-algebraic approach %A M\üller-Olm,M. %A Steffen,B. %A Cleaveland, Rance %B Lecture notes in computer science %P 161 - 175 %8 1999/// %G eng %0 Journal Article %J PROCEEDINGS OF THE ANNUAL MEETING-AMERICAN SOCIETY FOR INFORMATION SCIENCE %D 1999 %T An Exploratory Study of Video Browsing, User Interface Designs and Research Methodologies: Effectiveness in Information Seeking Tasks %A Tse,T. %A Vegh,S. %A Shneiderman, Ben %A Marchionini,G. %X The purpose of this exploratory study is to develop research methods to compare the effectiveness of two video browsing interface designs, or surrogates one static (storyboard) and one dynamic (slide show) on two distinct information seeking tasks (gist determination and object recognition). Although video data is multimodal, potentially consisting of images, speech, sound, and text, the surrogates tested depend on image data only and use key frames or stills extracted from source video. A test system was developed to determine the effects of different key frame displays on user performance in specified information seeking tasks. The independent variables were interface display and task type. The dependent variables were task accuracy and subjective satisfaction. Covariates included spatial visual ability and time-to-completion. The study used a repeated block factorial 2x2 design; each of 20 participants interacted with all four interface-task combinations. No statistically significant results for task accuracy were found. Statistically significant differences were found, however, for user satisfaction with the display types: users assessed the static display to be "easier" to use than the dynamic display for both task types, even though there were no performance differences. This methodological approach provides a useful way to learn about the relationship between surrogate types and user tasks during video browsing. %B PROCEEDINGS OF THE ANNUAL MEETING-AMERICAN SOCIETY FOR INFORMATION SCIENCE %V 36 %P 681 - 692 %8 1999/// %G eng %0 Conference Paper %D 1999 %T Fault injection based on a partial view of the global state of a distributed system %A Michel Cukier %A Chandra,R. %A Henke,D. %A Pistole,J. %A Sanders,W. H. %K bounding technique %K clock synchronization %K distributed programming %K distributed software systems %K fault injection %K Loki %K post-runtime analysis %K program testing %K program verification %K software reliability %K Synchronisation %X This paper describes the basis for and preliminary implementation of a new fault injector, called Loki, developed specifically for distributed systems. Loki addresses issues related to injecting correlated faults in distributed systems. In Loki, fault injection is performed based on a partial view of the global state of an application. In particular, facilities are provided to pass user-specified state information between nodes to provide a partial view of the global state in order to try to inject complex faults successfully. A post-runtime analysis, using an off-line clock synchronization and a bounding technique, is used to place events and injections on a single global time-line and determine whether the intended faults were properly injected. Finally, observations containing successful fault injections are used to estimate specified dependability measures. In addition to describing the details of our new approach, we present experimental results obtained from a preliminary implementation in order to illustrate Loki's ability to inject complex faults predictably %P 168 - 177 %8 1999/// %G eng %R 10.1109/RELDIS.1999.805093 %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Fisheye views %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %P 311 - 311 %8 1999/// %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Focus+ context %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %P 306 - 309 %8 1999/// %G eng %0 Journal Article %J Numerische Mathematik %D 1999 %T Four algorithms for the efficient computation of truncated pivoted QR approximations to a sparse matrix %A Stewart, G.W. %X In this paper we propose four algorithms to compute truncated pivoted QR approximations to a sparse matrix. Three are based on the Gram–Schmidt algorithm and the other on Householder triangularization. All four algorithms leave the original matrix unchanged, and the only additional storage requirements are arrays to contain the factorization itself. Thus, the algorithms are particularly suited to determining low-rank approximations to a sparse matrix. %B Numerische Mathematik %V 83 %P 313 - 323 %8 1999/// %G eng %N 2 %R 10.1007/s002110050451 %0 Conference Paper %B Simulation of Semiconductor Processes and Devices, 1999. SISPAD '99. 1999 International Conference on %D 1999 %T Gate leakage current simulation by Boltzmann transport equation and its dependence on the gate oxide thickness %A Han,Zhiyi %A Lin,Chung-Kai %A Goldsman,N. %A Mayergoyz, Issak D %A Yu,S. %A Stettler,M. %K 30 %K angstrom;60 %K angstrom;Boltzmann %K Bias %K calculations;leakage %K charges;spherical %K component;tunneling %K current %K currents;semiconductor %K dependence;method %K dependence;MOSFET;barrier %K device %K effect;distribution %K equation;DC %K equation;MOSFET;WKB %K function;first %K harmonic %K image %K leakage %K lowering %K method %K method;gate %K model;thermionic %K models;tunnelling; %K of %K oxide %K principle %K probability;Boltzmann %K simulation;gate %K thickness %K transport %K WKB %X As device dimensions shrink toward 0.1 mu;m, gate oxides are becoming so thin that MOSFET gate leakage current and oxide degradation are becoming limiting issues. We provide a more rigorous way to calculate gate leakage current. To achieve this we build upon the Spherical Harmonic Method of modeling, which deterministically solves the Boltzmann equation for an entire device. The method gives the distribution function and is 1000 times faster than MC. Once the distribution function is calculated, the tunneling probability is derived from the first principle WKB method. The barrier lowering effect is accounted for by the method of image charges. We calculate gate leakage current as a function of DC bias. The thermionic and tunneling components are compared at different DC bias points. The dependence of gate leakage current on gate oxide thickness is simulated %B Simulation of Semiconductor Processes and Devices, 1999. SISPAD '99. 1999 International Conference on %P 247 - 250 %8 1999/// %G eng %R 10.1109/SISPAD.1999.799307 %0 Conference Paper %B Geoscience and Remote Sensing Symposium, 1999. IGARSS '99 Proceedings. IEEE 1999 International %D 1999 %T A hierarchical data archiving and processing system to generate custom tailored products from AVHRR data %A Kalluri, SNV %A Zhang,Z. %A JaJa, Joseph F. %A Bader, D.A. %A Song,H. %A El Saleous,N. %A Vermote,E. %A Townshend,J.R.G. %K archiving;image %K AVHRR;GIS;PACS;custom %K data %K image;land %K image;remote %K mapping; %K mapping;PACS;geophysical %K measurement %K PROCESSING %K processing;geophysical %K product;data %K remote %K scheme;infrared %K sensing;optical %K sensing;terrain %K signal %K surface;multispectral %K system;indexing %K tailored %K technique;hierarchical %K techniques;remote %X A novel indexing scheme is described to catalogue satellite data on a pixel basis. The objective of this research is to develop an efficient methodology to archive, retrieve and process satellite data, so that data products can be generated to meet the specific needs of individual scientists. When requesting data, users can specify the spatial and temporal resolution, geographic projection, choice of atmospheric correction, and the data selection methodology. The data processing is done in two stages. Satellite data is calibrated, navigated and quality flags are appended in the initial processing. This processed data is then indexed and stored. Secondary processing such as atmospheric correction and projection are done after a user requests the data to create custom made products. By dividing the processing in to two stages saves time, since the basic processing tasks such as navigation and calibration which are common to all requests are not repeated when different users request satellite data. The indexing scheme described can be extended to allow fusion of data sets from different sensors %B Geoscience and Remote Sensing Symposium, 1999. IGARSS '99 Proceedings. IEEE 1999 International %V 5 %P 2374 -2376 vol.5 - 2374 -2376 vol.5 %8 1999/// %G eng %R 10.1109/IGARSS.1999.771514 %0 Journal Article %J ACM SIGCAS Computers and Society %D 1999 %T Human values and the future of technology: a declaration of responsibility %A Shneiderman, Ben %X We can make a difference in shaping the future by ensuring that computers "serve human needs (Mumford, 1934)." By making explicit the enduring values that we hold dear we can guide computer system designers and developers for the next decade, century, and thereafter. After setting our high-level goals we can pursue the components and seek the process for fulfilling them. High-level goals might include peace, excellent health care, adequate nutrition, accessible education, communication, freedom of expression, support for creative exploration, safety, and socially constructive entertainment. Computer technology can help attain these high-level goals if we clearly state measurable objectives, obtain participation of professionals, and design effective human-computer interfaces. Design considerations include adequate attention to individual differences among users, support of social and organizational structures, design for reliability and safety, provision of access by the elderly, handicapped, or illiterate, and appropriate user controlled adaptation. With suitable theories and empirical research we can achieve ease of learning, rapid performance, low error rates, and good retention over time, while preserving high subjective satisfaction. To raise the consciousness of designers and achieve these goals, we must generate an international debate, stimulate discussions within organizations, and interact with other intellectual communities. This paper calls for a focus on the "you" and "I" in developing improved user interface (UI) research and systems, offers a Declaration of Responsibility, and proposes a Social Impact Statement for major computing projects. %B ACM SIGCAS Computers and Society %V 29 %P 5 - 9 %8 1999/09// %@ 0095-2737 %G eng %U http://doi.acm.org/10.1145/572183.572185 %N 3 %R 10.1145/572183.572185 %0 Journal Article %J IEEE Computer Graphics and Applications %D 1999 %T Human-centered computing, online communities, and virtual environments %A Brown,J. R %A van Dam,A. %A Earnshaw,R. %A Encarnacao,J. %A Guedj,R. %A Preece,J. %A Shneiderman, Ben %A Vince,J. %K Books %K Collaboration %K Collaborative work %K Conferences %K EC/NSF joint Advanced Research Workshop %K Feeds %K Human computer interaction %K human-centered computing %K Internet %K Joining materials %K Laboratories %K Online communities %K Research initiatives %K USA Councils %K User interfaces %K Virtual environment %K virtual environments %K Virtual reality %X This report summarizes results of the first EC/NSF joint Advanced Research Workshop, which identified key research challenges and opportunities in information technology. The group agreed that the first joint research workshop should concentrate on the themes of human-centered computing and VEs. Human-centered computing is perceived as an area of strategic importance because of the move towards greater decentralization and decomposition in the location and provision of computation. The area of VEs is one where increased collaboration should speed progress in solving some of the more intractable problems in building effective applications %B IEEE Computer Graphics and Applications %V 19 %P 70 - 74 %8 1999/12//Nov %@ 0272-1716 %G eng %N 6 %R 10.1109/38.799742 %0 Journal Article %J SIAM Journal on Computing %D 1999 %T Improved Approximation Guarantees for Packing and Covering Integer Programs %A Srinivasan, Aravind %X Several important NP-hard combinatorial optimization problems can be posed as packing/covering integer programs; the randomized rounding technique of Raghavan and Thompson is a powerful tool with which to approximate them well. We present one elementary unifying property of all these integer linear programs and use the FKG correlation inequality to derive an improved analysis of randomized rounding on them. This yields a pessimistic estimator, thus presenting deterministic polynomial-time algorithms for them with approximation guarantees that are significantly better than those known. %B SIAM Journal on Computing %V 29 %P 648 - 648 %8 1999/// %@ 00975397 %G eng %U http://link.aip.org/link/SMJCAT/v29/i2/p648/s1&Agg=doi %N 2 %R 10.1137/S0097539796314240 %0 Journal Article %J Readings in information visualization %D 1999 %T Information workspaces %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %P 513 - 514 %8 1999/// %G eng %0 Journal Article %J Readings in information visualization %D 1999 %T Infosphere, workspace, tools, objects %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %P 463 - 464 %8 1999/// %G eng %0 Conference Paper %B IPPS %D 1999 %T Infrastructure for building parallel database systems for multi-dimensional data %A Chang,C. %A Ferreira,R. %A Sussman, Alan %A Saltz, J. %B IPPS %P 582 - 582 %8 1999/// %G eng %0 Conference Paper %B IEEE International Conference on Multimedia Computing and Systems, 1999 %D 1999 %T Integrated admission control in hierarchical video-on-demand systems %A Mundur, Padma %A Simon,R. %A Sood,A. %K Admission control %K Bandwidth %K blocking %K Computer science %K Design methodology %K end-to-end system %K hierarchical video-on-demand systems %K integrated admission control %K Intelligent networks %K Load management %K Motion pictures %K Network servers %K network subsystem %K performance %K Performance analysis %K performance evaluation %K quality of service %K request handling %K resource allocation %K Resource management %K simulation %K storage subsystem %K video on demand %K video servers %X We develop a unified model of a hierarchical video-on-demand (VoD) system by integrating the storage and the network subsystems. Rather than restricting the analysis to an isolated subsystem the performance of the VoD system is analyzed as an end-to-end system. On a system-wide basis, request handling and admission control policies are designed to minimize global performance metrics. Through our simulation, we compare different request handling policies and show that a hierarchical VoD architecture with request handling that allows retrials at more than one resource will minimize overall blocking %B IEEE International Conference on Multimedia Computing and Systems, 1999 %I IEEE %V 1 %P 220-225 vol.1 - 220-225 vol.1 %8 1999/07// %@ 0-7695-0253-9 %G eng %R 10.1109/MMCS.1999.779196 %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Interaction %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 231 - 234 %8 1999/// %@ 1-55860-533-9 %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Interactive analysis %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 261 - 262 %8 1999/// %@ 1-55860-533-9 %G eng %0 Journal Article %J ACM Trans. Inf. Syst. %D 1999 %T Interface and data architecture for query preview in networked information systems %A Plaisant, Catherine %A Shneiderman, Ben %A Doan,Khoa %A Bruns,Tom %K direct manipulation %K dynamic query %K EOSDIS %K graphical user interface %K query preview %K query refinement %K science data %X There are numerous problems associated with formulating queries onnetworked information systems. These include increased data volume and complexity, accompanied by slow network access. This article proposes a new approach to a network query user interfaces that consists of two phases: query preview and query refinement. This new approach is based on the concepts of dynamic queries and query previews, which guides users in rapidly and dynamically eliminating undesired records, reducing the data volume to a manageable size, and refining queries locally before submission over a network. Examples of two applications are given: a Restaurant Finder and a prototype for NASA's Earth Observing Systems Data Information Systems (EOSDIS). Data architecture is discussed, and user feedback is presented. %B ACM Trans. Inf. Syst. %V 17 %P 320 - 341 %8 1999/07// %@ 1046-8188 %G eng %U http://doi.acm.org/10.1145/314516.314522 %N 3 %R 10.1145/314516.314522 %0 Conference Paper %B CHI '99 extended abstracts on Human factors in computing systems %D 1999 %T An international SIGCHI research agenda %A Scholtz,Jean C. %A Muller,Michael %A Novick,David %A Olsen,Jr.,Dan R. %A Shneiderman, Ben %A Wharton,Cathleen %K agenda %K human-computer %K interaction %K research %X The goal of this workshop is to articulate an HCI research agenda statement, along with plans for continued refinement with the greater CHI community and plans for disseminating the information beyond the CHI community. There are been several similar prior efforts, including:• 1998 Universal Access program [2]• European community report, "Building the European Information Society for Us All" [1];• CHI 97 workshop [5] and paper [6], "HCI Research and Practice Agenda based on Human Needs …";• CHI 96 Workshop, "CHI Ten-Year View…" [3];• 1995 US National Science Foundation report, "New Directions in Human-Computer Interaction …" [9];• 1995 report from the US National Research Council [7];• 1991 report, "…HCI… Serving Human Needs" [4] %B CHI '99 extended abstracts on Human factors in computing systems %S CHI EA '99 %I ACM %C New York, NY, USA %P 171 - 171 %8 1999/// %@ 1-58113-158-5 %G eng %U http://doi.acm.org/10.1145/632716.632821 %R 10.1145/632716.632821 %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Internet and Infosphere %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 465 - 468 %8 1999/// %@ 1-55860-533-9 %G eng %0 Journal Article %J Empirical Software Engineering %D 1999 %T Introduction to Special Issue on Usability Engineering %A Scholtz,Jean %A Shneiderman, Ben %B Empirical Software Engineering %V 4 %P 5 - 10 %8 1999/// %@ 1382-3256 %G eng %U http://dx.doi.org/10.1023/A:1009852413783 %N 1 %0 Conference Paper %B Neural Networks, 1999. IJCNN'99. International Joint Conference on %D 1999 %T Lesion effects in a bihemispheric letter-identification model %A Shevtsova,N. %A Reggia, James A. %B Neural Networks, 1999. IJCNN'99. International Joint Conference on %V 1 %P 215 - 218 %8 1999/// %G eng %0 Book Section %B Readings in Information Visualization: Using Vision to ThinkReadings in Information Visualization: Using Vision to Think %D 1999 %T LifeLines: Visualizing Personal Histories %A Kumar,H. P %A Plaisant, Catherine %A Shneiderman, Ben %B Readings in Information Visualization: Using Vision to ThinkReadings in Information Visualization: Using Vision to Think %I Morgan Kaufmann Publishers Inc. %P 285 - 285 %8 1999/// %@ 9781558605336 %G eng %0 Journal Article %J International Journal on Software Tools for Technology Transfer (STTT) %D 1999 %T Local model checking and protocol analysis %A Du,X. %A Smolka,S. A %A Cleaveland, Rance %B International Journal on Software Tools for Technology Transfer (STTT) %V 2 %P 219 - 241 %8 1999/// %G eng %N 3 %0 Conference Paper %B GREC %D 1999 %T Model Based Graphics Recognition %A Stuckelberg,M.V. %A David Doermann %B GREC %P 136 - 143 %8 1999/// %G eng %0 Conference Paper %B Neural Networks, 1999. IJCNN'99. International Joint Conference on %D 1999 %T A model of lateralization and asymmetries in cortical maps %A Levitan,S. %A Stoica,I. %A Reggia, James A. %B Neural Networks, 1999. IJCNN'99. International Joint Conference on %V 1 %P 121 - 124 %8 1999/// %G eng %0 Journal Article %J Computer Methods in Applied Mechanics and Engineering %D 1999 %T Modified streamline diffusion schemes for convection-diffusion problems %A Shih,Yin-Tzer %A Elman, Howard %X We consider the design of robust and accurate finite element approximation methods for solving convection--diffusion problems. We develop some two-parameter streamline diffusion schemes with piecewise bilinear (or linear) trial functions and show that these schemes satisfy the necessary conditions for L2-uniform convergence of order greater than 1/2 introduced by Stynes and Tobiska M. Stynes and L. Tobiska, Necessary L2-uniform convergence conditions for difference schemes for two dimensional convection--diffusion problems, Comput. Math. Applic. 29 (1995) 45-53. For smooth problems, the schemes satisfy error bounds of the form O(h)u2 in an energy norm. In addition, extensive numerical experiments show that they effectively reproduce boundary layers and internal layers caused by discontinuities on relatively coarse grids, without any requirements on alignment of flow and grid. %B Computer Methods in Applied Mechanics and Engineering %V 174 %P 137 - 151 %8 1999/05/04/ %@ 0045-7825 %G eng %U http://www.sciencedirect.com/science/article/pii/S0045782598002837 %N 1-2 %R 16/S0045-7825(98)00283-7 %0 Journal Article %J D-Lib MagazineD-Lib Magazine %D 1999 %T Multilingual Information Discovery and AccesS (MIDAS): A Joint ACM DL'99/ ACM SIGIR'99 Workshop. %A Oard, Douglas %A Peters,Carol %A Ruiz,Miguel %A Frederking,Robert %A Klavans,Judith %A Sheridan,Paraic %K Distributed computing %K Electronic Media %X Discusses a multidisciplinary workshop that addressed issues concerning internationally distributed information networks. Highlights include multilingual information access in media other than character-coded text; cross-language information retrieval and multilingual metadata; and evaluation of multilingual systems. (LRW) %B D-Lib MagazineD-Lib Magazine %V 5 %P 1 - 12 %8 1999/// %@ ISSN-1082-9873 %G eng %U http://www.eric.ed.gov/ERICWebPortal/detail?accno=EJ601937 %N 10 %0 Conference Paper %B Readings in information visualization %D 1999 %T Multiple dimensions> 3 %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 93 - 95 %8 1999/// %@ 1-55860-533-9 %G eng %0 Conference Paper %B Document Analysis and Recognition, 1999. ICDAR '99. Proceedings of the Fifth International Conference on %D 1999 %T On musical score recognition using probabilistic reasoning %A Stuckelberg,M.V. %A David Doermann %K analysis;document %K attribute %K class;document %K descriptive %K document %K engine;local %K estimation;musical %K grammar;attribute %K grammars;character %K handling; %K image %K interpretation;stochastic %K Markov %K mechanisms;music;uncertainty %K model;global %K modeling %K models;inference %K parameter %K processing;image %K propagation;explicit %K reasoning;scanned %K recognition;document %K recognition;end-to-end %K recognition;inference %K recognition;probabilistic %K score %K structure;hidden %K Uncertainty %X We present a probabilistic framework for document analysis and recognition and illustrate it on the problem of musical score recognition. Our system uses an explicit descriptive model of the document class to find the most likely interpretation of a scanned document image. In contrast to the traditional pipeline architecture, we carry out all stages of the analysis with a single inference engine, allowing for an end-to-end propagation of the uncertainty. The global modeling structure is similar to a stochastic attribute grammar, and local parameters are estimated using hidden Markov models %B Document Analysis and Recognition, 1999. ICDAR '99. Proceedings of the Fifth International Conference on %P 115 - 118 %8 1999/09// %G eng %R 10.1109/ICDAR.1999.791738 %0 Journal Article %J Components and Packaging Technologies, IEEE Transactions on %D 1999 %T Nano-to-millimeter scale integrated systems %A Last,H.R. %A Deeds,M. %A Garvick,D. %A Kavetsky,R. %A Sandborn,P. A. %A Magrab,E. B. %A Gupta,S.K. %K fabrication technology %K In2m %K integrated nano to millimeter system %K micromechanical devices %K nanotechnology %X Over the last several years various industries have been developing nano, micro, and millimeter scale technologies, which have resulted in components ranging from quantum transistors, to widely commercialized integrated circuits, to microelectromechanical sensors. A common emphasis of these fabrication industries has been on the integration of different functions in miniaturized systems; however, the technology currently used to realize these systems is monolithic. A unique class of hybrid technology systems is Integrated nano to millimeter (In2m) systems. An In2m system typically has components spanning multiple sizes, diverse technology domains, and mixtures of electrical, mechanical, thermal, chemical, fluidic, and biological functions %B Components and Packaging Technologies, IEEE Transactions on %V 22 %P 338 - 343 %8 1999/06// %@ 1521-3331 %G eng %N 2 %R 10.1109/6144.774758 %0 Conference Paper %B Third International Conference on Computational Intelligence and Multimedia Applications, 1999. ICCIMA '99. Proceedings %D 1999 %T Network service selection for distributed multimedia applications %A Simon,R. %A Sood,A. %A Mundur, Padma %K Admission control %K Application software %K application-adequate end-to-end service %K Bandwidth %K Communication system traffic control %K Computer science %K Delay %K distributed processing %K end-to-end delivery delay control %K flexibility %K high-bandwidth distributed multimedia applications %K interactive multimedia %K multimedia systems %K network service selection %K network throughput %K nonpreemptive earliest deadline first %K queueing theory %K Regulators %K system support %K telecommunication services %K Throughput %K Traffic control %K weighted fair queueing %X An important question in the development of system support for distributed multimedia is the type of network service offered to applications. This paper compares two network service disciplines: weighted fair queueing (WFQ) and non-preemptive earliest deadline first (NEDF). We show that, for a broad class of high-bandwidth distributed multimedia applications, WFQ outperforms NEDF in terms of network throughput while still providing an application-adequate end-to-end service. This result holds despite the fact that NEDF offers applications far greater flexibility in terms of control over end-to-end delivery delay %B Third International Conference on Computational Intelligence and Multimedia Applications, 1999. ICCIMA '99. Proceedings %I IEEE %P 388 - 392 %8 1999/// %@ 0-7695-0300-4 %G eng %R 10.1109/ICCIMA.1999.798561 %0 Journal Article %J Journal of Cognitive Neuroscience %D 1999 %T A neural network model of lateralization during letter identification %A Shevtsova,N. %A Reggia, James A. %B Journal of Cognitive Neuroscience %V 11 %P 167 - 181 %8 1999/// %G eng %N 2 %0 Conference Paper %B Proceedings of the tenth annual ACM-SIAM symposium on Discrete algorithms %D 1999 %T New algorithmic aspects of the Local Lemma with applications to routing and partitioning %A Leighton,Tom %A Rao,Satish %A Srinivasan, Aravind %B Proceedings of the tenth annual ACM-SIAM symposium on Discrete algorithms %S SODA '99 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 643 - 652 %8 1999/// %@ 0-89871-434-6 %G eng %U http://dl.acm.org/citation.cfm?id=314500.314886 %0 Journal Article %J Parallel Processing Letters %D 1999 %T Object-relational queries into multidimensional databases with the active data repository %A Ferreira,R. %A Kurc, T. %A Beynon, M. %A Chang,C. %A Sussman, Alan %A Saltz,J. H %B Parallel Processing Letters %V 9 %P 173 - 195 %8 1999/// %G eng %N 2 %0 Report %D 1999 %T On Orthogonalization in the Inverse Power Method %A Stewart, G.W. %K Technical Report %X When the inverse power method is used to compute eigenvectors of asymmetric matrix corresponding to close eigenvalues, the computed eigenvectors may not be orthogonal. The cure for the problem is to orthogonalize the vectors using the Gram--Schmidt algorithm. In this note it is shown that the orthogonalization process does not cause the quality of the eigenvectors to deteriorate. Also cross-referenced as UMIACS-TR-99-64 %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-99-64 %8 1999/10/13/ %G eng %U http://drum.lib.umd.edu/handle/1903/1038 %0 Journal Article %J Technical Reports from UMIACS, UMIACS-TR-99-16 %D 1999 %T OSMA Software Program: Domain Analysis Guidebook %A Basili, Victor R. %A Seaman,Carolyn %A Tesoriero,Roseanne %A Zelkowitz, Marvin V %K Technical Report %X Domain analysis is the process of identifying and organizing knowledge abouta class of problems. This guidebook presents a method of performing experience domain analysis in software development organizations. The purpose of the guidebook is to facilitate the reader in characterizing two given development environments, applying domain analysis to model each, and then applying an evaluation process, based upon the Goal/Metric/Paradigm, to transfer a given development technology from one of the environments to the other. This guidebook describes this process and gives an example of its use within NASA. Also cross-referenced as UMIACS-TR-99-16 %B Technical Reports from UMIACS, UMIACS-TR-99-16 %8 1999/03/31/ %G eng %U http://drum.lib.umd.edu/handle/1903/1000 %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Overview+ detail %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %P 285 - 286 %8 1999/// %G eng %0 Book %D 1999 %T Perceptual Organization in Computer Vision %A Boyer,K.L. %A Sarkar,S. %A Feldman,J. %A Granlund,G. %A Horaud,R. %A Hutchinson,S. %A Jacobs, David W. %A Kak,A. %A Lowe,D. %A Malik,J. %I Academic Press %8 1999/// %G eng %0 Conference Paper %B Proceedings of the 13th international conference on Supercomputing %D 1999 %T Performance impact of proxies in data intensive client-server applications %A Beynon,M. D %A Sussman, Alan %A Saltz, J. %B Proceedings of the 13th international conference on Supercomputing %P 383 - 390 %8 1999/// %G eng %0 Journal Article %J Empirical Software Engineering %D 1999 %T Perspective-based Usability Inspection: An Empirical Validation of Efficacy %A Zhang,Zhijun %A Basili, Victor R. %A Shneiderman, Ben %X Inspection is a fundamental means of achieving software usability. Past research showed that the current usability inspection techniques were rather ineffective. We developed perspective-based usability inspection, which divides the large variety of usability issues along different perspectives and focuses each inspection session on one perspective. We conducted a controlled experiment to study its effectiveness, using a post-test only control group experimental design, with 24 professionals as subjects. The control group used heuristic evaluation, which is the most popular technique for usability inspection. The experimental design and the results are presented, which show that inspectors applying perspective-based inspection not only found more usability problems related to their assigned perspectives, but also found more overall problems. Perspective-based inspection was shown to be more effective for the aggregated results of multiple inspectors, finding about 30% more usability problems for 3 inspectors. A management implication of this study is that assigning inspectors more specific responsibilities leads to higher performance. Internal and external threats to validity are discussed to help better interpret the results and to guide future empirical studies. %B Empirical Software Engineering %V 4 %P 43 - 69 %8 1999/// %@ 1382-3256 %G eng %U http://dx.doi.org/10.1023/A:1009803214692 %N 1 %0 Conference Paper %B Readings in information visualization %D 1999 %T Physical data %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 37 - 38 %8 1999/// %@ 1-55860-533-9 %G eng %0 Journal Article %J Technical Reports from UMIACS %D 1999 %T Pixel Data Access for End-User Programming and Graphical Macros %A Potter,Richard %A Shneiderman, Ben %K Technical Report %X Pixel Data Access is an interprocess communication technique that enablesusers of graphical user interfaces to automate certain tasks. By accessing the contents of the display buffer, users can search for pixel representations of interface elements, and then initiate actions such as mouse clicks and keyboard entries. While this technique has limitations it offers users of current systems some unusually powerful features that are especially appealing in the area of end-user programming. Also cross-referenced as UMIACS-TR-99-27 %B Technical Reports from UMIACS %8 1999/05/25/ %G eng %U http://drum.lib.umd.edu/handle/1903/1009 %0 Conference Paper %B The Eighth International Symposium on High Performance Distributed Computing, 1999. Proceedings %D 1999 %T Predicting the CPU availability of time-shared Unix systems on the computational grid %A Wolski,R. %A Spring, Neil %A Hayes,J. %K accuracy %K Application software %K Autocorrelation %K Availability %K Central Processing Unit %K computational grid %K correlation methods %K CPU availability prediction %K CPU resources predictability %K CPU sensor %K Dynamic scheduling %K grid computing %K Load forecasting %K long-range autocorrelation dependence %K medium-term forecasts %K network operating systems %K Network Weather Service %K NWS %K performance evaluation %K self-similarity degree %K short-term forecasts %K successive CPU measurements %K Time measurement %K Time sharing computer systems %K time-shared Unix systems %K time-sharing systems %K Unix %K Unix load average %K vmstat utility %K Weather forecasting %X Focuses on the problem of making short- and medium-term forecasts of CPU availability on time-shared Unix systems. We evaluate the accuracy with which availability can be measured using the Unix load average, the Unix utility “vmstat” and the Network Weather Service (NWS) CPU sensor that uses both. We also examine the autocorrelation between successive CPU measurements to determine their degree of self-similarity. While our observations show a long-range autocorrelation dependence, we demonstrate how this dependence manifests itself in the short- and medium-term predictability of the CPU resources in our study %B The Eighth International Symposium on High Performance Distributed Computing, 1999. Proceedings %I IEEE %P 105 - 112 %8 1999/// %@ 0-7803-5681-0 %G eng %R 10.1109/HPDC.1999.805288 %0 Conference Paper %D 1999 %T Proteus: a flexible infrastructure to implement adaptive fault tolerance in AQuA %A Sabnis,C. %A Michel Cukier %A Ren,J. %A Rubel,P. %A Sanders,W. H. %A Bakken,D. E. %A Karr,D. %K adaptive fault tolerance %K AQuA %K commercial off-the-shelf components %K CORBA applications %K cost %K dependable distributed systems %K distributed object management %K object replication %K proteus %K reconfigurable architectures %K Runtime %K Software architecture %K software fault tolerance %X Building dependable distributed systems from commercial off-the-shelf components is of growing practical importance. For both cost and production reasons, there is interest in approaches and architectures that facilitate building such systems. The AQuA architecture is one such approach; its goal is to provide adaptive fault tolerance to CORBA applications by replicating objects, providing a high-level method for applications to specify their desired dependability, and providing a dependability manager that attempts to reconfigure a system at runtime so that dependability requests are satisfied. This paper describes how dependability is provided in AQuA. In particular it describes Proteus, the part of AQuA that dynamically manages replicated distributed objects to make them dependable. Given a dependability request, Proteus chooses a fault tolerance approach and reconfigures the system to try to meet the request. The infrastructure of Proteus is described in this paper, along with its use in implementing active replication and a simple dependability policy %P 149 - 168 %8 1999/11// %G eng %R 10.1109/DCFTS.1999.814294 %0 Journal Article %J SIAM Journal on Scientific Computing %D 1999 %T The QLP Approximation to the Singular Value Decomposition %A Stewart, G.W. %K pivoted QR decomposition %K QLP decomposition %K rank determination %K singular value decomposition %X In this paper we introduce a new decomposition called the pivoted QLP decomposition. It is computed by applying pivoted orthogonal triangularization to the columns of the matrix X in question to get an upper triangular factor R and then applying the same procedure to the rows of R to get a lower triangular matrix L. The diagonal elements of R are called the R-values of X; those of L are called the L-values. Numerical examples show that the L-values track the singular values of X with considerable fidelity---far better than the R-values. At a gap in the L-values the decomposition provides orthonormal bases of analogues of row, column, and null spaces provided of X. The decomposition requires no more than twice the work required for a pivoted QR decomposition. The computation of R and L can be interleaved, so that the computation can be terminated at any suitable point, which makes the decomposition especially suitable for low-rank determination problems. The interleaved algorithm also suggests a new, efficient 2-norm estimator. %B SIAM Journal on Scientific Computing %V 20 %P 1336 - 1348 %8 1999/// %G eng %U http://link.aip.org/link/?SCE/20/1336/1 %N 4 %R 10.1137/S1064827597319519 %0 Conference Paper %B In Proceedings of SC99: High Performance Networking and Computing %D 1999 %T Querying Very Large Multi-dimensional Datasets in ADR %A Chialin,T. K %A Kurc, T. %A Chang,C. %A Ferreira,R. %A Sussman, Alan %A Saltz, J. %B In Proceedings of SC99: High Performance Networking and Computing %8 1999/// %G eng %0 Book %D 1999 %T Readings in Information Visualization: Using Vision to Think %A Card,Stuart K. %A Mackinlay,Jock D. %A Shneiderman, Ben %K Computer Graphics %K Computers / Computer Engineering %K Computers / Computer Graphics %K Computers / Computer Graphics / General %K Computers / Computer Science %K Computers / General %K Computers / Information Technology %K Computers / Information Theory %K Computers / Intelligence (AI) & Semantics %K Computers / Social Aspects / Human-Computer Interaction %K IMAGE PROCESSING %K Information display systems %K Information Visualization %K Psychology / General %K Visualization %X This groundbreaking book defines the emerging field of information visualization and offers the first-ever collection of the classic papers of the discipline, with introductions and analytical discussions of each topic and paper. The authors' intention is to present papers that focus on the use of visualization to discover relationships, using interactive graphics to amplify thought. This book is intended for research professionals in academia and industry; new graduate students and professors who want to begin work in this burgeoning field; professionals involved in financial data analysis, statistics, and information design; scientific data managers; and professionals involved in medical, bioinformatics, and other areas.* Full-color reproduction throughout* Author power team - an exciting and timely collaboration between the field's pioneering, most-respected names* The only book on Information Visualization with the depth necessary for use as a text or as a reference for the information professional* Text includes the classic source papers as well as a collection of cutting edge work %I Morgan Kaufmann %8 1999/01/25/ %@ 9781558605336 %G eng %0 Conference Paper %B Proceedings of the 1999 ACM/IEEE conference on Supercomputing (CDROM) %D 1999 %T Running EveryWare on the computational grid %A Wolski,R. %A Brevik,J. %A Krintz,C. %A Obertelli,G. %A Spring, Neil %A Su,A. %B Proceedings of the 1999 ACM/IEEE conference on Supercomputing (CDROM) %P 6–es - 6–es %8 1999/// %G eng %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: the CHI is the limit %D 1999 %T Single display groupware: a model for co-present collaboration %A Stewart,J. %A Bederson, Benjamin B. %A Druin, Allison %B Proceedings of the SIGCHI conference on Human factors in computing systems: the CHI is the limit %P 286 - 293 %8 1999/// %G eng %0 Journal Article %J Languages and Compilers for Parallel Computing %D 1999 %T SIPR: A new framework for generating efficient code for sparse matrix computations %A Pugh, William %A Shpeisman,T. %B Languages and Compilers for Parallel Computing %P 213 - 229 %8 1999/// %G eng %0 Journal Article %J Technical Reports from UMIACS %D 1999 %T Snap-Together Visualization: Coordinating Multiple Views to Explore Information %A North,Chris %A Shneiderman, Ben %K Technical Report %X Information visualizations with multiple coordinated views enable users torapidly explore complex data and discover relationships. However, it is usually difficult for users to find or create the coordinated visualizations they need. Snap-Together Visualization allows users to coordinate multiple views that are customized to their needs. Users query their relational database and load results into desired visualizations. Then they specify coordinations between visualizations for selecting, navigating, or re-querying. Developers can make independent visualization tools 'snap-able' by including a few hooks. Also cross-referenced as UMIACS-TR-99-28 %B Technical Reports from UMIACS %8 1999/06/08/ %G eng %U http://drum.lib.umd.edu/handle/1903/1010 %0 Journal Article %J University of California, Berkeley %D 1999 %T A Study of the Structure of the Web %A Deshpande, Amol %A Huang,R. %A Raman,V. %A Riggs,T. %A Song,D. %A Subramanian,L. %X The World Wide Web is a huge, growing repository of information on a wide range of topics. It is alsobecoming important, commercially and sociologically, as a place of human interaction within different communities. In this paper we present an experimental study of the structure of the Web. We analyze link topologies of various communities, and patterns of mirroring of content, on 1997 and 1999 snapshots of the Web. Our results give insight into patterns of interaction within communities and how they evolve, as well as patterns of data replication. We also describe the techniques we have developed for performing complex processing on this large data set, and our experiences in doing so. We present new algorithms for finding partial and complete mirrors in URL hierarchies; these are also of independent interest for search and redirection. In order to study and visualize link topologies of different communities, we have developed techniques to compact these large link graphs without much information loss. %B University of California, Berkeley %8 1999/// %G eng %0 Journal Article %J American Mathematical Society, Series in Discrete Mathematics and Theoretical Computer Science %D 1999 %T A survey of the role of multicommodity flow and randomization in network design and routing %A Srinivasan, Aravind %B American Mathematical Society, Series in Discrete Mathematics and Theoretical Computer Science %V 43 %P 271 - 302 %8 1999/// %G eng %0 Conference Paper %B Proceedings of the 1999 workshop on new paradigms in information visualization and manipulation in conjunction with the eighth ACM internation conference on Information and knowledge management %D 1999 %T Temporal, geographical and categorical aggregations viewed through coordinated displays: a case study with highway incident data %A Fredrikson,Anna %A North,Chris %A Plaisant, Catherine %A Shneiderman, Ben %X Information visualization displays can hold a limited number of data points, typically a few thousand, before they get crowded. One way to solve this problem with larger data sets is to create aggregates. Aggregations were used together with the Snap-Together Visualization system to coordinate the visual displays of aggregates and their content. If two displays each hold one thousand items then rapid access and visibility can be maintained for a million points. This paper presents examples based on a database of highway incident data. %B Proceedings of the 1999 workshop on new paradigms in information visualization and manipulation in conjunction with the eighth ACM internation conference on Information and knowledge management %S NPIVM '99 %I ACM %C New York, NY, USA %P 26 - 34 %8 1999/// %@ 1-58113-254-9 %G eng %U http://doi.acm.org/10.1145/331770.331780 %R 10.1145/331770.331780 %0 Journal Article %J PROCEEDINGS OF THE ANNUAL MEETING-AMERICAN SOCIETY FOR INFORMATION SCIENCE %D 1999 %T Temporal visualization for legal case histories %A Harris,C. %A Allen,R.B. %A Plaisant, Catherine %A Shneiderman, Ben %X This paper discusses visualization of legal information using a tool for temporal information called LifeLines. The direct and indirect histories of cases can become very complex. We explored ways that LifeLines could aid in viewing the links between the original case and the direct and indirect histories. The Apple Computer, Inc. v. Microsoft Corporation and Hewlett Packard Company case is used to illustrate the prototype. For example, if users want to find out how the rulings or statutes changed throughout this case, they could retrieve this information within a single display. Using the timeline, users could also choose at which point in time they would like to begin viewing the case. LifeLines support various views of a case's history. For instance, users can view the trial history of a case, the references involved in a case, and citations made to a case. The paper describes improvements to LifeLines that could help in providing a more useful visualization of case history. %B PROCEEDINGS OF THE ANNUAL MEETING-AMERICAN SOCIETY FOR INFORMATION SCIENCE %V 36 %P 271 - 279 %8 1999/// %G eng %0 Journal Article %J Readings in information visualization: using vision to think %D 1999 %T TennisViewer: A Browser for Competition Trees %A Johnson,B. %A Shneiderman, Ben %A Baker,MJ %A Eick,SG %B Readings in information visualization: using vision to think %P 149 - 149 %8 1999/// %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Text in 1D %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 411 - 412 %8 1999/// %@ 1-55860-533-9 %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Text in 2D %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 431 - 431 %8 1999/// %@ 1-55860-533-9 %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Text in 3D %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 440 - 441 %8 1999/// %@ 1-55860-533-9 %G eng %0 Book Section %B Readings in information visualizationReadings in information visualization %D 1999 %T Text in 3D+ time %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualizationReadings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 451 - 451 %8 1999/// %@ 1-55860-533-9 %G eng %0 Conference Paper %B CHI '99 extended abstracts on Human factors in computing systems %D 1999 %T Trust me, I'm accountable: trust and accountability online %A Friedman,Batya %A Thomas,John C. %A Grudin,Jonathan %A Nass,Clifford %A Nissenbaum,Helen %A Schlager,Mark %A Shneiderman, Ben %K accountability %K anonymity %K Communication %K computers and society %K ethics %K Internet %K media effects %K privacy %K Reciprocity %K repute %K social actors %K social capital %K social impacts %K trust %K value-sensitive design %K wired world %K WWW %X We live in an increasingly wired world. According to Robert Putnam, people are spending less time in persistent personal face to face interactions and more time in pursuits such as watching TV and using the Internet. At the same time, independently measured "social capital" -- the extent to which we trust and work for a common good -- is declining. In this panel, we explore: the impacts of electronic media on trust and accountability; whether and how electronic media can be designed and used to increase deserved trust and accountability; the relationship between protecting privacy and increasing the efficacy of communication; and how people's tendency to treat computers as social actors impacts these issues. In brief, how can modern technology enhance humanity's humanity? %B CHI '99 extended abstracts on Human factors in computing systems %S CHI EA '99 %I ACM %C New York, NY, USA %P 79 - 80 %8 1999/// %@ 1-58113-158-5 %G eng %U http://doi.acm.org/10.1145/632716.632766 %R 10.1145/632716.632766 %0 Journal Article %J Tech Report HCIL-99-06 %D 1999 %T Understanding Patterns of User Visits to Web Sites: Interactive Starfield Visualization of WWW Log Data-Short Version %A Hochheiser,H. %A Shneiderman, Ben %X HTTP server log files provide Web site operators with substantial detail regarding the visitors to their sites. Interest in interpreting this data has spawned an active market for software packages that summarize and analyze this data, providing histograms, pie graphs, and other charts summarizing usage patterns. While useful, these summaries obscure useful information and restrict users to passive interpretation of static displays.Interactive starfield visualizations can be used to provide users with greater abilities to interpret and explore web log data. By combining two-dimensional displays of thousands of individual access requests, color and size coding for additional attributes, and facilities for zooming and filtering, these visualizations provide capabilities for examining data that exceed those of traditional web log analysis tools. We introduce a series of interactive starfield visualizations, which can be used to explore server data across various dimensions. Possible uses of these visualizations are discussed, and difficulties of data collection, presentation, and interpretation are explored. %B Tech Report HCIL-99-06 %8 1999/// %G eng %0 Conference Paper %B Proceedings of the 3rd conference on Creativity & cognition %D 1999 %T User interfaces for creativity support tools %A Shneiderman, Ben %B Proceedings of the 3rd conference on Creativity & cognition %S C&C '99 %I ACM %C New York, NY, USA %P 15 - 22 %8 1999/// %@ 1-58113-078-3 %G eng %U http://doi.acm.org/10.1145/317561.317565 %R 10.1145/317561.317565 %0 Conference Paper %B Software Engineering, 1999. Proceedings of the 1999 International Conference on %D 1999 %T Using a goal-driven approach to generate test cases for GUIs %A Memon, Atif M. %A Pollack,M. E %A Soffa,M. L %K Artificial intelligence %K automatic test case generation %K goal state %K goal-driven approach %K Graphical user interfaces %K GUIs %K hierarchical planning operators %K initial state %K Microsoft Word-Pad %K operators %K planning (artificial intelligence) %K program testing %K software %K verification commands %X The widespread use of GUIs for interacting with software is leading to the construction of more and more complex GUIs. With the growing complexity comes challenges in testing the correctness of a GUI and the underlying software. We present a new technique to automatically generate test cases for GUIs that exploits planning, a well developed and used technique in artificial intelligence. Given a set of operators, an initial state and a goal state, a planner produces a sequence of the operators that will change the initial state to the goal state. Our test case generation technique first analyzes a GUI and derives hierarchical planning operators from the actions in the GUI. The test designer determines the preconditions and effects of the hierarchical operators, which are then input into a planning system. With the knowledge of the GUI and the way in which the user will interact with the GUI, the test designer creates sets of initial and goal states. Given these initial and final states of the GUI, a hierarchical planner produces plans, or a set of test cases, that enable the goal state to be reached. Our technique has the additional benefit of putting verification commands into the test cases automatically. We implemented our technique by developing the GUI analyzer and extending a planner. We generated test cases for Microsoft's Word-Pad to demonstrate the viability and practicality of the approach. %B Software Engineering, 1999. Proceedings of the 1999 International Conference on %P 257 - 266 %8 1999/05// %G eng %0 Report %D 1999 %T Using Interactive Visualizations of WWW Log Data to Characterize Access Patterns and Inform Site Design %A Hochheiser,Harry %A Shneiderman, Ben %K Information Visualization %K log file analysis %K World Wide Web %X HTTP server log files provide Web site operators with substantial detail regarding the visitors to their sites. Interest ininterpreting this data has spawned an active market for software packages that summarize and analyze this data, providing histograms, pie graphs, and other charts summarizing usage patterns. While useful, these summaries obscureuseful information and restrict users to passive interpretation of static displays. Interactive visualizations can be used to provide users with greater abilities to interpret and explore web log data. Bycombining two-dimensional displays of thousands of individual access requests, color and size coding for additional attributes, and facilities for zooming and filtering, these visualizations provide capabilities for examining data that exceed those of traditional web log analysis tools.

We introduce a series of interactive visualizations that can be used to explore server data across various dimensions. Sample visualizations of server data from two web sites are presented. Coordinated, snap-together visualizations (STVs) of log data are introduced as a means of gaining additional expressive power. Possible uses of these visualizations are discussed, and difficulties of data collection, presentation, and interpretation are explored. %B Institute for Systems Research Technical Reports %8 1999/// %G eng %U http://drum.lib.umd.edu/handle/1903/6034 %0 Journal Article %J Proc. Workshop Web Usage Analysis and User Profiling (WEBKDD’99) %D 1999 %T Using probabilistic relational models for collaborative filtering %A Getoor, Lise %A Sahami,M. %X Recent projects in collaborative filtering and information filtering address the task of inferring user prefer-ence relationships for products or information. The data on which these inferences are based typically con- sists of pairs of people and items. The items may be information sources (such as web pages or newspaper articles) or products (such as books, software, movies or CDs). We are interested in making recommen- dations or predictions. Traditional approaches to the problem derive from classical algorithms in statistical pattern recognition and machine learning. The majority of these approaches assume a ”flat” data repre- sentation for each object, and focus on a single dyadic relationship between the objects. In this paper, we examine a richer model that allows us to reason about many different relations at the same time. We build on the recent work on probabilistic relational models (PRMs), and describe how PRMs can be applied to the task of collaborative filtering. PRMs allow us to represent uncertainty about the existence of relationships in the model and allow the properties of an object to depend probabilistically both on other properties of that object and on properties of related objects. %B Proc. Workshop Web Usage Analysis and User Profiling (WEBKDD’99) %8 1999/// %G eng %0 Journal Article %J Readings in information visualization: using vision to think %D 1999 %T Using vision to think %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization: using vision to think %P 579 - 581 %8 1999/// %G eng %0 Conference Paper %B Readings in information visualization %D 1999 %T Visually enhanced objects %A Card,S.K. %A Mackinlay,J.D. %A Shneiderman, Ben %B Readings in information visualization %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 561 - 561 %8 1999/// %@ 1-55860-533-9 %G eng %0 Journal Article %J University of Maryland Technical Report, CS-TR %D 1999 %T Zoom-Only Vs Overview-Detail Pair: A Study in Browsing Techniques as Applied to Patient Histories %A Ghosh,P. %A Shneiderman, Ben %B University of Maryland Technical Report, CS-TR %V 4028 %P 12 - 12 %8 1999/// %G eng %N 99 %0 Conference Paper %D 1998 %T 3D spatial layouts using a-teams %A Sachdev,S. %A Paredis,C. J. J. %A Gupta,S.K. %A Talukdar,S. N. %X Spatial layout is the problem of arranging a set of componentsin an enclosure such that a set of objectives and constraints is sat- isfied. The constraints may include non-interference of objects, accessibility requirements and connection cost limits. Spatial lay- out problems are found primarily in the domains of electrical en- gineering and mechanical engineering in the design of integrated circuits and mechanical or electromechanical artifacts. Traditional approaches include ad-hoc (or specialized) heuristics, Genetic Al- gorithms and Simulated Annealing. The A-Teams approach pro- vides a way of synergistically combining these approaches in a modular agent based fashion. A-Teams are also open to the addi- tion of new agents. Modifications in the task requirements trans- late to modifications in the agent mix. In this paper we describe how modular A-Team based optimization can be used to solve 3 dimensional spatial layout problems. %8 1998/// %G eng %U http://www-2.cs.cmu.edu/afs/cs.cmu.edu/user/cjp/www/pubs/DETC98.pdf %0 Journal Article %J Journal of the WAS, Special Issue on Communications, 85 (2) %D 1998 %T Access Policies for Distributed Video-on-Demand Systems %A Simon,R. %A Mundur, Padma %A Sood,A. %B Journal of the WAS, Special Issue on Communications, 85 (2) %8 1998/// %G eng %0 Journal Article %J Linear Algebra and its Applications %D 1998 %T On the adjugate matrix %A Stewart, G.W. %X The adjugate AA of a matrix A is the transpose of the matrix of the co-factors of the elements of A. The computation of the adjugate from its definition involves the computation of n2 determinants of order (n−1)—a prohibitively expensive O(n4) process. On the other hand, the computation from the formula AA = det (A)A−1 breaks down when A is singular and is potentially unstable when A is ill-conditioned with respect to inversion. In this paper we first show that the adjugate can be perfectly conditioned, even when A is ill-conditioned. We then show that if due care is taken the adjugate can be accurately computed from the inverse, even when the latter has been inaccurately computed. In Appendix A we give a formal derivation of an observation of Wilkinson on the accuracy of computed inverses. %B Linear Algebra and its Applications %V 283 %P 151 - 164 %8 1998/11/01/ %@ 0024-3795 %G eng %U http://www.sciencedirect.com/science/article/pii/S0024379598100988 %N 1–3 %R 10.1016/S0024-3795(98)10098-8 %0 Conference Paper %B Proceedings of the 12th international conference on Supercomputing %D 1998 %T Application level scheduling of gene sequence comparison on metacomputers %A Spring, Neil %A Wolski,R. %B Proceedings of the 12th international conference on Supercomputing %P 141 - 148 %8 1998/// %G eng %0 Journal Article %J PROCEEDINGS-SPIE THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING %D 1998 %T Approximating large convolutions in digital images %A Kanungo,T. %A Mount, Dave %A Netanyahu,N. S %A Piatko,C. %A Silverman,R. %A Wu,A. Y %X Computing discrete two-dimensional convolutions is an important problem in image processing. In mathematicalmorphology, an important variant is that of computing binary convolutions, where the kernel of the convolution is a 0{1 valued function. This operation can be quite costly, especially when large kernels are involved. In this paper, we present an algorithm for computing convolutions of this form, where the kernel of the binary convolution is derived from a convex polygon. Because the kernel is a geometric object, we allow the algorithm some exibility in how it elects to digitize the convex kernel at each placement, as long as the digitization satis es certain reasonable requirements. We say that such a convolution is valid. Given this exibility we show that it is possible to compute binary convolutions more e ciently than would normally be possible for large kernels. Our main result is an algorithm, which given an m n image and a k-sided convex polygonal kernel, computes a valid convolution in time O(kmn) time. Unlike standard algorithms for computing correlations and convolutions, the running time is independent of the area or perimeter of K, and our techniques do not rely on computing fast Fourier transforms. Our algorithm is based on a novel use of Bresenham's line-drawing algorithm and pre x-sums to update the convolution e ciently as the kernel is moved from one position to another across the image. %B PROCEEDINGS-SPIE THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING %P 216 - 227 %8 1998/// %G eng %0 Conference Paper %B Seventeenth IEEE Symposium on Reliable Distributed Systems %D 1998 %T AQuA: an adaptive architecture that provides dependable distributed objects %A Cukier, Michel %A Ren,J. %A Sabnis,C. %A Henke,D. %A Pistole,J. %A Sanders,W. H. %A Bakken,D. E. %A Berman,M.E. %A Karr,D. A. %A Schantz,R.E. %K adaptive architecture %K AQuA %K availability requests %K client-server systems %K commercial off-the-shelf hardware %K CORBA %K dependability manager %K dependability requirements %K dependable distributed objects %K distributed object management %K Ensemble protocol stack %K Fault tolerance %K group communication services %K middleware software %K Object Request Brokers %K process-level communication %K proteus %K Quality Objects %K replication %K software fault tolerance %K Software quality %X Dependable distributed systems are difficult to build. This is particularly true if they have dependability requirements that change during the execution of an application, and are built with commercial off-the-shelf hardware. In that case, fault tolerance must be achieved using middleware software, and mechanisms must be provided to communicate the dependability requirements of a distributed application to the system and to adapt the system's configuration to try to achieve the desired dependability. The AQuA architecture allows distributed applications to request a desired level of availability using the Quality Objects (QuO) framework and includes a dependability manager that attempts to meet requested availability levels by configuring the system in response to outside requests and changes in system resources due to faults. The AQuA architecture uses the QuO runtime to process and invoke availability requests, the Proteus dependability manager to configure the system in response to faults and availability requests, and the Ensemble protocol stack to provide group communication services. Furthermore, a CORBA interface is provided to application objects using the AQuA gateway. The gateway provides a mechanism to translate between process-level communication, as supported by Ensemble, and IIOP messages, understood by Object Request Brokers. Both active and passive replication are supported, and the replication type to use is chosen based on the performance and dependability requirements of particular distributed applications %B Seventeenth IEEE Symposium on Reliable Distributed Systems %P 245 - 253 %8 1998/10// %G eng %R 10.1109/RELDIS.1998.740506 %0 Report %D 1998 %T On the Area of Overlap of Translated Polygons %A Mount, Dave %A Silverman,Ruth %A Wu,Angela Y. %K Technical Report %X (Also cross-referenced as CAR-TR-699) Given two simple polygons P and Q in the plane and a translation vector t E R2, the area-oJ-overlap function of P and Q is the function Ar(t) = Area(P n (t + Q)), where t + Q denotes Q translated by t. This function has a number of applications in areas such as motion planning and object recognition. We present a number of mathematical results regarding this function. We also provide efficient algorithms for computing a representation of this function, and for tracing contour curves of constant area of o verlap. %I Department of Computer Science, University of Maryland, College Park %V CS-TR-3201 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu//handle/1903/403 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1998 %T Assessing users' subjective satisfaction with the Information System for Youth Services (ISYS) %A Slaughter,Laura %A Norman,Kent L %A Shneiderman, Ben %K Technical Report %X In this investigation, the Questionnaire for User Interaction Satisfaction(QUIS 5.5), a tool for assessing users' subjective satisfaction with specific aspects of the human/computer interface was used to assess the strengths and weaknesses of the Info rmation System for Youth Services (ISYS). ISYS is used by over 600 employees of the Maryland State Department of Juvenile Services (DJS) as a tracking device for juvenile offenders. Ratings and comments were collected from 254 DJS employees who use ISYS. The overall mean rating across all questions was 5.1 on a one to nine scale. The ten highest and lowest rated questions were identified. The QUIS allowed us to isolate subgroups which were compared with mean ratings from four measures of specific interfac e factors. The comments obtained from users provided suggestions, complaints and endorsements of the system. Also cross-referenced as CAR-TR-768 %B Institute for Systems Research Technical Reports %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/431 %0 Conference Paper %B Proceedings of the Symposium on Network and Distributed Systems Security (NDSS’98) %D 1998 %T Automated recovery in a secure bootstrap process %A Arbaugh, William A. %A Keromytis,A. D %A Farber,D. J %A Smith,J. M %B Proceedings of the Symposium on Network and Distributed Systems Security (NDSS’98) %P 155 - 167 %8 1998/// %G eng %0 Book %D 1998 %T Basic decompositions %A Stewart, G.W. %I Soc. for Industrial and Applied Mathematics %V 1 %8 1998/// %G eng %0 Report %D 1998 %T Bilingual Lexicon Construction Using Large Corpora %A Shen,Wade %A Dorr, Bonnie J %K Technical Report %X This paper introduces a method for learning bilingual term and sentencelevel alignments for the purpose of building lexicons. Combining statistical techniques with linguistic knowledge, a general algorithm is developed for learning term and sentence alignments from large bilingual corpora with high accuracy. This is achieved through the use of filtered linguistic feedback between term and sentence alignment processes. An implementation of this algorithm, TAG-ALIGN, is evaluated against approaches similar to [Brown et al. 1993] that apply Bayesian techniques for term alignment, and [Gale and Church 1991] a dynamic programming method for aligning sentences. The ultimate goal is to produce large bilingual lexicons with a high degree of accuracy from potentially noisy corpora. (Also cross-referenced as UMIACS-TR-97-50) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/832 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Browsing Unix Directories With Dynamic Queries: An Evaluation of Three Information Display Techniques %A Liao,Holmes %A Osada,Masakazu %A Shneiderman, Ben %K Technical Report %X We designed, implemented, and evaluated an innovative concept fordynamic queries which involves the direct manipulation of small databases. Our domain was directories in a Unix file system. Dynamic queries allow users to formulate queries and explore the databases with graphical widgets, such as sliders and buttons, without requiring them to have any knowledge about the underlying structure of the database query languages, or command language syntax. Three interfaces for presenting directories were developed and tested with eighteen subjects in a within-subject design. The results of the formative evaluation yielded some useful guidelines for software designers. (Also cross-referenced as CAR-TR-605) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/379 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Buttons vs. menus: An exploratory study of pull-down menu selection as compared to button bars %A Ellis,Jason %A Tran,Chi %A Ryoo,Jake %A Shneiderman, Ben %K Technical Report %X Button bars are a relatively new interaction method intended to speedup application use as compared to pull-down menus. This exploratory study compares three command selection methods: pull-down menus, button bars, and user choice of pull-down menus or button bars. Effectiveness was measured in two ways: speed of selection and error rate. 15 participants performed 15 word processor related tasks. Results show that in frequently used functions, such as character attribute selection (bold, italic, u nderline, etc.), button bars are faster. There were no statistically significant differences in error rates between the three interaction methods. (Also cross-referenced as CAR-TR-764) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/428 %0 Conference Paper %D 1998 %T Capturing articulation in assemblies from component geometry %A Sinha,R. %A Paredis,C. J. J. %A Gupta,S.K. %A Khosla,P. K. %X This paper presents a method to extract instantaneous artic-ulation from part geometry, based on surface mating constraints as well as constraints imposed by other incidental contacts. Many assemblies contain joints, each of which have degrees of freedom associated with them. These joints allow the relative positions of parts in the mechanism to change as the joints are articulated. Being able to represent these joints and their behav- ior is important from the designers perspective because it enables him or her to verify whether kinematic requirements have been met.Therefore, it is useful to be able to obtain such joint informa- tion directly from part geometry and contact physics. The method presented here handles all lower pairs of kine- matic joints. Surface mating contacts are classified into one of three types: planar, spherical and cylindrical. The contacts are represented by algebraic inequalities describing the translational and angular velocities at the contact. Non-penetration conditions are written for a finite set of points on the boundary of each contact face, and it is shown that the finite set of conditions is representative of the entire boundary and the region enclosed by the boundary. Simultaneous satisfaction of the non-penetration conditions at all the contact surfaces between a pair of bodies is represented by a 6-dimensional simplex, which can be solved using linear programming. %8 1998/// %G eng %U http://www-cgi.cs.cmu.edu/afs/cs.cmu.edu/user/cjp/www/pubs/DAC98.pdf %0 Journal Article %J International Journal of Human-Computer Interaction %D 1998 %T Codex, Memex, Genex: The Pursuit of Transformational Technologies %A Shneiderman, Ben %X Handwritten codexes or printed books transformed society by allowing users to preserve and transmit information. Today, leather-bound volumes and illuminated manuscripts are giving way to animated image maps and hot links. Vannevarr Bush's memex has inspired the World Wide Web, which provides users with vast information resources and convenient communications. In looking to the future, we might again transform society by building genexes?generators of excellence. Such inspirational environment would empower personal and collaborative creativity by enabling users to: Collect information from an existing domain of knowledge. Create innovations using advanced tools. Consult with peers or mentors in the field. Disseminate the results widely. This article describes how a framework for an integrated set of software tools might support this 4-phase model of creative in science, medicine, the arts, and beyond. Current initiatives are positive and encouraging, but they do not work in an integrated fashion often miss vital components, and are frequently poorly designed. A well-conceived and clearly stated framework could guide design efforts, coordinate planning, and speed development.Handwritten codexes or printed books transformed society by allowing users to preserve and transmit information. Today, leather-bound volumes and illuminated manuscripts are giving way to animated image maps and hot links. Vannevarr Bush's memex has inspired the World Wide Web, which provides users with vast information resources and convenient communications. In looking to the future, we might again transform society by building genexes?generators of excellence. Such inspirational environment would empower personal and collaborative creativity by enabling users to: Collect information from an existing domain of knowledge. Create innovations using advanced tools. Consult with peers or mentors in the field. Disseminate the results widely. This article describes how a framework for an integrated set of software tools might support this 4-phase model of creative in science, medicine, the arts, and beyond. Current initiatives are positive and encouraging, but they do not work in an integrated fashion often miss vital components, and are frequently poorly designed. A well-conceived and clearly stated framework could guide design efforts, coordinate planning, and speed development. %B International Journal of Human-Computer Interaction %V 10 %P 87 - 106 %8 1998/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1207/s15327590ijhc1002_1 %N 2 %R 10.1207/s15327590ijhc1002_1 %0 Journal Article %J Neural Computation %D 1998 %T Computational studies of lateralization of phoneme sequence generation %A Reggia, James A. %A Goodall,S. %A Shkuro,Y. %B Neural Computation %V 10 %P 1277 - 1297 %8 1998/// %G eng %N 5 %0 Journal Article %J Electronic Transactions on Numerical Analysis %D 1998 %T On the convergence of a new Rayleigh quotient method with applications to large eigenproblems %A Oleary,D. P. %A Stewart, G.W. %X In this paper we propose a variant of the Rayleigh quotient method to compute an eigenvalue and cor-responding eigenvectors of a matrix. It is based on the observation that eigenvectors of a matrix with eigenvalue zero are also singular vectors corresponding to zero singular values. Instead of computing eigenvector approximations by the inverse power method, we take them to be the singular vectors corresponding to the smallest singular value of the shifted matrix. If these singular vectors are computed exactly the method is quadratically convergent. However, ex- act singular vectors are not required for convergence, and the resulting method combined with Golub–Kahan–Krylov bidiagonalization looks promising for enhancement/refinement methods for large eigenvalue problems. %B Electronic Transactions on Numerical Analysis %V 7 %P 182 - 189 %8 1998/// %G eng %0 Conference Paper %B Proceedings of the 1998 workshop on New paradigms in information visualization and manipulation %D 1998 %T Data object and label placement for information abundant visualizations %A Li,Jia %A Plaisant, Catherine %A Shneiderman, Ben %K control panel %K data object placement %K Information Visualization %K label placement %K metrics %K timelines %K visual feedback %B Proceedings of the 1998 workshop on New paradigms in information visualization and manipulation %S NPIV '98 %I ACM %C New York, NY, USA %P 41 - 48 %8 1998/// %@ 1-58113-179-8 %G eng %U http://doi.acm.org/10.1145/324332.324341 %R 10.1145/324332.324341 %0 Journal Article %J Parallel Computing %D 1998 %T The design and evaluation of a high-performance earth science database %A Shock,Carter T. %A Chang,Chialin %A Moon,Bongki %A Acharya,Anurag %A Davis, Larry S. %A Saltz,Joel %A Sussman, Alan %K Communication %K High-performance I/O %K load balancing %K scalability %K Scientific databases %X Earth scientists have encountered two major obstacles in their attempts to use remotely sensed imagery to analyze the earth's land cover dynamics. First, the volume of data involved is very large and second, significant preprocessing is needed before the data can be used. This is particularly so for studies that analyze global trends using data sets that cover multiple years. In this paper, we present the design of an earth science database as well as our early experiences with it. The primary design goal of this database is to facilitate efficient access to and preprocessing of large volumes of satellite data. Our initial design assumed that the main bottleneck in the system would be retrieving data from the disks. However, experimental results show that precise identification of all the data values corresponding to a query can take a significant amount of time. The problem is even more pronounced in designing the system to attempt to minimize time spent performing I/O. We therefore discuss a major redesign of the system that includes a reworking of the indexing scheme and a reorganization of the data on disks. Experimental results show that the redesigned system performs significantly better than the original system, providing interactive response times for local queries. %B Parallel Computing %V 24 %P 65 - 89 %8 1998/01// %@ 0167-8191 %G eng %U http://www.sciencedirect.com/science/article/pii/S0167819197001178 %N 1 %R 10.1016/S0167-8191(97)00117-8 %0 Report %D 1998 %T Designing to facilitate browsing: A look back at the Hyperties workstation browser %A Shneiderman, Ben %A Plaisant, Catherine %A Botafogo,Rodrigo %A Hopkins,Don %A Weiland,William %K Technical Report %X Since browsing hypertext can present a formidable cognitivechallenge, user interface design plays a major role in determining acceptability. In the Unix workstation version of Hyperties, a research-oriented prototype, we focussed on design features that facilitate browsing. We first give a general overview of Hyperties and its markup language. Customizable documents can be generated by the conditional text feature that enables dynamic and selective display of text and graphics. In addition we present: - an innovative solution to link identification: pop-out graphical buttons of arbitrary shape. - application of pie menus to permit low cognitive load actions that reduce the distraction of common actions, such as page turning or window selection. - multiple window selection strategies that reduce clutter and housekeeping effort. We preferred piles-of-tiles, in which standard-sized windows were arranged in a consistent pattern on the display and actions could be done rapidly, allowing users to concentrate on the contents. (Also cross-referenced as CAR-TR-494) %B Technical Reports of the Computer Science Department, University of Maryland %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/362 %0 Journal Article %J Proceedings of Global Internet (GlobeCom)’98 %D 1998 %T DHCP++: Applying an efficient implementation method for fail-stop cryptographic protocols %A Arbaugh, William A. %A Keromytis,A. D %A Smith,J. M %B Proceedings of Global Internet (GlobeCom)’98 %P 59 - 65 %8 1998/// %G eng %0 Conference Paper %B Proceedings of the AMIA Symposium %D 1998 %T Digital dynamic telepathology–the Virtual Microscope. %A Afework,A. %A Beynon,M. D %A Bustamante,F. %A Cho,S. %A Demarzo,A. %A Ferreira,R. %A Miller,R. %A Silberman,M. %A Saltz, J. %A Sussman, Alan %A others %B Proceedings of the AMIA Symposium %P 912 - 912 %8 1998/// %G eng %0 Report %D 1998 %T Direction of Arrival and the Rank-Revealing URV Decomposition %A Boman,E. C %A Griffen,M. F %A Stewart, G.W. %K Technical Report %X In many practical direction-of-arrival (DOA) problems the numberof sources and their directions from an antenna array do not remain stationary. Hence a practical DOA algorithm must be able to track changes with a minimal number of snapshots. In this paper we describe DOA algorithms, based on a new decomposition, that are not expensive to compute or difficult to update. The algorithms are compared with algorithms based on the singular value decomposition (SVD). (Also cross-referenced as UMIACS-TR-91-166) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-91-166 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/561 %0 Conference Paper %B CHI 98 conference summary on Human factors in computing systems %D 1998 %T Distance learning: is it the end of education as most of us know it? %A Laurillard,Diana %A Preece,Jenny %A Shneiderman, Ben %A Neal,Lisa %A Wærn,Yvonne %K distance learning %K education %K Internet %K Web %B CHI 98 conference summary on Human factors in computing systems %S CHI '98 %I ACM %C New York, NY, USA %P 86 - 87 %8 1998/// %@ 1-58113-028-7 %G eng %U http://doi.acm.org/10.1145/286498.286542 %R 10.1145/286498.286542 %0 Journal Article %J Educom Review %D 1998 %T Educational Journeys on the Web Frontier %A Shneiderman, Ben %B Educom Review %V 33 %P 10 - 14 %8 1998/// %G eng %N 6 %0 Journal Article %J Software: Practice and Experience %D 1998 %T Elastic windows: design, implementation, and evaluation of multi‐window operations %A Kandogan,Eser %A Shneiderman, Ben %K hierarchical windows %K multi‐window operations %K personal role management %K window management %K windowing system evaluation %K World Wide Web %X Most windowing systems follow the independent overlapping windows approach, which emerged as an answer to the needs of early computer users. Due to advances in computers, display technology, and increased information needs, modern users demand more functionality from window management systems. We propose Elastic Windows with improved spatial layout and rapid multi-window operations as an alternative to current window management strategies. In this approach, multi-window operations are achieved by issuing operations on window groups hierarchically organized in a space-filling tiled layout similar to TreeMaps.1 Sophisticated multi-window operations have been developed to handle fast task-switching and to structure the work environment of users to their rapidly changing needs. We claim that these multi-window operations and the tiled spatial layout dynamics decrease the cognitive load on users by decreasing the number of window operations. This paper describes the Elastic Windows interface in detail and then presents a user study conducted to compare the performance of 12 users with Elastic Windows and traditional Independent Overlapping Windows. User performance was measured in terms of task environment setup, switching, and task execution for 2, 6, and 12 window situations. Elastic Windows users had statistically significantly faster performance for all tasks in 6 and 12 window situations. These results suggest promising possibilities for multiple window operations and hierarchical nesting, which can be applied to the next generation of tiled as well as overlapped window managers. © 1998 John Wiley & Sons, Ltd. %B Software: Practice and Experience %V 28 %P 225 - 248 %8 1998/03/01/ %@ 1097-024X %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/(SICI)1097-024X(199803)28:3%3C225::AID-SPE151%3E3.0.CO;2-D/abstract %N 3 %R 10.1002/(SICI)1097-024X(199803)28:3<225::AID-SPE151>3.0.CO;2-D %0 Journal Article %J Educational Technology Research and Development %D 1998 %T Emergent patterns of teaching/learning in electronic classrooms %A Shneiderman, Ben %A Borkowski,Ellen %A Alavi,Maryam %A Norman,Kent %X Novel patterns of teaching/learning have emerged from faculty and students who use our three teaching/learning theaters at the University of Maryland, College Park. These fully-equipped electronic classrooms have been used by 74 faculty in 264 semester-long courses since the fall of 1991 with largely enthusiastic reception by both faculty and students. The designers of the teaching/learning theaters sought to provide a technologically rich environment and a support staff so that faculty could concentrate on changing the traditional lecture from its unidirectional information flow to a more collaborative activity. As faculty have evolved their personal styles in using the electronic classrooms, novel patterns of teaching/learning have emerged. In addition to enhanced lectures, we identified three common patterns: (a) active individual learning, (b) small-group collaborative learning, and (c) entire-class collaborative learning. %B Educational Technology Research and Development %V 46 %P 23 - 42 %8 1998/// %@ 1042-1629 %G eng %U http://dx.doi.org/10.1007/BF02299671 %N 4 %0 Journal Article %J Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting %D 1998 %T An Empirical Study of Perspective-Based Usability Inspection %A Zhang,Zhijun %A Basili, Victor R. %A Shneiderman, Ben %X Inspection is a fundamental means of achieving software usability. Past research showed that during usability inspection the success rate (percentage of problems detected) of each individual inspector was rather low. We developed perspective-based usability inspection, which divides the large variety of usability issues along different perspectives and focuses each inspection session on one perspective. We conducted a controlled experiment to study its effectiveness, using a post-test only control group experimental design, with 24 professionals as subjects. The control group used heuristic evaluation, which is the most popular technique for usability inspection. The experimental results are that 1) for usability problems covered by each perspective, the inspectors using that perspective had higher success rate than others; 2) for all usability problems, perspective inspectors had higher average success rate than heuristic inspectors; 3) for all usability problems, the union of three perspective inspectors (with one from each perspective) had higher average success rate than the union of three heuristic inspectors. %B Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting %V 42 %P 1346 - 1350 %8 1998/10/01/ %@ 1071-1813, %G eng %U http://pro.sagepub.com/content/42/19/1346 %N 19 %R 10.1177/154193129804201904 %0 Journal Article %J Educational TechnologyEducational Technology %D 1998 %T Engagement Theory: A Framework for Technology-Based Teaching and Learning. %A Kearsley,Greg %A Shneiderman, Ben %K Computer Uses in Education %K Cooperation %K Educational Technology %K Educational Theories %K Instructional Effectiveness %K interaction %K Learning Activities %K Research Needs %K Student Motivation %K Student Projects %K Teaching Methods %X Discusses engagement theory (students meaningfully engaged in learning activities through interaction with others and worthwhile tasks) and the three components, collaboration, project orientation, and authentic focus, and outlines research questions to establish its efficacy. (PEN) %B Educational TechnologyEducational Technology %V 38 %P 20 - 23 %8 1998/// %@ ISSN-0013-1962 %G eng %U http://www.eric.ed.gov/ERICWebPortal/detail?accno=EJ573955 %N 5 %0 Journal Article %J Journal of the ACM (JACM) %D 1998 %T Explicit OR-dispersers with polylogarithmic degree %A Saks,Michael %A Srinivasan, Aravind %A Zhou,Shiyu %K derandomization %K expander graphs %K explicit constructions %K hardness of approximation %K hashing lemmas %K imperfect sources of randomness %K measures of information %K pseudo-random generators %K randomized computation %K time-space tradeoffs %X An (N, M, T)-OR-disperser is a bipartite multigraph G=(V, W, E) with |V| = N, and |W| = M, having the following expansion property: any subset of V having at least T vertices has a neighbor set of size at least M/2. For any pair of constants &xgr;, &lgr;, 1 ≥ &xgr; > &lgr; ≥ 0, any sufficiently large N, and for any T ≥ 2(logN) M ≤ 2(log N)&lgr;, we give an explicit elementary construction of an (N, M, T)-OR-disperser such that the out-degree of any vertex in V is at most polylogarithmic in N. Using this with known applications of OR-dispersers yields several results. First, our construction implies that the complexity class Strong-RP defined by Sipser, equals RP. Second, for any fixed &eegr; > 0, we give the first polynomial-time simulation of RP algorithms using the output of any “&eegr;-minimally random” source. For any integral R > 0, such a source accepts a single request for an R-bit string and generates the string according to a distribution that assigns probability at most 2−R&eegr; to any string. It is minimally random in the sense that any weaker source is insufficient to do a black-box polynomial-time simulation of RP algorithms. %B Journal of the ACM (JACM) %V 45 %P 123 - 154 %8 1998/01// %@ 0004-5411 %G eng %U http://doi.acm.org/10.1145/273865.273915 %N 1 %R 10.1145/273865.273915 %0 Journal Article %J Information Visualization %D 1998 %T The Eyes Have It: User Interfaces for Information Visualization %A Shneiderman, Ben %B Information Visualization %V 1452 %P 1519 - 1519 %8 1998/// %G eng %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Facilitating Network Data Exploration with Query Previews: A Study of User Performance and Preference %A Tanin,Egemen %A Lotem,Amnon %A Haddadin,Ihab %A Shneiderman, Ben %A Plaisant, Catherine %A Slaughter,Laura %K Technical Report %X Current network data exploration systems which use command languages (e.g.SQL) or form fill-in interfaces fail to give users an indication of the distribution of data items. This leads many users to waste time posing queries which have zero-hit or mega-hit result sets. Query previewing is a novel visual approach for browsing huge networked information warehouses. Query previews supply data distribution information about the database that is being searched and give continuous feedback about the size of the result set for the query as it is being formed. Our within-subjects empirical comparison studied 12 subjects using a form fill-in interface with and without query previews. We found statistically significant differences showing that query previews sped up performance 1.6 to 2.1 times and led to higher subjective satisfaction. (Also cross-referenced as UMIACS-98-14) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/488 %0 Journal Article %J Journal of Combinatorial Optimization %D 1998 %T Facility Location with Dynamic Distance Functions %A Khuller, Samir %A Sussmann,Y. J %A Bhatia,Randeep %A Guha,Sudipto %X Facility location problems have always been studied with theassumption that the edge lengths in the network are static anddo not change over time. The underlying network could be used to model a city street networkfor emergency facility location/hospitals, or an electronic network for locating information centers. In any case, it is clear that due to trafficcongestion the traversal time on links changes with time. Very often, we have estimates as to how the edge lengths change over time, and our objective is to choose a set of locations (vertices) ascenters, such that at every time instant each vertex has a center close to it (clearly, the center close to a vertex may change over time). We also provide approximation algorithms as well as hardness results forthe K-center problem under this model. This is the first comprehensive study regarding approximation algorithmsfor facility location for good time-invariant solutions. %B Journal of Combinatorial Optimization %V 2 %P 199 - 217 %8 1998/// %G eng %N 3 %R 10.1023/A:1009796525600 %0 Report %D 1998 %T Four algorithms for the efficient computation of truncated pivoted QR approximation to a sparse matrix %A Stewart, G.W. %X In this paper we propose four algorithms to compute truncated pivoted QR approximations to a sparse matrix. Three are based on the Gram–Schmidt algorithm and the other on Householder triangularization. All four algorithms leave the original matrix unchanged, and the only additional storage requirements are arrays to contain the factorization itself. Thus, the algorithms are particularly suited to determining low-rank approximations to a sparse matrix. %I University of Maryland, College Park %V TR-98-12 %8 1998/// %G eng %0 Report %D 1998 %T Gaussian Elimination, Perturbation Theory and Markov Chains %A Stewart, G.W. %K Technical Report %X The purpose of this paper is to describe the special problems thatemerge when Gaussian elimination is used to determinin the steady-state vector of a Markov chain. (Also cross-referenced as UMIACS-TR-92-23) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-92-23 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/564 %0 Journal Article %J SIAM journal on matrix analysis and applications %D 1998 %T On hyperbolic triangularization: Stability and pivoting %A Stewart,M. %A Stewart, G.W. %X This paper treats the problem of triangularizing a matrix by hyperbolic Householdertransformations. The stability of this method, which finds application in block updating and fast algorithms for Toeplitz-like matrices, has been analyzed only in special cases. Here we give a gen- eral analysis which shows that two distinct implementations of the individual transformations are relationally stable. The analysis also shows that pivoting is required for the entire triangularization algorithm to be stable. %B SIAM journal on matrix analysis and applications %V 19 %P 847 - 860 %8 1998/// %G eng %N 4 %0 Report %D 1998 %T Implementing an Algorithm for Solving Block Hessenberg Systems %A Stewart, G.W. %K Technical Report %X This paper describes the implementation of a recursive descent methodfor solving block Hessenberg systems. Although the algorithm is conceptually simple, its implementation in C (a natural choice of language given the recursive nature of the algorithm and its data) is nontrivial. Particularly important is the balance between ease of use, computational efficiency, and flexibility. (Also cross-referenced as UMIACS-TR-94-70) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-94-70 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/642 %0 Conference Paper %B 39th Annual Symposium on Foundations of Computer Science, 1998. Proceedings %D 1998 %T Improved bounds and algorithms for hypergraph two-coloring %A Radhakrishnan,J. %A Srinivasan, Aravind %K algorithms %K Application software %K Approximation algorithms %K bounds %K computational geometry %K Computer science %K Contracts %K Erbium %K graph colouring %K History %K hypergraph two-coloring %K Lab-on-a-chip %K MATHEMATICS %K n-uniform hypergraph %K Parallel algorithms %K Polynomials %K probability %X We show that for all large n, every n-uniform hypergraph with at most 0.7√(n/lnn)×2n edges can be two-colored. We, in fact, present fast algorithms that output a proper two-coloring with high probability for such hypergraphs. We also derandomize and parallelize these algorithms, to derive NC1 versions of these results. This makes progress on a problem of Erdos (1963), improving the previous-best bound of n1/3-0(1)×2n due to Beck (1978). We further generalize this to a “local” version, improving on one of the first applications of the Lovasz Local Lemma %B 39th Annual Symposium on Foundations of Computer Science, 1998. Proceedings %I IEEE %P 684 - 693 %8 1998/11/08/11 %@ 0-8186-9172-7 %G eng %R 10.1109/SFCS.1998.743519 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Incorporating String Search in a Hypertext System:User Interface and Signature File Design Issues %A Faloutsos,Christos %A Lee,Raymond %A Plaisant, Catherine %A Shneiderman, Ben %K Technical Report %X Hypertext systems provide an appealing mechanism forinformally browsing databases by traversing selectable links. However, in many fact finding situations string search is an effective complement to browsing. This paper describes the application of the signature file method to achieve rapid and convenient string search in small personal computer hypertext environments. The method has been implemented in a prototype, as well as in a commercial product. Performance data for search times and storage space are presented from a commercial hypertext database. User interface issues are then discussed. Experience with the string search interface indicates that it was used sucessfully by novice users. (Also cross-referenced as CAR-TR-448) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/359 %0 Report %D 1998 %T Incremental Condition Calculation and Column Selection %A Stewart, G.W. %K Technical Report %X This paper describes a method for calculating the condition number ofa matrix in the Frobenius norm that can be used to select columns in the course of computing a QR decomposition. When the number of rows of the matrix is much greater than the number of columns, the additional overhead is negligible. Limited numerical experiments suggest that the method is quite good at finding gaps in the singular values of the matrix. Additional files are available via anonymous ftp at: thales.cs.umd.edu in the directory pub/reports (Also cross-referenced as UMIACS-TR-90-87) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-90-87 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/550 %0 Report %D 1998 %T On an Inexpensive Triangular Approximation to the Singular Value Decomposition %A Stewart, G.W. %K Technical Report %X In this paper we introduce a new decomposition called the pivotedQLP~decomposition. It is computed by applying pivoted orthogonal triangularization to the columns of the matrix $X$ in question to get an upper triangular factor $R$ and then applying the same procedure to the rows of $R$ to get a lower triangular matrix $L$. The diagonal elements of $R$ are called the R-values of $X$; those of $L$ are called the L-values. Numerical examples show that the L-values track the singular values of $X$ with considerable fidelity\,---\,far better than the R-values. At a gap in the L-values the decomposition provides orthonormal bases of analogues of row, column, and null spaces provided of $X$. The decomposition requires no more than twice the work required for a pivoted QR~decomposition. The computation of $R$ and $L$ can be interleaved, so that the computation can be the rows of $R$ to get a lower triangular matrix $L$. The diagonal elements of $R$ are called the R-values of $X$; those of $L$ are called the L-values. Numerical examples show that the L-values track the singular values of $X$ with considerable fidelity\,---\,far better than the R-values. At a gap in the L-values the decomposition provides orthonormal bases of analogues of row, column, and null spaces provided of $X$. The decomposition requires no more than twice the work required for a pivoted QR~decomposition. The computation of $R$ and $L$ can be interleaved, so that the computation can be terminated at any suitable point, which makes the decomposition especially suitable for low-rank determination problems. The interleaved algorithm also suggests a new, efficient 2-norm estimator. (Also cross-referenced as UMIACS-TR-97-75) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-97-75 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/920 %0 Conference Paper %B Foundations of software technology and theoretical computer science: 18th Conference, Chennai, India, December 17-19, 1998: proceedings %D 1998 %T Infinite Probabilistic and Nonprobabilistic Testing⋆ %A Kumar,K. N %A Cleaveland, Rance %A Smolka,S. A %B Foundations of software technology and theoretical computer science: 18th Conference, Chennai, India, December 17-19, 1998: proceedings %P 209 - 209 %8 1998/// %G eng %0 Report %D 1998 %T On Infinitely Many Algorithms for Solving Equations %A Schroeder,E. %A Stewart, G.W. %K Technical Report %X Translated by G. W. StewartThis report contains a translation of ``Ueber unendlich viele Algorithmen zur Aufl\"osung der Gleichungen,'' a paper by E. Schr\"oder which appeared in {\it Mathematische Annalen\/} in 1870. (Also cross-referenced as UMIACS-TR-92-121) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-92-121 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/577 %0 Journal Article %J Information Processing & Management %D 1998 %T An information architecture to support the visualization of personal histories %A Plaisant, Catherine %A Shneiderman, Ben %A Mushlin,Rich %K Graphical user interfaces %K Information Visualization %K LifeLines %K medical patient record %K personal histories %K temporal data %K timelines %X This paper proposes an information architecture for personal history data and describes how the data model can be extended to a runtime model for a compact visualization using graphical timelines. Our information architecture was developed for juvenile justice and medical patient records, but is usable in other application domains such as personal resumes, financial histories, or customer support. Our model groups personal history events into aggregates that are contained in facets (e.g., doctor visits, hospitalizations, or lab tests). Crosslinks enable representation of arbitrary relationships across events and aggregates. Data attributes, such as severity, can be mapped by data administrators to visual attributes such as color and line thickness. End-users have powerful controls over the display contents, and they can modify the mapping to fit their tasks. %B Information Processing & Management %V 34 %P 581 - 597 %8 1998/09// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/S0306457398000247 %N 5 %R 10.1016/S0306-4573(98)00024-7 %0 Conference Paper %B CHI 98 conference summary on Human factors in computing systems %D 1998 %T Information visualization advanced interface and Web design %A Shneiderman, Ben %A Plaisant, Catherine %B CHI 98 conference summary on Human factors in computing systems %S CHI '98 %I ACM %C New York, NY, USA %P 145 - 146 %8 1998/// %@ 1-58113-028-7 %G eng %U http://doi.acm.org/10.1145/286498.286625 %R 10.1145/286498.286625 %0 Conference Paper %D 1998 %T An intelligent environment for simulating mechanical assembly operations %A Gupta,S.K. %A Paredis,C. J. J. %A Sinha,R. %A Wang,C. H. %A Brown,P. F. %X Rapid technical advances in many different areas of scientificcomputing provide the enabling technologies for creating a com- prehensive simulation and visualization environment for assembly design and planning. We have built an intelligent environment in which simple simulations can be composed together to create com- plex simulations for detecting potential assembly problems. Our goal in this project is to develop high fidelity assembly simulation and visualization tools that can detect assembly related problems without going through physical mock-ups. In addition, these tools can be used to create easy-to-visualize instructions for performing assembly and service operations. %P 13 - 16 %8 1998/// %G eng %U http://www.cs.cmu.edu/afs/cs.cmu.edu/Web/People/paredis/pubs/DFM98.pdf %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Interactive Smooth Zoomming in a Starfield Information Visualization %A Jog,Ninad %A Shneiderman, Ben %K Technical Report %X This paper discusses the design and implementation of interactivesmooth zooming of a starfield display. A starfield display is a two dimensional scatterplot of a multidimensional database where every item from the database is represented as a small colored glyph whose position is determined by its ranking along ordinal attributes of the items laid out on the axes. One way of navigating this visual information is by using a zooming tool to incrementally zoom in on the items by varying the attribute range on either axis independently - such zooming causes the glyphs to move continuously and to grow or shrink. To get a feeling of flying through the data, users should be able to track the motion of each glyph without getting distracted by flicker or large jumps - conditions that necessitate high display refresh rates and closely spaced glyphs on successive frames. Although the use of high-speed hardware can achieve the required visual effect for small databases, the twin software bottlenecks of rapidly accessing display items and constructing a new display image fundamentally retard the refresh rate. Our work explores several methods to overcome these bottlenecks, presents a taxonomy of various zooming methods and introduces a new widget, the zoom bar, that facilitates zooming. (Also cross-referenced as CAR-TR-714) (Also cross-referenced as ISR-TR-94-46) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/411 %0 Book Section %B Representation and processing of spatial expressions %D 1998 %T Lexical Allocation in Interlingua-Based Machine Translation of Spatial Expressions %A Dorr, Bonnie J %A Voss,C.R. %A Sencan,M.U. %B Representation and processing of spatial expressions %P 125 - 125 %8 1998/// %G eng %0 Journal Article %J In Proceedings of the 1998 American Medical Informatic Association Annual Fall SymposiumProc AMIA Symp %D 1998 %T LifeLines: using visualization to enhance navigation and analysis of patient records %A Plaisant, Catherine %A Mushlin,R. %A Snyder,A. %A Li,J. %A Heller,D. %A Shneiderman, Ben %X LifeLines provide a general visualization environment for personal histories. We explore its use for clinical patient records. A Java user interface is described, which presents a one-screen overview of a computerized patient record using timelines. Problems, diagnoses, test results or medications can be represented as dots or horizontal lines. Zooming provides more details; line color and thickness illustrate relationships or significance. The visual display acts as a giant menu, giving direct access to the data. %B In Proceedings of the 1998 American Medical Informatic Association Annual Fall SymposiumProc AMIA Symp %P 76 - 80 %8 1998/// %@ 1531-605X %G eng %0 Journal Article %J Proceedings of the AMIA Symposium %D 1998 %T LifeLines: using visualization to enhance navigation and analysis of patient records. %A Plaisant, Catherine %A Mushlin,R. %A Snyder,A. %A Li,J. %A Heller,D. %A Shneiderman, Ben %X LifeLines provide a general visualization environment for personal histories. We explore its use for clinical patient records. A Java user interface is described, which presents a one-screen overview of a computerized patient record using timelines. Problems, diagnoses, test results or medications can be represented as dots or horizontal lines. Zooming provides more details; line color and thickness illustrate relationships or significance. The visual display acts as a giant menu, giving direct access to the data. %B Proceedings of the AMIA Symposium %P 76 - 80 %8 1998/// %@ 1531-605X %G eng %0 Book Section %B Automata, Languages and ProgrammingAutomata, Languages and Programming %D 1998 %T Low-bandwidth routing and electrical power networks %A Cook,Doug %A Faber,Vance %A Marathe,Madhav %A Srinivasan, Aravind %A Sussmann,Yoram %E Larsen,Kim %E Skyum,Sven %E Winskel,Glynn %X Given a graph G and a (multi-)set of pairs of vertices in it, the classical NP-hard maximum edge-disjoint-paths problem (MDP) is to connect as many of the given pairs as possible using pairwise edge-disjoint paths in G . We study a relative of this problem: we have a network with fixed link capacities that may have to service large demands when necessary. In particular, individual demands are allowed to exceed capacities, and thus flows for some request pairs necessarily have to be split into different flow-paths. This is the framework for computational problems arising from: (i) electrical power networks due to the proposed deregulation of the electric utility industry in the USA, and (ii) applications such as real-time Internet services (e.g., telephone, fax, video). We show that these problems come in a few variants, some efficiently solvable and many NP -hard; we also present approximation algorithms for many of the NP -hard variants presented. Some of our approximation algorithms benefit from certain improved tail estimates that we derive; the latter also yield improved approximations for a family of packing integer programs. %B Automata, Languages and ProgrammingAutomata, Languages and Programming %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1443 %P 604 - 615 %8 1998/// %@ 978-3-540-64781-2 %G eng %U http://dx.doi.org/10.1007/BFb0055088 %0 Book %D 1998 %T Matrix Algorithms: Basic decompositions %A Stewart, G.W. %K Mathematics / Algebra / General %K Mathematics / Algebra / Linear %K Mathematics / Applied %K Mathematics / Mathematical Analysis %K Mathematics / Matrices %K Mathematics / Numerical Analysis %K Matrices %X This thorough, concise, and superbly written volume is the first in a self-contained five-volume series devoted to matrix algorithms. It focuses on the computation of matrix decompositions - the factorization of matrices into products of similar ones. The first two chapters provide the required background from mathematics and computer science needed to work effectively in matrix computations. The remaining chapters are devoted to the computation and applications of the LU and QR decompositions. The series is aimed at the nonspecialist who needs more than black-box proficiency with matrix computations. A certain knowledge of elementary analysis and linear algebra is assumed, as well as a reasonable amount of programming experience. The guiding principle, that if something is worth explaining, it is worth explaining fully, has necessarily restricted the scope of the series, but the selection of topics should give the reader a sound basis for further study. %I SIAM %8 1998/08/01/ %@ 9780898714142 %G eng %0 Report %D 1998 %T Minimum Enclosures with Specified Angles %A Mount, Dave %A Silverman,Ruth %K Technical Report %X Given a convex polygon P, an m-envelope is a convex m-sidedpolygon that contains P. Given any convex polygon P, and any sequence of m > 3 angles A = ((11Xct2X@..ckm) we consider the problem of computing the minimum area m-envelope for P whose counte rclockwise sequence of exterior angles is given by A. We show that such envelopes can be computed in O(nm log m) time. The main result on which the correctness of the algorithm rests is a flushness condition stating that for any locally minimum enclosure with specified angles, one of its sides must be collinear with one of the sides of P. (Also cross-referenced as CAR-TR-701) %I Department of Computer Science, University of Maryland, College Park %V CS-TR-3219 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu//handle/1903/405 %0 Conference Paper %B , Proceedings of the Thirty-First Hawaii International Conference on System Sciences, 1998 %D 1998 %T Multicommodity flow and circuit switching %A Leighton,T. %A Rao,S. %A Srinivasan, Aravind %K Application software %K Bandwidth %K circuit switching %K Computer science %K constant-degree expanders %K graph theory %K High speed integrated circuits %K Integrated circuit technology %K Laboratories %K low congestion %K MATHEMATICS %K multicommodity flow %K National electric code %K network routing %K path selections %K Routing %K short flow paths %K Switching circuits %K switching theory %K undirected graphs %K virtual circuit routing %X Given a set of request pairs in a network, the problem of routing virtual circuits with low congestion is to connect each pair by a path so that few paths use the same link in the network. We build on an earlier multicommodity flow based approach of Leighton and Rao (1996) to show that short flow paths lead to path selections with low congestion. This shows that such good path selections exist for constant-degree expanders with strong expansion, generalizing a result of (Broder et al., 1994). We also show, for infinitely many n, n-vertex undirected graphs Gn along with a set T of connection requests, such that: T is fractionally realizable using flow-paths that impose a (fractional) congestion of at most 1; but any rounding of such a flow to the given set of flow-paths, leads to a congestion of Ω(log n/log log n). This is progress on a long-standing open problem %B , Proceedings of the Thirty-First Hawaii International Conference on System Sciences, 1998 %I IEEE %V 7 %P 459-465 vol.7 - 459-465 vol.7 %8 1998/01/06/9 %@ 0-8186-8255-8 %G eng %R 10.1109/HICSS.1998.649241 %0 Journal Article %J Journal of the ACM (JACM) %D 1998 %T An optimal algorithm for approximate nearest neighbor searching fixed dimensions %A Arya,Sunil %A Mount, Dave %A Netanyahu,Nathan S. %A Silverman,Ruth %A Wu,Angela Y. %K Approximation algorithms %K Box-decomposition trees %K closet-point queries %K nearest neighbor searching %K post-office problem %K priority search %X Consider a set of S of n data points in real d-dimensional space, Rd, where distances are measured using any Minkowski metric. In nearest neighbor searching, we preprocess S into a data structure, so that given any query point q ∈ Rd, is the closest point of S to q can be reported quickly. Given any positive real &egr;, data point p is a (1 +&egr;)-approximate nearest neighbor of q if its distance from q is within a factor of (1 + &egr;) of the distance to the true nearest neighbor. We show that it is possible to preprocess a set of n points in Rd in O(dn log n) time and O(dn) space, so that given a query point q ∈ Rd, and &egr; > 0, a (1 + &egr;)-approximate nearest neighbor of q can be computed in O(cd, &egr; log n) time, where cd,&egr;≤d 1 + 6d/e;d is a factor depending only on dimension and &egr;. In general, we show that given an integer k ≥ 1, (1 + &egr;)-approximations to the k nearest neighbors of q can be computed in additional O(kd log n) time. %B Journal of the ACM (JACM) %V 45 %P 891 - 923 %8 1998/11// %@ 0004-5411 %G eng %U http://doi.acm.org/10.1145/293347.293348 %N 6 %R 10.1145/293347.293348 %0 Conference Paper %B ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI) %D 1998 %T Partial online cycle elimination in inclusion constraint graphs %A Aiken,A. %A Fhndrich,M. %A Foster, Jeffrey S. %A Su,Z. %B ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI) %8 1998/// %G eng %0 Conference Paper %B , The 19th IEEE Real-Time Systems Symposium, 1998. Proceedings %D 1998 %T Performance measurement using low perturbation and high precision hardware assists %A Mink, A. %A Salamon, W. %A Hollingsworth, Jeffrey K %A Arunachalam, R. %K Clocks %K Computerized monitoring %K Counting circuits %K Debugging %K Hardware %K hardware performance monitor %K high precision hardware assists %K low perturbation %K measurement %K MPI message passing library %K MultiKron hardware performance monitor %K MultiKron PCI %K NIST %K online performance monitoring tools %K Paradyn parallel performance measurement tools %K PCI bus slot %K performance bug %K performance evaluation %K performance measurement %K program debugging %K program testing %K real-time systems %K Runtime %K Timing %X We present the design and implementation of MultiKron PCI, a hardware performance monitor that can be plugged into any computer with a free PCI bus slot. The monitor provides a series of high-resolution timers, and the ability to monitor the utilization of the PCI bus. We also demonstrate how the monitor can be integrated with online performance monitoring tools such as the Paradyn parallel performance measurement tools to improve the overhead of key timer operations by a factor of 25. In addition, we present a series of case studies using the MultiKron hardware performance monitor to measure and tune high-performance parallel completing applications. By using the monitor, we were able to find and correct a performance bug in a popular implementation of the MPI message passing library that caused some communication primitives to run at one half their potential speed %B , The 19th IEEE Real-Time Systems Symposium, 1998. Proceedings %I IEEE %P 379 - 388 %8 1998/12/02/4 %@ 0-8186-9212-X %G eng %R 10.1109/REAL.1998.739771 %0 Report %D 1998 %T On the Perturbation of LU, Cholesky, and QR Factorizations %A Stewart, G.W. %K Technical Report %X To appear in SIMAXIn this paper error bounds are derived for a first order expansion of the LU factorization of a perturbation of the identity. The results are applied to obtain perturbation expansions of the LU, Cholesky, and QR factorizations. (Also cross-referenced as UMIACS-TR-92-24) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-92-24 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/565 %0 Report %D 1998 %T Perturbation Theory for the Singular Value Decomposition %A Stewart, G.W. %K Technical Report %X The singular value decomposition has a number of applications in digitalsignal processing. However, the the decomposition must be computed from a matrix consisting of both signal and noise. It is therefore important to be able to assess the effects of the noise on the singular values and singular vectors\,---\,a problem in classical perturbation theory. In this paper we survey the perturbation theory of the singular value decomposition. (Also cross-referenced as UMIACS-TR-90-124) Appeared in SVD and Signal Processing, II, R. J. Vacarro ed., Elsevier, Amsterdam, 1991. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-90-124 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/552 %0 Conference Paper %B Fourteenth International Conference on Pattern Recognition, 1998. Proceedings %D 1998 %T Pictorial query trees for query specification in image databases %A Soffer,A. %A Samet, Hanan %A Zotkin,Dmitry N %K Automation %K complex queries %K Computer science %K content-based retrieval %K Database systems %K database-image objects %K Educational institutions %K Electrical capacitance tomography %K grammars %K Image databases %K Image matching %K parsing %K pictorial query trees %K Postal services %K query specification %K query-image objects %K spatial constraints %K syntax %K visual databases %X A technique that enables specifying complex queries in image databases using pictorial query trees is presented. The leaves of a pictorial query tree correspond to individual pictorial queries that specify which objects should appear in the target images as well as how many occurrences of each object are required. In addition, the minimum required certainty of matching between query-image objects and database-image objects, as well as spatial constraints that specify bounds on the distance between objects and the relative direction between them are also specified. Internal nodes in the query tree represent logical operations (AND, OR, XOR) and their negations on the set of pictorial queries (or subtrees) represented by its children. The syntax of query trees is described. Algorithms for processing individual pictorial queries and for parsing and computing the overall result of a pictorial query tree are outlined %B Fourteenth International Conference on Pattern Recognition, 1998. Proceedings %I IEEE %V 1 %P 919-921 vol.1 - 919-921 vol.1 %8 1998/08// %@ 0-8186-8512-3 %G eng %R 10.1109/ICPR.1998.711383 %0 Journal Article %J CONCUR'98 Concurrency Theory %D 1998 %T Probabilistic resource failure in real-time process algebra %A Philippou,A. %A Cleaveland, Rance %A Lee,I. %A Smolka,S. %A Sokolsky,O. %B CONCUR'98 Concurrency Theory %P 465 - 472 %8 1998/// %G eng %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Providing Advisory Notices for UNIX Command Users: Design, Implementation, and Empirical Evaluations %A Kuah,Boon-Teck %A Shneiderman, Ben %K Technical Report %X UNIX Notices (UN) was developed to study the problems in providing advice tousers of complex systems. The issues studied were: what, when, and how to present the advice. The first experiment with 24 subjects examined how different presentation styles affect the effectiveness of UNs advice. The three presentation styles studied were: notice appears in separate window; notice appears only on request; notice appears in users window immediately. The results showed that the third style was significantly more effective than the first style. Furthermore, the results indicated that the most effective presentation method is also the most disruptive. The second experiment with 29 subjects studied how delay in the advice feedback affects the performance of UN. The treatments were: immediate feedback, feedback at end of session, and no feedback. Over a period of 6 weeks, the commands entered by the subjects were logged and studied. The results showed that immediate feedback caused subjects to repeat significantly fewer inefficient command sequences. However, immediate feedback and feedback at end of session may have given subjects a negative feeling towards UNIX. (Also cross-referenced as CAR-TR-651) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/387 %0 Report %D 1998 %T QR Sometimes Beats Jacobi %A Stewart, G.W. %K Technical Report %X This note exhibits a symmetric matrix having a small eigenvalue that iscomputed accurately by the QR algorithm but not by Jacobi's method. (Also cross-referenced as UMIACS-TR-95-32) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-95-32 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/709 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Reflections on authoring, editing and managing hypertex %A Shneiderman, Ben %K Technical Report %X This chapter offers recommendations for potential authors of hypertextdocuments based on the experience of designing a hypertext system and of creating a series of substantial hypertext databases on personal computers and larger workstations. Advice on choosing projects, identifying useful author tool features, and structuring knowledge is presented. Additional issues such as the design of the root document, article size, and conversion from existing databases are covered. While hypertext has exciting potentials, the dangers of poor design must be overcome to create attractive and effective products. (Also cross-referenced as CAR-TR-410) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/357 %0 Journal Article %J Computers & Education %D 1998 %T Relate–Create–Donate: a teaching/learning philosophy for the cyber-generation %A Shneiderman, Ben %B Computers & Education %V 31 %P 25 - 39 %8 1998/08// %@ 0360-1315 %G eng %U http://www.sciencedirect.com/science/article/pii/S0360131598000141 %N 1 %R 10.1016/S0360-1315(98)00014-1 %0 Report %D 1998 %T Rounding Errors in Solving Block Hessenberg Systems %A Von Matt,Urs %A Stewart, G.W. %K Technical Report %X A rounding error analysis is presented for a divide-and-conquer algorithm to solve linear systems with block Hessenberg matrices. Conditions are derived under which the algorithm computes a backward stable solution. The algorithm is shown to be stable for diagonally dominant matrices and for M-matrices. (Also cross-referenced as UMIACS-TR-94-105) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-94-105 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/661 %0 Journal Article %J IEEE Communications Magazine %D 1998 %T Safety and security of programmable network infrastructures %A Alexander,S. %A Arbaugh, William A. %A Keromytis,A. D %A Smith,J. M %K Access control %K error protection %K IP networks %K Multicast protocols %K network architecture %K network operating systems %K network service model %K operating system %K Power system dynamics %K Power system modeling %K Power system reliability %K programmable languages %K programmable network infrastructures %K programming languages %K Proposals %K Protection %K reliability properties %K Safety %K Secure Active Network Environment %K Security %K security of data %K service creation %K service providers %K Switches %K telecommunication computing %K telecommunication network reliability %K Web and internet services %X Safety and security are two reliability properties of a system. A “safe” system provides protection against errors of trusted users, while a “secure” system protects against errors introduced by untrusted users. There is considerable overlap between mechanisms to support each property. Requirements for rapid service creation have stimulated the development of programmable network infrastructures, where end users or service providers can customize the properties of a network infrastructure while it continues to operate. A central concern of potential users of such systems is their reliability and, most specifically, their safety and security. In this article we explain the impact the network service model and architecture have on safety and security, and provide a model with which policies can be translated into restrictions of a general system. We illustrate these ideas with the Secure Active Network Environment (SANE) architecture, which provides a means of controlling access to the functions provided by any programmable infrastructure %B IEEE Communications Magazine %V 36 %P 84 - 92 %8 1998/10// %@ 0163-6804 %G eng %N 10 %R 10.1109/35.722141 %0 Journal Article %J Scientific Visualization: Interactions, Features, Metaphors %D 1998 %T Salient Frame Detection for Molecular Dynamics Simulations %A Kim,Y. %A Patro,R. %A Ip,C. Y %A O’Leary,D. P %A Anishkin,A. %A Sukharev,S. %A Varshney, Amitabh %X Saliency-based analysis can be applied to time-varying 3D datasetsfor the purpose of summarization, abstraction, and motion analysis. As the sizes of time-varying datasets continue to grow, it becomes more and more difficult to comprehend vast amounts of data and information in a short period of time. Au- tomatically generated thumbnail images and previewing of time-varying datasets can help viewers explore and understand the datasets significantly faster as well as provide new insights. In this paper, we introduce a novel method for detect- ing salient frames for molecular dynamics simulations. Our method effectively detects crucial transitions in simulated mechanosensitive ion channel (MscS), in agreement with experimental data. %B Scientific Visualization: Interactions, Features, Metaphors %V 2 %P 160 - 175 %8 1998/// %G eng %0 Journal Article %J IEEE Network %D 1998 %T A secure active network environment architecture: realization in SwitchWare %A Alexander,D. S %A Arbaugh, William A. %A Keromytis,A. D %A Smith,J. M %K access protocols %K AEGIS secure bootstrap architecture %K architecture %K Authentication %K Collaboration %K Communication switching %K dynamic integrity checks %K extended LAN %K Functional programming %K implementation %K integrity %K Intelligent networks %K IP networks %K Local area networks %K network infrastructure %K network infrastructures %K network operating systems %K network-level solutions %K node %K node-to-node authentication %K packet switching %K Proposals %K ramming system %K SANE %K secure active network environment architecture %K security of data %K Switches %K SwitchWare %K trusted state %K Web and internet services %X An active network is a network infrastructure which is programmable on a per-user or even per-packet basis. Increasing the flexibility of such network infrastructures invites new security risks. Coping with these security risks represents the most fundamental contribution of active network research. The security concerns can be divided into those which affect the network as a whole and those which affect individual elements. It is clear that the element problems must be solved first, since the integrity of network-level solutions will be based on trust in the network elements. In this article we describe the architecture and implementation of a secure active network environment (SANE), which we believe provides a basis for implementing secure network-level solutions. We guarantee that a node begins operation in a trusted state with the AEGIS secure bootstrap architecture. We guarantee that the system remains in a trusted state by applying dynamic integrity checks in the network element's runtime system, using a novel naming system, and applying node-to-node authentication when needed. The construction of an extended LAN is discussed %B IEEE Network %V 12 %P 37 - 45 %8 1998/06//May %@ 0890-8044 %G eng %N 3 %R 10.1109/65.690960 %0 Journal Article %J Computer %D 1998 %T Security for virtual private intranets %A Arbaugh, William A. %A Davin,J. R %A Farber,D. J %A Smith,J. M %K businesses %K Clouds %K Companies %K core operating system components %K cryptography %K Data security %K employee homes %K encryption %K functional roles %K hard drive %K Home computing %K home working %K integrity checking %K Internet %K Local area networks %K multiple personalities %K network authentication %K network environment %K operating system modifications %K Operating systems %K Roads %K secure identity based lending %K security management %K security of data %K shared applications %K SIBL %K single hardware platform %K smart cards %K symmetric algorithm %K system partition %K telecommuting %K Teleworking %K trust relationship %K trustworthy system %K virtual private intranets %X As telecommuting grows, businesses must consider security when extending their network environment to employees' homes. Researchers at the University of Pennsylvania have addressed the problem with smart cards, operating system modifications, and network authentication. We note the distinction between trust and integrity: trust is determined through the verification of components and the dependencies among them, while integrity demonstrates that components haven't been modified. Thus integrity checking in a trustworthy system is about preserving an established trust or trust relationship. Our solution to the challenge of isolating functional roles that may share a single hardware platform is called secure identity based lending (SIBL). SIBL provides multiple personalities by partitioning the hard drive into n+1 partitions, where n is the number of supported personalities. All personalities use the system partition for core operating system components and shared applications. Each of the personalities is also associated with one of the remaining partitions, which are encrypted using a symmetric algorithm %B Computer %V 31 %P 48 - 55 %8 1998/09// %@ 0018-9162 %G eng %N 9 %R 10.1109/2.708450 %0 Journal Article %J Communications of the ACM %D 1998 %T Sorting out searching: a user-interface framework for text searches %A Shneiderman, Ben %A Byrd,Donald %A Croft,W. Bruce %B Communications of the ACM %V 41 %P 95 - 98 %8 1998/04// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/273035.273069 %N 4 %R 10.1145/273035.273069 %0 Report %D 1998 %T SRRIT--A FORTRAN Subroutine to Calculate the Dominant Invariant Subspace of a Nonsymmetric Matrix %A Bai,Z. %A Stewart, G.W. %K Technical Report %X SRRIT is a FORTRAN program to calculate an approximateorthonormal basis for a dominant invariant subspace of a real matrix $A$ by the method of simultaneous iteration \cite{stewart76a}. Specifically, given an integer $m$, {\sl SRRIT} attempts to compute a matrix $Q$ with $m$ orthonormal columns and real quasi-triangular matrix $T$ of order $m$ such that the equation \[ AQ = QT \] is satisfied up to a tolerance specified by the user. The eigenvalues of $T$ are approximations to the $m$ largest eigenvalues of $A$, and the columns of $Q$ span the invariant subspace corresponding to those eigenvalues. {\sl SRRIT} references $A$ only through a user provided subroutine to form the product $AQ$; hence it is suitable for large sparse problems. (Also cross-referenced as UMIACS-TR-92-61) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-92-61 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/572 %0 Report %D 1998 %T On Sublinear Convergence %A Stewart, G.W. %K Technical Report %X This note develops a theory of sublinearly convergingsequences, including a categorization of the rates of convergence and a method for determining the rate from an iteration function. (Also cross-referenced as UMIACS-TR-95-92) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-95-92 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/760 %0 Journal Article %J IEEE Network %D 1998 %T The SwitchWare active network architecture %A Alexander,D. S %A Arbaugh, William A. %A Hicks, Michael W. %A Kakkar,P. %A Keromytis,A. D %A Moore,J. T %A Gunter,C. A %A Nettles,S. M %A Smith,J. M %K active extensions %K active packets %K Authentication %K Computer languages %K Computer networks %K cryptography %K cryptography-based authentication %K high-integrity base %K integrity checking %K IP networks %K LAN interconnection %K mobile programs %K network operating systems %K packet switching %K programmable network infrastructure %K programming languages %K Protocols %K Safety %K safety requirements %K scalability %K secure active router infrastructure %K Security %K security requirements %K services %K strong type checking %K Switches %K SwitchWare active network architecture %K telecommunication network routing %K Tin %K usability %K verification techniques %X Active networks must balance the flexibility of a programmable network infrastructure against the safety and security requirements inherent in sharing that infrastructure. Furthermore, this balance must be achieved while maintaining the usability of the network. The SwitchWare active network architecture is a novel approach to achieving this balance using three layers: active packets, which contain mobile programs that replace traditional packets; active extensions, which provide services on the network elements and can be dynamically loaded; and a secure active router infrastructure, which forms a high-integrity base on which the security of the other layers depends. In addition to integrity checking and cryptography-based authentication, security in our architecture depends heavily on verification techniques from programming languages, such as strong type checking %B IEEE Network %V 12 %P 29 - 36 %8 1998/06//May %@ 0890-8044 %G eng %N 3 %R 10.1109/65.690959 %0 Journal Article %J The ML Workshop, International Conference on Functional Programming (ICFP) %D 1998 %T The switchware active network implementation %A Alexander,D. S %A Hicks, Michael W. %A Kakkar,P. %A Keromytis,A. D %A Shaw,M. %A Moore,J. T %A Gunter,C. A %A Jim,T. %A Nettles,S. M %A Smith,J. M %B The ML Workshop, International Conference on Functional Programming (ICFP) %8 1998/// %G eng %0 Journal Article %J ACM SIGMOD Record %D 1998 %T T2 %A Chang,Chialin %A Acharya,Anurag %A Sussman, Alan %A Saltz,Joel %B ACM SIGMOD Record %V 27 %P 58 - 66 %8 1998/03/01/ %@ 01635808 %G eng %U http://dl.acm.org/citation.cfm?id=273264 %R 10.1145/273244.273264 %0 Conference Paper %B Performance, Computing and Communications, 1998. IPCCC '98., IEEE International %D 1998 %T Threshold-based admission control for multi-class video-on-demand systems %A Mundur, Padma %A Sood,A. %A Simon,R. %K Admission control %K admission control tests %K analytical model %K Analytical models %K blocking probabilities %K buffer storage %K Computer architecture %K distributed digital VoD systems %K File servers %K guaranteed service %K High-speed networks %K interactive video %K Motion pictures %K multi-class video-on-demand systems %K multimedia systems %K Network servers %K newly arriving requests %K numerical analysis %K Predictive models %K probability %K recursive algorithm %K system blocking rates %K telecommunication congestion control %K threshold-based admission control %K unified admission control %X The next generation of distributed digital video-on-demand (VoD) systems will use admission control tests to ensure that users receive predictable and guaranteed service. If the system cannot support a new request then that request is blocked This paper presents an analytical model to evaluate unified admission control strategies for distributed VoD systems with multiple video classes. We prove that there exists a computationally efficient technique to determine the blocking probabilities for newly arriving requests. Through numerical analysis we show the effect of different admission control policies on overall system blocking rates %B Performance, Computing and Communications, 1998. IPCCC '98., IEEE International %I IEEE %P 154 - 160 %8 1998/02/16/18 %@ 0-7803-4468-5 %G eng %R 10.1109/PCCC.1998.659937 %0 Report %D 1998 %T Time-Domain Extraction of Broad-Band Sources by Tikhonov-Phillips Regularization of Triangular Toeplitz Kernels* %A Roginsky,Jacob %A Stewart, G.W. %K Technical Report %X Single receiver source deconvolution in a shallow waterenvironment is an ill-posed problem whose difficulty is compounded by the multipath nature of the propagation operator. If only sources that are quiescent prior to some initial time to are considered, the result of discretizatizing the problem in the time domain is an ill-conditioned triangular Toeplitz system. In this paper we show how an algorithm of Elden can be used to implement Tikhonov-Phillips regularization for this system. Unlike the multichannel deconvolution techniques used in underwater acoustics, this method can extract source signatures using the outputs of a single sensor. In addition, when the propagation is multipath and source signature extraction is performed as part of an optimization procedure for environmental inversion, we can work with shorter time windows so that the process becomes computationally more efficient than frequency domain deconvolution. A number of examples of the use of the Tikhonov-Philips regularization method for source series extraction are provided. (Also cross-referenced as UMIACS-TR-95-87) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-95-87 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/755 %0 Journal Article %J ACM Transactions on Graphics (TOG) Volume %D 1998 %T Treemaps for space-constrained visualization of hierarchies %A Shneiderman, Ben %B ACM Transactions on Graphics (TOG) Volume %V 11 %P 92 - 99 %8 1998/// %G eng %0 Book %D 1998 %T Trends in the Early Careers of Life Scientists %A Tilghman,S %A Astin,HS %A Brinkley,W %A Chilton,MD %A Cummings, Michael P. %A Ehrenberg,RG %A Fox,MF %A Glenn,K %A Green,PJ %A Hans,S %A Kelman,A %A LaPidus,J %A Levin,B %A McIntosh,JR %A Riecken,H %A Stephen,PE %I National Academy Press %C Washington, DC %8 1998/// %G eng %0 Journal Article %J Mol Biol Cell %D 1998 %T Trends in the early careers of life scientists - Preface and executive summary %A Tilghman,S %A Astin,HS %A Brinkley,W %A Chilton,MD %A Cummings, Michael P. %A Ehrenberg,RG %A Fox,MF %A Glenn,K %A Green,PJ %A Hans,S %A Kelman,A %A LaPidus,J %A Levin,B %A McIntosh,JR %A Riecken,H %A Stephen,PE %B Mol Biol Cell %V 9 %P 3007 - 3015 %8 1998/11// %G eng %N 11 %0 Report %D 1998 %T The Triangular Matrices of Gaussian Elimination and Related Decompositions %A Stewart, G.W. %K Technical Report %X It has become a commonplace that triangular systems are solved tohigher accuracy than their condition would warrant. This observation is not true in general, and counterexamples are easy to construct. However, it is often true of the triangular matrices from pivoted LU or QR decompositions. It is shown that this fact is closely connected with the rank-revealing character of these decompositions. (Also cross-referenced as UMIACS-TR-95-91) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-95-91 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/759 %0 Report %D 1998 %T Two Algorithms for the The Efficient Computation of Truncated Pivoted QR Approximations to a Sparse Matrix %A Stewart, G.W. %K Technical Report %X In this note we propose two algorithms to compute truncated pivoted QRapproximations to a sparse matrix. One is based on the Gram--Schmidt algorithm, and the other on Householder triangularization. Both algorithms leave the original matrix unchanged, and the only additional storage requirements are arrays to contain the factorization itself. Thus, the algorithms are particularly suited to determining low-rank approximations to a sparse matrix. (Also cross-referenced as UMIACS-TR-98-12) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-98-12 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/941 %0 Report %D 1998 %T Two Simple Residual Bounds for the Eigenvalues of Hermitian Matrices %A Stewart, G.W. %K Technical Report %X Let $A$ be Hermitian and let the orthonormal columns of $X$ span anapproximate invariant subspace of $X$. Then the residual $R = AX-XM$ $(M=X\ctp AX)$ will be small. The theorems of this paper bound the distance of the spectrum of $M$ from the spectrum of $A$ in terms of appropriate norms of $R$. Appeared in SIAM J. Mat. Anal. Appl. 12 (1991) 205--208. (Also cross-referenced as UMIACS-TR-89-123) %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V UMIACS-TR-89-123 %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/545 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Understanding Transportation Management Systems Performance with a Simulation-Based Learning Environment %A Plaisant, Catherine %A Tarnoff,Phil %A Saraf,Aditya %A Rose,Anne %K Technical Report %X We have developed a simulation-based learning environment to provide systemdesigners and operators with an appreciation of the impact of incidents on traffic delay. We used an application framework developed at the University of Maryland for constructing simulation-based learning environments called SimPLE (Simulated Processes in a Learning Environment). Environments developed with SimPLE use dynamic simulations and visualizations to represent realistic time-dependent behavior and are coupled with guidance material and other software aids that facilitate learning. The simulation allows learners to close freeway lanes and divert traffic to an arterial road. Users can see the effect of the detour on freeway and arterial delay. Users can then adjust signal timing interactively on a time space diagram and watch the effect of their adjustment on green band changes and on arterial delays and total delays. %B Technical Reports of the Computer Science Department %8 1998/12/05/ %G eng %U http://drum.lib.umd.edu/handle/1903/500 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T User Interface Reengineering: A Diagnostic Approach %A Vanniamparampil,Ajit J %A Shneiderman, Ben %A Plaisant, Catherine %A Rose,Anne %K Technical Report %X User interface technology has advanced rapidly in recent years.Incorporating new developments in existing systems could result in substantial improvements in usability, thereby improving performance and user satisfaction, while shortening training an d reducing error rates. Our focus is on low-effort high-payoff improvements to aspects such as data display and entry, consistency, messages, documentation, and system access. This paper provides guidelines for managers and designers responsible for use r interface reengineering, based on the experience we gained from six projects, and compiles our observations, recommendations and outcomes. (Also cross-referenced as CAR-TR-767) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/430 %0 Conference Paper %B CHI 98 conference summary on Human factors in computing systems %D 1998 %T Using elastic windows for World-Wide Web Browsing %A Kandogan,Eser %A Shneiderman, Ben %K elastic windows %K User interfaces %K window management %K World-wide web %B CHI 98 conference summary on Human factors in computing systems %S CHI '98 %I ACM %C New York, NY, USA %P 189 - 190 %8 1998/// %@ 1-58113-028-7 %G eng %U http://doi.acm.org/10.1145/286498.286666 %R 10.1145/286498.286666 %0 Conference Paper %B Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence %D 1998 %T Utility elicitation as a classification problem %A Chajewska,Urszula %A Getoor, Lise %A Norman,Joseph %A Shahar,Yuval %X We investigate the application of classification techniques to utility elicitation. In a decision problem, two sets of parameters must generally be elicited: the probabilities and the utilities. While the prior and conditional probabilities in the model do not change from user to user, the utility models do. Thus it is necessary to elicit a utility model separately for each new user. Elicitation is long and tedious, particularly if the outcome space is large and not decomposable. There are two common approaches to utility function elicitation. The first is to base the determination of the user's utility function solely on elicitation of qualitative preferences. The second makes assumptions about the form and decomposability of the utility function. Here we take a different approach: we attempt to identify the new user's utility function based on classification relative to a database of previously collected utility functions. We do this by identifying clusters of utility functions that minimize an appropriate distance measure. Having identified the clusters, we develop a classification scheme that requires many fewer and simpler assessments than full utility elicitation and is more robust than utility elicitation based solely on preferences. We have tested our algorithm on a small database of utility functions in a prenatal diagnosis domain and the results are quite promising. %B Proceedings of the Fourteenth conference on Uncertainty in artificial intelligence %S UAI'98 %I Morgan Kaufmann Publishers Inc. %C San Francisco, CA, USA %P 79 - 88 %8 1998/// %@ 1-55860-555-X %G eng %U http://dl.acm.org/citation.cfm?id=2074094.2074104 %0 Journal Article %J Technical Reports of the Computer Science Department %D 1998 %T Visual User Interfaces for Information Exploration %A Shneiderman, Ben %K Technical Report %X The next generation of database management, directory browsing,information retrieval, hypermedia, scientific data management, and library systems can enable convenient exploration of growing information spaces by a wider range of users. User interface designers can provide more powerful search techniques, more comprehensible query facilities, better presentation methods, and smoother integration of technology with task. This paper offers novel graphical and direct manipulation approaches to query formulation and information presentation/manipulation. These approaches include a graphical approach to restricted boolean query formulation based on generalization/aggregation hierarchies, a filter/flow metaphor for complete boolean expressions, dynamic query methods with continuous visual presentation of results as the query is changed (possibly employing parallel computation), and color-coded 2-dimensional space-filling tree-maps that present multiple-level hierarchies in a single display (hundreds of directories and more than a thousand files can be seen at once). (Also cross-referenced as CAR-TR-577) %B Technical Reports of the Computer Science Department %8 1998/10/15/ %G eng %U http://drum.lib.umd.edu/handle/1903/373 %0 Conference Paper %B CHI 98 conference summary on Human factors in computing systems %D 1998 %T Visualizing medical records with LifeLines %A Plaisant, Catherine %A Heller,Daniel %A Li,Jia %A Shneiderman, Ben %A Mushlin,Rich %A Karat,John %K healthcare %K History %K medical record %K overview %K timeline %K Visualization %B CHI 98 conference summary on Human factors in computing systems %S CHI '98 %I ACM %C New York, NY, USA %P 28 - 29 %8 1998/// %@ 1-58113-028-7 %G eng %U http://doi.acm.org/10.1145/286498.286513 %R 10.1145/286498.286513 %0 Journal Article %J ACM SIGCHI Bulletin %D 1998 %T Visualizing Personal Histories: A Workshop, July 21-22, 1997 %A Shneiderman, Ben %B ACM SIGCHI Bulletin %V 30 %P 34 - 35 %8 1998/01// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/280571.280585 %N 1 %R 10.1145/280571.280585 %0 Conference Paper %B CHI 98 conference summary on Human factors in computing systems %D 1998 %T Is the Web really different from everything else? %A Shneiderman, Ben %A Nielsen,Jakob %A Butler,Scott %A Levi,Michael %A Conrad,Frederick %K design %K user experience %K World Wide Web %B CHI 98 conference summary on Human factors in computing systems %S CHI '98 %I ACM %C New York, NY, USA %P 92 - 93 %8 1998/// %@ 1-58113-028-7 %G eng %U http://doi.acm.org/10.1145/286498.286545 %R 10.1145/286498.286545 %0 Conference Paper %B Extended Abstracts %D 1998 %T When Two Hands Are Better Than One: Enhancing Collaboration Using Single Display Groupware. CHI’98 %A Stewart,J. %A Raybourn,E. %A Bederson, Benjamin B. %A Druin, Allison %B Extended Abstracts %P 287 - 288 %8 1998/// %G eng %0 Journal Article %J SIAM Review %D 1997 %T Accuracy and Stability of Numerical Algorithms %A Higham,N. J %A Stewart, G.W. %B SIAM Review %V 39 %P 164 - 165 %8 1997/// %G eng %N 1 %0 Conference Paper %B ICDAR %D 1997 %T ADistributed Management System for Testing Document Image Database Analysis Algorithms %A David Doermann %A Sauvola,J. %A Haapakoski,S. %A Kauniskangas,H. %A Seppanen,T. %A Pietikainen,M. %B ICDAR %P 989 - 995 %8 1997/// %G eng %0 Journal Article %J ACM Trans. Math. Softw. %D 1997 %T Algorithm 776: SRRIT: a Fortran subroutine to calculate the dominant invariant subspace of a nonsymmetric matrix %A Bai,Z. %A Stewart, G.W. %K invariant subspace %K nonsymmetric eigenvalue problem %K project method %X SRRT is a Fortran program to calculate an approximate orthonomral basis fr a dominant invariant subspace of a real matrix A by the method of simultaneous iteration. Specifically, given an integer m, SRRIT computes a matrix Q with m orthonormal columns and real quasi-triangular matrix T or order m such that the equation AQ = QT is satisfied up to a tolerance specified by the user. The eigenvalues of T are approximations to the m eigenvalues of largest absolute magnitude of A and the columns of Q span the invariant subspace corresponding to those eigenvalues. SRRIT references A only through a user-provided subroutine to form the product AQ; hence it is suitable for large sparse problems. %B ACM Trans. Math. Softw. %V 23 %P 494 - 513 %8 1997/12// %@ 0098-3500 %G eng %U http://doi.acm.org/10.1145/279232.279234 %N 4 %R 10.1145/279232.279234 %0 Conference Paper %B Proceedings of the twenty-ninth annual ACM symposium on Theory of computing %D 1997 %T Approximating hyper-rectangles: learning and pseudo-random sets %A Auer,Peter %A Long,Philip M. %A Srinivasan, Aravind %K approximations of distributions %K derandomization %K explicit constructions %K machine learning %K multiple-instance learning %K PAC learning %K pseudorandomness %K Ramsey graphs %K random graphs %K rectangles %K sample complexity %B Proceedings of the twenty-ninth annual ACM symposium on Theory of computing %S STOC '97 %I ACM %C New York, NY, USA %P 314 - 323 %8 1997/// %@ 0-89791-888-6 %G eng %U http://doi.acm.org/10.1145/258533.258611 %R 10.1145/258533.258611 %0 Conference Paper %B Proceedings of the eighth annual ACM-SIAM symposium on Discrete algorithms %D 1997 %T Better approximation guarantees for job-shop scheduling %A Goldberg,Leslie Ann %A Paterson,Mike %A Srinivasan, Aravind %A Sweedyk,Elizabeth %B Proceedings of the eighth annual ACM-SIAM symposium on Discrete algorithms %S SODA '97 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 599 - 608 %8 1997/// %@ 0-89871-390-0 %G eng %U http://dl.acm.org/citation.cfm?id=314161.314395 %0 Journal Article %J Communications of the ACM %D 1997 %T Between hope and fear %A Shneiderman, Ben %B Communications of the ACM %V 40 %P 59 - 62 %8 1997/02// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/253671.253706 %N 2 %R 10.1145/253671.253706 %0 Journal Article %J Interactions-New York %D 1997 %T Book Preview-Designing the User Interface: Strategies for Effective Human-Computer Interaction %A Shneiderman, Ben %B Interactions-New York %V 4 %P 61 - 61 %8 1997/// %G eng %N 5 %0 Journal Article %J International Journal of Human-Computer Studies %D 1997 %T Browsing hierarchical data with multi-level dynamic queries and pruning %A Kumar,Harsha P. %A Plaisant, Catherine %A Shneiderman, Ben %X Users often must browse hierarchies with thousands of nodes in search of those that best match their information needs. ThePDQ Tree-browser(Pruning with Dynamic Queries) visualization tool was specified, designed and developed for this purpose. This tool presents trees in two tightly-coupled views, one a detailed view and the other an overview. Users can use dynamic queries, a method for rapidly filtering data, to filter nodes at each level of the tree. The dynamic query panels are user-customizable. Sub-trees of unselected nodes are pruned out, leading to compact views of relevant nodes. Usability testing of the PDQ Tree-browser, done with eight subjects, helped assess strengths and identify possible improvements. The PDQ Tree-browser was used in Network Management (600 nodes) and UniversityFinder (1100 nodes) applications. A controlled experiment, with 24 subjects, showed that pruning significantly improved performance speed and subjective user satisfaction. Future research directions are suggested. %B International Journal of Human-Computer Studies %V 46 %P 103 - 124 %8 1997/01// %@ 1071-5819 %G eng %U http://www.sciencedirect.com/science/article/pii/S1071581996900853 %N 1 %R 10.1006/ijhc.1996.0085 %0 Journal Article %J D-lib magazine %D 1997 %T Clarifying search: A user-interface framework for text searches %A Shneiderman, Ben %A Byrd,D. %A Croft, W.B. %X Current user interfaces for textual database searching leave much to be desired: individually, they are often confusing, and as a group, they are seriously inconsistent. We propose a four- phase framework for user-interface design: the framework provides common structure and terminology for searching while preserving the distinct features of individual collections and search mechanisms. Users will benefit from faster learning, increased comprehension, and better control, leading to more effective searches and higher satisfaction. %B D-lib magazine %V 3 %P 18 - 20 %8 1997/// %G eng %N 1 %0 Book Section %B Research and Advanced Technology for Digital LibrariesResearch and Advanced Technology for Digital Libraries %D 1997 %T Comprehension and object recognition capabilities for presentations of simultaneous video key frame surrogates %A Slaughters,Laura %A Shneiderman, Ben %A Marchionini,Gary %E Peters,Carol %E Thanos,Costantino %X The demand for more efficient browsing of video data is expected to increase as greater access to this type of data becomes available. This experiment looked at one technique for displaying video data using key frame surrogates that are presented as a ldquoslide showrdquo. Subjects viewed key frames for between one and four video clips simultaneously. Following this presentation, the subjects performed object recognition and gist comprehension tasks in order to determine human thresholds for divided attention between these multiple displays. It was our belief that subject performance would degrade as the number of slide shows shown simultaneously increased. For object recognition and gist comprehension tasks, a decrease in performance between the one slide show display and the two, three or four slide show displays was found. In the case of two or three video presentations, performance is about the same, and there remains adequate object recognition abilities and comprehension of the video clips. Performance drops off to unacceptable levels when four slide shows are displayed at once. %B Research and Advanced Technology for Digital LibrariesResearch and Advanced Technology for Digital Libraries %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1324 %P 41 - 54 %8 1997/// %@ 978-3-540-63554-3 %G eng %U http://dx.doi.org/10.1007/BFb0026720 %0 Conference Paper %B Proceedings of the twenty-ninth annual ACM symposium on Theory of computing %D 1997 %T A constant-factor approximation algorithm for packet routing, and balancing local vs. global criteria %A Srinivasan, Aravind %A Teo,Chung-Piaw %K Approximation algorithms %K covering integer programs %K discrete ham-sandwich theorems %K Linear programming %K packet routing %K randomized algorithms %K Randomized rounding %K rounding theorems %B Proceedings of the twenty-ninth annual ACM symposium on Theory of computing %S STOC '97 %I ACM %C New York, NY, USA %P 636 - 643 %8 1997/// %@ 0-89791-888-6 %G eng %U http://doi.acm.org/10.1145/258533.258658 %R 10.1145/258533.258658 %0 Conference Paper %B Proceedings of the 1997 Scandinavian Conference on Image Analysis %D 1997 %T Content-based Image Retrieval Using Composite Features %A Kauniskangas,H. %A Sauvola,J. %A Pietikainen,M. %A David Doermann %X In this paper, we demonstrate methods for content-based image retrieval problems utilizing a specially designed retrieval architecture implemented within the 'Intelligent Image Retrieval' system (IIR). The method consists of new image features in the retrieval context, including segmentation methods and use of image frames. They are combined in a unique way with color, texture, shape and localization information with a special data ion construction in a graphical user interface. The IIR system provides an efficient retrieval architecture utilizing a tailored query language, retrieval mechanisms and an object-oriented database enabling the use of complex data structures and relations needed for successful query processing. Functionality and performance of methods and architecture are illustrated with a series of tests using a database that consists of several hundred 'scene' images. %B Proceedings of the 1997 Scandinavian Conference on Image Analysis %P 35 - 42 %8 1997/// %G eng %0 Conference Paper %B Proceedings of the third IFIP WG2.6 working conference on Visual database systems 3 (VDB-3) %D 1997 %T Data models, visual representations, metaphors: how to solve the puzzle? (panel) %A Catarci,T. %A Costabile,M. F. %A Cruz,I. F. %A Ioannidis,Y. %A Shneiderman, Ben %K data model %K metaphor %K visual representation %B Proceedings of the third IFIP WG2.6 working conference on Visual database systems 3 (VDB-3) %I Chapman & Hall, Ltd. %C London, UK, UK %P 177 - 182 %8 1997/// %@ 0-412-72170-8 %G eng %U http://dl.acm.org/citation.cfm?id=265394.265440 %0 Conference Paper %B IEEE Symposium on Information Visualization, 1997. Proceedings %D 1997 %T Design and evaluation of incremental data structures and algorithms for dynamic query interfaces %A Tanin,E. %A Beigel,R. %A Shneiderman, Ben %K Algorithm design and analysis %K Bars %K Computer science %K continuous real-time feedback %K Data structures %K data visualisation %K Data visualization %K database access mechanism %K Displays %K DQI algorithms %K dynamic query interfaces %K Feedback %K Graphical user interfaces %K Heuristic algorithms %K incremental data structures %K Information Visualization %K large databases %K Manipulator dynamics %K NASA %K query formulation %K query languages %K Query processing %K real-time systems %K small databases %K User interfaces %K very large databases %K visual databases %K visual languages %X A dynamic query interface (DQI) is a database access mechanism that provides continuous real-time feedback to the user during query formulation. Previous work shows that DQIs are elegant and powerful interfaces to small databases. Unfortunately, when applied to large databases, previous DQI algorithms slow to a crawl. We present a new incremental approach to DQI algorithms and display updates that work well with large databases, both in theory and in practice. %B IEEE Symposium on Information Visualization, 1997. Proceedings %I IEEE %P 81 - 86 %8 1997/10/21/21 %@ 0-8186-8189-6 %G eng %R 10.1109/INFVIS.1997.636790 %0 Journal Article %J International Journal of Human-Computer Studies %D 1997 %T Designing information-abundant web sites: issues and recommendations %A Shneiderman, Ben %X “Gradually I began to feel that we were growing something almost organic in a new kind of reality, in cyberspace, growing it out of information…a pulsing tree of data that I loved to climb around in, scanning for new growth.”(Mickey Hart,Drumming at the Edge of Magic: A Journey into the Spirit of Percussion, 1990) “Look at every path closely and deliberately. Try it as many times as you think necessary. Then ask yourself, and yourself alone, one question… Does this path have a heart? If it does, the path is good; if it doesn’t it is of no use.” (Carlos CastanedaThe Teachings of Don Juan) The abundance of information on the World Wide Web has thrilled some, but frightened others. Improved web site design may increase users’ successful experiences and positive attitudes. This review of design issues identifies genres of web sites, goals of designers, communities of users and a spectrum of tasks. Then an Objects/Actions Interface Model is offered as a way to think about designing and evaluating web sites. Finally, search and navigation improvements are described to bring consistency, comprehensibility and user control. %B International Journal of Human-Computer Studies %V 47 %P 5 - 29 %8 1997/07// %@ 1071-5819 %G eng %U http://www.sciencedirect.com/science/article/pii/S1071581997901270 %N 1 %R 10.1006/ijhc.1997.0127 %0 Conference Paper %B Proceedings of the fifth ACM international conference on Multimedia %D 1997 %T Designing interactive multimedia (panel) %A Scarlatos,Lori L. %A Darken,Rudolph P. %A Harada,Komei %A Heeter,Carrie %A Muller,Richard %A Shneiderman, Ben %B Proceedings of the fifth ACM international conference on Multimedia %S MULTIMEDIA '97 %I ACM %C New York, NY, USA %P 215 - 218 %8 1997/// %@ 0-89791-991-2 %G eng %U http://doi.acm.org/10.1145/266180.266370 %R 10.1145/266180.266370 %0 Conference Paper %B Proceedings of the 2nd international conference on Intelligent user interfaces %D 1997 %T Direct manipulation for comprehensible, predictable and controllable user interfaces %A Shneiderman, Ben %B Proceedings of the 2nd international conference on Intelligent user interfaces %P 33 - 39 %8 1997/// %G eng %0 Book Section %B Software agentsSoftware agents %D 1997 %T Direct manipulation versus agents: Paths to predictable, controllable, and comprehensible interfaces %A Shneiderman, Ben %B Software agentsSoftware agents %P 97 - 106 %8 1997/// %G eng %0 Journal Article %J Interactions %D 1997 %T Direct manipulation vs. interface agents %A Shneiderman, Ben %A Maes,P. %B Interactions %V 4 %P 42 - 61 %8 1997/// %G eng %N 6 %0 Journal Article %J ACM Transactions on Graphics %D 1997 %T Efficient triangle strips for fast rendering %A Evans,F. %A Skiena,S. %A Varshney, Amitabh %B ACM Transactions on Graphics %8 1997/// %G eng %0 Conference Paper %B Proceedings of the 10th annual ACM symposium on User interface software and technology %D 1997 %T Elastic Windows: a hierarchical multi-window World-Wide Web browser %A Kandogan,E. %A Shneiderman, Ben %B Proceedings of the 10th annual ACM symposium on User interface software and technology %P 169 - 177 %8 1997/// %G eng %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 1997 %T Elastic Windows: evaluation of multi-window operations %A Kandogan,E. %A Shneiderman, Ben %B Proceedings of the SIGCHI conference on Human factors in computing systems %P 250 - 257 %8 1997/// %G eng %0 Journal Article %J Proceedings of ASME Design Engineering Technical Conferences, Sept %D 1997 %T Enabling virtual reality for large-scale mechanical CAD datasets %A Varshney, Amitabh %A El-Sana,J. %A Evans,F. %A Darsa,L. %A Costa,B. %A Skiena,S. %X Reconciling scene realism with interactivity has emerged asone of the most important areas in making virtual reality fea- sible for large-scale mechanical CAD datasets consisting of sev- eral millions of primitives. This paper surveys our research and related work for achieving interactivity without sacrificing real- ism in virtual reality walkthroughs and flythroughs of polygonal CAD datasets. We outline our recent work on efficient genera- tion of triangle strips from polygonal models that takes advan- tage of compression of connectivity information. This results in substantial savings in rendering, transmission, and storage. We outline our work on genus-reducing simplifications as well as real-time view-dependent simplifications that allow on-the-fly selection amongst multiple levels of detail, based upon lighting and viewing parameters. Our method allows multiple levels of detail to coexist on the same object at different regions and to merge seamlessly without any cracks or shading artifacts. We also present an overview of our work on hardware-assisted image- based rendering that allows interactive exploration of computer- generated scenes. %B Proceedings of ASME Design Engineering Technical Conferences, Sept %V 14 %P 17 - 17 %8 1997/// %G eng %0 Journal Article %J IEEE Transactions on Software Engineering %D 1997 %T An experiment to assess the cost-benefits of code inspections in large scale software development %A Porter, Adam %A Siy,H. P %A Toman,C. A %A Votta,L. G. %K Analysis of variance %K code inspection cost benefits %K code unit %K commercial software product %K Computer Society %K cost-benefit analysis %K Costs %K defect detection effectiveness %K defect detection rate %K Design optimization %K experienced software developers %K experiment design %K independent variables %K Inspection %K inspection interval %K inspection team %K Laboratories %K large scale software development %K Large-scale systems %K live development effort %K long term experiment %K professional aspects %K professional developers %K Programming %K reviewers %K simulation techniques %K software cost estimation %K software inspection methods %K software practitioners %K Software quality %K Switches %X We conducted a long term experiment to compare the costs and benefits of several different software inspection methods. These methods were applied by professional developers to a commercial software product they were creating. Because the laboratory for this experiment was a live development effort, we took special care to minimize cost and risk to the project, while maximizing our ability to gather useful data. The article has several goals: (1) to describe the experiment's design and show how we used simulation techniques to optimize it; (2) to present our results and discuss their implications for both software practitioners and researchers; and (3) to discuss several new questions raised by our findings. For each inspection, we randomly assigned three independent variables: (1) the number of reviewers on each inspection team (1, 2, or 4); (2) the number of teams inspecting the code unit (1 or 2); and (3) the requirement that defects be repaired between the first and second team's inspections. The reviewers for each inspection were randomly selected without replacement from a pool of 11 experienced software developers. The dependent variables for each inspection included inspection interval (elapsed time), total effort, and the defect detection rate. Our results showed that these treatments did not significantly influence the defect detection effectiveness, but that certain combinations of changes dramatically increased the inspection interval %B IEEE Transactions on Software Engineering %V 23 %P 329 - 346 %8 1997/06// %@ 0098-5589 %G eng %N 6 %R 10.1109/32.601071 %0 Journal Article %J IEEE Software %D 1997 %T A framework for search interfaces %A Shneiderman, Ben %K Abstracts %K Cities and towns %K Databases %K Delay %K design practice %K four-phase framework %K frequent users %K Information retrieval %K Information services %K Libraries %K multimedia libraries %K online front-ends %K popular search systems %K predictable design %K Protection %K relevance ranked list %K search interfaces %K search results %K stand alone systems %K textual database searching %K Thesauri %K User interfaces %K user performance %K word processing %K World Wide Web %X Searching textual databases can be confusing for users. Popular search systems for the World Wide Web and stand alone systems typically provide a simple interface: users type in keywords and receive a relevance ranked list of 10 results. This is appealing in its simplicity, but users are often frustrated because search results are confusing or aspects of the search are out of their control. If we are to improve user performance, reduce mistaken assumptions, and increase successful searches, we need more predictable design. To coordinate design practice, we suggest a four-phase framework that would satisfy first time, intermittent, and frequent users accessing a variety of textual and multimedia libraries %B IEEE Software %V 14 %P 18 - 20 %8 1997/04//Mar %@ 0740-7459 %G eng %N 2 %R 10.1109/52.582969 %0 Book Section %B Algebraic Frames for the Perception-Action CycleAlgebraic Frames for the Perception-Action Cycle %D 1997 %T The geometry of visual space distortion %A Fermüller, Cornelia %A LoongFah Cheong %A Aloimonos, J. %E Sommer,Gerald %E Koenderink,Jan %X The encounter of perception and action happens at the intermediate representations of space-time. In many of the computational models employed in the past, it has been assumed that a metric representation of physical space can be derived by visual means. Psychophysical experiments, as well as computational considerations, can convince us that the perception of space and shape has a much more complicated nature, and that only a distorted version of actual, physical space can be computed. This paper develops a computational geometric model that explains why such distortion might take place. The basic idea is that, both in stereo and motion, we perceive the world from multiple views. Given the rigid transformation between the views and the properties of the image correspondence, the depth of the scene can be obtained. Even a slight error in the rigid transformation parameters causes distortion of the computed depth of the scene. The unified framework introduced here describes this distortion in computational terms. We characterize the space of distortions by its level sets, that is, we characterize the systematic distortion via a family of iso-distortion surfaces which describes the locus over which depths are distorted by some multiplicative factor. Clearly, functions of the distorted space exhibiting some sort of invariance, produce desirable representations for biological and artificial systems [13]. Given that humans' estimation of egomotion or estimation of the extrinsic parameters of the stereo apparatus is likely to be imprecise, the framework is used to explain a number of psychophysical experiments on the perception of depth from motion or stereo. %B Algebraic Frames for the Perception-Action CycleAlgebraic Frames for the Perception-Action Cycle %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1315 %P 249 - 277 %8 1997/// %@ 978-3-540-63517-8 %G eng %U http://dx.doi.org/10.1007/BFb0017872 %0 Journal Article %J Educom review %D 1997 %T A grander goal: A thousand-fold increase in human capabilities %A Shneiderman, Ben %B Educom review %V 32 %P 4 - 10 %8 1997/// %G eng %N 6 %0 Conference Paper %B BSDIA %D 1997 %T Graphical Tools and Techniques for Querying Document Databases %A Sauvola,J. %A David Doermann %A Kauniskangas,H. %A Shin,C. %A Koivusaari,M. %A Pietikainen,M. %B BSDIA %P 213 - 224 %8 1997/// %G eng %0 Conference Paper %B ICSE Workshop on Process Modelling and Empirical Studies of Software Engineering %D 1997 %T If your version control system could talk %A Ball,T. %A Kim,J. M %A Porter, Adam %A Siy,H. P %B ICSE Workshop on Process Modelling and Empirical Studies of Software Engineering %8 1997/// %G eng %0 Conference Paper %B Proceedings of the 1997 ACM/IEEE conference on Supercomputing (CDROM) %D 1997 %T Implementing a performance forecasting system for metacomputing: the Network Weather Service %A Wolski,Rich %A Spring, Neil %A Peterson,Chris %X In this paper we describe the design and implementation of a system called the Network Weather Service (NWS) that takes periodic measurements of deliverable resource performance from distributed networked resources, and uses numerical models to dynamically generate forecasts of future performance levels. These performance forecasts, along with measures of performance fluctuation (e.g. the mean square prediction error) and forecast lifetime that the NWS generates, are made available to schedulers and other resource management mechanisms at runtime so that they may determine the quality-of-service that will be available from each resource.We describe the architecture of the NWS and implementations that we have developed and are currently deploying for the Legion [13] and Globus/Nexus [7] metacomputing infrastructures. We also detail NWS forecasts of resource performance using both the Legion and Globus/Nexus implementations. Our results show that simple forecasting techniques substantially outperform measurements of current conditions (commonly used to gauge resource availability and load) in terms of prediction accuracy. In addition, the techniques we have employed are almost as accurate as substantially more complex modeling methods. We compare our techniques to a sophisticated time-series analysis system in terms of forecasting accuracy and computational complexity. %B Proceedings of the 1997 ACM/IEEE conference on Supercomputing (CDROM) %S Supercomputing '97 %I ACM %C New York, NY, USA %P 1 - 19 %8 1997/// %@ 0-89791-985-8 %G eng %U http://doi.acm.org/10.1145/509593.509600 %R 10.1145/509593.509600 %0 Conference Paper %B , 38th Annual Symposium on Foundations of Computer Science, 1997. Proceedings %D 1997 %T Improved approximations for edge-disjoint paths, unsplittable flow, and related routing problems %A Srinivasan, Aravind %K Approximation algorithms %K Bandwidth %K Channel allocation %K computational complexity %K Computer science %K edge-disjoint paths %K graph theory %K High speed integrated circuits %K IEL %K Image motion analysis %K Information systems %K multi-commodity flow relaxation %K Multiprocessor interconnection networks %K network routing %K Optical fiber networks %K Routing %K routing problems %K unsplittable flow %X We present improved approximation algorithms for a family of problems involving edge-disjoint paths and unsplittable flow, and for some related routing problems. The central theme of all our algorithms is the underlying multi-commodity flow relaxation %B , 38th Annual Symposium on Foundations of Computer Science, 1997. Proceedings %I IEEE %P 416 - 425 %8 1997/10/20/22 %@ 0-8186-8197-7 %G eng %R 10.1109/SFCS.1997.646130 %0 Journal Article %J Algorithmica %D 1997 %T Improved parallel approximation of a class of integer programming problems %A Alon,N. %A Srinivasan, Aravind %X We present a method to derandomize RNC algorithms, converting them to NC algorithms. Using it, we show how to approximate a class of NP-hard integer programming problems in NC , to within factors better than the current-best NC algorithms (of Berger and Rompel and Motwani et al. ); in some cases, the approximation factors are as good as the best-known sequential algorithms, due to Raghavan. This class includes problems such as global wire-routing in VLSI gate arrays and a generalization of telephone network planning in SONET rings. Also for a subfamily of the “packing” integer programs, we provide the first NC approximation algorithms; this includes problems such as maximum matchings in hypergraphs, and generalizations. The key to the utility of our method is that it involves sums of superpolynomially many terms, which can however be computed in NC ; this superpolynomiality is the bottleneck for some earlier approaches, due to Berger and Rompel and Motwani et al. %B Algorithmica %V 17 %P 449 - 462 %8 1997/// %@ 0178-4617 %G eng %U http://dx.doi.org/10.1007/BF02523683 %N 4 %0 Conference Paper %B Proceedings of the eighth annual ACM-SIAM symposium on Discrete algorithms %D 1997 %T Improving the discrepancy bound for sparse matrices: better approximations for sparse lattice approximation problems %A Srinivasan, Aravind %B Proceedings of the eighth annual ACM-SIAM symposium on Discrete algorithms %S SODA '97 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 692 - 701 %8 1997/// %@ 0-89871-390-0 %G eng %U http://dl.acm.org/citation.cfm?id=314161.314418 %0 Conference Paper %B CHI '97 extended abstracts on Human factors in computing systems: looking to the future %D 1997 %T Intelligent software agents vs. user-controlled direct manipulation: a debate %A Maes,Pattie %A Shneiderman, Ben %A Miller,Jim %K agents %K direct manipulation %K graphical representation %K intelligent interfaces %X Critical issues in human-computer interaction - in particular, the advantages and disadvantages of intelligent agents and direct manipulation - will be discussed, debated, and hotly contested. The intent of the participants is to strike an appropriate balance between a serious discussion of the issues and an entertaining debate. %B CHI '97 extended abstracts on Human factors in computing systems: looking to the future %S CHI EA '97 %I ACM %C New York, NY, USA %P 105 - 106 %8 1997/// %@ 0-89791-926-2 %G eng %U http://doi.acm.org/10.1145/1120212.1120281 %R 10.1145/1120212.1120281 %0 Conference Paper %B IPPS %D 1997 %T Interoperability of data parallel runtime libraries %A Edjlali,G. %A Sussman, Alan %A Saltz, J. %B IPPS %P 451 - 451 %8 1997/// %G eng %0 Journal Article %J Iterative Methods in Scientific Computing %D 1997 %T Iterative methods for problems in computational fluid dynamics %A Elman, Howard %A Silvester, D. J %A Wathen, A. J %B Iterative Methods in Scientific Computing %P 271 - 327 %8 1997/// %G eng %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 1997 %T KidPad: a design collaboration between children, technologists, and educators %A Druin, Allison %A Stewart,J. %A Proft,D. %A Bederson, Benjamin B. %A Hollan,J. %B Proceedings of the SIGCHI conference on Human factors in computing systems %P 463 - 470 %8 1997/// %G eng %0 Journal Article %J Journal of Computational Biology %D 1997 %T Local Rules for Protein Folding on a Triangular Lattice and Generalized Hydrophobicity in the HP Model %A Agarwala,Richa %A Batzoglou,Serafim %A DančíK,Vlado %A Decatur,Scott E. %A Hannenhalli, Sridhar %A Farach,Martin %A Muthukrishnan,S. %A Skiena,Steven %X We consider the problem of determining the three-dimensional folding of a protein given its one-dimensional amino acid sequence. We use the HP model for protein folding proposed by Dill (1985), which models protein as a chain of amino acid residues that are either hydrophobic or polar, and hydrophobic interactions are the dominant initial driving force for the protein folding. Hart and Istrail (1996a) gave approximation algorithms for folding proteins on the cubic lattice under the HP model. In this paper, we examine the choice of a lattice by considering its algorithmic and geometric implications and argue that the triangular lattice is a more reasonable choice. We present a set of folding rules for a triangular lattice and analyze the approximation ratio they achieve. In addition, we introduce a generalization of the HP model to account for residues having different levels of hydrophobicity. After describing the biological foundation for this generalization, we show that in the new model we are able to achieve similar constant factor approximation guarantees on the triangular lattice as were achieved in the standard HP model. While the structures derived from our folding rules are probably still far from biological reality, we hope that having a set of folding rules with different properties will yield more interesting folds when combined. %B Journal of Computational Biology %V 4 %P 275 - 296 %8 1997/01// %@ 1066-5277, 1557-8666 %G eng %U http://www.liebertonline.com/doi/abs/10.1089/cmb.1997.4.275 %N 3 %R 10.1089/cmb.1997.4.275 %0 Journal Article %J Proceedings of SPIE %D 1997 %T Locally adaptive document skew detection %A Sauvola,Jaakko J %A David Doermann %A Pietikaeinen,Matti %X This paper proposes a new approach to the detection of local orientation and skew in document images. It is based on the observation that there are many documents where a single global estimate of the page skew is not sufficient. These documents require local adaptation to deal robustly with todays complex configurations of components on the page. The approach attempts to identify regions in the image which exhibit locally consistent physical properties and consistent physical properties and consistent orientation. To do this, we rapidly compute a coarse segmentation and delineate regions which differ with respect to layout and/or physical content. Each region is classified as text, graphics, mixed text/graphics, image or background using local features and additional features are extracted to estimate orientation. The local orientation decisions are propagated where appropriate to resolve ambiguity and to produce a global estimate of the skew for the page. The implementation of our algorithms is demonstrated on a set of images which have multiple regions with different orientations. %B Proceedings of SPIE %V 3027 %P 96 - 108 %8 1997/04/03/ %@ 0277786X %G eng %U http://spiedigitallibrary.org/proceedings/resource/2/psisdg/3027/1/96_1?isAuthorized=no %N 1 %R doi:10.1117/12.270063 %0 Journal Article %J IEEE Software %D 1997 %T Low-effort, high-payoff user interface reengineering %A Plaisant, Catherine %A Rose,A. %A Shneiderman, Ben %A Vanniamparampil,A. J %K Business process re-engineering %K complete system reengineering %K Design methodology %K Error analysis %K Hardware %K inadequate functionalities %K interface stability %K iterative methods %K low-effort high-payoff user interface reengineering %K short-term improvements %K short-term user interface reengineering %K software engineering %K Software testing %K System analysis and design %K System testing %K systems re-engineering %K User centered design %K user centred design %K User interfaces %X Although increasingly sophisticated design methodologies for developing new user interfaces exist, low-effort, high-payoff user interface reengineering represents a new direction-and opportunity. Yet reengineering a working system is complex and risky because of the potential disruption to users and managers, their justifiable fear of change, and the lack of guarantees that such changes will be for the better. Our largely positive experiences with the projects described here lead us to believe that user interface reengineering is a viable and important process. Low effort, high-payoff improvement recommendations can probably be made for most existing systems. Nevertheless, a narrowly focused user interface reengineering plan may be inappropriate when the major problems lie outside the scope of the user interface, such as inadequate functionalities, frequent crashes, and network problems. Attempts at improving less severe problems while ignoring deeper ones may be perceived as insensitive by the users. In such cases it is important to consider either making similar short-term improvements for other parts of the systems or postponing short-term user interface reengineering in favour of a more complete system reengineering. Similarly, the need for interface stability might outweigh the benefits of the short-term improvements if a complete reengineering is planned for the near future. But most likely these proposed diagnostic strategies and opportunities for improvement are only a prelude to the much larger task of business reengineering, which implies extensive user interface reengineering %B IEEE Software %V 14 %P 66 - 72 %8 1997/08//Jul %@ 0740-7459 %G eng %N 4 %R 10.1109/52.595958 %0 Journal Article %J Communications in Statistics. Stochastic Models %D 1997 %T On markov chains with sluggish transients %A Stewart, G.W. %X In this note it is shown how to construct a Markov chain whose subdominant eigenvalue does not predict the decay of its transient %B Communications in Statistics. Stochastic Models %V 13 %P 85 - 94 %8 1997/// %@ 0882-0287 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/15326349708807414 %N 1 %R 10.1080/15326349708807414 %0 Conference Paper %B Proceedings of the eighteenth international conference on Information systems %D 1997 %T Mechanism design for intellectual property rights protection %A Giridharan,P. S. %A Srinivasan, Aravind %B Proceedings of the eighteenth international conference on Information systems %S ICIS '97 %I Association for Information Systems %C Atlanta, GA, USA %P 448– - 448– %8 1997/// %G eng %U http://dl.acm.org/citation.cfm?id=353071.353206 %0 Journal Article %J The Journal of Logic Programming %D 1997 %T Model generation and state generation for disjunctive logic programs %A Seipel,Dietmar %A Minker, Jack %A Ruiz,Carolina %X This paper investigates two fixpoint approaches for minimal model reasoning with disjunctive logic programs P. The first one, called model generation, is based on an operator TPINT defined on sets of Herbrand interpretations whose least fixpoint is logically equivalent to the set of minimal Herbrand models of the program. The second approach, called state generation, uses a fixpoint operation TPs based on hyperresolution. It operates on disjunctive Herbrand states, and its least fixpoint is the set of logical consequences of P, the so-called minimal model state of the program. We establish a useful relationship between hyperresolution by PPs and model generation by TPINT. Then we investigate the problem of continuity of the two operators TPs and TPINT. It is known that the operator TPs is continuous, and so it reaches its least fixpoint in at most ω iterations. On the other hand, the question of whether TPINT is continuous has been open. We show by a counterexample that TPINT is not continuous. Nevertheless, we prove that it converges towards its least fixpoint in at most ω iterations, too, as follows from the relationship that we show exists between hyperresolution and model generation. We define an iterative version of TPINT that computes the perfect model semantics of stratified disjunctive logic programs. On each stratum of the program, this operator converges in at most ω iterations. Model generations for the stable semantics and the partial stable semantics are respectively achieved by using this iterative operator together with the evidential transformation and the 3-S transformation. %B The Journal of Logic Programming %V 32 %P 49 - 69 %8 1997/07// %@ 0743-1066 %G eng %U http://www.sciencedirect.com/science/article/pii/S0743106696001161 %N 1 %R 10.1016/S0743-1066(96)00116-1 %0 Journal Article %J SIGGRAPH 1997 Course Notes %D 1997 %T Multiresolution surface modeling %A Heckbert,P. %A Rossignac,J. %A Hoppe,H. %A Schroeder,W. %A Soucy,M. %A Varshney, Amitabh %B SIGGRAPH 1997 Course Notes %8 1997/// %G eng %0 Journal Article %J Displays %D 1997 %T The next generation of graphical user interfaces: information visualization and better window management %A Shneiderman, Ben %B Displays %V 17 %P 125 - 129 %8 1997/05/01/ %@ 0141-9382 %G eng %U http://www.sciencedirect.com/science/article/pii/S014193829700005X %N 3–4 %R 10.1016/S0141-9382(97)00005-X %0 Journal Article %J Signal Processing, IEEE Transactions on %D 1997 %T Optimizing synchronization in multiprocessor DSP systems %A Bhattacharyya, Shuvra S. %A Sriram,S. %A Lee,E. A %B Signal Processing, IEEE Transactions on %V 45 %P 1605 - 1618 %8 1997/// %G eng %N 6 %0 Journal Article %J SIAM Journal on Matrix Analysis and Applications %D 1997 %T Perturbation analysis for the QR decomposition %A Chang,X. C. %A Paige,C. C. %A Stewart, G.W. %B SIAM Journal on Matrix Analysis and Applications %V 18 %P 775 - 791 %8 1997/// %G eng %0 Journal Article %J IMA Journal of Numerical AnalysisIMA J Numer Anal %D 1997 %T On the Perturbation of LU and Cholesky Factors %A Stewart, G.W. %X In a recent paper, Chang and Paige have shown that the usual perturbation bounds for Cholesky factors can systematically overestimate the errors. In this note we sharpen their results and extend them to the factors of the LU decomposition. The results are based on a new formula for the first-order terms of the error in the factors. %B IMA Journal of Numerical AnalysisIMA J Numer Anal %V 17 %P 1 - 6 %8 1997/01/01/ %@ 0272-4979, 1464-3642 %G eng %U http://imajna.oxfordjournals.org/content/17/1/1 %N 1 %R 10.1093/imanum/17.1.1 %0 Conference Paper %B Proceedings of the eighth annual ACM-SIAM symposium on Discrete algorithms %D 1997 %T A practical approximation algorithm for the LMS line estimator %A Mount, Dave %A Netanyahu,Nathan S. %A Romanik,Kathleen %A Silverman,Ruth %A Wu,Angela Y. %K Approximation algorithms %K least median-of-squares regression %K line arrangements %K line fitting %K randomized algorithms %K robust estimation %B Proceedings of the eighth annual ACM-SIAM symposium on Discrete algorithms %S SODA '97 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 473 - 482 %8 1997/// %@ 0-89871-390-0 %G eng %U http://dl.acm.org/citation.cfm?id=314161.314349 %0 Conference Paper %D 1997 %T Probabilistic verification of a synchronous round-based consensus protocol %A Duggal,H.S. %A Michel Cukier %A Sanders,W. H. %K business-critical applications %K consensus protocol correctness %K crash failures %K formal verification %K network environment %K probabilistic verification %K probabilities %K probability %K program verification %K proper behavior %K protocol behavior %K Protocols %K realistic environment %K reliable distributed systems %K safety-critical applications %K simple consensus protocol %K software reliability %K stochastic assumptions %K synchronous round based consensus protocol %K synchronous round based consensus protocols %K traditional proof techniques %X Consensus protocols are used in a variety of reliable distributed systems, including both safety-critical and business-critical applications. The correctness of a consensus protocol is usually shown, by making assumptions about the environment in which it executes, and then proving properties about the protocol. But proofs about a protocol's behavior are only as good as the assumptions which were made to obtain them, and violation of these assumptions can lead to unpredicted and serious consequences. We present a new approach for the probabilistic verification of synchronous round based consensus protocols. In doing so, we make stochastic assumptions about the environment in which a protocol operates, and derive probabilities of proper and non proper behavior. We thus can account for the violation of assumptions made in traditional proof techniques. To obtain the desired probabilities, the approach enumerates possible states that can be reached during an execution of the protocol, and computes the probability of achieving the desired properties for a given fault and network environment. We illustrate the use of this approach via the evaluation of a simple consensus protocol operating under a realistic environment which includes performance, omission, and crash failures %P 165 - 174 %8 1997/10// %G eng %R 10.1109/RELDIS.1997.632812 %0 Journal Article %J SIGMOD Record %D 1997 %T Query previews for networked information systems: a case study with NASA environmental data %A Doan,K. %A Plaisant, Catherine %A Shneiderman, Ben %A Bruns,T. %X Formulating queries on networked information systemsis laden with problems: data diversity, data complexity, network growth, varied user base, and slow network access. This paper proposes a new approach to a network query user interface which consists of two phases: query preview and query refinement. This new approach is based on dynamic queries and tight coupling, guiding users to rapidly and dynamically eliminate undesired items, reduce the data volume to a manageable size, and refine queries locally before submission over a network. A two-phase dynamic query system for NASA’s Earth Observing Systems--Data Information Systems (EOSDIS) is presented. The prototype was well received by the team of scientists who evaluated the interface. %B SIGMOD Record %V 26 %P 75 - 81 %8 1997/// %G eng %N 1 %0 Conference Paper %B CHI '97 extended abstracts on Human factors in computing systems: looking to the future %D 1997 %T Query previews in networked information systems: the case of EOSDIS %A Plaisant, Catherine %A Bruns,Tom %A Shneiderman, Ben %A Doan,Khoa %K direct manipulation %K dynamic query %K earth science %K network information system %K query preview %K Visualization %X Dynamic queries have been shown to be an effective technique to browse information, and to find patterns and exceptions. Dynamic queries involve the interactive control by a user of visual query parameters that generate rapid (100 ms update), animated, and visual displays of database search results. The data of early implementations was stored in local memory to guarantee optimal speed. Problems arise when the data is very large and distributed over a network. To overcome the problems of slow networks and data volume we propose a two-phase approach to query formulation using query previews and query refinements [1]. Preview mechanisms have been used in the past [2] and we believe that their use will be a major component of successful networked information systems interfaces (e.g. [3]). %B CHI '97 extended abstracts on Human factors in computing systems: looking to the future %S CHI EA '97 %I ACM %C New York, NY, USA %P 202 - 203 %8 1997/// %@ 0-89791-926-2 %G eng %U http://doi.acm.org/10.1145/1120212.1120343 %R 10.1145/1120212.1120343 %0 Journal Article %J SIAM Journal on Computing %D 1997 %T Randomized Distributed Edge Coloring via an Extension of the Chernoff--Hoeffding Bounds %A Panconesi,Alessandro %A Srinivasan, Aravind %X Certain types of routing, scheduling, and resource-allocation problems in a distributed setting can be modeled as edge-coloring problems. We present fast and simple randomized algorithms for edge coloring a graph in the synchronous distributed point-to-point model of computation. Our algorithms compute an edge coloring of a graph $G$ with $n$ nodes and maximum degree $\Delta$ with at most $1.6 \Delta + O(\log^{1+ \delta} n)$ colors with high probability (arbitrarily close to 1) for any fixed $\delta > 0$; they run in polylogarithmic time. The upper bound on the number of colors improves upon the $(2 \Delta - 1)$-coloring achievable by a simple reduction to vertex coloring.To analyze the performance of our algorithms, we introduce new techniques for proving upper bounds on the tail probabilities of certain random variables. The Chernoff--Hoeffding bounds are fundamental tools that are used very frequently in estimating tail probabilities. However, they assume stochastic independence among certain random variables, which may not always hold. Our results extend the Chernoff--Hoeffding bounds to certain types of random variables which are not stochastically independent. We believe that these results are of independent interest and merit further study. %B SIAM Journal on Computing %V 26 %P 350 - 350 %8 1997/// %@ 00975397 %G eng %U http://link.aip.org/link/SMJCAT/v26/i2/p350/s1&Agg=doi %N 2 %R 10.1137/S0097539793250767 %0 Journal Article %J Scientific Programming %D 1997 %T Run-time and compiler support for programming in adaptive parallel environments %A Edjlali,G. %A Agrawal,G. %A Sussman, Alan %A Humphries,J. %A Saltz, J. %B Scientific Programming %V 6 %P 215 - 227 %8 1997/// %G eng %N 2 %0 Report %D 1997 %T Scheduling and load-balancing via randomization %A Srinivasan, Aravind %I DSpace at School of Computing, NUS %V TR11/97 %8 1997/11/01/undef %G eng %0 Conference Paper %B Proceedings of the Eighth SIAM Conference on Parallel Processing for Scientific Computing. SIAM %D 1997 %T Scheduling in a high performance remote-sensing data server %A Chang,C. %A Sussman, Alan %A Saltz, J. %B Proceedings of the Eighth SIAM Conference on Parallel Processing for Scientific Computing. SIAM %8 1997/// %G eng %0 Conference Paper %B , 1997 IEEE Symposium on Security and Privacy, 1997. Proceedings %D 1997 %T A secure and reliable bootstrap architecture %A Arbaugh, William A. %A Farber,D. J %A Smith,J. M %K active networks %K AEGIS architecture %K bootstrap architecture %K Computer architecture %K computer bootstrapping %K data integrity %K Distributed computing %K Hardware %K hardware validity %K initialization %K integrity chain %K integrity check failures %K Internet %K Internet commerce %K IP networks %K Laboratories %K lower-layer integrity %K Microprogramming %K Operating systems %K recovery process %K reliability %K robust systems %K Robustness %K Security %K security of data %K software reliability %K system integrity guarantees %K system recovery %K transitions %K Virtual machining %X In a computer system, the integrity of lower layers is typically treated as axiomatic by higher layers. Under the presumption that the hardware comprising the machine (the lowest layer) is valid, the integrity of a layer can be guaranteed if and only if: (1) the integrity of the lower layers is checked and (2) transitions to higher layers occur only after integrity checks on them are complete. The resulting integrity “chain” inductively guarantees system integrity. When these conditions are not met, as they typically are not in the bootstrapping (initialization) of a computer system, no integrity guarantees can be made, yet these guarantees are increasingly important to diverse applications such as Internet commerce, security systems and “active networks”. In this paper, we describe the AEGIS architecture for initializing a computer system. It validates integrity at each layer transition in the bootstrap process. AEGIS also includes a recovery process for integrity check failures, and we show how this results in robust systems %B , 1997 IEEE Symposium on Security and Privacy, 1997. Proceedings %I IEEE %P 65 - 71 %8 1997/05/04/7 %@ 0-8186-7828-3 %G eng %R 10.1109/SECPRI.1997.601317 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1997 %T A Taxonomy of Multiple Window Coordination %A North,Chris %A Shneiderman, Ben %K Human-Computer Interaction, %X In current windowing environments, individual windows are treated independently, making it difficult for users to coordinate information across multiple windows. While coordinated multi- window strategies are increasingly used in visualization and web user interfaces, designs are inflexible and haphazard. The space of such linked-window strategies is not well understood and largely unexplored. This paper presents a taxonomy of coordinations, identifies important components, and reviews example interfaces. This 2x3 taxonomy provides guidelines for designers of applications, user interface toolkits, and window managers. We hope to encourage construction of generalized, end- user programmable, robust, multiple-window coordination capabilities. %B Institute for Systems Research Technical Reports %8 1997/// %G eng %U http://drum.lib.umd.edu/handle/1903/5892 %0 Conference Paper %B Proceedings of the First Brazilian Symposium on Document Image Analysis %D 1997 %T Techniques for Automated Testing for Automated Testing of Document Analysis Algorithms %A Sauvola,J. %A Kauniskangas,H. %A David Doermann %A Pietikainen,M. %B Proceedings of the First Brazilian Symposium on Document Image Analysis %P 201 - 212 %8 1997/// %G eng %0 Journal Article %J Computational Geometry %D 1997 %T Testing simple polygons %A Arkin,Esther M. %A Belleville,Patrice %A Mitchell,Joseph S.B. %A Mount, Dave %A Romanik,Kathleen %A Salzberg,Steven %A Souvaine,Diane %K probing %K Testing %K Verifying %X We consider the problem of verifying a simple polygon in the plane using “test points”. A test point is a geometric probe that takes as input a point in Euclidean space, and returns “+” if the point is inside the object being probed or “−” if it is outside. A verification procedure takes as input a description of a target object, including its location and orientation, and it produces a set of test points that are used to verify whether a test object matches the description. We give a procedure for verifying an n-sided, non-degenerate, simple target polygon using 5n test points. This testing strategy works even if the test polygon has n + 1 vertices, and we show a lower bound of 3n + 1 test points for this case. We also give algorithms using O(n) test points for simple polygons that may be degenerate and for test polygons that may have up to n + 2 vertices. All of these algorithms work for polygons with holes. We also discuss extensions of our results to higher dimensions. %B Computational Geometry %V 8 %P 97 - 114 %8 1997/07// %@ 0925-7721 %G eng %U http://www.sciencedirect.com/science/article/pii/S0925772196000156 %N 2 %R 10.1016/S0925-7721(96)00015-6 %0 Conference Paper %B In Proceedings of the 1997 International Conference on Data Engineering %D 1997 %T Titan: a High-Performance Remote-sensing Database %A Bongki,C. C %A Chang,C. %A Moon,B. %A Acharya, A. %A Shock,C. %A Sussman, Alan %A Saltz, J. %B In Proceedings of the 1997 International Conference on Data Engineering %8 1997/// %G eng %0 Journal Article %J IMA Journal of Numerical Analysis %D 1997 %T The Triangular Matrices of Gaussian Elimination and Related Decompositions %A Stewart, G.W. %X It has become commonplace that triangular systems are solved to higher accuracy than their condition would warrant. This observation is not true in general, and counterexamples are easy to construct. However, it is often true of the triangular matrices from pivoted LU or QR decompositions. It is shown that this fact is closely connected with the rank-revealing character of these decompositions. %B IMA Journal of Numerical Analysis %V 17 %P 7 - 16 %8 1997/01/01/ %@ 0272-4979, 1464-3642 %G eng %U http://imajna.oxfordjournals.org/content/17/1/7 %N 1 %R 10.1093/imanum/17.1.7 %0 Journal Article %J IEEE Software %D 1997 %T User Interface Reengineering: Low-Effort, High-Payoff Strategies %A Plaisant, Catherine %A Rose,A. %A Shneiderman, Ben %A Vanniamparampil,A. J %X User interface technology has advanced rapidly in recent years. Incorporating new developments in existing systems could result in substantial improvements in usability, thereby improving performance and user satisfaction, while shortening training and reducing error rates. We describe low-effort, high-payoff strategies that focus attention on improvements to data display and entry, consistency, messages, documentation, system access and additional functionality. We report on experience from six projects, describing observations, recommendations and outcomes. We close with guidance for managers and designers who are responsible for user interface reengineering. %B IEEE Software %V 14 %P 66 - 72 %8 1997/// %G eng %N 4 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1997 %T User Interfaces for a Complex Robotic Task: A Comparison of Tiles vs. Overlapped Windows %A Lane,J. Corde %A Kuester,Steven P %A Shneiderman, Ben %K Human computer interaction %K overlapped %K Systems Integration Methodology %K tiled %K user interface design %K window management %X High complexity tasks, such as remote teleoperation of robotic vehicles, often require multiple windows. For these complex tasks, the windows necessary for task completion, may occupy more area than available on a single visual display unit (VDU). Since the focus of the robotic task constantly changes, modular control panels that can be opened, closed, and moved on the screen are invaluable to the operator. This study describes a specific robotic task and the need for a multi window interface that can be easily manipulated. This paper examines two multi-window management strategies: tiled (fixed size) and arbitrary overlap. Multi-window searches were performed using the two management styles and they were compared on the basis of search completion time and error rates. Results with 35 novice users showed faster completion times for the tiled management strategy than for the arbitrary overlap strategy. Over factors such as the number of windows available, the number of displayed windows workload of opening or closing windows, and effect of learning are discussed. %B Institute for Systems Research Technical Reports %8 1997/// %G eng %U http://drum.lib.umd.edu/handle/1903/5874 %0 Conference Paper %B Proceedings of the AMIA Annual Fall Symposium %D 1997 %T The virtual microscope. %A Ferreira,R. %A Moon,B. %A Humphries,J. %A Sussman, Alan %A Saltz, J. %A Miller,R. %A Demarzo,A. %B Proceedings of the AMIA Annual Fall Symposium %P 449 - 449 %8 1997/// %G eng %0 Journal Article %J IEEE Transactions on Software Engineering %D 1997 %T Visual and textual consistency checking tools for graphical user interfaces %A Mahajan,R. %A Shneiderman, Ben %K button analysis tools %K Color %K dialog box summary table %K experiment %K graphical analysis tools %K Graphical user interfaces %K human factors %K inconsistent interface terminology %K interface color %K interface spell checker %K Output feedback %K Prototypes %K SHERLOCK %K Software architecture %K Software design %K software metrics %K software prototyping %K software tools %K Terminology %K terminology analysis tools %K Testing %K textual consistency checking tools %K user interface management systems %K User interfaces %K user performance %K visual consistency checking tools %X Designing user interfaces with consistent visual and textual properties is difficult. To demonstrate the harmful effects of inconsistency, we conducted an experiment with 60 subjects. Inconsistent interface terminology slowed user performance by 10 to 25 percent. Unfortunately, contemporary software tools provide only modest support for consistency control. Therefore, we developed SHERLOCK, a family of consistency analysis tools, which evaluates visual and textual properties of user interfaces. It provides graphical analysis tools such as a dialog box summary table that presents a compact overview of visual properties of all dialog boxes. SHERLOCK provides terminology analysis tools including an interface concordance, an interface spellchecker, and terminology baskets to check for inconsistent use of familiar groups of terms. Button analysis tools include a button concordance and a button layout table to detect variant capitalization, distinct typefaces, distinct colors, variant button sizes, and inconsistent button placements. We describe the design, software architecture, and the use of SHERLOCK. We tested SHERLOCK with four commercial prototypes. The outputs, analysis, and feedback from designers of the applications are presented %B IEEE Transactions on Software Engineering %V 23 %P 722 - 735 %8 1997/11// %@ 0098-5589 %G eng %N 11 %R 10.1109/32.637386 %0 Journal Article %J Information in Images %D 1997 %T Visual Information Seeking in Digital Image Libraries: The Visible Human Explorer %A North,C. %A Shneiderman, Ben %A Plaisant, Catherine %X This chapter presents the Visible Human Explorer user interface, developed at theHuman-Computer Interaction Lab of the University of Maryland at College Park, for remotely accessing the National Library of Medicine’s Visible Human digital image library. With the interface, users can visualize the library, browse contents, locate data of interest, and retrieve and zoom on desired image details. The interface presents a pair of tightly coupled views of library data: an overview of the overall search space, and a preview of high-resolution images available for retrieval. To explore, the user sweeps the views through the search space and receives smooth, rapid, visual feedback of contents. Desired details are automatically downloaded over the internet from the library. The interface software is completely functional (runs on Sun Workstations) and freely available for public use, at: http://www.nlm.nih.gov/. We also present several human-computer interaction design principles used to create the Visible Human Explorer interface, describe how these were applied to the design, and discuss issues in employing these techniques in user interfaces for other information collections. These principles are direct manipulation, visual information seeking, query previews, and multiple tightly coupled views. We illustrate these concepts with a plethora of pictures of user interface screens. Please also check the included CD-ROM for additional illustration media. %B Information in Images %8 1997/// %G eng %0 Journal Article %J BIT Numerical Mathematics %D 1997 %T On the weighting method for least squares problems with linear equality constraints %A Stewart, G.W. %X The weighting method for solving a least squares problem with linear equality constraints multiplies the constraints by a large number and appends them to the top of the least squares problem, which is then solved by standard techniques. In this paper we give a new analysis of the method, based on the QR decomposition, that exhibits many features of the algorithm. In particular it suggests a natural criterion for chosing the weighting factor. %B BIT Numerical Mathematics %V 37 %P 961 - 967 %8 1997/// %G eng %N 4 %R 10.1007/BF02510363 %0 Journal Article %J Human Factors in Web Development %D 1997 %T A zooming web browser %A Bederson, Benjamin B. %A Hollan,J.D. %A Stewart,J. %A Rogers,D. %A Druin, Allison %A Vick,D. %A Ring,L. %A Grose,E. %A Forsythe,C. %B Human Factors in Web Development %8 1997/// %G eng %0 Journal Article %J ACM Comput. Surv. %D 1996 %T Advanced graphic user interfaces: elastic and tightly coupled windows %A Shneiderman, Ben %B ACM Comput. Surv. %V 28 %8 1996/12// %@ 0360-0300 %G eng %U http://doi.acm.org/10.1145/242224.242409 %N 4es %R 10.1145/242224.242409 %0 Journal Article %J Journal of Algorithms %D 1996 %T On the Complexity of Distributed Network Decomposition %A Panconesi,Alessandro %A Srinivasan, Aravind %X In this paper, we improve the bounds for computing a network decomposition distributively and deterministically. Our algorithm computes an (nϵ(n),nϵ(n))-decomposition innO(ϵ(n))time, where[formula]. As a corollary we obtain improved deterministic bounds for distributively computing several graph structures such as maximal independent sets and Δ-vertex colorings. We also show that the class of graphs G whose maximum degree isnO(δ(n))where δ(n)=1/log lognis complete for the task of computing a near-optimal decomposition, i.e., a (logn, logn)-decomposition, in polylog(n) time. This is a corollary of a more general characterization, which pinpoints the weak points of existing network decomposition algorithms. Completeness is to be intended in the following sense: if we have an algorithmAthat computes a near-optimal decomposition in polylog(n) time for graphs inG, then we can compute a near-optimal decomposition in polylog(n) time for all graphs. %B Journal of Algorithms %V 20 %P 356 - 374 %8 1996/03// %@ 0196-6774 %G eng %U http://www.sciencedirect.com/science/article/pii/S0196677496900176 %N 2 %R 10.1006/jagm.1996.0017 %0 Conference Paper %B Computer Aided Verification %D 1996 %T The Concurrency Factory: A development environment for concurrent systems %A Cleaveland, Rance %A Lewis,P. %A Smolka,S. %A Sokolsky,O. %B Computer Aided Verification %P 398 - 401 %8 1996/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 1996 %T Designing Information-Abundant Websites %A Shneiderman, Ben %K human computer interaction, Systems Integration Methodology %X The deluge of web pages has generated dystopian commentaries on the tragedy of the flood as well as utopian visions of harnessing the same flood for constructive purposes. Within this ocean of information there are also lifeboat web pages with design principles, but often the style parallels the early user interface writings in the 1970s. The well-intentioned Noahs who write from personal experience as website designers, often draw their wisdom from specific projects, making their advice incomplete or lacking in generalizability. Their experience is valuable but the paucity of empirical data to validate or sharpen insight means that some guidelines are misleading. As scientific evidence accumulates, foundational cognitive and perceptual theories will structure the discussion and guide designers in novel situations. %B Institute for Systems Research Technical Reports %8 1996/// %G eng %U http://drum.lib.umd.edu/handle/1903/5759 %0 Book %D 1996 %T Designing Multimedia Environments for Children: Computers, Creativity, and Kids. %A Druin, Allison %A Solomon,C. %I Wiley Computer Publishing, John Wiley and Sons, Inc., One Wiley Drive, Somerset, NJ 08875 ($44.95; Canada, $62.95); phone (Eastern Division): 201-469-4400;(Western Division): 801-972-5828. %8 1996/// %G eng %0 Conference Paper %B Proceedings in the International Workshop on Document Analysis Systems %D 1996 %T The Development of a General Framework for Intelligent Document Image Retrieval %A David Doermann %A Sauvola,J. %A Kauniskangas,H. %A Shin,C. %A Pietikainen,M. %A Rosenfeld, A. %B Proceedings in the International Workshop on Document Analysis Systems %P 605 - 632 %8 1996/// %G eng %0 Journal Article %J Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on %D 1996 %T Early detection of independent motion from active control of normal image flow patterns %A Sharma, R. %A Aloimonos, J. %B Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on %V 26 %P 42 - 52 %8 1996/// %G eng %N 1 %0 Conference Paper %B Foundations of Computer Science, 1996. Proceedings., 37th Annual Symposium on %D 1996 %T Efficient approximate and dynamic matching of patterns using a labeling paradigm %A Sahinalp,S. C %A Vishkin, Uzi %K algorithm;replaced %K algorithmics;substrings;suffix %K algorithms;pattern %K approximate %K characters;dynamic %K characters;labeling %K characters;string %K complexity;indexing;parallel %K construction;computational %K data %K dictionary %K dynamic %K indexing;efficient %K matching;deleted %K matching;dynamic %K matching;efficient %K matching;inserted %K matching;string %K matching;tree %K paradigm;optimal %K Parallel %K pattern %K PROCESSING %K string %K structures; %K text %K tree %X A key approach in string processing algorithmics has been the labeling paradigm which is based on assigning labels to some of the substrings of a given string. If these labels are chosen consistently, they can enable fast comparisons of substrings. Until the first optimal parallel algorithm for suffix tree construction was given by the authors in 1994 the labeling paradigm was considered not to be competitive with other approaches. They show that this general method is also useful for several central problems in the area of string processing: approximate string matching, dynamic dictionary matching, and dynamic text indexing. The approximate string matching problem deals with finding all substrings of a text which match a pattern ldquo;approximately rdquo;, i.e., with at most m differences. The differences can be in the form of inserted, deleted, or replaced characters. The text indexing problem deals with finding all occurrences of a pattern in a text, after the text is preprocessed. In the dynamic text indexing problem, updates to the text in the form of insertions and deletions of substrings are permitted. The dictionary matching problem deals with finding all occurrences of each pattern set of a set of patterns in a text, after the pattern set is preprocessed. In the dynamic dictionary matching problem, insertions and deletions of patterns to the pattern set are permitted %B Foundations of Computer Science, 1996. Proceedings., 37th Annual Symposium on %P 320 - 328 %8 1996/10// %G eng %R 10.1109/SFCS.1996.548491 %0 Journal Article %J Languages and Compilers for Parallel Computing %D 1996 %T Efficient distribution analysis via graph contraction %A Sheffler,T. %A Schreiber,R. %A Pugh, William %A Gilbert,J. %A Chatterjee,S. %B Languages and Compilers for Parallel Computing %P 377 - 391 %8 1996/// %G eng %0 Conference Paper %B Proceedings of the workshop on Advanced visual interfaces %D 1996 %T Elastic windows: improved spatial layout and rapid multiple window operations %A Kandogan,Eser %A Shneiderman, Ben %K CAD %K elastic windows %K multi-window operations %K personal role manager %K programming environment %K task switching %K window manager %X Most windowing systems follow the independent overlapping windows approach, which emerged as an answer to the needs of the 80s' application and technology. Advances in computers, display technology, and the applications demand more functionality from window management systems. Based on these changes and the problems of current windowing appraoches, we have updated the requirements for multiwindow systems to guide new methods of window management. We propose elastic windows with improved spatial layout and rapid multi-window operations. Multi-window operations are achieved by issuing operations on window groups hierachically organized in a space-filling tiled layout. Sophisticated multi-window operations and spatial layout dynamics helps users to handle fast task-switching and to structure thier work environment to their rapidly changing needs. We claim that these multi-window operations and the improved spatial layout decrease the cognitive load on users. Users found our prototype system to be comprehensible and enjoyable as they playfully explored the way multiple windows are reshaped. %B Proceedings of the workshop on Advanced visual interfaces %S AVI '96 %I ACM %C New York, NY, USA %P 29 - 38 %8 1996/// %@ 0-89791-834-7 %G eng %U http://doi.acm.org/10.1145/948449.948454 %R 10.1145/948449.948454 %0 Journal Article %J Empirical Software Engineering %D 1996 %T The empirical investigation of Perspective-Based Reading %A Basili, Victor R. %A Green,Scott %A Laitenberger,Oliver %A Lanubile,Filippo %A Shull, Forrest %A Sørumgård,Sivert %A Zelkowitz, Marvin V %K Computer %K Science %X We consider reading techniques a fundamental means of achieving high quality software. Due to the lack of research in this area, we are experimenting with the application and comparison of various reading techniques. This paper deals with our experiences with a family of reading techniques known as Perspective-Based Reading (PBR), and its application to requirements documents. The goal of PBR is to provide operational scenarios where members of a review team read a document from a particular perspective, e.g., tester, developer, user. Our assumption is that the combination of different perspectives provides better coverage of the document, i.e., uncovers a wider range of defects, than the same number of readers using their usual technique. %B Empirical Software Engineering %V 1 %P 133 - 164 %8 1996/// %@ 1382-3256 %G eng %U http://dx.doi.org/10.1007/BF00368702 %N 2 %0 Conference Paper %B Proceedings of the seventh annual ACM-SIAM symposium on Discrete algorithms %D 1996 %T An extension of the Lovász Local Lemma, and its applications to integer programming %A Srinivasan, Aravind %B Proceedings of the seventh annual ACM-SIAM symposium on Discrete algorithms %S SODA '96 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 6 - 15 %8 1996/// %@ 0-89871-366-8 %G eng %U http://dl.acm.org/citation.cfm?id=313852.313866 %0 Conference Paper %B , IEEE Symposium on Visual Languages, 1996. Proceedings %D 1996 %T The eyes have it: a task by data type taxonomy for information visualizations %A Shneiderman, Ben %K advanced graphical user interface design %K Art %K data mining %K data type taxonomy %K data visualisation %K Data visualization %K Displays %K Eyes %K Graphical user interfaces %K Information filtering %K Information filters %K information visualizations %K multi dimensional data %K Multimedia databases %K network data %K Taxonomy %K visual databases %K visual information seeking %K visual programming %X A useful starting point for designing advanced graphical user interfaces is the visual information seeking Mantra: overview first, zoom and filter, then details on demand. But this is only a starting point in trying to understand the rich and varied set of information visualizations that have been proposed in recent years. The paper offers a task by data type taxonomy with seven data types (one, two, three dimensional data, temporal and multi dimensional data, and tree and network data) and seven tasks (overview, zoom, filter, details-on-demand, relate, history, and extracts) %B , IEEE Symposium on Visual Languages, 1996. Proceedings %I IEEE %P 336 - 343 %8 1996/09/03/6 %@ 0-8186-7508-X %G eng %R 10.1109/VL.1996.545307 %0 Journal Article %J SIAM Journal on Scientific Computing %D 1996 %T Fast nonsymmetric iterations and preconditioning for Navier-Stokes equations %A Elman, Howard %A Silvester, D. %B SIAM Journal on Scientific Computing %V 17 %P 33 - 46 %8 1996/// %G eng %N 1 %0 Journal Article %J International Journal of Computational Geometry and Applications %D 1996 %T A fast parallel algorithm for finding the convex hull of a sorted point set %A Berkman,O. %A Schieber,B. %A Vishkin, Uzi %B International Journal of Computational Geometry and Applications %V 6 %P 231 - 242 %8 1996/// %G eng %0 Conference Paper %B Euro-Par'96 Parallel Processing %D 1996 %T A high performance image database system for remotely sensed imagery %A Shock,C. %A Chang,C. %A Davis, Larry S. %A Goward,S. %A Saltz, J. %A Sussman, Alan %B Euro-Par'96 Parallel Processing %P 109 - 122 %8 1996/// %G eng %0 Journal Article %J 16th AIAA International Communications Satellite Systems Conference %D 1996 %T Hybrid network management (communication systems) %A Baras,J. S %A Ball,M. %A Karne,R. K %A Kelley,S. %A Jang,K.D. %A Plaisant, Catherine %A Roussopoulos, Nick %A Stathatos,K. %A Vakhutinsky,A. %A Valluri,J. %X We describe our collaborative efforts towards the design and implementation of a next-generation integrated network management system for hybrid networks (INMS/HN). We describe the overall software architecture of the system at its current stage of development. This NMS is specifically designed to address issues relevant to complex heterogeneous networks consisting of seamlessly interoperable terrestrial and satellite networks. NMSs are a key element for interoperability in such networks. We describe the integration of configuration management and performance management. The next step in this integration is fault management. In particular, we describe the object model, issues concerning the graphical user interface, browsing tools, performance data graphical widget displays, and management information database organization issues. %B 16th AIAA International Communications Satellite Systems Conference %8 1996/// %G eng %0 Journal Article %J Journal of Molecular Biology %D 1996 %T Identification of the Binding Site for Acidic Phospholipids on the PH Domain of Dynamin: Implications for Stimulation of GTPase Activity %A Zheng,Jie %A Cahill,Sean M. %A Lemmon,Mark A. %A Fushman, David %A Schlessinger,Joseph %A Cowburn,David %K Fluorescence %K ligand binding %K NMR %K PH domain %K phospholipid %X It has recently been suggested that pleckstrin homology (PH) domains bind specifically to phospholipids, with phosphatidylinositol-4,5-bisphosphate (PtdIns(4,5)P2) being most strongly bound. This observation suggests that PH domains may be responsible for membrane association of proteins in which they occur. Further, this membrane association may be regulated by enzymes that modify lipid head groups to which PH domains may bind. We have studied the binding of phospholipids to the PH domain of human dynamin, a 100 kDa GTPase that is involved in the initial stages of endocytosis. We describe a rapid method for screening PH domain/ligand interactions that gives precise binding constants. We confirm that PtdIns(4,5)P2can bind to dynamin PH domain, although not in an aggregated state. Using NMR spectroscopy, we have mapped a specific site on the surface of dynamin PH domain of which binding of gIns(1,4,5)P3(the head-group skeleton of PtdIns(4,5)P2) occurs. The relative affinity of acidic phospholipids for dynamin PH domain correlates with their ability to activate the GTPase of dynamin. We propose, therefore, that the interaction of these phospholipids with dynamin is likely to occurviathe PH domain. Given the fact that PH domains are often found in pro- teins associated with GTPase activity, or in guanine nucleotide exchange factors, we suggest that one role of PH domains may be to couple phosphatidylinositol signalling to GTP hydrolysis. %B Journal of Molecular Biology %V 255 %P 14 - 21 %8 1996/01/12/ %@ 0022-2836 %G eng %U http://www.sciencedirect.com/science/article/pii/S0022283696900029 %N 1 %R 10.1006/jmbi.1996.0002 %0 Journal Article %J ACM SIGMOD Record %D 1996 %T Incremental data structures and algorithms for dynamic query interfaces %A Tanin,Egemen %A Beigel,Richard %A Shneiderman, Ben %K algorithm %K data structure %K database %K direct manipulation %K dynamic query %K Information Visualization %K user interface %X Dynamic query interfaces (DQIs) form a recently developed method of database access that provides continuous realtime feedback to the user during the query formulation process. Previous work shows that DQIs are elegant and powerful interfaces to small databases. Unfortunately, when applied to large databases, previous DQI algorithms slow to a crawl. We present a new approach to DQI algorithms that works well with large databases. %B ACM SIGMOD Record %V 25 %P 21 - 24 %8 1996/12// %@ 0163-5808 %G eng %U http://doi.acm.org/10.1145/245882.245891 %N 4 %R 10.1145/245882.245891 %0 Journal Article %J AIP Conference Proceedings %D 1996 %T Integrated network management of hybrid networks %A Baras,John S %A Ball,Mike %A Karne,Ramesh K %A Kelley,Steve %A Jang,Kap D %A Plaisant, Catherine %A Roussopoulos, Nick %A Stathatos,Kostas %A Vakhutinsky,Andrew %A Jaibharat,Valluri %A Whitefield,David %X We describe our collaborative efforts towards the design and implementation of a next generation integrated network management system for hybrid networks (INMS/HN). We describe the overall software architecture of the system at its current stage of development. This network management system is specifically designed to address issues relevant for complex heterogeneous networks consisting of seamlessly interoperable terrestrial and satellite networks. Network management systems are a key element for interoperability in such networks. We describe the integration of configuration management and performance management. The next step in this integration is fault management. In particular we describe the object model, issues of the Graphical User Interface (GUI), browsing tools and performance data graphical widget displays, management information database (MIB) organization issues. Several components of the system are being commercialized by Hughes Network Systems. © 1996 American Institute of Physics. %B AIP Conference Proceedings %V 361 %P 345 - 350 %8 1996/03/01/ %@ 0094243X %G eng %U http://proceedings.aip.org/resource/2/apcpcs/361/1/345_1?isAuthorized=no %N 1 %R doi:10.1063/1.50028 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: common ground %D 1996 %T LifeLines: visualizing personal histories %A Plaisant, Catherine %A Milash,Brett %A Rose,Anne %A Widoff,Seth %A Shneiderman, Ben %K History %K justice %K medical record %K overview %K personal record %K screen design %K screen management %K timeline %K Visualization %B Proceedings of the SIGCHI conference on Human factors in computing systems: common ground %S CHI '96 %I ACM %C New York, NY, USA %P 221 - 227 %8 1996/// %@ 0-89791-777-4 %G eng %U http://doi.acm.org/10.1145/238386.238493 %R 10.1145/238386.238493 %0 Conference Paper %B Proceedings of the 9th annual ACM symposium on User interface software and technology %D 1996 %T Local tools: an alternative to tool palettes %A Bederson, Benjamin B. %A Hollan,J.D. %A Druin, Allison %A Stewart,J. %A Rogers,D. %A Proft,D. %B Proceedings of the 9th annual ACM symposium on User interface software and technology %P 169 - 170 %8 1996/// %G eng %0 Journal Article %J Software - Concepts and Tools %D 1996 %T Modeling and verifying distributed systems using priorities: A case study %A Cleaveland, Rance %A Natarajan,V. %A Sims,S. %A Luettgen,G. %B Software - Concepts and Tools %V 17 %P 50 - 62 %8 1996/// %G eng %N 2 %0 Conference Paper %B Computer Aided Verification %D 1996 %T The NCSU concurrency workbench %A Cleaveland, Rance %A Sims,S. %B Computer Aided Verification %P 394 - 397 %8 1996/// %G eng %0 Conference Paper %B Visualization '96. Proceedings. %D 1996 %T Optimizing triangle strips for fast rendering %A Evans,F. %A Skiena,S. %A Varshney, Amitabh %K buffer %K data;triangulated %K disciplines;rendering %K model %K models;polygonal %K optimisation;triangulated %K partitioning;queuing %K rendering;graphics %K sizes;fast %K strip %K subsystem;interactive %K surfaces;data %K times;scientific %K triangulated %K visualisation; %K visualization;partially %K visualization;triangle %X Almost all scientific visualization involving surfaces is currently done via triangles. The speed at which such triangulated surfaces can be displayed is crucial to interactive visualization and is bounded by the rate at which triangulated data can be sent to the graphics subsystem for rendering. Partitioning polygonal models into triangle strips can significantly reduce rendering times over transmitting each triangle individually. We present new and efficient algorithms for constructing triangle strips from partially triangulated models, and experimental results showing these strips are on average 15% better than those from previous codes. Further, we study the impact of larger buffer sizes and various queuing disciplines on the effectiveness of triangle strips. %B Visualization '96. Proceedings. %P 319 - 326 %8 1996/11/27/1 %G eng %R 10.1109/VISUAL.1996.568125 %0 Journal Article %J Computer applications in the biosciences : CABIOS %D 1996 %T Positional sequencing by hybridization %A Hannenhalli, Sridhar %A Feldman,William %A Lewis,Herbert F. %A Skiena,Steven S. %A Pevzner,Pavel A. %X Sequencing by hybridization (SBH) is a promising alternative to the classical DNA sequencing approaches. However, the resolving power of SBH is rather low: with 64kb sequencing chips, unknown DNA fragments only as long as 200 bp can be reconstructed in a single SBH experiment. To improve the resolving power of SBH, positional SBH (PSBH) has recently been suggested; this allows (with additional experimental work) approximate positions of every l-tuple in a target DNA fragment to be measured. We study the positional Eulerian path problem motivated by PSBH. The input to the positional eulerian path problem is an Eulerian graph G( V, E) in which every edge has an associated range of integers and the problem is to find an Eulerian path el, …, e|E| in G such that the range of ei, contains i. We show that the positional Eulerian path problem is NP-complete even when the maximum out-degree (in-degree) of any vertex in the graph is 2. On a positive note we present polynomial algorithms to solve a special case of PSBH (bounded PSBH), where the range of the allowed positions for any edge is bounded by a constant (it corresponds to accurate experimental measurements of positions in PSBH). Moreover, if the positions of every l-tuple in an unknown DNA fragment of length n are measured with O(log n) error, then our algorithm runs in polynomial time. We also present an estimate of the resolving power of PSBH for a more realistic case when positions are measured with Θ(n) error. %B Computer applications in the biosciences : CABIOS %V 12 %P 19 - 24 %8 1996/02/01/ %G eng %U http://bioinformatics.oxfordjournals.org/content/12/1/19.abstract %N 1 %R 10.1093/bioinformatics/12.1.19 %0 Journal Article %J Tools and Algorithms for the Construction and Analysis of Systems %D 1996 %T Priorities for modeling and verifying distributed systems %A Cleaveland, Rance %A L\üttgen,G. %A Natarajan,V. %A Sims,S. %B Tools and Algorithms for the Construction and Analysis of Systems %P 278 - 297 %8 1996/// %G eng %0 Conference Paper %B Proceedings of the 20th workshop of the Austrian Association for Pattern Recognition (OAGM/AAPR) on Pattern recognition 1996 %D 1996 %T Qualitative depth and FOE information via smoothed flow %A Shulman,David %A Fermüller, Cornelia %B Proceedings of the 20th workshop of the Austrian Association for Pattern Recognition (OAGM/AAPR) on Pattern recognition 1996 %I R. Oldenbourg Verlag GmbH %C Munich, Germany, Germany %P 207 - 213 %8 1996/// %@ 3-486-23865-5 %G eng %U http://dl.acm.org/citation.cfm?id=258081.258100 %0 Conference Paper %B Proceedings of the Third Forum on Research and Technology Advances in Digital Libraries, 1996. ADL '96 %D 1996 %T Query previews in networked information systems %A Donn,K. %A Plaisant, Catherine %A Shneiderman, Ben %K Computer networks %K data complexity %K data mining %K data patterns %K data volume %K dynamic query user interfaces %K Educational institutions %K EOS-DIS %K exploratory method %K Information retrieval %K Information services %K Information systems %K Intelligent networks %K interactive systems %K Laboratories %K Manipulator dynamics %K matching data sets %K NASA Earth Observing System-Data Information System %K Network performance %K networked environment %K networked information systems %K query formulation %K query preview %K query refinement %K querying process %K rough attribute values %K User interfaces %K visual databases %K zero hit queries %X In a networked information system (such as the NASA Earth Observing System-Data Information System (EOS-DIS)), there are three major obstacles facing users in a querying process: network performance, data volume and data complexity. In order to overcome these obstacles, we propose a two phase approach to query formulation. The two phases are the Query Preview and the Query Refinement. In the Query Preview phase, users formulate an initial query by selecting rough attribute values. The estimated number of matching data sets is shown, graphically on preview bars which allows users to rapidly focus on a manageable number of relevant data sets. Query previews also prevent wasted steps by eliminating zero hit queries. When the estimated number of data sets is long enough, the initial query is submitted to the network which returns the metadata of the data sets for further refinement in the Query Refinement phase. The two phase approach to query formulation overcomes slow network performance, and reduces the data volume and data complexity, problems. This approach is especially appropriate for users who do not have extensive knowledge about the data and who prefer an exploratory method to discover data patterns and exceptions. Using this approach, we have developed dynamic query user interfaces to allow users to formulate their queries across a networked environment %B Proceedings of the Third Forum on Research and Technology Advances in Digital Libraries, 1996. ADL '96 %I IEEE %P 120 - 129 %8 1996/05/13/15 %@ 0-8186-7403-2 %G eng %R 10.1109/ADL.1996.502522 %0 Conference Paper %B Research and Technology Advances in Digital Libraries, 1996. ADL '96., Proceedings of the Third Forum on %D 1996 %T Query previews in networked information systems %A Donn,K. %A Plaisant, Catherine %A Shneiderman, Ben %K attribute %K complexity;data %K data %K Earth %K environment;networked %K EOS-DIS;NASA %K formulation;querying %K formulation;user %K hit %K information %K interfaces; %K interfaces;exploratory %K method;matching %K networks;information %K Observing %K patterns;data %K performance;networked %K Preview;Query %K process;rough %K queries;computer %K query %K refinement;data %K retrieval;information %K services;interactive %K sets;network %K System-Data %K System;Query %K systems;query %K user %K values;zero %K volume;dynamic %X In a networked information system (such as the NASA Earth Observing System-Data Information System (EOS-DIS)), there are three major obstacles facing users in a querying process: network performance, data volume and data complexity. In order to overcome these obstacles, we propose a two phase approach to query formulation. The two phases are the Query Preview and the Query Refinement. In the Query Preview phase, users formulate an initial query by selecting rough attribute values. The estimated number of matching data sets is shown, graphically on preview bars which allows users to rapidly focus on a manageable number of relevant data sets. Query previews also prevent wasted steps by eliminating zero hit queries. When the estimated number of data sets is long enough, the initial query is submitted to the network which returns the metadata of the data sets for further refinement in the Query Refinement phase. The two phase approach to query formulation overcomes slow network performance, and reduces the data volume and data complexity, problems. This approach is especially appropriate for users who do not have extensive knowledge about the data and who prefer an exploratory method to discover data patterns and exceptions. Using this approach, we have developed dynamic query user interfaces to allow users to formulate their queries across a networked environment %B Research and Technology Advances in Digital Libraries, 1996. ADL '96., Proceedings of the Third Forum on %P 120 - 129 %8 1996/05// %G eng %R 10.1109/ADL.1996.502522 %0 Journal Article %J Mathematics of Computation %D 1996 %T Rounding errors in solving block Hessenberg systems %A Von Matt,Urs %A Stewart, G.W. %K block diagonally dominant matrices %K block Hessenberg matrices %K Linear systems %K M-matrices %K rounding error analysis %X A rounding error analysis is presented for a divide-and-conquer algorithm to solve linear systems with block Hessenberg matrices. Conditions are derived under which the algorithm computes a stable solution. The algorithm is shown to be stable for block diagonally dominant matrices and for M-matrices. %B Mathematics of Computation %V 65 %P 115 - 135 %8 1996/01// %@ 0025-5718 %G eng %U http://dx.doi.org/10.1090/S0025-5718-96-00667-9 %N 213 %R 10.1090/S0025-5718-96-00667-9 %0 Conference Paper %B Proceedings of the 10th international conference on Supercomputing %D 1996 %T Runtime coupling of data-parallel programs %A Ranganathan,M. %A Acharya, A. %A Edjlali,G. %A Sussman, Alan %A Saltz, J. %B Proceedings of the 10th international conference on Supercomputing %P 229 - 236 %8 1996/// %G eng %0 Conference Paper %B Proceedings of the symposium on Computers and the quality of life %D 1996 %T Social impact statements: engaging public participation in information technology design %A Shneiderman, Ben %A Rose,Anne %B Proceedings of the symposium on Computers and the quality of life %S CQL '96 %I ACM %C New York, NY, USA %P 90 - 96 %8 1996/// %@ 0-89791-827-4 %G eng %U http://doi.acm.org/10.1145/238339.238378 %R 10.1145/238339.238378 %0 Book Section %B Dependable Computing — EDCC-2 %D 1996 %T On stratified sampling for high coverage estimations %A Powell,David %A Michel Cukier %A Arlat,Jean %E Hlawiczka,Andrzej %E Silva,João %E Simoncini,Luca %K Computer science %X This paper addresses the problem of estimating the coverage of a fault tolerance mechanism through statistical processing of observations collected in faultinjection experiments. In an earlier paper, several techniques for sampling the fault/activity input space of a fault tolerance mechanism were presented. Various estimators based on simple sampling in the whole space and stratified sampling in a partitioned space were studied; confidence limits were derived based on a normal approximation. In this paper, the validity of this approximation is analyzed, especially for high coverage systems. The theory of confidence regions is then introduced to estimate the coverage without approximation when, for practical reasons, stratification is used. Three statistics are considered for defining confidence regions. It is shown that one of these statistics — a vectorial statistic — is often more conservative than the other two. However, only the vectorial statistic is computationally tractable. The results obtained are compared with those based on approximation by means of three hypothetical example systems. %B Dependable Computing — EDCC-2 %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 1150 %P 35 - 54 %8 1996/// %@ 978-3-540-61772-3 %G eng %U http://www.springerlink.com/content/7t2w2u472601h730/abstract/ %0 Conference Paper %B ACM SIGGRAPH 96 Visual Proceedings: The art and interdisciplinary programs of SIGGRAPH '96 %D 1996 %T Stripe: a software tool for efficient triangle strips %A Evans,Francine %A Skiena,Steven %A Varshney, Amitabh %B ACM SIGGRAPH 96 Visual Proceedings: The art and interdisciplinary programs of SIGGRAPH '96 %S SIGGRAPH '96 %I ACM %C New York, NY, USA %P 153– - 153– %8 1996/// %@ 0-89791-784-7 %G eng %U http://doi.acm.org/10.1145/253607.253894 %R 10.1145/253607.253894 %0 Journal Article %J Getty Art History Information Program (1996) Research Agenda for Networked Cultural Heritage, Santa Monica. CA: Getty AHIP %D 1996 %T Tools for creating and exploiting content %A Kolker,R. %A Shneiderman, Ben %B Getty Art History Information Program (1996) Research Agenda for Networked Cultural Heritage, Santa Monica. CA: Getty AHIP %P 27 - 30 %8 1996/// %G eng %0 Journal Article %J Languages and Compilers for Parallel Computing %D 1996 %T Transitive closure of infinite graphs and its applications %A Kelly,W. %A Pugh, William %A Rosser,E. %A Shpeisman,T. %B Languages and Compilers for Parallel Computing %P 126 - 140 %8 1996/// %G eng %0 Conference Paper %B Proceedings of the fourth workshop on I/O in parallel and distributed systems: part of the federated computing research conference %D 1996 %T Tuning the performance of I/O-intensive parallel applications %A Acharya, A. %A Uysal, M. %A Bennett, R. %A Mendelson, A. %A Beynon, M. %A Hollingsworth, Jeffrey K %A Saltz, J. %A Sussman, Alan %B Proceedings of the fourth workshop on I/O in parallel and distributed systems: part of the federated computing research conference %P 15 - 27 %8 1996/// %G eng %0 Conference Paper %B Proceedings of the first ACM international conference on Digital libraries %D 1996 %T User controlled overviews of an image library: a case study of the visible human %A North,Chris %A Shneiderman, Ben %A Plaisant, Catherine %K Browsing %K digital library %K image database %K information exploration %K Information retrieval %K Internet %K medical image %K remote access %K user interface %K Visualization %K World-wide web %B Proceedings of the first ACM international conference on Digital libraries %S DL '96 %I ACM %C New York, NY, USA %P 74 - 82 %8 1996/// %@ 0-89791-830-4 %G eng %U http://doi.acm.org/10.1145/226931.226946 %R 10.1145/226931.226946 %0 Conference Paper %B Proceedings of the the seventh ACM conference on Hypertext %D 1996 %T Visual metaphor and the problem of complexity in the design of Web sites: techniques for generating, recognizing and visualizing structure %A Joyce,Michael %A Kolker,Robert %A Moulthrop,Stuart %A Shneiderman, Ben %A Unsworth,John Merritt %B Proceedings of the the seventh ACM conference on Hypertext %S HYPERTEXT '96 %I ACM %C New York, NY, USA %P 257– - 257– %8 1996/// %@ 0-89791-778-2 %G eng %U http://doi.acm.org/10.1145/234828.234854 %R 10.1145/234828.234854 %0 Conference Paper %B Proceedings of the 1st conference on Designing interactive systems: processes, practices, methods, & techniques %D 1995 %T An applied ethnographic method for redesigning user interfaces %A Rose,Anne %A Shneiderman, Ben %A Plaisant, Catherine %B Proceedings of the 1st conference on Designing interactive systems: processes, practices, methods, & techniques %S DIS '95 %I ACM %C New York, NY, USA %P 115 - 122 %8 1995/// %@ 0-89791-673-5 %G eng %U http://doi.acm.org/10.1145/225434.225447 %R 10.1145/225434.225447 %0 Journal Article %J Molecular and Biochemical Parasitology %D 1995 %T cDNA expressed sequence tags of Trypanosoma brucei rhodesiense provide new insights into the biology of the parasite %A El‐Sayed, Najib M. %A Alarcon,Clara M. %A Beck,John C. %A Sheffield,Val C. %A Donelson,John E. %K cDNA %K Expressed sequence tag %K Trypanosoma brucei rhodesiense %X A total of 518 expressed sequence tags (ESTs) have been generated from clones randomly selected from a cDNA library and a spliced leader sub-library of a Trypanosoma brucei rhodesiense bloodstream clone. 205 (39%) of the clones were identified based on matches to 113 unique genes in the public databases. Of these, 71 cDNAs display significant similarities to genes in unrelated organisms encoding metabolic enzymes, signal transduction proteins, transcription factors, ribosomal proteins, histones, a proliferation-associated protein and thimet oligopeptidase, among others. 313 of the cDNAs are not related to any other sequences in the databases. These cDNA ESTs provide new avenues of research for exploring both the novel trypanosome-specific genes and the genome organization of this parasite, as well as a resource for identifying trypanosome homologs to genes expressed in other organisms. %B Molecular and Biochemical Parasitology %V 73 %P 75 - 90 %8 1995/07// %@ 0166-6851 %G eng %U http://www.sciencedirect.com/science/article/pii/016668519500098L %N 1-2 %R 16/0166-6851(95)00098-L %0 Journal Article %J Integrated Planning Applications %D 1995 %T The COLLAGE/KHOROS link: Planning for image processing tasks %A Lansky,A. %A Friedman,M. %A Getoor, Lise %A Schmidler,S. %A Short Jr,N. %B Integrated Planning Applications %P 67 - 76 %8 1995/// %G eng %0 Journal Article %J Journal of Intelligent Information Systems %D 1995 %T Combining databases with prioritized information %A Pradhan,S. %A Minker, Jack %A V.S. Subrahmanian %X To solve a problem one may need to combine the knowledge of several different experts. It can happen that some of the claims of one or more experts may be in conflict with the claims of other experts. There may be several such points of conflict and any claim may be involved in several different such points of conflict. In that case, the user of the knowledge of experts may prefer a certain claim to another in one conflict-point without necessarily preferring that statement in another conflict-point.Our work constructs a framework within which the consequences of a set of such preferences (expressed as priorities among sets of statements) can be computed. We give four types of semantics for priorities, three of which are shown to be equivalent to one another. The fourth type of semantics for priorities is shown to be more cautious than the other three. In terms of these semantics for priorities, we give a function for combining knowledge from different sources such that the combined knowledge is conflict-free and satisfies all the priorities. %B Journal of Intelligent Information Systems %V 4 %P 231 - 260 %8 1995/// %G eng %N 3 %R 10.1007/BF00961654 %0 Journal Article %J AMERICAN PROGRAMMER %D 1995 %T Comprehensible, Predictable, and Controllable User Interfaces %A Shneiderman, Ben %B AMERICAN PROGRAMMER %V 8 %P 2 - 2 %8 1995/// %G eng %0 Journal Article %J TechnicM report, University of Maryland Department of Computer Science %D 1995 %T Data compression using locally consistent parsing %A Sahinalp,S. C %A Vishkin, Uzi %B TechnicM report, University of Maryland Department of Computer Science %8 1995/// %G eng %0 Conference Paper %B Parallel Processing Symposium, 1995. Proceedings., 9th International %D 1995 %T Data parallel programming in an adaptive environment %A Edjlali,G. %A Agrawal,G. %A Sussman, Alan %A Saltz, J. %B Parallel Processing Symposium, 1995. Proceedings., 9th International %P 827 - 832 %8 1995/// %G eng %0 Conference Paper %B Conference companion on Human factors in computing systems %D 1995 %T Dealing with complexity: uniting agents and direct manipulation (panel session) %A Riecken,Doug %A Maes,Pattie %A Shneiderman, Ben %A Smith,David Canfield %B Conference companion on Human factors in computing systems %S CHI '95 %I ACM %C New York, NY, USA %P 191 - 192 %8 1995/// %@ 0-89791-755-3 %G eng %U http://doi.acm.org/10.1145/223355.223504 %R 10.1145/223355.223504 %0 Book %D 1995 %T Development of interlingual lexical conceptual structures with syntactic markers for machine translation %A Dorr, Bonnie J %A Lee,J. %A Voss,C. %A Suh,S. %X This document reports on research conducted at the University of Mary-land for the Korean/English Machine Translation (MT) project. Our pri- mary objective was to develop an interlingual representation based on lexical conceptual structure (LCS) and to examine the relation between this repre- sentation and a set of linguistically motivated semantic classes. We view the work of the past year as a critical step toward achieving our goal of building a generator: the classification of LCS's into a semantic hierarchy provides a systematic mapping between semantic knowledge about verbs and their surface syntactic structures. We have focused on several areas in support of our objectives: (1) inves- tigation of morphological structure including distinctions between Korean and English; (2) porting a fast, message-passing parser to Korean (and to the IBM PC); (3) study of free word order and development of the associ- ated processing algorithm; (4) investigation of the aspectual dimension as it impacts morphology, syntax, and lexical semantics; (5) investigation of the relation between semantic classes and syntactic structure; (6) develop- ment of theta-role and lexical-semantic templates through lexical acquisition techniques; (7) definition a mapping between KR concepts and interlingual representations; (8) formalization of the lexical conceptual structure %I University of Maryland %8 1995/// %G eng %0 Journal Article %J ACM SIGCHI Bulletin %D 1995 %T Dissertations %E Humphrey,Susanne M. %E Shneiderman, Ben %X The following are citations selected by title and abstract as being related to Computer-Human Interaction, resulting from a computer search, using Dialog Information Services, of the Dissertation Abstracts Online database produced by University Microfilms International (UMI). Included are UMI order number, title, author, degree, year, institution; number of pages, Dissertation Abstracts International (DAI) subject category chosen by the author, and abstract. Unless otherwise specified, paper or microform copies of dissertations may be ordered from University Microfilms International, Dissertation Copies, Post Office Box 1764, Ann Arbor, MI 48106; telephone for U.S. (except Michigan, Hawaii, Alaska): 1-800-521-3042, for Canada: 1-800-343-5299. Price lists and other ordering and shipping information are in the introduction to the published DAI. An alternate source for copies is sometimes provided. Dissertation titles and abstracts contained here are published with permission of University Microfilms International, publishers of Dissertation Abstracts International (copyright by University Microfilms International), and may not be reproduced without their prior permission. %B ACM SIGCHI Bulletin %V 27 %P 73 - 75 %8 1995/01// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/202642.570129 %N 1 %R 10.1145/202642.570129 %0 Journal Article %J Computational Linguistics %D 1995 %T Efficient parsing for Korean and English: A parameterized message-passing approach %A Dorr, Bonnie J %A Lee,J. %A Lin,D. %A Suh,S. %B Computational Linguistics %V 21 %P 255 - 263 %8 1995/// %G eng %N 2 %0 Conference Paper %B Proceedings of the twenty-seventh annual ACM symposium on Theory of computing %D 1995 %T Euclidean spanners: short, thin, and lanky %A Arya,Sunil %A Das,Gautam %A Mount, Dave %A Salowe,Jeffrey S. %A Smid,Michiel %B Proceedings of the twenty-seventh annual ACM symposium on Theory of computing %S STOC '95 %I ACM %C New York, NY, USA %P 489 - 498 %8 1995/// %@ 0-89791-718-9 %G eng %U http://doi.acm.org/10.1145/225058.225191 %R 10.1145/225058.225191 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1995 %T Evaluating Spatial and Textual Style of Displays %A Shneiderman, Ben %A Chimera,Richard %A Jog,Ninog %A Stimart,Ren %A White,David %K human-computer interaction, Systems Integration Methodology %X The next generation of Graphic User Interfaces (GUIs) will offer rapid access to perceptually-rich, information abundant, and cognitively consistent interfaces. These new GUIs will be subjected to usability tests and expert reviews, plus new analysis methods and novel metrics to help guide designers. We have developed and tested first generation concordance tools to help developers to review terminology, capitalization, and abbreviation. We have also developed a dialog box summary table to help developers spot patterns and identify possible inconsistencies in layout, color fonts, font size, font style, and ordering of widgets. In this study we also explored the use of metrics such as widget counts, balance, alignment, density, and aspect ratios to provide further clues about where redesigns might be appropriate. Preliminary experience with several commercial projects is encouraging. %B Institute for Systems Research Technical Reports %8 1995/// %G eng %U http://drum.lib.umd.edu/handle/1903/5639 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1995 %T A Family of User Interface Consistency Checking Tools %A Mahajan,Rohit %A Shneiderman, Ben %K automated metrics %K concordance tools %K consistency checking tools %K Human-computer interaction %K spatial %K Systems Integration Methodology %K textual evaluation tools %K user interface %X Incorporating evaluation metrics with GUI development tools will help designers create consistent interfaces in the future. Complexity in design of interfaces makes efficient evaluation impossible by a single consistency checking evaluation tool. Our focus is on developing a family of evaluation tools in order to make the evaluation process less cumbersome. We have developed a dialog box typeface and color table to facilitate detection of anomalies in color, font, font size, and font style. Concordance tools have been developed to spot variant capitalization and abbreviations globally in the interface and specifically in the button widgets. As buttons are frequently used widgets, a button layout table has been created to spot any inconsistencies in height, width and relative position between a given group of buttons if present. Finally, a terminology basket tool has been created to identify unwanted synonyms of computer related terms used in the interface which may be misleading to the end user. %B Institute for Systems Research Technical Reports %8 1995/// %G eng %U http://drum.lib.umd.edu/handle/1903/5640 %0 Journal Article %J Proceedings of the Twentieth Annual Software Engineering Workshop %D 1995 %T A Family of User Interface Consistency Checking Tools: Design and Development of SHERLOCK %A Mahajan,R. %A Shneiderman, Ben %X Incorporating evaluation metrics with GUI development tools will help designers create consistent interfaces in the future. Complexity in design of interfaces makes efficient evaluation infeasible by a single consistency checking evaluation tool. Our focus is on developing a family of consistency checking tools to evaluate spatial layout and terminology in user interfaces and make the evaluation process less cumbersome. We have Created a dialog box summary table to provide a compact overview of spatial and visual properties of dozens or hundreds of dialog boxes of the interface. Interface concordance tool has been developed to spot variant capitalization and abbreviations in interface terminology. As buttons are most frequent used widgets, a button concordance tool and a button layout table has been constructed. Button concordance identifies variant capitalization, distinct typefaces, distinct background colors and variant sizes in buttons. Button layout table spots any inconsistencies in height, width and relative position between a given group of buttons. A spell checking tools which detects spelling errors in interface terms has also been included in the tool set. Finally, a terminology basket tool has been created to identify unwanted synonyms of computer related terms used in the interface. These tools are integrated together as SHERLOCK, a family of six consistency checking tools to expedite the evaluation process and provide feedback to the designers plus aid Usability Testing. %B Proceedings of the Twentieth Annual Software Engineering Workshop %8 1995/// %G eng %N NASA no. 19990027806 %0 Journal Article %J Tools and Algorithms for the Construction and Analysis of Systems %D 1995 %T A front-end generator for verification tools %A Cleaveland, Rance %A Madelaine,E. %A Sims,S. %B Tools and Algorithms for the Construction and Analysis of Systems %P 153 - 173 %8 1995/// %G eng %0 Journal Article %J Journal of Computational and Graphical Statistics %D 1995 %T Gauss, Statistics, and Gaussian Elimination %A Stewart, G.W. %X Gaussian elimination is the algorithm of choice for the solution of dense linear systems of equations. However, Gauss himself originally introduced his elimination procedure as a way of determining the precision of least squares estimates and only later described the computational algorithm. This article tells the story of Gauss, his algorithm, and its relation to his probabilistic development of least squares. %B Journal of Computational and Graphical Statistics %V 4 %P 1 - 11 %8 1995/03/01/ %@ 1061-8600 %G eng %U http://www.jstor.org/stable/1390624 %N 1 %R 10.2307/1390624 %0 Journal Article %J Electronic Transactions on Numerical Analysis %D 1995 %T On graded QR decompositions of products of matrices %A Stewart, G.W. %X This paper is concerned with the singular values and vectors of a product Mm =A1A2 ... Am of matrices of order n. The chief difficulty with computing them directly from Mm is that with increasing m the ratio of the small to the large singular values of Mm may fall below the rounding unit, so that the former are computed inaccurately. The solution proposed here is to compute recursively the factorization Mm = QRPT, where Q is orthogonal, R is a graded upper triangular, and PT is a permutation. %B Electronic Transactions on Numerical Analysis %V 3 %P 39 - 49 %8 1995/// %G eng %0 Journal Article %J Human factors in information systems: emerging theoretical bases %D 1995 %T HUMAN VALUE AND THE FUTURE OF TECHNOLOGY %A Shneiderman, Ben %B Human factors in information systems: emerging theoretical bases %P 355 - 355 %8 1995/// %G eng %0 Book %D 1995 %T Human-Computer Interaction Laboratory 1995 Video Reports %A Plaisant, Catherine %A Morrison,S. %A Skokowski,C. %A Reesch,J. %A Shneiderman, Ben %A Laboratory,University of Maryland at College Park. Human/Computer Interaction %A Channel,F. %I University of Maryland at College Park, Human/Computer Interaction Laboratory %8 1995/// %G eng %0 Journal Article %J IEEE Software %D 1995 %T Image-browser taxonomy and guidelines for designers %A Plaisant, Catherine %A Carr,D. %A Shneiderman, Ben %K analysis %K Computer Graphics %K design %K designer guidelines %K Equations %K Europe %K Evaluation %K Formal specifications %K Graphical user interfaces %K Guidelines %K IMAGE PROCESSING %K image-browser taxonomy %K informal specification technique %K Laboratories %K large image browsing %K Layout %K Road transportation %K selected image exploration %K SHAPE %K Software design %K task taxonomy %K Taxonomy %K tools %K two-dimensional browsing %K user interface management systems %K visual databases %X In many applications users must browse large images. Most designers merely use two one-dimensional scroll bars or ad hoc designs for two-dimensional scroll bars. However, the complexity of two-dimensional browsing suggests that more careful analysis, design, and evaluation might lead to significant improvements. Our exploration of existing 2D browsers has led us to identify many features and a wide variety of tasks performed with the browsers. We introduce an informal specification technique to describe 2D browsers and a task taxonomy, suggest design features and guidelines, and assess existing strategies. We focus on the tools to explore a selected image and so do not cover techniques to browse a series of images or to browse large-image databases %B IEEE Software %V 12 %P 21 - 32 %8 1995/03// %@ 0740-7459 %G eng %N 2 %R 10.1109/52.368260 %0 Report %D 1995 %T Improved approximation guarantees for packing and covering integer programs %A Srinivasan, Aravind %X Several important NP-hard combinatorial optimization problems canbe posed as packing/covering integer programs; the randomized rounding technique of Raghavan & Thompson is a powerful tool to approximate them well. We present one elementary unifying property of all these in- teger programs (IPs), and use the FKG correlation inequality to derive an improved analysis of randomized rounding on them. This also yields a pessimistic estimator, thus presenting deterministic polynomial-time algo- rithms for them with approximation guarantees signi cantly better than those known. %B DIMACS Technical Report %8 1995/09// %G eng %0 Conference Paper %B Proceedings of the twenty-seventh annual ACM symposium on Theory of computing %D 1995 %T Improved approximations of packing and covering problems %A Srinivasan, Aravind %B Proceedings of the twenty-seventh annual ACM symposium on Theory of computing %S STOC '95 %I ACM %C New York, NY, USA %P 268 - 276 %8 1995/// %@ 0-89791-718-9 %G eng %U http://doi.acm.org/10.1145/225058.225138 %R 10.1145/225058.225138 %0 Journal Article %J Communications of the ACM %D 1995 %T The info superhighway: for the people %A Shneiderman, Ben %X The opportunities are attractive, but some pavers of the Information Superhighway (ISH) are too eager to pour concrete. They risk making rough roads that will alienate the very users they seek. These technologically oriented ISH devotees may be building dramatic overpasses and painting stripes without figuring out where the highway should be going. I believe greater attention should be paid to identifying appropriate services, designing a consistent user interface, and developing a clearer model of the diverse user communities. %B Communications of the ACM %V 38 %P 162– - 162– %8 1995/01// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/204865.204898 %N 1 %R 10.1145/204865.204898 %0 Journal Article %J IEEE Transactions on Parallel and Distributed Systems %D 1995 %T An integrated runtime and compile-time approach for parallelizing structured and block structured applications %A Agrawal,G. %A Sussman, Alan %A Saltz, J. %K Bandwidth %K block structured applications %K block structured codes %K compile-time approach %K compiling applications %K data access patterns %K Data analysis %K Delay %K distributed memory machines %K distributed memory systems %K FORTRAN %K Fortran 90D/HPF compiler %K High performance computing %K HPF-like parallel programming languages %K integrated runtime approach %K irregularly coupled regular mesh problems %K multigrid code %K Navier-Stokes solver template %K Parallel machines %K parallel programming %K Pattern analysis %K performance evaluation %K program compilers %K Program processors %K Runtime library %K Uninterruptible power systems %X In compiling applications for distributed memory machines, runtime analysis is required when data to be communicated cannot be determined at compile-time. One such class of applications requiring runtime analysis is block structured codes. These codes employ multiple structured meshes, which may be nested (for multigrid codes) and/or irregularly coupled (called multiblock or irregularly coupled regular mesh problems). In this paper, we present runtime and compile-time analysis for compiling such applications on distributed memory parallel machines in an efficient and machine-independent fashion. We have designed and implemented a runtime library which supports the runtime analysis required. The library is currently implemented on several different systems. We have also developed compiler analysis for determining data access patterns at compile time and inserting calls to the appropriate runtime routines. Our methods can be used by compilers for HPF-like parallel programming languages in compiling codes in which data distribution, loop bounds and/or strides are unknown at compile-time. To demonstrate the efficacy of our approach, we have implemented our compiler analysis in the Fortran 90D/HPF compiler developed at Syracuse University. We have experimented with a multi-bloc Navier-Stokes solver template and a multigrid code. Our experimental results show that our primitives have low runtime communication overheads and the compiler parallelized codes perform within 20% of the codes parallelized by manually inserting calls to the runtime library %B IEEE Transactions on Parallel and Distributed Systems %V 6 %P 747 - 754 %8 1995/07// %@ 1045-9219 %G eng %N 7 %R 10.1109/71.395403 %0 Book Section %B Human-computer interactionHuman-computer interaction %D 1995 %T Interacting with computers %A Shneiderman, Ben %B Human-computer interactionHuman-computer interaction %I Morgan Kaufmann Publishers Inc. %C San Francisco %P 399 - 400 %8 1995/// %@ 1-55860-246-1 %G eng %0 Report %D 1995 %T An Iterative Method for Solving Linear Inequalities %A Stewart, G.W. %K Technical Report %X This paper describes and analyzes a method for finding nontrivialsolutions of the inequality $Ax \geq 0$, where $A$ is an $m \times n$ matrix of rank $n$. The method is based on the observation that a certain function $f$ has a unique minimum if and only if the inequality {\it fails to have} a nontrivial solution. Moreover, if there is a solution, an attempt to minimize $f$ will produce a sequence that will diverge in a direction that converges to a solution of the inequality. The technique can also be used to solve inhomogeneous inequalities and hence linear programming problems, although no claims are made about competitiveness with existing methods. %I Department of Computer Science, University of Maryland, College Park %V CS-TR-1833 %8 1995/02/06/ %G eng %U http://drum.lib.umd.edu/handle/1903/355 %0 Conference Paper %B Geoscience and Remote Sensing Symposium, 1995. IGARSS '95. 'Quantitative Remote Sensing for Science and Applications', International %D 1995 %T Land cover dynamics investigation using parallel computers %A Liang, S. %A Davis, Larry S. %A Townshend,J. %A Chellapa, Rama %A DeFries, R. %A Dubayah, R. %A Goward,S. %A JaJa, Joseph F. %A Krishnamachar, S. %A Roussopoulos, Nick %A Saltz, J. %A Samet, Hanan %A Shock,T. %A Srinivasan, M. %K ; geographic information system; geophysical measurement technique; geophysics computing; image classification; image map database; image segmentation; land cover dynamics; land surface; mixture modeling; object oriented programming; optical imaging; para %K GIS; IR imaging; Markovian random fields; atmospheric correction %X A comprehensive and highly interdisciplinary research program is being carried out to investigate global land cover dynamics in heterogeneous parallel computing environments. Some of the problems are addressed including atmospheric correction, mixture modeling, image classifications by Markovian random fields and by segmentation, global image/map databases, object oriented parallel programming and parallel/IO. During the initial two years project, significant progress has been made in all of these areas %B Geoscience and Remote Sensing Symposium, 1995. IGARSS '95. 'Quantitative Remote Sensing for Science and Applications', International %V 1 %P 332 -334 vol.1 - 332 -334 vol.1 %8 1995//10/14 %G eng %R 10.1109/IGARSS.1995.520273 %0 Journal Article %J Working Notes for IJCAI-95 Workshop on the Representation and Processing of Spatial Expressions, Montreal, Canada %D 1995 %T Lexical allocation in interlingua-based machine translation of spatial expressions %A Voss,C.R. %A Dorr, Bonnie J %A Sencan,M.U. %X Given a spatial expression, or its computationalsemantic form, how is the expression’s spatial semantics to be allocated lexically, i.e., among entries in the lexicon? In interlingua-based ma- chine translation (MT) research, lexical alloca- tion is the problem of allocating or subdividing a linguistic expression’s full interlingual (IL) structure into the substructures that are lexical IL forms, i.e., in the lexicon. Here we present our work developing IL forms and an IL lexicon for translating English spatial expressions into Turkish. We examine several co-occurrence patterns between motion verbs (spatial place- ment and displacement) and directional adpo- sitions (particles in English, postpositions in Turkish) and the lexical allocation of spatial vectors in these patterns %B Working Notes for IJCAI-95 Workshop on the Representation and Processing of Spatial Expressions, Montreal, Canada %8 1995/// %G eng %0 Journal Article %J Combinatorica %D 1995 %T The local nature of Δ-coloring and its algorithmic applications %A Panconesi,Alessandro %A Srinivasan, Aravind %X Given a connected graph G =( V, E ) with | V ⊧ n and maximum degree Δ such that G is neither a complete graph nor an odd cycle, Brooks' theorem states that G can be colored with Δ colors. We generalize this as follows: let G - v be Δ-colored; then, v can be colored by considering the vertices in an O (log Δ n ) radius around v and by recoloring an O (log Δ n ) length “augmenting path” inside it. Using this, we show that Δ-coloring G is reducible in O (log 3 n /logΔ) time to (Δ+1)-vertex coloring G in a distributed model of computation. This leads to fast distributed algorithms and a linear-processor NC algorithm for Δ-coloring. %B Combinatorica %V 15 %P 255 - 280 %8 1995/// %@ 0209-9683 %G eng %U http://dx.doi.org/10.1007/BF01200759 %N 2 %0 Conference Paper %B In: 27th ACM Symposium on Theory of Computing %D 1995 %T M.H.M: Euclidean spanners: short, thin, and lanky %A Arya,Sunil %A Dast,Gautam %A Mount, Dave %A Salowe,Jeffrey S. %A Smid,Michiel %X Euclidean spanners are important data structures in geometric algorithm design, because they provide a means of approximating the complete Euclidean graph with only O(n) edges, so that the shortest path length between each pair of points is not more than a constant factor longer than the Euclidean distance between the points. In many applications of spanners, it is important that the spanner possess a number of additional properties: low tot al edge weight, bounded degree, and low diameter. Existing research on spanners has considered one property or the other. We show that it is possible to build spanners in optimal O (n log n) time and O(n) space that achieve optimal or near optimal tradeoffs between all combinations of these *Max-Planck-Institut fiir Informatik, D-66123 Saarbrucken, %B In: 27th ACM Symposium on Theory of Computing %I ACM press %P 489 - 498 %8 1995/// %G eng %0 Journal Article %J Machine Vision and applications %D 1995 %T A miniaturized space-variant active vision system: Cortex-I %A Bederson, Benjamin B. %A Wallace,R.S. %A Schwartz,E. %B Machine Vision and applications %V 8 %P 101 - 109 %8 1995/// %G eng %N 2 %0 Journal Article %J AIP Conference Proceedings %D 1995 %T Next generation network management technology %A Baras,John S %A Atallah,George C %A Ball,Mike %A Goli,Shravan %A Karne,Ramesh K %A Kelley,Steve %A Kumar,Harsha %A Plaisant, Catherine %A Roussopoulos, Nick %A Schneiderman,Ben %A Srinivasarao,Mulugu %A Stathatos,Kosta %A Teittinen,Marko %A Whitefield,David %X Today’s telecommunications networks are becoming increasingly large, complex, mission critical and heterogeneous in several dimensions. For example, the underlying physical transmission facilities of a given network may be ‘‘mixed media’’ (copper, fiber‐optic, radio, and satellite); the subnetworks may be acquired from different vendors due to economic, performance, or general availability reasons; the information being transmitted over the network may be ‘‘multimedia’’ (video, data, voice, and images) and, finally, varying performance criteria may be imposed e.g., data transfer may require high throughput while the others, whose concern is voice communications, may require low call blocking probability. For these reasons, future telecommunications networks are expected to be highly complex in their services and operations. Due to this growing complexity and the disparity among management systems for individual sub‐networks, efficient network management systems have become critical to the current and future success of telecommunications companies. This paper addresses a research and development effort which focuses on prototyping configuration management, since that is the central process of network management and all other network management functions must be built upon it. Our prototype incorporates ergonomically designed graphical user interfaces tailored to the network configuration management subsystem and to the proposed advanced object‐oriented database structure. The resulting design concept follows open standards such as Open Systems Interconnection (OSI) and incorporates object oriented programming methodology to associate data with functions, permit customization, and provide an open architecture environment. © 1995 American Institute of Physics %B AIP Conference Proceedings %V 325 %P 75 - 82 %8 1995/01/25/ %@ 0094243X %G eng %U http://proceedings.aip.org/resource/2/apcpcs/325/1/75_1?isAuthorized=no %N 1 %R doi:10.1063/1.47255 %0 Conference Paper %B Proceedings of the Fourth Workshop on Enabling Technologies: Infrastructure for Collaborative Enterprises, 1995 %D 1995 %T Organization overviews and role management: inspiration for future desktop environments %A Plaisant, Catherine %A Shneiderman, Ben %K Asia %K bank data processing %K Databases %K Environmental economics %K Environmental management %K future desktop environments %K Graphical user interfaces %K human resource management %K management information systems %K Management training %K multiple personal roles %K office automation %K organization overviews %K personnel %K Project management %K Prototypes %K role management %K role-centered approach %K scheduling %K semi-automated searches %K User interfaces %K Utility programs %K World Bank %X In our exploration of future work environments for the World Bank we proposed two concepts. First, organization overviews provide a consistent support to present the results of a variety of manual or semi-automated searches. Second this view can be adapted or expanded for each class of users to finally map the multiple personal roles an individual has in an organization. After command line interfaces, graphical user interfaces, and the current “docu-centric” designs, a natural direction is towards a role-centered approach where we believe the emphasis is on the management of those multiple roles. Large visual overviews of the organization can be rapidly manipulated and zoomed in on to reveal the multiple roles each individual plays. Each role involves coordination with groups of people and accomplishment of tasks within a schedule %B Proceedings of the Fourth Workshop on Enabling Technologies: Infrastructure for Collaborative Enterprises, 1995 %I IEEE %P 14 - 22 %8 1995/04/20/22 %@ 0-8186-7019-3 %G eng %R 10.1109/ENABL.1995.484544 %0 Journal Article %J IEEE Transactions on ComputersIEEE Trans. Comput. %D 1995 %T Parametric dispatching of hard real-time tasks %A Gerber,R. %A Pugh, William %A Saksena,M. %B IEEE Transactions on ComputersIEEE Trans. Comput. %V 44 %P 471 - 479 %8 1995/03// %@ 00189340 %G eng %U http://dl.acm.org/citation.cfm?id=626999 %N 3 %R 10.1109/12.372041 %0 Report %D 1995 %T On the perturbation of Schur complement in positive semidefinite matrix %A Stewart, G.W. %X This note gives perturbation bounds for the Schur complement of apositive definite matrix in a positive semidefinite matrix. %I Instititue for Advanced Computer Studies, Univ of Maryland, College Park %V TR-95-38 %8 1995/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 1995 %T QUERY PREVIEWS IN NETWORKED INFORMATION SYSTEMS %A Doan,K. %A Plaisant, Catherine %A Shneiderman, Ben %X In a networked information system, there are three major obstacles facing users in a querying process: network performance, data volume and data complexity. In order to overcome these obstacles, we propose a two-phase approach to dynamic query formulation by volume preview. The two phases are the Query Preview and Query Refinement. In the Query Preview phase, users formulate an initial query by selecting desired attribute values. The volume of matching data sets is shown graphically on preview bars which aid users to rapidly eliminate undesired data sets, and focus on a manageable number of relevant data sets. Query previews also prevent wasted steps by eliminating zero-hit queries. When the estimated number of data sets is low enough, the initial query is submitted to the network, which returns the metadata of the data sets for further refinement in the Query Refinement phase. The two-phase approach to query formulation overcomes slow network performance, and reduces the data volume and data complexity problems. This approach is especially appropriate for users who prefer the exploratory method to discover data patterns and exceptions during the query formulation process. Using this approach, we have developed dynamic query user interfaces to allow users to formulate their queries across a networked environment. %B Institute for Systems Research Technical Reports %8 1995/10// %G eng %0 Journal Article %J SIAM Journal on Computing %D 1995 %T Randomness-Optimal Unique Element Isolation with Applications to Perfect Matching and Related Problems %A Chari,Suresh %A Rohatgi,Pankaj %A Srinivasan, Aravind %X In this paper, we precisely characterize the randomness complexity of the unique element isolation problem, a crucial step in the $RNC$ algorithm for perfect matching by Mulmuley, Vazirani, and Vazirani [Combinatorica, 7 (1987), pp. 105–113] and in several other applications. Given a set $S$ and an unknown family $\mathcal{F} \subseteq 2^{S}$ with $|\mathcal{F}| \leq Z$, we present a scheme for assigning polynomially bounded weights to the elements of $S$ using only $O(\log Z + \log |S|)$ random bits, such that the minimum weight set in $\mathcal{F}$ is unique with high probability. This generalizes the solution of Mulmuley, Vazirani, and Vazirani, who use $O(S \log S)$ bits, independent of $Z$. We also provide a matching lower bound for the randomness complexity of this problem. The new weight assignment scheme yields a randomness-efficient $RNC^{2}$ algorithm for perfect matching which uses $O(\log Z + \log n)$ random bits, where $Z$ is any given upper bound on the number of perfect matchings in the input graph. This generalizes the result of Grigoriev and Karpinski [Proc. IEEE Symposium on Foundations of computer Science, 1987, pp. 166–172], who presentan $NC^{3}$ algorithm when $Z$ is polynomial and improves the running time in this case. The worst-case randomness complexity of our algorithm is $O(n \log (m/n))$ random bits improving on the previous bound of $O(m \log n)$. Our scheme also gives randomness-efficient solutions for several problems where unique element isolation is used, such as $RNC$ algorithms for variants of matching and basic problems on linear matroids. We obtain a randomness-efficient random reduction from SAT to USAT, the language of uniquely satisfiable formulas, which can be derandomized in the case of languages in Few $P$ to yield new proofs of the results Few $P \subseteq \oplus P$ and Few $P \subseteq C_{=} P$. %B SIAM Journal on Computing %V 24 %P 1036 - 1036 %8 1995/// %@ 00975397 %G eng %U http://link.aip.org/link/SMJCAT/v24/i5/p1036/s1&Agg=doi %N 5 %R 10.1137/S0097539793250330 %0 Report %D 1995 %T The representation of document structure: A generic object-process approach %A Dori,D. %A David Doermann %A Shin,C. %A Haralick,R. %A Phillips,I. %A Buchman,M. %A Ross,D. %I University of Maryland, College Park %V CAR-TR-785 %8 1995/// %G eng %0 Journal Article %J American Association of Artificial Intelligence Spring Symposium on Representing Mental States and Mechanisms %D 1995 %T Representing a student’s learning states and transitions %A Gurer,D. %A desJardins, Marie %A Schlager,M. %X We describe an ongoing project to develop an adaptive training system (ATS) that dynamically models a student’s learning processes and can provide specialized tutoring adapted to a student’s knowledge state and learning style. The student modeling component of the ATS, ML-Modeler, uses machine learning (ML) techniques to emulate the student’s novice-toexpert transition. ML-Modeler infers which learning methods the student has used to reach the current knowledge state by comparing the student’s solution trace to an expert solution and generating plausible hypotheses about what misconceptions and errors the student has made. A case-based approach is used to generate hypotheses through incorrectly applying analogy, overgeneralization, and overspecialization. The student and expert models use a network-based representation that includes abstract concepts and relationships as well as strategies for problem solving. Fuzzy methods are used to represent the uncertainty in the student model. This paper describes the design of the ATS and ML-Modeler, and gives a detailed example of how the system would model and tutor the student in a typical session. The domain we use for this example is high-school level chemistry. %B American Association of Artificial Intelligence Spring Symposium on Representing Mental States and Mechanisms %8 1995/// %G eng %0 Journal Article %J Numerical Linear Algebra with Applications %D 1995 %T On the solution of block Hessenberg systems %A Stewart, G.W. %K block Hessenberg matrix %K Block Toeplitz matrix %K linear system %K queue %X This paper describes a divide-and-conquer strategy for solving block Hessenberg systems. For dense matrices the method is as efficient as Gaussian elimination; however, because it works almost entirely with the original blocks, it is much more efficient for sparse matrices or matrices whose blocks can be generated on the fly. For Toeplitz matrices, the algorithm can be combined with the fast Fourier transform. %B Numerical Linear Algebra with Applications %V 2 %P 287 - 296 %8 1995/05/01/ %@ 1099-1506 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/nla.1680020309/abstract %N 3 %R 10.1002/nla.1680020309 %0 Conference Paper %B , 36th Annual Symposium on Foundations of Computer Science, 1995. Proceedings %D 1995 %T Splitters and near-optimal derandomization %A Naor,M. %A Schulman,L. J %A Srinivasan, Aravind %K Boosting %K Circuit testing %K computational complexity %K computational linguistics %K Computer science %K Contracts %K derandomization %K deterministic constructions %K Educational institutions %K Engineering profession %K exhaustive testing %K fairly general method %K fixed-subgraph finding algorithms %K hardness of approximation %K Information systems %K k-restrictions %K learning %K local-coloring protocol %K MATHEMATICS %K near-optimal constructions %K near-optimal derandomization %K Parallel algorithms %K probabilistic bound %K probability %K Protocols %K randomised algorithms %K Set cover %K splitters %X We present a fairly general method for finding deterministic constructions obeying what we call k-restrictions; this yields structures of size not much larger than the probabilistic bound. The structures constructed by our method include (n,k)-universal sets (a collection of binary vectors of length n such that for any subset of size k of the indices, all 2k configurations appear) and families of perfect hash functions. The near-optimal constructions of these objects imply the very efficient derandomization of algorithms in learning, of fixed-subgraph finding algorithms, and of near optimal ΣIIΣ threshold formulae. In addition, they derandomize the reduction showing the hardness of approximation of set cover. They also yield deterministic constructions for a local-coloring protocol, and for exhaustive testing of circuits %B , 36th Annual Symposium on Foundations of Computer Science, 1995. Proceedings %I IEEE %P 182 - 191 %8 1995/10/23/25 %@ 0-8186-7183-1 %G eng %R 10.1109/SFCS.1995.492475 %0 Journal Article %J Signal Processing, IEEE Transactions on %D 1995 %T On the stability of sequential updates and downdates %A Stewart, G.W. %K algorithm;Cholesky %K algorithm;downdating %K algorithm;error %K algorithm;URV %K algorithms;hyperbolic %K analysis;matrix %K analysis;sequential %K Chambers' %K condition;rounding %K decomposition;backward %K decomposition;numerical %K decomposition;plane %K decompositions;LINPACK %K decompositions;updating %K downdates;sequential %K error %K errors;sequences; %K orthogonal %K rotations;relational %K stability %K stability;roundoff %K stable %K transformations;matrix %K updates;stability;two-sided %X The updating and downdating of Cholesky decompositions has important applications in a number of areas. There is essentially one standard updating algorithm, based on plane rotations, which is backward stable. Three downdating algorithms have been treated in the literature: the LINPACK algorithm, the method of hyperbolic transformations, and Chambers' (1971) algorithm. Although none of these algorithms is backward stable, the first and third satisfy a relational stability condition. It is shown that relational stability extends to a sequence of updates and downdates. In consequence, other things being equal, if the final decomposition in the sequence is well conditioned, it will be accurately computed, even though intermediate decompositions may be almost completely inaccurate. These results are also applied to the two-sided orthogonal decompositions, such as the URV decomposition %B Signal Processing, IEEE Transactions on %V 43 %P 2642 - 2648 %8 1995/11// %@ 1053-587X %G eng %N 11 %R 10.1109/78.482114 %0 Journal Article %J Proc. IFIP 2.6 Visual Databases Systems %D 1995 %T Starfield information visualization with interactive smooth zooming %A Jog,N. %A Shneiderman, Ben %X This paper discusses the design and implementation of interactive smooth zooming of a starfield display (which is a visualization of a multi-attribute database) and introduces the zoom bar, a new widget for zooming and panning. Whereas traditional zoom techniques are based on zooming towards or away from a focal point, this paper introduces a novel approach based on zooming towards or away from a fixed line.Starfield displays plot items from a database as small selectable glyphs using two of the ordinal attributes of the data as the variables along the display axes. One way of filtering this visual information is by changing the range of displayed values on either of the display axes. If this is done incrementally and smoothly, the starfield display appears to zoom in and out, and users can track the motion of the glyphs without getting disoriented by sudden, large changes in context. %B Proc. IFIP 2.6 Visual Databases Systems %P 3 - 14 %8 1995/// %G eng %0 Book Section %B Visual database systems, IIIVisual database systems, III %D 1995 %T Starfield visualization with interactive smooth zooming %A Jogt,NK %A Shneiderman, Ben %X This paper discusses the design and implementation of interactive smooth zooming of a starfield display (which is a visualization of a multi-attribute database) and introduces the zoom bar, a new widget for zooming and panning. Whereas traditional zoom techniques are based on zooming towards or away from a focal point, this paper introduces a novel approach based on zooming towards or away from a fixed line.Starfield displays plot items from a database as small selectable glyphs using two of the ordinal attributes of the data as the variables along the display axes. One way of filtering this visual information is by changing the range of displayed values on either of the display axes. If this is done incrementally and smoothly, the starfield display appears to zoom in and out, and users can track the motion of the glyphs without getting disoriented by sudden, large changes in context. %B Visual database systems, IIIVisual database systems, III %V 3 %P 1 - 1 %8 1995/// %G eng %0 Journal Article %J ACM Computing Surveys %D 1995 %T Survival of the fittest: the evolution of multimedia user interfaces %A Preece,Jenny %A Shneiderman, Ben %B ACM Computing Surveys %V 27 %P 557 - 559 %8 1995/12// %@ 0360-0300 %G eng %U http://doi.acm.org/10.1145/234782.234789 %N 4 %R 10.1145/234782.234789 %0 Journal Article %J SIAM, Philadelphia.[Translation of Gauss (1821, 1823, 1826).] %D 1995 %T Theory of the Combination of Observations Least Subject to Errors %A Stewart, G.W. %B SIAM, Philadelphia.[Translation of Gauss (1821, 1823, 1826).] %8 1995/// %G eng %0 Journal Article %J Information processing letters %D 1995 %T Using synthetic perturbations and statistical screening to assay shared-memory programs %A Snelick,R. %A JaJa, Joseph F. %A Kacker,R. %A Lyon,G. %B Information processing letters %V 54 %P 147 - 153 %8 1995/// %G eng %N 3 %0 Journal Article %J Information Systems ResearchInformation Systems Research %D 1995 %T Using Treemaps to Visualize the Analytic Hierarchy Process %A Asahi,Toshiyuki %A Turo,David %A Shneiderman, Ben %K AHP %K analytic hierarchy process %K decision support %K treemap %K User interfaces %K Visualization %X Treemaps, a visualization method for large hierarchical data spaces, are used to augment the capabilities of the Analytic Hierarchy Process (AHP) for decision-making. Two direct manipulation tools, presented metaphorically as a “pump” and a “hook,” were developed and applied to the treemap to support AHP sensitivity analysis. Users can change the importance of criteria dynamically on the two-dimensional treemap and immediately see the impact on the outcome of the decision. This fluid process dramatically speeds up exploration and provides a better understanding of the relative impact of the component criteria. A usability study with six subjects using a prototype AHP application showed that treemap representation was acceptable from a visualization and data operation standpoint. %B Information Systems ResearchInformation Systems Research %V 6 %P 357 - 375 %8 1995/12/01/ %@ 1047-7047, 1526-5536 %G eng %U http://isr.journal.informs.org/content/6/4/357 %N 4 %R 10.1287/isre.6.4.357 %0 Journal Article %J Readings in human-computer interaction: toward the year 2000 %D 1995 %T Vision, Graphic Design, and Visual Display %A Marcus,A. %A Murch,G.M. %A Baecker,R. %A Small,I. %A Mander,R. %A Ahlberg,C. %A Shneiderman, Ben %B Readings in human-computer interaction: toward the year 2000 %P 411 - 411 %8 1995/// %G eng %0 Conference Paper %B Conference companion on Human factors in computing systems %D 1995 %T Visual decision-making: using treemaps for the analytic hierarchy process %A Asahi,Toshiyuki %A Turo,David %A Shneiderman, Ben %B Conference companion on Human factors in computing systems %S CHI '95 %I ACM %C New York, NY, USA %P 405 - 406 %8 1995/// %@ 0-89791-755-3 %G eng %U http://doi.acm.org/10.1145/223355.223747 %R 10.1145/223355.223747 %0 Book Section %B Proceedings of the Second International Workshop on Markov ChainsProceedings of the Second International Workshop on Markov Chains %D 1995 %T What is Fundamental for Markov Chains: First Passage Times, Fundamental Matrices, and Group Generalized Inverses %A Heyman,D. P. %A O'Leary, Dianne P. %E Stewart,W. %B Proceedings of the Second International Workshop on Markov ChainsProceedings of the Second International Workshop on Markov Chains %I Kluwer Academic Publishers %P 151 - 161 %8 1995/// %G eng %0 Journal Article %J Communications of the ACM %D 1995 %T Windows of opportunity in electronic classrooms %A Shneiderman, Ben %A Alavi,Maryann %A Norman,Kent %A Borkowski,Ellen Yu %X Paradigm-shifting landmark buildings are cherished by their occupants and remembered because they reshape our expectations of schools, homes, or offices. Classic examples include Thomas Jefferson's communal design of the “academical village” at the University of Virginia where faculty and students lived close to classrooms, Frank Lloyd Wright's organic harmony with nature in Fallingwater (in western Pennsylvania) where the waterfall sounds and leafy surroundings offered a stress-reducing getaway for an urban executive, or Kevin Roche's open glass-walled Ford Foundation (in New York City) that promoted new team-oriented management strategies. %B Communications of the ACM %V 38 %P 19 - 24 %8 1995/11// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/219717.219725 %N 11 %R 10.1145/219717.219725 %0 Journal Article %J Journal of Artificial Neural Networks %D 1994 %T Adaptation of noncompetitive and competitive neural networks to focal lesions %A Weinrich,M. %A Sutton,G. G %A Reggia, James A. %A D'Autrechy,C. L %B Journal of Artificial Neural Networks %V 1 %P 51 - 60 %8 1994/// %G eng %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: celebrating interdependence %D 1994 %T The alphaslider: a compact and rapid selector %A Ahlberg,Christopher %A Shneiderman, Ben %K Alphaslider %K dynamic queries %K menus %K selection technology %K widget %B Proceedings of the SIGCHI conference on Human factors in computing systems: celebrating interdependence %S CHI '94 %I ACM %C New York, NY, USA %P 365 - 371 %8 1994/// %@ 0-89791-650-6 %G eng %U http://doi.acm.org/10.1145/191666.191790 %R 10.1145/191666.191790 %0 Journal Article %J Journal of Heredity %D 1994 %T An Atlas of Drosophila Genes: Sequences and Molecular Features %A Maroni,G. %A Mount, Stephen M. %A Cavener,DR %A Cavener,BA %A Sharp,PM %A Lloyd,AT %A Schaeffer,SW %B Journal of Heredity %V 85 %P 495 - 496 %8 1994/// %G eng %N 6 %0 Journal Article %J ACM SIGCHI Bulletin %D 1994 %T AVI '94: An International Workshop %A Shneiderman, Ben %A Badre,AI %A Santos,Paulo %X The fresh Mediterranean breezes of innovation enlivened the workshop conversations at Advanced Visual Interfaces '94 (Bari, Italy June 1--4, 1994). More than 100 participants met on the flower-gardened grounds of the Grande Albergo di Villa Romanazzi Carducci, whose modern hotel plus conference center complemented the beautifully restored old villa and the marble-lined swimming pool. Being together with active professionals from Europe and America for 4 days gave us an in-depth chance to learn about each other's professional and personal outlooks. The ancient Roman architecture and the traditional Italian surroundings reminded us that the impact of our work could also last for two thousand years and shape many people's daily lives. %B ACM SIGCHI Bulletin %V 26 %P 54 - 55 %8 1994/10// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/191642.1047934 %N 4 %R 10.1145/191642.1047934 %0 Conference Paper %B Conference companion on Human factors in computing systems %D 1994 %T Beyond accuracy, reliability, and efficiency: criteria for a good computer system %A Friedman,Batya %A Levenson,Nancy %A Shneiderman, Ben %A Suchman,Lucy %A Winograd,Terry %B Conference companion on Human factors in computing systems %S CHI '94 %I ACM %C New York, NY, USA %P 195 - 198 %8 1994/// %@ 0-89791-651-4 %G eng %U http://doi.acm.org/10.1145/259963.260253 %R 10.1145/259963.260253 %0 Conference Paper %B , 35th Annual Symposium on Foundations of Computer Science, 1994 Proceedings %D 1994 %T Computing with very weak random sources %A Srinivasan, Aravind %A Zuckerman,D. %K Application software %K BPP simulations %K Chor-Goldreich sources %K computational complexity %K Computational modeling %K Computer science %K Computer simulation %K cryptography %K distributed algorithms %K expander constructions %K hardness %K MATHEMATICS %K min-entropy %K Physics computing %K Polynomials %K probability %K R-bit string %K randomness-efficient Leftover Hash Lemma %K RP algorithms simulation %K Testing %K time-space tradeoffs %K very weak random sources %X For any fixed ε>0, we show how to simulate RP algorithms in time nO(log n) using the output of a δ-source with min-entropy R(ε). Such a weak random source is asked once for R(ε) bits; it outputs an R-bit string such that any string has probability at most 2-R(ε). If ε>1-1/(k+1), our BPP simulations take time nO(log(k n)) (log(k) is the logarithm iterated k times). We also give a polynomial-time BPP simulation using Chor-Goldreich sources of min-entropy RΩ(1), which is optimal. We present applications to time-space tradeoffs, expander constructions, and the hardness of approximation. Also of interest is our randomness-efficient Leftover Hash Lemma, found independently by Goldreich and Wigderson %B , 35th Annual Symposium on Foundations of Computer Science, 1994 Proceedings %I IEEE %P 264 - 275 %8 1994/11/20/22 %@ 0-8186-6580-7 %G eng %R 10.1109/SFCS.1994.365688 %0 Journal Article %J Numerische Mathematik %D 1994 %T On the convergence of line iterative methods for cyclically reduced non-symmetrizable linear systems %A Elman, Howard %A Golub, G. H %A Starke, G. %B Numerische Mathematik %V 67 %P 177 - 190 %8 1994/// %G eng %N 2 %0 Journal Article %J Neural Computation %D 1994 %T Cortical map reorganization as a competitive process %A Sutton III,G. G. %A Reggia, James A. %A Armentrout,S. L %A D'Autrechy,C. L %B Neural Computation %V 6 %P 1 - 13 %8 1994/// %G eng %N 1 %0 Conference Paper %B Proceedings of the workshop on Advanced visual interfaces %D 1994 %T Data structures for dynamic queries: an analytical and experimental evaluation %A Jain,Vinit %A Shneiderman, Ben %X Dynamic Queries is a querying technique for doing range search on multi-key data sets. It is a direct manipulation mechanism where the query is formulated using graphical widgets and the results are displayed graphically preferably within 100 milliseconds.This paper evaluates four data structures, the multilist, the grid file, k-d tree and the quad tree used to organize data in high speed storage for dynamic queries. The effect of factors like size, distribution and dimensionality of data on the storage overhead and the speed of search is explored. Analytical models for estimating the storage and the search overheads are presented, and verified to be correct by empirical data. Results indicate that multilists are suitable for small (few thousand points) data sets irrespective of the data distribution. For large data sets the grid files are excellent for uniformly distributed data, and trees are good for skewed data distributions. There was no significant difference in performance between the tree structures. %B Proceedings of the workshop on Advanced visual interfaces %S AVI '94 %I ACM %C New York, NY, USA %P 1 - 11 %8 1994/// %@ 0-89791-733-2 %G eng %U http://doi.acm.org/10.1145/192309.192313 %R 10.1145/192309.192313 %0 Journal Article %J Principles of Real-Time Systems. Prentice Hall %D 1994 %T Design and implementation of Maruti-II %A Saksena,M. %A da Silva,J. %A Agrawala, Ashok K. %B Principles of Real-Time Systems. Prentice Hall %8 1994/// %G eng %0 Journal Article %J Computational Geometry: Theory and Applications %D 1994 %T Dynamic algorithms for geometric spanners of small diameter: Randomized solutions %A Arya,Sunil %A Mount, Dave %A Smid,Michiel %X Let S be a set of n points in IR d and let t ? 1 be a real number. A t-spanner for S is a directed graph having the points of S as its vertices, such that for any pair p and q of points there is a path from p to q of length at most t times the Euclidean distance between p and q. Such a path is called a t-spanner path. The spanner diameter of such a spanner is defined as the smallest integer D such that for any pair p and q of points there is a t-spanner path from p to q containing at most D edges. A randomized algorithm is given for constructing a t-spanner that, with high probability, contains O(n) edges and has spanner diameter O(log n). A data structure of size O(n log d n) is given that maintains this t-spanner in O(log d n log log n) expected amortized time per insertion and deletion, in the model of random updates, as introduced by Mulmuley. Keywords: Computational geometry, proximity problems, skip lists, randomization, dynamic data structures. Preprint submitted to Els... %B Computational Geometry: Theory and Applications %V 13 %P 13 - 91 %8 1994/// %G eng %0 Journal Article %J IEEE Software %D 1994 %T Dynamic queries for visual information seeking %A Shneiderman, Ben %K Algorithm design and analysis %K animated results %K animation %K Application software %K Command languages %K complex queries %K database management systems %K Databases %K display algorithms %K Displays %K dynamic queries %K Educational institutions %K Games %K Graphical user interfaces %K human factors %K Query processing %K retrieval %K Runtime %K Technological innovation %K user-interface design %K visual databases %K visual information seeking %K visual interfaces %K widgets %X Considers how dynamic queries allow users to "fly through" databases by adjusting widgets and viewing the animated results. In studies, users reacted to this approach with an enthusiasm more commonly associated with video games. Adoption requires research into retrieval and display algorithms and user-interface design. The author discusses how experts may benefit from visual interfaces because they will be able to formulate more complex queries and interpret intricate results.<> %B IEEE Software %V 11 %P 70 - 77 %8 1994/11// %@ 0740-7459 %G eng %N 6 %R 10.1109/52.329404 %0 Journal Article %J Unpublished manuscript, University of Maryland, College Park %D 1994 %T Education by engagement and construction: Can distance learning be better than face-to-face %A Shneiderman, Ben %X This essay describes an emerging theory of "education by engagement and construction," in which students work in teams to create ambitious projects with results that are presented to someone other than the professor. The video shows how a distance learning Graduate Computer Science Seminar titled "Virtual Reality, Telepresence and Beyond" was conducted according to this theory. The intense interactions by satellite TV and electronic mail may have created a greater sense of interaction and intimacy among the students than many face-to-face courses. (Video available as part of the HCIL 1994 Video Reports, contact hcil-info@cs.umd.edu for ordering information) %B Unpublished manuscript, University of Maryland, College Park %8 1994/// %G eng %0 Journal Article %J Neural Networks, IEEE Transactions on %D 1994 %T Effects of normalization constraints on competitive learning %A Sutton,G. G %A Reggia, James A. %B Neural Networks, IEEE Transactions on %V 5 %P 502 - 504 %8 1994/// %G eng %N 3 %0 Conference Paper %B Scalable High-Performance Computing Conference, 1994., Proceedings of the %D 1994 %T Efficient runtime support for parallelizing block structured applications %A Agrawal,G. %A Sussman, Alan %A Saltz, J. %K application programmers %K block structured applications %K distributed memory parallel machines %K distributed memory systems %K engineering applications %K irregularly coupled regular mesh problems %K machine-independent %K multiblock %K multiblock computational fluid dynamics code %K multiblock template %K multigrid codes %K Parallel machines %K parallel programming %K Physics computing %K runtime communication overhead %K Runtime library %K runtime support %K scientific applications %K software reusability %K structured meshes %X Scientific and engineering applications often involve structured meshes. These meshes may be nested (for multigrid codes) and/or irregularly coupled (called multiblock or irregularly coupled regular mesh problems). We describe a runtime library for parallelizing these applications on distributed memory parallel machines in an efficient and machine-independent fashion. This runtime library is implemented on several different systems. This library can be used by application programmers to port applications by hand and can also be used by a compiler to handle communication for these applications. Our experimental results show that our primitives have low runtime communication overheads. We have used this library to port a multiblock template and a multigrid code. Effort is also underway to port a complete multiblock computational fluid dynamics code using our library %B Scalable High-Performance Computing Conference, 1994., Proceedings of the %P 158 - 167 %8 1994/05// %G eng %R 10.1109/SHPCC.1994.296639 %0 Journal Article %J ACM Trans. Inf. Syst. %D 1994 %T An exploratory evaluation of three interfaces for browsing large hierarchical tables of contents %A Chimera,Richard %A Shneiderman, Ben %K Browsing %K hierarchies %K table of contents %K User interfaces %X Three different interfaces were used to browse a large (1296 items) table of contents. A fully expanded stable interface, expand/contract interface, and multipane interface were studied in a between-groups experiment with 41 novice participants. Nine timed fact retrieval tasks were performed; each task is analyzed and discussed separately. We found that both the expand/contract and multipane interfaces produced significantly faster times than the stable interface for many tasks using this large hierarchy; other advantages of the expand/contract and multipane interfaces over the stable interface are discussed. The animation characteristics of the expand/contract interface appear to play a major role. Refinements to the multipane and expand/contract interfaces are suggested. A predictive model for measuring navigation effort of each interface is presented. %B ACM Trans. Inf. Syst. %V 12 %P 383 - 406 %8 1994/10// %@ 1046-8188 %G eng %U http://doi.acm.org/10.1145/185462.185483 %N 4 %R 10.1145/185462.185483 %0 Journal Article %J CONCUR'94: Concurrency Theory %D 1994 %T Fully abstract characterizations of testing preorders for probabilistic processes %A Yuen,S. %A Cleaveland, Rance %A Dayar,Z. %A Smolka,S. %B CONCUR'94: Concurrency Theory %P 497 - 512 %8 1994/// %G eng %0 Book Section %B People and Computers %D 1994 %T The future of graphic user interfaces: Personal role managers %A Shneiderman, Ben %A Plaisant, Catherine %B People and Computers %I Cambridge University Press %P 444 - 444 %8 1994/// %@ 9780521485579 %G eng %0 Conference Paper %B Pattern Recognition, 1994. Vol. 3-Conference C: Signal Processing, Proceedings of the 12th IAPR International Conference on %D 1994 %T High performance computing for land cover dynamics %A Parulekar,R. %A Davis, Larry S. %A Chellapa, Rama %A Saltz, J. %A Sussman, Alan %A Townshend,J. %B Pattern Recognition, 1994. Vol. 3-Conference C: Signal Processing, Proceedings of the 12th IAPR International Conference on %P 234 - 238 %8 1994/// %G eng %0 Journal Article %J Institute for Systems Research Technical Reports %D 1994 %T Image Browsers: Taxonomy, Guidelines, and Informal Specifications %A Plaisant, Catherine %A Carr,David A %A Shneiderman, Ben %K Graphics %K Systems Integration %K use interface %X Image browsing is necessary in numerous applications. Designers have merely used two one-dimensional scroll bars or they have made ad hoc designs for a two-dimensional scroll bar. However, the complexity of two-dimensional browsing suggests that more careful analysis, design, and evaluation might lead to significant improvements. We present a task taxonomy for image browsing, suggest design features and guidelines, assess existing strategies, and introduce an informal specification %B Institute for Systems Research Technical Reports %8 1994/// %G eng %U http://drum.lib.umd.edu/handle/1903/5591 %0 Book %D 1994 %T Interacting with virtual environments %A MacDonald,L. %A Vince,J. %A Shneiderman, Ben %I John Wiley & Sons %8 1994/// %G eng %0 Conference Paper %B Scalable Parallel Libraries Conference, 1994., Proceedings of the 1994 %D 1994 %T Jovian: A framework for optimizing parallel I/O %A Bennett, R. %A Bryant,K. %A Sussman, Alan %A Das,R. %A Saltz, J. %B Scalable Parallel Libraries Conference, 1994., Proceedings of the 1994 %P 10 - 20 %8 1994/// %G eng %0 Conference Paper %B Handbook of logic in artificial intelligence and logic programming %D 1994 %T Meta-languages, reflection principles and self-reference %A Perlis, Don %A V.S. Subrahmanian %B Handbook of logic in artificial intelligence and logic programming %P 323 - 358 %8 1994/// %G eng %0 Journal Article %J IEEE Transactions on Robotics and Automation %D 1994 %T A miniature pan-tilt actuator: the spherical pointing motor %A Bederson, Benjamin B. %A Wallace,R.S. %A Schwartz,E.L. %K absolute positioning device %K active vision %K Actuators %K Application software %K Assembly %K automatic calibration %K camera sensor %K CAMERAS %K CCD image sensors %K CCD sensor chip %K Charge coupled devices %K Charge-coupled image sensors %K closed-loop control strategies %K computerised control %K Lenses %K Magnetic fields %K miniature lens assembly %K miniature pan-tilt actuator %K open-loop control strategies %K orthogonal motor windings %K pan-tilt mechanism %K Permanent magnet motors %K position control %K Scanning probe microscopy %K spherical pointing motor %K two-degree-of-freedom device %X A pan-tilt mechanism is a computer-controlled actuator designed to point an object such as a camera sensor. For applications in active vision, a pan-tilt mechanism should be accurate, fast, small, inexpensive and have low power requirements. The authors have designed and constructed a new type of actuator meeting these requirements, which incorporates both pan and tilt into a single, two-degree-of-freedom device. The spherical pointing motor (SPM) consists of three orthogonal motor windings in a permanent magnetic field, configured to move a small camera mounted on a gimbal. It is an absolute positioning device and is run open-loop. The SPM is capable of panning and tilting a load of 15 grams, for example a CCD image sensor, at rotational velocities of several hundred degrees per second with a repeatability of .15°. The authors have also built a miniature camera consisting of a single CCD sensor chip and miniature lens assembly that fits on the rotor of this motor. In this paper, the authors discuss the theory of the SPM, which includes its basic electromagnetic principles, and derive the relationship between applied currents and resultant motor position. The authors present an automatic calibration procedure and discuss open- and closed-loop control strategies. Finally, the authors present the physical characteristics and results of their prototype %B IEEE Transactions on Robotics and Automation %V 10 %P 298 - 308 %8 1994/06// %@ 1042-296X %G eng %N 3 %R 10.1109/70.294205 %0 Journal Article %J Communications of the ACM %D 1994 %T Navigating in hyperspace: designing a structure-based toolbox %A Rivlin,Ehud %A Botafogo,Rodrigo %A Shneiderman, Ben %B Communications of the ACM %V 37 %P 87 - 96 %8 1994/02// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/175235.175242 %N 2 %R 10.1145/175235.175242 %0 Report %D 1994 %T Next Generation Network Management Technology %A Atallah,George C %A Ball,Michael O %A Baras,John S %A Goli,Shravan K %A Karne,Ramesh K %A Kelley,Stephen %A Kumar,Harsha P. %A Plaisant, Catherine %A Roussopoulos, Nick %A Shneiderman, Ben %A Srinivasarao,Mulugu %A Stathatos,Kostas %A Teittinen,Marko %A Whitefield,David %K Constraints for Network Management. %K Network Configuration Management %K network management %K Object Oriented Data Base Model for Network Management %K Rules %K Systems Integration %K Visual Information Management for Network Configuration Management %X Today's telecommunications networks are becoming increasingly large, complex, mission critical and heterogeneous in several dimensions. For example, the underlying physical transmission facilities of a given network may be ﲭixed media (copper, fiber- optic, radio, and satellite); the sub networks may be acquired from different vendors due to economic, performance, or general availability reasons; the information being transmitted over the network may be ﲭultimedia (video, data, voice, and images) and, finally, varying performance criteria may be imposed e.g. data transfer may require high throughput while the others, whose concern is voice communications, may require low call blocking probability. For these reasons, future telecommunications networks are expected to be highly complex in their services and operations. Due to this growing complexity and the disparity among management systems for individual sub networks, efficient network management systems have become critical to the current and future success of telecommunications companies. This paper addresses a research and development effort which focuses on prototyping configuration management, since that is the central process of network management and all other network management functions must be built upon it. Our prototype incorporates ergonomically designed graphical user interfaces tailored to the network configuration management subsystem and to the proposed advanced object-oriented database structure. The resulting design concept follows open standards such as Open Systems Interconnection (OSI) and incorporates object oriented programming methodology to associate data with functions, permit customization, and provide an open architecture environment.

%B Institute for Systems Research Technical Reports %8 1994/// %G eng %U http://drum.lib.umd.edu/handle/1903/5519 %0 Journal Article %J Proceedings of the International Conference on New Methods in Language Processing, Manchester, UK %D 1994 %T A paradigm for non-head-driven parsing: Parameterized message-passing %A Dorr, Bonnie J %A Lin,D. %A Lee,J. %A Suh,S. %B Proceedings of the International Conference on New Methods in Language Processing, Manchester, UK %8 1994/// %G eng %0 Journal Article %J Lecture Notes in Computer Science %D 1994 %T On a parallel-algorithms method for string matching problems %A Sahinalp,S. C %A Vishkin, Uzi %B Lecture Notes in Computer Science %V 778 %P 22 - 32 %8 1994/// %G eng %0 Book Section %B Algorithms and ComplexityAlgorithms and Complexity %D 1994 %T On a parallel-algorithms method for string matching problems (overview) %A Sahinalp,Suleyman %A Vishkin, Uzi %E Bonuccelli,M. %E Crescenzi,P. %E Petreschi,R. %K Computer %K Science %B Algorithms and ComplexityAlgorithms and Complexity %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 778 %P 22 - 32 %8 1994/// %@ 978-3-540-57811-6 %G eng %U http://dx.doi.org/10.1007/3-540-57811-0_3 %0 Conference Paper %B Proceedings of the tenth annual symposium on Computational geometry %D 1994 %T Query-sensitive ray shooting %A Mitchell,Joseph S.B. %A Mount, Dave %A Suri,Subhash %B Proceedings of the tenth annual symposium on Computational geometry %S SCG '94 %I ACM %C New York, NY, USA %P 359 - 368 %8 1994/// %@ 0-89791-648-4 %G eng %U http://doi.acm.org/10.1145/177424.178094 %R 10.1145/177424.178094 %0 Journal Article %J Digital Libraries' 94 %D 1994 %T QUEST—Query Environment for Science Teaching %A Shneiderman, Ben %A Rosenfeld, A. %A Marchioni,G. %A Holliday,W. %A Ricart, G. %A Faloutsos,С. %A Dick,J. %X QUery Environment for Science Teaching (QUEST) is a proposed digital library implementation consisting of a set of research projects dealing with data capture and organization, content analysis, information seeking and visual interfaces.The QUEST team includes a large number of renown technical collaborators and prominent source collaborators, as well as a significant number of contributors in the University of Maryland, the central, co-ordinating agency.A large collection of multidisciplinary materials in visual and textual formats, made accessible to us by our source collaborators, will be organized to allow integrated access by users from the science education community, that is elementary school through college level teachers. QUEST will be structured so as to provide seamless access to widespread resources on disparate subjects. We intend to provide first-rate subject analysis and representation in order to provide ready access. QUEST will be accessible nationally by means of Mosaic. We propose to provide highly sophisticated querying, browsing and information investigation facilities which will handle integrated textual and visual materials without difficulty. They will be augmented by online reference and referral services, immediately accessible by the user. QUEST will provide a comprehensive information resource for science education accessible through a dynamic, visual user interface. %B Digital Libraries' 94 %8 1994/// %G eng %0 Conference Paper %B Foundations of Computer Science, 1994 Proceedings., 35th Annual Symposium on %D 1994 %T Randomized and deterministic algorithms for geometric spanners of small diameter %A Arya,S. %A Mount, Dave %A Smid,M. %K computational geometry %K deletions %K deterministic algorithms %K directed graph %K directed graphs %K geometric spanners %K insertions %K randomised algorithms %K randomized algorithms %X Let S be a set of n points in IRd and let t gt;1 be a real number. A t-spanner for S is a directed graph having the points of S as its vertices, such that for any pair p and q of points there is a path from p to q of length at most t times the Euclidean distance between p and p. Such a path is called a t-spanner path. The spanner diameter of such a spanner is defined as the smallest integer D such that for any pair p and q of points there is a t-spanner path from p to q containing at most D edges. Randomized and deterministic algorithms are given for constructing t-spanners consisting of O(n) edges and having O(log n) diameter. Also, it is shown how to maintain the randomized t-spanner under random insertions and deletions. Previously, no results were known for spanners with low spanner diameter and for maintaining spanners under insertions and deletions %B Foundations of Computer Science, 1994 Proceedings., 35th Annual Symposium on %P 703 - 712 %8 1994/11// %G eng %R 10.1109/SFCS.1994.365722 %0 Journal Article %J Connecting the DOE community: partnerships in information %D 1994 %T The river beyond the rapids: responsive services for responsible users %A Shneiderman, Ben %X Providing responsive services to all members of our society will require new corporate alliances that put the users and their needs ahead of the pursuit of new technology. Responsive systems are the boats on which responsible users will travel to their work, social communities, and entertainment. With responsive systems users can take responsibility for what they promise to do and derive satisfaction from accomplishing their tasks. The keys to user responsibility are comprehensible, predictable, and controllable interfaces. %B Connecting the DOE community: partnerships in information %P 25 - 26 %8 1994/// %G eng %0 Journal Article %J Journal of Parallel and Distributed computing %D 1994 %T A simulation-based study on the concurrent execution of rules in a database environment %A Raschid, Louiqa %A Sellis,T. %A Delis,A. %B Journal of Parallel and Distributed computing %V 20 %P 20 - 42 %8 1994/// %G eng %N 1 %0 Journal Article %J International Journal of Computer Vision %D 1994 %T Space variant image processing %A Wallace,R.S. %A Ong,P.W. %A Bederson, Benjamin B. %A Schwartz,E.L. %B International Journal of Computer Vision %V 13 %P 71 - 90 %8 1994/// %G eng %N 1 %0 Journal Article %J ACM Transactions on Computer-Human Interaction %D 1994 %T Split menus: effectively using selection frequency to organize menus %A Sears,Andrew %A Shneiderman, Ben %K Human-computer interaction %K menus %K selection frequency %K split menus %K user interface %X When some items in a menu are selected more frequently than others, as is often the case, designers or individual users may be able to speed performance and improve preference ratings by placing several high-frequency items at the top of the menu. Design guidelines for split menus were developed and applied. Split menus were implemented and tested in two in situ usability studies and a controlled experiment. In the usability studies performance times were reduced by 17 to 58% depending on the site and menus. In the controlled experiment split menus were significantly faster than alphabetic menus and yielded significantly higher subjective preferences. A possible resolution to the continuing debate among cognitive theorists about predicting menu selection times is offered. We conjecture and offer evidence that, at least when selecting items from pull-down menus, a logarithmic model applies to familiar (high-frequency) items, and a linear model to unfamiliar (low-frequency) items. %B ACM Transactions on Computer-Human Interaction %V 1 %P 27 - 51 %8 1994/03// %@ 1073-0516 %G eng %U http://doi.acm.org/10.1145/174630.174632 %N 1 %R 10.1145/174630.174632 %0 Conference Paper %B Proceedings of the twenty-sixth annual ACM symposium on Theory of computing %D 1994 %T Symmetry breaking for suffix tree construction %A Sahinalp,S. C %A Vishkin, Uzi %B Proceedings of the twenty-sixth annual ACM symposium on Theory of computing %P 300 - 309 %8 1994/// %G eng %0 Journal Article %J Signal Processing, IEEE Transactions on %D 1994 %T URV ESPRIT for tracking time-varying signals %A Liu,K. J.R %A O'Leary, Dianne P. %A Stewart, G.W. %A Wu,Y.-J.J. %K arrival;narrowband %K decomposition;real %K directions %K ESPRIT;fixed %K estimation;real-time %K of %K processing;time-varying %K signals;performance;rank-revealing %K signals;tracking;direction-of-arrival %K systems;time-varying %K systems;tracking; %K TIME %K URV %X ESPRIT is an algorithm for determining the fixed directions of arrival of a set of narrowband signals at an array of sensors. Unfortunately, its computational burden makes it unsuitable for real time processing of signals with time-varying directions of arrival. The authors develop a new implementation of ESPRIT that has potential for real time processing. It is based on a rank-revealing URV decomposition, rather than the eigendecomposition or singular value decomposition used in previous ESPRIT algorithms. The authors demonstrate its performance on simulated data representing both constant and time-varying signals. They find that the URV-based ESPRIT algorithm is effective for estimating time-varying directions-of-arrival at considerable computational savings over the SVD-based algorithm %B Signal Processing, IEEE Transactions on %V 42 %P 3441 - 3448 %8 1994/12// %@ 1053-587X %G eng %N 12 %R 10.1109/78.340778 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1994 %T User Controlled Smooth Zooming for Information Visualization %A Jog,Ninog %A Shneiderman, Ben %K animation %K dynamic queries %K starfield display %K Systems Integration %K visualization zooming %K zoom bar %X This paper discusses the design and implementation of user controlled smooth zooming of a starfield display. A starfield display is a two dimensional graphical visualization of a multidimensional database where every item from the database is represented as a small colored rectangle whose position is determined by its ranking along ordinal attributes of the items laid out on the axes. One way of navigating this visual information is by using a zooming tool to incrementally zoom in on the items by varying the attribute range on either axis independently - such zooming causes the rectangles to move continuously and to grow or shrink. To get a feeling of flying through the data, users should be able to track the motion of each rectangle without getting distracted by flicker or large jumps - conditions that necessitate high display refresh rates and closely spaced rectangles on successive frames. Although the use of high-speed hardware can achieve the required visual effect for small databases, the twin software bottlenecks of rapidly accessing display items and constructing a new display image fundamentally retard the refresh rate. Our work explores several methods to overcome these bottlenecks, presents a taxonomy of various zooming methods and introduces a new widget, the zoom bar, that facilitates zooming.

%B Institute for Systems Research Technical Reports %8 1994/// %G eng %U http://drum.lib.umd.edu/handle/1903/5520 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: celebrating interdependence %D 1994 %T Visual information seeking: tight coupling of dynamic query filters with starfield displays %A Ahlberg,Christopher %A Shneiderman, Ben %K database query %K dynamic queries %K information seeking %K starfield displays %K tight coupling %B Proceedings of the SIGCHI conference on Human factors in computing systems: celebrating interdependence %S CHI '94 %I ACM %C New York, NY, USA %P 313 - 317 %8 1994/// %@ 0-89791-650-6 %G eng %U http://doi.acm.org/10.1145/191666.191775 %R 10.1145/191666.191775 %0 Journal Article %J ACM CHI94 Conference, Boston %D 1994 %T Visual information seeking using the FilmFinder %A Ahlberg,C. %A Shneiderman, Ben %B ACM CHI94 Conference, Boston %8 1994/// %G eng %0 Conference Paper %B Proceedings of the conference on Visualization '94 %D 1994 %T Visualization in the information highway %A Gershon,Nahum %A Card,Stuart K. %A Roth,Steven F. %A Shneiderman, Ben %B Proceedings of the conference on Visualization '94 %S VIS '94 %I IEEE Computer Society Press %C Los Alamitos, CA, USA %P 4 - 7 %8 1994/// %@ 0-7803-2521-4 %G eng %U http://dl.acm.org/citation.cfm?id=951087.951091 %0 Conference Paper %B Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on %D 1993 %T An adaptive ESPRIT based on URV decomposition %A Liu,K. J.R %A O'Leary, Dianne P. %A Stewart, G.W. %A Wu,Y.-J.J. %K algorithm;array %K complexity;parameter %K decomposition;real-time %K ESPRIT %K estimation;real-time %K filters;array %K of %K processing;adaptive %K processing;computational %K savings;performance;rank-revealing %K sensors;computational %K signal %K systems; %K systems;time-varying %K URV %K URV-based %X ESPRIT is an algorithm for determining the fixed directions of arrival of a set of narrowband signals at an array of sensors. Its computational burden makes it unsuitable for real-time processing of signals with time-varying directions of arrival. The authors develop a new implementation of ESPRIT that has potential for real-time processing. It is based on a rank-revealing URV decomposition, rather than the eigendecomposition or singular value decomposition (SVD) used in previous ESPRIT algorithms. Its performance is demonstrated on simulated data representing both constant and time-varying signals. It is shown that the URV-based ESPRIT algorithm is effective for estimating time-varying directions-of-arrival at considerable computational savings over the SVD-based algorithm.≪ETX≫ %B Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on %V 4 %P 37 -40 vol.4 - 37 -40 vol.4 %8 1993/04// %G eng %R 10.1109/ICASSP.1993.319588 %0 Journal Article %J IEEE Transactions on Knowledge and Data Engineering %D 1993 %T ADMS: a testbed for incremental access methods %A Roussopoulos, Nick %A Economou,N. %A Stamenas,A. %K Add-drop multiplexers %K ADMS %K advanced database management system %K client-server architecture %K commercial database management systems %K Computational modeling %K Database systems %K distributed databases %K heterogeneous DBMS %K incremental access methods %K incremental gateway %K Information retrieval %K interoperability %K join index %K large databases %K Navigation %K network operating systems %K Object oriented databases %K Object oriented modeling %K Query processing %K System testing %K very large databases %K view index %K Workstations %X ADMS is an advanced database management system developed-to experiment with incremental access methods for large and distributed databases. It has been developed over the past eight years at the University of Maryland. The paper provides an overview of ADMS, and describes its capabilities and the performance attained by its incremental access methods. This paper also describes an enhanced client-server architecture that allows an incremental gateway access to multiple heterogeneous commercial database management systems %B IEEE Transactions on Knowledge and Data Engineering %V 5 %P 762 - 774 %8 1993/10// %@ 1041-4347 %G eng %N 5 %R 10.1109/69.243508 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1993 %T Alpha Slider: Searching Textual Lists with Sliders %A Osada,M. %A Liao,Holmes %A Shneiderman, Ben %K direct manipulation %K Human-computer interaction %K keyboard data entry %K sliders %K Systems Integration %K touchscreen %X AlphaSlider is a query interface that uses a direct manipulation slider to select words, phrases, or names from an existing list. This paper introduces a prototype of AlphaSlider, describes the design issues, reports on an experimental evaluation, and offers directions for further research. The experiment tested 24 subjects selecting items from lists of 40, 80, 160, and 320 entries. Mean selection times only doubled with the 8-fold increase in list length. Users quickly accommodated to this selection method. %B Institute for Systems Research Technical Reports %8 1993/// %G eng %U http://drum.lib.umd.edu/handle/1903/5397 %0 Journal Article %J Ninth Symposium on Human Interface %D 1993 %T Alphaslider: Development and Evaluation of Text Retrieval Methods Using Sliders %A Osada,M. %A Liao,H. %A Shneiderman, Ben %X AlphaSlider is a query interface that uses a direct manipulation slider to select words, phrases, or names from an existing list. This paper introduces a prototype of AlphaSlider, describes the design issues, reports on an experimental evaluation, and offers directions for further research. The experiment tested 24 subjects selecting items from lists of 40, 80, 160, and 320 entries. Mean selection times only doubled with the 8-fold increase in list length. Users quickly accommodated to this selection method. %B Ninth Symposium on Human Interface %P 91 - 94 %8 1993/// %G eng %0 Conference Paper %B Scalable Parallel Libraries Conference, 1993., Proceedings of the %D 1993 %T Applying the CHAOS/PARTI library to irregular problems in computational chemistry and computational aerodynamics %A Das,R. %A Hwang,Y. S %A Uysal, M. %A Saltz, J. %A Sussman, Alan %B Scalable Parallel Libraries Conference, 1993., Proceedings of the %P 45 - 56 %8 1993/// %G eng %0 Journal Article %J SPIE Vision Geometry II %D 1993 %T On the area of overlap of translated polygons %A Mount, Dave %A Silverman,R. %A Wu,A. %X Given two simple polygons P and Q in the plane and a translation vector t 2R2, the area-of-overlap function of P and Q is the function Ar(t) = Area(P \(t + Q)), where t + Q denotes Q translated by t. This function has a number of applications in areas such as motion planning and object recognition. We present a number of mathematical results regarding this function. We also provide e cient algorithms for computing a representation of this function, and for tracing contour curves of constant area-of-overlap. %B SPIE Vision Geometry II %V 2060 %P 254 - 264 %8 1993/// %G eng %0 Journal Article %J IEEE Software %D 1993 %T Beyond intelligent machines: just do it %A Shneiderman, Ben %K Artificial intelligence %K Computer aided instruction %K Computer errors %K controllable interfaces %K Displays %K dynamic queries %K Educational institutions %K History %K HUMANS %K intelligent machines %K learning systems %K Machine intelligence %K predictable interfaces %K Speech recognition %K tree maps %K User interfaces %X The author argues that users want a sense of direct and immediate control over computers that differs from how they interact with people. He presents several examples of these predictable and controllable interfaces developed in the lab. The examples include tree maps and dynamic queries.<> %B IEEE Software %V 10 %P 100 - 103 %8 1993/01// %@ 0740-7459 %G eng %N 1 %R 10.1109/52.207235 %0 Conference Paper %B Proceedings of the 1993 ACM conference on Computer science %D 1993 %T Beyond the graphical user interface (abstract) %A Shneiderman, Ben %X The wildly popular graphical user interfaces are a dramatic improvement over earlier interaction styles, but the next generation of user interfaces is already being fashioned. The future will be dynamic, spatial, gestural, colorful, ubiquitous, often auditory, and sometimes virtual. The dominance of visual information with hi-res images and full-motion video will push the hardware requirements, absorb network capacity, and challenge the algorithm designers. As human-computer interaction researchers accumulate experimental data and theory matures, novel input devices, information visualization displays, and interaction strategies will emerge. %B Proceedings of the 1993 ACM conference on Computer science %S CSC '93 %I ACM %C New York, NY, USA %P 510– - 510– %8 1993/// %@ 0-89791-558-5 %G eng %U http://doi.acm.org/10.1145/170791.171139 %R 10.1145/170791.171139 %0 Conference Paper %B Proc. Ninth Japanese Symp. Human Interface %D 1993 %T Browsing Unix directories with dynamic queries: An analytical and experimental evaluation %A Liao,H. %A Osada,M. %A Shneiderman, Ben %X We designed, implemented, and evaluated an innovative concept for dynamic queries which involves the direct manipulation of small databases. Our domain was directories in a Unix file system. Dynamic queries allow users to formulate queries and explore the databases with graphical widgets, such as sliders and buttons, without requiring them to have any knowledge about the underlying structure of the database query languages, or command language syntax. Three interfaces for presenting directories were developed and tested with eighteen subjects in a within-subject design. The results of the formative evaluation yielded some useful guidelines for software designers. %B Proc. Ninth Japanese Symp. Human Interface %P 95 - 98 %8 1993/// %G eng %0 Journal Article %J Behaviour & Information Technology %D 1993 %T Building HCI partnerships and infrastructure %A Shneiderman, Ben %A LEWIS,CLAYTON %X As policymakers and technology planners respond to the growing activity in human?computer interaction, a broad perspective may be helpful. This article offers a top-down view of current activities and suggests opportunities and challenges for the continued growth of HCI. Partnerships among universities, corporations, government agencies, and professional societies are proposed. Infrastructure needs to support this new discipline are outlined.Abstract As policymakers and technology planners respond to the growing activity in human?computer interaction, a broad perspective may be helpful. This article offers a top-down view of current activities and suggests opportunities and challenges for the continued growth of HCI. Partnerships among universities, corporations, government agencies, and professional societies are proposed. Infrastructure needs to support this new discipline are outlined. %B Behaviour & Information Technology %V 12 %P 130 - 135 %8 1993/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449299308924373 %N 2 %R 10.1080/01449299308924373 %0 Journal Article %J Communications of the ACM %D 1993 %T Building Partnerships and Infrastructure %A Shneiderman, Ben %A Lewis,C. %X As policymakers and technology planners respond to the growingactivity in human-computer interaction, a broad perspective may be helpful. This article offers a top-down view of current activities and suggests opportunities and challenges for the continued growth of HCI. Partnerships among universities, corporations, government agencies, and professional societies are proposed. Infrastructure needs to support this new discipline are outlined. %B Communications of the ACM %V 32 %P 9 - 23 %8 1993/// %G eng %N 1 %0 Conference Paper %B Proceedings of the fourth annual ACM-SIAM Symposium on Discrete algorithms %D 1993 %T Chernoff-Hoeffding bounds for applications with limited independence %A Schmidt,Jeanette P. %A Siegel,Alan %A Srinivasan, Aravind %B Proceedings of the fourth annual ACM-SIAM Symposium on Discrete algorithms %S SODA '93 %I Society for Industrial and Applied Mathematics %C Philadelphia, PA, USA %P 331 - 340 %8 1993/// %@ 0-89871-313-7 %G eng %U http://dl.acm.org/citation.cfm?id=313559.313797 %0 Journal Article %J ACM Transactions on Programming Languages and Systems %D 1993 %T The concurrency workbench: a semantics-based tool for the verification of concurrent systems %A Cleaveland, Rance %A Parrow,Joachim %A Steffen,Bernhard %B ACM Transactions on Programming Languages and Systems %V 15 %P 36 - 72 %8 1993/01/01/ %@ 01640925 %G eng %U http://dl.acm.org/citation.cfm?id=151648 %R 10.1145/151646.151648 %0 Conference Paper %B , 1993 IEEE International Conference on Robotics and Automation, 1993. Proceedings %D 1993 %T Control and design of the spherical pointing motor %A Bederson, Benjamin B. %A Wallace,R.S. %A Schwartz,E.L. %K absolute positioning device %K Actuators %K camera sensor %K CAMERAS %K Coils %K electric actuators %K electric motors %K Magnetic field measurement %K Magnetic sensors %K Open loop systems %K open-loop control %K orthogonal coils %K pan-tilt actuator %K permanent magnet %K Permanent magnet motors %K Permanent magnets %K position control %K Scanning probe microscopy %K second order system %K spherical pointing motor %K velocity control %K Velocity measurement %K velocity measurements %X A miniature pan-tilt actuator, the spherical pointing motor (SPM) has been built. The SPM is an absolute positioning device, designed to orient a small camera sensor in two degrees-of-rotational-freedom. It is 4 cm×5 cm×6 cm and weighs 160 g. The operating principle of the SPM is to orient a permanent magnet to the magnetic field induced by three orthogonal coils by applying the appropriate ratio of currents to the coils. The open-loop control and design of the SPM are described. Physical characteristics and velocity measurements of the prototype are reported. The SPM is modeled with a simple second order system. Without control, it oscillates substantially when a single impulse is applied. By applying a second impulse without feedback, the oscillation can be reduced by as much as a factor of ten %B , 1993 IEEE International Conference on Robotics and Automation, 1993. Proceedings %I IEEE %P 630-636 vol.2 - 630-636 vol.2 %8 1993/05/02/6 %@ 0-8186-3450-2 %G eng %R 10.1109/ROBOT.1993.291888 %0 Journal Article %J Knowledge and Data Engineering, IEEE Transactions on %D 1993 %T Coupling production systems and database systems: A homogeneous approach %A Sellis,T. %A Lin,C. C %A Raschid, Louiqa %B Knowledge and Data Engineering, IEEE Transactions on %V 5 %P 240 - 256 %8 1993/// %G eng %N 2 %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Declaration in Apple vs. Microsoft/Hewlett-Packard %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 355 - 355 %8 1993/// %G eng %0 Conference Paper %B , Fifth International Conference on Computing and Information, 1993. Proceedings ICCI '93 %D 1993 %T A distributed algorithm for ear decomposition %A Hannenhalli, Sridhar %A Perumalla,K. %A Chandrasekharan,N. %A Sridhar,R. %K Asynchronous communication %K asynchronous communication network %K Automata %K Communication networks %K computational complexity %K Computer networks %K Computer science %K decomposition graph %K distributed algorithm %K distributed algorithms %K Distributed computing %K Ear %K ear decomposition %K graph theory %K message-optimal %K network decomposition %K sorting %K Testing %K time-optimal %X A distributed algorithm for finding an ear decomposition of an asynchronous communication network with n nodes and m links is presented. At the completion of the algorithm either the ears are correctly labeled or the nodes are informed that there exists no ear decomposition. First we present a novel algorithm to check the existence of an ear decomposition which uses O(m) messages. We also present two other algorithms, one which is time-optimal and the other which is message-optimal to determine the actual ears and their corresponding numbers after determining the existence of an ear decomposition %B , Fifth International Conference on Computing and Information, 1993. Proceedings ICCI '93 %I IEEE %P 180 - 184 %8 1993/05/27/29 %@ 0-8186-4212-2 %G eng %R 10.1109/ICCI.1993.315382 %0 Journal Article %J Institute for Systems Research Technical Reports %D 1993 %T Dynamic Queries: A step Beyond Database Languages %A Shneiderman, Ben %K human-computer interaction database query, Systems Integration %X The capacity to incrementally adjust a query (with sliders, buttons, selections from a set of discrete attribute values, etc.) coupled with a visual display of results that are rapidly updated, dramatically changes the information seeking process. Dynamic queries on the chemical table of elements, computer directions, and a real estate database were built and tested in three separate exploratory experiments. Preliminary results show highly significant performance improvements and user enthusiasm more commonly seen with video games. Widespread application seems possible but research issues abound in the areas of: (1) graphic visualization design, (2) database and display algorithms, and (3) user interface requirements. Challenges include methods for rapidly displaying and changing many points, colors, and areas; multi-dimensional pointing and exploring using 6 degree of freedom input/output devices; incorporation of sound and visual display techniques that increase user comprehension; and integration with existing database systems. %B Institute for Systems Research Technical Reports %8 1993/// %G eng %U http://drum.lib.umd.edu/handle/1903/5350 %0 Journal Article %J Educational Multimedia and Hypermedia Annual. Association for the Advancement of Computing in Education, Charlottesville, Virg %D 1993 %T Education by engagement and construction: Experiences in the AT&T teaching theater %A Shneiderman, Ben %B Educational Multimedia and Hypermedia Annual. Association for the Advancement of Computing in Education, Charlottesville, Virg %P 471 - 479 %8 1993/// %G eng %0 Journal Article %J Journal of Computing in Higher Education %D 1993 %T Engagement and construction: Educational strategies for the post-TV era %A Shneiderman, Ben %X THE POST-TV MEDIA of computers (hypermedia, graphic user interfaces, word processors, spreadsheets, databases, sound, music, animation, etc.) and communications (electronic mail, bulletin boards, groupware, fax, video, etc.) enables teachers, students, and parents to creatively develop education by engagement and construction. Students should be given the chance to engage with each other in team projects, preferably situated in the world outside the classroom, with the goal of constructing a product that is useful or interesting to someone other than the teacher. Challenges remain such as scaling up from small class projects to lecture sections with hundreds of students, covering the curriculum that is currently required by many school districts, and evaluating performance. However, there seems to be no turning back and, anyway, the children of the Nintendo and Video Age are eager to press fast forward. %B Journal of Computing in Higher Education %V 4 %P 106 - 116 %8 1993/// %@ 1042-1726 %G eng %U http://dx.doi.org/10.1007/BF02941067 %N 2 %0 Book Section %B Sparks of innovation in human-computer interaction %D 1993 %T Evaluating three museum installations of a hypertext system %A Shneiderman, Ben %A Brethauer,D. %A Plaisant, Catherine %A Potter,R. %B Sparks of innovation in human-computer interaction %I Intellect Books %V 40 %P 404 - 404 %8 1993/// %@ 9781567500783 %G eng %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Finding facts vs. browsing knowledge in hypertext systems %A Marchionini,G. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 103 - 103 %8 1993/// %G eng %0 Journal Article %J Communications of the ACM %D 1993 %T First steps toward positive contributions %A Shneiderman, Ben %B Communications of the ACM %V 36 %P 15 - 16 %8 1993/// %G eng %N 1 %0 Journal Article %J Journal of the American Society for Information Science %D 1993 %T A graphical filter/flow representation of Boolean queries: A prototype implementation and evaluation %A Young,Degi %A Shneiderman, Ben %X One of the powerful applications of Boolean expression is to allow users to extract relevant information from a database. Unfortunately, previous research has shown that users have difficulty specifying Boolean queries. In an attempt to overcome this limitation, a graphical Filter/Flow representation of Boolean queries was designed to provide users with an interface that visually conveys the meaning of the Boolean operators (AND, OR, and NOT). This was accomplished by implementing a graphical interface prototype that uses the metaphor of water flowing through filters. Twenty subjects having no experience with Boolean logic participated in an experiment comparing the Boolean operations represented in the Filter/Flow interface with a text-only SQL interface. The subjects independently performed five comprehension tasks and five composition tasks in each of the interfaces. A significant difference (p < 0.05) in the total number of correct queries in each of the comprehension and composition tasks was found favoring Filter/Flow. © 1993 John Wiley & Sons, Inc. %B Journal of the American Society for Information Science %V 44 %P 327 - 339 %8 1993/07/01/ %@ 1097-4571 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/(SICI)1097-4571(199307)44:6%3C327::AID-ASI3%3E3.0.CO;2-J/abstract;jsessionid=76B714851135D8044D4E03335B4D39A0.d01t04 %N 6 %R 10.1002/(SICI)1097-4571(199307)44:6<327::AID-ASI3>3.0.CO;2-J %0 Journal Article %J Information Systems %D 1993 %T A graphical query interface based on aggregation/generalization hierarchies %A Weiland,William J %A Shneiderman, Ben %K aggregation/generalization hierarchies %K empirical test %K Graphical user interfaces %K visual queries %X In order for automated information systems to be used effectively, they must be made easily accessible to a wide range of users and with short training periods. This work proposes a method of organizing documents based on the concepts of aggregation and generalization hierarchies. We propose a graphical user interface to provide a more intuitive form of Boolean query. This design is based on mapping the nodes of the aggregation hierarchy to Boolean intersection operations, mapping the nodes of the generalization hierarchy to Boolean union operations, and providing a concrete, graphical, manipulable representation of both of these node types. Finally, a working prototype interface was constructed and evaluated experimentally against a classical command-line Boolean query interface. In this formative evaluation with sixteen subjects, the graphical interface produced less than one-tenth the errors of the textual interface, on average. Significant differences in time spent specifying queries were not found. Observations and comments provide guidance for designers. %B Information Systems %V 18 %P 215 - 232 %8 1993/06// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/030643799390018V %N 4 %R 10.1016/0306-4379(93)90018-V %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Improving the accuracy of touchscreens: an experimental evaluation of three strategies %A Potter,R.L. %A Weldon,L.J. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 161 - 161 %8 1993/// %G eng %0 Journal Article %J Behaviour & Information Technology %D 1993 %T Investigating touchscreen typing: the effect of keyboard size on typing speed %A Sears,Andrew %A REVIS,DOREEN %A SWATSKI,JANET %A CRITTENDEN,ROB %A Shneiderman, Ben %X Abstract Two studies investigated the effect keyboard size has on typing speed and error rates for touchscreen keyboards using the lift-off strategy. A cursor appeared when users touched the screen and a key was selected when they lifted their finger from the screen. Four keyboard sizes were investigated ranging from 24.6 cm to 6.8 cm wide. Results indicate that novices can type approximately 10 words per minute (WPM) on the smallest keyboard and 20 WPM on the largest. Experienced users improved to 21 WPM on the smallest keyboard and 32 WPM on the largest. These results indicate that, although slower, small touchscreen keyboards can be used for limited data entry when the presence of a regular keyboard is not practical. Applications include portable pocket-sized or palmtop computers, messaging systems, and personal information resources. Results also suggest the increased importance of experience on these smaller keyboards. Research directions are suggested.Abstract Two studies investigated the effect keyboard size has on typing speed and error rates for touchscreen keyboards using the lift-off strategy. A cursor appeared when users touched the screen and a key was selected when they lifted their finger from the screen. Four keyboard sizes were investigated ranging from 24.6 cm to 6.8 cm wide. Results indicate that novices can type approximately 10 words per minute (WPM) on the smallest keyboard and 20 WPM on the largest. Experienced users improved to 21 WPM on the smallest keyboard and 32 WPM on the largest. These results indicate that, although slower, small touchscreen keyboards can be used for limited data entry when the presence of a regular keyboard is not practical. Applications include portable pocket-sized or palmtop computers, messaging systems, and personal information resources. Results also suggest the increased importance of experience on these smaller keyboards. Research directions are suggested. %B Behaviour & Information Technology %V 12 %P 17 - 22 %8 1993/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449299308924362 %N 1 %R 10.1080/01449299308924362 %0 Journal Article %J Formal methods in system design %D 1993 %T A linear-time model-checking algorithm for the alternation-free modal mu-calculus %A Cleaveland, Rance %A Steffen,B. %B Formal methods in system design %V 2 %P 121 - 147 %8 1993/// %G eng %N 2 %0 Journal Article %J Proceedings of the Fifth Canadian Conference on Computational Geometry, Waterloo, Ontario %D 1993 %T On The Maximum Number of Intersections of Two Polyhedra in 2 and 3 Dimensions %A Dillencourt,M. B %A Mount, Dave %A Saalfeld,A.J. %X We obtain exact bounds on the maximum number of in-tersections between two polygons with prescribed edge counts. In three dimensions, we obtain exact bounds on the number of intersecting face-pairs of two polyhe- dra with prescribed face counts when the polyhedra are convex, and lower bounds in the general case. %B Proceedings of the Fifth Canadian Conference on Computational Geometry, Waterloo, Ontario %P 49 - 54 %8 1993/// %G eng %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T A nonanthropomorphic style guide: overcoming the Humpty Dumpty syndrome %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 331 - 331 %8 1993/// %G eng %0 Journal Article %J Journal of Algorithms %D 1993 %T Optimal doubly logarithmic parallel algorithms based on finding all nearest smaller values %A Berkman,O. %A Schieber,B. %A Vishkin, Uzi %B Journal of Algorithms %V 14 %P 344 - 370 %8 1993/// %G eng %N 3 %0 Journal Article %J Algorithms and Data Structures %D 1993 %T Point probe decision trees for geometric concept classes %A Arkin,E. %A Goodrich,M. %A Mitchell,J. %A Mount, Dave %A Piatko,C. %A Skiena,S. %X A fundamental problem in model-based computer vision is that of identifying to which of a given set of concept classes of geometric models an observed model belongs. Considering a ldquoproberdquo to be an oracle that tells whether or not the observed model is present at a given point in an image, we study the problem of computing efficient strategies (ldquodecision treesrdquo) for probing an image, with the goal to minimize the number of probes necessary (in the worst case) to determine in which class the observed model belongs. We prove a hardness result and give strategies that obtain decision trees whose height is within a log factor of optimal.These results grew out of discussions that began in a series of workshops on Geometric Probing in Computer Vision, sponsored by the Center for Night Vision and Electro-Optics, Fort Belvoir, Virginia, and monitored by the U.S. Army Research Office. The views, opinions, and/or findings contained in this report are those of the authors and should not be construed as an official Department of the Army position, policy, or decision, unless so designated by other documentation. %B Algorithms and Data Structures %P 95 - 106 %8 1993/// %G eng %R 10.1007/3-540-57155-8_239 %0 Conference Paper %B Proceedings of the 1993 ACM conference on Computer science %D 1993 %T Potentials and limitations of pen-based computers %A Citrin,Wayne %A Halbert,Dan %A Hewitt,Carl %A Meyrowitz,Norm %A Shneiderman, Ben %X There are four possible genres of input devices that can be attached to personal workstations; keyboard, mouse, pen, and voice. For investigating potentials and limitations of pen-based computers, we propose to compare those four categories as different types of man-machine communication channel.Even though arrow keys allow a limited scope of 2D capability in keyboard usage, the primary use of a keyboard is typing. Typing generates a linear sequence of discrete characters. The maximum speed of typing is roughly 10 characters per second, hence the bandwidth of 100 (10×10) bps. A mouse is used primarily for pointing (menu item/object selection) and then for dragging (moving and re-sizing objects). Mouse pointing generates a discrete information of a point location on a 2D (planer) plane with the maximum speed of 2 clicking per second. Each pointing may generate 20 bits of information with the bandwidth of 40 (2×20) bps. Mouse dragging generates a continuous geometric pattern in 2D. Assuming the maximum rate of 40 selections per second in the eight (3 bits) possible directions of dragging, the bandwidth of mouse dragging peaks at 120 (40×3) bps. The usage of pen on a planer flat surface of LCD unit can be divided into scribing and tapping. By scribing we mean a generation of continuous pen strokes forming a character, a gesture, or a picture. Scribing includes drawing and gesturing. Tapping corresponds to a mouse clicking. Tapping may be considered as a special kind of gesture as in the PenPoint operating system. The bandwidth of scribing can be calculated in the same manner as for mouse dragging. With the maximum rate of 100 selections of direction per second for pen, scribing may produce strokes with the speed of 300 (100×3) bps. The bandwidth of pen tapping is almost the same as that of a mouse clicking except that a selection of a point is easier by a pen (3 tappings per second) than by a mouse (2 clickings per second). Talking through a microphone generates a linear sequence of continuous speech with a high degree of redundancy. Using the CELP speech compression algorithm, the bandwidth of normal speech can be reduced to 4800 bps. By vocal signaling we mean a generation of a sequence of discrete messages each of which consists of different pitches and loudness levels. The maximum rate of signaling could be 5 messages per second with 10 differentiable pitches and 10 levels of loudness producing the bandwidth of 35 (5×7) bps. Note that mouse dragging, pen scribing, and voice talking, each produces continuous data objects. Only after quantization by sampling, the data objects can be represented by discrete data structure. By precision we mean the degree of ease in duplicating the identical information using the same input technology. Keyboard is a high precision device because there is no difficulty in generating the same character over and over again. Drawing by a mouse is more difficult than drawing by a pen, because pen is easier to control than mouse. Precision of voice is low because it is difficult to duplicate the sound of the same pitch and the same volume. By latency we mean the set-up time necessary to start generating a stream of information. The latency of using keyboard and mouse is larger than the latency of using pen and voice. By translation we mean a process of converting the information generated by the input device into a sequence of discrete symbols, i.e., a transduction of a continuous data type to a discrete data type. Translation for a keyboard is not necessary. Translation of a mouse click on a menu item requires a finite table look-up which is rather simple operation. Translation of pen scribing and mouse dragging involves a handwriting recognition algorithm which is still a difficult problem at present time. Voice recognition is a very difficult problem. With the assumption that real-time translation is feasible for handwriting recognition and speech recognition, the efficiency of input device for text entry can be measured by how many characters can be entered in a second (cps). A simulated keyboard on CRT is used for entering text by a mouse. Long handwriting on a pen computer is used for a pen. When personal workstations become down-sized, the physical dimension of an input/output device becomes a dominant factor for the mobility of workstations. Keyboard and mouse are portable but intrusive. Wireless pen is mobile and less intrusive. Voice can be ubiquitous but intrusive. One conclusion we can draw from the above analysis is that pen is mightier than mouse. A pen can replace a mouse any time any place. However, keyboard, pen, and voice have different strong points and weak points. They compensate with each other. Therefore, we predict that future workstations will carry multi-modal user-interface with any combination of keyboard, pen, and voice. %B Proceedings of the 1993 ACM conference on Computer science %S CSC '93 %I ACM %C New York, NY, USA %P 536 - 539 %8 1993/// %@ 0-89791-558-5 %G eng %U http://doi.acm.org/10.1145/170791.171171 %R 10.1145/170791.171171 %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Preface to Sparks of Innovation in Human-Computer Interaction %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %8 1993/// %G eng %0 Journal Article %J Systems, Man and Cybernetics, IEEE Transactions on %D 1993 %T Probabilistic analysis of some navigation strategies in a dynamic environment %A Sharma, R. %A Mount, Dave %A Aloimonos, J. %B Systems, Man and Cybernetics, IEEE Transactions on %V 23 %P 1465 - 1474 %8 1993/// %G eng %N 5 %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Protecting rights in user interface designs %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 351 - 351 %8 1993/// %G eng %0 Book Section %B Sparks of innovation in human-computer interactionSparks of innovation in human-computer interaction %D 1993 %T Remote direct manipulation: A case study of a telemedicine workstation %A Keil-Slawik,R. %A Plaisant, Catherine %A Shneiderman, Ben %B Sparks of innovation in human-computer interactionSparks of innovation in human-computer interaction %I Intellect Books %P 51 - 51 %8 1993/// %@ 9781567500783 %G eng %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Restructuring knowledge for an electronic encyclopedia %A Kreitzberg,C.B. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %V 615 %P 123 - 123 %8 1993/// %G eng %0 Book Section %B Sparks of innovation in human-computer interactionSparks of innovation in human-computer interaction %D 1993 %T Scheduling home control devices: a case study of the transition from the research project to a product %A Plaisant, Catherine %A Shneiderman, Ben %A Battagtia,J. %B Sparks of innovation in human-computer interactionSparks of innovation in human-computer interaction %I Intellect Books %P 205 - 205 %8 1993/// %@ 9781567500783 %G eng %0 Book %D 1993 %T Sparks of innovation in human-computer interaction %A Shneiderman, Ben %K Computers / Interactive & Multimedia %K Computers / Social Aspects / Human-Computer Interaction %K Human-computer interaction %X These key papers from the 10th anniversary of the Human-Computer Interaction Laboratory (HCIL) at the University of Maryland, exemplify different research methodologies, and show the maturation of human-computer interaction research. The first section introduces how HCIL does what they do, including some of their failures and background stories that are not appropriate for journal papers. This book is a tribute to the faculty, staff, visitors, and students who have shared in a decade of work. %I Intellect Books %8 1993/// %@ 9781567500783 %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1993 %T Speech versus mouse commands for word processing: an empirical evaluation %A Karl,Lewis R. %A Pettey,Michael %A Shneiderman, Ben %X Despite advances in speech technology, human factors research since the late 1970s has provided only weak evidence that automatic speech recognition devices are superior to conventional input devices such as keyboards and mice. However, recent studies indicate that there may be advantages to providing an additional input channel based on speech input to supplement the more common input modes. Recently the authors conducted an experiment to demonstrate the advantages of using speech-activated commands over mouse-activated commands for word processing applications when, in both cases, the keyboard is used for text entry and the mouse for direct manipulation. Sixteen experimental subjects, all professionals and all but one novice users of speech input, performed four simple word processing tasks using both input groups in this counterbalanced experiment. Performance times for all tasks were significantly faster when using speech to activate commands as opposed to using the mouse. On average, the reduction in task time due to using speech was 18·7%. The error rates due to subject mistakes were roughly the same for both input groups, and recognition errors, averaged over all the tasks, occurred for 6·3% of the speech-activated commands. Subjects made significantly more memorization errors when using speech as compared with the mouse for command activation. Overall, the subjects reacted positively to using speech input and preferred it over the mouse for command activation; however, they also voiced concerns about recognition accuracy, the interference of background noise, inadequate feedback and slow response time. The authors believe that the results of the experiment provide evidence for the utility of speech input for command activation in application programs. %B International Journal of Man-Machine Studies %V 39 %P 667 - 687 %8 1993/10// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020737383710783 %N 4 %R 10.1006/imms.1993.1078 %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T A study of file manipulation by novices using commands vs, direct manipulation %A Margono,S. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 39 - 39 %8 1993/// %G eng %0 Journal Article %J Computer Science Technical Reports, Cornell University %D 1993 %T Techniques for probabilistic analysis and randomness-efficient computation %A Srinivasan, Aravind %X Randomness is well-recognized as an important computational resource in theoretical computer science. Not only are there classical problems for which the only known "efficient" solutions are randomized, but there are problems for which randomness can be used to circumvent deterministic lower bounds. However, standard tools available for probabilistic analysis such as the Chernoff-Hoeffding bounds are not sufficiently powerful in all situations; second, since there are no known "true" sources of randomness, there is a need to develop general techniques for reducing/removing randomness from randomized algorithms. This thesis addresses these two issues. Certain types of scheduling and resource allocation problems in a distributed setting can be modeled as edge coloring problems. In Chapter 2, we present a fast and simple randomized algorithm for edge coloring a graph, in the synchronous distributed point-to-point model of computation. Our algorithm computes an edge-coloring of a graph $G$ with $n$ nodes and maximum degree $\Delta$ with at most $(1.6 + \epsilon) \Delta + O(\log^{2+\delta} n)$ colors with high probability, for any fixed $\epsilon, \delta, greater than 0$. To analyze our algorithm, we introduce techniques for proving upper bounds on the tails of random variables, extending the Chernoff-Hoeffding (CH) bounds to some types of dependent random variables. This is joint work with A. Panconesi [91]. In Chapter 3, we show that the CH bounds for the tails of the sums of bounded and independent random variables $X_{1}, \ldots, X_{n}$ require only limited independence among the $X_{i}$s. We show that if $X_{1}, \ldots, X_{n}$ lie in [0,1] and are $k$-wise independent, then $Pr(X \geq E[\sum_{i}X_{i}](1 + \delta))$ can be upper bounded by the CH bound, if $k \geq \mu \cdot min\{\delta,\delta^{2}\}$. This leads to algorithms that require a reduced amount of randomness for any analysis which uses the CH bounds. This is joint work with J.P. Schmidt and A. Siegel [108]. In Chapter 4, we present an $RNC^{2}$ algorithm for the perfect matching problem which uses $O(\log Z)$ random bits where $Z$ is any given upper bound on the number of perfect matchings in the graph, generalizing results of Grigoriev and Karpinski. Underlying our algorithm is a randomness-optimal generalization of the Isolating Lemma of Mulmuley, Vazirani and Vazirani, which also leads to other applications. This is joint work with S. Chari and P. Rohatgi [26]. %B Computer Science Technical Reports, Cornell University %8 1993/08//undefin %G eng %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Touchscreens now offer compelling uses %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 187 - 187 %8 1993/// %G eng %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T Treemaps %A Johnson,B. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %V 284 %P 309 - 309 %8 1993/// %G eng %0 Journal Article %J Sparks of innovation in human-computer interaction %D 1993 %T User interface consistency: an evaluation of original and revised interfaces for a videodisk library %A Chimera,R. %A Shneiderman, Ben %B Sparks of innovation in human-computer interaction %P 259 - 259 %8 1993/// %G eng %0 Journal Article %J The VLDB JournalVLDB Journal %D 1993 %T Using differential technlques to efficiently support transaction time %A Jensen,Christian S. %A Mark,Leo %A Roussopoulos, Nick %A Sells,Timos %B The VLDB JournalVLDB Journal %V 2 %P 75 - 111 %8 1993/01// %@ 1066-8888 %G eng %U http://dl.acm.org/citation.cfm?id=615162 %N 1 %R 10.1007/BF01231799 %0 Conference Paper %B Parallel Processing, 1993. ICPP 1993. International Conference on %D 1993 %T Using Synthetic-Perturbation Techniques for Tuning Shared Memory Programs %A Snelick,Robert %A JaJa, Joseph F. %A Kacker,Raghu %A Lyon,Gordon %X The Synthetic-Perturbation Tuning (SPT) methodology is base d on compirical approach that introduces artificial delays into the MIMD program and captures the effects of such delays by using the modrn branch statistics called design of experiments. %B Parallel Processing, 1993. ICPP 1993. International Conference on %V 2 %P 2 - 10 %8 1993/08// %G eng %R 10.1109/ICPP.1993.184 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 1992 %T Anthropomorphism: from Eliza to Terminator 2 %A Don,Abbe %A Brennan,Susan %A Laurel,Brenda %A Shneiderman, Ben %B Proceedings of the SIGCHI conference on Human factors in computing systems %S CHI '92 %I ACM %C New York, NY, USA %P 67 - 70 %8 1992/// %@ 0-89791-513-5 %G eng %U http://doi.acm.org/10.1145/142750.142760 %R 10.1145/142750.142760 %0 Journal Article %J Proceedings, DARPA Image Understanding Workshop %D 1992 %T Behavioral Visual Motion Analysis," %A Aloimonos, J. %A Duriç,Z. %A Fermüller, Cornelia %A Huang, L. %A Rivlin,E. %A Sharma, R. %B Proceedings, DARPA Image Understanding Workshop %P 521 - 541 %8 1992/// %G eng %0 Journal Article %J Neural Computation %D 1992 %T A competitive distribution theory of neocortical dynamics %A Reggia, James A. %A D'Autrechy,C. L %A Sutton III,G. G. %A Weinrich,M. %B Neural Computation %V 4 %P 287 - 317 %8 1992/// %G eng %N 3 %0 Journal Article %J Systems, Man and Cybernetics, IEEE Transactions on %D 1992 %T Coordinated motion planning: the warehouseman's problem with constraints on free space %A Sharma, R. %A Aloimonos, J. %B Systems, Man and Cybernetics, IEEE Transactions on %V 22 %P 130 - 141 %8 1992/// %G eng %N 1 %0 Journal Article %J ACM SIGCHI Bulletin %D 1992 %T DISSERTATIONS: ABSTRACTS OF INTEREST %A Humphrey,Susanne M. %A Shneiderman, Ben %X The following abstracts were selected from a computer search using the BRS Information Technologies retrieval services of the Dissertation Abstracts International (DAI) database produced by University Microfilms International.Unless otherwise specified, paper or microform copies of dissertations may be ordered, using the UM order number, from University Microfilms International, Dissertation Copies, Post Office Box 1794, Ann Arbor, MI 488106; telephone for U.S. (except Michigan, Hawaii, or Alaska): 800/521-3042, for Canada: 800/268-6090. Price lists, ordering, and shipping information are in the introduction to the published DAI. An alternative source for copies is sometimes provided at the end of the abstract.The dissertation titles and abstracts contained here are published with permission of University Microfilms International, publishers of Dissertation Abstracts International (copyright 1985) by University Microfilms International), and may not be reproduced without their prior permission. %B ACM SIGCHI Bulletin %V 24 %P 58 - 75 %8 1992/04// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/142386.1055546 %N 2 %R 10.1145/142386.1055546 %0 Conference Paper %B Proceedings of the 15th annual international ACM SIGIR conference on Research and development in information retrieval %D 1992 %T The Dynamic HomeFinder: Evaluating dynamic queries in a real-estate information exploration system %A Williamson,C. %A Shneiderman, Ben %B Proceedings of the 15th annual international ACM SIGIR conference on Research and development in information retrieval %P 338 - 346 %8 1992/// %G eng %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 1992 %T Dynamic queries: database searching by direct manipulation %A Shneiderman, Ben %A Williamson,Christopher %A Ahlberg,Christopher %B Proceedings of the SIGCHI conference on Human factors in computing systems %S CHI '92 %I ACM %C New York, NY, USA %P 669 - 670 %8 1992/// %@ 0-89791-513-5 %G eng %U http://doi.acm.org/10.1145/142750.143082 %R 10.1145/142750.143082 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 1992 %T Dynamic queries for information exploration: an implementation and evaluation %A Ahlberg,Christopher %A Williamson,Christopher %A Shneiderman, Ben %X We designed, implemented and evaluated a new concept for direct manipulation of databases, called dynamic queries, that allows users to formulate queries with graphical widgets, such as sliders. By providing a graphical visualization of the database and search results, users can find trends and exceptions easily. Eighteen undergraduate chemistry students performed statistically significantly faster using a dynamic queries interface compared to two interfaces both providing form fill-in as input method, one with graphical visualization output and one with all-textual output. The interfaces were used to explore the periodic table of elements and search on their properties. %B Proceedings of the SIGCHI conference on Human factors in computing systems %S CHI '92 %I ACM %C New York, NY, USA %P 619 - 626 %8 1992/// %@ 0-89791-513-5 %G eng %U http://doi.acm.org/10.1145/142750.143054 %R 10.1145/142750.143054 %0 Book Section %B Sociomedia: Hypermedia, Multimedia and the Social Creation of KnowledgeSociomedia: Hypermedia, Multimedia and the Social Creation of Knowledge %D 1992 %T Education by engagement and construction: a strategic education initiative for a multimedia renewal of American education %A Shneiderman, Ben %B Sociomedia: Hypermedia, Multimedia and the Social Creation of KnowledgeSociomedia: Hypermedia, Multimedia and the Social Creation of Knowledge %P 13 - 26 %8 1992/// %G eng %0 Conference Paper %B Foundations of Computer Science, 1992. Proceedings., 33rd Annual Symposium on %D 1992 %T Efficient minimum cost matching using quadrangle inequality %A Aggarwal,A. %A Bar-Noy,A. %A Khuller, Samir %A Kravets,D. %A Schieber,B. %K algorithm; %K array; %K bipartite %K bitonic %K blue %K complexity; %K computational %K cost %K distance; %K Euclidean %K function; %K geometry; %K graph %K graphs; %K inequality; %K linear %K MATCHING %K matching; %K minimisation; %K minimum %K Monge %K perfect %K points; %K polynomial %K problem; %K quadrangle %K red %K theory; %K TIME %K transportation %K transportation; %K weakly %X The authors present efficient algorithms for finding a minimum cost perfect matching, and for solving the transportation problem in bipartite graphs, G = (Red cup; Blue, Red times; Blue), where |Red| = n, |Blue| = m, n les; m, and the cost function obeys the quadrangle inequality. The first results assume that all the red points and all the blue points lie on a curve that is homeomorphic to either a line or a circle and the cost function is given by the Euclidean distance along the curve. They present a linear time algorithm for the matching problem. They generalize the method to solve the corresponding transportation problem in O((m+n)log(m+n)) time. The next result is an O(n log m) algorithm for minimum cost matching when the cost array is a bitonic Monge array. An example of this is when the red points lie on one straight line and the blue points lie on another straight line (that is not necessarily parallel to the first one). Finally, they provide a weakly polynomial algorithm for the transportation problem in which the associated cost array is a bitonic Monge array %B Foundations of Computer Science, 1992. Proceedings., 33rd Annual Symposium on %P 583 - 592 %8 1992/10// %G eng %R 10.1109/SFCS.1992.267793 %0 Conference Paper %B Tools with Artificial Intelligence, 1992. TAI'92, Proceedings., Fourth International Conference on %D 1992 %T Experiments on the concurrent rule execution in database systems %A Delis,A. %A Raschid, Louiqa %A Sellis,T. %B Tools with Artificial Intelligence, 1992. TAI'92, Proceedings., Fourth International Conference on %P 405 - 412 %8 1992/// %G eng %0 Conference Paper %B Proceedings of the eleventh annual ACM symposium on Principles of distributed computing %D 1992 %T Fast randomized algorithms for distributed edge coloring %A Panconesi,Alessandro %A Srinivasan, Aravind %B Proceedings of the eleventh annual ACM symposium on Principles of distributed computing %S PODC '92 %I ACM %C New York, NY, USA %P 251 - 262 %8 1992/// %@ 0-89791-495-3 %G eng %U http://doi.acm.org/10.1145/135419.135465 %R 10.1145/135419.135465 %0 Book %D 1992 %T A formative evaluation of three interfaces for browsing directories using dynamic queries %A Liao,H.S. %A Osada,M. %A Shneiderman, Ben %I University of Maryland, Center for Automation Research, Computer Vision Laboratory %8 1992/// %G eng %0 Journal Article %J Computer Science Technical Reports %D 1992 %T A Generalization of Brooks' Theorem %A Srinivasan, Aravind %X Given a connected graph $G = (V, E)$ with $n$ vertices and maximum degree $\Delta$ such that $\Delta \geq$ 3 and $G$ is not a complete graph, Brooks' theorem shows that $G$ is $\Delta$-colorable. We prove a generalization of this theorem: assume inductively that all but one vertex $v$ is colored; then, $v$ can be colored by considering the vertices (and their colors) in just an $O$ (log $n$) radius around $v$. Our proof uses a probabilistic technique to link the connectivity and diameter of "almost-regular" graphs, which could have other applications too. %B Computer Science Technical Reports %8 1992/09// %G eng %0 Book %D 1992 %T A Graphical Filter/flow Representation of Boolean Queries: A Prototype Implementation %A Young,D. %A Shneiderman, Ben %I University of Maryland, Center for Automation Research, Computer Vision Laboratory %8 1992/// %G eng %0 Conference Paper %B Proceedings of the twenty-fourth annual ACM symposium on Theory of computing %D 1992 %T Improved distributed algorithms for coloring and network decomposition problems %A Panconesi,Alessandro %A Srinivasan, Aravind %B Proceedings of the twenty-fourth annual ACM symposium on Theory of computing %S STOC '92 %I ACM %C New York, NY, USA %P 581 - 592 %8 1992/// %@ 0-89791-511-9 %G eng %U http://doi.acm.org/10.1145/129712.129769 %R 10.1145/129712.129769 %0 Journal Article %J Information Processing Letters %D 1992 %T On independent spanning trees %A Khuller, Samir %A Schieber,Baruch %K combinatorial problems %K edge connectivity %K spanning trees %K vertex connectivity %X We prove that if any k-vertex connected graph has k vertex independent spanning trees, then any k-edge connected graph has k edge independent spanning trees. Thus, answering a question raised by Zehavi and Itai [J. Graph Theory 13 (1989)] in the affirmative. %B Information Processing Letters %V 42 %P 321 - 323 %8 1992/07/24/ %@ 0020-0190 %G eng %U http://www.sciencedirect.com/science/article/pii/002001909290230S %N 6 %R 10.1016/0020-0190(92)90230-S %0 Conference Paper %B Neural Networks, 1992. IJCNN., International Joint Conference on %D 1992 %T Lateral cortical inhibition without lateral inhibitory connections %A Reggia, James A. %A Sutton,G. %A D'Autrechy,C. L %A Weinrich,M. %K brain models %K competitive distribution of activation %K digital simulation %K excitatory neurons %K lateral cortical inhibition %K neural circuitry %K neurophysiology %K physiological models %X The authors consider how a set of neurons forming a cortical column, none of which individually can competitively distribute its activity, can function collectively to produce competitive distribution of activation. To address this issue, they describe a model cortical column composed of three populations of excitatory neurons that collectively distribute their activity competitively. The idea of competitive distribution and the neural circuitry that could bring it about are described. Computer simulation results, implications, and testable predictions of the theory are presented %B Neural Networks, 1992. IJCNN., International Joint Conference on %V 3 %P 415 -420 vol.3 - 415 -420 vol.3 %8 1992/06// %G eng %R 10.1109/IJCNN.1992.227139 %0 Conference Paper %B Proceedings of the 1992 ACM/IEEE conference on Supercomputing %D 1992 %T Model-driven mapping onto distributed memory parallel computers %A Sussman, Alan %B Proceedings of the 1992 ACM/IEEE conference on Supercomputing %P 818 - 829 %8 1992/// %G eng %0 Conference Paper %B Computer Software and Applications Conference, 1992. COMPSAC '92. Proceedings., Sixteenth Annual International %D 1992 %T On the nature of bias and defects in the software specification process %A Straub,P. %A Zelkowitz, Marvin V %K Center;NASA;Software %K defects;software %K engineering %K Flight %K Goddard %K Laboratory;bias;coding %K phase;design %K processes;multiattribute %K reliability; %K reliability;formal %K software %K space %K specification;software %X Implementation bias in a specification is an arbitrary constraint in the solution space. The authors describe the problem of bias and then present a model of the specification and design processes describing individual subprocesses in terms of precision/detail programs, and a model of bias in multi-attribute software specifications. While studying how bias is introduced into a specification it was realized that software defects and bias are dual problems of a single phenomenon. This has been used to explain the large proportion of faults found during the coding phase at the Software Engineering Laboratory at NASA Goddard Space Flight Center %B Computer Software and Applications Conference, 1992. COMPSAC '92. Proceedings., Sixteenth Annual International %P 17 - 24 %8 1992/09// %G eng %R 10.1109/CMPSAC.1992.217609 %0 Journal Article %J Advances in human-computer interaction %D 1992 %T A new era for high precision touchscreens %A Sears,A. %A Plaisant, Catherine %A Shneiderman, Ben %X While many input devices allow interfaces to be customized, increased directness distinguishes touchscreens. Touchscreens are easy to learn to use, fast, and result in low error rates when interfaces are designed carefully. Many actions which are diffi cult with a mouse, joystick, or keyboard are simple when using a touchscreen. Making rapid selections at widely separated locations on the screen, signing your name, dragging the hands of a clock in a circular motion are all simple when using a touchscre en, but may be awkward using other devices. This paper presents recent empirical research which can provide a basis for theories of touchscreen usage. We believe recent improvements warrant increased use of touchscreens. %B Advances in human-computer interaction %V 3 %P 1 - 33 %8 1992/// %G eng %0 Journal Article %J Advances in Human-Computer Interaction %D 1992 %T A new era for touchscreen applications: High precision, dragging icons, and refined feedback %A Sears,A. %A Plaisant, Catherine %A Shneiderman, Ben %B Advances in Human-Computer Interaction %V 3 %8 1992/// %G eng %0 Journal Article %J J. Parallel Distrib. Comput. %D 1992 %T A parallel pipelined strategy for evaluating linear recursive predicates in a multiprocessor environment. %A Raschid, Louiqa %A Su,S. Y.W %B J. Parallel Distrib. Comput. %V 14 %P 146 - 162 %8 1992/// %G eng %N 2 %0 Journal Article %J Computing Systems in Engineering %D 1992 %T PARTI primitives for unstructured and block structured problems %A Sussman, Alan %A Saltz, J. %A Das,R. %A Gupta,S. %A Mavriplis,D. %A Ponnusamy,R. %A Crowley,K. %X This paper describes a set of primitives (PARTI) developed to efficiently execute unstructured and block structured problems on distributed memory parallel machines. We present experimental data from a three-dimensional unstructured Euler solver run on the Intel Touchstone Delta to demonstrate the usefulness of our methods. %B Computing Systems in Engineering %V 3 %P 73 - 86 %8 1992/// %@ 0956-0521 %G eng %U http://www.sciencedirect.com/science/article/pii/0956052192900962 %N 1-4 %R 10.1016/0956-0521(92)90096-2 %0 Conference Paper %B Proceedings of the 1992 symposium on Interactive 3D graphics %D 1992 %T Real-time procedural textures %A Rhoades,John %A Turk,Greg %A Bell,Andrew %A State,Andrei %A Neumann,Ulrich %A Varshney, Amitabh %B Proceedings of the 1992 symposium on Interactive 3D graphics %S I3D '92 %I ACM %C New York, NY, USA %P 95 - 100 %8 1992/// %@ 0-89791-467-8 %G eng %U http://doi.acm.org/10.1145/147156.147171 %R 10.1145/147156.147171 %0 Journal Article %J International Journal of Man-Machine Studies %D 1992 %T Scheduling home control devices: design issues and usability evaluation of four touchscreen interfaces %A Plaisant, Catherine %A Shneiderman, Ben %X This article describes four different user interfaces supporting scheduling two-state (ON/OFF) devices over time periods ranging from minutes to days. The touchscreen-based user interfaces including a digital 12-h clock, 24-h linear and 24-h dial prototypes are described and compared on a feature by feature basis. A formative usability test with 14 subjects, feedback from more than 30 reviewers, and the flexibility to add functions favour the 24-h linear version. %B International Journal of Man-Machine Studies %V 36 %P 375 - 393 %8 1992/03// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/002073739290040R %N 3 %R 10.1016/0020-7373(92)90040-R %0 Report %D 1992 %T Six Generations of Building Walkthrough: Final Technical Report to the National Science Foundation %A Brooks,Frederick P. %A Airey,John %A Alspaugh,John %A Bell,Andrew %A Brown,Randolph %A Hill,Curtis %A Nimscheck,Uwe %A Rheingans,Penny %A Rohlf,John %A Smith,Dana %A Turner,Douglass %A Varshney, Amitabh %A Wang,Yulan %A Weber,Hans %A Yuan,Xialin %I University of North Carolina at Chapel Hill %C Chapel Hill, NC, USA %8 1992/// %G eng %0 Journal Article %J ACM SIGCHI Bulletin %D 1992 %T Socially responsible computing I: a call to action following the LA riots %A Shneiderman, Ben %B ACM SIGCHI Bulletin %V 24 %P 14 - 15 %8 1992/// %G eng %N 3 %0 Journal Article %J ACM SIGCHI Bulletin %D 1992 %T Socially responsible computing II: first steps on the path to positive contributions %A Shneiderman, Ben %X Now that we are past the initial emotional responses to the LA riots, maybe we can use these tragic events as a stimulus for innovative and constructive efforts. Computing professionals have been active in social causes on an individual basis and in some large projects, but with the growing maturity of our industry we can promote larger initiatives and accept greater responsibility. %B ACM SIGCHI Bulletin %V 24 %P 16 - 17 %8 1992/07// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/134421.134432 %N 3 %R 10.1145/134421.134432 %0 Journal Article %J Nucleic Acids ResearchNucl. Acids Res. %D 1992 %T Splicing signals in Drosophila: intron size, information content, and consensus sequences %A Mount, Stephen M. %A Burks,Christian %A Herts,Gerald %A Stormo,Gary D. %A White,Owen %A Fields,Chris %X A database of 209 Drosophila Introns was extracted from Genbank (release number 64.0) and examined by a number of methods in order to characterize features that might serve as signals for messenger RNA splicing. A tight distribution of sizes was observed: while the smallest introns in the database are 51 nucleotides, more than half are less than 80 nucleotides in length, and most of these have lengths in the range of 59 – 67 nucleotides. Drosophila splice sites found in large and small introns differ in only minor ways from each other and from those found in vertebrate Introns. However, larger introns have greater pyrimidlne-richness in the region between 11 and 21 nucleotides upstream of 3′ splice sites. The Drosophila branchpoint consensus matrix resembles C T A A T (in which branch formation occurs at the underlined A), and differs from the corresponding mammalian signal in the absence of G at the position immediately preceding the branchpoint. The distribution of occurrences of this sequence suggests a minimum distance between 5′ splice shies and branchpoints of about 38 nucleotides, and a minimum distance between 3′ splice sites and branchpoints of 15 nucleotides. The methods we have used detect no information in exon sequences other than in the few nucleotides immediately adjacent to the splice sites. However, Drosophila resembles many other species in that there is a discontinuity in A + T content between exons and introns, which are A + T rich. %B Nucleic Acids ResearchNucl. Acids Res. %V 20 %P 4255 - 4262 %8 1992/08/25/ %@ 0305-1048, 1362-4962 %G eng %U http://nar.oxfordjournals.org/content/20/16/4255 %N 16 %R 10.1093/nar/20.16.4255 %0 Journal Article %J ACM Transactions on Information Systems %D 1992 %T Structural analysis of hypertexts: identifying hierarchies and useful metrics %A Botafogo,Rodrigo A. %A Rivlin,Ehud %A Shneiderman, Ben %K graph theory %K hierarchies %K hypertext %K metrics %K structural analysis %X Hypertext users often suffer from the “lost in hyperspace” problem: disorientation from too many jumps while traversing a complex network. One solution to this problem is improved authoring to create more comprehensible structures. This paper proposes several authoring tools, based on hypertext structure analysis.In many hypertext systems authors are encouraged to create hierarchical structures, but when writing, the hierarchy is lost because of the inclusion of cross-reference links. The first part of this paper looks at ways of recovering lost hierarchies and finding new ones, offering authors different views of the same hypertext. The second part helps authors by identifying properties of the hypertext document. Multiple metrics are developed including compactness and stratum. Compactness indicates the intrinsic connectedness of the hypertext, and stratum reveals to what degree the hypertext is organized so that some nodes must be read before others. Several existing hypertexts are used to illustrate the benefits of each technique. The collection of techniques provides a multifaceted view of the hypertext, which should allow authors to reduce undesired structural complexity and create documents that readers can traverse more easily. %B ACM Transactions on Information Systems %V 10 %P 142 - 180 %8 1992/04// %@ 1046-8188 %G eng %U http://doi.acm.org/10.1145/146802.146826 %N 2 %R 10.1145/146802.146826 %0 Journal Article %J Automata, Languages and Programming %D 1992 %T Testing preorders for probabilistic processes %A Cleaveland, Rance %A Smolka,S. %A Zwarico,A. %B Automata, Languages and Programming %P 708 - 719 %8 1992/// %G eng %0 Journal Article %J Human Factors and Ergonomics Society Annual Meeting Proceedings %D 1992 %T Touchscreen interfaces for alphanumeric data entry %A Plaisant, Catherine %A Sears,A. %B Human Factors and Ergonomics Society Annual Meeting Proceedings %V 36 %P 293 - 297 %8 1992/// %G eng %N 4 %0 Journal Article %J ACM Transactions on Graphics %D 1992 %T Tree visualization with tree-maps: 2-d space-filling approach %A Shneiderman, Ben %B ACM Transactions on Graphics %V 11 %P 92 - 99 %8 1992/01// %@ 0730-0301 %G eng %U http://doi.acm.org/10.1145/102377.115768 %N 1 %R 10.1145/102377.115768 %0 Conference Paper %B Robotics and Automation, 1992. Proceedings., 1992 IEEE International Conference on %D 1992 %T Two miniature pan-tilt devices %A Bederson, Benjamin B. %A Wallace,R.S. %A Schwartz,E.L. %B Robotics and Automation, 1992. Proceedings., 1992 IEEE International Conference on %P 658 - 663 %8 1992/// %G eng %0 Conference Paper %B Applications of Computer Vision, Proceedings, 1992., IEEE Workshop on %D 1992 %T Voice-bandwidth visual communication through logmaps: the Telecortex %A Wallace,R.S. %A Bederson, Benjamin B. %A Schwartz,E.L. %B Applications of Computer Vision, Proceedings, 1992., IEEE Workshop on %P 4 - 10 %8 1992/// %G eng %0 Conference Paper %B , First International Workshop on Interoperability in Multidatabase Systems, 1991. IMS '91. Proceedings %D 1991 %T An algebra and calculus for relational multidatabase systems %A Grant,J. %A Litwin,W. %A Roussopoulos, Nick %A Sellis,T. %K Algebra %K autonomous databases %K Calculus %K Computer networks %K Computer science %K Data models %K Data structures %K Database systems %K database theory %K distributed databases %K Military computing %K multidatabase manipulation language %K multidatabase system %K multirelational algebra %K query languages %K relational algebra %K Relational databases %K Spatial databases %K theoretical foundation %X With the existence of many autonomous databases widely accessible through computer networks, users will require the capability to jointly manipulate data in different databases. A multidatabase system provides such a capability through a multidatabase manipulation language. The authors propose a theoretical foundation for such languages by presenting a multirelational algebra and calculus based on the relational algebra and calculus. The proposal is illustrated by various queries on an example multidatabase %B , First International Workshop on Interoperability in Multidatabase Systems, 1991. IMS '91. Proceedings %I IEEE %P 118 - 124 %8 1991/04// %@ 0-8186-2205-9 %G eng %R 10.1109/IMS.1991.153694 %0 Journal Article %J Contemporary Mathematics %D 1991 %T Combinatorial and computational aspects of Minkowski decompositions %A Mount, Dave %A Silverman,R. %B Contemporary Mathematics %V 119 %P 107 - 124 %8 1991/// %G eng %0 Journal Article %J Advances in control networks and large-scale parallel distributed processing models %D 1991 %T Competitive activation mechanisms in connectionist models %A Reggia, James A. %A Sutton,G. %A Cho,S. %B Advances in control networks and large-scale parallel distributed processing models %V 1 %P 119 - 119 %8 1991/// %G eng %0 Journal Article %J Automata, Languages and Programming %D 1991 %T Computing behavioural relations, logically %A Cleaveland, Rance %A Steffen,B. %B Automata, Languages and Programming %P 127 - 138 %8 1991/// %G eng %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: Reaching through technology %D 1991 %T Demonstrational interfaces: Coming soon? %A Myers,Brad A. %A Cypher,Allen %A Maulsby,David %A Smith,David C. %A Shneiderman, Ben %B Proceedings of the SIGCHI conference on Human factors in computing systems: Reaching through technology %S CHI '91 %I ACM %C New York, NY, USA %P 393 - 396 %8 1991/// %@ 0-89791-383-3 %G eng %U http://doi.acm.org/10.1145/108844.108966 %R 10.1145/108844.108966 %0 Conference Paper %B In Proceedings of the IFIP TC6 Working Conference on ComACM Computing Surveys %D 1991 %T DTP: An Efficient Transport Protocol %A Sanghi,Dheeraj %A Agrawala, Ashok K. %X We recently introduced a new flow control scheme, called send-time control, which is based on a deterministic model of virtual circuits in a computer network. In this scheme, the time at which a packet is sent by a source is computed from estimates of round-trip time, traffic in the network and bottleneck service time. In this paper, we describe a new transport protocol, called DTP, which uses send-time control as its flow control scheme. Preliminary measurements of coast-to-coast connections over the Internet show significant performance improvement over TCP, which is the most commonly used transport protocol in the Internet today. Keyword Codes: C.2.2 Keywords: Computer-Communication Networks, Network Protocols 1. Introduction Flow control is one of the most important function at the transport layer of the protocol hierarchy. The function of flow control schemes is to regulate the traffic entering into the network in order to avoid overloading of network resources, thus preventing c... %B In Proceedings of the IFIP TC6 Working Conference on ComACM Computing Surveys %P 91 - 133 %8 1991/// %G eng %0 Book Section %B Hypertext/hypermedia handbookHypertext/hypermedia handbook %D 1991 %T Editing to structure a reader's experience %A Shneiderman, Ben %A Kreitzberg,C. %A Berk,E. %B Hypertext/hypermedia handbookHypertext/hypermedia handbook %P 143 - 164 %8 1991/// %G eng %0 Journal Article %J Theoretical Computer Science %D 1991 %T Efficient algorithms for the minimum weighted dominating clique problem on permutation graphs %A Srinivasan, Aravind %A Pandu Rangan,C. %X Given a graph G=(V, E) with real weights assigned to its vertices, a clique of G that also dominates its vertex set V, is called a dominating clique (DC) of G. Given a permutation graph G with all its vertices having nonnegative weights, and its permutation representation, the problem addressed in this paper is that of finding any minimum weight DC of G. We improve the existing O(|V|2) algorithm for this problem to O(|V|log|V|). The space complexity of our algorithm is O(|V|). We also present a |V| processor, O(log|V|) time, O(|V|log|V|) space parallel EREW PRAM algorithm for this problem. %B Theoretical Computer Science %V 91 %P 1 - 21 %8 1991/12/09/ %@ 0304-3975 %G eng %U http://www.sciencedirect.com/science/article/pii/0304397591902654 %N 1 %R 10.1016/0304-3975(91)90265-4 %0 Journal Article %J Empirical studies of programmers: fourth workshop %D 1991 %T ESP'91 Panel on The Best of ESP %A Zweben,S. %A Curtis,B. %A Shneiderman, Ben %B Empirical studies of programmers: fourth workshop %P 1 - 1 %8 1991/// %G eng %0 Journal Article %J Information and Computation %D 1991 %T On finding the minimum bandwidth of interval graphs %A Mahesh,R. %A Rangan,C.Pandu %A Srinivasan, Aravind %X Let G = (V, E) be an interval graph, with |V| = n, and |E| = m. An O(n2 log n) algorithm was proposed in Kratsch (Inform Comput. 74, 140–158 (1987)) to find the bandwidth of G. We show that this algorithm is wrong, and provide a corrected version of the same. Also, it was observed in [4] that the bandwidth of a proper interval graph can be computed in O(n log n + m) time. We show how this idea can be modified slightly to yield an O(n = m) algorithm. %B Information and Computation %V 95 %P 218 - 224 %8 1991/12// %@ 0890-5401 %G eng %U http://www.sciencedirect.com/science/article/pii/0890540191900454 %N 2 %R 10.1016/0890-5401(91)90045-4 %0 Book %D 1991 %T Gdr: A visualization tool for graph algorithms %A Stallmann,M. %A Cleaveland, Rance %A Hebbar,P. %I Citeseer %8 1991/// %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1991 %T High precision touchscreens: design strategies and comparisons with a mouse %A Sears,Andrew %A Shneiderman, Ben %X Three studies were conducted comparing speed of performance, error rates and user preference ratings for three selection devices. The devices tested were a touchscreen, a touchscreen with stabilization (stabilization software filters and smooths raw data from hardware), and a mouse. The task was the selection of rectangular targets 1, 4, 16 and 32 pixels per side (0·4 × 0·6, 1·7 × 2·2, 6·9 × 9·0, 13·8 × 17·9 mm respectively). Touchscreen users were able to point at single pixel targets, thereby countering widespread expectations of poor touchscreen resolution. The results show no difference in performance between the mouse and touchscreen for targets ranging from 32 to 4 pixels per side. In addition, stabilization significantly reduced the error rates for the touchscreen when selecting small targets. These results imply that touchscreens, when properly used, have attractive advantages in selecting targets as small as 4 pixels per size (approximately one-quarter of the size of a single character). A variant of Fitts' Law is proposed to predict touchscreen pointing times. Ideas for future research are also presented. %B International Journal of Man-Machine Studies %V 34 %P 593 - 613 %8 1991/04// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/0020737391900378 %N 4 %R 10.1016/0020-7373(91)90037-8 %0 Journal Article %J ACM SIGCHI Bulletin %D 1991 %T Human values and the future of technology: a declaration of responsibility %A Shneiderman, Ben %X "We must learn to balance the material wonders of technology with the spiritual demands of our human nature."---John Naisbitt (1982).We can make a difference in shaping the future by ensurin g that computers "serve human needs (Mumford, 1934)." By making explicit the enduring values that we hold dear we can guide computer system designers and developers for the next decade, century, and thereafter. After setting our high-level goals we can pursue the components and seek the process for fulfilling them.High-level goals might include peace, excellent health care, adequate nutrition, accessible education, communication, freedom of expression, support for creative exploration, safety, and socially constructive entertainment. Computer technology can help attain these high-level goals if we clearly state measurable objectives, obtain participation of professionals, and design effective human-computer interfaces. Design considerations include adequate attention to individual differences among users, support of social and organizational structures, design for reliability and safety, provision of access by the elderly, handicapped, or illiterate, and appropriate user controlled adaptation. With suitable theories and empirical research we can achieve ease of learning, rapid performance, low error rates, and good retention over time, while preserving high subjective satisfaction.To raise the consciousness of designers and achieve these goals, we must generate an international debate, stimulate discussions within organizations, and interact with other intellectual communities. This paper calls for a focus on the "you" and "I" in developing improved user interface (UI) research and systems, offers a Declaration of Responsibility, and proposes a Social Impact Statement for major computing projects. %B ACM SIGCHI Bulletin %V 23 %P 11 - 16 %8 1991/01// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/122672.122674 %N 1 %R 10.1145/122672.122674 %0 Conference Paper %B Proceedings of the third annual ACM conference on Hypertext %D 1991 %T Identifying aggregates in hypertext structures %A Botafogo,Rodrigo A. %A Shneiderman, Ben %B Proceedings of the third annual ACM conference on Hypertext %S HYPERTEXT '91 %I ACM %C New York, NY, USA %P 63 - 74 %8 1991/// %@ 0-89791-461-9 %G eng %U http://doi.acm.org/10.1145/122974.122981 %R 10.1145/122974.122981 %0 Conference Paper %B Proceedings of the 13th international conference on Software engineering %D 1991 %T Metric-driven analysis and feedback systems for enabling empirically guided software development %A Selby,R. W %A Porter, Adam %A Schmidt,D. C %A Berney,J. %B Proceedings of the 13th international conference on Software engineering %P 288 - 298 %8 1991/// %G eng %0 Report %D 1991 %T Modeling of cross traffic in conjunction with deterministic analysis of queues %A Agrawala, Ashok K. %A Sanghi,Dheeraj %I University of Maryland at College Park %C College Park, MD, USA %8 1991/// %G eng %0 Journal Article %J International Journal of Computer Vision %D 1991 %T A multi-frame approach to visual motion perception %A Spetsakis, M. %A Aloimonos, J. %B International Journal of Computer Vision %V 6 %P 245 - 255 %8 1991/// %G eng %N 3 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: Reaching through technology %D 1991 %T Scheduling on-off home control devices %A Plaisant, Catherine %A Shneiderman, Ben %B Proceedings of the SIGCHI conference on Human factors in computing systems: Reaching through technology %S CHI '91 %I ACM %C New York, NY, USA %P 459 - 460 %8 1991/// %@ 0-89791-383-3 %G eng %U http://doi.acm.org/10.1145/108844.109002 %R 10.1145/108844.109002 %0 Journal Article %J Human factors for informatics usability %D 1991 %T A taxonomy and rule base for the selection of interaction styles %A Shneiderman, Ben %B Human factors for informatics usability %P 325 - 325 %8 1991/// %G eng %0 Journal Article %J IEEE Software %D 1991 %T Touch screens now offer compelling uses %A Shneiderman, Ben %K Application software %K Computer aided instruction %K Computer displays %K Computer science %K Cultural differences %K Fingers %K future developments %K information resources %K Marketing and sales %K Mice %K Psychology %K touch screen applications %K touch sensitive screens %K User interfaces %X Research on improving the user interfaces of touch screen applications is described. The advantages of touch screens are discusses, their current capabilities are examined, and possible future developments are considered.<> %B IEEE Software %V 8 %P 93 - 94 %8 1991/03// %@ 0740-7459 %G eng %N 2 %R 10.1109/52.73754 %0 Conference Paper %B , IEEE Conference on Visualization, 1991. Visualization '91, Proceedings %D 1991 %T Tree-maps: a space-filling approach to the visualization of hierarchical information structures %A Johnson,B. %A Shneiderman, Ben %K Computer displays %K Computer Graphics %K Computer science %K Data analysis %K display space %K Educational institutions %K Feedback %K hierarchical information structures %K HUMANS %K Laboratories %K Libraries %K Marine vehicles %K rectangular region %K semantic information %K space-filling approach %K tree-map visualization technique %K trees (mathematics) %K Two dimensional displays %K Visualization %X A method for visualizing hierarchically structured information is described. The tree-map visualization technique makes 100% use of the available display space, mapping the full hierarchy onto a rectangular region in a space-filling manner. This efficient use of space allows very large hierarchies to be displayed in their entirety and facilitates the presentation of semantic information. Tree-maps can depict both the structure and content of the hierarchy. However, the approach is best suited to hierarchies in which the content of the leaf nodes and the structure of the hierarchy are of primary importance, and the content information associated with internal nodes is largely derived from their children %B , IEEE Conference on Visualization, 1991. Visualization '91, Proceedings %I IEEE %P 284 - 291 %8 1991/10/22/25 %@ 0-8186-2245-8 %G eng %R 10.1109/VISUAL.1991.175815 %0 Journal Article %J Proc. 29th Annual ACM DC Technical Symposium %D 1991 %T Window control strategies for hypertext traversal: An empirical study %A Lifshitz,J. %A Shneiderman, Ben %X Nowadays, larger and higher resolution screens supporting multiple windows are widely available. They are often used to traverse hypertext databases for fact retrieval, education or casual browsing. This paper describes a study comparing two methods of window control strategies for article placement in four, equal sized and tiled windows (By tiled we mean non-overlapping, non-movable and non-resizable windows, which together cover the entire screen.) The first placement strategy - Automatic Panning (A P) - was controlled by the system software and the second placement strategy - User Controlled Placement (UCP) - trasferred control to the users. The task required subjects to answer questions about the content of a hypertext. Statistical analysis shows that subjective preference was significantly higher and amount of backtracking was significantly lower for the user controlled placement. Speed differences did not prove to be significant, although there was consistent tendency favoring the automatic pa nning strategy. %B Proc. 29th Annual ACM DC Technical Symposium %8 1991/// %G eng %0 Journal Article %J Hypertext: concepts, systems and applications: proceedings of the First European Conference on Hypertext, INRIA, France, November 1990 %D 1990 %T The advantages of hypertext for large information spaces; where are the big systems? %A Shneiderman, Ben %A Deter,R. %B Hypertext: concepts, systems and applications: proceedings of the First European Conference on Hypertext, INRIA, France, November 1990 %P 343 - 343 %8 1990/// %G eng %N 5 %0 Conference Paper %B Neural Networks, 1990., 1990 IJCNN International Joint Conference on %D 1990 %T Competitive learning using competitive activation rules %A Sutton III,G. G. %A Reggia, James A. %A Maisog,J. M %B Neural Networks, 1990., 1990 IJCNN International Joint Conference on %P 285 - 291 %8 1990/// %G eng %0 Journal Article %J Journal of Computational Physics %D 1990 %T Computing the eigenvalues and eigenvectors of symmetric arrowhead matrices %A O'Leary, Dianne P. %A Stewart, G.W. %B Journal of Computational Physics %V 90 %P 497 - 505 %8 1990/10// %@ 0021-9991 %G eng %U http://www.sciencedirect.com/science/article/pii/0021999190901773 %N 2 %R 10.1016/0021-9991(90)90177-3 %0 Conference Paper %B Automatic Verification Methods for Finite State Systems %D 1990 %T The concurrency workbench %A Cleaveland, Rance %A Parrow,J. %A Steffen,B. %B Automatic Verification Methods for Finite State Systems %P 24 - 37 %8 1990/// %G eng %0 Journal Article %J Information Systems %D 1990 %T Efficient compilation of large rule bases using logical access paths %A Sellis,Timos K. %A Roussopoulos, Nick %A Ng,Raymond T. %X This paper presents an efficient way to compile and store large deductive databases. Deductive rules are modeled through logical access paths and then integrated into a global structure, the Rule-base Access Path Schema (RAP Schema). The RAP Schema stores information on the interrelationships that exist among the computations required for rule executions. Ground data as well as data derived through some rule executions and cached for later reuse are also modeled in the RAP Schema. Some of the advantages of using this rule base organization scheme include the automatic detection of shared computations, the inclusion of cached results of rule executions and query results in the knowledge base. The details of the compilation and query processing phases are the focus of the paper. %B Information Systems %V 15 %P 73 - 84 %8 1990/// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/030643799090017J %N 1 %R 16/0306-4379(90)90017-J %0 Journal Article %J IEEE Software %D 1990 %T Empirically guided software development using metric-based classification trees %A Porter, Adam %A Selby,R. W %K Application software %K Area measurement %K automatic generation %K classification problem %K Classification tree analysis %K Costs %K empirically guided software development %K Error correction %K life cycle %K measurable attributes %K metric-based classification trees %K Predictive models %K Programming %K software engineering %K Software measurement %K software metrics %K Software systems %X The identification of high-risk components early in the life cycle is addressed. A solution that casts this as a classification problem is examined. The proposed approach derives models of problematic components, based on their measurable attributes and those of their development processes. The models provide a basis for forecasting which components are likely to share the same high-risk properties, such as being error-prone or having a high development cost. Developers can use these classification techniques to localize the troublesome 20% of the system. The method for generating the models, called automatic generation of metric-based classification trees, uses metrics from previous releases or projects to identify components that are historically high-risk. %B IEEE Software %V 7 %P 46 - 54 %8 1990/03// %@ 0740-7459 %G eng %N 2 %R 10.1109/52.50773 %0 Journal Article %J Human factors in computing systems: CHI'90 Conference proceedings, April 1-5, Seattle %D 1990 %T EVALUATING HYPERMEDIA SYSTEMS %A Ehrlich,K. %A Microsytems,S. %A Marchionini,G. %A Nielsen,J. %A Shneiderman, Ben %B Human factors in computing systems: CHI'90 Conference proceedings, April 1-5, Seattle %P 387 - 387 %8 1990/// %G eng %0 Journal Article %J Journal of Systems and Software %D 1990 %T Evaluating techniques for generating metric-based classification trees %A Porter, Adam %A Selby,Richard W. %X Metric-based classification trees provide an approach for identifying user-specified classes of high-risk software components throughout the software lifecycle. Based on measurable attributes of software components and processors, this empirically guided approach derives models of problematic software components. These models, which are represented as classification trees, are used on future systems to identify components likely to share the same high-risk properties. Example high-risk component properties include being fault-prone, change-prone, or effort-prone, or containing certain types of faults. Identifying these components allows developers to focus the application of specialized techniques and tools for analyzing, testing, and constructing software. A validation study using metric data from 16 NASA systems showed that the trees had an average classification accuracy of 79.3% for fault-prone and effort-prone components in that environment.One fundamental feature of the classification tree generation algorithm is the method used for partitioning the metric data values into mutually exclusive and exhaustive ranges. This study compares the accuracy and the complexity of trees resulting from five techniques for partitioning metric data values. The techniques are quartiles, octiles, and three methods based on least weight subsequence (LWS-[chi]) analysis, where [chi] is the upper bound on the number of partitions. The LWS-3 and LWS-5 partition techniques resulted in trees with higher accuracy (in terms of completeness and consistency) than did quartiles and octiles. LWS-3 and LWS-5 trees were not statistically different in terms of accuracy, but LWS-3 trees had lower complexity than all other methods in terms of the number of unique metrics required. The trees from the three LWS methods (LWS-3, LWS-5, and LWS-8) had lower complexity than did the trees from quartiles and octiles. In general, the results indicate that distribution-sensitive partition techniques that use only relatively few partitions, such as the least weight subsequence techniques LWS-3 and LWS-5, can increase accuracy and decrease complexity in classification trees. Classification analysis techniques, along with other empirically based analysis techniques for large-scale software, will be supported in the Amadeus measurement and empirical analysis system. %B Journal of Systems and Software %V 12 %P 209 - 218 %8 1990/07// %@ 0164-1212 %G eng %U http://www.sciencedirect.com/science/article/pii/016412129090041J %N 3 %R 16/0164-1212(90)90041-J %0 Journal Article %J Electronic Publishing %D 1990 %T Examining usability for a training-oriented hypertext: Can hyper-activity be good? %A Jones,T. %A Shneiderman, Ben %X We describe the design and evaluation of a hypertext-based tutorial for hypertext authors. This85-article tutorial represents an innovative application of hypertext to procedural learning. The work has been influenced by Carroll’s minimalist model, and by the syntactic/semantic semantic model of user behavior. The usability study involved eight subjects who studied the Hyperties Author Tutorial (HAT) for approximately one hour and then performed a set of authoring tasks in an average of 21 minutes. All users successfully completed the tasks. As a result of the study, we provide a characterization of appropriate uses of hypertext for training, and describe the meaning of a hyper-active environment. %B Electronic Publishing %V 3 %P 207 - 225 %8 1990/// %G eng %N 4 %0 Journal Article %J Discrete Applied Mathematics %D 1990 %T Finding all nearest neighbors for convex polygons in parallel: a new lower bound technique and a matching algorithm %A Schieber,B. %A Vishkin, Uzi %B Discrete Applied Mathematics %V 29 %P 97 - 111 %8 1990/// %G eng %N 1 %0 Book %D 1990 %T FORTRAN Programming: A Spiral Approach %A Kreitzberg,C.B. %A Shneiderman, Ben %I Harcourt Brace College Publishers %8 1990/// %@ 0155280155 %G eng %0 Journal Article %J International Journal of Human-Computer Interaction %D 1990 %T Future directions for human‐computer interaction %A Shneiderman, Ben %X This article offers a set of goals for user interface development, followed by scenarios of future developments. The applications include home control, hypermedia, office automation, digital photography, collaborative meetings/classrooms, public access, professional workstations, and medical record keeping. Also, predictions are made for some of the underlying technologies such as User Interface Management Systems, remote control, flexible search, display devices, and touchscreens.This article offers a set of goals for user interface development, followed by scenarios of future developments. The applications include home control, hypermedia, office automation, digital photography, collaborative meetings/classrooms, public access, professional workstations, and medical record keeping. Also, predictions are made for some of the underlying technologies such as User Interface Management Systems, remote control, flexible search, display devices, and touchscreens. %B International Journal of Human-Computer Interaction %V 2 %P 73 - 90 %8 1990/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/10447319009525971 %N 1 %R 10.1080/10447319009525971 %0 Conference Paper %B Proceedings of the conference on Computers and the quality of life %D 1990 %T Human values and the future of technology: a declaration of empowerment %A Shneiderman, Ben %B Proceedings of the conference on Computers and the quality of life %S CQL '90 %I ACM %C New York, NY, USA %P 1 - 6 %8 1990/// %@ 0-89791-403-1 %G eng %U http://doi.acm.org/10.1145/97344.97360 %R 10.1145/97344.97360 %0 Journal Article %J Journal of Algorithms %D 1990 %T Packing and covering the plane with translates of a convex polygon %A Mount, Dave %A Silverman,Ruth %X A covering of the Euclidean plane by a polygon P is a system of translated copies of P whose union is the plane, and a packing of P in the plane is a system of translated copies of P whose interiors are disjoint. A lattice covering is a covering in which the translates are defined by the points of a lattice, and a lattice packing is defined similarly. We show that, given a convex polygon P with n vertices, the densest lattice packing of P in the plane can be found in O(n) time. We also show that the sparsest lattice covering of the plane by a centrally symmetric convex polygon can be solved in O(n) time. Our approach utilizes results from classical geometry that reduce these packing and covering problems to the problems of finding certain extremal enclosed figures within the polygon. %B Journal of Algorithms %V 11 %P 564 - 580 %8 1990/12// %@ 0196-6774 %G eng %U http://www.sciencedirect.com/science/article/pii/019667749090010C %N 4 %R 10.1016/0196-6774(90)90010-C %0 Journal Article %J CONCUR'90 Theories of Concurrency: Unification and Extension %D 1990 %T A preorder for partial process specifications %A Cleaveland, Rance %A Steffen,B. %B CONCUR'90 Theories of Concurrency: Unification and Extension %P 141 - 151 %8 1990/// %G eng %0 Journal Article %J International Journal of Computer VisionInt J Comput Vision %D 1990 %T Structure from motion using line correspondences %A Spetsakis, Minas E. %A Aloimonos, J. %B International Journal of Computer VisionInt J Comput Vision %V 4 %P 171 - 183 %8 1990/06// %@ 0920-5691 %G eng %U http://www.springerlink.com/content/r4h10053528335k2/ %N 3 %R 10.1007/BF00054994 %0 Conference Paper %B Neural Networks, 1990., 1990 IJCNN International Joint Conference on %D 1990 %T Supervised and reinforced competitive learning %A Sutton III,G. G. %A Reggia, James A. %A Maisog,J. M %B Neural Networks, 1990., 1990 IJCNN International Joint Conference on %P 563 - 567 %8 1990/// %G eng %0 Conference Paper %B Proceedings of the 1990 ACM annual conference on Cooperation %D 1990 %T Touchscreen field specification for public access database queries: let your fingers do the walking %A Sears,Andrew %A Kochavy,Yoram %A Shneiderman, Ben %X Database query is becoming a common task in public access systems; touchscreens can provide an appealing interface for such a system. This paper explores three interfaces for constructing queries on alphabetic field values with a touchscreen interface; including a QWERTY keyboard, an Alphabetic keyboard, and a Reduced Input Data Entry (RIDE) interface. The RIDE interface allows field values to be entered with fewer “keystrokes” (touches) than either keyboard while eliminating certain errors. In one test database, the RIDE interface required 69% fewer keystrokes than either keyboard interface. %B Proceedings of the 1990 ACM annual conference on Cooperation %S CSC '90 %I ACM %C New York, NY, USA %P 1 - 7 %8 1990/// %@ 0-89791-348-5 %G eng %U http://doi.acm.org/10.1145/100348.100349 %R 10.1145/100348.100349 %0 Conference Paper %B Proceedings of DARPA Image Understanding Workshop %D 1990 %T A Unified Theory Of Structure From Motion %A Spetsakis, M. %A Aloimonos, J. %B Proceedings of DARPA Image Understanding Workshop %P 271 - 283 %8 1990/// %G eng %0 Conference Paper %B Logic in Computer Science, 1990. LICS '90, Proceedings., Fifth Annual IEEE Symposium on e %D 1990 %T When is `partial' adequate? A logic-based proof technique using partial specifications %A Cleaveland, Rance %A Steffen,B. %K Calculus %K Carbon capture and storage %K compositional proof rules %K Computer science %K Concurrent computing %K Context %K correctness %K formal specification %K logic-based proof technique %K modal formula %K parallel processes %K partial process specification %K partial specifications %K specification adequacy %K State-space methods %K Technological innovation %X A technique is presented for ascertaining when a (finite-state) partial process specification is adequate, in the sense of being specified enough, for contexts in which it is to be used. The method relies on the automatic generation of a modal formula from the partial specification; if the remainder of the network satisfies this formula, then any process that meets the specification is guaranteed to ensure correct behavior of the overall system. Using the results, the authors develop compositional proof rules for establishing the correctness of networks of parallel processes and illustrate their use with several examples %B Logic in Computer Science, 1990. LICS '90, Proceedings., Fifth Annual IEEE Symposium on e %I IEEE %P 440 - 449 %8 1990/06/04/7 %@ 0-8186-2073-0 %G eng %R 10.1109/LICS.1990.113768 %0 Journal Article %J Electronic Publishing %D 1989 %T Automatically transforming regularly structured linear documents into hypertext. %A Furuta,R. %A Plaisant, Catherine %A Shneiderman, Ben %X Fully automatic conversion of a paper-based document into hypertext can be achieved in manycases if the original document is naturally partitioned into a collection of small-sized pieces that are unambiguously and consistently structured. We describe the methodology that we have used successfully to design and implement several straightforward conversions from the original document’s machine-readable markup. %B Electronic Publishing %V 2 %P 211 - 229 %8 1989/// %G eng %N 4 %0 Journal Article %J IEEE Trans. Softw. Eng. %D 1989 %T Distributed Checkpointing for Globally Consistent States of Databases %A Son,Sang Hyuk %A Agrawala, Ashok K. %K Availability %K checkpoint %K consistency %K distributed database %K noninterference %K recovery %K transaction %B IEEE Trans. Softw. Eng. %V 15 %P 1157 - 1167 %8 1989/10// %@ 0098-5589 %G eng %U http://dx.doi.org/10.1109/TSE.1989.559763 %N 10 %R 10.1109/TSE.1989.559763 %0 Book Section %B Parallel Processing and Medium-Scale Multiprocessors (Proceedings of a 1986 Conference)Parallel Processing and Medium-Scale Multiprocessors (Proceedings of a 1986 Conference) %D 1989 %T Domino: A Transportable System for Parallel Processing %A O'Leary, Dianne P. %A Stewart, G.W. %A Geijn,Robert van de %E Wouk,Arthur %B Parallel Processing and Medium-Scale Multiprocessors (Proceedings of a 1986 Conference)Parallel Processing and Medium-Scale Multiprocessors (Proceedings of a 1986 Conference) %I SIAM Press %C Philadelphia %P 25 - 34 %8 1989/// %G eng %0 Journal Article %J ACM SIGCHI Bulletin %D 1989 %T Dynamic versus static menus: an exploratory comparison %A Mitchell,J. %A Shneiderman, Ben %X Sixty-three subjects completed 24 tasks using a menu driven computer program. The menu items appeared in a fixed (static) order during 12 of the tasks. During the other 12 tasks the menu item order changed dynamically such that the most frequently selected items always appeared at the top of the menu. All the subjects tried both dynamic and static menus.The subjects that used adaptive dynamic menus for the first set of tasks were significantly slower than those who used static menus on the first set of tasks. Subjects' performance during the second set of tasks was not affected by menu style. Eighty-one percent of the subjects preferred working with static menus to working with dynamic menus. %B ACM SIGCHI Bulletin %V 20 %P 33 - 37 %8 1989/04// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/67243.67247 %N 4 %R 10.1145/67243.67247 %0 Conference Paper %B Foundations of Computer Science, 1989., 30th Annual Symposium on %D 1989 %T Efficient parallel algorithms for testing connectivity and finding disjoint s-t paths in graphs %A Khuller, Samir %A Schieber,B. %K algorithm;parallel %K algorithms;random-access %K algorithms;testing %K complexity;graph %K connectivity;computational %K connectivity;optimal %K CRCW %K disjoint %K paths;graphs;k-edge %K paths;k-vertex %K PRAM;disjoint %K s-t %K speedup %K storage; %K theory;parallel %X An efficient parallel algorithm for testing whether a graph G is K-vertex connected, for any fixed k, is presented. The algorithm runs in O(log n) time and uses nC(n,m) processors on a concurrent-read, concurrent-write parallel random-access machine (CRCW PRAM), where n and m are the number of vertices and edges of G and C(n,m) is the number of processors required to compute the connected components of G in logarithmic time. An optimal speedup algorithm for computing connected components would induce an optimal speedup algorithm for testing k -vertex connectivity, for any k gt;4. To develop the algorithm, an efficient parallel algorithm is designed for the following disjoint s-t paths problem: Given a graph G and two specified vertices s and t, find k-vertex disjoint paths between s and t, if they exist. If no such paths exist, find a set of at most k-1 vertices whose removal disconnects s and t. The parallel algorithm for this problem runs in O(log n) time using C(n,m) processors. It is shown how to modify the algorithm to find k-edge disjoint paths, if they exist. This yields an efficient parallel algorithm for testing whether a graph G is k-edge connected, for any fixed k. The algorithm runs in O(log n) time and uses nC (n,n) processors on a CRCW PRAM. Again, an optimal speedup algorithm for computing connected components would induce an optimal speedup algorithm for testing k-edge connectivity %B Foundations of Computer Science, 1989., 30th Annual Symposium on %P 288 - 293 %8 1989/11/01/oct %G eng %R 10.1109/SFCS.1989.63492 %0 Journal Article %J International Journal of Human-Computer Interaction %D 1989 %T An experimental evaluation of three touch screen strategies within a hypertext database %A Potter,Richard %A Berman,Mitchell %A Shneiderman, Ben %X High resolution touch screens and novel usage strategies have overcome earlier problems with parallax and inaccurate pointing. A study testing the utility of three touch screen strategies within the Hyperties hypertext environment was performed. This provided a replication and extension of an earlier touch screen strategy comparison that focused on small closely?spaced targets. The experiment compared three touch screen strategies in three experimental tasks that reflect hypertext usage. The results showed that a strategy that only uses the initial impact with the touch screen causes the user to miss the target more than other touch strategies. A statistically significant difference in errors was found. Our results should encourage system implementers and touch screen hardware designers to support ?touch mouse? strategies that enable cursor dragging on the touch screen surface.High resolution touch screens and novel usage strategies have overcome earlier problems with parallax and inaccurate pointing. A study testing the utility of three touch screen strategies within the Hyperties hypertext environment was performed. This provided a replication and extension of an earlier touch screen strategy comparison that focused on small closely?spaced targets. The experiment compared three touch screen strategies in three experimental tasks that reflect hypertext usage. The results showed that a strategy that only uses the initial impact with the touch screen causes the user to miss the target more than other touch strategies. A statistically significant difference in errors was found. Our results should encourage system implementers and touch screen hardware designers to support ?touch mouse? strategies that enable cursor dragging on the touch screen surface. %B International Journal of Human-Computer Interaction %V 1 %P 41 - 52 %8 1989/// %@ 1044-7318 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/10447318909525956 %N 1 %R 10.1080/10447318909525956 %0 Journal Article %J Journal of Clinical Oncology %D 1989 %T Functional and psychosocial effects of multimodality limb-sparing therapy in patients with soft tissue sarcomas %A Chang,A. E %A Steinberg,S. M %A Culnane,M. %A Lampert,M. H %A Reggia, James A. %A Simpson,C. G %A Hicks,J. E %A White,D. E %A Yang,J. J %A Glatstein,E. %B Journal of Clinical Oncology %V 7 %P 1217 - 1217 %8 1989/// %G eng %N 9 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems: Wings for the mind %D 1989 %T Human-computer interaction lab, University of Maryland %A Shneiderman, Ben %B Proceedings of the SIGCHI conference on Human factors in computing systems: Wings for the mind %S CHI '89 %I ACM %C New York, NY, USA %P 309 - 310 %8 1989/// %@ 0-89791-301-9 %G eng %U http://doi.acm.org/10.1145/67449.67509 %R 10.1145/67449.67509 %0 Journal Article %J CHI %D 1989 %T Human-Computer Interaction Laboratory, University of Maryland, Center for Automation Research %A Shneiderman, Ben %B CHI %V 89 %P 309 - 310 %8 1989/// %G eng %0 Conference Paper %B Proceedings of the second annual ACM conference on Hypertext %D 1989 %T Hypertext and software engineering %A Balzer,R. %A Begeman,M. %A Garg,P. K. %A Schwartz,M. %A Shneiderman, Ben %X The purpose of this panel is to bring together researchers in software engineering and hypertext and help identify the major issues in the application of hypertext technology and concepts to software engineering and vice versa. %B Proceedings of the second annual ACM conference on Hypertext %S HYPERTEXT '89 %I ACM %C New York, NY, USA %P 395 - 396 %8 1989/// %@ 0-89791-339-6 %G eng %U http://doi.acm.org/10.1145/74224.74259 %R 10.1145/74224.74259 %0 Journal Article %J Tech Report HCIL-89-06 %D 1989 %T Hypertext hands-on! %A Shneiderman, Ben %A Kearsley,G. %X This innovative book/software package provides the first hands-on nontechnical introduction to hypertext. Hypertext is a computer technology for manipulating information; in a grander sense, it is a new way of reading and writing. With the IBM-PC disket tes provided in this package, you will learn about hypertext by experiencing it. You will discover what it is like to read interactively, to find information according to your own needs and interests. Both the book and the software versions cover the basic concepts of hypertext, typical hypertext applications, and currently available authoring systems. They also raise important design and implementations issues. The book is self-contained an can be read from beginning to end without a computer. The software is also self-contained and, presenting hypertext in hypertext form, can be read in any order you choose. Since the two versions contain largely similar material, they provide an interesting basis for comparison between conventional text presentation and hypertext. %B Tech Report HCIL-89-06 %8 1989/// %G eng %0 Journal Article %J Proc. 28th Annual ACM DC Technical Symposium %D 1989 %T Interactive graphics interfaces in hypertext systems %A Weiland,W. %A Shneiderman, Ben %X One of the key design aspects of hypertext systems is the rapid selection of items displayed on the screen in a direct manipulation manner. The user can select next or back page turning icons or buttons, or select another item that might jump to a remote destination. Authors and browsers are confronted with the problem of recognizing and selecting these choices in textual and graphic databases. This paper discusses the problems and offers a variety of solutions. %B Proc. 28th Annual ACM DC Technical Symposium %V 23 %P 28 - 28 %8 1989/// %G eng %0 Journal Article %J JOSA A %D 1989 %T Learning early-vision computations %A Aloimonos, J. %A Shulman, D. %B JOSA A %V 6 %P 908 - 919 %8 1989/// %G eng %N 6 %0 Conference Paper %B Proceedings of the second annual ACM conference on Hypertext %D 1989 %T Lessons learned from the ACM hypertext on hypertext project %A Rous,B. %A Shneiderman, Ben %A Yankelovich,N. %A Yoder,E. %B Proceedings of the second annual ACM conference on Hypertext %S HYPERTEXT '89 %I ACM %C New York, NY, USA %P 385 - 386 %8 1989/// %@ 0-89791-339-6 %G eng %U http://doi.acm.org/10.1145/74224.74255 %R 10.1145/74224.74255 %0 Conference Paper %B Workshop on Visual Motion, 1989.,Proceedings %D 1989 %T Optimal motion estimation %A Spetsakis, M. E %A Aloimonos, J. %K 3D motion interpretation %K Automation %K Computer vision %K computerised pattern recognition %K computerised picture processing %K constraint minimization %K dynamic imagery %K Educational institutions %K feature correspondences %K Gaussian noise %K Image motion analysis %K Laboratories %K maximum-likelihood-principle %K Minimization methods %K Motion analysis %K Motion estimation %K motion parameters %K moving object %K multiple frames %K nonlinear constraint %K Optical computing %K optimal motion estimation %K parameter estimation %K quadratic minimization %K quadratic programming %K redundancy %K rigidity assumption %K robotics applications %K structure-from-motion %K successive images %K two-frame %X The problem of using feature correspondences to recover the structure and 3D motion of a moving object from its successive images is analyzed. They formulate the problem as a quadratic minimization problem with a nonlinear constraint. Then they derive the condition for the solution to be optimal under the assumption of Gaussian noise in the input, in the maximum-likelihood-principle sense. The authors present two efficient ways to approximate it and discuss some inherent limitations of the structure-from-motion problem when two frames are used that should be taken into account in robotics applications that involve dynamic imagery. Finally, it is shown that some of the difficulties inherent in the two-frame approach disappear when redundancy in the data is introduced. This is concluded from experiments using a structure-from-motion algorithm that is based on multiple frames and uses only the rigidity assumption %B Workshop on Visual Motion, 1989.,Proceedings %I IEEE %P 229 - 237 %8 1989/03/20/22 %@ 0-8186-1903-1 %G eng %R 10.1109/WVM.1989.47114 %0 Conference Paper %B Proceedings of the IFIP WG6 %D 1989 %T A semantics based verification tool for finite state systems %A Cleaveland, Rance %A Parrow,J. %A Steffen,B. %B Proceedings of the IFIP WG6 %V 1 %P 287 - 302 %8 1989/// %G eng %0 Journal Article %J Educational Media International %D 1989 %T Social and Individual Impact %A Shneiderman, Ben %B Educational Media International %V 26 %P 101 - 106 %8 1989/// %@ 0952-3987 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/0952398890260208 %N 2 %R 10.1080/0952398890260208 %0 Conference Paper %B , Conference on Software Maintenance, 1989., Proceedings %D 1989 %T Software metric classification trees help guide the maintenance of large-scale systems %A Selby,R. W %A Porter, Adam %K automated method %K automatic programming %K classification %K Classification tree analysis %K classification trees %K Computer errors %K empirically-based models %K error-prone software objects %K Fault diagnosis %K feasibility study %K high development effort %K Large-scale systems %K multivalued functions %K NASA %K NASA projects %K recursive algorithm %K Software algorithms %K software engineering %K Software maintenance %K Software measurement %K software metrics %K software modules %K Software systems %K trees (mathematics) %X The 80:20 rule states that approximately 20% of a software system is responsible for 80% of its errors. The authors propose an automated method for generating empirically-based models of error-prone software objects. These models are intended to help localize the troublesome 20%. The method uses a recursive algorithm to automatically generate classification trees whose nodes are multivalued functions based on software metrics. The purpose of the classification trees is to identify components that are likely to be error prone or costly, so that developers can focus their resources accordingly. A feasibility study was conducted using 16 NASA projects. On average, the classification trees correctly identified 79.3% of the software modules that had high development effort or faults %B , Conference on Software Maintenance, 1989., Proceedings %I IEEE %P 116 - 123 %8 1989/10/16/19 %@ 0-8186-1965-1 %G eng %R 10.1109/ICSM.1989.65202 %0 Journal Article %J Hypermedia %D 1989 %T A spectrum of automatic hypertext constructions %A Furuta,R. %A Plaisant, Catherine %A Shneiderman, Ben %X We describe our experiences with four separate conversions from paper documents into hypertext and discuss the lessons we have learned. The paper document's organization affects the ease with which it can be converted and the appropriateness of the resu lting hypertext. The form of the paper document's machine-readable `markup' description affects the ability to transform the structure automatically. Designing the link structures that tie together the parts of the hypertext takes special care in automa ting, as badly-designed and incorrectly-formed links destroy the integrity of the hypertext. Overall, each of the conversions followed the same basic methodology, providing the handle for the development of `power tools' that can be applied to simplify s ubsequent conversions. %B Hypermedia %V 1 %P 179 - 195 %8 1989/// %G eng %N 2 %0 Journal Article %J Interacting with computers %D 1989 %T The user interface in a hypertext, multiwindow program browser %A Seabrook,R.H.C. %A Shneiderman, Ben %B Interacting with computers %V 1 %P 301 - 337 %8 1989/// %G eng %N 3 %0 Book %D 1989 %T User-Friendly Computer Interfaces %A Shneiderman, Ben %I Chantico Pub. Co. %C Carrollton, Texas %8 1989/// %@ 1557110956 %G eng %0 Journal Article %J International journal of man-machine studies %D 1988 %T Analysis of competition-based spreading activation in connectionist models %A Wang,P. Y %A Seidman,S. B %A Reggia, James A. %B International journal of man-machine studies %V 28 %P 77 - 97 %8 1988/// %G eng %N 1 %0 Journal Article %J SPE Reservoir EngineeringSPERE %D 1988 %T Block-Preconditioned Conjugate-Gradient-Like Methods for Numerical Reservoir Simulation %A Eisenstat, Stanley %A Elman, Howard %A Schlutz, Martin %B SPE Reservoir EngineeringSPERE %V 3 %8 1988/02// %@ 0885-9248 %G eng %U http://www.onepetro.org/mslib/servlet/onepetropreview?id=00013534&soc=SPE %N 1 %R 10.2118/13534-PA %0 Conference Paper %B Proceedings 2nd International Conference on Expert Database Systems %D 1988 %T Deep compilation of large rule bases %A Sellis,T. %A Roussopoulos, Nick %B Proceedings 2nd International Conference on Expert Database Systems %8 1988/// %G eng %0 Journal Article %J Commun. ACM %D 1988 %T Ecological studies of professional programmers %A Shneiderman, Ben %A Carroll,John M. %X For over two decades, software psychology researchers have been developing insights to software productivity and quality by investigating builders and users of software. This research has been diverse in both its approach and its impacts. It has introduced systematic behavioral measurement into the software development process and into research on new software techniques and technologies, and has also opened up new social and cognitive interpretations of software processes [5, 12].We now see evidence of a new thrust in software psychology coming to the fore, one in which usability researchers are direct participants in the definition and creation of new software artifacts. We call this paradigm Ecological Design, to emphasize (1) that realistic software situations are being confronted on their own terms, and (2) that the work is directed toward design results, not merely toward evaluation and description in the service of design goals. The reorientation towards studying teamwork was prompted in 1971 by Weinberg and followed by a few researchers at that time, but the movement has accelerated with the recent and intense interest in computer supported collaborative work [15]. This was apparent in the papers presented at the two workshops on Empirical Studies of Programmers [10, 13]. An accompanying shift has also occurred in the software engineering community. The traditional waterfall model of software development with the precise specification of a provable topdown design is giving way to newer exploratory styles of program development that emphasize rapid prototyping and iterative refinement. The shift from product to process also puts greater emphasis on team organization, group processes, management policies, reusability, development tools, design methods, debugging strategies, and maintenance [6]. The three papers in this special section exemplify this new paradigm. Rosson, Maass, and Kellogg and Curtis, Krasner, and Iscoe describe highly qualitative studies of professional designers that produced specific technical proposals for improving software tools and the coordination of project management, an assessment of major bottlenecks, and a new framework for thinking about software design as a learning and communication process. Soloway, Pinto, Letovsky, Littman, and Lampert describe the design and exploration of software documentation that grew out of similarly qualitative studies of program maintenance. We caution that this research paradigm is still in its infancy: setting design requirements and developing prototypes are not traditional activities of psychological researchers. These roles are still emerging, still being reconciled with the earlier paradigms. The particular projects highlighted here are only the beginning; the field continues to evolve, as more researchers are attracted, as more topics are explored, as more methods are developed. Thus, despite the shortcomings of any particular project, the trajectory of this paradigm seems clear to us: it is the development of ideas that directly impact productivity and quality in software. Indeed, part of our intention in presenting this special section is to encourage more and more rapid development of the new paradigm. %B Commun. ACM %V 31 %P 1256 - 1258 %8 1988/11// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/50087.214900 %N 11 %R 10.1145/50087.214900 %0 Journal Article %J IEEE Transactions on Software Engineering %D 1988 %T An efficient pictorial database system for PSQL %A Roussopoulos, Nick %A Faloutsos,C. %A Sellis,T. %K alphanumeric encodings %K Computer science %K Data structures %K Database languages %K database management systems %K Database systems %K Encoding %K Image coding %K Image databases %K multidimensional B-trees %K Object oriented databases %K pictorial database system %K PSQL %K query language %K query languages %K R+-trees %K Relational databases %K Spatial databases %K spatial objects %K spatial search %K User interfaces %X Pictorial databases require efficient and direct spatial search based on the analog form of spatial objects and relationships instead of search based on some cumbersome alphanumeric encodings of the pictures. A description is given of PSQL, a query language that allows pictorial domains to be presented to the user in their analog form and allows him or her to do direct manipulation on the objects found on those domains. Direct spatial search and computation on the pictures is done using efficient data structures, R- and R+-trees (multidimensional B-trees), which are excellent devices for searching spatial objects and relationships found on pictures %B IEEE Transactions on Software Engineering %V 14 %P 639 - 650 %8 1988/05// %@ 0098-5589 %G eng %N 5 %R 10.1109/32.6141 %0 Conference Paper %B Proceedings of the SIGCHI conference on Human factors in computing systems %D 1988 %T An empirical comparison of pie vs. linear menus %A Callahan,J. %A Hopkins,D. %A Weiser,M. %A Shneiderman, Ben %X Menus are largely formatted in a linear fashion listing items from the top to bottom of the screen or window. Pull down menus are a common example of this format. Bitmapped computer displays, however, allow greater freedom in the placement, font, and general presentation of menus. A pie menu is a format where the items are placed along the circumference of a circle at equal radial distances from the center. Pie menus gain over traditional linear menus by reducing target seek time, lowering error rates by fixing the distance factor and increasing the target size in Fitts's Law, minimizing the drift distance after target selection, and are, in general, subjectively equivalent to the linear style. %B Proceedings of the SIGCHI conference on Human factors in computing systems %S CHI '88 %I ACM %C New York, NY, USA %P 95 - 100 %8 1988/// %@ 0-201-14237-6 %G eng %U http://doi.acm.org/10.1145/57167.57182 %R 10.1145/57167.57182 %0 Journal Article %J Data & Knowledge Engineering %D 1988 %T Expert database systems: Efficient support for engineering environments %A Sellis,T. %A Roussopoulos, Nick %A Mark,L. %A Faloutsos,C. %K engineering databases %K Expert database systems %K high performance systems %K incremental computation models %X Manufacturing and Engineering processes use both large scale data and knowledge bases, and the use of expert systems in such environments has become a necessity. Expert Database Systems have evolved from conventional database systems to meet the requirements of current Artificial Intelligence applications. However, future Expert Database Systems will contain knowledge bases of significant size which makes main memory insufficient and the use of a database system a necessity. We propose an effective way of building High Performance Expert Database Systems to support manufacturing and engineering environments. These systems are based on Incremental Computation Models; such models utilize results of previous computations by merging them with newly derived results of computations on small increments representing changes in the database. Our system will be able to support very large knowledge bases by utilizing novel structures and access methods and by using a very sophisticated inference engine based on incremental computation models. %B Data & Knowledge Engineering %V 3 %P 71 - 85 %8 1988/09// %@ 0169-023X %G eng %U http://www.sciencedirect.com/science/article/pii/0169023X88900079 %N 2 %R 16/0169-023X(88)90007-9 %0 Journal Article %J Computer %D 1988 %T Finding facts vs. browsing knowledge in hypertext systems %A Marchionini,G. %A Shneiderman, Ben %K browsing knowledge %K Cognitive science %K cognitive systems %K Electronic publishing %K hypertext research %K hypertext systems %K Information retrieval %K information retrieval systems %K information-seeking %K interface design %K User interfaces %K word processing %X The authors discuss the role of information retrieval, interface design, and cognitive science in hypertext research. They present a user-centered framework for information-seeking that has been used in evaluating two hypertext systems. They apply the framework to key design issues related to information retrieval in hypertext systems.<> %B Computer %V 21 %P 70 - 80 %8 1988/01// %@ 0018-9162 %G eng %N 1 %R 10.1109/2.222119 %0 Conference Paper %B VLSI Algorithms and Architectures %D 1988 %T On finding lowest common ancestors: simplification and parallelization %A Schieber,B. %A Vishkin, Uzi %B VLSI Algorithms and Architectures %P 111 - 123 %8 1988/// %G eng %0 Journal Article %J Simulation %D 1988 %T A general-purpose simulation environment for developing connectionist models %A D'Autrechy,C. L %A Reggia, James A. %A Sutton,G. G %A Goodall,S. M %B Simulation %V 51 %P 5 - 5 %8 1988/// %G eng %N 1 %0 Conference Paper %B Proceedings of the fourth annual symposium on Computational geometry %D 1988 %T Globally-Equiangular triangulations of co-circular points in 0(n log n) time %A Mount, Dave %A Saalfeld,A. %B Proceedings of the fourth annual symposium on Computational geometry %S SCG '88 %I ACM %C New York, NY, USA %P 143 - 152 %8 1988/// %@ 0-89791-270-5 %G eng %U http://doi.acm.org/10.1145/73393.73408 %R 10.1145/73393.73408 %0 Book %D 1988 %T Hypertext on hypertext: IBM PC & compatibles %E Shneiderman, Ben %I ACM %C New York, NY, USA %8 1988/// %@ 0-89791-280-2 %G eng %0 Conference Paper %B VLSI Algorithms and Architectures %D 1988 %T Input sensitive VLSI layouts for graphs of arbitrary degree %A Sherlekar,D. %A JaJa, Joseph F. %X A general method to find area-efficient VLSI layouts of graphs of arbitrary degree is presented. For graphs of maximum degree Δ, the layouts obtained are smaller by a factor of Δ2 than those obtained using existing methods.Optimal planar layouts, and near-optimal nonplanar layouts are also derived for planar graphs of arbitrary degree and gauge. The results span the spectrum between outerplanar graphs (gauge 1), and arbitrary planar graphs (gauge O(n)). Optimality is established by developing families of planar graphs of varying gauge and degree, and proving lower bounds on their layout area. These techniques can be combined to exhibit a trade-off between area and the number of contact cuts. The resulting scheme is sensitive to all three parameters that affect the area: the maximum degree, the gauge, and the number of contact cuts. %B VLSI Algorithms and Architectures %P 268 - 277 %8 1988/// %G eng %R 10.1007/BFb0040394 %0 Journal Article %J IEEE Transactions on Software Engineering %D 1988 %T Learning from examples: generation and evaluation of decision trees for software resource analysis %A Selby,R. W %A Porter, Adam %K Analysis of variance %K Artificial intelligence %K Classification tree analysis %K Data analysis %K decision theory %K Decision trees %K Fault diagnosis %K Information analysis %K machine learning %K metrics %K NASA %K production environment %K software engineering %K software modules %K software resource analysis %K Software systems %K Termination of employment %K trees (mathematics) %X A general solution method for the automatic generation of decision (or classification) trees is investigated. The approach is to provide insights through in-depth empirical characterization and evaluation of decision trees for one problem domain, specifically, that of software resource data analysis. The purpose of the decision trees is to identify classes of objects (software modules) that had high development effort, i.e. in the uppermost quartile relative to past data. Sixteen software systems ranging from 3000 to 112000 source lines have been selected for analysis from a NASA production environment. The collection and analysis of 74 attributes (or metrics), for over 4700 objects, capture a multitude of information about the objects: development effort, faults, changes, design style, and implementation style. A total of 9600 decision trees are automatically generated and evaluated. The analysis focuses on the characterization and evaluation of decision tree accuracy, complexity, and composition. The decision trees correctly identified 79.3% of the software modules that had high development effort or faults, on the average across all 9600 trees. The decision trees generated from the best parameter combinations correctly identified 88.4% of the modules on the average. Visualization of the results is emphasized, and sample decision trees are included %B IEEE Transactions on Software Engineering %V 14 %P 1743 - 1757 %8 1988/12// %@ 0098-5589 %G eng %N 12 %R 10.1109/32.9061 %0 Conference Paper %B Proceedings of Second International Conference on Computer Vision %D 1988 %T Optimal Computing Of Structure From Motion Using Point Correspondences In Two Frames %A Spetsakis, M. E %A Aloimonos, J. %K Automation %K Computer vision %K Educational institutions %K Gaussian noise %K Image motion analysis %K Laboratories %K Least squares approximation %K Least squares methods %K Motion estimation %K Optical computing %B Proceedings of Second International Conference on Computer Vision %I IEEE %P 449 - 453 %8 1988/12/05/8 %@ 0-8186-0883-8 %G eng %R 10.1109/CCV.1988.590022 %0 Journal Article %J Algorithmica %D 1988 %T Parallel construction of a suffix tree with applications %A Apostolico,A. %A Iliopoulos,C. %A Landau,G. M %A Schieber,B. %A Vishkin, Uzi %B Algorithmica %V 3 %P 347 - 365 %8 1988/// %G eng %N 1 %0 Conference Paper %B Artificial Intelligence Applications, 1988., Proceedings of the Fourth Conference on %D 1988 %T Parallel set covering algorithms %A Sekar,S. %A Reggia, James A. %K Butterfly parallel processor system %K irredundancy %K Parallel algorithms %K parsimonious set covering theory %K set covering %K set theory %X The authors develop some parallel algorithms for set covering. A brief introduction is given into the parsimonious set covering theory, and algorithms using one type of parsimony called irredundancy are developed. They also discuss several machine-independent parallel constructs that are used to express the parallel algorithms. The algorithms were tested on the Butterfly parallel processor system. The authors present some of the tests conducted and their analyses. Finally, the merits and limitations of the algorithms that were identified during the tests are presented %B Artificial Intelligence Applications, 1988., Proceedings of the Fourth Conference on %P 274 - 279 %8 1988/03// %G eng %R 10.1109/CAIA.1988.196115 %0 Journal Article %J Information Processing & Management %D 1988 %T Selection devices for user of an electronic encyclopedia: An empirical comparison of four possibilities %A Ostroff,Daniel %A Shneiderman, Ben %X This study measured the speed, error rates, and subjective evaluation of arrow-jump keys, a jump-mouse, number keys, and a touch screen in an interactive encyclopedia. A summary of previous studies comparing selection devices and strategies is presented to provide the background for this study. We found the touch screen to be the fastest in time, the least accurate but the overall favorite of the participants. The results are discussed and improvements are suggested accordingly. %B Information Processing & Management %V 24 %P 665 - 680 %8 1988/// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/0306457388900040 %N 6 %R 10.1016/0306-4573(88)90004-0 %0 Journal Article %J Proceedings of the IEEE %D 1988 %T Self-processing networks and their biomedical implications %A Reggia, James A. %A Sutton III,G. G. %B Proceedings of the IEEE %V 76 %P 680 - 692 %8 1988/// %G eng %N 6 %0 Journal Article %J International journal of computer vision %D 1988 %T Shape from patterns: Regularization %A Aloimonos, J. %A Swain, M. %B International journal of computer vision %V 2 %P 171 - 187 %8 1988/// %G eng %N 2 %0 Book %D 1988 %T User interface strategies '88 (videotape) %A Shneiderman, Ben %A Malone,Thomas %A Norman,Donald %A Foley,James %I University of Maryland at College Park %C College Park, MD, USA %8 1988/// %G eng %0 Journal Article %J Ergonomics %D 1988 %T We can design better user interfaces: A review of human-computer interaction styles %A Shneiderman, Ben %X The widespread use of computers has opened a new dimension of application for the ergonomic sciences. This review recommends three pillars to support the user interface design process: guidelines documents, User Interface Management Systems, and usability labs for iterative testing. Then it presents five primary interaction styles: menu selection, form fill-in, command language, natural language and direct manipulation. The author encourages greater attention to direct manipulation in which the objects and actions are visible, the actions are invoked by selection or pointing, and the impact is immediately visible and reversible.The widespread use of computers has opened a new dimension of application for the ergonomic sciences. This review recommends three pillars to support the user interface design process: guidelines documents, User Interface Management Systems, and usability labs for iterative testing. Then it presents five primary interaction styles: menu selection, form fill-in, command language, natural language and direct manipulation. The author encourages greater attention to direct manipulation in which the objects and actions are visible, the actions are invoked by selection or pointing, and the impact is immediately visible and reversible. %B Ergonomics %V 31 %P 699 - 710 %8 1988/// %@ 0014-0139 %G eng %U http://www.tandfonline.com/doi/abs/10.1080/00140138808966713 %N 5 %R 10.1080/00140138808966713 %0 Conference Paper %B Proceedings of the 15th annual conference on Computer Science %D 1987 %T Algorithms for covering and packing and applications to CAD/CAM (abstract only): preliminary results %A Mount, Dave %A Silverman,Ruth %X Computational geometry deals with the computational complexity of geometric problems within the framework of the analysis of algorithms. Numerous applications have been found to computer graphics, computer-aided design, pattern recognition and robotics.Among the geometric problems are bin packing problems. C.A. Roger “Packing and Covering”, Cambridge University Press (1964) studied these problems. Many of these are hard problems, and very few have been solved. Some packing problems are the restriction of a very interesting problem in CAM (computer-aided manufacture). In the manufacture of clothing, certain parts, such as sleeves, are laid out as repeated polygonal copies and combinations on a bolt of cloth. This pattern layout is traditionally (and currently) done by hand. It's a candidate for automation. We propose some special cases of this problem whose solution can be extended to practical use. We are working on several geometric questions involving packing translations of objects. Classical results in mathematics demonstrate equivalence between certain packing problems and finding optimal enclosed figures. J. O'Rourke et al [“An optimal algorithm for finding minimal enclosing triangles”, J. of Algorithms 7, 258-264 (1986)], as well as V. Klee and M. Laskowski [“Finding the smallest triangles containing a given convex polygon”, J. of Algorithms 6, 359-375 (1985)] have produced algorithms for finding a smallest area triangle containing (or enclosing) a given polygon. Smallest area convex polygons of a certain type, containing a given set, can be viewed as larger pieces of fabric containing pattern to be cut out. We have an optimal algorithm for finding largest enclosed polygons for a given set. We also have preliminary results for smallest enclosing polygons. These results have theoretical interest, and have applications to robotics and collision avoidance, as well as to manufacture. %B Proceedings of the 15th annual conference on Computer Science %S CSC '87 %I ACM %C New York, NY, USA %P 439– - 439– %8 1987/// %@ 0-89791-218-7 %G eng %U http://doi.acm.org/10.1145/322917.323100 %R 10.1145/322917.323100 %0 Journal Article %J ACM SIGMOD Record %D 1987 %T Analysis of object oriented spatial access methods %A Faloutsos,C. %A Sellis,T. %A Roussopoulos, Nick %B ACM SIGMOD Record %V 16 %P 426 - 439 %8 1987/// %G eng %N 3 %0 Conference Paper %B Proceedings of the SIGCHI/GI conference on Human factors in computing systems and graphics interface %D 1987 %T Computer-supported cooperative work (panel): is this REALLY a new field of research? %A Greif,Irene %A Curtis,Bill %A Krasner,Herb %A Malone,Thomas W. %A Shneiderman, Ben %B Proceedings of the SIGCHI/GI conference on Human factors in computing systems and graphics interface %S CHI '87 %I ACM %C New York, NY, USA %P 227 - 228 %8 1987/// %@ 0-89791-213-6 %G eng %U http://doi.acm.org/10.1145/29933.275634 %R 10.1145/29933.275634 %0 Book %D 1987 %T Design Guidebook for Interaction Styles: A Taxonomy, Rule-base, and Some Opinions %A Shneiderman, Ben %I University of Maryland %8 1987/// %G eng %0 Book %D 1987 %T Designing the user interface: professional development courses from the University of Maryland %A Shneiderman, Ben %I University of Maryland at College Park %C College Park, MD, USA %8 1987/// %G eng %0 Book Section %B Expert systems: the user interfaceExpert systems: the user interface %D 1987 %T Direct manipulation user interfaces for expert systems %A Baroff,J. %A Simon,R. %A Gilman,F. %A Shneiderman, Ben %B Expert systems: the user interfaceExpert systems: the user interface %P 99 - 125 %8 1987/// %G eng %0 Journal Article %J Behaviour & Information Technology %D 1987 %T An evaluation of jump-ahead techniques in menu selection %A Laverson,Alan %A Norman,Kent %A Shneiderman, Ben %X Menu selection systems provide a means of selecting operations and retrieving information which requires little training and reduces the need for memorizing complex command sequences. However, a major disadvantage of many menu selection systems is that experienced users cannot traverse the menu tree significantly faster than novices. A common solution to this problem is to provide the menu selection system with a jump-ahead capability.The purpose of this research was to evaluate two jump-ahead methods (type-ahead and direct-access). In the type-ahead method the user anticipates a selection on each of several successive menus and enters as many selections at one time as desired. In the direct-access method, each menu frame is assigned a unique name which the user must enter to locate it. Thirty-two students were given training on an information retrieval system for college course information and were required to learn the two jump-ahead methods in a counterbalanced design. The direct-access method resulted in fewer traversals to learn the system, lower error rates, and reduced learning time. The subjective impressions, obtained from post-experiment questionnaires and oral comments, indicated that the direct-access jump-ahead method was also preferred in a frequently used menu selection system. %B Behaviour & Information Technology %V 6 %P 97 - 108 %8 1987/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449298708901820 %N 2 %R 10.1080/01449298708901820 %0 Journal Article %J Computers, IEEE Transactions on %D 1987 %T From Determinacy to Systaltic Arrays %A O'Leary, Dianne P. %A Stewart, G.W. %X In this paper we extend a model of Karp and Miller for parallel computation. We show that the extended model is deterministic, in the sense that under different scheduling regimes each process in the computation consumes the same input and generates the same output. Moreover, if the computation halts, the final state is independent of scheduling. The model is applied to the generation of precedence graphs, from which lower time bounds may be deduced, and to the synchronization of systolic arrays by local rather than global control. %B Computers, IEEE Transactions on %V C-36 %P 1355 - 1359 %8 1987/11// %@ 0018-9340 %G eng %N 11 %R 10.1109/TC.1987.5009475 %0 Journal Article %J Proc. DARPA Image %D 1987 %T Learning shape computations %A Aloimonos, J. %A Shulman, D. %B Proc. DARPA Image %8 1987/// %G eng %0 Conference Paper %B Functional Programming Languages and Computer Architecture %D 1987 %T Mapping a single-assignment language onto the Warp systolic array %A Gross,T. %A Sussman, Alan %B Functional Programming Languages and Computer Architecture %P 347 - 363 %8 1987/// %G eng %0 Journal Article %J Automata, Languages and Programming %D 1987 %T Parallel construction of a suffix tree %A Landau,G. %A Schieber,B. %A Vishkin, Uzi %B Automata, Languages and Programming %P 314 - 325 %8 1987/// %G eng %0 Journal Article %J Proceedings of VLDB 1987 %D 1987 %T The R+-tree: A dynamic index for multi-dimensional data %A Sellis,T. %A Roussopoulos, Nick %A Faloutsos,C. %B Proceedings of VLDB 1987 %P 507 - 518 %8 1987/// %G eng %0 Journal Article %J Journal of the Optical Society of America A %D 1987 %T Spatiotemporal blur paths for image flow estimation (A) %A Spetsakis, M. E %A Aloimonos, J. %B Journal of the Optical Society of America A %V 4 %P 35 - 35 %8 1987/// %G eng %0 Book %D 1987 %T Structure from Motion from Line Correspondencies: New Results %A Science, University of Maryland at College Park. Dept. of Computer %A Spetsakis, M. E %A Aloimonos, J. %A Research, University of Maryland at College Park. Center for Automation %8 1987/// %G eng %0 Journal Article %J Department of Computer Science and Human-Computer Interaction Laboratory Working Paper %D 1987 %T Subjective user evaluation of CF PASCAL programming tools %A Chin,JP %A Norman,K. L %A Shneiderman, Ben %X This study investigated subjective evaluations of two programming environments: 1) SUPPORT, an interactive programming environment with a syntax directed editor on a personal computer and 2) a batch run environment on a large mainframe computer. Participants were students in a 15 week introductory computer science course. In Part 1, one group of 128 first used SUPPORT, while another group of 85 programmed on a mainframe environment. After 6 weeks they were given an evaluative questionnaire and then switched programming environments. In Part 2, 68 used SUPPORT and 60 used the mainframe. At the twelfth week of the course, they were given two questionnaires, one evaluating the environment they had used in the last 6 weeks and one comparing both enviro nments. A measure of programming performance (exam and programming project grades) was also collected. SUPPORT was predicted to reduce the burden of remembering syntactic details resulting in better performance and higher subjective evaluations. Unexpectedly, the SUPPORT users did not earn statistically significantly higher grades. Furthermore, participants expressed a preference for the mainframe over SUPPORT. Specific items on the questionnaires were used to diagnose the strengths and weaknesses of each environment. Designers of syntax directed editors should focus on reducing the syntactic burden not only in programming , but also in the user interface of these tools. %B Department of Computer Science and Human-Computer Interaction Laboratory Working Paper %8 1987/// %G eng %0 Journal Article %J Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting %D 1987 %T Time Stress Effects on Two Menu Selection Systems %A Wallace,Daniel F %A Anderson,Nancy S %A Shneiderman, Ben %X The optimal number of menu items per display screen has been the topic of considerable debate and study. On the one hand, some designers have packed many items into each menu to conserve space and reduce the number of menus, whereas on the other hand there are designers who prefer a sparse display for menu structures and other videotex information. This study evaluated the effects of a broad/shallow menu compared to a narrow/deep menu structure under two conditions of time stress for inexperienced users. Results indicated that time stress both slowed performance, and increased errors. In addition, it was demonstrated that the broad/shallow menu was faster and resulted in fewer errors. Implications for menu design are discussed. %B Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting %V 31 %P 727 - 731 %8 1987/09/01/ %@ 1071-1813, %G eng %U http://pro.sagepub.com/content/31/7/727 %N 7 %R 10.1177/154193128703100708 %0 Conference Paper %B Proc. of the 2nd International Conference on Human-Computer Interaction %D 1987 %T User interface design and evaluation for an electronic encyclopedia %A Shneiderman, Ben %X The Interactive Encyclopedia System (TIES) has been under development since Fall 1983. It enables users to easily traverse a database of articles by merely pointing at highlighted words in context. This embedded menus approach to hypertext and its user interface design are described with three exploratory studies of TIES use. Plans for future development and studies are offered. %B Proc. of the 2nd International Conference on Human-Computer Interaction %8 1987/// %G eng %0 Conference Paper %B Proceedings of the ACM conference on Hypertext %D 1987 %T User interface design for the Hyperties electronic encyclopedia (panel session) %A Shneiderman, Ben %B Proceedings of the ACM conference on Hypertext %S HYPERTEXT '87 %I ACM %C New York, NY, USA %P 189 - 194 %8 1987/// %@ 0-89791-340-X %G eng %U http://doi.acm.org/10.1145/317426.317441 %R 10.1145/317426.317441 %0 Journal Article %J Department of Computer Science and Human Computer Interaction Laboratory, University of Maryland, College Park, MD (correspondence with B. Schneiderman) %D 1987 %T Window control strategies for on-line text traversal %A Lifshitz,K. %A Shneiderman, Ben %X Larger and higher resolution screens that support multiple windows are now widely available. They are often used to traverse hypertext databases for fact retrieval, education or casual browsing. This paper introduces window control strategies that we implemented, describes their cognitive complexity, and characterizes the tasks that users might encounter. An infomal usage by dozens of visitors and a usability test with four subjects performing information search tasks revealed the limitations of several strategies and guided us to select user control of article placement in a tiled non-overlapping multiple window display. %B Department of Computer Science and Human Computer Interaction Laboratory, University of Maryland, College Park, MD (correspondence with B. Schneiderman) %8 1987/// %G eng %0 Journal Article %J COMP. SYST. SCI. ENG. %D 1986 %T Algorithm for concurrency control in replicated databases. %A Son,S. H %A Agrawala, Ashok K. %X A replicated database system is a distributed database system in which some data objects are stored redundantly at multiple sites to improve the reliability of the system. Without proper control mechanisms, the consistency of a replicated database system could be violated. A new scheme to increase the reliability as well as the degree of concurrency is described. It allows transactions to operate on a data object so long as one or more token copies are available. The scheme also exploits the fact that, for recovery reasons, there are two values for one data object. It is proved that transactions executed according to this scheme preserve consistency. %B COMP. SYST. SCI. ENG. %V 1 %P 75 - 81 %8 1986/// %G eng %N 2 %0 Journal Article %J SIGSOFT Softw. Eng. Notes %D 1986 %T Applying direct manipulation concepts: direct manipulation dik operating system (DMDOS) %A Iseki,Osamu %A Shneiderman, Ben %X Software engineers are often called upon to design user interfaces, but strategies and guidelines are only beginning to emerge. Shneiderman (1983) introduced the term "Direct Manipulation" to describe user interfaces which have:1) continuous representation of the objects of interest.2) physical actions (movement and selection by mouse, joystick, touch screen, etc.) or labeled button presses instead of complex Syntax.3) rapid, incremental, reversible operations whose impact on the object of interest is immediately visible.4) layered or spiral approach to learning that permits usage with minimal knowledge.The concepts of direct manipulation has been applied in some distinctive systems such as XEROX STAR and APPLE Macintosh, and many application software products such as spread sheets, word processors, drawing tools, desk-top managers, etc.However, the basic software of personal computers, the operating system, is still often based on command language concepts. This paper describes DMDOS (Direct Manipulation Disk Operating System), that we designed by applying the concepts of direct manipulation. to MS-DOS on the IBM PC. %B SIGSOFT Softw. Eng. Notes %V 11 %P 22 - 26 %8 1986/04// %@ 0163-5948 %G eng %U http://doi.acm.org/10.1145/382248.382815 %N 2 %R 10.1145/382248.382815 %0 Journal Article %J Linear Algebra and its Applications %D 1986 %T Assignment and scheduling in parallel matrix factorization %A O'Leary, Dianne P. %A Stewart, G.W. %X We consider the problem of factoring a dense n×n matrix on a network consisting of P MIMD processors, with no shared memory, when the network is smaller than the number of elements in the matrix (P<n2). The specific example analyzed is a computational network that arises in computing the LU, QR, or Cholesky factorizations. We prove that if the nodes of the network are evenly distributed among processors and if computations are scheduled by a round-robin or a least-recently-executed scheduling algorithm, then optimal order of speedup is achieved. However, such speedup is not necessarily achieved for other scheduling algorithms or if the computation for the nodes is inappropriately split across processors, and we give examples of these phenomena. Lower bounds on execution time for the algorithm are established for two important node-assignment strategies. %B Linear Algebra and its Applications %V 77 %P 275 - 299 %8 1986/05// %@ 0024-3795 %G eng %U http://www.sciencedirect.com/science/article/pii/0024379586901722 %R 10.1016/0024-3795(86)90172-2 %0 Journal Article %J Tech Report HCIL-86-08 %D 1986 %T Is Bigger Better?: The Effects of Display Size on Program Reading %A Reisel,J.F. %A Shneiderman, Ben %X An experiment was conducted in which four window sizes (10, 22, 60 and 120 lines) were compared for their effectiveness for program reading. The largest window, in which the entire program could be seen at one time, was found significantly better, in both time to complete the reading task and the number of correct answers in the reading task. Subjects also preferred the larger windows to the smaller windows. %B Tech Report HCIL-86-08 %8 1986/// %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1986 %T Cognitive layouts of windows and multiple screens for user interfaces %A Norman,Kent L %A Weldon,Linda J. %A Shneiderman, Ben %X In order to make computers easier to use and more versatile many system designers are exploring the use of multiple windows on a single screen and multiple coordinated screens in a single work station displaying linked or related information. The designers of such systems attempt to take into account the characteristics of the human user and the structure of the tasks to be performed. Central to this design issue is the way in which the user views and cognitively processes information presented in the windows or in multiple screens. This paper develops a theory of the “cognitive layout” of information presented in multiple windows or screens. It is assumed that users adopt a cognitive representation or layout of the type of information to be presented and the relationships among the windows or screens and the information they contain. A number of cognitive layouts are derived from theories in cognitive psychology and are discussed in terms of the intent of the software driving the system and congruence with the cognitive processing of the information. It is hypothesized that the particular layout adopted by a user will drastically affect the user's understanding and expectation of events at the human-computer interface and could either greatly facilitate or frustrate the interaction. Ways of ensuring the former and avoiding the latter are discussed in terms of implementations on existing multiple-window and multiple-screen systems. %B International Journal of Man-Machine Studies %V 25 %P 229 - 248 %8 1986/08// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020737386800773 %N 2 %R 10.1016/S0020-7373(86)80077-3 %0 Journal Article %J Proc. 29th Conference of the Association for the Development of Computer Based Instructional Systems %D 1986 %T Design and research on the interactive encyclopedia system (TIES) %A Morariu,J. %A Shneiderman, Ben %X Printed books were an enormous stimulus to science, culture, commerce, and entertainment. Electronic books and hypertext systems may produce a similar stimulus in the next century, but current designs are poor. Typical screens are too small, too slow, too complicated, and too hard to read. With careful attention to the user interface and the underlying technology, we have a chance to create a new medium that is potentially more attractive and effective than printed books in many situations. %B Proc. 29th Conference of the Association for the Development of Computer Based Instructional Systems %P 19 - 21 %8 1986/// %G eng %0 Journal Article %J Journal of the American Society for Information Science %D 1986 %T Designing menu selection systems %A Shneiderman, Ben %X Menu selection systems reduce training and memorization, simplify entry of choices, and structure the user's tasks. However, the use of menu selection is no guarantee that novices or experts will be satisfied or able to carry out their work. This article focuses on the multiple design issues in creating successful menu selection systems. These include the primary issue of semantic organization and the host of secondary issues such as response time and display rates, shortcuts for frequent users, titles, phrasing of menu items, graphic layout, and selection mechanisms. Novel approaches such as pop-up menus and embedded menus are covered. Experimental results and design guidelines are presented. © 1986 John Wiley & Sons, Inc. %B Journal of the American Society for Information Science %V 37 %P 57 - 70 %8 1986/03/01/ %@ 1097-4571 %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/(SICI)1097-4571(198603)37:2%3C57::AID-ASI2%3E3.0.CO;2-S/abstract %N 2 %R 10.1002/(SICI)1097-4571(198603)37:2<57::AID-ASI2>3.0.CO;2-S %0 Journal Article %J IEEE Software %D 1986 %T Display Strategies for Program Browsing: Concepts and Experiment %A Shneiderman, Ben %A Shafer,P. %A Simon,R. %A Weldon,L. %K Computer errors %K Content management %K Fault detection %K Information analysis %K Large screen displays %K Microscopy %K Phase detection %K Programming profession %K User interfaces %X The new, larger display screens can improve program comprehension—if the added space is used for mome effective presentation, not just more code or larger type. %B IEEE Software %V 3 %P 7 - 15 %8 1986/05// %@ 0740-7459 %G eng %N 3 %R 10.1109/MS.1986.233405 %0 Journal Article %J Empirical studies of programmers %D 1986 %T Empirical studies of programmers: The territory, paths, and destinations %A Shneiderman, Ben %X This paper attempts to describe the varied intellectual territory thatprogrammers work in. It offers several paths for researchers who wish to explore this territory: controlled experiments, observational or field studies, surveys, and cognitive theories. Finally, this paper suggests several important destinations for researchers: refining the use of current languages, improving present and future languages, developing special purpose languages, and improving tools and methods. %B Empirical studies of programmers %P 1 - 12 %8 1986/// %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1986 %T An experimental comparison of a mouse and arrow-jump keys for an interactive encyclopedia %A Ewing,John %A Mehrabanzad,Simin %A Sheck,Scott %A Ostroff,Dan %A Shneiderman, Ben %X This paper reports on an experiment which was conducted to examine relative merits of using a mouse or arrow-jump keys to select text in an interactive encyclopedia. Timed path traversais were performed by subjects using each device, and were followed by subjective questions. Personality and background of the subjects were recorded to see if those attributes would affect device preference and performance. The arrow-jump keys were found to have the quickest traversal times for paths with either short or long target distances. The subjective responses indicated that the arrow-jump method was overwhelmingly preferred over the mouse method. Personality type was not found to play a critical role. %B International Journal of Man-Machine Studies %V 24 %P 29 - 45 %8 1986/01// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020737386800384 %N 1 %R 10.1016/S0020-7373(86)80038-4 %0 Journal Article %J ACM SIGCHI Bulletin %D 1986 %T Human-computer interaction research at the University of Maryland %A Shneiderman, Ben %X The Human-Computer Interaction Laboratory (HCIL) is a unit of the Center for Automation Research at the Univesity of Maryland. HCIL is an interdisciplinary research group whose participants are faculty in the Departments of Computer Science and Psychology and the Colleges of Library and Information Services, Business, and Education. In addition, staff scienctists, graduate students, and undergraduates contribute to this small, but lively community that pursues empirical studies of people using computers.Our support comes from industrial research projects, government grants, the State of Maryland, and the University of Maryland. Projects often become interrelated in surprising ways enabling individuals to cooperate constructively. Some of our efforts during the past year are described below. %B ACM SIGCHI Bulletin %V 17 %P 27 - 32 %8 1986/01// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/15671.15673 %N 3 %R 10.1145/15671.15673 %0 Journal Article %J SIAM Journal on Scientific and Statistical ComputingSIAM J. Sci. and Stat. Comput. %D 1986 %T A Hybrid Chebyshev Krylov Subspace Algorithm for Solving Nonsymmetric Systems of Linear Equations %A Elman, Howard %A Saad, Youcef %A Saylor, Paul E. %B SIAM Journal on Scientific and Statistical ComputingSIAM J. Sci. and Stat. Comput. %V 7 %P 840 - 840 %8 1986/// %@ 01965204 %G eng %U http://link.aip.org/link/SJOCE3/v7/i3/p840/s1&Agg=doi %N 3 %R 10.1137/0907057 %0 Journal Article %J Computer Supported Cooperative Work: Proceedings of the 1986 ACM conference on Computer-supported cooperative work %D 1986 %T Interfaces: multi-media and multi-user %A Shneiderman, Ben %B Computer Supported Cooperative Work: Proceedings of the 1986 ACM conference on Computer-supported cooperative work %8 1986/// %G eng %0 Journal Article %J ACM SIGCHI Bulletin %D 1986 %T Learning disabled students' difficulties in learning to use a word processor: implications for design %A MacArthur,Charles A. %A Shneiderman, Ben %X Learning disabled students can derive great benefits from using work processors. The ability to produce a neat, printed copy can increase motivation and encourage writing for a wider audience. The editing power makes revision possible without tedious re-copying, thus freeing students and teachers to approach writing as a process involving repeated drafts. Specific problems with handwriting and spelling can also be circumvented. However, learning to use the editing capabilities often presents problems for students, especially those with learning difficulties. Word processors must be designed that are simple, easy to learn, and yet powerful. This study makes software design recommendations based on a study of learning disabled students learning to use word processing.Two groups of four LD students (4th-6th grade) were given twelve hours of word processing instruction using two word processors. Detailed records of progess and errors were made during learning and a final assessment task. Specific design problems are reported and recommendations are made for tasks such as cursor movement, insertion/deletion, use of nulls, blanks, and formatting characters, and overall organization. %B ACM SIGCHI Bulletin %V 17 %P 41 - 46 %8 1986/01// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/15671.15675 %N 3 %R 10.1145/15671.15675 %0 Journal Article %J Journal of Learning DisabilitiesJ Learn Disabil %D 1986 %T Learning Disabled Students' Difficulties in Learning to Use A Word Processor: Implications for Instruction and Software Evaluation %A MacArthur,Charles A. %A Shneiderman, Ben %X Learning disabled (LD) students can derive great benefits from using word processors. The ability to produce a neat, printed copy can increase motivation and encourage writing for a wider audience. The editing power makes revision possible without tedious recopying, thus freeing students and teachers to approach writing as a process involving repeated drafts. Specific problems with handwriting and spelling can also be circumvented. However, learning to use a word processor often presents problems. Based on a study of LD students learning to use word processing, this paper makes recommendations for evaluating word processing software and designing instruction that is sensitive to students difficulties. %B Journal of Learning DisabilitiesJ Learn Disabil %V 19 %P 248 - 253 %8 1986/04/01/ %@ 0022-2194, %G eng %U http://ldx.sagepub.com/content/19/4/248 %N 4 %R 10.1177/002221948601900414 %0 Journal Article %J ACM SIGCHI Bulletin %D 1986 %T No members, no officers, no dues: A ten year history of the software psychology society %A Shneiderman, Ben %X For ten years we have joyfully overcome our insecurity with anarchy and successfully conducted the "business" of the Software Psychology Society. Our business has always been science; to improve our understanding of how people use computers. The two-horse team of computer science and psychology usually pulled in the same direction. Sometimes it leaned towards design guidelines, software, and hardware, other times it leaned towards cognitive models, personality theory, and human problem-solving. Mostly our travel was steadied by an appreciation of the importance of empirical studies, data collection, and controlled experiments with an occasional diversion into theoretical models or inspirational system designs. Our destinations have included programming and command language design, human-computer interaction models, menu selection strategies, natural language interaction, software engineering metrics and methods, educational packages, expert system user interfaces, voice communications and conferencing systems, screen readability, use of color, user interface management systems, and text editor design. %B ACM SIGCHI Bulletin %V 18 %P 14 - 16 %8 1986/10// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/15683.15685 %N 2 %R 10.1145/15683.15685 %0 Journal Article %J Theoretical Computer Science %D 1986 %T Parallel ear decomposition search (EDS) and st-numbering in graphs %A Maon,Y. %A Schieber,B. %A Vishkin, Uzi %B Theoretical Computer Science %V 47 %P 277 - 298 %8 1986/// %G eng %0 Conference Paper %B Proceedings of the 12th International Conference on Very Large Data Bases %D 1986 %T A parallel processing strategy for evaluating recursive queries %A Raschid, Louiqa %A Su,S. Y.W %B Proceedings of the 12th International Conference on Very Large Data Bases %P 412 - 419 %8 1986/// %G eng %0 Conference Paper %B AAAI %D 1986 %T A Parallel Self-Modifying Default Reasoning System %A Minker, Jack %A Perlis, Don %A Subramanian,K. %B AAAI %P 923 - 927 %8 1986/// %G eng %0 Journal Article %J Numerical Analysis %D 1986 %T Polynomial iteration for nonsymmetric indefinite linear systems %A Elman, Howard %A Streit, R. %B Numerical Analysis %P 103 - 117 %8 1986/// %G eng %0 Conference Paper %B IEEE Real-Time Systems Symposium, New Orleans, Louisiana %D 1986 %T Practicality of non-interfering checkpoints in distributed database systems %A Son,S. H %A Agrawala, Ashok K. %B IEEE Real-Time Systems Symposium, New Orleans, Louisiana %P 234 - 241 %8 1986/// %G eng %0 Journal Article %J SIAM Journal on Numerical Analysis %D 1986 %T Preconditioning by Fast Direct Methods for Nonself-Adjoint Nonseparable Elliptic Equations %A Elman, Howard %A Schultz, Martin H. %X We consider the use of fast direct methods as preconditioners for iterative methods for computing the numerical solution of nonself-adjoint elliptic boundary value problems. We derive bounds on convergence rates that are independent of discretization mesh size. For two-dimensional problems on rectangular domains, discretized on an n × n grid, these bounds lead to asymptotic operation counts of O(n2 log n log ε-1) to achieve relative error ε and O(n2(log n)2) to reach truncation error. %B SIAM Journal on Numerical Analysis %V 23 %P 44 - 57 %8 1986/02/01/ %@ 0036-1429 %G eng %U http://www.jstor.org/stable/2157450 %N 1 %0 Journal Article %J ACM SIGCHI Bulletin - Special issue: CHI '86 Conference Proceedings %D 1986 %T Seven plus or minus two central issues in human-computer interaction %A Shneiderman, Ben %X This paper offers seven issues and specific challenges for researchers and developers of human-computer interaction. These issues are: interaction styles, input techniques, output organization, response time, error handling, individual differences, explanatory and predictive theories. %B ACM SIGCHI Bulletin - Special issue: CHI '86 Conference Proceedings %V 17 %P 343 - 349 %8 1986/04// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/22339.22394 %N 4 %R 10.1145/22339.22394 %0 Journal Article %J IEEE Transactions on Computers %D 1986 %T A Special-Function Unit for Sorting and Sort-Based Database Operations %A Raschid, Louiqa %A Fei,T. %A Lam,H. %A Su,S. Y.W %K Application software %K Computer applications %K Database machines %K Hardware %K hardware sorter %K Microelectronics %K Software algorithms %K Software design %K Software systems %K sort-based algorithms for database operations %K sorting %K special-function processor %K Technology management %X Achieving efficiency in database management functions is a fundamental problem underlying many computer applications. Efficiency is difficult to achieve using the traditional general-purpose von Neumann processors. Recent advances in microelectronic technologies have prompted many new research activities in the design, implementation, and application of database machines which are tailored for processing database management functions. To build an efficient system, the software algorithms designed for this type of system need to be tailored to take advantage of the hardware characteristics of these machines. Furthermore, special hardware units should be used, if they are cost- effective, to execute or to assist the execution of these software algorithms. %B IEEE Transactions on Computers %V C-35 %P 1071 - 1077 %8 1986/12// %@ 0018-9340 %G eng %N 12 %R 10.1109/TC.1986.1676715 %0 Book %D 1985 %T Cognitive Representations of Windows and Multiple Screen Layouts of Computer Interfaces %A Norman,K. L %A Weldon,L.J. %A Shneiderman, Ben %I University of Maryland %8 1985/// %G eng %0 Journal Article %J Communications of the ACM %D 1985 %T Data-flow algorithms for parallel matrix computation %A O'Leary, Dianne P. %A Stewart, G.W. %X In this article we develop some algorithms and tools for solving matrix problems on parallel processing computers. Operations are synchronized through data-flow alone, which makes global synchronization unnecessary and enables the algorithms to be implemented on machines with very simple operating systems and communication protocols. As examples, we present algorithms that form the main modules for solving Liapounov matrix equations. We compare this approach to wave front array processors and systolic arrays, and note its advantages in handling missized problems, in evaluating variations of algorithms or architectures, in moving algorithms from system to system, and in debugging parallel algorithms on sequential machines. %B Communications of the ACM %V 28 %P 840 - 853 %8 1985/08// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/4021.4025 %N 8 %R 10.1145/4021.4025 %0 Book %D 1985 %T Detection of Surface Orientation and Motion from Texture: The Case of Planes %A Aloimonos, J. %A Chou, P. B %A Science, University of Rochester. Dept. of Computer %I Dept. of Computer Science, University of Rochester %8 1985/// %G eng %0 Journal Article %J Empirical foundations of information and software science %D 1985 %T Human factors issues of manuals, online help, and tutorials %A Shneiderman, Ben %B Empirical foundations of information and software science %V 1984 %P 107 - 107 %8 1985/// %G eng %0 Journal Article %J Behaviour & Information Technology %D 1985 %T Learning a menu selection tree: training methods compared %A Parton,Diana %A Huffman,Keith %A Pridgen,Patty %A Norman,Kent %A Shneiderman, Ben %X Abstract Abstract. Menu selection systems sometimes present learning problems for novice users. This comparison of four training methods for novice users found that the global tree diagram of the menu system was superior to command sequence and frame presentation methods, and somewhat better than trial and error. Methods were evaluated on the basis of (1) number of target nodes found, (2) mean number of selections to a target node, (3) recall of the menu structure, and (4) subjective rating of ease of learning.Abstract Abstract. Menu selection systems sometimes present learning problems for novice users. This comparison of four training methods for novice users found that the global tree diagram of the menu system was superior to command sequence and frame presentation methods, and somewhat better than trial and error. Methods were evaluated on the basis of (1) number of target nodes found, (2) mean number of selections to a target node, (3) recall of the menu structure, and (4) subjective rating of ease of learning. %B Behaviour & Information Technology %V 4 %P 81 - 91 %8 1985/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449298508901790 %N 2 %R 10.1080/01449298508901790 %0 Journal Article %J Advances in Human-Computer Interaction %D 1985 %T A model programming environment %A Shneiderman, Ben %B Advances in Human-Computer Interaction %V 1 %P 105 - 132 %8 1985/// %G eng %0 Journal Article %J Tech Report HCIL-85-03 %D 1985 %T Performance on content-free menus as a function of study method %A Schwartz,J.P. %A Norman,K. L %A Shneiderman, Ben %B Tech Report HCIL-85-03 %8 1985/// %G eng %0 Journal Article %J Conference on Software Maintenance, 1985, Sheraton Inn Washington-Northwest, November 11-13, 1985 %D 1985 %T The Psychology of Program Documentation %A Brooks,R. %A Sheppard,S. %A Shneiderman, Ben %B Conference on Software Maintenance, 1985, Sheraton Inn Washington-Northwest, November 11-13, 1985 %P 191 - 191 %8 1985/// %G eng %0 Journal Article %J Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting %D 1985 %T The Structure of Information in Online and Paper Technical Manuals %A Weldon,Linda J. %A Koved,Larry %A Shneiderman, Ben %X An experiment was conducted to compare online computer manuals to paper manuals. For each type of manual there were two different database structures – a linear (sequential) structure and a tree structure. The results showed that people using the paper manuals were faster at performing a switch setting task based on information in the manual than were people using the online manuals. No significant differences were found in speed of performance between the linear and tree structures. Nor were there any differences in the number of correct switch settings for the different types of manuals. The subjective evaluation data revealed that the online manuals were rated as better and judged to be more organized than the paper manuals. %B Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting %V 29 %P 1110 - 1113 %8 1985/10/01/ %@ 1071-1813, %G eng %U http://pro.sagepub.com/content/29/12/1110 %N 12 %R 10.1177/154193128502901207 %0 Book %D 1984 %T Designing Interactive Computer Systems: A Software Psychology Workshop %A Shneiderman, Ben %I University of California, Los Angeles, University Extension, Dept. of Engineering, Science, and Mathematics %8 1984/// %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1984 %T An experimental comparison of tabular and graphic data presentation %A Powers,Matthew %A Lashley,Conda %A Sanchez,Pamela %A Shneiderman, Ben %X We present the results of our experiment designed to test the hypothesis that more usable information can be conveyed using a combination of graphical and tabular data then by using either form alone. Our independent variables were memory (recall and non-recall) and form (tables, graphs, or both). Comprehension was measured with a multiple choice exam consisting of three types of questions (retrieve, compare, or compare/calculate answers). Both non-recall and tabular treatments significantly increased comprehension. Combinations of graphs and tables produced slower but more accurate performance. An executive should use the form with which he/she is most familiar and comfortable. %B International Journal of Man-Machine Studies %V 20 %P 545 - 566 %8 1984/06// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020737384800292 %N 6 %R 10.1016/S0020-7373(84)80029-2 %0 Journal Article %J Proceedings of the Annual Symposium on Computer Application in Medical CareProc Annu Symp Comput Appl Med Care %D 1984 %T Human Factors in Interactive Medical Systems %A Shneiderman, Ben %B Proceedings of the Annual Symposium on Computer Application in Medical CareProc Annu Symp Comput Appl Med Care %P 96 - 96 %8 1984/11/07/ %@ 0195-4210 %G eng %0 Book %D 1984 %T Let's Learn BASIC: A Kids' Introduction to BASIC Programming on IBM Personal Computers %A Shneiderman, Ben %I Little Brown and Company %8 1984/// %@ 0316787264 %G eng %0 Journal Article %J University of Maryland Computer Science Technical Report Series, CS-TR-1412 %D 1984 %T Memory For Menus: Effects of Study Mode %A Norman,K. L %A Schwartz,J.P. %A Shneiderman, Ben %X Menu selection systems require a certain amount of training in order for users to remember how to access target functions. In this experiment users studied four different types of documentation before searching for target words in a content-free menu containing no meaningful relationships among terms. Documentation was presented to subjects in one of four modes of study. In the Command Mode subjects studied sequences of choices to arrive at a desired target. In the Menu Frame Mode subjects studied individual frames in the menu system. In the Global Tree Mode subjects studied a diagram of the menu tree. Finally, in the Trial and Error Mode subjects studied the menu system by actually selecting alternatives. Although the Global Tree and Command Sequence Groups found the most target words during the test phase, the four groups did not differ significantly. On the other hand, differences in the ability to recall menu terms were significant with the command Sequence and Menu Frame Groups recalling the most terms. Furthermore, results indicated that the type of training fundamentally affected the type of information used in order to find targets. Subjects in the Command Sequence and Menu Frame Groups tended to rely on recall of menu terms. Furthermore, all groups relied heavily on memory of the menu tree except for the Command Squence Group. Results of this study are related to practical considerations for the design of menu selection systems. %B University of Maryland Computer Science Technical Report Series, CS-TR-1412 %8 1984/// %G eng %0 Journal Article %J ACM SIG Ada Letters %D 1984 %T Monitoring an Ada software development %A Basili, Victor R. %A Chang,Shih %A Gannon,John %A Katz,Elizabeth %A Panlilio-Yap,N. Monina %A Ramsey,Connie Loggia %A Zelkowitz, Marvin V %A Bailey,John %A Kruesi,Elizabeth %A Sheppard,Sylvia %B ACM SIG Ada Letters %V IV %P 32 - 39 %8 1984/07// %@ 1094-3641 %G eng %U http://doi.acm.org/10.1145/998401.998402 %N 1 %R 10.1145/998401.998402 %0 Journal Article %J ACM Comput. Surv. %D 1984 %T Response time and display rate in human performance with computers %A Shneiderman, Ben %B ACM Comput. Surv. %V 16 %P 265 - 285 %8 1984/09// %@ 0360-0300 %G eng %U http://doi.acm.org/10.1145/2514.2517 %N 3 %R 10.1145/2514.2517 %0 Journal Article %J Modern cell biology %D 1984 %T RNA splicing and the involvement of small ribonucleoproteins %A Mount, Stephen M. %A Steitz,J. A. %K Article synthèse %K Articulo sintesis %K review %K Ribonucleoprotein %K Ribonucleoproteina %K Ribonucléoprotéine %K RNA %K Sn-RNA %K splicing %B Modern cell biology %V 3 %P 249 - 297 %8 1984/// %@ 0745-3000 %G eng %U http://cat.inist.fr/?aModele=afficheN&cpsidt=9051421 %0 Journal Article %J Journal of the ACM (JACM) %D 1984 %T Solving NP-hard problems on graphs that are almost trees and an application to facility location problems %A Gurevich,Y. %A Stockmeyer,L. %A Vishkin, Uzi %B Journal of the ACM (JACM) %V 31 %P 459 - 473 %8 1984/// %G eng %N 3 %0 Journal Article %J Software: Practice and Experience %D 1983 %T The effect of scope delimiters on program comprehension %A Sykes,Floyd %A Tillman,Raymond T %A Shneiderman, Ben %K Programming language design %K Scope delimiters %K Software psychology %X Scope delimiters, such as BEGIN-END or DO-END, are used in many programming languages, but they can lengthen and clutter a program listing. This paper provides experimental evidence that ENDIF or ENDWHILE statement terminators make for easier to comprehend programs than BEGIN-END pairs surrounding compound statements. %B Software: Practice and Experience %V 13 %P 817 - 824 %8 1983/09/01/ %@ 1097-024X %G eng %U http://onlinelibrary.wiley.com/doi/10.1002/spe.4380130908/abstract %N 9 %R 10.1002/spe.4380130908 %0 Journal Article %J Journal of Computer-Based Instruction %D 1983 %T An empirical comparison of two PLATO text editors. %A Shneiderman, Ben %A Hill,R. %A Jacob,R. %A Mah,W. %X Two PLATO system text editors were evaluated with 14 nonprogrammers and 14 programmers who were either university staff or college students. Half of each group learned the "plain" editor, which had 8 commands and 3 screens of HELP material; while the other half of each group learned the "fancy" editor, which had 15 commands and 1 screen of HELP material. The fancy editor is a subset of the widely used TUTOR editor, but nonprogrammers preferred the plain editor. Faster learning, faster performance, fewer invocations of HELP material, and fewer requests for human assistance were characteristic of the plain editor. %B Journal of Computer-Based Instruction %V 10 %P 43 - 50 %8 1983/// %G eng %0 Journal Article %J NASA. Goddard Space Flight Center Proc. of the Eighth Ann. Software Eng. Workshop 11 p(SEE N 84-23137 13-61) %D 1983 %T Evaluating multiple coordinated windows for programming workstations %A Shneiderman, Ben %A GRANTHAM,C. %A Norman,K. %A ROGERS,J. %A Roussopoulos, Nick %X Programmers might benefit from larger screens with multiple windows or multiple screens, especially if convenient coordination among screens are arranged. Uses for multiple coordinated displays in a programmers workstation are explored. Initial efforts focus on the potential applications, a command language for coordinating the displays, and the psychological basis for effective utilization so as to avoid information overload. Subsequent efforts will be devoted to implementing the concepts and performing controlled psychologically oriented experiments to validate the hypotheses. %B NASA. Goddard Space Flight Center Proc. of the Eighth Ann. Software Eng. Workshop 11 p(SEE N 84-23137 13-61) %8 1983/// %G eng %0 Journal Article %J SIGCHI Bull. %D 1983 %T High-tech can stimulate creative action: the increased ease-of-use of computers supports individual competence and productive work %A Shneiderman, Ben %X Jokes about the complexity and error-prone use of computer systems reflect the experience of many people. Computer anxiety, terminal terror, and network neurosis are modern maladies - but help is on the way! %B SIGCHI Bull. %V 14 %P 6 - 7 %8 1983/04// %@ 0736-6906 %G eng %U http://doi.acm.org/10.1145/1044188.1044189 %N 4 %R 10.1145/1044188.1044189 %0 Journal Article %J Proceedings COMPCON %D 1983 %T Human engineering management plan for interactive systems %A Shneiderman, Ben %B Proceedings COMPCON %V 83 %8 1983/// %G eng %0 Book Section %B Enduser Systems and Their Human FactorsEnduser Systems and Their Human Factors %D 1983 %T Human factors of interactive software %A Shneiderman, Ben %E Blaser,Albrecht %E Zoeppritz,Magdalena %X There is intense interest about human factors issues in interactive computer systems for life-critical applications, industrial/commercial uses, and personal computing in the office or home. Primary design goals include proper functionality, adequate reliability, suitable cost, and adherence to schedule. Measurable human factors issues include time to learn, speed of performance, rate of errors, subjective satisfaction, and retention over time. Specific human factors acceptance tests are described as a natural complement to hardware and software acceptance tests. Project management ideas, information resources, and potential research directions are presented. %B Enduser Systems and Their Human FactorsEnduser Systems and Their Human Factors %S Lecture Notes in Computer Science %I Springer Berlin / Heidelberg %V 150 %P 9 - 29 %8 1983/// %@ 978-3-540-12273-9 %G eng %U http://dx.doi.org/10.1007/3-540-12273-7_16 %0 Journal Article %J Elliptic problem solvers II %D 1983 %T The (new) Yale sparse matrix package %A Eisenstat, S. C %A Elman, Howard %A Schultz, M. H %A Sherman, A. H %B Elliptic problem solvers II %P 45 - 52 %8 1983/// %G eng %0 Journal Article %J Communications of the ACM %D 1983 %T Program indentation and comprehensibility %A Miara,Richard J. %A Musselman,Joyce A. %A Navarro,Juan A. %A Shneiderman, Ben %K indentation %K program format %K program readability %B Communications of the ACM %V 26 %P 861 - 867 %8 1983/11// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/182.358437 %N 11 %R 10.1145/182.358437 %0 Journal Article %J Nature %D 1983 %T RNA processing: Lessons from mutant globins %A Mount, Stephen M. %A Steitz,Joan %B Nature %V 303 %P 380 - 381 %8 1983/06/02/ %@ 0028-0836 %G eng %U http://ukpmc.ac.uk/abstract/MED/6855891/reload=0;jsessionid=6LgsyNwDXt7LrUcZkgTq.12 %N 5916 %R 10.1038/303380a0 %0 Journal Article %J Cold Spring Harbor Symposia on Quantitative BiologyCold Spring Harb Symp Quant Biol %D 1983 %T Small Ribonucleoproteins from Eukaryotes: Structures and Roles in RNA Biogenesis %A Steitz,J. A. %A Wolin,S. L. %A Rinke,J. %A Pettersson,I. %A Mount, Stephen M. %A Lerner,E. A. %A Hinterberger,M. %A Gottlieb,E. %X Detailed reviews describing work presented at the annual Cold Spring Harbor Symposia on Quantitative Biology %B Cold Spring Harbor Symposia on Quantitative BiologyCold Spring Harb Symp Quant Biol %V 47 %P 893 - 900 %8 1983/01/01/ %@ 0091-7451, 1943-4456 %G eng %U http://symposium.cshlp.org/content/47/893 %R 10.1101/SQB.1983.047.01.103 %0 Journal Article %J Cell %D 1983 %T Splicing of messenger RNA precursors is inhibited by antisera to small nuclear ribonucleoprotein %A Padgett,Richard A. %A Mount, Stephen M. %A Steitz,Joan A. %A Sharp,Phillip A. %B Cell %V 35 %P 101 - 107 %8 1983/11// %@ 00928674 %G eng %U http://ukpmc.ac.uk/abstract/MED/6194895 %N 1 %R 10.1016/0092-8674(83)90212-X %0 Journal Article %J Cell %D 1983 %T The U1 small nuclear RNA-protein complex selectively binds a 5′ splice site in vitro %A Mount, Stephen M. %A Pettersson,Ingvar %A Hinterberger,Monique %A Karmas,Aavo %A Steitz,Joan A. %X The ability of purified U1 small nuclear RNA-protein complexes (U1 snRNPs) to bind in vitro to two RNAs transcribed from recombinant DNA clones by bacteriophage T7 RNA polymerase has been studied. A transcript which contains sequences corresponding to the small intron and flanking exons of the major mouse beta-globin gene is bound in marked preference to an RNA devoid of splice site sequences. The site of U1 snRNP binding to the globin RNA has been defined by T1 ribonuclease digestion of the RNA-U1 snRNP complex. A 15-17-nucleotide region, including the 5' splice site, remains undigested and complexed with the snRNP such that it can be co-precipitated by antibodies directed against the U1 snRNP. Partial proteinase K digestion of the U1 snRNP abolishes interaction with the globin RNA, indicating that the snRNP proteins contribute significantly to RNA binding. No RNA cleavage, splicing, or recognition of the 3' splice site by U1 snRNPs has been detected. Our results are discussed in terms of the probable role of U1 snRNPs in the messenger RNA splicing of eucaryotic cell nuclei. %B Cell %V 33 %P 509 - 518 %8 1983/06// %@ 00928674 %G eng %U http://ukpmc.ac.uk/abstract/MED/6190573 %N 2 %R 10.1016/0092-8674(83)90432-4 %0 Journal Article %J SIAM Journal on Numerical Analysis %D 1983 %T Variational Iterative Methods for Nonsymmetric Systems of Linear Equations %A Eisenstat, Stanley C. %A Elman, Howard %A Schultz, Martin H. %X We consider a class of iterative algorithms for solving systems of linear equations where the coefficient matrix is nonsymmetric with positive-definite symmetric part. The algorithms are modelled after the conjugate gradient method, and are well suited for large sparse systems. They do not make use of any associated symmetric problems. Convergence results and error bounds are presented. %B SIAM Journal on Numerical Analysis %V 20 %P 345 - 357 %8 1983/04/01/ %@ 0036-1429 %G eng %U http://www.jstor.org/stable/2157222 %N 2 %0 Journal Article %J ACM Trans. Database Syst. %D 1982 %T An architecture for automatic relational database sytem conversion %A Shneiderman, Ben %A Thomas,Glenn %K automatic conversion %K Database systems %K relational model %K transformations %X Changes in requirements for database systems necessitate schema restructuring, database translation, and application or query program conversion. An alternative to the lengthy manual revision process is proposed by offering a set of 15 transformations keyed to the relational model of data and the relational algebra. Motivations, examples, and detailed descriptions are provided. %B ACM Trans. Database Syst. %V 7 %P 235 - 257 %8 1982/06// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/319702.319724 %N 2 %R 10.1145/319702.319724 %0 Conference Paper %B Proceedings of the June 7-10, 1982, national computer conference %D 1982 %T Automatic database system conversion: schema revision, data translation, and source-to-source program transformation %A Shneiderman, Ben %A Thomas,Glenn %X Changing data requirements present database administrators with a difficult problem: the revision of the schema, the translation of the stored database, and the conversion of the numerous application programs. This paper describes an automatic database system conversion facility which provides one approach for coping with this problem. The Pure Definition Language and the Pure Manipulation Language have been designed to facilitate the conversions specified in the Pure Transformation Language. Two conversions and their effect on retrievals are demonstrated. %B Proceedings of the June 7-10, 1982, national computer conference %S AFIPS '82 %I ACM %C New York, NY, USA %P 579 - 587 %8 1982/// %@ 0-88283-035-X %G eng %U http://doi.acm.org/10.1145/1500774.1500849 %R 10.1145/1500774.1500849 %0 Conference Paper %B 23rd Annual Symposium on Foundations of Computer Science %D 1982 %T A complexity theory for unbounded fan-in parallelism %A Chandra,A. K %A Stockmeyer,L. J %A Vishkin, Uzi %B 23rd Annual Symposium on Foundations of Computer Science %P 1 - 13 %8 1982/// %G eng %0 Journal Article %J Commun. ACM %D 1982 %T Control flow and data structure documentation: two experiments %A Shneiderman, Ben %K data structure diagrams %K pseudocode %X Two experiments were carried out to assess the utility of external documentation aids such as macro flowcharts, pseudocode, data structure diagrams, and data structure descriptions. A 223 line Pascal program which manipulates four arrays was used. The program interactively handles commands that allow the user to manage five lists of items. A comprehension test was given to participants along with varying kinds of external documentation. The results indicate that for this program the data structure information was more helpful than the control flow information, independently of whether textual or graphic formats were used. %B Commun. ACM %V 25 %P 55 - 63 %8 1982/01// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/358315.358391 %N 1 %R 10.1145/358315.358391 %0 Journal Article %J Communications of the ACM %D 1982 %T Designing computer system messages %A Shneiderman, Ben %K error messages %K human/computer interaction %K system messages %X tive computer systems and studies of their users, we have become increasingly aware of the importance of system messages. Novice users are unimpressed with CPU speeds, disk storage capabilities, or elegant file structures. For them, the system appears only in the form of the messages on their screens or printers. So when novices encounter violent messages such as “FATAL ERROR, RUN ABORTED”, vague phases like “ILLEGAL CMD”, or obscure codes such as “OC7” or “IEH2191”, they are understandably shaken, confused, dismayed, and discouraged from continuing. The negative image that computer systems sometimes generate is, we believe, largely due to the difficulties users experience when they make mistakes or are unsure about what to do next. %B Communications of the ACM %V 25 %P 610 - 611 %8 1982/09// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/358628.358639 %N 9 %R 10.1145/358628.358639 %0 Book %D 1982 %T Directions in human/computer interaction. %A Badre,A. %A Shneiderman, Ben %X This book presents current research in interactive system design and human information processing. The contributions offer a mixture of survey, theory, practical recommendations and experimental results reflecting the diversity of research efforts in human/computer interaction. %I ABLEX PUBL. CORP. %C 355 CHESTNUT ST., NORWOOD, NJ 07648, USA %8 1982/// %G eng %0 Journal Article %J Bulletin of the American Society for Information ScienceBulletin of the American Society for Information Science %D 1982 %T Fighting for the User. %A Shneiderman, Ben %K Automation %K design %K Formative Evaluation %K Human Factors Engineering %K Information systems %K Input Output Devices %K Man Machine Systems %K Online Systems %K performance %K Summative Evaluation %K Systems Development %K Technological Advancement %X Redesign of human-computer interface for users of computerized information systems can make substantial difference in training time, performance speed, error rates, and user satisfaction. Information and computer scientists are using controlled psychologically oriented experimentation, and evaluations during system development and active use to test design impact on human performance measures. (EJS) %B Bulletin of the American Society for Information ScienceBulletin of the American Society for Information Science %V 9 %P 27 - 29 %8 1982/// %G eng %U http://www.eric.ed.gov/ERICWebPortal/detail?accno=EJ276739 %N 2 %0 Journal Article %J Behaviour & Information Technology %D 1982 %T The future of interactive systems and the emergence of direct manipulation %A Shneiderman, Ben %X Abstract This paper suggests three motivations for the strong interest in human factors' aspects of user interfaces and reviews five design issues: command language versus menu selection, response time and display rates, wording of system messages, on-line tutorials, explanations and help messages and hardware devices. Five methods and tools for system development are considered: participatory design, specification methods, software implementation tools, pilot studies and acceptance tests and evolutionary refinement based on user feedback. The final portion of the paper presents direct manipulation, an approach which promises to become widely used in interactive systems. Direct manipulation involves representation of the object of interest, rapid incremental reversible actions and physical action instead of complex syntax.Abstract This paper suggests three motivations for the strong interest in human factors' aspects of user interfaces and reviews five design issues: command language versus menu selection, response time and display rates, wording of system messages, on-line tutorials, explanations and help messages and hardware devices. Five methods and tools for system development are considered: participatory design, specification methods, software implementation tools, pilot studies and acceptance tests and evolutionary refinement based on user feedback. The final portion of the paper presents direct manipulation, an approach which promises to become widely used in interactive systems. Direct manipulation involves representation of the object of interest, rapid incremental reversible actions and physical action instead of complex syntax. %B Behaviour & Information Technology %V 1 %P 237 - 256 %8 1982/// %@ 0144-929X %G eng %U http://www.tandfonline.com/doi/abs/10.1080/01449298208914450 %N 3 %R 10.1080/01449298208914450 %0 Journal Article %J IEEE Transactions on Systems, Man and Cybernetics %D 1982 %T Multiparty Grammars and Related Features for Defining Interactive Systems %A Shneiderman, Ben %K Application software %K Computer aided instruction %K Computer displays %K Computer languages %K Computer networks %K Debugging %K HUMANS %K interactive systems %K Protocols %K Writing %X Multiparty grammars are introduced which contain labeled nonterminals to indicate the party that produces the terminal string. For interactive person-computer systems, both the user commands and system responses can be described by the linked BNF grammars. Multiparty grammars may also be used to describe communication among several people (by way of computers or in normal dialogue), network protocols among several machines, or complex interactions involving several people and machines. Visual features such as underlining, reversal, blinking, and color, window declarations, and dynamic operations dependent on cursor movement are also covered. %B IEEE Transactions on Systems, Man and Cybernetics %V 12 %P 148 - 154 %8 1982/03// %@ 0018-9472 %G eng %N 2 %R 10.1109/TSMC.1982.4308798 %0 Journal Article %J Journal of Algorithms %D 1982 %T An O (logn) parallel connectivity algorithm %A Shiloach,Y. %A Vishkin, Uzi %B Journal of Algorithms %V 3 %P 57 - 67 %8 1982/// %G eng %N 1 %0 Journal Article %J Journal of Algorithms %D 1982 %T An O (n2log n) parallel max-flow algorithm %A Shiloach,Y. %A Vishkin, Uzi %B Journal of Algorithms %V 3 %P 128 - 146 %8 1982/// %G eng %N 2 %0 Journal Article %J Journal of Algorithms %D 1982 %T parallel MAX-FLOW algorithm %A Shiloach,Y. %A Vishkin, Uzi %A An,O. %B Journal of Algorithms %V 3 %P 128 - 146 %8 1982/// %G eng %0 Journal Article %J ACM SIGMOD Record %D 1982 %T Response to N. C. Rowe's review %A Shneiderman, Ben %B ACM SIGMOD Record %V 13 %P 98 - 98 %8 1982/09// %@ 0163-5808 %G eng %U http://doi.acm.org/10.1145/984514.984521 %N 1 %R 10.1145/984514.984521 %0 Journal Article %J Princess Takamatsu symposiaInt. Symp. Princess Takamatsu Cancer Res. Fund %D 1982 %T Structure and function of small ribonucleoproteins from eukaryotic cells %A Steitz,J. A. %A Berg,C %A Gottlieb,E. %A Hardin,J A %A Hashimoto,C %A Hendrick,J P %A Hinterberger,M. %A Krikeles,M %A Lerner,M R %A Mount, Stephen M. %K Antigen-Antibody Complex %K Autoantibodies %K HUMANS %K Lupus Erythematosus, Systemic %K Nucleoproteins %K Ribonucleoproteins %K RNA Polymerase III %K Transcription, Genetic %X Autoantibodies from patients with systemic lupus erythematosus and other related diseases have been used to identify and study small RNA-protein complexes from mammalian cells. Properties of three previously described and several new classes of small ribonucleoproteins (RNPs) are reviewed. The sequence of Drosophila U1 RNA reveals that the region proposed to pair with 5' splice junctions is conserved, while that proposed to interact with 3' junctions diverges; this forces some revision of the model for U1 small nuclear (sn)RNP participation in hnRNA splicing. Further characterization of the Ro and La small RNPs has shown that the Ro small cytoplasmic (sc)RNPs are a subclass of La RNPs. Both tRNA and 5S rRNA precursors are at least transiently associated with the La protein. This raises the possibility that the La protein may be an RNA polymerase III transcription factor. %B Princess Takamatsu symposiaInt. Symp. Princess Takamatsu Cancer Res. Fund %V 12 %P 101 - 107 %8 1982/// %G eng %U http://www.ncbi.nlm.nih.gov/pubmed/7166547 %0 Book %D 1982 %T System message design: Guidelines and experimental results %A Shneiderman, Ben %I Ablex Publishing Company: Norwood, NJ %8 1982/// %G eng %0 Journal Article %J Gathering Information for Problem Formulation %D 1982 %T SYSTEM MESSAGE GUIDELINES: POSITIVE TONE, CONSTRUCTIVE, SPECIFIC, AND USER CENTERED %A Shneiderman, Ben %B Gathering Information for Problem Formulation %8 1982/// %G eng %0 Journal Article %J ACM SIGCSE Bulletin %D 1982 %T Teaching software psychology experimentation through team projects %A Shneiderman, Ben %B ACM SIGCSE Bulletin %V 14 %P 38 - 40 %8 1982/09// %@ 0097-8418 %G eng %U http://doi.acm.org/10.1145/989322.989331 %N 3 %R 10.1145/989322.989331 %0 Journal Article %J SIAM Journal on Scientific and Statistical Computing %D 1981 %T A Bidiagonalization-Regularization Procedure for Large Scale Discretizations of Ill-Posed Problems %A O'Leary, Dianne P. %A Simmons,John A. %K first kind integral equation %K ill-posed problems %K Lanczos algorithm %K regularization %X In this paper, we consider ill-posed problems which discretize to linear least squares problems with matrices $K$ of high dimensions. The algorithm proposed uses $K$ only as an operator and does not need to explicitly store or modify it. A method related to one of Lanczos is used to project the problem onto a subspace for which $K$ is bidiagonal. It is then an easy matter to solve the projected problem by standard regularization techniques. These ideas are illustrated with some integral equations of the first kind with convolution kernels, and sample numerical results are given. %B SIAM Journal on Scientific and Statistical Computing %V 2 %P 474 - 489 %8 1981/// %G eng %U http://link.aip.org/link/?SCE/2/474/1 %N 4 %R 10.1137/0902037 %0 Journal Article %J Software Engineering, IEEE Transactions on %D 1981 %T Comments from a Letter Written on July 19, 1978 %A Shneiderman, Ben %B Software Engineering, IEEE Transactions on %P 370 - 371 %8 1981/// %G eng %N 4 %0 Journal Article %J ACM SIGSOC Bulletin %D 1981 %T Direct manipulation: A step beyond programming languages (abstract only) %A Shneiderman, Ben %X Direct manipulation is a style of interaction which has been used by implementers of widely varying systems. Direct manipulation permits novice users access to powerful facilities without the burden of learning to use a complex syntax and lengthy list of commands. Display editors use direct manipulation more than line editors. Form-fill-in is more direct than tag fields and delimiters. Spatial data management is more direct than query-by-example, which is more direct than SEQUEL. Computer arcade games and Visicalc are further examples.Direct manipulation involves three interrelated techniques:1. Provide a physically direct way of moving a cursor or manipulating the objects of interest.2. Present a concrete visual representation of the objects of interest and immediately change the view to reflect operations.3. Avoid using a command language and depend? on operations applied to the cognitive model which is shown on the display. %B ACM SIGSOC Bulletin %V 13 %P 143 - 143 %8 1981/05// %@ 0163-5794 %G eng %U http://doi.acm.org/10.1145/1015579.810991 %N 2-3 %R 10.1145/1015579.810991 %0 Journal Article %J Proceedings of the 44th ASIS Annual Meeting %D 1981 %T Exploratory research on training aids for naive users of interactive systems %A Stevens,P. %A Shneiderman, Ben %B Proceedings of the 44th ASIS Annual Meeting %V 18 %P 65 - 67 %8 1981/// %G eng %0 Journal Article %J Journal of Algorithms %D 1981 %T Finding the maximum, merging, and sorting in a parallel computation model %A Shiloach,Y. %A Vishkin, Uzi %B Journal of Algorithms %V 2 %P 88 - 102 %8 1981/// %G eng %N 1 %0 Journal Article %J Human factors in software development %D 1981 %T Human Aspects of Computing Editor Control Flow and Data Structure Documentation: Two Experiments %A Shneiderman, Ben %B Human factors in software development %P 365 - 365 %8 1981/// %G eng %0 Journal Article %J Proceedings-Compcon %D 1981 %T HUMAN FACTORS ISSUES IN DESIGNING INTERACTIVE SYSTEMS %A Shneiderman, Ben %B Proceedings-Compcon %V 25 %P 116 - 116 %8 1981/// %G eng %0 Journal Article %J ACM SIGSOC Bulletin %D 1981 %T Human factors studies with system message styles (abstract only) %A Shneiderman, Ben %X Computer systems often contain messages which are imprecise ('SYNTAX ERROR'), hostile ('FATAL ERROR, RUN ABORTED'), cryptic ('IEH291H'), or obscure ('CTL DAMAGE, TRANS ERR'). Such messages may be acceptable to computer professionals who regularly use a specific system, but they lead to frustration for novices and for professionals who are using new features or facilities.We have conducted five studies using COBOL compiler syntax errors and text editor command errors to measure the impact of improving the wording of system messages. The results indicate that increased specificity, more positive tone, and greater clarity can improve correction rates and user satisfaction.An overview of the experimental results will be presented along with guidelines for writing system messages. %B ACM SIGSOC Bulletin %V 13 %P 138– - 138– %8 1981/05// %@ 0163-5794 %G eng %U http://doi.acm.org/10.1145/1015579.810979 %N 2-3 %R 10.1145/1015579.810979 %0 Journal Article %J Information Systems %D 1981 %T A note on human factors issues of natural language interaction with database systems %A Shneiderman, Ben %X Efforts at constructing natural language computer systems have turned to implementing database query facilities. If the application domain is limited there is renewed hope for the success of natural language interfaces. This paper reviews arguments for and against such facilities, stresses research in human semantic knowledge, and emphasizes controlled psychologically oriented experimentation. The results of recent experiments are reported. %B Information Systems %V 6 %P 125 - 129 %8 1981/// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/030643798190034X %N 2 %R 10.1016/0306-4379(81)90034-X %0 Conference Paper %B Proceedings of the eighteenth annual computer personnel research conference %D 1981 %T Putting the human factor into systems development %A Shneiderman, Ben %X As the community of computer users expands beyond experienced professionals to encompass novice users with little technical training, human factors considerations must play a larger role. “Computer shock” and “terminal terror” cannot be cured, they must be prevented by more careful human engineering during the system design phase. This paper offers four approaches to including human factors considerations during system design. These approaches focus on increasing user involvement and emphasize extensive pilot testing. Human factors cannot be added as refinements to a completed design; they must be a central concern during the initial requirements analysis and through every design stage. %B Proceedings of the eighteenth annual computer personnel research conference %S SIGCPR '81 %I ACM %C New York, NY, USA %P 1 - 13 %8 1981/// %@ 0-89791-044-3 %G eng %U http://doi.acm.org/10.1145/800051.801845 %R 10.1145/800051.801845 %0 Journal Article %J Nucleic Acids ResearchNucl. Acids Res. %D 1981 %T Sequence of U1 RNA from Drosophila melanogaster: implications for U1 secondary structure and possible involvement in splicing %A Mount, Stephen M. %A Steitz,Joan A. %X U1 RNA from cultured Drosophila melanogaster cells (Kc) was identified by its ability to be recognized, as an RNP, by anti-(U1)RNP antibodies from human lupus patients. Its sequence was deduced largely from direct analysis of the RNA molecule and then confirmed by DNA sequence determinations on a genomic clone isolated by hybridization to Drosophila U1 RNA. The Drosophila U1 RNA sequence exhibits 72% agreement with human U1 RNA. Nucleotides 3-11, which are complementary to the entire consensus sequence for donor (5′) splice junctions in hnRNA, and to part of the acceptor (3′) consensus, are exactly conserved. However, nucleotides 14-21, postulated to interact only with acceptor junctions, differ. Comparison of the Drosophila U1 sequence with vertebrate U1 sequences allows a particular secondary structure model to be preferred over others. These results are consistent with the hypothesis that U1 snRNPs are involved in splicing, but suggest specific modifications of the model detailing molecular interactions between U1 RNA and hnRNA during the splicing reaction. %B Nucleic Acids ResearchNucl. Acids Res. %V 9 %P 6351 - 6368 %8 1981/12/11/ %@ 0305-1048, 1362-4962 %G eng %U http://nar.oxfordjournals.org/content/9/23/6351 %N 23 %R 10.1093/nar/9.23.6351 %0 Journal Article %J Nucleic Acids ResearchNucl. Acids Res. %D 1981 %T Transcription of cloned tRNA and 5S RNA genes in a Drosophila cell free extract %A Dingermann,Theodor %A Sharp,Stephen %A Appel,Bernd %A DeFranco,Donald %A Mount, Stephen M. %A Heiermann,Reinhard %A Pongs,Olaf %A Söll,Dieter %X We describe the preparation of a cell-free extract from Drosophila Kc cells which allows transcription of a variety of cloned eukaryotic RNA polymerase III genes. The extract has low RNA-processing nuclease activity and thus the major products obtained are primary transcripts. %B Nucleic Acids ResearchNucl. Acids Res. %V 9 %P 3907 - 3918 %8 1981/08/25/ %@ 0305-1048, 1362-4962 %G eng %U http://nar.oxfordjournals.org/content/9/16/3907 %N 16 %R 10.1093/nar/9.16.3907 %0 Journal Article %J Nature %D 1980 %T Are snRNPs involved in splicing? %A Lerner,Michael R. %A Boyle,John A. %A Mount, Stephen M. %A Wolin,Sandra L. %A Steitz,Joan A. %X Discrete, stable small RNA molecules are found in the nuclei of cells from a wide variety of eukaryotic organisms. Many of these small nuclear RNA (snRNA) species, which range in size from about 90 to 220 nucleotides, have been well-characterised biochemically, and some sequenced. However, their function has remained obscure. The most abundant snRNA species exist as a closely related set of RNA–protein complexes called small nuclear ribonucleoproteins (snRNPs). snRNPs are the antigens recognised by antibodies from some patients with lupus erythematosus (LE), an autoimmune rheumatic disease. Anti-RNP antibodies from lupus sera selectively precipitate snRNP species containing Ula7 and Ulb9 RNAs from mouse Ehrlich ascites cell nuclei, whereas anti-Sm antibodies bind these snRNPs and four others containing U2, U4, US and U6 RNAs. Both antibody systems precipitate the same seven prominent nuclear proteins (molecular weight 12,000–32,000). All molecules of the snRNAs U1, U2, U4, U5 and U6 appear to exist in the form of antigenic snRNPs9. The particles sediment at about 10S and each probably contains a single snRNA molecule. Indirect immunofluorescence studies using anti-RNP and anti-Sm sera confirm the nuclear (but non-nucleolar) location of the antigenic snRNPs. Here we present several lines of evidence that suggest a direct involvement of snRNPs in the splicing of hnRNA. Most intriguing is the observation that the nucleotide sequence at the 5′ end of U1 RNA exhibits extensive complementarity to those across splice junctions in hnRNA molecules. %B Nature %V 283 %P 220 - 224 %8 1980/01/10/ %G eng %U http://www.nature.com/nature/journal/v283/n5743/abs/283220a0.html %N 5743 %R 10.1038/283220a0 %0 Journal Article %J IEEE Computer Software and Applications Conference %D 1980 %T Automatic database system conversion: A transformation language approach to sub-schema implementation %A Thomas,G. %A Shneiderman, Ben %B IEEE Computer Software and Applications Conference %P 80 - 88 %8 1980/// %G eng %0 Journal Article %J International Journal of Man-Machine Studies %D 1980 %T An experiment using memorization/reconstruction as a measure of programmer ability %A Di Persio,Tom %A Isbister,Dan %A Shneiderman, Ben %X Measuring the abilities of programmers in a classroom or organizational setting is not a trivial task. Current approaches are not always accurate or reliable. This paper describes an experiment which provides evidence that performance on a memorization/recon- struction test can be used as a measure or predictor of programmer ability. The contribution of indentation in program comprehension is also examined. %B International Journal of Man-Machine Studies %V 13 %P 339 - 354 %8 1980/10// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020737380800472 %N 3 %R 10.1016/S0020-7373(80)80047-2 %0 Journal Article %J Information & Management %D 1980 %T Hardware options, evaluation metrics, and a design sequence for interactive information systems %A Shneiderman, Ben %K casual users %K design sequence %K evaluation metrics %K ferminal design %K human factor %K Interactive information systems %X Interactive information systems must satisfy a wide variety of users, serve a broad range of tasks, and be suited to diverse hardware environments. This paper concentrates on three aspects of interactive information systems design: hardware options, evaluation metrics, and a possible design sequence. Rigorous pilot studies are emphasized, and supporting experimental evidence is offered. %B Information & Management %V 3 %P 3 - 18 %8 1980/// %@ 0378-7206 %G eng %U http://www.sciencedirect.com/science/article/pii/0378720680900269 %N 1 %R 10.1016/0378-7206(80)90026-9 %0 Journal Article %J Proceedings-Compcon %D 1980 %T HUMAN FACTORS EXPERIMENTS FOR REFINING INTERACTIVE SYSTEM DESIGNS %A Shneiderman, Ben %B Proceedings-Compcon %V 21 %P 123 - 123 %8 1980/// %G eng %0 Journal Article %J Software Configuration Management %D 1980 %T Human Factors of Software Design and Development %A Weiser,M. %A Shneiderman, Ben %B Software Configuration Management %P 67 - 67 %8 1980/// %G eng %0 Conference Paper %B Proceedings of the 18th annual meeting on Association for Computational Linguistics %D 1980 %T Natural vs. precise concise languages for human operation of computers: research issues and experimental approaches %A Shneiderman, Ben %X This paper raises concerns that natural language front ends for computer systems can limit a researcher's scope of thinking, yield inappropriately complex systems, and exaggerate public fear of computers. Alternative modes of computer use are suggested and the role of psychologically oriented controlled experimentation is emphasized. Research methods and recent experimental results are briefly reviewed. %B Proceedings of the 18th annual meeting on Association for Computational Linguistics %S ACL '80 %I Association for Computational Linguistics %C Stroudsburg, PA, USA %P 139 - 141 %8 1980/// %G eng %U http://dx.doi.org/10.3115/981436.981478 %R 10.3115/981436.981478 %0 Conference Paper %B Proceedings of the sixth international conference on Very Large Data Bases - Volume 6 %D 1980 %T Path expressions for complex queries and automatic database program conversion %A Shneiderman, Ben %A Thomas,Glenn %K data definition language %K data manipulation language %K Database conversion %K Database systems %K path expressions %K program conversion %K query languages %K transformation language %X Our efforts to develop an automatic database system conversion facility yielded a powerful, yet simple query language which was designed for ease of conversion. The path expression of this query language is a convenient and appealing notation for describing complex traversals with multiple boolean qualifications. This paper describes the path expression, shows how automatic conversions can be done, introduces the boolean functions as part of the basic path expression, offers four extensions (path macros, implied path, path replacement, and path optimization), and discusses some implementation issues. %B Proceedings of the sixth international conference on Very Large Data Bases - Volume 6 %S VLDB '80 %I VLDB Endowment %P 33 - 44 %8 1980/// %G eng %U http://dl.acm.org/citation.cfm?id=1286887.1286891 %0 Book %D 1980 %T Software psychology: Human factors in computer and information systems (Winthrop computer systems series) %A Shneiderman, Ben %I Winthrop Publishers %8 1980/// %@ 0876268165 %G eng %0 Conference Paper %B Fifth International Conference on Very Large Data Bases, 1979 %D 1979 %T Database Program Conversion: A Framework For Research %A Taylor,R. W %A Fry,J. P %A Shneiderman, Ben %A Smith,D. C.P %A Su,S. Y.W %K Application software %K Costs %K Data conversion %K Data structures %K Databases %K Delay %K Prototypes %K Technology planning %K US Government %K Writing %B Fifth International Conference on Very Large Data Bases, 1979 %I IEEE %P 299 - 312 %8 1979/10/03/5 %G eng %R 10.1109/VLDB.1979.718145 %0 Journal Article %J Mathematics of Computation %D 1979 %T Estimating the Largest Eigenvalue of a Positive Definite Matrix %A O'Leary, Dianne P. %A Stewart, G.W. %A Vandergraft,James S. %B Mathematics of Computation %V 33 %P 1289 - 1292 %8 1979/// %G eng %0 Journal Article %J Computer %D 1979 %T Human Factors Experiments in Designing Interactive Systems %A Shneiderman, Ben %K Application software %K Computer languages %K Design engineering %K Home computing %K human factors %K interactive systems %K Process design %K Testing %X Successful industrial design gracefully unites esthetics and function at minimum cost. However, designers face special problems when they apply their skills to interactive computer systems. %B Computer %V 12 %P 9 - 19 %8 1979/12// %@ 0018-9162 %G eng %N 12 %R 10.1109/MC.1979.1658571 %0 Book Section %B Sparse Matrix Proceedings 1978Sparse Matrix Proceedings 1978 %D 1979 %T Linear programming Problems Arising from Partial Differential Equations %A O'Leary, Dianne P. %E Duff,Iain S. %E Stewart, G.W. %B Sparse Matrix Proceedings 1978Sparse Matrix Proceedings 1978 %I SIAM Press %C Philadelphia %P 25 - 40 %8 1979/// %G eng %0 Book %D 1979 %T Principles of software engineering and design %A Zelkowitz, Marvin V %A Shaw,A. C %A Gannon,J. D %I Prentice Hall Professional Technical Reference %8 1979/// %@ 013710202X %G eng %0 Journal Article %J International Journal of Parallel Programming %D 1979 %T Syntactic/semantic interactions in programmer behavior: A model and experimental results %A Shneiderman, Ben %A Mayer,Richard %X This paper presents a cognitive framework for describing behaviors involved in program composition, comprehension, debugging, modification, and the acquisition of new programming concepts, skills, and knowledge. An information processing model is presented which includes a long-term store of semantic and syntactic knowledge, and a working memory in which problem solutions are constructed. New experimental evidence is presented to support the model of syntactic/semantic interaction. %B International Journal of Parallel Programming %V 8 %P 219 - 238 %8 1979/// %@ 0885-7458 %G eng %U http://dx.doi.org/10.1007/BF00977789 %N 3 %0 Book %D 1978 %T Data Bases: Improving Usability and Effectiveness %A Shneiderman, Ben %I Academic Press, Inc. %C Orlando, FL, USA %8 1978/// %@ 0126421501 %G eng %0 Journal Article %J ACM Trans. Database Syst. %D 1978 %T Improving the human factors aspect of database interactions %A Shneiderman, Ben %K Data models %K Database systems %K experimentation %K human factors %K natural language interfaces %K Psychology %K query languages %X The widespread dissemination of computer and information systems to nontechnically trained individuals requires a new approach to the design and development of database interfaces. This paper provides the motivational background for controlled psychological experimentation in exploring the person/machine interface. Frameworks for the reductionist approach are given, research methods discussed, research issues presented, and a small experiment is offered as an example of what can be accomplished. This experiment is a comparison of natural and artificial language query facilities. Although subjects posed approximately equal numbers of valid queries with either facility, natural language users made significantly more invalid queries which could not be answered from the database that was described. %B ACM Trans. Database Syst. %V 3 %P 417 - 439 %8 1978/12// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/320289.320295 %N 4 %R 10.1145/320289.320295 %0 Journal Article %J Information & Management %D 1978 %T Information policy issues: selecting a policy framework and defining the schema horizon %A Shneiderman, Ben %K conceptual schema %K data independence %K data mode theory %K data submodels %K database administrator %K Database management %K decision making %K extended set theory %K external schema %K functional model %K internal schema %K management information systems %K organizational behavior %K organizational design %K schema %X Technical advances in database management systems are rebalancing organizational structures as management seeks to accommodate these powerful new tools. Managers must participate in establishing a framework for policy decisions and in setting goals for database usage. First, responsibility and authority must be delegated to management and not technical personnel. Second, the bounds of the application, called the schema horizon, should be carefully considered. This paper covers ten information policy issues within these two caregories and attempts to clarify management tasks. %B Information & Management %V 1 %P 207 - 218 %8 1978/// %@ 0378-7206 %G eng %U http://www.sciencedirect.com/science/article/pii/0378720678900277 %N 4 %R 10.1016/0378-7206(78)90027-7 %0 Journal Article %J Communications of the ACM %D 1978 %T Jump searching: a fast sequential search technique %A Shneiderman, Ben %K database structures %K file management %K index searching %K jump searching %K search strategies %K sequential files %X When sequential file structures must be used and binary searching is not feasible, jump searching becomes an appealing alternative. This paper explores variants of the classic jump searching scheme where the optimum jump size is the square root of the number of records. Multiple level and variable size jump strategies are explored, appropriate applications are discussed and performance is evaluated. %B Communications of the ACM %V 21 %P 831 - 834 %8 1978/10// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/359619.359623 %N 10 %R 10.1145/359619.359623 %0 Conference Paper %B Proceedings of the 1978 annual conference - Volume 2 %D 1978 %T Personality and programming: Time-sharing vs. batch preference %A Lee,Jeanne M. %A Shneiderman, Ben %K Assertive/passive %K batch processing %K Locus of control %K Personality %K Programming %K Psychology %K Time sharing %X Only within the past ten years has some attention been given to psychological concerns of human-machine interface. A review of the literature in this area reveals that personality has received the least attention, but interest is growing. If critical personality factors can be isolated and associated with particular programming tasks, such information could be a useful tool for education as well as management. The hypothesis of this exploratory study was that two personality dimensions, assertiveness and locus of control, influence a programmer's choice of batch or interactive processing for program development. Locus of control relates to the perception an individual has of his/her influence over events. Assertiveness allows an individual expression in a manner that fully communicates personal desires without infringing upon the rights of others. These two dimensions and the programmer's preference for batch or interactive mode were studied through a questionnaire survey of experienced programmers. %B Proceedings of the 1978 annual conference - Volume 2 %S ACM '78 %I ACM %C New York, NY, USA %P 561 - 569 %8 1978/// %@ 0-89791-000-1 %G eng %U http://doi.acm.org/10.1145/800178.810092 %R 10.1145/800178.810092 %0 Journal Article %J International Journal of Man-Machine Studies %D 1978 %T Two experimental comparisons of relational and hierarchical database models %A Brosey,Margaret %A Shneiderman, Ben %X The data model is a central feature which characterizes and distinguishes database management systems. This paper presents two experimental comparative studies of two prominent data models; the relational and hierarchical models. Comprehension, problem solving situation and memorization tasks were performed by undergraduate subjects. Significant effects were found for the data model, presentation order, subject background and tasks. %B International Journal of Man-Machine Studies %V 10 %P 625 - 637 %8 1978/11// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S0020737378800248 %N 6 %R 10.1016/S0020-7373(78)80024-8 %0 Journal Article %J Information Processing & Management %D 1977 %T Design, development and utilization perspectives on database management systems %A Shneiderman, Ben %X This paper reviews the historical development of integrated database management systems and examines the competing approaches. Topics covered include management and utilization perspectives, implementation and design issues, query languages, security, integrity, privacy and concurrency. Extensive references are provided. %B Information Processing & Management %V 13 %P 23 - 33 %8 1977/// %@ 0306-4573 %G eng %U http://www.sciencedirect.com/science/article/pii/0306457377900280 %N 1 %R 10.1016/0306-4573(77)90028-0 %0 Conference Paper %B SIGCSE '77 Proceedings of the eighth SIGCSE technical symposium on Computer science education %D 1977 %T Evaluating introductory programming textbooks: A guide for students, instructors, authors and publishers %A Shneiderman, Ben %X My father counseled me not to give advice to others. Inspite of that admonition, I have an irrepressible desire to make some highly subjective, potentially pompous remarks about introductory programming language textbooks. During the past nine years I have evaluated dozens of manuscripts for eight publishers and hundreds of books for teaching term-length introductory courses in FORTRAN, BASIC, PL/I, PASCAL, COBOL and assembly languages. I have co-authored two FORTRAN texts and developed two independent study guides to programming. Each time I see a text, I make judgments by reacting to the material, rather than by comparing the material to a pre-determined set of criteria. In order to provide a “structured” review process (no computer-science oriented paper is complete without a reference to “structure”), I offer the following criteria for evaluating texts. This list is far from complete, but it is a beginning. %B SIGCSE '77 Proceedings of the eighth SIGCSE technical symposium on Computer science education %V 9 %P 56 - 58 %8 1977/02// %G eng %U http://doi.acm.org/10.1145/382175.803434 %R 10.1145/382175.803434 %0 Journal Article %J Communications of the ACM %D 1977 %T Experimental investigations of the utility of detailed flowcharts in programming %A Shneiderman, Ben %A Mayer,Richard %A McKay,Don %A Heller,Peter %K Debugging %K experimental testing %K flowcharts %K human factors %K modification %K program composition %K program comprehension %X This paper describes previous research on flowcharts and a series of controlled experiments to test the utility of detailed flowcharts as an aid to program composition, comprehension, debugging, and modification. No statistically significant difference between flowchart and nonflowchart groups has been shown, thereby calling into question the utility of detailed flowcharting. A program of further research is suggested. %B Communications of the ACM %V 20 %P 373 - 381 %8 1977/06// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/359605.359610 %N 6 %R 10.1145/359605.359610 %0 Journal Article %J International Journal of Man-Machine Studies %D 1977 %T Measuring computer program quality and comprehension %A Shneiderman, Ben %K cognitive psychology %K commenting %K memorization %K modification %K program design %K programmer behavior %K Programming %K programming experiments %K programming languages %K RECALL %K structured programming %X Judging the relative quality of computer programs and a programmer's comprehension of a given program has proven to be a difficult task. Ability to debug, modify, hand simulate execution or respond to questions about the program all have their weaknesses as comprehension metrics. This paper offers experimental evidence which supports the use of memorization/recall tasks as a further basis for judging program quality and programmer comprehension. A syntactic/semantic model of programmer behavior is presented to provide a basis for this hypothesis. %B International Journal of Man-Machine Studies %V 9 %P 465 - 478 %8 1977/07// %@ 0020-7373 %G eng %U http://www.sciencedirect.com/science/article/pii/S002073737780014X %N 4 %R 10.1016/S0020-7373(77)80014-X %0 Journal Article %J Information Systems %D 1977 %T Reduced combined indexes for efficient multiple attribute retrieval %A Shneiderman, Ben %X Combined indexes were proposed by Lum[4] as an alternative to the traditional approach of single attribute indexes. The combined index approach is appealing for queries requiring conjunctions of attribute values since it eliminates the time consuming intersections. The penalty of wasted auxiliary storage space in the combined index approach can be minimized by adopting the Reduced Combined Index technique proposed in this paper. %B Information Systems %V 2 %P 149 - 154 %8 1977/// %@ 0306-4379 %G eng %U http://www.sciencedirect.com/science/article/pii/0306437977900035 %N 4 %R 10.1016/0306-4379(77)90003-5 %0 Journal Article %J Computers & Education %D 1977 %T Teaching programming: A spiral approach to syntax and semantics %A Shneiderman, Ben %X Coupling the recently proposed syntactic/semantic model of programmer behavior [1] with classic educational psychological theories yields new insights to teaching programming to novices. These new insights should make programming education more natural to students. alleviate “computer shock” (the analog of “math anxiety” [2]) and promote the development of widespread “computer literacy”.The spiral approach is the parallel acquisition of syntactic and semantic knowledge in a sequence which provokes student interest by using meaningful examples, builds on previous knowledge, is in harmony with the student's cognitive skills, provides reinforcement of recently acquired material and develops confidence through successful accomplishment of increasingly difficult tasks. The relationship of structured programming and flowcharts to the spiral approach is discussed. %B Computers & Education %V 1 %P 193 - 197 %8 1977/// %@ 0360-1315 %G eng %U http://www.sciencedirect.com/science/article/pii/0360131577900082 %N 4 %R 10.1016/0360-1315(77)90008-2 %0 Conference Paper %B Proceedings of the fifteenth annual SIGCPR conference %D 1977 %T Use of peer ratings in evaluating computer program quality %A Anderson,Nancy %A Shneiderman, Ben %X Peer review techniques can be useful tools for supplementing programmer education, improving cooperation and communication within programming teams, and providing programmer self-evaluation. This paper will explore the benefits of peer review for practicing professional programmers and describe a framework for administration of an annual or semi-annual peer review process. %B Proceedings of the fifteenth annual SIGCPR conference %S SIGCPR '77 %I ACM %C New York, NY, USA %P 218 - 226 %8 1977/// %G eng %U http://doi.acm.org/10.1145/800100.803247 %R 10.1145/800100.803247 %0 Journal Article %J ACM Trans. Database Syst. %D 1976 %T Batched searching of sequential and tree structured files %A Shneiderman, Ben %K B-trees %K batched searching %K Data structures %K Database systems %K file management %K Indexes %K informational retrieval %K sequential files %K tree structures %X The technique of batching searches has been ignored in the context of disk based online data retrieval systems. This paper suggests that batching be reconsidered for such systems since the potential reduction in processor demand may actually reduce response time. An analysis with sample numerical results and algorithms is presented. %B ACM Trans. Database Syst. %V 1 %P 268 - 275 %8 1976/09// %@ 0362-5915 %G eng %U http://doi.acm.org/10.1145/320473.320487 %N 3 %R 10.1145/320473.320487 %0 Book %D 1976 %T The EXEC-8 Log System: Error Analysis %A Mohr,J. M %A Agrawala, Ashok K. %A Flanagan,J. F. %A Science,University of Maryland. Dept. of Computer %I University of Maryland, Dept. of Computer Science %8 1976/// %G eng %0 Journal Article %J International Journal of Parallel Programming %D 1976 %T Exploratory experiments in programmer behavior %A Shneiderman, Ben %X The techniques of cognitive psychological experimentation can help resolve specific issues in programming and explore the broader issues of programmer behavior. This paper describes the methodological questions of such experimentation and presents two exploratory experiments: a memorization task and a comparison of the arithmetic and logical IF statements infortran. %B International Journal of Parallel Programming %V 5 %P 123 - 143 %8 1976/// %@ 0885-7458 %G eng %U http://dx.doi.org/10.1007/BF00975629 %N 2 %0 Journal Article %J SIGCSE Bull. %D 1976 %T An internship in information systems: Combining computer science education with realistic problems %A Buck,John %A Shneiderman, Ben %X Computer science graduates who become professional programmers will have a direct and substantial influence on the impact of applications, but little in traditional computer science training curriculum prepares them for this serious responsibility. Recognizing this situation, we designed a two term sequence for advanced undergraduates and masters students which would not only provide them with the required academic knowledge. The educational atmosphere that we tried to create resembles the internship phase followed in teacher training, medical schools, law schools, clinical psychology and other disciplines. %B SIGCSE Bull. %V 8 %P 80 - 83 %8 1976/07// %@ 0097-8418 %G eng %U http://doi.acm.org/10.1145/952991.804761 %N 3 %R 10.1145/952991.804761 %0 Journal Article %J International Journal of Parallel Programming %D 1976 %T Toward a theory of encoded data structures and data translation %A Shneiderman, Ben %A Shapiro,Stuart C. %X Several models of data base systems have distinguished levels of abstraction ranging from the high-level entity set model down to the low-level physical device level. This paper presents a model for describing data encodings, an intermediate level which focuses on the relationship among data items as demonstrated by contiguity or by pointer connections. Multiple data encodings for a file are shown and transformation functions that describe the translation between data encodings are discussed. %B International Journal of Parallel Programming %V 5 %P 33 - 43 %8 1976/// %@ 0885-7458 %G eng %U http://dx.doi.org/10.1007/BF00991070 %N 1 %0 Journal Article %J Information Processing & Management %D 1976 %T Travels in computerland or incompatibilities and interfaces::, Addison-Wesley, Reading, Mass., 1974 %A Shneiderman, Ben %B Information Processing & Management %V 12 %P 230 - 230 %8 1976/// %G eng %N 3 %0 Journal Article %J ACM SIGPLAN Notices %D 1975 %T Cognitive psychology and programming language design %A Shneiderman, Ben %B ACM SIGPLAN Notices %V 10 %P 46 - 47 %8 1975/07// %@ 0362-1340 %G eng %U http://doi.acm.org/10.1145/987305.987314 %N 7 %R 10.1145/987305.987314 %0 Conference Paper %B Proceedings of the May 19-22, 1975, national computer conference and exposition %D 1975 %T Experimental testing in programming languages, stylistic considerations and design techniques %A Shneiderman, Ben %X In the early stages of the development of high-level languages, radically differing alternatives were often promulgated. Each language had a dedicated corps of adherents who advocated the primacy of their facility. Turbulent debates among the protagonists were a common affair at conferences and in the trade journals. Now as the field matures, the vehement discussions have subsided and there is a widespread recognition of the usefulness of a variety of languages. Even the proponents of a single universal language have softened their tone and have accepted the multiple language condition. %B Proceedings of the May 19-22, 1975, national computer conference and exposition %S AFIPS '75 %I ACM %C New York, NY, USA %P 653 - 656 %8 1975/// %G eng %U http://doi.acm.org/10.1145/1499949.1500087 %R 10.1145/1499949.1500087 %0 Book %D 1975 %T Two Exploratory Experiments in Program Comprehension %A Shneiderman, Ben %A Ho,M.H. %I Indiana University, Computer Science Department %8 1975/// %G eng %0 Journal Article %J ACM SIGPLAN Notices %D 1974 %T The chemistry of control structures %A Shneiderman, Ben %B ACM SIGPLAN Notices %V 9 %P 29 - 34 %8 1974/12// %@ 0362-1340 %G eng %U http://doi.acm.org/10.1145/953246.953249 %N 12 %R 10.1145/953246.953249 %0 Journal Article %J International Journal of Parallel Programming %D 1974 %T A model for optimizing indexed file structures %A Shneiderman, Ben %X Multilevel indexes have long been used for accessing records in sorted files. Given the access cost at each level, the total cost of retrieving a record from the file can be substantially reduced by selecting the proper size of the index at each level. Organizations involving a variable number of levels are covered and binary searching is compared to sequential searching. %B International Journal of Parallel Programming %V 3 %P 91 - 101 %8 1974/// %@ 0885-7458 %G eng %U http://dx.doi.org/10.1007/BF01355589 %N 1 %0 Journal Article %J ACM SIGMOD Record %D 1974 %T Opportunities for data base reorganization %A Shneiderman, Ben %B ACM SIGMOD Record %V 6 %P 1 - 8 %8 1974/12// %@ 0163-5808 %G eng %U http://doi.acm.org/10.1145/983082.983083 %N 4 %R 10.1145/983082.983083 %0 Journal Article %J Communications of the ACM %D 1974 %T Structured data structures %A Shneiderman, Ben %A Scheuermann,Peter %K data base management system %K Data structures %K structured programming %X Programming systems which permit arbitrary linked list structures enable the user to create complicated structures without sufficient protection. Deletions can result in unreachable data elements, and there is no guarantee that additions will be performed properly. To remedy this situation, this paper proposes a Data Structure Description and Manipulation Language which provides for the creation of a restricted class of data structures but ensures the correctness of the program. This is accomplished by an explicit structure declaration facility, a restriction on the permissible operations, and execution-time checks. %B Communications of the ACM %V 17 %P 566 - 574 %8 1974/10// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/355620.361170 %N 10 %R 10.1145/355620.361170 %0 Journal Article %J ACM SIGPLAN Notices %D 1973 %T Flowchart techniques for structured programming %A Nassi,I. %A Shneiderman, Ben %X With the advent of structured programming and GOTO-less programming a method is needed to model computation in simply ordered structures, each representing a complete thought possibly defined in terms of other thoughts as yet undefined. A model is needed which prevents unrestricted transfers of control and has a control structure closer to languages amenable to structured programming. We present an attempt at such a model. %B ACM SIGPLAN Notices %V 8 %P 12 - 26 %8 1973/08// %@ 0362-1340 %G eng %U http://doi.acm.org/10.1145/953349.953350 %N 8 %R 10.1145/953349.953350 %0 Journal Article %J Communications of the ACM %D 1973 %T Optimum data base reorganization points %A Shneiderman, Ben %K data base %K files %K Information retrieval %K reorganization %X In certain data base organization schemes the cost per access may increase due to structural inefficiencies caused by update. By reorganizing the data base the cost per access may be reduced. However, the high cost of a reorganization prohibits frequent reorganizations. This paper examines strategies for selecting the optimum reorganization points. %B Communications of the ACM %V 16 %P 362 - 365 %8 1973/06// %@ 0001-0782 %G eng %U http://doi.acm.org/10.1145/362248.362267 %N 6 %R 10.1145/362248.362267 %0 Book %D 1972 %T The Elements of FORTRAN Style: Techniques for Effective Programming %A Kreitzberg,Charles B %A Shneiderman, Ben %I Harcourt Brace College Publishers %8 1972/// %@ 0155221566 %G eng %0 Journal Article %J SIGIR Forum %D 1972 %T A graph theoretic model of data structures %A Heller,J. %A Shneiderman, Ben %X The widespread development of large shared data bases has proceeded mainly on empirical notions. Only in the past few years have there been attempts at formalizing the underlying concepts into a theory of data structures. The present models have been founded on the well-established mathematical theories of sets, relations and graphs, but few descriptions have been sufficiently developed to meet the needs of a complete theory of data structures. The set theoretic formalism offered by Childs(1) or Schwartz(2) or the relational model by Codd(3) are useful but fail to give a complete description of the complex structures found in modern data bases. The graph theoretic concepts found in papers by Harary and Hsiao(4) and Earley(5) are more appealing but need further elaboration. This paper is an attempt to develop a graph theoretic model into a useful formalism for describing and manipulating data structures. %B SIGIR Forum %V 7 %P 36 - 44 %8 1972/12// %@ 0163-5840 %G eng %U http://doi.acm.org/10.1145/1095495.1095499 %N 4 %R 10.1145/1095495.1095499 %0 Journal Article %J ACM SIGCSE Bulletin %D 1971 %T Computer science education and social relevance %A Shneiderman, Ben %X The rise of computer science as a theoretical discipline should not be allowed to proceed without promoting the study of the social implications and applications of the field. This paper describes an undergraduate course whose primary goal is to foster an understanding of how computers can be used for socially relevant purposes. The students were required to propose and execute a project which could benefit people directly. The projects are described and suggestions for further work are given.The rapid rise of Computer Science as a major field of university study is unprecedented. In 25 years digital computers have risen from a laboratory experiment to a full-fledged academic discipline. Most major universities have created Departments of Computer Science and offer advanced degrees. %B ACM SIGCSE Bulletin %V 3 %P 21 - 24 %8 1971/03// %@ 0097-8418 %G eng %U http://doi.acm.org/10.1145/873674.873677 %N 1 %R 10.1145/873674.873677