@article {20391, title = {Metagenome sequencing-based strain-level and functional characterization of supragingival microbiome associated with dental caries in children}, journal = {Journal of Oral Microbiology}, year = {2019}, month = {12/2018}, pages = {1557986}, abstract = {Studies of the microbiome associated with dental caries have largely relied on 16S rRNA sequence analysis, which is associated with PCR biases, low taxonomic resolution, and inability to accurately study functions. Here, we employed whole metagenome shotgun sequencing, coupled with high-resolution analysis algorithm, to analyze supragingival microbiomes from 30 children with or without dental caries. A total of 726 bacterial strains belonging to 406 species, in addition to 34 bacteriophages were identified. A core bacteriome was identified at the species and strain levels. Species of Prevotella, Veillonella, as yet unnamed Actinomyces, and Atopobium showed strongest association with caries; Streptococcus sp. AS14 and Leptotrichia sp. Oral taxon 225, among others, were overabundant in caries-free. For several species, the association was strain-specific. Furthermore, for some species, e.g. Streptococcus mitis and Streptococcus sanguinis, sister strains showed differential associations. Noteworthy, associations were also identified for phages: Streptococcus phage M102 with caries and Haemophilus phage HP1 with caries-free. Functionally, potentially relevant features were identified including urate, vitamin K2, and polyamine biosynthesis in association with caries; and three deiminases and lactate dehydrogenase with health. The results demonstrate new associations between the microbiome and dental caries at the strain and functional levels that need further investigation.}, doi = {10.1080/20002297.2018.1557986}, url = {https://www.tandfonline.com/doi/full/10.1080/20002297.2018.1557986}, author = {Al-Hebshi, Nezar Noor and Baraniya, Divyashri and Chen, Tsute and Hill, Jennifer and Puri, Sumant and Tellez, Marisol and Hassan, Nur A. and Rita R Colwell and Ismail, Amid} } @article {20449, title = {Quantum Capacitance-Limited MoS2 Biosensors Enable Remote Label-Free Enzyme Measurements}, journal = {Nanoscale}, year = {2019}, month = {Jan-01-2019}, issn = {2040-3364}, doi = {10.1039/C9NR03171E}, url = {http://pubs.rsc.org/en/Content/ArticleLanding/2019/NR/C9NR03171Ehttp://pubs.rsc.org/en/content/articlepdf/2019/NR/C9NR03171E}, author = {Le, Son T and Guros, Nicholas B and Bruce, Robert C and Cardone, Antonio and Amin, Niranjana D and Zhang, Siyuan and Klauda, Jeffery and Pant, Harish C and Richter, Curt A and Balijepalli, Arvind} } @article {20295, title = {Biofilms Comprise a Component of the Annual Cycle of Vibrio cholerae in the Bay of Bengal Estuary}, journal = {mBio}, year = {2018}, month = {Feb-05-2018}, pages = {e00483-18}, abstract = {Vibrio cholerae, an estuarine bacterium, is the causative agent of cholera, a severe diarrheal disease that demonstrates seasonal incidence in Bangladesh. In an extensive study of V. cholerae occurrence in a natural aquatic environment, water and plankton samples were collected biweekly between December 2005 and November 2006 from Mathbaria, an estuarine village of Bangladesh near the mangrove forests of the Sundarbans. Toxigenic V. cholerae exhibited two seasonal growth peaks, one in spring (March to May) and another in autumn (September to November), corresponding to the two annual seasonal outbreaks of cholera in this region. The total numbers of bacteria determined by heterotrophic plate count (HPC), representing culturable bacteria, accounted for 1\% to 2.7\% of the total numbers obtained using acridine orange direct counting (AODC). The highest bacterial culture counts, including toxigenic V. cholerae, were recorded in the spring. The direct fluorescent antibody (DFA) assay was used to detect V. cholerae O1 cells throughout the year, as free-living cells, within clusters, or in association with plankton. V. cholerae O1 varied significantly in morphology, appearing as distinctly rod-shaped cells in the spring months, while small coccoid cells within thick clusters of biofilm were observed during interepidemic periods of the year, notably during the winter months. Toxigenic V. cholerae O1 was culturable in natural water during the spring when the temperature rose sharply. The results of this study confirmed biofilms to be a means of persistence for bacteria and an integral component of the annual life cycle of toxigenic V. cholerae in the estuarine environment of Bangladesh.}, doi = {10.1128/mBio.00483-18}, url = {https://mbio.asm.org/content/9/2/e00483-18}, author = {Sultana, Marzia and Nusrin, Suraia and Hasan, Nur A. and Sadique, Abdus and Ahmed, Kabir U. and Islam, Atiqul and Hossain, Anwar and Longini, Ira and Nizam, Azhar and Huq, Anwar and Siddique, Abul K. and Sack, David A. and Sack, Richard B. and Rita R Colwell and Alam, Munirul}, editor = {Vidaver, Anne K.} } @article {20289, title = {Environmental and hydroclimatic factors influencing Vibrio populations in the estuarine zone of the Bengal delta}, journal = {Environmental Monitoring and Assessment}, year = {2018}, month = {Jan-10-2018}, abstract = {The objective of this study was to determine environmental parameters driving Vibrio populations in the estuarine zone of the Bengal delta. Spatio-temporal data were collected at river estuary, mangrove, beach, pond, and canal sites. Effects of salinity, tidal amplitude, and a cyclone and tsunami were included in the study. Vibrio population shifts were found to be correlated with tide-driven salinity and suspended particulate matter (SPM). Increased abundance of Vibrio spp. in surface water was observed after a cyclone, attributed to re-suspension of benthic particulate organic carbon (POC), and increased availability of chitin and dissolved organic carbon (DOC). Approximately a two log10 increase in the (p < 0.05) number of Vibrio spp. was observed in < 20 μm particulates, compared with microphytoplankton (20{\textendash}60 μm) and zooplankton > 60 μm fractions. Benthic and suspended sediment comprised a major reservoir of Vibrio spp. Results of microcosm experiments showed enhanced growth of vibrios was related to concentration of organic matter in SPM. It is concluded that SPM, POC, chitin, and salinity significantly influence abundance and distribution of vibrios in the Bengal delta estuarine zone. Keywords Vibrio Salinity Cyclone Tide Chitin Sediment dynamics }, issn = {0167-6369}, doi = {10.1007/s10661-018-6925-7}, url = {https://link.springer.com/article/10.1007\%2Fs10661-018-6925-7}, author = {Neogi, Sucharit and Rubin, Lara and Alam, Munirul and Harder, Jens and Yamasaki, Shinji and Rita R Colwell} } @article {20283, title = {A Metagenomic Approach to Evaluating Surface Water Quality in Haiti}, journal = {International Journal of Environmental Research and Public Health}, volume = {1542}, year = {2018}, month = {Jan-10-2018}, pages = {2211}, abstract = {The cholera epidemic that occurred in Haiti post-earthquake in 2010 has resulted in over 9000 deaths during the past eight years. Currently, morbidity and mortality rates for cholera have declined, but cholera cases still occur on a daily basis. One continuing issue is an inability to accurately predict and identify when cholera outbreaks might occur. To explore this surveillance gap, a metagenomic approach employing environmental samples was taken. In this study, surface water samples were collected at two time points from several sites near the original epicenter of the cholera outbreak in the Central Plateau of Haiti. These samples underwent whole genome sequencing and subsequent metagenomic analysis to characterize the microbial community of bacteria, fungi, protists, and viruses, and to identify antibiotic resistance and virulence associated genes. Replicates from sites were analyzed by principle components analysis, and distinct genomic profiles were obtained for each site. Cholera toxin converting phage was detected at one site, and Shiga toxin converting phages at several sites. Members of the Acinetobacter family were frequently detected in samples, including members implicated in waterborne diseases. These results indicate a metagenomic approach to evaluating water samples can be useful for source tracking and the surveillance of pathogens such as Vibrio cholerae over time, as well as for monitoring virulence factors such as cholera toxin.}, url = {https://www.ncbi.nlm.nih.gov/pubmed/30309013}, author = {Roy, Monika and Arnaud, Jean and Jasmin, Paul and Hamner, Steve and Hasan, Nur and Rita R Colwell and Ford, Timothy} } @article {20296, title = {Satellites and Cell Phones Form a Cholera Early-Warning System}, journal = {Eos}, volume = {99}, year = {2018}, month = {Mar-03-2020}, doi = {10.1029/2018EO094839}, url = {https://eos.org/project-updates/satellites-and-cell-phones-form-a-cholera-early-warning-system}, author = {Akanda, Ali and Aziz, Sonia and Jutla, Antarpreet and Huq, Anwar and Alam, Munirul and Ahsan, Gias and Rita R Colwell} } @article {20298, title = {Virulence-Related Genes Identified from the Genome Sequence of the Non-O1/Non-O139 Vibrio cholerae Strain VcN1, Isolated from Dhaka, Bangladesh}, journal = {Genome Announcements}, year = {2018}, month = {Aug-03-2018}, abstract = {We report here the first draft genome sequence of the non-O1/non-O139 Vibrio cholerae strain VcN1, isolated from Dhaka, Bangladesh. The data submitted to GenBank for this strain will contribute to advancing our understanding of this environmentally disseminated bacterium, including its virulence and its evolution as an important pathogen.}, doi = {10.1128/genomeA.01513-17}, url = {https://mra.asm.org/content/6/10/e01513-17}, author = {Hossain, Maqsud and Alam, Munirul and Khaleque, Abdul and Islam, Sohidul and Sadique, Abdus and Khan, Nayeim and Halim, Zahra and Sarker, Mrinmoy and El-Sayed, Najib M. and Huq, Anwar and Ahsan, Gias Uddin and Rita R Colwell} } @article {20315, title = {Assessment of Risk of Cholera in Haiti following Hurricane Matthew}, journal = {The American Journal of Tropical Medicine and Hygiene}, volume = {97}, year = {2017}, month = {Jul-09-2017}, pages = {896 - 903}, abstract = {Damage to the inferior and fragile water and sanitation infrastructure of Haiti after Hurricane Matthew has created an urgent public health emergency in terms of likelihood of cholera occurring in the human population. Using satellite-derived data on precipitation, gridded air temperature, and hurricane path and with information on water and sanitation (WASH) infrastructure, we tracked changing environmental conditions conducive for growth of pathogenic vibrios. Based on these data, we predicted and validated the likelihood of cholera cases occurring past hurricane. The risk of cholera in the southwestern part of Haiti remained relatively high since November 2016 to the present. Findings of this study provide a contemporary process for monitoring ground conditions that can guide public health intervention to control cholera in human population by providing access to vaccines, safe WASH facilities. Assuming current social and behavioral patterns remain constant, it is recommended that WASH infrastructure should be improved and considered a priority especially before 2017 rainy season.}, issn = {0002-9637}, doi = {10.4269/ajtmh.17-0048}, url = {http://www.ajtmh.org/content/journals/10.4269/ajtmh.17-0048}, author = {Huq, Anwar and Anwar, Rifat and Rita R Colwell and McDonald, Michael D. and Khan, Rakib and Jutla, Antarpreet and Akanda, Shafqat} } @article {20311, title = {Comparative Genomics of Escherichia coli Isolated from Skin and Soft Tissue and Other Extraintestinal Infections}, journal = {mBio}, year = {2017}, month = {Jun-09-2017}, abstract = {Escherichia coli, an intestinal Gram-negative bacterium, has been shown to be associated with a variety of diseases in addition to intestinal infections, such as urinary tract infections (UTIs), meningitis in neonates, septicemia, skin and soft tissue infections (SSTIs), and colisepticemia. Thus, for nonintestinal infections, it is categorized as extraintestinal pathogenic E. coli (ExPEC). It is also an opportunistic pathogen, causing cross infections, notably as an agent of zoonotic diseases. However, comparative genomic data providing functional and genetic coordinates for ExPEC strains associated with these different types of infections have not proven conclusive. In the study reported here, ExPEC E. coli isolated from SSTIs was characterized, including virulence and drug resistance profiles, and compared with isolates from patients suffering either pyelonephritis or septicemia. Results revealed that the majority of the isolates belonged to two pathogenic phylogroups, B2 and D. Approximately 67\% of the isolates were multidrug resistant (MDR), with 85\% producing extended-spectrum beta-lactamase (ESBL) and 6\% producing metallo-beta-lactamase (MBL). The blaCTX-M-15 genotype was observed in at least 70\% of the E. coli isolates in each category, conferring resistance to an extended range of beta-lactam antibiotics. Whole-genome sequencing and comparative genomics of the ExPEC isolates revealed that two of the four isolates from SSTIs, NA633 and NA643, belong to pandemic sequence type ST131, whereas functional characteristics of three of the ExPEC pathotypes revealed that they had equal capabilities to form biofilm and were resistant to human serum. Overall, the isolates from a variety of ExPEC infections demonstrated similar resistomes and virulomes and did not display any disease-specific functional or genetic coordinates. IMPORTANCE Infections caused by extraintestinal pathogenic E. coli (ExPEC) are of global concern as they result in significant costs to health care facilities management. The recent emergence of a multidrug-resistant pandemic clone, Escherichia coli ST131, is of primary concern as a global threat. In developing countries, such as India, skin and soft tissue infections (SSTIs) associated with E. coli are marginally addressed. In this study, we employed both genomic analysis and phenotypic assays to determine relationships, if any, among the ExPEC pathotypes. Similarity between antibiotic resistance and virulence profiles was observed, ST131 isolates from SSTIs were reported, and genomic similarities among strains isolated from different disease conditions were detected. This study provides functional molecular infection epidemiology insight into SSTI-associated E. coli compared with ExPEC pathotypes.}, doi = {10.1128/mBio.01070-17}, url = {http://mbio.asm.org/lookup/doi/10.1128/mBio.01070-17}, author = {Ranjan, Amit and Shaik, Sabiha and Nandanwar, Nishant and Hussain, Arif and Tiwari, Sumeet K. and Semmler, Torsten and Jadhav, Savita and Wieler, Lothar H. and Alam, Munirul and Rita R Colwell and Ahmed, Niyaz}, editor = {Cossart, Pascale F.} } @article {20308, title = {Comprehensive benchmarking and ensemble approaches for metagenomic classifiers}, journal = {Genome Biology}, year = {2017}, month = {Jan-12-2017}, doi = {10.1186/s13059-017-1299-7}, url = {http://genomebiology.biomedcentral.com/articles/10.1186/s13059-017-1299-7}, author = {McIntyre, Alexa B. R. and Ounit, Rachid and Afshinnekoo, Ebrahim and Prill, Robert J. and H{\'e}naff, Elizabeth and Alexander, Noah and Minot, Samuel S. and Danko, David and Foox, Jonathan and Ahsanuddin, Sofia and Tighe, Scott and Hasan, Nur A. and Subramanian, Poorani and Moffat, Kelly and Levy, Shawn and Lonardi, Stefano and Greenfield, Nick and Rita R Colwell and Rosen, Gail L. and Mason, Christopher E.} } @article {20313, title = {Genomic Methods and Microbiological Technologies for Profiling Novel and Extreme Environments for the Extreme Microbiome Project (XMP)}, journal = {Journal of Biomolecular Techniques : JBT}, volume = {28}, year = {2017}, month = {Jan-04-2017}, pages = {31 - 39}, abstract = {The Extreme Microbiome Project (XMP) is a project launched by the Association of Biomolecular Resource Facilities Metagenomics Research Group (ABRF MGRG) that focuses on whole genome shotgun sequencing of extreme and unique environments using a wide variety of biomolecular techniques. The goals are multifaceted, including development and refinement of new techniques for the following: 1) the detection and characterization of novel microbes, 2) the evaluation of nucleic acid techniques for extremophilic samples, and 3) the identification and implementation of the appropriate bioinformatics pipelines. Here, we highlight the different ongoing projects that we have been working on, as well as details on the various methods we use to characterize the microbiome and metagenome of these complex samples. In particular, we present data of a novel multienzyme extraction protocol that we developed, called Polyzyme or MetaPolyZyme. Presently, the XMP is characterizing sample sites around the world with the intent of discovering new species, genes, and gene clusters. Once a project site is complete, the resulting data will be publically available. Sites include Lake Hillier in Western Australia, the {\textquotedblleft}Door to Hell{\textquotedblright} crater in Turkmenistan, deep ocean brine lakes of the Gulf of Mexico, deep ocean sediments from Greenland, permafrost tunnels in Alaska, ancient microbial biofilms from Antarctica, Blue Lagoon Iceland, Ethiopian toxic hot springs, and the acidic hypersaline ponds in Western Australia.}, issn = {1524-0215}, doi = {10.7171/jbt.17-2801-004}, url = {https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5345951/}, author = {Tighe, Scott and Afshinnekoo, Ebrahim and Rock, Tara M. and McGrath, Ken and Alexander, Noah and McIntyre, Alexa and Ahsanuddin, Sofia and Bezdan, Daniela and Green, Stefan J. and Joye, Samantha and Stewart Johnson, Sarah and Baldwin, Don A. and Bivens, Nathan and Ajami, Nadim and Carmical, Joseph R. and Herriott, Ian Charold and Rita R Colwell and Donia, Mohamed and Foox, Jonathan and Greenfield, Nick and Hunter, Tim and Hoffman, Jessica and Hyman, Joshua and Jorgensen, Ellen and Krawczyk, Diana and Lee, Jodie and Levy, Shawn and Garcia-Reyero, {\`a}lia and Settles, Matthew and Thomas, Kelley and {\'o}mez, Felipe and Schriml, Lynn and Kyrpides, Nikos and Zaikova, Elena and Penterman, Jon and Mason, Christopher E.} } @article {20322, title = {Hydroclimatic sustainability assessment of changing climate on cholera in the Ganges-Brahmaputra basin}, journal = {Advances in Water Resources}, volume = {108}, year = {2017}, month = {Jan-10-2017}, pages = {332 - 344}, abstract = {The association of cholera and climate has been extensively documented. However, determining the effects of changing climate on the occurrence of disease remains a challenge. Bimodal peaks of cholera in Bengal Delta are hypothesized to be linked to asymmetric flow of the Ganges and Brahmaputra rivers. Spring cholera is related to intrusion of bacteria-laden coastal seawater during low flow seasons, while autumn cholera results from cross-contamination of water resources when high flows in the rivers cause massive inundation. Coarse resolution of General Circulation Model (GCM) output (usually at 100 {\textendash} 300 km)cannot be used to evaluate variability at the local scale(10{\textendash}20 km),hence the goal of this study was to develop a framework that could be used to understand impacts of climate change on occurrence of cholera. Instead of a traditional approach of downscaling precipitation, streamflow of the two rivers was directly linked to GCM outputs, achieving reasonable accuracy (R2 = 0.89 for the Ganges and R2 = 0.91 for the Brahmaputra)using machine learning algorithms (Support Vector Regression-Particle Swarm Optimization). Copula methods were used to determine probabilistic risks of cholera under several discharge conditions. Key results, using model outputs from ECHAM5, GFDL, andHadCM3for A1B and A2 scenarios, suggest that the combined low flow of the two rivers may increase in the future, with high flows increasing for first half of this century, decreasing thereafter. Spring and autumn cholera, assuming societal conditions remain constant e.g., at the current rate, may decrease. However significant shifts were noted in the magnitude of river discharge suggesting that cholera dynamics of the delta may well demonstrate an uncertain predictable pattern of occurrence over the next century.}, issn = {03091708}, doi = {10.1016/j.advwatres.2016.11.018}, url = {https://linkinghub.elsevier.com/retrieve/pii/S030917081630728X}, author = {Nasr-Azadani, Fariborz and Khan, Rakibul and Rahimikollu, Javad and Unnikrishnan, Avinash and Akanda, Ali and Alam, Munirul and Huq, Anwar and Jutla, Antarpreet and Rita R Colwell} } @article {20317, title = {Vibrio cholerae O1 with Reduced Susceptibility to Ciprofloxacin and Azithromycin Isolated from a Rural Coastal Area of Bangladesh}, journal = {Frontiers in Microbiology}, year = {2017}, month = {Sep-02-2018}, abstract = {Cholera outbreaks occur each year in the remote coastal areas of Bangladesh and epidemiological surveillance and routine monitoring of cholera in these areas is challenging. In this study, a total of 97 Vibrio cholerae O1 isolates from Mathbaria, Bangladesh, collected during 2010 and 2014 were analyzed for phenotypic and genotypic traits, including antimicrobial susceptibility. Of the 97 isolates, 95 possessed CTX-phage mediated genes, ctxA, ace, and zot, and two lacked the cholera toxin gene, ctxA. Also both CTX+ and CTX- V. cholerae O1 isolated in this study carried rtxC, tcpAET, and hlyA. The classical cholera toxin gene, ctxB1, was detected in 87 isolates, while eight had ctxB7. Of 95 CTX+ V. cholerae O1, 90 contained rstRET and 5 had rstRCL. All isolates, except two, contained SXT related integrase intSXT. Resistance to penicillin, streptomycin, nalidixic acid, sulfamethoxazole-trimethoprim, erythromycin, and tetracycline varied between the years of study period. Most importantly, 93\% of the V. cholerae O1 were multidrug resistant. Six different resistance profiles were observed, with resistance to streptomycin, nalidixic acid, tetracycline, and sulfamethoxazole-trimethoprim predominant every year. Ciprofloxacin and azithromycin MIC were 0.003{\textendash}0.75 and 0.19{\textendash}2.00 μg/ml, respectively, indicating reduced susceptibility to these antibiotics. Sixteen of the V. cholerae O1 isolates showed higher MIC for azithromycin (>=0.5 μg/ml) and were further examined for 10 macrolide resistance genes, erm(A), erm(B), erm(C), ere(A), ere(B), mph(A), mph(B), mph(D), mef(A), and msr(A) with none testing positive for the macrolide resistance genes.}, doi = {10.3389/fmicb.2017.00252}, url = {http://journal.frontiersin.org/article/10.3389/fmicb.2017.00252}, author = {Rashed, Shah M. and Hasan, Nur A. and Alam, Munirul and Sadique, Abdus and Sultana, Marzia and Hoq, Md. Mozammel and Sack, R. Bradley and Rita R Colwell and Huq, Anwar} } @article {20309, title = {The microbiomes of blowflies and houseflies as bacterial transmission reservoirs}, journal = {Scientific Reports}, year = {2017}, month = {Jan-12-2017}, abstract = {Blowflies and houseflies are mechanical vectors inhabiting synanthropic environments around the world. They feed and breed in fecal and decaying organic matter, but the microbiome they harbour and transport is largely uncharacterized. We sampled 116 individual houseflies and blowflies from varying habitats on three continents and subjected them to high-coverage, whole-genome shotgun sequencing. This allowed for genomic and metagenomic analyses of the host-associated microbiome at the species level. Both fly host species segregate based on principal coordinate analysis of their microbial communities, but they also show an overlapping core microbiome. Legs and wings displayed the largest microbial diversity and were shown to be an important route for microbial dispersion. The environmental sequencing approach presented here detected a stochastic distribution of human pathogens, such as Helicobacter pylori, thereby demonstrating the potential of flies as proxies for environmental and public health surveillance.}, doi = {10.1038/s41598-017-16353-x}, url = {http://www.nature.com/articles/s41598-017-16353-x}, author = {Junqueira, AC and Ratan, Aakrosh and Acerbi, Enzo and Drautz-Moses, Daniela I. and Premkrishnan, BNV and Costea, PI and Linz, Bodo and Purbojati, Rikky W. and Paulo, Daniel F. and Gaultier, Nicolas E. and Subramanian, Poorani and Hasan, Nur A. and Rita R Colwell and Bork, Peer and Azeredo-Espin, Ana Maria L. and Bryant, Donald A. and Schuster, Stephan C.} } @article {20332, title = {Comparison of inferred relatedness based on multilocus variable-number tandem-repeat analysis and whole genome sequencing of Vibrio cholerae O1}, journal = {FEMS Microbiology Letters}, volume = {36389}, year = {2016}, month = {Nov-06-2017}, pages = {fnw116}, abstract = {Vibrio cholerae causes cholera, a severe diarrheal disease. Understanding the local genetic diversity and transmission of V. cholerae will improve our ability to control cholera. Vibrio cholerae isolates clustered in genetically related groups (clonal complexes, CC) by multilocus variable tandem-repeat analysis (MLVA) were compared by whole genome sequencing (WGS). Isolates in CC1 had been isolated from two geographical locations. Isolates in a second genetically distinct group, CC2, were isolated only at one location. Using WGS, CC1 isolates from both locations revealed, on average, 43.8 nucleotide differences, while those strains comprising CC2 averaged 19.7 differences. Strains from both MLVA-CCs had an average difference of 106.6. Thus, isolates comprising CC1 were more closely related (P < 10-6) to each other than to isolates in CC2. Within a MLVA-CC, after removing all paralogs, alternative alleles were found in all possible combinations on separate chromosomes indicative of recombination within the core genome. Including recombination did not affect the distinctiveness of the MLVA-CCs when measured by WGS. We found that WGS generally reflected the same genetic relatedness of isolates as MLVA, indicating that isolates from the same MLVA-CC shared a more recent common ancestor than isolates from the same location that clustered in a distinct MLVA-CC.}, doi = {10.1093/femsle/fnw116}, url = {https://academic.oup.com/femsle/article-lookup/doi/10.1093/femsle/fnw116}, author = {Rashid, Mahamud-ur and Almeida, Mathieu and Azman, Andrew S. and Lindsay, Brianna R. and Sack, David A. and Rita R Colwell and Huq, Anwar and Morris, J. Glenn and Alam, Munirul and Stine, O. Colin}, editor = {Winstanley, Craig} } @article {20336, title = {Cross-talk among flesh-eating Aeromonas hydrophila strains in mixed infection leading to necrotizing fasciitis}, journal = {Proceedings of the National Academy of Sciences}, volume = {11312161268}, year = {2016}, month = {Jul-01-2017}, pages = {722 - 727}, abstract = {Necrotizing fasciitis (NF) caused by flesh-eating bacteria is associated with high case fatality. In an earlier study, we reported infection of an immunocompetent individual with multiple strains of Aeromonas hydrophila (NF1{\textendash}NF4), the latter three constituted a clonal group whereas NF1 was phylogenetically distinct. To understand the complex interactions of these strains in NF pathophysiology, a mouse model was used, whereby either single or mixed A. hydrophila strains were injected intramuscularly. NF2, which harbors exotoxin A (exoA) gene, was highly virulent when injected alone, but its virulence was attenuated in the presence of NF1 (exoA-minus). NF1 alone, although not lethal to animals, became highly virulent when combined with NF2, its virulence augmented by cis-exoA expression when injected alone in mice. Based on metagenomics and microbiological analyses, it was found that, in mixed infection, NF1 selectively disseminated to mouse peripheral organs, whereas the other strains (NF2, NF3, and NF4) were confined to the injection site and eventually cleared. In vitro studies showed NF2 to be more effectively phagocytized and killed by macrophages than NF1. NF1 inhibited growth of NF2 on solid media, but ExoA of NF2 augmented virulence of NF1 and the presence of NF1 facilitated clearance of NF2 from animals either by enhanced priming of host immune system or direct killing via a contact-dependent mechanism.}, issn = {0027-8424}, doi = {10.1073/pnas.1523817113}, url = {http://www.pnas.org/lookup/doi/10.1073/pnas.1523817113}, author = {Ponnusamy, Duraisamy and Kozlova, Elena V. and Sha, Jian and Erova, Tatiana E. and Azar, Sasha R. and Fitts, Eric C. and Kirtley, Michelle L. and Tiner, Bethany L. and Andersson, Jourdan A. and Grim, Christopher J. and Isom, Richard P. and Hasan, Nur A. and Rita R Colwell and Chopra, Ashok K.} } @article {20325, title = {Enrichment dynamics of Listeria monocytogenes and the associated microbiome from naturally contaminated ice cream linked to a listeriosis outbreak}, journal = {BMC Microbiology}, year = {2016}, month = {Jan-12-2016}, doi = {10.1186/s12866-016-0894-1}, url = {http://bmcmicrobiol.biomedcentral.com/articles/10.1186/s12866-016-0894-1}, author = {Ottesen, Andrea and Ramachandran, Padmini and Reed, Elizabeth and White, James R. and Hasan, Nur and Subramanian, Poorani and Ryan, Gina and Jarvis, Karen and Grim, Christopher and Daquiqan, Ninalynn and Hanes, Darcy and Allard, Marc and Rita R Colwell and Brown, Eric and Chen, Yi} } @article {20321, title = {Modeling Sustainability: Population, Inequality, Consumption, and Bidirectional Coupling of the Earth and Human Systems}, journal = {National Science Review}, year = {2016}, month = {Nov-12-2016}, pages = {nww081}, abstract = {Over the last two centuries, the impact of the Human System has grown dramatically, becoming strongly dominant within the Earth System in many different ways. Consumption, inequality, and population have increased extremely fast, especially since about 1950, threatening to overwhelm the many critical functions and ecosystems of the Earth System. Changes in the Earth System, in turn, have important feedback effects on the Human System, with costly and potentially serious consequences. However, current models do not incorporate these critical feedbacks. We argue that in order to understand the dynamics of either system, Earth System Models must be coupled with Human System Models through bidirectional couplings representing the positive, negative, and delayed feedbacks that exist in the real systems. In particular, key Human System variables, such as demographics, inequality, economic growth, and migration, are not coupled with the Earth System but are instead driven by exogenous estimates, such as United Nations population projections. This makes current models likely to miss important feedbacks in the real Earth{\textendash}Human system, especially those that may result in unexpected or counterintuitive outcomes, and thus requiring different policy interventions from current models. The importance and imminence of sustainability challenges, the dominant role of the Human System in the Earth System, and the essential roles the Earth System plays for the Human System, all call for collaboration of natural scientists, social scientists, and engineers in multidisciplinary research and modeling to develop coupled Earth{\textendash}Human system models for devising effective science-based policies and measures to benefit current and future generations.}, issn = {2095-5138}, doi = {10.1093/nsr/nww081}, url = {https://academic.oup.com/nsr/article-lookup/doi/10.1093/nsr/nww081}, author = {Motesharrei, Safa and Rivas, Jorge and Kalnay, Eugenia and Asrar, Ghassem R. and Busalacchi, Antonio J. and Cahalan, Robert F. and Cane, Mark A. and Rita R Colwell and Feng, Kuishuang and Franklin, Rachel S. and Hubacek, Klaus and Miralles-Wilhelm, Fernando and Miyoshi, Takemasa and Ruth, Matthias and Sagdeev, Roald and Shirmohammadi, Adel and Shukla, Jagadish and Srebric, Jelena and Yakovenko, Victor M. and Zeng, Ning} } @article {20335, title = {Phylogenetic Diversity of Vibrio cholerae Associated with Endemic Cholera in Mexico from 1991 to 2008}, journal = {mBio}, volume = {7}, year = {2016}, month = {Apr-05-2016}, abstract = {An outbreak of cholera occurred in 1991 in Mexico, where it had not been reported for more than a century and is now endemic. Vibrio cholerae O1 prototype El Tor and classical strains coexist with altered El Tor strains (1991 to 1997). Nontoxigenic (CTX-) V. cholerae El Tor dominated toxigenic (CTX+) strains (2001 to 2003), but V. cholerae CTX+ variant El Tor was isolated during 2004 to 2008, outcompeting CTX- V. cholerae. Genomes of six Mexican V. cholerae O1 strains isolated during 1991 to 2008 were sequenced and compared with both contemporary and archived strains of V. cholerae. Three were CTX+ El Tor, two were CTX- El Tor, and the remaining strain was a CTX+ classical isolate. Whole-genome sequence analysis showed the six isolates belonged to five distinct phylogenetic clades. One CTX- isolate is ancestral to the 6th and 7th pandemic CTX+ V. cholerae isolates. The other CTX- isolate joined with CTX- non-O1/O139 isolates from Haiti and seroconverted O1 isolates from Brazil and Amazonia. One CTX+ isolate was phylogenetically placed with the sixth pandemic classical clade and the V. cholerae O395 classical reference strain. Two CTX+ El Tor isolates possessing intact Vibrio seventh pandemic island II (VSP-II) are related to hybrid El Tor isolates from Mozambique and Bangladesh. The third CTX+ El Tor isolate contained West African-South American (WASA) recombination in VSP-II and showed relatedness to isolates from Peru and Brazil. Except for one isolate, all Mexican isolates lack SXT/R391 integrative conjugative elements (ICEs) and sensitivity to selected antibiotics, with one isolate resistant to streptomycin. No isolates were related to contemporary isolates from Asia, Africa, or Haiti, indicating phylogenetic diversity.}, doi = {10.1128/mBio.02160-15}, url = {http://mbio.asm.org/lookup/doi/10.1128/mBio.02160-15}, author = {Choi, Seon Y and Rashed, Shah M. and Hasan, Nur A. and Alam, Munirul and Islam, Tarequl and Sadique, Abdus and Johura, Fatema-Tuz and Eppinger, Mark and Ravel, Jacques and Huq, Anwar and Cravioto, Alejandro and Rita R Colwell} } @article {20326, title = {Reduced Susceptibility to Extended-Spectrum β-Lactams in Vibrio cholerae Isolated in Bangladesh}, journal = {Frontiers in Public Health}, year = {2016}, month = {Jun-10-2017}, doi = {10.3389/fpubh.2016.00231}, url = {http://journal.frontiersin.org/article/10.3389/fpubh.2016.00231/full}, author = {Ceccarelli, Daniela and Alam, Munirul and Huq, Anwar and Rita R Colwell} } @article {20343, title = {Downscaling river discharge to assess the effects of climate change on cholera outbreaks in the Bengal Delta}, journal = {Climate Research}, volume = {64}, year = {2015}, month = {Jul-08-2017}, pages = {257 - 274}, abstract = {Endemic cholera in the Bengal Delta region of South Asia has been associated with asymmetric and episodic variability of river discharge. Spring cholera was found to be related to intrusion of bacteria-laden coastal seawater during low flow seasons. Autumn cholera was hypothesized to result from cross-contamination of water resources when high river discharge causes massive inland inundation. The effect of climate change on diarrheal diseases has not been explored, because of the difficulties in establishing linkages between coarse-resolution global climate model outputs with localized disease outbreaks. Since rivers act as corridors for transport of cholera bacteria, the first step is to understand the discharge variability that may occur with climate change and whether it is linked to cholera. Here, we present a framework for downscaling precipitation from global climate models for river discharge in the Ganges-Brahmaputra-Meghna basin. Using a data-mining method that includes particle swarm optimization-based support vector regression, precipitation was downscaled for a statistical multiple regressive model to estimate river discharge in the basin. Key results from an ensemble of HadCM3, GFDL, and ECHAM5 models indicated 8 and 7.5\% increase in flows for the IPCC A1B and A2 scenarios, respectively. The majority of the changes are attributable to increases in flows from February through August for both scenarios, with little to no change in seasonality of high and low flows during the next century. The probability of spring and autumn cholera is likely to increase steadily in the endemic region of the Bengal Delta. }, issn = {0936-577X}, doi = {10.3354/cr01310}, url = {http://www.int-res.com/abstracts/cr/v64/n3/p257-274/}, author = {Nasr-Azadani, F and Unnikrishnan, A and Akanda, A and Islam, S and Alam, M and Huq, A and Jutla, A and Rita R Colwell} } @article {20347, title = {Hybrid Vibrio~cholerae El Tor Lacking SXT Identified as the Cause of a Cholera Outbreak in the Philippines}, journal = {mBio}, year = {2015}, month = {Jan-05-2015}, abstract = {Cholera continues to be a global threat, with high rates of morbidity and mortality. In 2011, a cholera outbreak occurred in Palawan, Philippines, affecting more than 500 people, and 20 individuals died. Vibrio cholerae O1 was confirmed as the etiological agent. Source attribution is critical in cholera outbreaks for proper management of the disease, as well as to control spread. In this study, three V. cholerae O1 isolates from a Philippines cholera outbreak were sequenced and their genomes analyzed to determine phylogenetic relatedness to V. cholerae O1 isolates from recent outbreaks of cholera elsewhere. The Philippines V. cholerae O1 isolates were determined to be V. cholerae O1 hybrid El Tor belonging to the seventh-pandemic clade. They clustered tightly, forming a monophyletic clade closely related to V. cholerae O1 hybrid El Tor from Asia and Africa. The isolates possess a unique multilocus variable-number tandem repeat analysis (MLVA) genotype (12-7-9-18-25 and 12-7-10-14-21) and lack SXT. In addition, they possess a novel 15-kb genomic island (GI-119) containing a predicted type I restriction-modification system. The CTXΦ-RS1 array of the Philippines isolates was similar to that of V. cholerae O1 MG116926, a hybrid El Tor strain isolated in Bangladesh in 1991. Overall, the data indicate that the Philippines V. cholerae O1 isolates are unique, differing from recent V. cholerae O1 isolates from Asia, Africa, and Haiti. Furthermore, the results of this study support the hypothesis that the Philippines isolates of V. cholerae O1 are indigenous and exist locally in the aquatic ecosystem of the Philippines.}, doi = {10.1128/mBio.00047-15}, url = {http://mbio.asm.org/lookup/doi/10.1128/mBio.00047-15}, author = {Klinzing, David C. and Choi, Seon Young and Hasan, Nur A. and Matias, Ronald R. and Tayag, Enrique and Geronimo, Josefina and Skowronski, Evan and Rashed, Shah M. and Kawashima, Kent and Rosenzweig, C. Nicole and Gibbons, Henry S. and Torres, Brian C. and Liles, Veni and Alfon, Alicia C. and Juan, Maria Luisa and Natividad, Filipinas F. and Cebula, Thomas A. and Rita R Colwell} } @article {20345, title = {Molecular tools in understanding the evolution of Vibrio cholerae}, journal = {Frontiers in Microbiology}, year = {2015}, month = {Jun-10-2015}, abstract = {Vibrio cholerae, the etiological agent of cholera, has been a scourge for centuries. Cholera remains a serious health threat for developing countries and has been responsible for millions of deaths globally over the past 200 years. Identification of V. cholerae has been accomplished using a variety of methods, ranging from phenotypic strategies to DNA based molecular typing and currently whole genomic approaches. This array of methods has been adopted in epidemiological investigations, either singly or in the aggregate, and more recently for evolutionary analyses of V. cholerae. Because the new technologies have been developed at an ever increasing pace, this review of the range of fingerprinting strategies, their relative advantages and limitations, and cholera case studies was undertaken. The task was challenging, considering the vast amount of the information available. To assist the study, key references representative of several areas of research are provided with the intent to provide readers with a comprehensive view of recent advances in the molecular epidemiology of V. cholerae. Suggestions for ways to obviate many of the current limitations of typing techniques are also provided. In summary, a comparative report has been prepared that includes the range from traditional typing to whole genomic strategies.}, doi = {10.3389/fmicb.2015.01040}, url = {http://journal.frontiersin.org/Article/10.3389/fmicb.2015.01040/abstract}, author = {Rahaman, Md. Habibur and Islam, Tarequl and Rita R Colwell and Alam, Munirul} } @article {20339, title = {Predictive Time Series Analysis Linking Bengal Cholera with Terrestrial Water Storage Measured from Gravity Recovery and Climate Experiment Sensors}, journal = {The American Journal of Tropical Medicine and Hygiene}, volume = {93}, year = {2015}, month = {Sep-12-2015}, pages = {1179 - 1186}, abstract = {Outbreaks of diarrheal diseases, including cholera, are related to floods and droughts in regions where water and sanitation infrastructure are inadequate or insufficient. However, availability of data on water scarcity and abundance in transnational basins, are a prerequisite for developing cholera forecasting systems. With more than a decade of terrestrial water storage (TWS) data from the Gravity Recovery and Climate Experiment, conditions favorable for predicting cholera occurrence may now be determined. We explored lead{\textendash}lag relationships between TWS in the Ganges{\textendash}Brahmaputra{\textendash}Meghna basin and endemic cholera in Bangladesh. Since bimodal seasonal peaks in cholera in Bangladesh occur during spring and autumn seasons, two separate logistical models between TWS and disease time series (2002{\textendash}2010) were developed. TWS representing water availability showed an asymmetrical, strong association with cholera prevalence in the spring (τ = -0.53; P < 0.001) and autumn (τ = 0.45; P < 0.001) up to 6 months in advance. One unit (centimeter of water) decrease in water availability in the basin increased odds of above normal cholera by 24\% (confidence interval [CI] = 20{\textendash}31\%; P < 0.05) in the spring, while an increase in regional water by 1 unit, through floods, increased odds of above average cholera in the autumn by 29\% (CI = 22{\textendash}33\%; P < 0.05).}, issn = {0002-9637}, doi = {10.4269/ajtmh.14-0648}, url = {http://www.ajtmh.org/content/journals/10.4269/ajtmh.14-0648}, author = {Rita R Colwell and Unnikrishnan, Avinash and Jutla, Antarpreet and Huq, Anwar and Akanda, Ali} } @article {20342, title = {Satellite Based Assessment of Hydroclimatic Conditions Related to Cholera in Zimbabwe}, journal = {PLOS ONE}, year = {2015}, month = {May-09-2017}, pages = {e0137828}, abstract = {Introduction Cholera, an infectious diarrheal disease, has been shown to be associated with large scale hydroclimatic processes. The sudden and sporadic occurrence of epidemic cholera is linked with high mortality rates, in part, due to uncertainty in timing and location of outbreaks. Improved understanding of the relationship between pathogenic abundance and climatic processes allows prediction of disease outbreak to be an achievable goal. In this study, we show association of large scale hydroclimatic processes with the cholera epidemic in Zimbabwe reported to have begun in Chitungwiza, a city in Mashonaland East province, in August, 2008. Principal Findings Climatic factors in the region were found to be associated with triggering cholera outbreak and are shown to be related to anomalies of temperature and precipitation, validating the hypothesis that poor conditions of sanitation, coupled with elevated temperatures, and followed by heavy rainfall can initiate outbreaks of cholera. Spatial estimation by satellite of precipitation and global gridded air temperature captured sensitivities in hydroclimatic conditions that permitted identification of the location in the region where the disease outbreak began. Discussion Satellite derived hydroclimatic processes can be used to capture environmental conditions related to epidemic cholera, as occurred in Zimbabwe, thereby providing an early warning system. Since cholera cannot be eradicated because the causative agent, Vibrio cholerae, is autochthonous to the aquatic environment, prediction of conditions favorable for its growth and estimation of risks of triggering the disease in a given population can be used to alert responders, potentially decreasing infection and saving lives. }, doi = {10.1371/journal.pone.0137828}, url = {https://dx.plos.org/10.1371/journal.pone.0137828}, author = {Jutla, Antarpreet and Aldaach, Haidar and Billian, Hannah and Akanda, Ali and Huq, Anwar and Rita R Colwell}, editor = {Schumann, Guy J-P.} } @article {20340, title = {A unified initiative to harness Earth{\textquoteright}s microbiomes}, journal = {Science}, year = {2015}, month = {Jun-10-2017}, pages = {507 - 508}, issn = {0036-8075}, doi = {10.1126/science.aac8480}, url = {http://www.sciencemag.org/cgi/doi/10.1126/science.aac8480}, author = {Alivisatos, A. P. and Blaser, M. J. and Brodie, E. L. and Chun, M. and Dangl, J. L. and Donohue, T. J. and Dorrestein, P. C. and Gilbert, J. A. and Green, J. L. and Jansson, J. K. and Knight, R. and Maxon, M. E. and McFall-Ngai, M. J. and Miller, J. F. and Pollard, K. S. and Ruby, E. G. and Taha, S. A. and Rita R Colwell} } @article {20365, title = {Global diarrhoea action plan needs integrated climate-based surveillance}, journal = {The Lancet Global Health}, volume = {2}, year = {2014}, month = {Jan-02-2014}, pages = {e69 - e70}, issn = {2214109X}, doi = {10.1016/S2214-109X(13)70155-4}, url = {https://linkinghub.elsevier.com/retrieve/pii/S2214109X13701554}, author = {Akanda, Ali S and Jutla, Antarpreet S and Rita R Colwell} } @article {20361, title = {Microbial Community Profiling of Human Saliva Using Shotgun Metagenomic Sequencing}, journal = {PLoS ONE}, year = {2014}, month = {Aug-05-2015}, pages = {e97699}, abstract = {Human saliva is clinically informative of both oral and general health. Since next generation shotgun sequencing (NGS) is now widely used to identify and quantify bacteria, we investigated the bacterial flora of saliva microbiomes of two healthy volunteers and five datasets from the Human Microbiome Project, along with a control dataset containing short NGS reads from bacterial species representative of the bacterial flora of human saliva. GENIUS, a system designed to identify and quantify bacterial species using unassembled short NGS reads was used to identify the bacterial species comprising the microbiomes of the saliva samples and datasets. Results, achieved within minutes and at greater than 90\% accuracy, showed more than 175 bacterial species comprised the bacterial flora of human saliva, including bacteria known to be commensal human flora but also Haemophilus influenzae, Neisseria meningitidis, Streptococcus pneumoniae, and Gamma proteobacteria. Basic Local Alignment Search Tool (BLASTn) analysis in parallel, reported ca. five times more species than those actually comprising the in silico sample. Both GENIUS and BLAST analyses of saliva samples identified major genera comprising the bacterial flora of saliva, but GENIUS provided a more precise description of species composition, identifying to strain in most cases and delivered results at least 10,000 times faster. Therefore, GENIUS offers a facile and accurate system for identification and quantification of bacterial species and/or strains in metagenomic samples.}, doi = {10.1371/journal.pone.0097699}, url = {https://dx.plos.org/10.1371/journal.pone.0097699}, author = {Hasan, Nur A. and Young, Brian A. and Minard-Smith, Angela T. and Saeed, Kelly and Li, Huai and Heizer, Esley M. and McMillan, Nancy J. and Isom, Richard and Abdullah, Abdul Shakur and Bornman, Daniel M. and Faith, Seth A. and Choi, Seon Young and Dickens, Michael L. and Cebula, Thomas A. and Rita R Colwell}, editor = {Ahmed, Niyaz} } @article {20334, title = {Occurrence in Mexico, 1998{\textendash}2008, of Vibrio cholerae CTX + El Tor carrying an additional truncated CTX prophage}, journal = {Proceedings of the National Academy of Sciences}, volume = {111}, year = {2014}, month = {Aug-07-2014}, pages = {9917 - 9922}, issn = {0027-8424}, doi = {10.1073/pnas.1323408111}, url = {http://www.pnas.org/lookup/doi/10.1073/pnas.1323408111}, author = {Alam, Munirul and Rashed, Shah M and Mannan, Shahnewaj Bin and Islam, Tarequl and Lizarraga-Partida, Marcial L. and Delgado, Gabriela and Morales-Espinosa, Rosario and Mendez, Jose Luis and Navarro, Armando and Watanabe, Haruo and Ohnishi, Makoto and Hasan, Nur A. and Huq, Anwar and Sack, R. Bradley and Rita R Colwell and Cravioto, Alejandro} } @article {20357, title = {Phylodynamic Analysis of Clinical and Environmental Vibrio cholerae Isolates from Haiti Reveals Diversification Driven by Positive Selection}, journal = {mBio}, year = {2014}, month = {Jul-12-2016}, abstract = {Phylodynamic analysis of genome-wide single-nucleotide polymorphism (SNP) data is a powerful tool to investigate underlying evolutionary processes of bacterial epidemics. The method was applied to investigate a collection of 65 clinical and environmental isolates of Vibrio cholerae from Haiti collected between 2010 and 2012. Characterization of isolates recovered from environmental samples identified a total of four toxigenic V. cholerae O1 isolates, four non-O1/O139 isolates, and a novel nontoxigenic V. cholerae O1 isolate with the classical tcpA gene. Phylogenies of strains were inferred from genome-wide SNPs using coalescent-based demographic models within a Bayesian framework. A close phylogenetic relationship between clinical and environmental toxigenic V. cholerae O1 strains was observed. As cholera spread throughout Haiti between October 2010 and August 2012, the population size initially increased and then fluctuated over time. Selection analysis along internal branches of the phylogeny showed a steady accumulation of synonymous substitutions and a progressive increase of nonsynonymous substitutions over time, suggesting diversification likely was driven by positive selection. Short-term accumulation of nonsynonymous substitutions driven by selection may have significant implications for virulence, transmission dynamics, and even vaccine efficacy.}, doi = {10.1128/mBio.01824-14}, url = {http://mbio.asm.org/lookup/doi/10.1128/mBio.01824-14}, author = {Azarian, Taj and Ali, Afsar and Johnson, Judith A. and Mohr, David and Prosperi, Mattia and Veras, Nazle M. and Jubair, Mohammed and Strickland, Samantha L. and Rashid, Mohammad H. and Alam, Meer T. and Weppelmann, Thomas A. and Katz, Lee S. and Tarr, Cheryl L. and Rita R Colwell and Morris, J. Glenn and Salemi, Marco} } @article {19704, title = {Bear-with-me: an embodied prototype to explore tangible two-way exchanges of emotional language}, journal = {CHI{\textquoteright}13 Extended Abstracts on Human Factors in Computing Systems}, year = {2013}, month = {2013/00/01}, abstract = {Abstract Given the busy day-to-day schedule of families and couples, communication between loved ones is often limited to text-messaging, email, or phone calls. These forms of communication do not allow for more tangible modes of intimate expression like hugging. ...}, url = {http://dl.acm.org/citation.cfm?id=2468537}, author = {Fong, A and Ashktorab, Z and Jon Froehlich} } @article {19696, title = {Clutter noise removal in binary document images}, journal = {International Journal on Document Analysis and Recognition (IJDAR)}, volume = {16}, year = {2013}, month = {2013/00/01}, pages = {351 - 369}, publisher = {Springer Berlin Heidelberg}, abstract = {Abstract The paper presents a clutter detection and removal algorithm for complex document images. This distance transform based technique aims to remove irregular and independent unwanted clutter while preserving the text content. The novelty of this approach is in its ...}, doi = {10.1007/s10032-012-0196-6}, url = {http://link.springer.com/article/10.1007/s10032-012-0196-6/fulltext.html}, author = {Agrawal, Mudit and David Doermann} } @article {20378, title = {Detection of Vibrio cholerae in environmental waters including drinking water reservoirs of Azerbaijan}, journal = {Environmental Microbiology Reports}, year = {2013}, month = {Jan-02-2013}, pages = {30 - 38}, doi = {10.1111/j.1758-2229.2012.00369.x }, url = {https://onlinelibrary.wiley.com/doi/10.1111/j.1758-2229.2012.00369.x}, author = {Rashid, Ahmadov and Haley, Bradd J. and Rajabov, Mukhtar and Ahmadova, Sevinj and Gurbanov, Shair and Rita R Colwell and Huq, Anwar} } @article {19160, title = {Dirichlet Mixtures, the Dirichlet Process, and the Structure of Protein Space}, journal = {Journal of Computational Biology}, volume = {20}, number = {1}, year = {2013}, abstract = {The Dirichlet process is used to model probability distributions that are mixtures of an unknown number of components. Amino acid frequencies at homologous positions within related proteins have been fruitfully modeled by Dirichlet mixtures, and we use the Dirichlet process to derive such mixtures with an unbounded number of components. This application of the method requires several technical innovations to sample an unbounded number of Dirichlet-mixture components. The resulting Dirichlet mixtures model multiple-alignment data substantially better than do previously derived ones. They consist of over 500 components, in contrast to fewer than 40 previously, and provide a novel perspective on the structure of proteins. Individual protein positions should be seen not as falling into one of several categories, but rather as arrayed near probability ridges winding through amino acid multinomial space.}, keywords = {ALIGNMENT, computational molecular biology, dynamic programming, multiple alignment, sequence analysis}, doi = {10.1089/cmb.2012.0244}, author = {Nguyen,Viet-An and Jordan Boyd-Graber and Altschul, Stephen F.} } @article {20370, title = {Distribution of Virulence Genes in Clinical and Environmental Vibrio cholerae Strains in Bangladesh}, journal = {Applied and Environmental Microbiology}, year = {2013}, month = {Mar-09-2014}, pages = {5782 - 5785}, abstract = {Vibrio cholerae, an environmental organism, is a facultative human pathogen. Here, we report the virulence profiles, comprising 18 genetic markers, of 102 clinical and 692 environmental V. cholerae strains isolated in Bangladesh between March 2004 and January 2006, showing the variability of virulence determinants within the context of public health. }, issn = {0099-2240}, doi = {10.1128/AEM.01113-13}, url = {http://aem.asm.org/lookup/doi/10.1128/AEM.01113-13}, author = {Hasan, Nur A. and Ceccarelli, Daniela and Grim, Christopher J. and Taviani, Elisa and Choi, Jinna and Sadique, Abdus and Alam, Munirul and Siddique, Abul K. and Sack, R. Bradley and Huq, Anwar and Rita R Colwell} } @article {20367, title = {Drug response and genetic properties of Vibrio cholerae associated with endemic cholera in north-eastern Thailand, 2003-2011}, journal = {Journal of Medical Microbiology}, year = {2013}, month = {Jan-04-2013}, pages = {599 - 609}, abstract = {Cholera, caused by Vibrio cholerae, results in significant morbidity and mortality worldwide, including Thailand. Representative V. cholerae strains associated with endemic cholera (n = 32), including strains (n = 3) from surface water sources, in Khon Kaen, Thailand (2003{\textendash}2011), were subjected to microbiological, molecular and phylogenetic analyses. According to phenotypic and related genetic data, all tested V. cholerae strains belonged to serogroup O1, biotype El Tor (ET), Inaba (IN) or Ogawa (OG). All of the strains were sensitive to gentamicin and ciprofloxacin, while multidrug-resistant (MDR) strains showing resistance to erythromycin, tetracycline, trimethoprim/sulfamethoxazole and ampicillin were predominant in 2007. V. cholerae strains isolated before and after 2007 were non-MDR. All except six diarrhoeal strains possessed ctxA and ctxB genes and were toxigenic altered ET, confirmed by MAMA-PCR and DNA sequencing. Year-wise data revealed that V. cholerae INET strains isolated between 2003 and 2004, plus one strain isolated in 2007, lacked the RS1 sequence (rstC) and toxin-linked cryptic plasmid (TLC)-specific genetic marker, but possessed CTXCL prophage genes ctxB CL and rstR CL. A sharp genetic transition was noted, namely the majority of V. cholerae strains in 2007 and all in 2010 and 2011 were not repressor genotype rstR CL but instead were rstR ET, and all ctx + strains possessed RS1 and TLC-specific genetic markers. DNA sequencing data revealed that strains isolated since 2007 had a mutation in the tcpA gene at amino acid position 64 (N{\textrightarrow}S). Four clonal types, mostly of environmental origin, including subtypes, reflected genetic diversity, while distinct signatures were observed for clonally related, altered ET from Thailand, Vietnam and Bangladesh, confirmed by distinct subclustering patterns observed in the PFGE (NotI)-based dendrogram, suggesting that endemic cholera is caused by V. cholerae indigenous to Khon Kaen.}, issn = {0022-2615}, doi = {10.1099/jmm.0.053801-0}, url = {http://jmm.microbiologyresearch.org/content/journal/jmm/10.1099/jmm.0.053801-0}, author = {Chomvarin, C. and Johura, F.-T. and Mannan, S. B. and Jumroenjit, W. and Kanoktippornchai, B. and Tangkanakul, W. and Tantisuwichwong, N. and Huttayananont, S. and Watanabe, H. and Hasan, N. A. and Huq, A. and Cravioto, A. and Rita R Colwell and Alam, M.} } @article {20363, title = {Engaging Actively with Issues in the Responsible Conduct of Science: Lessons from International Efforts Are Relevant for Undergraduate Education in the United States}, journal = {CBE{\textemdash}Life Sciences Education}, year = {2013}, month = {Jan-12-2013}, pages = {596 - 603}, abstract = {Numerous studies are demonstrating that engaging undergraduate students in original research can improve their achievement in the science, technology, engineering, and mathematics (STEM) fields and increase the likelihood that some of them will decide to pursue careers in these disciplines. Associated with this increased prominence of research in the undergraduate curriculum are greater expectations from funders, colleges, and universities that faculty mentors will help those students, along with their graduate students and postdoctoral fellows, develop an understanding and sense of personal and collective obligation for responsible conduct of science (RCS). This Feature describes an ongoing National Research Council (NRC) project and a recent report about educating faculty members in culturally diverse settings (Middle East/North Africa and Asia) to employ active-learning strategies to engage their students and colleagues deeply in issues related to RCS. The NRC report describes the first phase of this project, which took place in Aqaba and Amman, Jordan, in September 2012 and April 2013, respectively. Here we highlight the findings from that report and our subsequent experience with a similar interactive institute in Kuala Lumpur, Malaysia. Our work provides insights and perspectives for faculty members in the United States as they engage undergraduate and graduate students, as well as postdoctoral fellows, to help them better understand the intricacies of and connections among various components of RCS. Further, our experiences can provide insights for those who may wish to establish {\textquotedblleft}train-the-trainer{\textquotedblright} programs at their home institutions.}, doi = {10.1187/cbe.13-09-0184}, url = {https://www.lifescied.org/doi/10.1187/cbe.13-09-0184}, author = {Clements, John D. and Connell, Nancy D. and Dirks, Clarissa and El-Faham, Mohamed and Hay, Alastair and Heitman, Elizabeth and Stith, James H. and Bond, Enriqueta C. and Rita R Colwell and Anestidou, Lida and Husbands, Jo L. and Labov, Jay B.} } @article {20372, title = {Environmental Factors Influencing Epidemic Cholera}, journal = {The American Journal of Tropical Medicine and Hygiene}, volume = {89}, year = {2013}, month = {Apr-09-2013}, pages = {597 - 607}, abstract = {Cholera outbreak following the earthquake of 2010 in Haiti has reaffirmed that the disease is a major public health threat. Vibrio cholerae is autochthonous to aquatic environment, hence, it cannot be eradicated but hydroclimatology-based prediction and prevention is an achievable goal. Using data from the 1800s, we describe uniqueness in seasonality and mechanism of occurrence of cholera in the epidemic regions of Asia and Latin America. Epidemic regions are located near regional rivers and are characterized by sporadic outbreaks, which are likely to be initiated during episodes of prevailing warm air temperature with low river flows, creating favorable environmental conditions for growth of cholera bacteria. Heavy rainfall, through inundation or breakdown of sanitary infrastructure, accelerates interaction between contaminated water and human activities, resulting in an epidemic. This causal mechanism is markedly different from endemic cholera where tidal intrusion of seawater carrying bacteria from estuary to inland regions, results in outbreaks.}, issn = {0002-9637}, doi = {10.4269/ajtmh.12-0721}, url = {http://www.ajtmh.org/content/journals/10.4269/ajtmh.12-0721}, author = {Huq, Anwar and Hasan, Nur and Akanda, Ali and Whitcombe, Elizabeth and Rita R Colwell and Haley, Bradd and Alam, Munir and Jutla, Antarpreet and Sack, R. Bradley} } @article {19700, title = {Improving public transit accessibility for blind riders by crowdsourcing bus stop landmark locations with Google street view}, journal = {The 15th International ACM SIGACCESS Conference}, year = {2013}, month = {2013/00/21}, pages = {16 - 8}, publisher = {SIGACCESS, ACM Special Interest Group on Accessible ComputingACM}, address = {New York, New York, USA}, abstract = {Abstract Low-vision and blind bus riders often rely on known physical landmarks to help locate and verify bus stop locations (eg., by searching for a shelter, bench, newspaper bin). However, there are currently few, if any, methods to determine this information a priori via ...}, isbn = {9781450324052}, doi = {10.1145/2513383.2513448}, url = {http://dl.acm.org/citation.cfm?doid=2513383.2513448}, author = {Hara, Kotaro and Azenkot, Shiri and Campbell, Megan and Bennett, Cynthia L and Le, Vicki and Pannella, Sean and Moore, Robert and Minckler, Kelly and Ng, Rochelle H and Jon Froehlich} } @conference {19695, title = {Large-Scale Signature Matching Using Multi-stage Hashing}, booktitle = {Document Analysis and Recognition (ICDAR), 2013 12th International Conference on}, year = {2013}, month = {2013/00/01}, pages = {976 - 980}, publisher = {IEEE}, organization = {IEEE}, abstract = {In this paper, we propose a fast large-scale signature matching method based on locality sensitive hashing (LSH). Shape Context features are used to describe the structure of signatures. Two stages of hashing are performed to find the nearest neighbours for query signatures. In the first stage, we use M randomly generated hyper planes to separate shape context feature points into different bins, and compute a term-frequency histogram to represent the feature point distribution as a feature vector. In the second stage we again use LSH to categorize the high-level features into different classes. The experiments are carried out on two datasets - DS-I, a small dataset contains 189 signatures, and DS-II, a large dataset created by our group which contains 26,000 signatures. We show that our algorithm can achieve a high accuracy even when few signatures are collected from one same person and perform fast matching when dealing with a large dataset. View full abstract}, isbn = {978-0-7695-4999-6}, doi = {10.1109/ICDAR.2013.197}, url = {http://ieeexplore.ieee.org/lpdocs/epic03/wrapper.htm?arnumber=6628762}, author = {Du, Xianzhi and Abdalmageed, W and David Doermann} } @article {20373, title = {Species composition of limnetic zooplankton from the southern coastal areas (Mathbaria and Bakerganj) in Bangladesh}, journal = {Journal of the Asiatic Society of Bangladesh, Science}, volume = {38}, year = {2013}, month = {Apr-06-2014}, abstract = {We studied the coastal zooplankton community structure of six waterbodies of Mathbaria and Bakerganj from January 2008 to June 2009. In total 35 zooplankton species were identified under 26 genera under 20 families under 8 orders from Mathbaria. Among them 6 were protozoans, 24 were rotifers, 3 were copepods, 1 was cladocerans and 1 was ostracods. From Bakergonj a total of 42 zooplankton species were identified under 23 genera under 17 Families under 7 orders. Among them 3 were protozoans, 25 were rotifers, 7 were copepods, 6 were cladocerans and 1 was ostracods. Results showed that abundance of rotifera group from both area was higher in Mathbaria (64.86\%) and Bakerganj (60.98\%) than other groups while abundance of ostracoda of Bakerganj area was lowest.}, issn = {1016-6947}, doi = {10.3329/jasbs.v38i1.15326}, url = {http://banglajol.info/index.php/JASBS/article/view/15326}, author = {Mozumder, Pronob K and Nahar, Samsun and Naser, M Niamul and Alam, Munirul and Huq, Anwar and Sack, R Bradley and Rita R Colwell} } @article {20377, title = {A water marker monitored by satellites to predict seasonal endemic cholera}, journal = {Remote Sensing Letters}, volume = {4472741982394456}, year = {2013}, month = {Jan-08-2013}, pages = {822 - 831}, abstract = {The ability to predict an occurrence of cholera, a water-related disease, offers a significant public health advantage. Satellite-based estimates of chlorophyll, a surrogate for plankton abundance, have been linked to cholera incidence. However, cholera bacteria can survive under a variety of coastal ecological conditions, thus constraining the predictive ability of the chlorophyll, since it provides only an estimate of greenness of seawater. Here, a new remote-sensing-based index is proposed: Satellite Water Marker (SWM), which estimates the condition of coastal water, based on observed variability in the difference between blue (412 nm) and green (555 nm) wavelengths that can be related to seasonal cholera incidence. The index is bounded between physically separable wavelengths for relatively clear (blue) and turbid (green) water. Using SWM, prediction of cholera with reasonable accuracy, at least two months in advance, can potentially be achieved in the endemic coastal regions. }, issn = {2150-704X}, doi = {10.1080/2150704X.2013.802097}, url = {http://www.tandfonline.com/doi/abs/10.1080/2150704X.2013.802097}, author = {Jutla, Antarpreet and Akanda, Ali Shafqat and Huq, Anwar and Faruque, Abu Syed Golam and Rita R Colwell and Islam, Shafiqul} } @article {20375, title = {An additional step in the transmission of Yersinia pestis?}, journal = {The ISME Journal}, year = {2012}, month = {Jan-02-2012}, pages = {231 - 236}, abstract = {Plague, caused by the bacterium Yersinia pestis, is a mammalian vector-borne disease, transmitted by fleas that serve as the vector between rodent hosts. For many pathogens, including Y. pestis, there are strong evolutionary pressures that lead to a reduction in {\textquoteleft}useless genes{\textquoteright}, with only those retained that reflect function in the specific environment inhabited by the pathogen. Genetic traits critical for survival and transmission between two environments, the rodent and the flea, are conserved in epizootic/epidemic plague strains. However, there are genes that remain conserved for which no function in the flea{\textendash}rodent cycle has yet been observed, indicating an additional environment may exist in the transmission cycle of plague. Here, we present evidence for highly conserved genes that suggests a role in the persistence of Y. pestis after death of its host. Furthermore, maintenance of these genes points to Y. pestis traversing a post-mortem path between, and possibly within, epizootic periods and offering insight into mechanisms that may allow Y. pestis an alternative route of transmission in the natural environment.}, issn = {1751-7362}, doi = {10.1038/ismej.2011.105}, url = {http://www.nature.com/articles/ismej2011105}, author = {Easterday, W Ryan and Kausrud, Kyrre L and Star, Bastiaan and Heier, Lise and Haley, Bradd J and Ageyev, Vladimir and Rita R Colwell and Stenseth, Nils Chr} } @article {12980, title = {BEAGLE: An Application Programming Interface and High-Performance Computing Library for Statistical Phylogenetics}, journal = {Systematic BiologySyst Biol}, volume = {61}, year = {2012}, month = {2012/01/01/}, pages = {170 - 173}, abstract = {Phylogenetic inference is fundamental to our understanding of most aspects of the origin and evolution of life, and in recent years, there has been a concentration of interest in statistical approaches such as Bayesian inference and maximum likelihood estimation. Yet, for large data sets and realistic or interesting models of evolution, these approaches remain computationally demanding. High-throughput sequencing can yield data for thousands of taxa, but scaling to such problems using serial computing often necessitates the use of nonstatistical or approximate approaches. The recent emergence of graphics processing units (GPUs) provides an opportunity to leverage their excellent floating-point computational performance to accelerate statistical phylogenetic inference. A specialized library for phylogenetic calculation would allow existing software packages to make more effective use of available computer hardware, including GPUs. Adoption of a common library would also make it easier for other emerging computing architectures, such as field programmable gate arrays, to be used in the future. We present BEAGLE, an application programming interface (API) and library for high-performance statistical phylogenetic inference. The API provides a uniform interface for performing phylogenetic likelihood calculations on a variety of compute hardware platforms. The library includes a set of efficient implementations and can currently exploit hardware including GPUs using NVIDIA CUDA, central processing units (CPUs) with Streaming SIMD Extensions and related processor supplementary instruction sets, and multicore CPUs via OpenMP. To demonstrate the advantages of a common API, we have incorporated the library into several popular phylogenetic software packages. The BEAGLE library is free open source software licensed under the Lesser GPL and available from http://beagle-lib.googlecode.com. An example client program is available as public domain software.}, keywords = {Bayesian phylogenetics, gpu, maximum likelihood, parallel computing}, isbn = {1063-5157, 1076-836X}, doi = {10.1093/sysbio/syr100}, url = {http://sysbio.oxfordjournals.org/content/61/1/170}, author = {Ayres,Daniel L and Darling,Aaron and Zwickl,Derrick J and Beerli,Peter and Holder,Mark T and Lewis,Paul O and Huelsenbeck,John P and Ronquist,Fredrik and Swofford,David L and Cummings, Michael P. and Rambaut,Andrew and Suchard,Marc A} } @article {20063, title = {Biases from model assumptions in texture sub-cellular image segmentation}, journal = {SPIE Newsroom}, volume = {13}, year = {2012}, author = {Cardone, Antonio and Amelot, Julien and Li-Baboud, Ya-Shian and Brady, Mary and Bajcsy, Peter} } @article {14328, title = {Disaggregated water sensing from a single, pressure-based sensor: An extended analysis of HydroSense using staged experiments}, journal = {Pervasive and Mobile Computing}, volume = {8}, year = {2012}, month = {2012/02//}, pages = {82 - 102}, abstract = {We present an extended analysis of our previous work on the HydroSense technology, which is a low-cost and easily installed single-point sensor of pressure for automatically disaggregating water usage activities in the home (Froehlich et~al., 2009~[53]). We expand upon this work by providing a survey of existing and emerging water disaggregation techniques, a more comprehensive description of the theory of operation behind our approach, and an expanded analysis section that includes hot versus cold water valve usage classification and a comparison between two classification approaches: the template-based matching scheme used in Froehlich et~al. (2009)~[53] and a new stochastic approach using a Hidden Markov Model. We show that both are successful in identifying valve- and fixture-level water events with greater than 90\% accuracies. We conclude with a discussion of the limitations in our experimental methodology and open problems going forward.}, keywords = {Activity sensing, H1.2, H5.2, Infrastructure-mediated sensing, Sensors, Water usage}, isbn = {1574-1192}, doi = {10.1016/j.pmcj.2010.08.008}, url = {http://www.sciencedirect.com/science/article/pii/S1574119210000842}, author = {Larson,Eric and Jon Froehlich and Campbell,Tim and Haggerty,Conor and Atlas,Les and Fogarty,James and Patel,Shwetak N.} } @article {14330, title = {Fast track article: Disaggregated water sensing from a single, pressure-based sensor: An extended analysis of HydroSense using staged experiments}, journal = {Pervasive Mob. Comput.}, volume = {8}, year = {2012}, month = {2012/02//}, pages = {82 - 102}, abstract = {We present an extended analysis of our previous work on the HydroSense technology, which is a low-cost and easily installed single-point sensor of pressure for automatically disaggregating water usage activities in the home (Froehlich et al., 2009 [53]). We expand upon this work by providing a survey of existing and emerging water disaggregation techniques, a more comprehensive description of the theory of operation behind our approach, and an expanded analysis section that includes hot versus cold water valve usage classification and a comparison between two classification approaches: the template-based matching scheme used in Froehlich et al. (2009) [53] and a new stochastic approach using a Hidden Markov Model. We show that both are successful in identifying valve- and fixture-level water events with greater than 90\% accuracies. We conclude with a discussion of the limitations in our experimental methodology and open problems going forward.}, keywords = {Activity sensing, H1.2, H5.2, Infrastructure-mediated sensing, Sensors, Water usage}, isbn = {1574-1192}, doi = {10.1016/j.pmcj.2010.08.008}, url = {http://dx.doi.org/10.1016/j.pmcj.2010.08.008}, author = {Larson,Eric and Jon Froehlich and Campbell,Tim and Haggerty,Conor and Atlas,Les and Fogarty,James and Patel,Shwetak N.} } @article {19131, title = {A framework for human microbiome research}, journal = {Nature}, volume = {486}, year = {2012}, month = {2012}, pages = {215 - 221}, abstract = {A variety of microbial communities and their genes (the microbiome) exist throughout the human body, with fundamental roles in human health and disease. The National Institutes of Health (NIH)-funded Human Microbiome Project Consortium has established a population-scale framework to develop metagenomic protocols, resulting in a broad range of quality-controlled resources and data including standardized methods for creating, processing and interpreting distinct types of high-throughput metagenomic data available to the scientific community. Here we present resources from a population of 242 healthy adults sampled at 15 or 18 body sites up to three times, which have generated 5,177 microbial taxonomic profiles from 16S ribosomal RNA genes and over 3.5 terabases of metagenomic sequence so far. In parallel, approximately 800 reference strains isolated from the human body have been sequenced. Collectively, these data represent the largest resource describing the abundance and variety of the human microbiome, while providing a framework for current and future studies.}, author = {Meth{\'e}, B.A. and Nelson,K. E and Pop, Mihai and Creasy, H.H. and Giglio, M.G. and Huttenhower, C. and Gevers, D. and Petrosino, J.F. and Abubucker, S. and Badger, J.H.} } @article {20374, title = {Genetic characteristics of drug-resistant Vibrio cholerae O1 causing endemic cholera in Dhaka, 2006-2011}, journal = {Journal of Medical Microbiology}, year = {2012}, month = {Jan-12-2012}, pages = {1736 - 1745}, abstract = {Vibrio cholerae O1 biotype El Tor (ET), causing the seventh cholera pandemic, was recently replaced in Bangladesh by an altered ET possessing ctxB of the Classical (CL) biotype, which caused the first six cholera pandemics. In the present study, V. cholerae O1 strains associated with endemic cholera in Dhaka between 2006 and 2011 were analysed for major phenotypic and genetic characteristics. Of 54 representative V. cholerae isolates tested, all were phenotypically ET and showed uniform resistance to trimethoprim/sulfamethoxazole (SXT) and furazolidone (FR). Resistance to tetracycline (TE) and erythromycin (E) showed temporal fluctuation, varying from year to year, while all isolates were susceptible to gentamicin (CN) and ciprofloxacin (CIP). Year-wise data revealed erythromycin resistance to be 33.3 \% in 2006 and 11 \% in 2011, while tetracycline resistance accounted for 33, 78, 0, 100 and 27 \% in 2006, 2007, 2008, 2009 and 2010, respectively; interestingly, all isolates tested were sensitive to TE in 2011, as observed in 2008. All V. cholerae isolates tested possessed genetic elements such as SXT, ctxAB, tcpA ET, rstR ET and rtxC; none had IntlI (Integron I). Double mismatch amplification mutation assay (DMAMA)-PCR followed by DNA sequencing and analysis of the ctxB gene revealed a point mutation at position 58 (C{\textrightarrow}A), which has resulted in an amino acid substitution from histidine (H) to asparagine (N) at position 20 (genotype 7) since 2008. Although the multi-resistant strains having tetracycline resistance showed minor genetic divergence, V. choleraestrains were clonal, as determined by a PFGE (NotI)-based dendrogram. This study shows 2008{\textendash}2010 to be the time of transition from ctxB genotype 1 to genotype 7 in V. cholerae ET causing endemic cholera in Dhaka, Bangladesh.}, issn = {0022-2615}, doi = {10.1099/jmm.0.049635-0}, url = {http://jmm.microbiologyresearch.org/content/journal/jmm/10.1099/jmm.0.049635-0}, author = {Rashed, S. M. and Mannan, S. B. and Johura, F.-T. and Islam, M. T. and Sadique, A. and Watanabe, H. and Sack, R. B. and Huq, A. and Rita R Colwell and Cravioto, A. and Alam, M.} } @article {20386, title = {Genomic diversity of 2010 Haitian cholera outbreak strains}, journal = {Proceedings of the National Academy of Sciences}, year = {2012}, month = {05/2012}, pages = {E2010 - E2017}, issn = {0027-8424}, doi = {10.1073/pnas.1207359109}, url = {http://www.pnas.org/cgi/doi/10.1073/pnas.1207359109}, author = {Hasan, N. A. and Choi, S. Y. and Eppinger, M. and Clark, P. W. and Chen, A. and Alam, M. and Haley, B. J. and Taviani, E. and Hine, E. and Su, Q. and Tallon, L. J. and Prosper, J. B. and Furth, K. and Hoq, M. M. and Li, H. and Fraser-Liggett, C. M. and Cravioto, A. and Huq, A. and Ravel, J. and Cebula, T. A. and Rita R Colwell} } @article {19660, title = {Global secretome analysis identifies novel mediators of bone metastasis}, journal = {Cell Research}, volume = {22}, year = {2012}, month = {2012/09//}, pages = {1339 - 1355}, abstract = {Bone is the one of the most common sites of distant metastasis of solid tumors. Secreted proteins are known to influence pathological interactions between metastatic cancer cells and the bone stroma. To comprehensively profile secreted proteins associated with bone metastasis, we used quantitative and non-quantitative mass spectrometry to globally analyze the secretomes of nine cell lines of varying bone metastatic ability from multiple species and cancer types. By comparing the secretomes of parental cells and their bone metastatic derivatives, we identified the secreted proteins that were uniquely associated with bone metastasis in these cell lines. We then incorporated bioinformatic analyses of large clinical metastasis datasets to obtain a list of candidate novel bone metastasis proteins of several functional classes that were strongly associated with both clinical and experimental bone metastasis. Functional validation of selected proteins indicated that in vivo bone metastasis can be promoted by high expression of (1) the salivary cystatins CST1, CST2, and CST4; (2) the plasminogen activators PLAT and PLAU; or (3) the collagen functionality proteins PLOD2 and COL6A1. Overall, our study has uncovered several new secreted mediators of bone metastasis and therefore demonstrated that secretome analysis is a powerful method for identification of novel biomarkers and candidate therapeutic targets.}, keywords = {bone, cancer, metastasis, proteomics, secretome}, isbn = {1001-0602}, url = {http://www.nature.com/cr/journal/v22/n9/abs/cr201289a.html}, author = {Blanco, Mario Andres and LeRoy, Gary and Zia Khan and Ale{\v c}kovi{\'c}, Ma{\v s}a and Zee, Barry M. and Garcia, Benjamin A. and Kang, Yibin} } @article {15427, title = {Grey-box GUI Testing: Efficient Generation of Event Sequences}, journal = {arXiv:1205.4928}, year = {2012}, month = {2012/05/22/}, abstract = {Graphical user interfaces (GUIs), due to their event driven nature, present a potentially unbounded space of all possible ways to interact with software. During testing it becomes necessary to effectively sample this space. In this paper we develop algorithms that sample the GUI{\textquoteright}s input space by only generating sequences that (1) are allowed by the GUI{\textquoteright}s structure, and (2) chain together only those events that have data dependencies between their event handlers. We create a new abstraction, called an event-dependency graph (EDG) of the GUI, that captures data dependencies between event handler code. We develop a mapping between EDGs and an existing black-box user-level model of the GUI{\textquoteright}s workflow, called an event-flow graph (EFG). We have implemented automated EDG construction in a tool that analyzes the bytecode of each event handler. We evaluate our "grey-box" approach using four open-source applications and compare it with the current state-of-the-art EFG approach. Our results show that using the EDG reduces the number of test cases while still achieving at least the same coverage. Furthermore, we were able to detect 2 new bugs in the subject applications.}, keywords = {68N30, Computer Science - Software Engineering}, url = {http://arxiv.org/abs/1205.4928}, author = {Arlt,Stephan and Banerjee,Ishan and Bertolini,Cristiano and Memon, Atif M. and Sch{\"a}f,Martin} } @article {19721, title = {InterPro in 2011: new developments in the family and domain prediction database.}, journal = {Nucleic Acids Res}, volume = {40}, year = {2012}, month = {2012 Jan}, pages = {D306-12}, abstract = {

InterPro (http://www.ebi.ac.uk/interpro/) is a database that integrates diverse information about protein families, domains and functional sites, and makes it freely available to the public via Web-based interfaces and services. Central to the database are diagnostic models, known as signatures, against which protein sequences can be searched to determine their potential function. InterPro has utility in the large-scale analysis of whole genomes and meta-genomes, as well as in characterizing individual protein sequences. Herein we give an overview of new developments in the database and its associated software since 2009, including updates to database content, curation processes and Web and programmatic interfaces.

}, keywords = {Databases, Protein, Protein Structure, Tertiary, Proteins, Sequence Analysis, Protein, software, Terminology as Topic, User-Computer Interface}, issn = {1362-4962}, doi = {10.1093/nar/gkr948}, author = {Hunter, Sarah and Jones, Philip and Mitchell, Alex and Apweiler, Rolf and Attwood, Teresa K and Bateman, Alex and Bernard, Thomas and Binns, David and Bork, Peer and Burge, Sarah and de Castro, Edouard and Coggill, Penny and Corbett, Matthew and Das, Ujjwal and Daugherty, Louise and Duquenne, Lauranne and Finn, Robert D and Fraser, Matthew and Gough, Julian and Haft, Daniel and Hulo, Nicolas and Kahn, Daniel and Kelly, Elizabeth and Letunic, Ivica and Lonsdale, David and Lopez, Rodrigo and Madera, Martin and Maslen, John and McAnulla, Craig and McDowall, Jennifer and McMenamin, Conor and Mi, Huaiyu and Mutowo-Muellenet, Prudence and Mulder, Nicola and Natale, Darren and Orengo, Christine and Pesseat, Sebastien and Punta, Marco and Quinn, Antony F and Rivoire, Catherine and Sangrador-Vegas, Amaia and Jeremy D Selengut and Sigrist, Christian J A and Scheremetjew, Maxim and Tate, John and Thimmajanarthanan, Manjulapramila and Thomas, Paul D and Wu, Cathy H and Yeats, Corin and Yong, Siew-Yit} } @article {12854, title = {Vibrio Cholerae Classical Biotype Strains Reveal Distinct Signatures in Mexico}, journal = {Journal of Clinical Microbiology}, year = {2012}, month = {04/2012}, abstract = {Vibrio cholerae O1 Classical (CL) biotype caused the 5th and 6th, and probably the earlier cholera pandemics, before the El Tor (ET) biotype initiated the 7th pandemic in Asia in the 1970{\textquoteright}s by completely displacing the CL biotype. Although the CL biotype was thought to be extinct in Asia, and it had never been reported from Latin America, V. cholerae CL and ET biotypes, including hybrid ET were found associated with endemic cholera in Mexico between 1991 and 1997. In this study, CL biotype strains isolated from endemic cholera in Mexico, between 1983 and 1997 were characterized in terms of major phenotypic and genetic traits, and compared with CL biotype strains isolated in Bangladesh between 1962 and 1989. According to sero- and bio-typing data, all V. cholerae strains tested had the major phenotypic and genotypic characteristics specific for the CL biotype. Antibiograms revealed the majority of the Bangladeshi strains to be resistant to trimethoprim/sulfamethoxazole, furazolidone, ampicillin, and gentamycin, while the Mexican strains were sensitive to all of these drugs, as well as to ciprofloxacin, erythromycin, and tetracycline. Pulsed-field gel electrophoresis (PFGE) of NotI-digested genomic DNA revealed characteristic banding patterns for all the CL biotype strains, although the Mexican strains differed with the Bangladeshi strains in 1-2 DNA bands. The difference may be subtle, but consistent, as confirmed by the sub-clustering patterns in the PFGE-based dendrogram, and can serve as regional signature, suggesting pre-1991 existence and evolution of the CL biotype strains in the Americas, independent from that of Asia.}, isbn = {0095-1137, 1098-660X}, doi = {10.1128/JCM.00189-12}, url = {http://jcm.asm.org/content/early/2012/04/12/JCM.00189-12}, author = {Alam,Munirul and Islam,M. Tarequl and Rashed,Shah Manzur and Johura,Fatema-Tuz and Bhuiyan,Nurul A. and Delgado,Gabriela and Morales,Rosario and Mendez,Jose Luis and Navarro,Armando and Watanabe,Haruo and Hasan,Nur-A. and Rita R Colwell and Cravioto,Alejandro} } @conference {12418, title = {Mr. LDA: A Flexible Large Scale Topic Modeling Package using Variational Inference in MapReduce}, booktitle = {Proceedings of ACM International Conference on World Wide Web, 2012}, year = {2012}, month = {2012///}, abstract = {Latent Dirichlet Allocation (LDA) is a popular topic modeling tech- nique for exploring document collections. Because of the increasing prevalence of large datasets, there is a need to improve the scal- ability of inference for LDA. In this paper, we introduce a novel and flexible large scale topic modeling package in MapReduce (Mr. LDA). As opposed to other techniques which use Gibbs sampling, our proposed framework uses variational inference, which easily fits into a distributed environment. More importantly, this variational implementation, unlike highly tuned and specialized implementa- tions based on Gibbs sampling, is easily extensible. We demonstrate two extensions of the models possible with this scalable framework: informed priors to guide topic discovery and extracting topics from a multilingual corpus. We compare the scalability of Mr. LDA against Mahout, an existing large scale topic modeling package. Mr. LDA out-performs Mahout both in execution speed and held-out likelihood.}, author = {Zhai,Ke and Jordan Boyd-Graber and Asadi,Nima and Alkhouja,Mohamad} } @article {12857, title = {Occurrence of protozoans \& their limnological relationships in some ponds of Mathbaria, Bangladesh}, journal = {University Journal of Zoology, Rajshahi University}, volume = {29}, year = {2012}, month = {2012///}, pages = {69 - 71}, isbn = {1023-6104}, url = {http://journals.sfu.ca/bd/index.php/UJZRU}, author = {Mozumder,P. K. and Banu,M. A. and Naser,M. N. and Ali,M. S. and Alam,M. and Sack,R. B. and Rita R Colwell and Huq,A.} } @conference {15650, title = {Polytope approximation and the Mahler volume}, booktitle = {Proceedings of the Twenty-Third Annual ACM-SIAM Symposium on Discrete Algorithms}, series = {SODA {\textquoteright}12}, year = {2012}, month = {2012///}, pages = {29 - 42}, publisher = {SIAM}, organization = {SIAM}, abstract = {The problem of approximating convex bodies by polytopes is an important and well studied problem. Given a convex body K in Rd, the objective is to minimize the number of vertices (alternatively, the number of facets) of an approximating polytope for a given Hausdorff error ε. Results to date have been of two types. The first type assumes that K is smooth, and bounds hold in the limit as ε tends to zero. The second type requires no such assumptions. The latter type includes the well known results of Dudley (1974) and Bronshteyn and Ivanov (1976), which show that in spaces of fixed dimension, O((diam(K)/ε)(d-1)/2) vertices (alt., facets) suffice. Our results are of this latter type. In our first result, under the assumption that the width of the body in any direction is at least ε, we strengthen the above bound to [EQUATION]. This is never worse than the previous bound (by more than logarithmic factors) and may be significantly better for skinny bodies. Our analysis exploits an interesting analogy with a classical concept from the theory of convexity, called the Mahler volume. This is a dimensionless quantity that involves the product of the volumes of a convex body and its polar dual. In our second result, we apply the same machinery to improve upon the best known bounds for answering ε-approximate polytope membership queries. Given a convex polytope P defined as the intersection of halfspaces, such a query determines whether a query point q lies inside or outside P, but may return either answer if q{\textquoteright}s distance from P{\textquoteright}s boundary is at most ε. We show that, without increasing storage, it is possible to reduce the best known search times for ε-approximate polytope membership significantly. This further implies improvements to the best known search times for approximate nearest neighbor searching in spaces of fixed dimension.}, url = {http://dl.acm.org/citation.cfm?id=2095116.2095119}, author = {Arya,Sunil and da Fonseca,Guilherme D. and Mount, Dave} } @article {19668, title = {Quantitative measurement of allele-specific protein expression in a diploid yeast hybrid by LC-MS}, journal = {Molecular Systems Biology}, volume = {8}, year = {2012}, month = {2012/01/01/}, abstract = {Understanding the genetic basis of gene regulatory variation is a key goal of evolutionary and medical genetics. Regulatory variation can act in an allele-specific manner (cis-acting) or it can affect both alleles of a gene (trans-acting). Differential allele-specific expression (ASE), in which the expression of one allele differs from another in a diploid, implies the presence of cis-acting regulatory variation. While microarrays and high-throughput sequencing have enabled genome-wide measurements of transcriptional ASE, methods for measurement of protein ASE (pASE) have lagged far behind. We describe a flexible, accurate, and scalable strategy for measurement of pASE by liquid chromatography-coupled mass spectrometry (LC-MS). We apply this approach to a hybrid between the yeast species Saccharomyces cerevisiae and Saccharomyces bayanus. Our results provide the first analysis of the relative contribution of cis-acting and trans-acting regulatory differences to protein expression divergence between yeast species.Synopsis A novel strategy for the quantitative measurement of allele-specific protein expression is used to infer the contributions of cis- and trans-acting factors influencing the divergence of protein levels between yeast species. Rigorous experimental controls and analyses confirm the accuracy of the new strategy for the quantitative measurement of allele-specific protein expression by high-throughput mass spectrometry.Analysis of allele-specific protein expression in an interspecies yeast hybrid and protein expression differences between species reveals that both cis-effects and trans-effects contribute to protein expression divergence between two yeast species, Saccharomyces cerevisiae and Saccharomyces bayanus. }, keywords = {allele specific, divergence, mass spectrometry, protein expression, proteomics}, url = {http://msb.embopress.org/content/8/1/602}, author = {Zia Khan and Bloom, Joshua S. and Amini, Sasan and Singh, Mona and Perlman, David H. and Caudy, Amy A. and Kruglyak, Leonid} } @article {12855, title = {Role of Shrimp Chitin in the Ecology of Toxigenic Vibrio cholerae and Cholera Transmission}, journal = {Frontiers in Microbiology}, volume = {2:260}, year = {2012}, month = {01/2012}, abstract = {Seasonal plankton blooms correlate with occurrence of cholera in Bangladesh, although the mechanism of how dormant Vibrio cholerae, enduring interepidemic period in biofilms and plankton, initiates seasonal cholera is not fully understood. In this study, laboratory microcosms prepared with estuarine Mathbaria water (MW) samples supported active growth of toxigenic V. cholerae O1 up to 7 weeks as opposed to 6 months when microcosms were supplemented with dehydrated shrimp chitin chips (CC) as the single source of nutrient. Bacterial counting and detection of wbe and ctxA genes were done employing culture, direct fluorescent antibody (DFA) assay, and multiplex-polymerase chain reaction methods. In MW microcosm, the aqueous phase became clear as the non-culturable cells settled, whereas the aqueous phase of the MW{\textendash}CC microcosm became turbid from bacterial growth stimulated by chitin. Bacterial chitin degradation and biofilm formation proceeded from an initial steady state to a gradually declining bacterial culturable count. V. cholerae within the microenvironments of chitin and chitin-associated biofilms remained metabolically active even in a high acidic environment without losing either viability or virulence. It is concluded that the abundance of chitin that occurs during blooms plays an important role in the aquatic life cycle of V. cholerae and, ultimately, in the seasonal transmission of cholera.}, isbn = {1664-302X}, doi = {10.3389/fmicb.2011.00260}, author = {Nahar,Shamsun and Sultana,Marzia and Naser,M. Niamul and Nair,Gopinath B. and Watanabe,Haruo and Ohnishi,Makoto and Yamamoto,Shouji and Endtz,Hubert and Cravioto,Alejandro and Sack,R. Bradley and Hasan,Nur A. and Sadique,Abdus and Huq,Anwar and Rita R Colwell and Alam,Munirul} } @article {19132, title = {Structure, function and diversity of the healthy human microbiome}, journal = {Nature}, volume = {486}, year = {2012}, month = {2012}, pages = {207 - 214}, abstract = {Studies of the human microbiome have revealed that even healthy individuals differ remarkably in the microbes that occupy habitats such as the gut, skin and vagina. Much of this diversity remains unexplained, although diet, environment, host genetics and early microbial exposure have all been implicated. Accordingly, to characterize the ecology of human-associated microbial communities, the Human Microbiome Project has analysed the largest cohort and set of distinct, clinically relevant body habitats so far. We found the diversity and abundance of each habitat{\textquoteright}s signature microbes to vary widely even among healthy subjects, with strong niche specialization both within and among individuals. The project encountered an estimated 81{\textendash}99\% of the genera, enzyme families and community configurations occupied by the healthy Western microbiome. Metagenomic carriage of metabolic pathways was stable among individuals despite variation in community structure, and ethnic/racial background proved to be one of the strongest associations of both pathways and microbes with clinical metadata. These results thus delineate the range of structural and functional configurations normal in the microbial communities of a healthy population, enabling future characterization of the epidemiology, ecology and translational applications of the human microbiome.}, author = {Huttenhower, C. and Gevers, D. and Knight,R. and Abubucker, S. and Badger, J.H. and Chinwalla, A.T. and Creasy, H.H. and Earl, A.M. and Fitzgerald, M.G. and Fulton, R.S. and others} } @article {19220, title = {Systems and Methods for Energy Harvesting in a Contained Fluid Circuit}, year = {2012}, month = {2012}, publisher = {UNIVERSITY OF WASHINGTON THROUGH ITS CENTER FOR COMMERCIALIZATION, CAMPBELL, Tim, LARSON, Eric, COHN, Gabriel, ALCAIDE, Ramses, FROEHLICH, Jon, PATEL, Shwetak}, abstract = {Systems and methods for harvesting energy from a closed fluid circuit, such as a water pipe system in a building. An energy harvester can be installed at a point in a water circuit and can generate energy from pressure differentials caused when a valve is opened or closed at any other point in the water circuit that is in fluid communication with the energy harvester. The energy can be used to power, for example, a sensor and/or a transmitter., L{\textquoteright}invention concerne des syst{\`e}mes et des proc{\'e}d{\'e}s de collecte de l{\textquoteright}{\'e}nergie qui provient d{\textquoteright}un circuit de fluide ferm{\'e}, comme un syst{\`e}me de conduites d{\textquoteright}eau dans un immeuble. Un collecteur d{\textquoteright}{\'e}nergie peut {\^e}tre install{\'e} {\`a} un certain point au sein d{\textquoteright}un circuit d{\textquoteright}eau et peut g{\'e}n{\'e}rer de l{\textquoteright}{\'e}nergie {\`a} partir des diff{\'e}rentiels de pression provoqu{\'e}s lorsqu{\textquoteright}une soupape est ouverte ou ferm{\'e}e {\`a} n{\textquoteright}importe quel autre point au sein du circuit d{\textquoteright}eau qui est en communication fluidique avec le collecteur d{\textquoteright}{\'e}nergie. L{\textquoteright}{\'e}nergie peut {\^e}tre utilis{\'e}e pour alimenter, par exemple, un capteur et/ou un {\'e}metteur.}, keywords = {Systems and Methods for Energy Harvesting in a Contained Fluid Circuit}, isbn = {WO/2012/021551}, url = {http://patentscope.wipo.int/search/en/WO2012021551}, author = {Campbell,Tim and Larson,Eric and Cohn, Gabriel and Alcaide,Ramses and Jon Froehlich and Patel,Shwetak} } @article {14592, title = {Accelerated evolution of 3{\textquoteright}avian FOXE1 genes, and thyroid and feather specific expression of chicken FoxE1}, journal = {BMC Evolutionary Biology}, volume = {11}, year = {2011}, month = {2011/10/15/}, pages = {302 - 302}, abstract = {The forkhead transcription factor gene E1 (FOXE1) plays an important role in regulation of thyroid development, palate formation and hair morphogenesis in mammals. However, avian FOXE1 genes have not been characterized and as such, codon evolution of FOXE1 orthologs in a broader evolutionary context of mammals and birds is not known.}, isbn = {1471-2148}, doi = {10.1186/1471-2148-11-302}, url = {http://www.biomedcentral.com/1471-2148/11/302}, author = {Yaklichkin,Sergey Yu and Darnell,Diana K and Pier,Maricela V and Antin,Parker B and Hannenhalli, Sridhar} } @article {19650, title = {Accurate proteome-wide protein quantification from high-resolution 15N mass spectra}, journal = {Genome Biology}, volume = {12}, year = {2011}, month = {2011/12/19/}, abstract = {In quantitative mass spectrometry-based proteomics, the metabolic incorporation of a single source of 15N-labeled nitrogen has many advantages over using stable isotope-labeled amino acids. However, the lack of a robust computational framework for analyzing the resulting spectra has impeded wide use of this approach. We have addressed this challenge by introducing a new computational methodology for analyzing 15N spectra in which quantification is integrated with identification. Application of this method to an Escherichia coli growth transition reveals significant improvement in quantification accuracy over previous methods.PMID: 22182234 }, isbn = {1465-6906}, url = {http://genomebiology.com/2011/12/12/R122/abstract}, author = {Zia Khan and Amini, Sasan and Bloom, Joshua S. and Ruse, Cristian and Caudy, Amy A. and Kruglyak, Leonid and Singh, Mona and Perlman, David H. and Tavazoie, Saeed} } @conference {14192, title = {Active scene recognition with vision and language}, booktitle = {2011 IEEE International Conference on Computer Vision (ICCV)}, year = {2011}, month = {2011/11/06/13}, pages = {810 - 817}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper presents a novel approach to utilizing high level knowledge for the problem of scene recognition in an active vision framework, which we call active scene recognition. In traditional approaches, high level knowledge is used in the post-processing to combine the outputs of the object detectors to achieve better classification performance. In contrast, the proposed approach employs high level knowledge actively by implementing an interaction between a reasoning module and a sensory module (Figure 1). Following this paradigm, we implemented an active scene recognizer and evaluated it with a dataset of 20 scenes and 100+ objects. We also extended it to the analysis of dynamic scenes for activity recognition with attributes. Experiments demonstrate the effectiveness of the active paradigm in introducing attention and additional constraints into the sensing process.}, keywords = {accuracy, active scene recognition, classification performance, Cognition, Computer vision, Detectors, Equations, high level knowledge utilization, HUMANS, image classification, inference mechanisms, object detectors, Object recognition, reasoning module, sensing process, sensory module, support vector machines, Training}, isbn = {978-1-4577-1101-5}, doi = {10.1109/ICCV.2011.6126320}, author = {Yu,Xiaodong and Ferm{\"u}ller, Cornelia and Ching Lik Teo and Yezhou Yang and Aloimonos, J.} } @conference {16815, title = {Adapting a map query interface for a gesturing touch screen interface}, booktitle = {Proceedings of the 20th international conference companion on World wide web}, series = {WWW {\textquoteright}11}, year = {2011}, month = {2011///}, pages = {257 - 260}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {NewsStand is an example application of a general framework that we are developing to enable searching for information using a map query interface, where the information results from monitoring the output of over 8,000 RSS news sources and is available for retrieval within minutes of publication. The user interface of NewsStand was recently adapted so that NewsStand can execute on mobile and tablet devices with a gesturing touch screen interface such as the iPhone, iPod Touch, and iPad. This action led to a discovery of some shortcomings of current mapping APIs as well as devising some interesting new widgets. These issues are discussed, and the realization can be seen by a demo at http://newsstand.umiacs.umd.edu on any of the above Apple devices as well as other devices that support gestures such as an Android phone.}, keywords = {map query interface, newsstand, touch screen gesturing interface}, isbn = {978-1-4503-0637-9}, doi = {10.1145/1963192.1963303}, url = {http://doi.acm.org/10.1145/1963192.1963303}, author = {Samet, Hanan and Teitler,Benjamin E. and Adelfio,Marco D. and Lieberman,Michael D.} } @article {15616, title = {Approximate polytope membership queries}, journal = {Proceedings of 43rd Annual ACM Symposium on Theory of Computing}, year = {2011}, month = {2011///}, pages = {579 - 586}, abstract = {We consider an approximate version of a fundamental geo-metric search problem, polytope membership queries. Given a convex polytope P in Rd, presented as the intersection of halfspaces, the objective is to preprocess P so that, given a query point q, it is possible to determine efficiently whether q lies inside P subject to an error bound ε. Previous solutions to this problem were based on straight- forward applications of classic polytope approximation tech- niques by Dudley (1974) and Bentley et al. (1982). The former yields minimum storage, and the latter yields con- stant query time. A space-time tradeoff can be obtained by interpolating between the two. We present the first sig- nificant improvements to this tradeoff. For example, using the same storage as Dudley, we reduce the query time from O(1/ε (d-1)/2 ) to O(1/ε (d-1)/4 ). Our approach is based on a very simple algorithm. Both lower bounds and upper bounds on the performance of the algorithm are presented. To establish the relevance of our results, we introduce a reduction from approximate nearest neighbor searching to approximate polytope membership queries. We show that our tradeoff provides significant improvements to the best known space-time tradeoffs for approximate nearest neigh- bor searching. Furthermore, this is achieved with construc- tions that are much simpler than existing methods }, author = {Arya,S. and da Fonseca,G. D and Mount, Dave} } @conference {17564, title = {Approximation algorithms for throughput maximization in wireless networks with delay constraints}, booktitle = {2011 Proceedings IEEE INFOCOM}, year = {2011}, month = {2011/04/10/15}, pages = {1116 - 1124}, publisher = {IEEE}, organization = {IEEE}, abstract = {We study the problem of throughput maximization in multi-hop wireless networks with end-to-end delay constraints for each session. This problem has received much attention starting with the work of Grossglauser and Tse (2002), and it has been shown that there is a significant tradeoff between the end-to-end delays and the total achievable rate. We develop algorithms to compute such tradeoffs with provable performance guarantees for arbitrary instances, with general interference models. Given a target delay-bound Δ(c) for each session c, our algorithm gives a stable flow vector with a total throughput within a factor of O (logΔm/log log Δm) of the maximum, so that the per-session (end-to-end) delay is O ((logΔm/log log Δm Δ(c))2), where Δm = maxc{Δ(c)}; note that these bounds depend only on the delays, and not on the network size, and this is the first such result, to our knowledge.}, keywords = {Approximation algorithms, Approximation methods, approximation theory, Delay, delay constraints, delays, general interference model, Interference, multihop wireless networks, optimisation, Optimized production technology, radio networks, radiofrequency interference, target delay bound, Throughput, throughput maximization, Wireless networks}, isbn = {978-1-4244-9919-9}, doi = {10.1109/INFCOM.2011.5934887}, author = {Guanhong Pei and Anil Kumar,V. S and Parthasarathy,S. and Srinivasan, Aravind} } @article {18547, title = {Architecting for innovation}, journal = {SIGCOMM Comput. Commun. Rev.}, volume = {41}, year = {2011}, month = {2011///}, pages = {24 - 36}, abstract = {We argue that the biggest problem with the current Internet architecture is not a particular functional deficiency, but its inability to accommodate innovation. To address this problem we propose a minimal architectural "framework" in which comprehensive architectures can reside. The proposed Framework for Internet Innovation (FII) --- which is derived from the simple observation that network interfaces should be extensible and abstract --- allows for a diversity of architectures to coexist, communicate, and evolve. We demonstrate FII{\textquoteright}s ability to accommodate diversity and evolution with a detailed examination of how information flows through the architecture and with a skeleton implementation of the relevant interfaces.}, keywords = {diversity, Evolution, innovation, internet architecture}, isbn = {0146-4833}, doi = {10.1145/2002250.2002256}, url = {http://doi.acm.org/10.1145/2002250.2002256}, author = {Koponen,Teemu and Shenker,Scott and Balakrishnan,Hari and Feamster, Nick and Ganichev,Igor and Ghodsi,Ali and Godfrey,P. Brighten and McKeown,Nick and Parulkar,Guru and Raghavan,Barath and Rexford,Jennifer and Arianfar,Somaya and Kuptsov,Dmitriy} } @conference {13069, title = {AVSS 2011 demo session: A large-scale benchmark dataset for event recognition in surveillance video}, booktitle = {Advanced Video and Signal-Based Surveillance (AVSS), 2011 8th IEEE International Conference on}, year = {2011}, month = {2011/09/30/2}, pages = {527 - 528}, abstract = {We introduce to the surveillance community the VIRAT Video Dataset[1], which is a new large-scale surveillance video dataset designed to assess the performance of event recognition algorithms in realistic scenes1.}, doi = {10.1109/AVSS.2011.6027400}, author = {Oh,Sangmin and Hoogs,Anthony and Perera,Amitha and Cuntoor,Naresh and Chen,Chia-Chih and Lee,Jong Taek and Mukherjee,Saurajit and Aggarwal, JK and Lee,Hyungtae and Davis, Larry S. and Swears,Eran and Wang,Xiaoyang and Ji,Qiang and Reddy,Kishore and Shah,Mubarak and Vondrick,Carl and Pirsiavash,Hamed and Ramanan,Deva and Yuen,Jenny and Torralba,Antonio and Song,Bi and Fong,Anesco and Roy-Chowdhury,Amit and Desai,Mita} } @article {16247, title = {Bacillus Anthracis Comparative Genome Analysis in Support of the Amerithrax Investigation}, journal = {Proceedings of the National Academy of SciencesPNAS}, volume = {108}, year = {2011}, month = {2011/03/22/}, pages = {5027 - 5032}, abstract = {Before the anthrax letter attacks of 2001, the developing field of microbial forensics relied on microbial genotyping schemes based on a small portion of a genome sequence. Amerithrax, the investigation into the anthrax letter attacks, applied high-resolution whole-genome sequencing and comparative genomics to identify key genetic features of the letters{\textquoteright} Bacillus anthracis Ames strain. During systematic microbiological analysis of the spore material from the letters, we identified a number of morphological variants based on phenotypic characteristics and the ability to sporulate. The genomes of these morphological variants were sequenced and compared with that of the B. anthracis Ames ancestor, the progenitor of all B. anthracis Ames strains. Through comparative genomics, we identified four distinct loci with verifiable genetic mutations. Three of the four mutations could be directly linked to sporulation pathways in B. anthracis and more specifically to the regulation of the phosphorylation state of Spo0F, a key regulatory protein in the initiation of the sporulation cascade, thus linking phenotype to genotype. None of these variant genotypes were identified in single-colony environmental B. anthracis Ames isolates associated with the investigation. These genotypes were identified only in B. anthracis morphotypes isolated from the letters, indicating that the variants were not prevalent in the environment, not even the environments associated with the investigation. This study demonstrates the forensic value of systematic microbiological analysis combined with whole-genome sequencing and comparative genomics.}, isbn = {0027-8424, 1091-6490}, doi = {10.1073/pnas.1016657108}, url = {http://www.pnas.org/content/108/12/5027}, author = {Rasko,David A and Worsham,Patricia L and Abshire,Terry G and Stanley,Scott T and Bannan,Jason D and Wilson,Mark R and Langham,Richard J and Decker,R. Scott and Jiang,Lingxia and Read,Timothy D. and Phillippy,Adam M and Salzberg,Steven L. and Pop, Mihai and Van Ert,Matthew N and Kenefic,Leo J and Keim,Paul S and Fraser-Liggett,Claire M and Ravel,Jacques} } @conference {12398, title = {Believe Me{\textemdash}We Can Do This! Annotating Persuasive Acts in Blog Text}, booktitle = {Workshops at the Twenty-Fifth AAAI Conference on Artificial Intelligence}, year = {2011}, month = {2011///}, author = {Anand,P. and King,J. and Jordan Boyd-Graber and Wagner,E. and Martell,C. and Oard, Douglas and Resnik, Philip} } @article {17571, title = {Capacity of wireless networks under SINR interference constraints}, journal = {Wireless Networks}, volume = {17}, year = {2011}, month = {2011///}, pages = {1605 - 1624}, abstract = {A fundamental problem in wireless networks is to estimate their throughput capacity{\textemdash}given a set of wireless nodes and a set of connections, what is the maximum rate at which data can be sent on these connections. Most of the research in this direction has focused either on random distributions of points, or has assumed simple graph-based models for wireless interference. In this paper, we study the capacity estimation problem using a realistic Signal to Interference Plus Noise Ratio (SINR) model for interference, on arbitrary wireless networks without any assumptions on node distributions. The problem becomes much more challenging for this setting, because of the non-locality of the SINR model. Recent work by Moscibroda et al. (IEEE INFOCOM 2006, ACM MobiHoc 2006) has shown that the throughput achieved by using SINR models can differ significantly from that obtained by using graph-based models. In this work, we develop polynomial time algorithms to provably approximate the throughput capacity of wireless network under the SINR model.}, isbn = {1022-0038}, url = {http://dx.doi.org/10.1007/s11276-011-0367-2}, author = {Chafekar,Deepti and Anil Kumar,V. and Marathe,Madhav and Parthasarathy,Srinivasan and Srinivasan, Aravind} } @article {12869, title = {Clonal transmission, dual peak, and off-season cholera in Bangladesh}, journal = {Infection Ecology \& Epidemiology}, volume = {1}, year = {2011}, month = {08/2011}, abstract = {Vibrio cholerae is an estuarine bacterium associated with a single peak of cholera (March{\textendash}May) in coastal villages of Bangladesh. For an unknown reason, however, cholera occurs in a unique dual peak (March{\textendash}May and September{\textendash}November) pattern in the city of Dhaka that is bordered by a heavily polluted freshwater river system and flood embankment. In August 2007, extreme flooding was accompanied by an unusually severe diarrhea outbreak in Dhaka that resulted in a record high illness. This study was aimed to understand the unusual outbreak and if it was related to the circulation of a new V. cholerae clone. Nineteen V. cholerae isolated during the peak of the 2007 outbreak were subjected to extensive phenotypic and molecular analyses, including multi-locus genetic screening by polymerase chain reaction (PCR), sequence-typing of the ctxB gene, and pulsed-field gel electrophoresis (PFGE). Factors associated with the unusual incidence of cholera were determined and analysis of the disease severity was done. Overall, microbiological and molecular data confirmed that the hypervirulent V. cholerae was O1 biotype El Tor (ET) that possessed cholera toxin (CT) of the classical biotype. The PFGE (NotI) and dendrogram clustering confirmed that the strains were clonal and related to the pre-2007 variant ET from Dhaka and Matlab and resembled one of two distinct clones of the variant ET confirmed to be present in the estuarine ecosystem of Bangladesh. Results of the analyses of both diarrheal case data for three consecutive years (2006{\textendash}2008) and regional hydroclimatology over three decades (1980{\textendash}2009) clearly indicate that the pattern of cholera occurring in Dhaka, and not seen at other endemic sites, was associated with flood waters transmitting the infectious clone circulating via the fecal-oral route during and between the dual seasonal cholera peaks in Dhaka. Circular river systems and flood embankment likely facilitate transmission of infectious V. cholerae throughout the year that leads to both sudden and off-season outbreaks in the densely populated urban ecosystem of Dhaka. Clonal recycling of hybrid El Tor with increasing virulence in a changing climate and in a region with a growing urban population represents a serious public health concern for Bangladesh.}, doi = {10.3402/iee.v1i0.7273}, author = {Alam,M. and Islam,A. and Bhuiyan,N. A. and Rahim,N. and Hossain,A. and Khan,G. Y. and Ahmed,D. and Watanabe,H. and Izumiya,H. and Faruque,A. S. G. and Rita R Colwell} } @conference {14183, title = {A Corpus-Guided Framework for Robotic Visual Perception}, booktitle = {Workshops at the Twenty-Fifth AAAI Conference on Artificial Intelligence}, year = {2011}, month = {2011///}, author = {Teo,C.L. and Yang, Y. and Daum{\'e}, Hal and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @book {15188, title = {Decision and Game Theory for Security: Second International Conference, GameSec 2011, College Park, MD, Maryland, USA, November 14-15, 2011, Proceedings}, year = {2011}, month = {2011/12/27/}, publisher = {Springer}, organization = {Springer}, abstract = {This book constitutes the refereed proceedings of the Second International Conference on Decision and Game Theory for Security, GameSec 2011, held in College Park, Maryland, USA, in November 2011. The 16 revised full papers and 2 plenary keynotes presented were carefully reviewed and selected from numerous submissions. The papers are organized in topical sections on attacks, adversaries, and game theory, wireless adhoc and sensor networks, network games, security insurance, security and trust in social networks and security investments.}, keywords = {Business \& Economics / Information Management, Computers / Hardware / Network Hardware, Computers / Information Technology, Computers / Programming / Algorithms, Computers / Security / General, Mathematics / Applied, Mathematics / Game Theory, Mathematics / General}, isbn = {9783642252792}, author = {Baras,John S and Katz, Jonathan and Altman,Eitan} } @article {18786, title = {Design of Revolute Joints for In-Mold Assembly Using Insert Molding}, journal = {Journal of Mechanical Design}, volume = {133}, year = {2011}, month = {2011///}, pages = {121010 - 121010}, author = {Ananthanarayanan,A. and Ehrlich,L. and Desai,J. P. and Gupta,S.K.} } @article {19656, title = {Direct targeting of Sec23a by miR-200s influences cancer cell secretome and promotes metastatic colonization}, journal = {Nature Medicine}, volume = {17}, year = {2011}, month = {2011/09//}, pages = {1101 - 1108}, abstract = {Although the role of miR-200s in regulating E-cadherin expression and epithelial-to-mesenchymal transition is well established, their influence on metastatic colonization remains controversial. Here we have used clinical and experimental models of breast cancer metastasis to discover a pro-metastatic role of miR-200s that goes beyond their regulation of E-cadherin and epithelial phenotype. Overexpression of miR-200s is associated with increased risk of metastasis in breast cancer and promotes metastatic colonization in mouse models, phenotypes that cannot be recapitulated by E-cadherin expression alone. Genomic and proteomic analyses revealed global shifts in gene expression upon miR-200 overexpression toward that of highly metastatic cells. miR-200s promote metastatic colonization partly through direct targeting of Sec23a, which mediates secretion of metastasis-suppressive proteins, including Igfbp4 and Tinagl1, as validated by functional and clinical correlation studies. Overall, these findings suggest a pleiotropic role of miR-200s in promoting metastatic colonization by influencing E-cadherin{\textendash}dependent epithelial traits and Sec23a-mediated tumor cell secretome.View full text }, isbn = {1078-8956}, url = {http://www.nature.com/nm/journal/v17/n9/abs/nm.2401.html}, author = {Korpal, Manav and Ell, Brian J. and Buffa, Francesca M. and Ibrahim, Toni and Blanco, Mario A. and Celi{\`a}-Terrassa, Toni and Mercatali, Laura and Zia Khan and Goodarzi, Hani and Hua, Yuling and Wei, Yong and Hu, Guohong and Garcia, Benjamin A. and Ragoussis, Jiannis and Amadori, Dino and Harris, Adrian L. and Kang, Yibin} } @inbook {12455, title = {Distributed Sensing and Processing for Multi-Camera Networks}, booktitle = {Distributed Video Sensor NetworksDistributed Video Sensor Networks}, year = {2011}, month = {2011///}, pages = {85 - 101}, publisher = {Springer London}, organization = {Springer London}, abstract = {Sensor networks with large numbers of cameras are becoming increasingly prevalent in a wide range of applications, including video conferencing, motion capture, surveillance, and clinical diagnostics. In this chapter, we identify some of the fundamental challenges in designing such systems: robust statistical inference, computationally efficiency, and opportunistic and parsimonious sensing. We show that the geometric constraints induced by the imaging process are extremely useful for identifying and designing optimal estimators for object detection and tracking tasks. We also derive pipelined and parallelized implementations of popular tools used for statistical inference in non-linear systems, of which multi-camera systems are examples. Finally, we highlight the use of the emerging theory of compressive sensing in reducing the amount of data sensed and communicated by a camera network.}, isbn = {978-0-85729-127-1}, url = {http://dx.doi.org/10.1007/978-0-85729-127-1_6}, author = {Sankaranarayanan,Aswin C. and Chellapa, Rama and Baraniuk,Richard G.}, editor = {Bhanu,Bir and Ravishankar,Chinya V. and Roy-Chowdhury,Amit K. and Aghajan,Hamid and Terzopoulos,Demetri} } @conference {19024, title = {Do You Know Where Your Data Are? Secure Data Capsules for Deployable Data Protection}, year = {2011}, month = {2011}, url = {http://static.usenix.org/event/hotos11/tech/final_files/ManiatisAkhawe.pdf}, author = {Maniatis, P. and Akhawe, D. and Fall,K. and Elaine Shi and McCamant, S. and Song,D.} } @conference {14315, title = {Dynamic inference of static types for ruby}, booktitle = {Proceedings of the 38th annual ACM SIGPLAN-SIGACT symposium on Principles of programming languages}, series = {POPL {\textquoteright}11}, year = {2011}, month = {2011///}, pages = {459 - 472}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {There have been several efforts to bring static type inference to object-oriented dynamic languages such as Ruby, Python, and Perl. In our experience, however, such type inference systems are extremely difficult to develop, because dynamic languages are typically complex, poorly specified, and include features, such as eval and reflection, that are hard to analyze. In this paper, we introduce constraint-based dynamic type inference, a technique that infers static types based on dynamic program executions. In our approach, we wrap each run-time value to associate it with a type variable, and the wrapper generates constraints on this type variable when the wrapped value is used. This technique avoids many of the often overly conservative approximations of static tools, as constraints are generated based on how values are used during actual program runs. Using wrappers is also easy to implement, since we need only write a constraint resolution algorithm and a transformation to introduce the wrappers. The best part is that we can eat our cake, too: our algorithm will infer sound types as long as it observes every path through each method body---note that the number of such paths may be dramatically smaller than the number of paths through the program as a whole. We have developed Rubydust, an implementation of our algorithm for Ruby. Rubydust takes advantage of Ruby{\textquoteright}s dynamic features to implement wrappers as a language library. We applied Rubydust to a number of small programs and found it to be both easy to use and useful: Rubydust discovered 1 real type error, and all other inferred types were correct and readable.}, keywords = {dynamic languages, dynamic type inference, ruby, static types}, isbn = {978-1-4503-0490-0}, doi = {10.1145/1926385.1926437}, url = {http://doi.acm.org/10.1145/1926385.1926437}, author = {An,Jong-hoon (David) and Chaudhuri,Avik and Foster, Jeffrey S. and Hicks, Michael W.} } @article {18526, title = {The Evolution of Network Configuration: A Tale of Two Campuses}, journal = {Networks}, volume = {7}, year = {2011}, month = {2011///}, pages = {18 - 18}, abstract = {Studying network configuration evolution can improve our under- standing of the evolving complexity of networks and can be helpful in making network configuration less error-prone. Unfortunately, the nature of changes that operators make to network configuration is poorly understood. Towards improving our understanding, we examine and analyze five years of router, switch, and firewall con- figurations from two large campus networks using the logs from version control systems used to store the configurations. We study how network configuration is distributed across different network operations tasks and how the configuration for each task evolves over time, for different types of devices and for different loca- tions in the network. To understand the trends of how configura- tion evolves over time, we study the extent to which configuration for various tasks are added, modified, or deleted. We also study whether certain devices experience configuration changes more fre- quently than others, as well as whether configuration changes tend to focus on specific portions of the configuration (or on specific tasks). We also investigate when network operators make configu- ration changes of various types. Our results concerning configura- tion changes can help the designers of configuration languages un- derstand which aspects of configuration might be more automated or tested more rigorously and may ultimately help improve config- uration languages.}, author = {Kim,H. and Benson,T. and Akella,A. and Feamster, Nick} } @book {15187, title = {GameSec{\textquoteright}11: Proceedings of the Second international conference on Decision and Game Theory for Security}, year = {2011}, month = {2011///}, publisher = {Springer-Verlag}, organization = {Springer-Verlag}, address = {Berlin, Heidelberg}, isbn = {978-3-642-25279-2}, editor = {Baras,John S and Katz, Jonathan and Altman,Eitan} } @article {14613, title = {Gene Coexpression Network Topology of Cardiac Development, Hypertrophy, and FailureClinical Perspective}, journal = {Circulation: Cardiovascular GeneticsCirc Cardiovasc Genet}, volume = {4}, year = {2011}, month = {2011/02/01/}, pages = {26 - 35}, abstract = {Background{\textemdash} Network analysis techniques allow a more accurate reflection of underlying systems biology to be realized than traditional unidimensional molecular biology approaches. Using gene coexpression network analysis, we define the gene expression network topology of cardiac hypertrophy and failure and the extent of recapitulation of fetal gene expression programs in failing and hypertrophied adult myocardium.Methods and Results{\textemdash} We assembled all myocardial transcript data in the Gene Expression Omnibus (n=1617). Because hierarchical analysis revealed species had primacy over disease clustering, we focused this analysis on the most complete (murine) dataset (n=478). Using gene coexpression network analysis, we derived functional modules, regulatory mediators, and higher-order topological relationships between genes and identified 50 gene coexpression modules in developing myocardium that were not present in normal adult tissue. We found that known gene expression markers of myocardial adaptation were members of upregulated modules but not hub genes. We identified ZIC2 as a novel transcription factor associated with coexpression modules common to developing and failing myocardium. Of 50 fetal gene coexpression modules, 3 (6\%) were reproduced in hypertrophied myocardium and 7 (14\%) were reproduced in failing myocardium. One fetal module was common to both failing and hypertrophied myocardium. Conclusions{\textemdash} Network modeling allows systems analysis of cardiovascular development and disease. Although we did not find evidence for a global coordinated program of fetal gene expression in adult myocardial adaptation, our analysis revealed specific gene expression modules active during both development and disease and specific candidates for their regulation. }, keywords = {fetal, Gene expression, heart failure, hypertrophy, myocardium}, isbn = {1942-325X, 1942-3268}, doi = {10.1161/CIRCGENETICS.110.941757}, url = {http://circgenetics.ahajournals.org/content/4/1/26}, author = {Dewey,Frederick E and Perez,Marco V and Wheeler,Matthew T and Watt,Clifton and Spin,Joshua and Langfelder,Peter and Horvath,Steve and Hannenhalli, Sridhar and Cappola,Thomas P. and Ashley,Euan A} } @conference {18677, title = {How secure are networked office devices?}, year = {2011}, month = {2011/06//}, pages = {465 - 472}, abstract = {Many office devices have a history of being networked (such as printers) and others without the same past are increasingly becoming networked (such as photocopiers). The modern networked versions of previously non-networked devices have much in common with traditional networked servers in terms of features and functions. While an organization may have policies and procedures for securing traditional network servers, securing networked office devices providing similar services can easily be overlooked. In this paper we present an evaluation of privacy and security risks found when examining over 1,800 networked office devices connected to a large university network. We use the STRIDE threat model to categorize threats and vulnerabilities and then we group the devices according to assessed risk from the perspective of the university. We found that while steps had been taken to secure some devices, many were using default or unsecured configurations.}, keywords = {computer network security, data privacy, networked office device security, privacy risk assessment, Risk management, security risk assessment, STRIDE threat model, university network}, doi = {10.1109/DSN.2011.5958259}, author = {Condon,E. and Cummins,E. and Afoulki,Z. and Michel Cukier} } @article {12866, title = {The Importance of Chitin in the Marine Environment}, journal = {Marine Biotechnology}, year = {2011}, month = {2011///}, pages = {1 - 8}, abstract = {Chitin is the most abundant renewable polymer in the oceans and is an important source of carbon and nitrogen for marine organisms. The process of chitin degradation is a key step in the cycling of nutrients in the oceans and chitinolytic bacteria play a significant role in this process. These bacteria are autochthonous to both marine and freshwater ecosystems and produce chitinases that degrade chitin, an insoluble polysaccharide, to a biologically useful form. In this brief review, a description of the structure of chitin and diversity of chitinolytic bacteria in the oceans is provided, in the context of the significance of chitin degradation for marine life.}, doi = {10.1007/s10126-011-9388-1}, author = {Souza,C. P. and Almeida,B. C. and Rita R Colwell and Rivera,I. N. G.} } @conference {13074, title = {A large-scale benchmark dataset for event recognition in surveillance video}, booktitle = {Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on}, year = {2011}, month = {2011/06//}, pages = {3153 - 3160}, abstract = {We introduce a new large-scale video dataset designed to assess the performance of diverse visual event recognition algorithms with a focus on continuous visual event recognition (CVER) in outdoor areas with wide coverage. Previous datasets for action recognition are unrealistic for real-world surveillance because they consist of short clips showing one action by one individual [15, 8]. Datasets have been developed for movies [11] and sports [12], but, these actions and scene conditions do not apply effectively to surveillance videos. Our dataset consists of many outdoor scenes with actions occurring naturally by non-actors in continuously captured videos of the real world. The dataset includes large numbers of instances for 23 event types distributed throughout 29 hours of video. This data is accompanied by detailed annotations which include both moving object tracks and event examples, which will provide solid basis for large-scale evaluation. Additionally, we propose different types of evaluation modes for visual recognition tasks and evaluation metrics along with our preliminary experimental results. We believe that this dataset will stimulate diverse aspects of computer vision research and help us to advance the CVER tasks in the years ahead.}, keywords = {algorithm;evaluation, CVER, databases;, databases;video, dataset;moving, event, metrics;large-scale, object, recognition, recognition;diverse, recognition;video, scenes;surveillance, surveillance;visual, tasks;computer, tracks;outdoor, video, video;computer, vision;continuous, vision;image, visual}, doi = {10.1109/CVPR.2011.5995586}, author = {Oh,Sangmin and Hoogs, A. and Perera,A. and Cuntoor, N. and Chen,Chia-Chih and Lee,Jong Taek and Mukherjee,S. and Aggarwal, JK and Lee,Hyungtae and Davis, Larry S. and Swears,E. and Wang,Xioyang and Ji,Qiang and Reddy,K. and Shah,M. and Vondrick,C. and Pirsiavash,H. and Ramanan,D. and Yuen,J. and Torralba,A. and Song,Bi and Fong,A. and Roy-Chowdhury, A. and Desai,M.} } @inbook {19186, title = {A Longitudinal Study of Pressure Sensing to Infer Real-World Water Usage Events in the Home}, booktitle = {Pervasive Computing}, series = {Lecture Notes in Computer Science}, volume = {6696}, year = {2011}, month = {2011}, pages = {50 - 69}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We present the first longitudinal study of pressure sensing to infer real-world water usage events in the home (e.g., dishwasher, upstairs bathroom sink, downstairs toilet). In order to study the pressure-based approach out in the wild , we deployed a ground truth sensor network for five weeks in three homes and two apartments that directly monitored valve-level water usage by fixtures and appliances . We use this data to, first, demonstrate the practical challenges in constructing water usage activity inference algorithms and, second, to inform the design of a new probabilistic-based classification approach. Inspired by algorithms in speech recognition, our novel Bayesian approach incorporates template matching, a language model, grammar, and prior probabilities. We show that with a single pressure sensor, our probabilistic algorithm can classify real-world water usage at the fixture level with 90\% accuracy and at the fixturecategory level with 96\% accuracy. With two pressure sensors, these accuracies increase to 94\% and 98\%. Finally, we show how our new approach can be trained with fewer examples than a strict template-matching approach alone.}, isbn = {978-3-642-21725-8}, url = {http://dx.doi.org/10.1007/978-3-642-21726-5_4}, author = {Jon Froehlich and Larson,Eric and Saba,Elliot and Campbell,Tim and Atlas,Les and Fogarty,James and Patel,Shwetak}, editor = {Lyons,Kent and Hightower,Jeffrey and Huang,Elaine} } @inbook {14324, title = {A Longitudinal Study of Pressure Sensing to Infer Real-World Water Usage Events in the Home}, booktitle = {Pervasive ComputingPervasive Computing}, series = {Lecture Notes in Computer Science}, volume = {6696}, year = {2011}, month = {2011///}, pages = {50 - 69}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We present the first longitudinal study of pressure sensing to infer real-world water usage events in the home (e.g., dishwasher, upstairs bathroom sink, downstairs toilet). In order to study the pressure-based approach out in the wild , we deployed a ground truth sensor network for five weeks in three homes and two apartments that directly monitored valve-level water usage by fixtures and appliances . We use this data to, first, demonstrate the practical challenges in constructing water usage activity inference algorithms and, second, to inform the design of a new probabilistic-based classification approach. Inspired by algorithms in speech recognition, our novel Bayesian approach incorporates template matching, a language model, grammar, and prior probabilities. We show that with a single pressure sensor, our probabilistic algorithm can classify real-world water usage at the fixture level with 90\% accuracy and at the fixturecategory level with 96\% accuracy. With two pressure sensors, these accuracies increase to 94\% and 98\%. Finally, we show how our new approach can be trained with fewer examples than a strict template-matching approach alone.}, isbn = {978-3-642-21725-8}, url = {http://dx.doi.org/10.1007/978-3-642-21726-5_4}, author = {Jon Froehlich and Larson,Eric and Saba,Elliot and Campbell,Tim and Atlas,Les and Fogarty,James and Patel,Shwetak}, editor = {Lyons,Kent and Hightower,Jeffrey and Huang,Elaine} } @article {17622, title = {Maximum bipartite flow in networks with adaptive channel width}, journal = {Theoretical Computer Science}, volume = {412}, year = {2011}, month = {2011/05/27/}, pages = {2577 - 2587}, abstract = {Traditionally, network optimization problems assume that each link in the network has a fixed capacity. Recent research in wireless networking has shown that it is possible to design networks where the capacity of the links can be changed adaptively to suit the needs of specific applications. In particular, one gets a choice of having a few high capacity outgoing links or many low capacity ones at any node of the network. This motivates us to have a re-look at classical network optimization problems and design algorithms to solve them in this new framework. In particular, we consider the problem of maximum bipartite flow, which has been studied extensively in the fixed-capacity network model. One of the motivations for studying this problem arises from the need to maximize the throughput of an infrastructure wireless network comprising base-stations (one set of vertices in the bipartition) and clients (the other set of vertices in the bipartition). We show that this problem has a significantly different combinatorial structure in this new network model from the fixed-capacity one. While there are several polynomial time algorithms for the maximum bipartite flow problem in traditional networks, we show that the problem is NP-hard in the new model. In fact, our proof extends to showing that the problem is APX-hard. We complement our lower bound by giving two algorithms for solving the problem approximately. The first algorithm is deterministic and achieves an approximation factor of O ( log N ) , where N is the number of nodes in the network, while the second algorithm is randomized and achieves an approximation factor of e e - 1 .}, keywords = {Adaptive channel width, graph algorithm, Linear program rounding, Maximum flow, Wireless networks}, isbn = {0304-3975}, doi = {10.1016/j.tcs.2010.10.023}, url = {http://www.sciencedirect.com/science/article/pii/S0304397510005852}, author = {Azar,Yossi and M{\k a}dry,Aleksander and Moscibroda,Thomas and Panigrahi,Debmalya and Srinivasan, Aravind} } @article {12859, title = {Metagenomic 16S rDNA Targeted PCR-DGGE in Determining Bacterial Diversity in Aquatic Ecosystem}, journal = {Bangladesh Journal of Microbiology}, volume = {27}, year = {2011}, month = {2011/12/13/}, abstract = {Bacterial numbers in surface water samples, collected randomly from six different water bodies, were estimated by acridine orange direct counting (AODC) and conventional culture-based heterotrophic plate counting (HPC). Bacterial genomic DNA was prepared from water samples by employing methods used for stool samples, including the population dynamics, were determined by primer extension of the 16S rDNA (V6/V8 region) using polymerase chain reaction (PCR), followed by denaturing gradient gel electrophoresis (DGGE), a metagenomic tool that is capable of separating unrelated DNAs based on the differences in their sequences and GC contents. The bacterial numbers in water samples ranged from 103 {\textendash} 106 CFU/ mL for HPC and 104 {\textendash} 107 cells/ mL for AODC, showing that a great majority of bacteria prevail as uncultivable which do not respond to culture methods that are used widely for tracking bacterial pathogens. The acridine orange-stained bacteria varied in sizes and shapes, and appeared either as planktonic (solitary) cells or as clusters of biofilms, showing the presence of diverse community under the epifluorescence microscope. The DGGE of the ca. 457 bp amplicons, as confirmed by agarose gel electrophoresis, produced bands that ranged in intensities and numbers from 18 to 31, with each band possibly indicating the presence of one or more closely related bacterial species. The enrichment of pathogenic bacteria in the aquatic ecosystem is known to precede the seasonal diarrhoeal outbreaks; therefore, bacterial community dynamics determined by Metagenomic 16S PCR-DGGE during pre-epidemic enrichment appears promising in predicting the upcoming diarrheal outbreaks.}, isbn = {1011-9981}, doi = {10.3329/bjm.v27i2.9171}, url = {http://www.banglajol.info/bd/index.php/BJM/article/view/9171}, author = {Hasan,Nur A. and Chowdhury,W Bari and Rahim,Niaz and Sultana,Marzia and Shabnam,S Antara and Mai,Volker and Ali,Afsar and Morris,Glen J and Sack,R. Bradley and Huq,Anwar and Rita R Colwell and Endtz,Hubert Ph and Cravioto,Alejandro and Alam,Munirul} } @article {19111, title = {Next Generation Sequence Assembly with AMOS}, journal = {Current Protocols in Bioinformatics}, volume = {11}, year = {2011}, month = {2011}, pages = {1 - 11}, author = {Treangen, T.J. and Sommer, D.D. and Angly, F.E. and Koren, S. and Pop, Mihai} } @article {20383, title = {Occurrence of Vibrio cholerae in Municipal and Natural Waters and Incidence of Cholera in Azerbaijan}, journal = {EcoHealth}, year = {2011}, month = {Jan-12-2011}, pages = {468 - 477}, abstract = {Cholera, a waterborne disease caused by Vibrio cholerae, is an autochthonous member of the aquatic environment and predominantly reported from developing countries. Technical reports and proceedings were reviewed to determine the relationship between occurrence of V. cholerae in natural waters, including sources of municipal water, and cases of cholera in Azerbaijan. Water samples collected from different environmental sources from 1970 to 1998 were tested for V. cholerae and 0.73\% (864/117,893) were positive. The results showed that in April of each year, when the air temperature rose by approximately 5{\textdegree}C, V. cholerae could be isolated. With each increase in air temperature, 6{\textendash}8 weeks after, impact on cases of cholera was recorded. The incidence of cholera peaked when the air temperature reached >25{\textdegree}C during the month of September. It is concluded that a distinct seasonality in cholera incidence exists in Azerbaijan, with increased occurrence during warmer months.}, issn = {1612-9202}, doi = {10.1007/s10393-012-0756-8}, url = {http://link.springer.com/10.1007/s10393-012-0756-8}, author = {Gurbanov, Shair and Akhmadov, Rashid and Shamkhalova, Gulnara and Akhmadova, Sevinj and Haley, Bradd J. and Rita R Colwell and Huq, Anwar} } @article {12861, title = {Role of Zooplankton Diversity in Vibrio Cholerae Population Dynamics and in the Incidence of Cholera in the Bangladesh Sundarbans}, journal = {Applied and Environmental Microbiology}, volume = {77}, year = {2011}, month = {09/2011}, pages = {6125 - 6132}, abstract = {Vibrio cholerae, a bacterium autochthonous to the aquatic environment, is the causative agent of cholera, a severe watery, life-threatening diarrheal disease occurring predominantly in developing countries. V. cholerae, including both serogroups O1 and O139, is found in association with crustacean zooplankton, mainly copepods, and notably in ponds, rivers, and estuarine systems globally. The incidence of cholera and occurrence of pathogenic V. cholerae strains with zooplankton were studied in two areas of Bangladesh: Bakerganj and Mathbaria. Chitinous zooplankton communities of several bodies of water were analyzed in order to understand the interaction of the zooplankton population composition with the population dynamics of pathogenic V. cholerae and incidence of cholera. Two dominant zooplankton groups were found to be consistently associated with detection of V. cholerae and/or occurrence of cholera cases, namely, rotifers and cladocerans, in addition to copepods. Local differences indicate there are subtle ecological factors that can influence interactions between V. cholerae, its plankton hosts, and the incidence of cholera.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.01472-10}, url = {http://aem.asm.org/content/77/17/6125}, author = {De Magny,Guillaume Constantin and Mozumder,Pronob K. and Grim,Christopher J. and Hasan,Nur A. and Naser,M. Niamul and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @conference {13642, title = {Segmentation of Handwritten Textlines in Presence of Touching Components}, booktitle = {Intl. Conf. on Document Analysis and Recognition (ICDAR 11)}, year = {2011}, month = {2011///}, pages = {109 - 113}, abstract = {This paper presents an approach to textline extraction in handwritten document images which combines local and global techniques. We propose a graph-based technique to detect touching and proximity errors that are common with handwritten text lines. In a refinement step, we use Expectation-Maximization (EM) to iteratively split the error segments to obtain correct text-lines. We show improvement in accuracies using our correction method on datasets of Arabic document images. Results on a set of artificially generated proximity images show that the method is effective for handling touching errors in handwritten document images.}, author = {Kumar,Jayant and Kang,Le and David Doermann and Abd-Almageed, Wael} } @conference {13643, title = {Shape Codebook based Handwritten and Machine Printed Text Zone Extraction}, booktitle = {Document Recognition and Retrieval}, year = {2011}, month = {2011/01//}, pages = {7874:1-8 - 7874:1-8}, address = {San Francisco}, abstract = {We present a novel method for extracting handwritten and printed text zones from noisy document images with mixed content. We use Triple-Adjacent-Segment (TAS) based features which encode local shape characteristics of text in a consistent manner. We first construct two different codebooks of the shape features extracted from a set of handwritten and printed text documents. In the next step, we compute the normalized histogram of codewords for each segmented zone and use it to train Support Vector Machine (SVM) classifier. Due to a codebook based approach, our method is robust to the background noise present in the image. The TAS features used are invariant to translation, scale and rotation of text. In our experimental results, we show that a pixel-weighted zone classification accuracy of 98\% can be achieved for noisy Arabic documents. Further, we demonstrate the effectiveness of our method for document page classification and show that a high precision can be achieved for machine printed documents. The proposed method is robust to the size of zones, which may contain text content at word, line or paragraph level.}, author = {Kumar,Jayant and Prasad,Rohit and Cao,Huiagu and Abd-Almageed, Wael and David Doermann and Natarajan,Prem} } @article {11906, title = {Silhouette-based gesture and action recognition via modeling trajectories on Riemannian shape manifolds}, journal = {Computer Vision and Image Understanding}, volume = {115}, year = {2011}, month = {2011/03//}, pages = {439 - 455}, abstract = {This paper addresses the problem of recognizing human gestures from videos using models that are built from the Riemannian geometry of shape spaces. We represent a human gesture as a temporal sequence of human poses, each characterized by a contour of the associated human silhouette. The shape of a contour is viewed as a point on the shape space of closed curves and, hence, each gesture is characterized and modeled as a trajectory on this shape space. We propose two approaches for modeling these trajectories. In the first template-based approach, we use dynamic time warping (DTW) to align the different trajectories using elastic geodesic distances on the shape space. The gesture templates are then calculated by averaging the aligned trajectories. In the second approach, we use a graphical model approach similar to an exemplar-based hidden Markov model, where we cluster the gesture shapes on the shape space, and build non-parametric statistical models to capture the variations within each cluster. We model each gesture as a Markov model of transitions between these clusters. To evaluate the proposed approaches, an extensive set of experiments was performed using two different data sets representing gesture and action recognition applications. The proposed approaches not only are successfully able to represent the shape and dynamics of the different classes for recognition, but are also robust against some errors resulting from segmentation and background subtraction.}, keywords = {Action recognition, Gesture recognition, Riemannian manifolds, Shape space, Silhouette-based approaches}, isbn = {1077-3142}, doi = {16/j.cviu.2010.10.006}, url = {http://www.sciencedirect.com/science/article/pii/S1077314210002377}, author = {Abdelkader,Mohamed F. and Abd-Almageed, Wael and Srivastava,Anuj and Chellapa, Rama} } @conference {13651, title = {Stroke-like Pattern Noise Removal in Binary Document Images}, booktitle = {International Conference on Document Analysis and Recognition}, year = {2011}, month = {2011///}, pages = {17 - 21}, abstract = {This paper presents a two-phased stroke-like pattern noise (SPN) removal algorithm for binary document images. The proposed approach aims at understanding script-independent prominent text component features using supervised classification as a first step. It then uses their cohesiveness and stroke-width properties to filter and associate smaller text components with them using an unsupervised classification technique. In order to perform text extraction, and hence noise removal, at diacritic-level, this divide-and-conquer technique does not assume the availability of accurate and large amounts of ground-truth data at component-level for training purposes. The method was tested on a collection of degraded and noisy, machine-printed and handwritten binary Arabic text documents. Results show pixel-level precision and recall of 98\% and 97\% respectively.}, author = {Agrawal,Mudit and David Doermann} } @article {16045, title = {A task taxonomy of network evolution analysis}, journal = {University of Maryland, Human-Computer Interaction Lab Tech Report HCIL-2011-09}, year = {2011}, month = {2011///}, abstract = {Visualization is a useful tool for understanding the nature of networks. The recent growth of social media requires morepowerful visualization techniques beyond static network diagrams. One of the most important challenges is the visualization of temporal network evolution. In order to provide strong temporal visualization methods, we need to understand what tasks users accomplish. This study provides a taxonomy of the temporal network visualization tasks. We identify (1) the entities, (2) the properties to be visualized, and (3) the hierarchy of temporal features, which were extracted by surveying existing temporal network visualization systems. By building and examining the task taxonomy, we report which tasks have been covered so far and suggest additions for designing the future visualizations. We also present example visualizations constructed using the task taxonomy for a social networking site in order to validate the quality of the taxonomy. }, author = {Ahn,JW and Plaisant, Catherine and Shneiderman, Ben} } @inbook {16046, title = {Temporal Visualization of Social Network Dynamics: Prototypes for Nation of Neighbors}, booktitle = {Social Computing, Behavioral-Cultural Modeling and PredictionSocial Computing, Behavioral-Cultural Modeling and Prediction}, series = {Lecture Notes in Computer Science}, volume = {6589}, year = {2011}, month = {2011///}, pages = {309 - 316}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Information visualization is a powerful tool for analyzing the dynamic nature of social communities. Using Nation of Neighbors community network as a testbed, we propose five principles of implementing temporal visualizations for social networks and present two research prototypes: NodeXL and TempoVis. Three different states are defined in order to visualize the temporal changes of social networks. We designed the prototypes to show the benefits of the proposed ideas by letting users interactively explore temporal changes of social networks.}, isbn = {978-3-642-19655-3}, url = {http://dx.doi.org/10.1007/978-3-642-19656-0_43}, author = {Ahn,Jae-wook and Taieb-Maimon,Meirav and Sopan,Awalin and Plaisant, Catherine and Shneiderman, Ben}, editor = {Salerno,John and Yang,Shanchieh and Nau, Dana S. and Chai,Sun-Ki} } @article {11903, title = {Threat-Aware Clustering in Wireless Sensor Networks}, journal = {IFIP Advances in Information and Communication Technology (AICT)}, volume = {264}, year = {2011}, month = {2011///}, pages = {1 - 12}, author = {Abd-Almageed, Wael and Blace,R.E. and Eltoweissy,M.} } @article {12417, title = {Using Variational Inference and MapReduce to Scale Topic Modeling}, journal = {arXiv:1107.3765}, year = {2011}, month = {2011/07/19/}, abstract = {Latent Dirichlet Allocation (LDA) is a popular topic modeling technique for exploring document collections. Because of the increasing prevalence of large datasets, there is a need to improve the scalability of inference of LDA. In this paper, we propose a technique called ~\emph{MapReduce LDA} (Mr. LDA) to accommodate very large corpus collections in the MapReduce framework. In contrast to other techniques to scale inference for LDA, which use Gibbs sampling, we use variational inference. Our solution efficiently distributes computation and is relatively simple to implement. More importantly, this variational implementation, unlike highly tuned and specialized implementations, is easily extensible. We demonstrate two extensions of the model possible with this scalable framework: informed priors to guide topic discovery and modeling topics from a multilingual corpus.}, keywords = {Computer Science - Artificial Intelligence, Computer Science - Distributed, Parallel, and Cluster Computing}, url = {http://arxiv.org/abs/1107.3765}, author = {Zhai,Ke and Jordan Boyd-Graber and Asadi,Nima} } @conference {17484, title = {Visual Analysis of Temporal Trends in Social Networks Using Edge Color Coding and Metric Timelines}, booktitle = {Privacy, Security, Risk and Trust (PASSAT), 2011 IEEE Third International Conference on and 2011 IEEE Third International Confernece on Social Computing (SocialCom)}, year = {2011}, month = {2011/10/09/11}, pages = {549 - 554}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present Net EvViz, a visualization tool for analysis and exploration of a dynamic social network. There are plenty of visual social network analysis tools but few provide features for visualization of dynamically changing networks featuring the addition or deletion of nodes or edges. Our tool extends the code base of the Node XL template for Microsoft Excel, a popular network visualization tool. The key features of this work are (1) The ability of the user to specify and edit temporal annotations to the network components in an Excel sheet, (2) See the dynamics of the network with multiple graph metrics plotted over the time span of the graph, called the Timeline, and (3) Temporal exploration of the network layout using an edge coloring scheme and a dynamic Time slider. The objectives of the new features presented in this paper are to let the data analysts, computer scientists and others to observe the dynamics or evolution in a network interactively. We presented Net EvViz to five users of Node XL and received positive responses.}, keywords = {Color, computer scientists, data analysts, data visualisation, Data visualization, dynamic social network, dynamic timeslider, edge color coding, excel sheet, Image coding, image colour analysis, Layout, measurement, metric timelines, Microsoft excel, multiple graph metrics, Net EvViz, network components, network layout, network visualization tool, NodeXL template, social networking (online), temporal trends, Twitter, Visualization}, isbn = {978-1-4577-1931-8}, doi = {10.1109/PASSAT/SocialCom.2011.212}, author = {Khurana,U. and Nguyen,Viet-An and Hsueh-Chien Cheng and Ahn,Jae-wook and Chen,Xi and Shneiderman, Ben} } @article {17487, title = {Visual Exploration across Biomedical Databases}, journal = {IEEE/ACM Transactions on Computational Biology and Bioinformatics}, volume = {8}, year = {2011}, month = {2011/04//March}, pages = {536 - 550}, abstract = {Though biomedical research often draws on knowledge from a wide variety of fields, few visualization methods for biomedical data incorporate meaningful cross-database exploration. A new approach is offered for visualizing and exploring a query-based subset of multiple heterogeneous biomedical databases. Databases are modeled as an entity-relation graph containing nodes (database records) and links (relationships between records). Users specify a keyword search string to retrieve an initial set of nodes, and then explore intra- and interdatabase links. Results are visualized with user-defined semantic substrates to take advantage of the rich set of attributes usually present in biomedical data. Comments from domain experts indicate that this visualization method is potentially advantageous for biomedical knowledge exploration.}, keywords = {Bioinformatics, Biomedical computing, biomedical databases, cross-database exploration, Data exploration and discovery, Data visualization, database management systems, Databases, Factual, DNA, graph theory, Information Storage and Retrieval, information visualization., Keyword search, medical computing, natural language processing, Proteins, semantic networks, semantics, sequences, text mining, User-Computer Interface, user-defined semantics, visual databases}, isbn = {1545-5963}, doi = {10.1109/TCBB.2010.1}, author = {Lieberman,M.D. and Taheri, S. and Guo,Huimin and Mirrashed,F. and Yahav,I. and Aris,A. and Shneiderman, Ben} } @conference {14274, title = {Visual Scene Interpretation as a Dialogue between Vision and Language}, booktitle = {Workshops at the Twenty-Fifth AAAI Conference on Artificial Intelligence}, year = {2011}, month = {2011/08/24/}, abstract = {We present a framework for semantic visual scene interpretation in a system with vision and language. In this framework the system consists of two modules, a language module and a vision module that communicate with each other in a form of a dialogue to actively interpret the scene. The language module is responsible for obtaining domain knowledge from linguistic resources and reasoning on the basis of this knowledge and the visual input. It iteratively creates questions that amount to an attention mechanism for the vision module which in turn shifts its focus to selected parts of the scene and applies selective segmentation and feature extraction. As a formalism for optimizing this dialogue we use information theory. We demonstrate the framework on the problem of recognizing a static scene from its objects and show preliminary results for the problem of human activity recognition from video. Experiments demonstrate the effectiveness of the active paradigm in introducing attention and additional constraints into the sensing process.}, url = {http://www.aaai.org/ocs/index.php/WS/AAAIW11/paper/viewPaper/3989}, author = {Yu,Xiaodong and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {12862, title = {Warming Oceans, Phytoplankton, and River Discharge: Implications for Cholera Outbreaks}, journal = {The American Journal of Tropical Medicine and Hygiene}, volume = {85}, year = {2011}, month = {08/2011}, pages = {303 - 308}, abstract = {Phytoplankton abundance is inversely related to sea surface temperature (SST). However, a positive relationship is observed between SST and phytoplankton abundance in coastal waters of Bay of Bengal. This has led to an assertion that in a warming climate, rise in SST may increase phytoplankton blooms and, therefore, cholera outbreaks. Here, we explain why a positive SST-phytoplankton relationship exists in the Bay of Bengal and the implications of such a relationship on cholera dynamics. We found clear evidence of two independent physical drivers for phytoplankton abundance. The first one is the widely accepted phytoplankton blooming produced by the upwelling of cold, nutrient-rich deep ocean waters. The second, which explains the Bay of Bengal findings, is coastal phytoplankton blooming during high river discharges with terrestrial nutrients. Causal mechanisms should be understood when associating SST with phytoplankton and subsequent cholera outbreaks in regions where freshwater discharge are a predominant mechanism for phytoplankton production.}, isbn = {0002-9637,}, doi = {10.4269/ajtmh.2011.11-0181}, url = {http://www.ajtmh.org/content/85/2/303}, author = {Jutla,Antarpreet S. and Akanda,Ali S. and Griffiths,Jeffrey K. and Rita R Colwell and Islam,Shafiqul} } @conference {16417, title = {When does simulated data match real data?}, booktitle = {Proceedings of the 13th annual conference companion on Genetic and evolutionary computation}, series = {GECCO {\textquoteright}11}, year = {2011}, month = {2011///}, pages = {231 - 232}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Agent-based models can replicate real-world patterns, but finding parameters that achieve the best match can be difficult. To validate a model, a real-world dataset is often divided into a training set (to calibrate the parameters) and a test set (to validate the calibrated model). The difference between the training and test data and the simulated data is determined using an error measure. In the context of evolutionary computation techniques, the error measure also serves as a fitness function, and thus affects evolutionary search dynamics. We survey the effect of five different error measures on both a toy problem and a real world problem of matching a model to empirical online news consumption behavior. We use each error measure separately for calibration on the training dataset, and then examine the results of all five error measures on both the training and testing datasets. We show that certain error measures sometimes serve as better fitness functions than others, and in fact using one error measure may result in better calibration (on a different measure) than using the different measure directly. For the toy problem, the Pearson{\textquoteright}s correlation measure dominated all other measures, but no single error measure was Pareto dominant for the real world problem.}, keywords = {Agent-based modeling, business, Calibration, Genetic algorithms, information search, network analysis}, isbn = {978-1-4503-0690-4}, doi = {10.1145/2001858.2001988}, url = {http://doi.acm.org/10.1145/2001858.2001988}, author = {Stonedahl,Forrest and Anderson,David and Rand, William} } @article {15262, title = {Achieving anonymity via clustering}, journal = {ACM Trans. Algorithms}, volume = {6}, year = {2010}, month = {2010/07//}, pages = {49:1{\textendash}49:19 - 49:1{\textendash}49:19}, abstract = {Publishing data for analysis from a table containing personal records, while maintaining individual privacy, is a problem of increasing importance today. The traditional approach of deidentifying records is to remove identifying fields such as social security number, name, etc. However, recent research has shown that a large fraction of the U.S. population can be identified using nonkey attributes (called quasi-identifiers) such as date of birth, gender, and zip code. The k-anonymity model protects privacy via requiring that nonkey attributes that leak information are suppressed or generalized so that, for every record in the modified table, there are at least k-1 other records having exactly the same values for quasi-identifiers. We propose a new method for anonymizing data records, where quasi-identifiers of data records are first clustered and then cluster centers are published. To ensure privacy of the data records, we impose the constraint that each cluster must contain no fewer than a prespecified number of data records. This technique is more general since we have a much larger choice for cluster centers than k-anonymity. In many cases, it lets us release a lot more information without compromising privacy. We also provide constant factor approximation algorithms to come up with such a clustering. This is the first set of algorithms for the anonymization problem where the performance is independent of the anonymity parameter k. We further observe that a few outlier points can significantly increase the cost of anonymization. Hence, we extend our algorithms to allow an ε fraction of points to remain unclustered, that is, deleted from the anonymized publication. Thus, by not releasing a small fraction of the database records, we can ensure that the data published for analysis has less distortion and hence is more useful. Our approximation algorithms for new clustering objectives are of independent interest and could be applicable in other clustering scenarios as well.}, keywords = {anonymity, Approximation algorithms, clustering, privacy}, isbn = {1549-6325}, doi = {10.1145/1798596.1798602}, url = {http://doi.acm.org/10.1145/1798596.1798602}, author = {Aggarwal,Gagan and Panigrahy,Rina and Feder,Tom{\'a}s and Thomas,Dilys and Kenthapadi,Krishnaram and Khuller, Samir and Zhu,An} } @article {18499, title = {Building a dynamic reputation system for DNS}, journal = {19th Usenix Security Symposium}, year = {2010}, month = {2010///}, abstract = {The Domain Name System (DNS) is an essential protocolused by both legitimate Internet applications and cyber at- tacks. For example, botnets rely on DNS to support agile com- mand and control infrastructures. An effective way to disrupt these attacks is to place malicious domains on a {\textquotedblleft}blocklist{\textquotedblright} (or {\textquotedblleft}blacklist{\textquotedblright}) or to add a filtering rule in a firewall or net- work intrusion detection system. To evade such security coun- termeasures, attackers have used DNS agility, e.g., by using new domains daily to evade static blacklists and firewalls. In this paper we propose Notos, a dynamic reputation system for DNS. The premise of this system is that malicious, agile use of DNS has unique characteristics and can be distinguished from legitimate, professionally provisioned DNS services. No- tos uses passive DNS query data and analyzes the network and zone features of domains. It builds models of known legit- imate domains and malicious domains, and uses these models to compute a reputation score for a new domain indicative of whether the domain is malicious or legitimate. We have eval- uated Notos in a large ISP{\textquoteright}s network with DNS traffic from 1.4 million users. Our results show that Notos can identify malicious domains with high accuracy (true positive rate of 96.8\%) and low false positive rate (0.38\%), and can identify these domains weeks or even months before they appear in public blacklists. }, author = {Antonakakis,M. and Perdisci,R. and Dagon,D. and Lee,W. and Feamster, Nick} } @article {18553, title = {Building a fast, virtualized data plane with programmable hardware}, journal = {SIGCOMM Comput. Commun. Rev.}, volume = {40}, year = {2010}, month = {2010/01//}, pages = {75 - 82}, abstract = {Network virtualization allows many networks to share the same underlying physical topology; this technology has offered promise both for experimentation and for hosting multiple networks on a single shared physical infrastructure. Much attention has focused on virtualizing the network control plane, but, ultimately, a limiting factor in the deployment of these virtual networks is data-plane performance: Virtual networks must ultimately forward packets at rates that are comparable to native, hardware-based approaches. Aside from proprietary solutions from vendors, hardware support for virtualized data planes is limited. The advent of open, programmable network hardware promises flexibility, speed, and resource isolation, but, unfortunately, hardware does not naturally lend itself to virtualization. We leverage emerging trends in programmable hardware to design a flexible, hardware-based data plane for virtual networks. We present the design, implementation, and preliminary evaluation of this hardware-based data plane and show how the proposed design can support many virtual networks without compromising performance or isolation.}, keywords = {netfpga, network virtualization}, isbn = {0146-4833}, doi = {10.1145/1672308.1672323}, url = {http://doi.acm.org/10.1145/1672308.1672323}, author = {Anwer,Muhammad Bilal and Feamster, Nick} } @article {12870, title = {Comparative Genomics of Clinical and Environmental Vibrio Mimicus}, journal = {Proceedings of the National Academy of SciencesPNAS}, volume = {107}, year = {2010}, month = {2010/12/07/}, pages = {21134 - 21139}, abstract = {Whether Vibrio mimicus is a variant of Vibrio cholerae or a separate species has been the subject of taxonomic controversy. A genomic analysis was undertaken to resolve the issue. The genomes of V. mimicus MB451, a clinical isolate, and VM223, an environmental isolate, comprise ca. 4,347,971 and 4,313,453 bp and encode 3,802 and 3,290 ORFs, respectively. As in other vibrios, chromosome I (C-I) predominantly contains genes necessary for growth and viability, whereas chromosome II (C-II) bears genes for adaptation to environmental change. C-I harbors many virulence genes, including some not previously reported in V. mimicus, such as mannose-sensitive hemagglutinin (MSHA), and enterotoxigenic hemolysin (HlyA); C-II encodes a variant of Vibrio pathogenicity island 2 (VPI-2), and Vibrio seventh pandemic island II (VSP-II) cluster of genes. Extensive genomic rearrangement in C-II indicates it is a hot spot for evolution and genesis of speciation for the genus Vibrio. The number of virulence regions discovered in this study (VSP-II, MSHA, HlyA, type IV pilin, PilE, and integron integrase, IntI4) with no notable difference in potential virulence genes between clinical and environmental strains suggests these genes also may play a role in the environment and that pathogenic strains may arise in the environment. Significant genome synteny with prototypic pre-seventh pandemic strains of V. cholerae was observed, and the results of phylogenetic analysis support the hypothesis that, in the course of evolution, V. mimicus and V. cholerae diverged from a common ancestor with a prototypic sixth pandemic genomic backbone.}, isbn = {0027-8424, 1091-6490}, doi = {10.1073/pnas.1013825107}, url = {http://www.pnas.org/content/107/49/21134}, author = {Hasan,Nur A. and Grim,Christopher J. and Haley,Bradd J. and Jongsik Chun and Alam,Munirul and Taviani,Elisa and Mozammel Hoq and Munk,A. Christine and Rita R Colwell} } @conference {13530, title = {Context-Aware and Content-Based Dynamic Voronoi Page Segmentation}, booktitle = {The Nineth IAPRInternational Workshop on Document Analysis Systems}, year = {2010}, month = {2010///}, pages = {73 - 80}, abstract = {This paper presents a dynamic approach to document page segmentation based on inter-component relationships, local patterns and context features. State-of-the art page segmentation algorithms segment zones based on local properties of neighboring connected components such as distance and orientation, and do not typically consider additional properties other than size. Our proposed approach uses a contextually aware and dynamically adaptive page segmentation scheme. The page is first over-segmented using a dynamically adaptive scheme of separation features based on [2] and adapted from [13]. A decision to form zones is then based on the context built from these local separation features and high-level content features. Zone-based evaluation was performed on sets of printed and handwritten documents in English and Arabic scripts with multiple font types, sizes and we achieved an increase of 15\% over the accuracy reported in [2].}, author = {Agrawal,Mudit and David Doermann} } @conference {19445, title = {Deliberate Interactions: Characterizing Technology Use in Nairobi, Kenya}, booktitle = {SIGCHI {\textquoteright}10}, series = {CHI {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {2593 - 2602}, publisher = {ACM}, organization = {ACM}, abstract = {We present results from a qualitative study examining how professionals living and working in Nairobi, Kenya regularly use ICT in their everyday lives. There are two contributions of this work for the HCI community. First, we provide empirical evidence demonstrating constraints our participants encountered when using technology in an infrastructure-poor setting. These constraints are limited bandwidth, high costs, differing perceptions of responsiveness, and threats to physical and virtual security. Second, we use our findings to critically evaluate the "access, anytime and anywhere" construct shaping the design of future technologies. We present an alternative vision called deliberate interactions--a planned and purposeful interaction style that involves offline preparation and discuss ways ICT can support this online usage behavior.}, keywords = {everyday technology, hci4d, Kenya, urban computing}, isbn = {978-1-60558-929-9}, url = {http://doi.acm.org/10.1145/1753326.1753719}, author = {Wyche, Susan P. and Smyth, Thomas N. and Marshini Chetty and Aoki, Paul M. and Grinter, Rebecca E.} } @article {18539, title = {Don{\textquoteright}t Configure the Network, Program It! Domain-Specific Programming Languages for Network Systems}, volume = {YALEU/DCS/RR-1432}, year = {2010}, month = {2010///}, institution = {Yale University}, abstract = {Network operators must configure networks to accomplish critical, complex, and often conflicting requirements: they must ensure good performance while maintaining security, and satisfy contractual obligations while ensuring profitable use of interdomain connections. Unfortunately, today they have no choice but to implement these high-level goals by configuring hundreds of individual network devices. These interact in complex and unexpected ways, often resulting in misconfigurations or downtime. We propose a new approach: rather than configure individual network devices, operators should program the network holistically, according to high-level policies. Towards this goal, we present Nettle, a system for clearly and concisely expressing network requirements together with mechanisms to control the network accordingly. At the lowest level, we rely on OpenFlow switches for programmable network hardware. On top of this layer, we build an extensible family of embedded domain-specific languages (EDSLs), each aimed at different operational concerns and provide convenient ways to sensibly combine expressions in these languages. We present a case study demonstrating a DSL for networks that provides fine-grained, dynamic access control policies.}, author = {Feamster, Nick and Voellmy,A. and Agarwal,A. and Hudak,P. and Burnett,S. and Launchbury,J.} } @conference {14688, title = {Dynamically checking ownership policies in concurrent c/c++ programs}, booktitle = {Proceedings of the 37th annual ACM SIGPLAN-SIGACT symposium on Principles of programming languages}, series = {POPL {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {457 - 470}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Concurrent programming errors arise when threads share data incorrectly. Programmers often avoid these errors by using synchronization to enforce a simple ownership policy: data is either owned exclusively by a thread that can read or write the data, or it is read owned by a set of threads that can read but not write the data. Unfortunately, incorrect synchronization often fails to enforce these policies and memory errors in languages like C and C++ can violate these policies even when synchronization is correct. In this paper, we present a dynamic analysis for checking ownership policies in concurrent C and C++ programs despite memory errors. The analysis can be used to find errors in commodity multi-threaded programs and to prevent attacks that exploit these errors. We require programmers to write ownership assertions that describe the sharing policies used by different parts of the program. These policies may change over time, as may the policies{\textquoteright} means of enforcement, whether it be locks, barriers, thread joins, etc. Our compiler inserts checks in the program that signal an error if these policies are violated at runtime. We evaluated our tool on several benchmark programs. The run-time overhead was reasonable: between 0 and 49\% with an average of 26\%. We also found the tool easy to use: the total number of ownership assertions is small, and the asserted specification and implementation can be debugged together by running the instrumented program and addressing the errors that arise. Our approach enjoys a pleasing modular soundness property: if a thread executes a sequence of statements on variables it owns, the statements are serializable within a valid execution, and thus their effects can be reasoned about in isolation from other threads in the program.}, keywords = {concurrency, Debugging, Dynamic analysis, Security, Testing, tools}, isbn = {978-1-60558-479-9}, doi = {10.1145/1706299.1706351}, url = {http://doi.acm.org/10.1145/1706299.1706351}, author = {Martin,Jean-Phillipe and Hicks, Michael W. and Costa,Manuel and Akritidis,Periklis and Castro,Miguel} } @article {12477, title = {An Efficient and Robust Algorithm for Shape Indexing and Retrieval}, journal = {Multimedia, IEEE Transactions on}, volume = {12}, year = {2010}, month = {2010/08//}, pages = {372 - 385}, abstract = {Many shape matching methods are either fast but too simplistic to give the desired performance or promising as far as performance is concerned but computationally demanding. In this paper, we present a very simple and efficient approach that not only performs almost as good as many state-of-the-art techniques but also scales up to large databases. In the proposed approach, each shape is indexed based on a variety of simple and easily computable features which are invariant to articulations, rigid transformations, etc. The features characterize pairwise geometric relationships between interest points on the shape. The fact that each shape is represented using a number of distributed features instead of a single global feature that captures the shape in its entirety provides robustness to the approach. Shapes in the database are ordered according to their similarity with the query shape and similar shapes are retrieved using an efficient scheme which does not involve costly operations like shape-wise alignment or establishing correspondences. Depending on the application, the approach can be used directly for matching or as a first step for obtaining a short list of candidate shapes for more rigorous matching. We show that the features proposed to perform shape indexing can be used to perform the rigorous matching as well, to further improve the retrieval performance.}, keywords = {activity, algorithm;shape, classification;human, databases;robust, estimation;large, indexing;shape, MATCHING, matching;image, methods;shape, pose, recognition;, retrieval;image, retrieval;indexing;shape}, isbn = {1520-9210}, doi = {10.1109/TMM.2010.2050735}, author = {Biswas,S. and Aggarwal,G. and Chellapa, Rama} } @article {12426, title = {Evaluation of the Interaction of Cyclin-Dependent Kinase 5 with Activator p25 and with p25-Derived Inhibitor CIP}, journal = {Journal of Computational Biology}, volume = {17}, year = {2010}, month = {2010///}, pages = {707 - 721}, author = {Cardone, Antonio and Albers,R.W. and Sriram,R.D. and Pant,H.C.} } @inbook {14196, title = {An Experimental Study of Color-Based Segmentation Algorithms Based on the Mean-Shift Concept}, booktitle = {Computer Vision {\textendash} ECCV 2010Computer Vision {\textendash} ECCV 2010}, series = {Lecture Notes in Computer Science}, volume = {6312}, year = {2010}, month = {2010///}, pages = {506 - 519}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We point out a difference between the original mean-shift formulation of Fukunaga and Hostetler and the common variant in the computer vision community, namely whether the pairwise comparison is performed with the original or with the filtered image of the previous iteration. This leads to a new hybrid algorithm, called Color Mean Shift, that roughly speaking, treats color as Fukunaga{\textquoteright}s algorithm and spatial coordinates as Comaniciu{\textquoteright}s algorithm. We perform experiments to evaluate how different kernel functions and color spaces affect the final filtering and segmentation results, and the computational speed, using the Berkeley and Weizmann segmentation databases. We conclude that the new method gives better results than existing mean shift ones on four standard comparison measures ( improvement on RAND and BDE measures respectively for color images), with slightly higher running times ( ). Overall, the new method produces segmentations comparable in quality to the ones obtained with current state of the art segmentation algorithms.}, isbn = {978-3-642-15551-2}, url = {http://dx.doi.org/10.1007/978-3-642-15552-9_37}, author = {Bitsakos,K. and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Daniilidis,Kostas and Maragos,Petros and Paragios,Nikos} } @conference {13564, title = {Handwritten Arabic text line segmentation using affinity propagation}, booktitle = {Proceedings of the 9th IAPR International Workshop on Document Analysis Systems}, series = {DAS {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {135 - 142}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {In this paper, we present a novel graph-based method for extracting handwritten text lines in monochromatic Arabic document images. Our approach consists of two steps - Coarse text line estimation using primary components which define the line and assignment of diacritic components which are more difficult to associate with a given line. We first estimate local orientation at each primary component to build a sparse similarity graph. We then, use a shortest path algorithm to compute similarities between non-neighboring components. From this graph, we obtain coarse text lines using two estimates obtained from Affinity propagation and Breadth-first search. In the second step, we assign secondary components to each text line. The proposed method is very fast and robust to non-uniform skew and character size variations, normally present in handwritten text lines. We evaluate our method using a pixel-matching criteria, and report 96\% accuracy on a dataset of 125 Arabic document images. We also present a proximity analysis on datasets generated by artificially decreasing the spacings between text lines to demonstrate the robustness of our approach.}, keywords = {affinity propagation, arabic, arabic documents, breadth-first search, clustering, dijkstra{\textquoteright}s shortest path algorithm, handwritten documents, line detection, text line segmentation}, isbn = {978-1-60558-773-8}, doi = {10.1145/1815330.1815348}, url = {http://doi.acm.org/10.1145/1815330.1815348}, author = {Kumar,Jayant and Abd-Almageed, Wael and Kang,Le and David Doermann} } @article {14487, title = {On the importance of sharing negative results}, journal = {SIGKDD explorations}, volume = {12}, year = {2010}, month = {2010///}, pages = {3 - 4}, author = {Giraud-Carrier,C. and Dunham,M.H. and Atreya,A. and Elkan,C. and Perlich,C. and Swirszcz,G. and Shi,X. and Philip,S.Y. and F{\"u}rnkranz,J. and Sima,J.F.} } @article {13957, title = {Insights into head-related transfer function: Spatial dimensionality and continuous representation}, journal = {The Journal of the Acoustical Society of America}, volume = {127}, year = {2010}, month = {2010///}, pages = {2347 - 2357}, abstract = {This paper studies head-related transfer function (HRTF) sampling and synthesis in a three-dimensional auditory scene based on a general modal decomposition of the HRTF in all frequency-range-angle domains. The main finding is that the HRTF decomposition with the derived spatial basis function modes can be well approximated by a finite number, which is defined as the spatial dimensionality of the HRTF. The dimensionality determines the minimum number of parameters to represent the HRTF corresponding to all directions and also the required spatial resolution in HRTF measurement. The general model is further developed to a continuous HRTF representation, in which the normalized spatial modes can achieve HRTF near-field and far-field representations in one formulation. The remaining HRTF spectral components are compactly represented using a Fourier spherical Bessel series, where the aim is to generate the HRTF with much higher spectral resolution in fewer parameters from typical measurements, which usually have limited spectral resolution constrained by sampling conditions. A low-computation algorithm is developed to obtain the model coefficients from the existing measurements. The HRTF synthesis using the proposed model is validated by three sets of data: (i) synthetic HRTFs from the spherical head model, (ii) the MIT KEMAR (Knowles Electronics Mannequin for Acoustics Research) data, and (iii) 45-subject CIPIC HRTF measurements.}, keywords = {acoustic signal processing, Bessel functions, Fourier series, hearing, Transfer functions}, doi = {10.1121/1.3336399}, url = {http://link.aip.org/link/?JAS/127/2347/1}, author = {Zhang,Wen and Abhayapala,Thushara D. and Kennedy,Rodney A. and Duraiswami, Ramani} } @article {16052, title = {Interactive information visualization for exploring and querying electronic health records: A systematic review}, year = {2010}, month = {2010///}, institution = {Human-Computer Interaction Lab, University of Maryland}, abstract = {To overcome the complexity and scale of making medical decisions based on electronic health records (EHRs) a variety of visual methods have been proposed. This paper surveys twelve state-of-the-art information visualization systems described in the scientific literature and compares them based on a set of carefully selected criteria. It aims to systematically examine the systems{\textquoteright} strengths and weaknesses to inform future information visualization designs.We select twelve state-of-the-art information visualization systems from information visualization, medical information systems and human-computer interaction conferences and journals. We compare the systems using the following criteria: (1) data types covered, (2) multivariate analysis support, (3) number of patients records used (one or many), and (4) user intents addressed. The review describes the twelve systems in detail and evaluates the systems using the aforementioned criteria. We discuss how the systems differ in their features and highlight how these differences are related to their design and affect the user intent model. Examples of findings include: (1) most systems handle numerical or categorical data but not both, (2) most systems are specifically designed for looking at a single patient or multiple patients but not both, (3) most systems utilize horizontal time lines to represent time, (4) only systems that handle multiple patient records have good support for Filter, and (5) some specific user intents (e.g. the Encode and Connect intents) are rarely addressed. Based on our review results, we believe that effective information visualization can facilitate analysis of patient records, and we encourage the information visualization community to study the application of their systems and conduct more in depth evaluations. We identify potential future research topics in interactive support for data abstraction and medical tasks that involve looking at a single or multiple records. Finally, we propose to create a repository for data and tasks so benchmarks can be established for both academic and commercial patient record visualization systems. }, author = {Rind,A. and Wang,T. D and Aigner,W. and Miksh,S. and Wongsuphasawat,K. and Plaisant, Catherine and Shneiderman, Ben} } @conference {14227, title = {Learning shift-invariant sparse representation of actions}, booktitle = {2010 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, year = {2010}, month = {2010/06/13/18}, pages = {2630 - 2637}, publisher = {IEEE}, organization = {IEEE}, abstract = {A central problem in the analysis of motion capture (MoCap) data is how to decompose motion sequences into primitives. Ideally, a description in terms of primitives should facilitate the recognition, synthesis, and characterization of actions. We propose an unsupervised learning algorithm for automatically decomposing joint movements in human motion capture (MoCap) sequences into shift-invariant basis functions. Our formulation models the time series data of joint movements in actions as a sparse linear combination of short basis functions (snippets), which are executed (or {\textquotedblleft}activated{\textquotedblright}) at different positions in time. Given a set of MoCap sequences of different actions, our algorithm finds the decomposition of MoCap sequences in terms of basis functions and their activations in time. Using the tools of L1 minimization, the procedure alternately solves two large convex minimizations: Given the basis functions, a variant of Orthogonal Matching Pursuit solves for the activations, and given the activations, the Split Bregman Algorithm solves for the basis functions. Experiments demonstrate the power of the decomposition in a number of applications, including action recognition, retrieval, MoCap data compression, and as a tool for classification in the diagnosis of Parkinson (a motion disorder disease).}, keywords = {action characterization, Action recognition, action retrieval, action synthesis, Character recognition, data compression, human motion capture, HUMANS, Image matching, Image motion analysis, image representation, Image sequences, Information retrieval, joint movements, large convex minimizations, learning (artificial intelligence), learning shift-invariant sparse representation, Matching pursuit algorithms, minimisation, Minimization methods, MoCap data compression, Motion analysis, motion capture analysis, motion disorder disease, motion sequences, orthogonal matching pursuit, Parkinson diagnosis, Parkinson{\textquoteright}s disease, Pursuit algorithms, shift-invariant basis functions, short basis functions, snippets, sparse linear combination, split Bregman algorithm, time series, time series data, Unsupervised learning, unsupervised learning algorithm}, isbn = {978-1-4244-6984-0}, doi = {10.1109/CVPR.2010.5539977}, author = {Li,Yi and Ferm{\"u}ller, Cornelia and Aloimonos, J. and Hui Ji} } @conference {16025, title = {The Metacognitive Loop: An Architecture for Building Robust Intelligent Systems}, booktitle = {2010 AAAI Fall Symposium Series}, year = {2010}, month = {2010/03/11/}, abstract = {The Metacognitive Loop: An Architecture for Building Robust Intelligent Systems}, keywords = {commonsense, ontologies, robust intelligent systems}, url = {http://www.aaai.org/ocs/index.php/FSS/FSS10/paper/view/2161}, author = {Shahri,Hamid Haidarian and Dinalankara,Wikum and Fults,Scott and Wilson,Shomir and Perlis, Don and Schmill,Matt and Oates,Tim and Josyula,Darsana and Anderson,Michael} } @inbook {16823, title = {Multidimensional data structures for spatial applications}, booktitle = {Algorithms and theory of computation handbookAlgorithms and theory of computation handbook}, year = {2010}, month = {2010///}, pages = {6 - 6}, publisher = {Chapman \& Hall/CRC}, organization = {Chapman \& Hall/CRC}, abstract = {An overview is presented of a number of representations of multidimensional data that arise in spatial applications. Multidimensional spatial data consists of points as well as objects that have extent such as line segments, rectangles, regions, and volumes. The points may have locational as well as nonlocational attributes. The focus is on spatial data which is a subset of multidimensional data consisting of points with locational attributes and objects with extent. The emphasis is on hierarchical representations based on the "divide-and-conquer" problem-solving paradigm. They are of interest because they enable focusing computational resources on the interesting subsets of data. Thus, there is no need to expend work where the payoff is small. These representations are of use in operations such as range searching and finding nearest neighbors.}, isbn = {978-1-58488-822-2}, url = {http://dl.acm.org/citation.cfm?id=1882757.1882763}, author = {Samet, Hanan}, editor = {Atallah,Mikhail J. and Blanton,Marina} } @conference {18589, title = {Network I/O fairness in virtual machines}, booktitle = {Proceedings of the second ACM SIGCOMM workshop on Virtualized infrastructure systems and architectures}, series = {VISA {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {73 - 80}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {We present a mechanism for achieving network I/O fairness in virtual machines, by applying flexible rate limiting mechanisms directly to virtual network interfaces. Conventional approaches achieve this fairness by implementing rate limiting either in the virtual machine monitor or hypervisor, which generates considerable CPU interrupt and instruction overhead for forwarding packets. In contrast, our design pushes per-VM rate limiting as close as possible to the physical hardware themselves, effectively implementing per-virtual interface rate limiting in hardware. We show that this design reduces CPU overhead (both interrupts and instructions) by an order of magnitude. Our design can be applied either to virtual servers for cloud-based services, or to virtual routers}, keywords = {netfpga, network virtualization, xen}, isbn = {978-1-4503-0199-2}, doi = {10.1145/1851399.1851412}, url = {http://doi.acm.org/10.1145/1851399.1851412}, author = {Anwer,Muhammad Bilal and Nayak,Ankur and Feamster, Nick and Liu,Ling} } @article {14513, title = {Object Classification Using Taxonomies}, volume = {12/414,065}, year = {2010}, month = {2010/07/22/}, abstract = {As provided herein objects from a source catalog, such as a provider{\textquoteright}s catalog, can be added to a target catalog, such as an enterprise master catalog, in a scalable manner utilizing catalog taxonomies. A baseline classifier determines probabilities for source objects to target catalog classes. Source objects can be assigned to those classes with probabilities that meet a desired threshold and meet a desired rate. A classification cost for target classes can be determined for respective unassigned source objects, which can comprise determining an assignment cost and separation cost for the source objects for respective desired target classes. The separation and assignment costs can be combined to determine the classification cost, and the unassigned source objects can be assigned to those classes having a desired classification cost.}, url = {http://www.google.com/patents?id=oXDSAAAAEBAJ}, author = {Tsaparas,Panayiotis and Papadimitriou,Panagiotis and Fuxman,Ariel D. and Getoor, Lise and Agrawal,Rakesh}, editor = {Microsoft Corporation} } @conference {16826, title = {Ontuition: intuitive data exploration via ontology navigation}, booktitle = {Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems}, series = {GIS {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {540 - 541}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Ontuition, a system for mapping ontologies, is presented. Transforming data to a usable format for Ontuition involves recognizing and resolving data values corresponding to concepts in multiple ontological domains. In particular, for datasets with a geographic component an attempt is made to identify and extract enough spatio-textual data that specific lat/long values to dataset entries can be assigned. Next, a gazetteer is used to transform the textually-specified locations into lat/long values that can be displayed on a map. Non-spatial ontological concepts are also discovered. This methodology is applied to the National Library of Medicine{\textquoteright}s very popular clinical trials website (http://clinicaltrials.gov/) whose users are generally interested in locating trials near where they live. The trials are specified using XML files. The location data is extracted and coupled with a disease ontology to enable general queries on the data with the result being of use to a very large group of people. The goal is to do this automatically for such ontology datasets with a locational component.}, keywords = {MAPPING, ontology, ontuition, spatio-textual}, isbn = {978-1-4503-0428-3}, doi = {10.1145/1869790.1869887}, url = {http://doi.acm.org/10.1145/1869790.1869887}, author = {Adelfio,Marco D. and Lieberman,Michael D. and Samet, Hanan and Firozvi,Kashif A.} } @article {19373, title = {Overlap-based cell tracker}, journal = {The Journal of Research of the National Institute of Standards and Technology}, volume = {115}, year = {2010}, month = {11/2010}, chapter = {477}, abstract = { order to facilitate the extraction of quantitative data from live cell image sets, automated image analysis methods are needed. This paper presents an introduction to the general principle of an overlap cell tracking software developed by the National Institute of Standards and Technology (NIST). This cell tracker has the ability to track cells across a set of time lapse images acquired at high rates based on the amount of overlap between cellular regions in consecutive frames. It is designed to be highly flexible, requires little user parameterization, and has a fast execution time. }, author = {Chalfoun, J and Cardone, Antonio and Dima, A.A. and Allen, D.P. and Halter, M.W.} } @article {12466, title = {PADS: A Probabilistic Activity Detection Framework for Video Data}, journal = {Pattern Analysis and Machine Intelligence, IEEE Transactions on}, volume = {32}, year = {2010}, month = {2010/12//}, pages = {2246 - 2261}, abstract = {There is now a growing need to identify various kinds of activities that occur in videos. In this paper, we first present a logical language called Probabilistic Activity Description Language (PADL) in which users can specify activities of interest. We then develop a probabilistic framework which assigns to any subvideo of a given video sequence a probability that the subvideo contains the given activity, and we finally develop two fast algorithms to detect activities within this framework. OffPad finds all minimal segments of a video that contain a given activity with a probability exceeding a given threshold. In contrast, the OnPad algorithm examines a video during playout (rather than afterwards as OffPad does) and computes the probability that a given activity is occurring (even if the activity is only partially complete). Our prototype Probabilistic Activity Detection System (PADS) implements the framework and the two algorithms, building on top of existing image processing algorithms. We have conducted detailed experiments and compared our approach to four different approaches presented in the literature. We show that-for complex activity definitions-our approach outperforms all the other approaches.}, keywords = {Automated;Programming Languages;Video Recording;, Computer-Assisted;Models, PADL;PADS;image processing algorithms;offPad algorithm;onPad algorithm;probabilistic activity description language;probabilistic activity detection framework;video data;video sequence;image sequences;object detection;probability;video surveillance;Algorit, Statistical;Movement;Pattern Recognition}, isbn = {0162-8828}, doi = {10.1109/TPAMI.2010.33}, author = {Albanese, M. and Chellapa, Rama and Cuntoor, N. and Moscato, V. and Picariello, A. and V.S. Subrahmanian and Udrea,O.} } @conference {13624, title = {Performance Evaluation Tools for Zone Segmentation and Classification (PETS)}, booktitle = {International Conference on Pattern Recognition}, year = {2010}, month = {2010///}, pages = {503 - 506}, abstract = {This paper overviews a set of Performance Evaluation ToolS (PETS) for zone segmentation and classification. The tools allow researchers and developers to evaluate, optimize and compare their algorithms by providing a variety of quantitative performance metrics. The evaluation of segmentation quality is based on the pixel-based overlaps between two sets of regions proposed by Randriamasy and Vincent. PETS extends the approach by providing a set of metrics for overlap analysis, RLE and polygonal representation of regions and introduces type-matching to evaluate zone classification. The software is available for research use.}, author = {Seo,W. and Agrawal,Mudit and David Doermann} } @article {18708, title = {Perturbing the Ubiquitin Pathway Reveals How Mitosis Is Hijacked to Denucleate and Regulate Cell Proliferation and Differentiation In Vivo}, journal = {PLoS ONEPLoS ONE}, volume = {5}, year = {2010}, month = {2010/10/20/}, pages = {e13331 - e13331}, abstract = {BackgroundThe eye lens presents a unique opportunity to explore roles for specific molecules in cell proliferation, differentiation and development because cells remain in place throughout life and, like red blood cells and keratinocytes, they go through the most extreme differentiation, including removal of nuclei and cessation of protein synthesis. Ubiquitination controls many critical cellular processes, most of which require specific lysines on ubiquitin (Ub). Of the 7 lysines (K) least is known about effects of modification of K6. Methodology and Principal Findings We replaced K6 with tryptophan (W) because K6 is the most readily modified K and W is the most structurally similar residue to biotin. The backbone of K6W-Ub is indistinguishable from that of Wt-Ub. K6W-Ub is effectively conjugated and deconjugated but the conjugates are not degraded via the ubiquitin proteasome pathways (UPP). Expression of K6W-ubiquitin in the lens and lens cells results in accumulation of intracellular aggregates and also slows cell proliferation and the differentiation program, including expression of lens specific proteins, differentiation of epithelial cells into fibers, achieving proper fiber cell morphology, and removal of nuclei. The latter is critical for transparency, but the mechanism by which cell nuclei are removed has remained an age old enigma. This was also solved by expressing K6W-Ub. p27kip, a UPP substrate accumulates in lenses which express K6W-Ub. This precludes phosphorylation of nuclear lamin by the mitotic kinase, a prerequisite for disassembly of the nuclear membrane. Thus the nucleus remains intact and DNAseIIβ neither gains entry to the nucleus nor degrades the DNA. These results could not be obtained using chemical proteasome inhibitors that cannot be directed to specific tissues. Conclusions and Significance K6W-Ub provides a novel, genetic means to study functions of the UPP because it can be targeted to specific cells and tissues. A fully functional UPP is required to execute most stages of lens differentiation, specifically removal of cell nuclei. In the absence of a functional UPP, small aggregate prone, cataractous lenses are formed. }, doi = {10.1371/journal.pone.0013331}, url = {http://dx.doi.org/10.1371/journal.pone.0013331}, author = {Caceres,Andrea and Shang,Fu and Wawrousek,Eric and Liu,Qing and Avidan,Orna and Cvekl,Ales and Yang,Ying and Haririnia,Aydin and Storaska,Andrew and Fushman, David and Kuszak,Jer and Dudek,Edward and Smith,Donald and Taylor,Allen} } @conference {12486, title = {Pose estimation in heavy clutter using a multi-flash camera}, booktitle = {Robotics and Automation (ICRA), 2010 IEEE International Conference on}, year = {2010}, month = {2010/05//}, pages = {2028 - 2035}, abstract = {We propose a novel solution to object detection, localization and pose estimation with applications in robot vision. The proposed method is especially applicable when the objects of interest may not be richly textured and are immersed in heavy clutter. We show that a multi-flash camera (MFC) provides accurate separation of depth edges and texture edges in such scenes. Then, we reformulate the problem, as one of finding matches between the depth edges obtained in one or more MFC images to the rendered depth edges that are computed offline using 3D CAD model of the objects. In order to facilitate accurate matching of these binary depth edge maps, we introduce a novel cost function that respects both the position and the local orientation of each edge pixel. This cost function is significantly superior to traditional Chamfer cost and leads to accurate matching even in heavily cluttered scenes where traditional methods are unreliable. We present a sub-linear time algorithm to compute the cost function using techniques from 3D distance transforms and integral images. Finally, we also propose a multi-view based pose-refinement algorithm to improve the estimated pose. We implemented the algorithm on an industrial robot arm and obtained location and angular estimation accuracy of the order of 1 mm and 2 $\#$x00B0; respectively for a variety of parts with minimal texture.}, keywords = {3D, algorithm;object, based, camera;multiview, depth, detection;object, detection;pose, distance, edge, edges;cameras;image, edges;integral, estimation;binary, estimation;multiflash, estimation;robot, function;depth, images;location, localization;pose, maps, matching;cost, matching;image, pose-refinement, texture;object, transforms;angular, vision;texture, vision;transforms;}, doi = {10.1109/ROBOT.2010.5509897}, author = {Ming-Yu Liu and Tuzel, O. and Veeraraghavan,A. and Chellapa, Rama and Agrawal,A. and Okuda, H.} } @article {18711, title = {Reversible Post-Translational Carboxylation Modulates the Enzymatic Activity of N-Acetyl-l-ornithine Transcarbamylase}, journal = {Biochemistry}, volume = {49}, year = {2010}, month = {2010///}, pages = {6887 - 6895}, abstract = {N-Acetyl-l-ornithine transcarbamylase (AOTCase), rather than ornithine transcarbamylase (OTCase), is the essential carbamylase enzyme in the arginine biosynthesis of several plant and human pathogens. The specificity of this unique enzyme provides a potential target for controlling the spread of these pathogens. Recently, several crystal structures of AOTCase from Xanthomonas campestris (xc) have been determined. In these structures, an unexplained electron density at the tip of the Lys302 side chain was observed. Using 13C NMR spectroscopy, we show herein that Lys302 is post-translationally carboxylated. The structure of wild-type AOTCase in a complex with the bisubstrate analogue Nδ-(phosphonoacetyl)-Nα-acetyl-l-ornithine (PALAO) indicates that the carboxyl group on Lys302 forms a strong hydrogen bonding network with surrounding active site residues, Lys252, Ser253, His293, and Glu92 from the adjacent subunit either directly or via a water molecule. Furthermore, the carboxyl group is involved in binding N-acetyl-l-ornithine via a water molecule. Activity assays with the wild-type enzyme and several mutants demonstrate that the post-translational modification of lysine 302 has an important role in catalysis.N-Acetyl-l-ornithine transcarbamylase (AOTCase), rather than ornithine transcarbamylase (OTCase), is the essential carbamylase enzyme in the arginine biosynthesis of several plant and human pathogens. The specificity of this unique enzyme provides a potential target for controlling the spread of these pathogens. Recently, several crystal structures of AOTCase from Xanthomonas campestris (xc) have been determined. In these structures, an unexplained electron density at the tip of the Lys302 side chain was observed. Using 13C NMR spectroscopy, we show herein that Lys302 is post-translationally carboxylated. The structure of wild-type AOTCase in a complex with the bisubstrate analogue Nδ-(phosphonoacetyl)-Nα-acetyl-l-ornithine (PALAO) indicates that the carboxyl group on Lys302 forms a strong hydrogen bonding network with surrounding active site residues, Lys252, Ser253, His293, and Glu92 from the adjacent subunit either directly or via a water molecule. Furthermore, the carboxyl group is involved in binding N-acetyl-l-ornithine via a water molecule. Activity assays with the wild-type enzyme and several mutants demonstrate that the post-translational modification of lysine 302 has an important role in catalysis. }, isbn = {0006-2960}, doi = {10.1021/bi1007386}, url = {http://dx.doi.org/10.1021/bi1007386}, author = {Li,Yongdong and Yu,Xiaolin and Ho,Jeremy and Fushman, David and Allewell,Norma M. and Tuchman,Mendel and Shi,Dashuang} } @inbook {12507, title = {Semantic Video Content Analysis}, booktitle = {Video Search and MiningVideo Search and Mining}, series = {Studies in Computational Intelligence}, volume = {287}, year = {2010}, month = {2010///}, pages = {147 - 176}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {In recent years, there has been significant interest in the area of automatically recognizing activities occurring in a camera{\textquoteright}s field of view and detecting abnormalities. The practical applications of such a system could include airport tarmac monitoring, or monitoring of activities in secure installations, to name a few. The difficulty of the problem is compounded by several factors: detection of primitive actions in spite of changes in illumination, occlusions and noise; complexmultiagent interaction;mapping of higher-level activities to lower-level primitive actions; variations in which the same semantic activity can be performed. In this chapter, we develop a theory of semantic activity analysis that addresses each of these issues in an integrated manner. Specifically, we discuss ontological representations of knowledge of a domain, integration of domain knowledge and statistical models for achieving semantic mappings, definition of logical languages to describe activities, and design of frameworks which integrate all the above aspects in a coherent way, thus laying the foundations of effective Semantic Video Content Analysis systems.}, isbn = {978-3-642-12899-8}, url = {http://dx.doi.org/10.1007/978-3-642-12900-1_6}, author = {Albanese,Massimiliano and Turaga,Pavan and Chellapa, Rama and Pugliese,Andrea and Subrahmanian,V.}, editor = {Schonfeld,Dan and Shan,Caifeng and Tao,Dacheng and Wang,Liang} } @article {16397, title = {THE SHAKY LADDER HYPERPLANE-DEFINED FUNCTIONS AND CLASSIC DYNAMIC PROBLEMS}, journal = {International Journal of Computational Intelligence and Applications}, volume = {09}, year = {2010}, month = {2010///}, pages = {33 - 33}, abstract = {To improve the understanding of the GA in dynamic environments we explore a set of test problems, the shaky ladder hyper-defined functions (sl-hdf), and extend these functions to create versions that are equivalent to many classical dynamic problems. We do this by constraining the space of all sl-hdfs to create representations of these classical functions. We have examined three classical problems, and compared sl-hdf versions of these problems with their standard representations. These results show that the sl-hdfs are representative of a larger class of problems, and can represent a larger class of test suite. Previous results on sl-hdf showed that GA performance is best in the Defined Cliffs variant of the sl-hdf. We build upon these results to improve GA performance in several classes of real world dynamic problems by modifying the problem representation. These results lend insight into dynamic problems where the GA will perform well.}, isbn = {1469-0268}, doi = {10.1142/S1469026810002756}, url = {http://www.worldscinet.com/ijcia/09/0901/S1469026810002756.html}, author = {Alharbi,Abir and Rand, William} } @article {19368, title = {Structural and dynamic determinants of ligand binding and regulation of cyclin-dependent kinase 5 by pathological activator p25 and inhibitory peptide CIP.}, journal = {Journal of molecular biology}, volume = {401}, year = {2010}, month = {2010 Aug 20}, pages = {478-92}, abstract = {The crystal structure of the cdk5/p25 complex has provided information on possible molecular mechanisms of the ligand binding, specificity, and regulation of the kinase. Comparative molecular dynamics simulations are reported here for physiological conditions. This study provides new insight on the mechanisms that modulate such processes, which may be exploited to control pathological activation by p25. The structural changes observed in the kinase are stabilized by a network of interactions involving highly conserved residues within the cyclin-dependent kinase (cdk) family. Collective motions of the proteins (cdk5, p25, and CIP) and their complexes are identified by principal component analysis, revealing two conformational states of the activation loop upon p25 complexation, which are absent in the uncomplexed kinase and not apparent from the crystal. Simulations of the uncomplexed inhibitor CIP show structural rearrangements and increased flexibility of the interfacial loop containing the critical residue E240, which becomes fully hydrated and available for interactions with one of several positively charged residues in the kinase. These changes provide a rationale for the observed high affinity and enhanced inhibitory action of CIP when compared to either p25 or the physiological activators of cdk5.}, keywords = {Crystallography, X-Ray, Cyclin-Dependent Kinase 5, Cyclin-Dependent Kinase Inhibitor Proteins, HUMANS, Ligands, molecular dynamics simulation, Nerve Tissue Proteins, Principal component analysis, Protein Binding, Protein Conformation}, issn = {1089-8638}, doi = {10.1016/j.jmb.2010.06.040}, author = {Cardone, Antonio and Hassan, S A and Albers,R.W. and Sriram,R.D. and Pant,H.C.} } @article {18601, title = {SwitchBlade: a platform for rapid deployment of network protocols on programmable hardware}, journal = {SIGCOMM Comput. Commun. Rev.}, volume = {40}, year = {2010}, month = {2010/08//}, pages = {183 - 194}, abstract = {We present SwitchBlade, a platform for rapidly deploying custom protocols on programmable hardware. SwitchBlade uses a pipeline-based design that allows individual hardware modules to be enabled or disabled on the fly, integrates software exception handling, and provides support for forwarding based on custom header fields. SwitchBlade{\textquoteright}s ease of programmability and wire-speed performance enables rapid prototyping of custom data-plane functions that can be directly deployed in a production network. SwitchBlade integrates common packet-processing functions as hardware modules, enabling different protocols to use these functions without having to resynthesize hardware. SwitchBlade{\textquoteright}s customizable forwarding engine supports both longest-prefix matching in the packet header and exact matching on a hash value. SwitchBlade{\textquoteright}s software exceptions can be invoked based on either packet or flow-based rules and updated quickly at runtime, thus making it easy to integrate more flexible forwarding function into the pipeline. SwitchBlade also allows multiple custom data planes to operate in parallel on the same physical hardware, while providing complete isolation for protocols running in parallel. We implemented SwitchBlade using NetFPGA board, but SwitchBlade can be implemented with any FPGA. To demonstrate SwitchBlade{\textquoteright}s flexibility, we use SwitchBlade to implement and evaluate a variety of custom network protocols: we present instances of IPv4, IPv6, Path Splicing, and an OpenFlow switch, all running in parallel while forwarding packets at line rate.}, keywords = {netfpga, network virtualization}, isbn = {0146-4833}, doi = {10.1145/1851275.1851206}, url = {http://doi.acm.org/10.1145/1851275.1851206}, author = {Anwer,Muhammad Bilal and Motiwala,Murtaza and Bin Tariq,Mukarram and Feamster, Nick} } @conference {18785, title = {A Systematic Methodology for Accurate Design-Stage Estimation of Energy Consumption for Injection Molded Parts}, year = {2010}, month = {2010///}, abstract = {Today{\textquoteright}s ubiquitous use of plastics in product designand manufacturing presents significant environmental and human health challenges. Injection molding, one of the most commonly used processes for making plastic products, consumes a significant amount of energy. A methodology for accurately estimating the energy consumed to injection-mold a part would enable environmentally conscious decision making during the product design. Unfortunately, only limited information is available at the design stage. Therefore, accurately estimating energy consumption before the part has gone into production can be challenging. In this paper, we describe a methodology for energy estimation that works with the limited amount of data available during the design stage, namely the CAD model of the part, the material name, and the production requirements. This methodology uses this data to estimate the parameters of the runner system and an appropriately sized molding machine. It then uses these estimates to compute the machine setup time and the cycle time required for the injection molding operation. This is done by appropriately abstracting information available from the mold flow simulation tools and analytical models that are traditionally used during the manufacturing stage. These times are then multiplied by the power consumed by the appropriately sized machine during each stage of the molding cycle to compute the estimated energy consumption per part. }, url = {http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.165.4139\&rep=rep1\&type=pdf}, author = {Weissman,A. and Ananthanarayanan,A. and Gupta,S.K. and Sriram,R.D.} } @inbook {17395, title = {Technology-Mediated Social Participation: Deep Science and Extreme Technology}, booktitle = {Active Media TechnologyActive Media Technology}, series = {Lecture Notes in Computer Science}, volume = {6335}, year = {2010}, month = {2010///}, pages = {1 - 4}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {The dramatic success of social media such as Facebook, Twitter, YouTube, Flickr, blogs, and traditional discussion groups empowers individuals to become active in local and global communities. With modest redesign, these technologies can be harnessed to support national priorities such as healthcare/wellness, disaster response, community safety, energy sustainability, etc. This talk describes a research agenda for these topics that develops deep science questions and extreme technology challenges.}, isbn = {978-3-642-15469-0}, url = {http://dx.doi.org/10.1007/978-3-642-15470-6_1}, author = {Shneiderman, Ben}, editor = {An,Aijun and Lingras,Pawan and Petty,Sheila and Huang,Runhe} } @conference {15620, title = {Tight lower bounds for halfspace range searching}, booktitle = {Proceedings of the 2010 annual symposium on Computational geometry}, series = {SoCG {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {29 - 37}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {We establish two new lower bounds for the halfspace range searching problem: Given a set of n points in ℜd, where each point is associated with a weight from a commutative semigroup, compute the semigroup sum of the weights of the points lying within any query halfspace. Letting $m$ denote the space requirements, we prove a lower bound for general semigroups of Ω(n1-1/(d+1)/m1/(d+1)) and for integral semigroups of Ω(n/m1/d). Our lower bounds are proved in the semigroup arithmetic model. Neglecting logarithmic factors, our result for integral semigroups matches the best known upper bound due to Matou{\v s}ek. Our result for general semigroups improves upon the best known lower bound due to Br{\"o}nnimann, Chazelle, and Pach. Moreover, Fonseca and Mount have recently shown that, given uniformly distributed points, halfspace range queries over idempotent semigroups can be answered in O(n1-1/(d+1)/m1/(d+1)) time in the semigroup arithmetic model. As our lower bounds are established for uniformly distributed point sets, it follows that they also resolve the computational complexity of halfspace range searching over idempotent semigroups in this important special case.}, keywords = {Idempotence, lower bounds, Range searching}, isbn = {978-1-4503-0016-2}, doi = {10.1145/1810959.1810964}, url = {http://doi.acm.org/10.1145/1810959.1810964}, author = {Arya,Sunil and Mount, Dave and Xia,Jian} } @conference {13925, title = {Toque: designing a cooking-based programming language for and with children}, booktitle = {Proceedings of the 28th international conference on Human factors in computing systems}, year = {2010}, month = {2010///}, pages = {2417 - 2426}, author = {Tarkan,S. and Sazawal,V. and Druin, Allison and Golub,E. and Bonsignore,E. M and Walsh,G. and Atrash,Z.} } @article {15558, title = {A unified approach to approximate proximity searching}, journal = {Algorithms{\textendash}ESA 2010}, year = {2010}, month = {2010///}, pages = {374 - 385}, author = {Arya,S. and da Fonseca,G. and Mount, Dave} } @inbook {17503, title = {Visualizing Threaded Conversation Networks: Mining Message Boards and Email Lists for Actionable Insights}, booktitle = {Active Media TechnologyActive Media Technology}, series = {Lecture Notes in Computer Science}, volume = {6335}, year = {2010}, month = {2010///}, pages = {47 - 62}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Analyzing complex online relationships is a difficult job, but new information visualization tools are enabling a wider range of users to make actionable insights from the growing volume of online data. This paper describes the challenges and methods for conducting analyses of threaded conversations such as found in enterprise message boards, email lists, and forums. After defining threaded conversation, we characterize the types of networks that can be extracted from them. We then provide 3 mini case studies to illustrate how actionable insights for community managers can be gained by applying the network analysis metrics and visualizations available in the free, open source NodeXL tool, which is a powerful, yet easy-to-use tool embedded in Excel 2007/2010.}, isbn = {978-3-642-15469-0}, url = {http://dx.doi.org/10.1007/978-3-642-15470-6_7}, author = {Hansen,Derek and Shneiderman, Ben and Smith,Marc}, editor = {An,Aijun and Lingras,Pawan and Petty,Sheila and Huang,Runhe} } @conference {14351, title = {WATTR: a method for self-powered wireless sensing of water activity in the home}, booktitle = {Proceedings of the 12th ACM international conference on Ubiquitous computing}, series = {Ubicomp {\textquoteright}10}, year = {2010}, month = {2010///}, pages = {169 - 172}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, keywords = {power harvesting, sensing, water conservation}, isbn = {978-1-60558-843-8}, doi = {10.1145/1864349.1864378}, url = {http://doi.acm.org/10.1145/1864349.1864378}, author = {Campbell,Tim and Larson,Eric and Cohn,Gabe and Jon Froehlich and Alcaide,Ramses and Patel,Shwetak N.} } @article {12527, title = {3D Face Recognition: Technology and Applications}, journal = {Handbook of Remote Biometrics}, year = {2009}, month = {2009///}, pages = {217 - 246}, author = {Tistarelli,M. and Li, S.Z. and Chellapa, Rama and Akarun, L.} } @conference {14193, title = {Active segmentation for robotics}, booktitle = {IEEE/RSJ International Conference on Intelligent Robots and Systems, 2009. IROS 2009}, year = {2009}, month = {2009/10/10/15}, pages = {3133 - 3139}, publisher = {IEEE}, organization = {IEEE}, abstract = {The semantic robots of the immediate future are robots that will be able to find and recognize objects in any environment. They need the capability of segmenting objects in their visual field. In this paper, we propose a novel approach to segmentation based on the operation of fixation by an active observer. Our approach is different from current approaches: while existing works attempt to segment the whole scene at once into many areas, we segment only one image region, specifically the one containing the fixation point. Furthermore, our solution integrates monocular cues (color, texture) with binocular cues (stereo disparities and optical flow). Experiments with real imagery collected by our active robot and from the known databases demonstrate the promise of the approach.}, keywords = {binocular cues, image colour analysis, Image segmentation, Image texture, Intelligent robots, Layout, Machine vision, monocular cues, Navigation, optical flow, Orbital robotics, Robot sensing systems, robot vision, Robot vision systems, robotics active segmentation, Robotics and automation, semantic robot, Simultaneous localization and mapping, stereo disparity, stereo image processing}, isbn = {978-1-4244-3803-7}, doi = {10.1109/IROS.2009.5354325}, author = {Mishra,A. and Aloimonos, J. and Ferm{\"u}ller, Cornelia} } @conference {18659, title = {Analyzing the process of installing rogue software}, year = {2009}, month = {2009/07/29/2}, pages = {560 - 565}, abstract = {This practical experience report presents the results of an experiment aimed at understanding the sequence of malicious actions following a remote compromise. The type of rogue software installed during attacks was used to classify and understand sequences of malicious actions. For this experiment, we used four Linux target computers running SSH with simple passwords. During the eight-month data collection period, we recorded a total of 1,171 attack sessions. In these sessions, attackers typed a total of 20,335 commands that we categorized into 24 specific actions. These actions were analyzed based on the type of rogue software installed by attackers.}, keywords = {Linux, Linux target computers, malicious actions, rogue software installation, security of data}, doi = {10.1109/DSN.2009.5270293}, author = {Berthier,R. and Arjona,J. and Michel Cukier} } @conference {11902, title = {Approximate kernel matrix computation on GPUs forlarge scale learning applications}, booktitle = {Proceedings of the 23rd international conference on Supercomputing}, series = {ICS {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {511 - 512}, publisher = {ACM}, organization = {ACM}, address = {Yorktown Heights, NY, USA}, abstract = {Kernel-based learning methods require quadratic space and time complexities to compute the kernel matrix. These complexities limit the applicability of kernel methods to large scale problems with millions of data points. In this paper, we introduce a novel representation of kernel matrices on Graphics Processing Units (GPU). The novel representation exploits the sparseness of the kernel matrix to address the space complexity problem. It also respects the guidelines for memory access on GPUs, which are critical for good performance, to address the time complexity problem. Our representation utilizes the locality preserving properties of space filling curves to obtain a band approximation of the kernel matrix. To prove the validity of the representation, we use Affinity Propagation, an unsupervised clustering algorithm, as an example of kernel methods. Experimental results show a 40x speedup of AP using our representation without degradation in clustering performance.}, keywords = {affinity propagation, algorithms, arrays, gpu, kernel methods, parallel programming, performance, space filling curves, sparse matrices}, isbn = {978-1-60558-498-0}, doi = {10.1145/1542275.1542355}, author = {Hussein,Mohamed E and Abd-Almageed, Wael} } @conference {15855, title = {Arabic Cross-Document Coreference Resolution}, booktitle = {Proceedings of the ACL-IJCNLP 2009 Conference Short Papers$}$}, year = {2009}, month = {2009///}, pages = {357 - 360}, author = {Sayeed,A. and Elsayed,T. and Garera,N. and Alexander,D. and Xu,T. and Oard, Douglas and Yarowsky,D. and Piatko,C.} } @article {12531, title = {An architectural level design methodology for smart camera applications}, journal = {International Journal of Embedded Systems}, volume = {4}, year = {2009}, month = {2009///}, pages = {83 - 97}, abstract = {Today{\textquoteright}s embedded computing applications are characterised by increased functionality, and hence increased design complexity and processing requirements. The resulting design spaces are vast and designers are typically able to evaluate only small subsets of solutions due to lack of efficient design tools. In this paper, we propose an architectural level design methodology that provides a means for a comprehensive design space exploration for smart camera applications and enable designers to select higher quality solutions and provides substantial savings on the overall cost of the system. We present efficient, accurate and intuitive models for performance estimation and validate them with experiments.}, author = {Saha,S. and Kianzad,V. and Schlessman,J. and Aggarwal,G. and Bhattacharyya, Shuvra S. and Chellapa, Rama and Wolf,W.} } @conference {15136, title = {Attacking cryptographic schemes based on "perturbation polynomials"}, booktitle = {Proceedings of the 16th ACM conference on Computer and communications security}, series = {CCS {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {1 - 10}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {We show attacks on several cryptographic schemes that have recently been proposed for achieving various security goals in sensor networks. Roughly speaking, these schemes all use "perturbation polynomials" to add "noise" to polynomialbased systems that offer information-theoretic security, in an attempt to increase the resilience threshold while maintaining efficiency. We show that the heuristic security arguments given for these modified schemes do not hold, and that they can be completely broken once we allow even a slight extension of the parameters beyond those achieved by the underlying information-theoretic schemes. Our attacks apply to the key predistribution scheme of Zhang et al. (MobiHoc 2007), the access-control schemes of Subramanian et al. (PerCom 2007), and the authentication schemes of Zhang et al. (INFOCOM 2008). Our results cast doubt on the viability of using "perturbation polynomials" for designing secure cryptographic schemes.}, keywords = {pairwise key establishment, random perturbation polynomial, sensor network security}, isbn = {978-1-60558-894-0}, doi = {10.1145/1653662.1653664}, url = {http://doi.acm.org/10.1145/1653662.1653664}, author = {Albrecht,Martin and Gentry,Craig and Halevi,Shai and Katz, Jonathan} } @article {18772, title = {Characterization and control of pin diameter during in-mold assembly of mesoscale revolute joints}, journal = {North American Manufacturing Research Institute}, volume = {37}, year = {2009}, month = {2009///}, abstract = {Macro-scale revolute joints can be formed byfirst molding the hole and then molding the pin inside the hole. As the pin shrinks during the solidification process, it moves away from the hole and provides the clearance for the joint to function. The value of clearance in the macro- scale joint can be controlled by carefully selecting the process parameters and the material for the pin. However, in order for this strategy to work at the mesoscale, it requires the use of very thin cores to form sub-millimeter holes. Such thin cores are very difficult to make, are easily damaged during the molding process, and very difficult to retract from the hole. Our previous work has shown that making the pins first and then creating holes on the top of pins leads to successful mesoscale joints. This strategy is counter intuitive based on our experiences at the macro-scale. At the macroscale, as the hole shrinks on top of the pin, the joint is jammed. So a fundamental question is why this counter-intuitive strategy works at the mesoscale. In this paper we show that at the mesoscale, the joint jamming is prevented because of the deformation of the pin under the compressive loading during the second stage molding. We also describe features in the mold that can control the pin deformation and hence control the joint parameters. We present experimental data and computational models to show how mesoscale revolute joints can be formed. }, url = {http://web.mit.edu/arvinda/www/NAMRI_2009_draft.pdf}, author = {Ananthanarayanan,A. and Gupta,S.K. and Bruck,H. A.} } @conference {18555, title = {Characterizing VLAN-induced sharing in a campus network}, booktitle = {Proceedings of the 9th ACM SIGCOMM conference on Internet measurement conference}, series = {IMC {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {116 - 121}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Many enterprise, campus, and data-center networks have complex layer-2 virtual LANs ("VLANs") below the IP layer. The interaction between layer-2 and IP topologies in these VLANs introduces hidden dependencies between IP level network and the physical infrastructure that has implications for network management tasks such as planning for capacity or reliability, and for fault diagnosis. This paper characterizes the extent and effect of these dependencies in a large campus network. We first present the design and implementation of EtherTrace, a tool that we make publicly available, which infers the layer-2 topology using data passively collected from Ethernet switches. Using this tool, we infer the layer-2 topology for a large campus network and compare it with the IP topology. We find that almost 70\% of layer-2 edges are shared by 10 or more IP edges, and a single layer-2 edge may be shared by as many as 34 different IP edges. This sharing of layer-2 edges and switches among IP paths commonly results from trunking multiple VLANs to the same access router, or from colocation of academic departments that share layer-2 infrastructure, but have logically separate IP subnet and routers. We examine how this sharing affects the accuracy and specificity of fault diagnosis. For example, applying network tomography to the IP topology to diagnose failures caused by layer-2 devices results in only 54\% accuracy, compared to 100\% accuracy when our tomography algorithm takes input across layers.}, keywords = {network diagnosis, network virtualization, VLAN, vlan-induced dependency}, isbn = {978-1-60558-771-4}, doi = {10.1145/1644893.1644907}, url = {http://doi.acm.org/10.1145/1644893.1644907}, author = {Bin Tariq,Mukarram and Mansy,Ahmed and Feamster, Nick and Ammar,Mostafa} } @conference {13521, title = {Clutter Noise Removal in Binary Document Images}, booktitle = {International Conference on Document Analysis and Recognition (ICDAR {\textquoteright}09)}, year = {2009}, month = {2009///}, pages = {556 - 560}, abstract = {The paper presents a clutter detection and removal algorithm for complex document images. The distance transform based approach is independent of clutter{\textquoteright}s position, size, shape and connectivity with text. Features are based on a new technique called {\textquoteleft}nth erosion{\textquoteright} and clutter elements are identified with an SVM classifier. Removal is restrictive, so text attached to the clutter is not deleted in the process. The method was tested on a mix of degraded and noisy, machine-printed and handwritten Arabic and English text documents. Results show pixel-level accuracies of 97.5\% and 95\% for clutter detection and removal respectively. This approach was also extended with a noise detection and removal model for documents having a mix of clutter and salt-n-pepper noise.}, author = {Agrawal,Mudit and David Doermann} } @article {15077, title = {Collusion-free multiparty computation in the mediated model}, journal = {Advances in Cryptology-CRYPTO 2009}, year = {2009}, month = {2009///}, pages = {524 - 540}, abstract = {Collusion-free protocols prevent subliminal communication (i.e., covert channels) between parties running the protocol. In the standard communication model, if one-way functions exist, then protocols satisfying any reasonable degree of privacy cannot be collusion-free. To circumvent this impossibility, Alwen, shelat and Visconti (CRYPTO 2008) recently suggested the mediated model where all communication passes through a mediator. The goal is to design protocols where collusion-freeness is guaranteed as long as the mediator is honest, while standard security guarantees hold if the mediator is dishonest. In this model, they gave constructions of collusion-free protocols for commitments and zero-knowledge proofs in the two-party setting.We strengthen the definition of Alwen et al., and resolve the main open questions in this area by showing a collusion-free protocol (in the mediated model) for computing any multi-party functionality. }, doi = {10.1007/978-3-642-03356-8_31}, author = {Alwen,J. and Katz, Jonathan and Lindell,Y. and Persiano,G. and Shelat,A. and Visconti,I.} } @article {15298, title = {Computational linguistics for metadata building (CLiMB): using text mining for the automatic identification, categorization, and disambiguation of subject terms for image metadata}, journal = {Multimedia Tools and Applications}, volume = {42}, year = {2009}, month = {2009///}, pages = {115 - 138}, abstract = {In this paper, we present a system using computational linguistic techniques to extract metadata for image access. We discuss the implementation, functionality and evaluation of an image catalogers{\textquoteright} toolkit, developed in the Computational Linguistics for Metadata Building (CLiMB) research project. We have tested components of the system, including phrase finding for the art and architecture domain, functional semantic labeling using machine learning, and disambiguation of terms in domain-specific text vis a vis a rich thesaurus of subject terms, geographic and artist names. We present specific results on disambiguation techniques and on the nature of the ambiguity problem given the thesaurus, resources, and domain-specific text resource, with a comparison of domain-general resources and text. Our primary user group for evaluation has been the cataloger expert with specific expertise in the fields of painting, sculpture, and vernacular and landscape architecture.}, doi = {10.1007/s11042-008-0253-9}, author = {Klavans,J. L and Sheffield,C. and Abels,E. and Jimmy Lin and Passonneau,R. and Sidhu,T. and Soergel,D.} } @conference {15891, title = {Cross-document coreference resolution: A key technology for learning by reading}, booktitle = {AAAI Spring Symposium on Learning by Reading and Learning to Read}, year = {2009}, month = {2009///}, author = {Mayfield,J. and Alexander,D. and Dorr, Bonnie J and Eisner,J. and Elsayed,T. and Finin,T. and Fink,C. and Freedman,M. and Garera,N. and McNamee,P. and others} } @article {14603, title = {CTCF binding site classes exhibit distinct evolutionary, genomic, epigenomic and transcriptomic features}, journal = {Genome Biology}, volume = {10}, year = {2009}, month = {2009/11/18/}, pages = {R131 - R131}, abstract = {CTCF (CCCTC-binding factor) is an evolutionarily conserved zinc finger protein involved in diverse functions ranging from negative regulation of MYC, to chromatin insulation of the beta-globin gene cluster, to imprinting of the Igf2 locus. The 11 zinc fingers of CTCF are known to differentially contribute to the CTCF-DNA interaction at different binding sites. It is possible that the differences in CTCF-DNA conformation at different binding sites underlie CTCF{\textquoteright}s functional diversity. If so, the CTCF binding sites may belong to distinct classes, each compatible with a specific functional role.}, isbn = {1465-6906}, doi = {10.1186/gb-2009-10-11-r131}, url = {http://genomebiology.com/2009/10/11/R131}, author = {Essien,Kobby and Vigneau,Sebastien and Apreleva,Sofia and Singh,Larry N. and Bartolomei,Marisa S. and Hannenhalli, Sridhar} } @conference {18562, title = {Detecting network neutrality violations with causal inference}, booktitle = {Proceedings of the 5th international conference on Emerging networking experiments and technologies}, series = {CoNEXT {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {289 - 300}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {We present NANO, a system that detects when ISPs apply policies that discriminate against specific classes of applications, users, or destinations. Existing systems for detecting discrimination are typically specific to an application or to a particular discrimination mechanism and rely on active measurement tests. Unfortunately, ISPs can change discrimination policies and mechanisms, and they can evade these tests by giving probe traffic higher priority. NANO detects ISP discrimination by passively collecting performance data from clients. To distinguish discrimination from other causes of degradation (e.g., overload, misconfiguration, failure), NANO establishes a causal relationship between an ISP and observed performance by adjusting for confounding factors. NANO agents deployed at participating clients across the Internet collect performance data for selected services and report this information to centralized servers, which analyze the measurements to establish causal relationship between an ISP and performance degradations. We have implemented NANO and deployed clients in a controlled environment on Emulab. We run a combination of controlled experiments on Emulab and wide-area experiments on PlanetLab that show that NANO can determine the extent and criteria for discrimination for a variety of discrimination policies and applications.}, keywords = {causal inference, network neutrality}, isbn = {978-1-60558-636-6}, doi = {10.1145/1658939.1658972}, url = {http://doi.acm.org/10.1145/1658939.1658972}, author = {Bin Tariq,Mukarram and Motiwala,Murtaza and Feamster, Nick and Ammar,Mostafa} } @article {12890, title = {Diversity and Seasonality of Bioluminescent Vibrio Cholerae Populations in Chesapeake Bay}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {75}, year = {2009}, month = {2009/01/01/}, pages = {135 - 146}, abstract = {Association of luminescence with phenotypic and genotypic traits and with environmental parameters was determined for 278 strains of Vibrio cholerae isolated from the Chesapeake Bay during 1998 to 2000. Three clusters of luminescent strains (A, B, and C) and two nonluminescent clusters (X and Y) were identified among 180 clonal types. V. cholerae O1 strains isolated during pandemics and endemic cholera in the Ganges Delta were related to cluster Y. Heat-stable enterotoxin (encoded by stn) and the membrane protein associated with bile resistance (encoded by ompU) were found to be linked to luminescence in strains of cluster A. Succession from nonluminescent to luminescent populations of V. cholerae occurred during spring to midsummer. Occurrence of cluster A strains in water with neutral pH was contrasted with that of cluster Y strains in water with a pH of >8. Cluster A was found to be associated with a specific calanoid population cooccurring with cyclopoids. Cluster B was related to cluster Y, with its maximal prevalence at pH 8. Occurrence of cluster B strains was more frequent with warmer water temperatures and negatively correlated with maturity of the copepod community. It is concluded that each cluster of luminescent V. cholerae strains occupies a distinct ecological niche. Since the dynamics of these niche-specific subpopulations are associated with zooplankton community composition, the ecology of luminescent V. cholerae is concluded to be related to its interaction with copepods and related crustacean species.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.02894-07}, url = {http://aem.asm.org/content/75/1/135}, author = {Zo,Young-Gun and Chokesajjawatee,Nipa and Grim,Christopher and Arakawa,Eiji and Watanabe,Haruo and Rita R Colwell} } @article {15579, title = {The effect of corners on the complexity of approximate range searching}, journal = {Discrete \& Computational Geometry}, volume = {41}, year = {2009}, month = {2009///}, pages = {398 - 443}, abstract = {Given an n-element point set in ℝ d , the range searching problem involves preprocessing these points so that the total weight, or for our purposes the semigroup sum, of the points lying within a given query range η can be determined quickly. In ε-approximate range searching we assume that η is bounded, and the sum is required to include all the points that lie within η and may additionally include any of the points lying within distance ε.diam(η) of η{\textquoteright}s boundary.In this paper we contrast the complexity of approximate range searching based on properties of the semigroup and range space. A semigroup (S,+) is idempotent if x+x=x for all x∈S, and it is integral if for all k>=2, the k-fold sum x+...+x is not equal to x. Recent research has shown that the computational complexity of approximate spherical range searching is significantly lower for idempotent semigroups than it is for integral semigroups in terms of the dependencies on ε. In this paper we consider whether these results can be generalized to other sorts of ranges. We show that, as with integrality, allowing sharp corners on ranges has an adverse effect on the complexity of the problem. In particular, we establish lower bounds on the worst-case complexity of approximate range searching in the semigroup arithmetic model for ranges consisting of d-dimensional unit hypercubes under rigid motions. We show that for arbitrary (including idempotent) semigroups and linear space, the query time is at least \varOmega(1/{\varepsilon }^{d-2\sqrt{d}}) . In the case of integral semigroups we prove a tighter lower bound of Ω(1/ε d-2). These lower bounds nearly match existing upper bounds for arbitrary semigroups. In contrast, we show that the improvements offered by idempotence do apply to smooth convex ranges. We say that a range is smooth if at every boundary point there is an incident Euclidean sphere that lies entirely within the range whose radius is proportional to the range{\textquoteright}s diameter. We show that for smooth ranges and idempotent semigroups, ε-approximate range queries can be answered in O(log n+(1/ε)(d-1)/2log (1/ε)) time using O(n/ε) space. We show that this is nearly tight by presenting a lower bound of Ω(log n+(1/ε)(d-1)/2). This bound is in the decision-tree model and holds irrespective of space. }, doi = {10.1007/s00454-009-9140-z}, author = {Arya,S. and Malamatos,T. and Mount, Dave} } @conference {11907, title = {Efficient band approximation of Gram matrices for large scale kernel methods on GPUs}, booktitle = {Proceedings of the Conference on High Performance Computing Networking, Storage and Analysis}, series = {SC {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {31:1{\textendash}31:10 - 31:1{\textendash}31:10}, publisher = {ACM}, organization = {ACM}, address = {Portland, Oregon}, abstract = {Kernel-based methods require O(N2) time and space complexities to compute and store non-sparse Gram matrices, which is prohibitively expensive for large scale problems. We introduce a novel method to approximate a Gram matrix with a band matrix. Our method relies on the locality preserving properties of space filling curves, and the special structure of Gram matrices. Our approach has several important merits. First, it computes only those elements of the Gram matrix that lie within the projected band. Second, it is simple to parallelize. Third, using the special band matrix structure makes it space efficient and GPU-friendly. We developed GPU implementations for the Affinity Propagation (AP) clustering algorithm using both our method and the COO sparse representation. Our band approximation is about 5 times more space efficient and faster to construct than COO. AP gains up to 6x speedup using our method without any degradation in its clustering performance.}, keywords = {affinity propagation, algorithms, gpu}, isbn = {978-1-60558-744-8}, doi = {10.1145/1654059.1654091}, author = {Hussein,Mohamed and Abd-Almageed, Wael} } @article {12095, title = {Efficient peer-to-Peer lookup in multi-hop wireless networks}, journal = {KSII Trans. Internet and Information Systems}, volume = {3}, year = {2009}, month = {2009///}, pages = {5 - 25}, author = {Shin,M. and Arbaugh, William A.} } @inbook {19635, title = {Efficient Robust Private Set Intersection}, booktitle = {Applied Cryptography and Network Security}, series = {Lecture Notes in Computer Science}, year = {2009}, month = {2009/01/01/}, pages = {125 - 142}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, abstract = {Computing Set Intersection privately and efficiently between two mutually mistrusting parties is an important basic procedure in the area of private data mining. Assuring robustness, namely, coping with potentially arbitrarily misbehaving (i.e., malicious) parties, while retaining protocol efficiency (rather than employing costly generic techniques) is an open problem. In this work the first solution to this problem is presented.}, keywords = {Coding and Information Theory, Computer Communication Networks, Cryptographic protocols, Data Encryption, Data Structures, Cryptology and Information Theory, Information Systems Applications (incl.Internet), Privacy Preserving Data Mining, Secure Two-party Computation, Set Intersection, Systems and Data Security}, isbn = {978-3-642-01956-2, 978-3-642-01957-9}, url = {http://link.springer.com/chapter/10.1007/978-3-642-01957-9_8}, author = {Dana Dachman-Soled and Malkin, Tal and Raykova, Mariana and Yung, Moti}, editor = {Abdalla, Michel and Pointcheval, David and Fouque, Pierre-Alain and Vergnaud, Damien} } @conference {12517, title = {Enforcing integrability by error correction using l1-minimization}, booktitle = {Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on}, year = {2009}, month = {2009/06//}, pages = {2350 - 2357}, abstract = {Surface reconstruction from gradient fields is an important final step in several applications involving gradient manipulations and estimation. Typically, the resulting gradient field is non-integrable due to linear/non-linear gradient manipulations, or due to presence of noise/outliers in gradient estimation. In this paper, we analyze integrability as error correction, inspired from recent work in compressed sensing, particulary lscr0 - lscr1 equivalence. We propose to obtain the surface by finding the gradient field which best fits the corrupted gradient field in lscr1 sense. We present an exhaustive analysis of the properties of lscr1 solution for gradient field integration using linear algebra and graph analogy. We consider three cases: (a) noise, but no outliers (b) no-noise but outliers and (c) presence of both noise and outliers in the given gradient field. We show that lscr1 solution performs as well as least squares in the absence of outliers. While previous lscr0 - lscr1 equivalence work has focused on the number of errors (outliers), we show that the location of errors is equally important for gradient field integration. We characterize the lscr1 solution both in terms of location and number of outliers, and outline scenarios where lscr1 solution is equivalent to lscr0 solution. We also show that when lscr1 solution is not able to remove outliers, the property of local error confinement holds: i.e., the errors do not propagate to the entire surface as in least squares. We compare with previous techniques and show that lscr1 solution performs well across all scenarios without the need for any tunable parameter adjustments.}, keywords = {-, algebra;lscr0, algebra;minimisation;, analogy;integrability;least, compressed, correction;gradient, equivalence;lscr1-minimization;noise-outlier;surface, estimation;gradient, field, integration;gradient, lscr1, manipulation;graph, methods;graph, reconstruction;error, reconstruction;linear, sensing;error, squares;linear, theory;image}, doi = {10.1109/CVPR.2009.5206603}, author = {Reddy, D. and Agrawal,A. and Chellapa, Rama} } @conference {13054, title = {Exponential family hybrid semi-supervised learning}, booktitle = {Proceedings of the 21st International Joint Conference on Artifical Intelligence (IJCAI-09)}, year = {2009}, month = {2009///}, pages = {974 - 979}, author = {Agarwal,A. and Daum{\'e}, Hal} } @inbook {12526, title = {Face Recognition from Video}, booktitle = {The Essential Guide to Video Processing (Second Edition)The Essential Guide to Video Processing (Second Edition)}, year = {2009}, month = {2009///}, pages = {653 - 688}, publisher = {Academic Press}, organization = {Academic Press}, address = {Boston}, isbn = {978-0-12-374456-2}, url = {http://www.sciencedirect.com/science/article/pii/B9780123744562000232}, author = {Zhou,Shaohua Kevin and Chellapa, Rama and Aggarwal,Gaurav}, editor = {Al Bovik} } @article {18523, title = {A Fast, Virtualized Data Plane for the NetFPGA}, journal = {NetFPGA Developers Workshop, Stanford, California}, year = {2009}, month = {2009///}, pages = {90 - 94}, abstract = {Network virtualization allows many networks to share thesame underlying physical topology; this technology has of- fered promise both for experimentation and for hosting mul- tiple networks on a single shared physical infrastructure. Much attention has focused on virtualizing the network con- trol plane, but, ultimately, a limiting factor in the deployment of these virtual networks is data-plane performance: Virtual networks must ultimately forward packets at rates that are comparable to native, hardware-based approaches. Aside from proprietary solutions from vendors, hardware support for virtualized data planes is limited. The advent of open, programmable network hardware promises flexibility, speed, and resource isolation, but, unfortunately, hardware does not naturally lend itself to virtualization. We leverage emerg- ing trends in programmable hardware to design a flexible, hardware-based data plane for virtual networks. We present the design, implementation, and preliminary evaluation of this hardware-based data plane and show how the proposed design can support many virtual networks without compro- mising performance or isolation. }, author = {Anwer,M.B. and Feamster, Nick} } @article {12029, title = {Image Transformations and Blurring}, journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence}, volume = {31}, year = {2009}, month = {2009///}, pages = {811 - 823}, abstract = {Since cameras blur the incoming light during measurement, different images of the same surface do not contain the same information about that surface. Thus, in general, corresponding points in multiple views of a scene have different image intensities. While multiple-view geometry constrains the locations of corresponding points, it does not give relationships between the signals at corresponding locations. This paper offers an elementary treatment of these relationships. We first develop the notion of "ideal{\textquotedblright} and "real{\textquotedblright} images, corresponding to, respectively, the raw incoming light and the measured signal. This framework separates the filtering and geometric aspects of imaging. We then consider how to synthesize one view of a surface from another; if the transformation between the two views is affine, it emerges that this is possible if and only if the singular values of the affine matrix are positive. Next, we consider how to combine the information in several views of a surface into a single output image. By developing a new tool called "frequency segmentation,{\textquotedblright} we show how this can be done despite not knowing the blurring kernel.}, keywords = {reconstruction, restoration, sharpening and deblurring, smoothing.}, isbn = {0162-8828}, doi = {http://doi.ieeecomputersociety.org/10.1109/TPAMI.2008.133}, author = {Domke, Justin and Aloimonos, J.} } @conference {14560, title = {Improved approximation algorithms for prize-collecting Steiner tree and TSP}, booktitle = {2009 50th Annual IEEE Symposium on Foundations of Computer Science}, year = {2009}, month = {2009///}, pages = {427 - 436}, author = {Archer,A. and Bateni,M. H and Hajiaghayi, Mohammad T. and Karloff,H.} } @article {18751, title = {Integrated product and process design for a flapping wing drive mechanism}, journal = {Journal of Mechanical Design}, volume = {131}, year = {2009}, month = {2009///}, pages = {061006 - 061006}, author = {Bejgerowski,W. and Ananthanarayanan,A. and Mueller,D. and Gupta,S.K.} } @conference {17242, title = {Interacting with eHealth: towards grand challenges for HCI}, booktitle = {Proceedings of the 27th international conference extended abstracts on Human factors in computing systems}, year = {2009}, month = {2009///}, pages = {3309 - 3312}, abstract = {While health records are increasingly stored electronically, we, as citizens, have little access to this data about ourselves. We are not used to thinking of these official records either as ours or as useful to us. We increasingly turn to the Web, however, to query any ache, pain or health goal we may have before consulting with health care professionals. Likewise, for proactive health care such as nutrition or fitness, or to find fellow-sufferers for post diagnosis support, we turn to online resources. There is a potential disconnect between points at which professional and lay eHealth data and resources intersect for preventative or proactive health care. Such gaps in information sharing may have direct impact on practices we decide to take up, the care we seek, or the support professionals offer. In this panel, we consider several places within proactive, preventative health care in particular HCI has a role towards enhancing health knowledge discovery and health support interaction. Our goal is to demonstrate how now is the time for eHealth to come to the forefront of the HCI research agenda.}, author = {Andr{\'e},P. and White,R. and Tan,D. and Berners-Lee,T. and Consolvo,S. and Jacobs,R. and Kohane,I. and Le Dantec,C.A. and Mamykina,L. and Marsden,G.} } @article {14228, title = {MEASURING 1ST ORDER STRETCH WITH A SINGLE FILTER}, journal = {Relation}, volume = {10}, year = {2009}, month = {2009///}, pages = {691 - 691}, abstract = {We analytically develop a filter that is able to measurethe linear stretch of the transformation around a point, and present results of applying it to real signals. We show that this method is a real-time alternative solution for measuring local signal transformations. Experimentally, this method can accurately measure stretch, however, it is sensitive to shift. }, author = {Bitsakos,K. and Domke, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {13453, title = {Measuring General Relational Structure Using the Block Modularity Clustering Objective}, booktitle = {Twenty-Second International FLAIRS Conference}, year = {2009}, month = {2009/03/16/}, abstract = {The performance of all relational learning techniques has an implicit dependence on the underlying connectivity structure of the relations that are used as input. In this paper, we show how clustering can be used to develop an efficient optimization strategy can be used to effectively measure the structure of a graph in the absence of labeled instances.}, url = {http://www.aaai.org/ocs/index.php/FLAIRS/2009/paper/viewPaper/46}, author = {Anthony,Adam Paul and desJardins, Marie and Lombardi,Michael} } @conference {18513, title = {Measuring vlan-induced dependencies on a campus network}, booktitle = {Proceedings of the ACM SIGCOMM Internet measurement conference}, year = {2009}, month = {2009///}, abstract = {Many enterprise, campus, and data-center networks have complexlayer-2 virtual LANs ({\textquotedblleft}VLANs{\textquotedblright}) underlying the layer-3 network. The interac- tion between layer-2 and layer-3 topologies can introduce dependencies that are not visible when analyzed solely from layer 3; this paper studies the extent and impact of these dependencies. We first present EtherTrace, a tool that infers the layer-2 topology using data passively collected from Ethernet switches. Using this tool, we infer the layer-2 topology for a large campus network and compare the two topologies. Our comparison yields some striking initial findings: almost 70\% of layer-2 edges are shared by 10 or more IP edges, and a single layer-2 edge may be shared by as many as 34 different IP edges. This sharing has significant implications for both robustness and for network tomography. Applying network tomography to the IP topology to diagnose failures caused by layer-2 devices results in only 54\% accuracy, compared to 100\% accuracy when tomography is performed at layer 2. }, author = {Mansy,A. and Tariq,M.B. and Feamster, Nick and Ammar,M.} } @conference {13970, title = {Modal expansion of HRTFs: Continuous representation in frequency-range-angle}, booktitle = {Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on}, year = {2009}, month = {2009/04//}, pages = {285 - 288}, abstract = {This paper proposes a continuous HRTF representation in both 3D spatial and frequency domains. The method is based on the acoustic reciprocity principle and a modal expansion of the wave equation solution to represent the HRTF variations with different variables in separate basis functions. The derived spatial basis modes can achieve HRTF near-field and far-field representation in one formulation. The HRTF frequency components are expanded using Fourier Spherical Bessel series for compact representation. The proposed model can be used to reconstruct HRTFs at any arbitrary position in space and at any frequency point from a finite number of measurements. Analytical simulated and measured HRTFs from a KEMAR are used to validate the model.}, keywords = {analysis;signal, bessel, domains;frequency-range-angle;head, expansion;modal, Fourier, function;modal, functions;, reconstruction;signal, related, representation;transfer, series;HRTF;frequency, spherical, transfer}, doi = {10.1109/ICASSP.2009.4959576}, author = {Zhang,Wen and Abhayapala,T.D. and Kennedy,R.A. and Duraiswami, Ramani} } @article {18534, title = {A Narrow Waist for Multipath Routing}, year = {2009}, month = {2009///}, institution = {School of Computer Science, Georgia Tech}, abstract = {Many applications can use multipath routing to improvereliability or throughput, and many multipath routing pro- tocols exist. Despite this diversity of mechanisms and applications, no common interface exists to allow an ap- plication to select these paths. This paper presents a de- sign for such a common interface, called path bits. Path bits are a small string of opaque, semantic-free bits in a packet header; these bits have a simple property: changing a packet{\textquoteright}s path bits should, with high probability, result in the packet taking a different path to the destination. This paper presents the design of path bits and demonstrates that they are simple enough to be easily implemented in both hardware and software and expressive enough to sup- port a variety of applications that can benefit from mul- tipath routing. We present both hardware and software implementations of multipath routing protocols that im- plement the path bits abstraction, as well as implementa- tions of applications that can use this abstraction with only small modifications. }, author = {Motiwala,M. and bin Tariq,M. and Anwer,B. and Andersen,D. and Feamster, Nick} } @article {12887, title = {New records of phytoplankton for Bangladesh. 9. Some rare and a new species}, journal = {Bangladesh Journal of Plant Taxonomy}, volume = {16}, year = {2009}, month = {2009/06/25/}, abstract = {Ten taxa belonging to Chlorophyceae, Cyanophyceae, Bacillariophyceae and Euglenophyceae, and one with an uncertain taxonomic position have been described in this paper. Of these, 10 taxa have been found to be globally rare and new records for Bangladesh, whereas Strombomonas islamii Khondker sp. nov. has been described as new to science.}, isbn = {1028-2092}, doi = {10.3329/bjpt.v16i1.2734}, url = {http://www.banglajol.info/bd/index.php/BJPT/article/viewArticle/2734}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasmin,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @conference {15258, title = {Online allocation of display advertisements subject to advanced sales contracts}, booktitle = {Proceedings of the Third International Workshop on Data Mining and Audience Intelligence for Advertising}, series = {ADKDD {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {69 - 77}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {In this paper we propose a utility model that accounts for both sales and branding advertisers. We first study the computational complexity of optimization problems related to both online and offline allocation of display advertisements. Next, we focus on a particular instance of the online allocation problem, and design a simple online algorithm with provable approximation guarantees. Our algorithm is near optimal as is shown by a matching lower bound. Finally, we report on experiments to establish actual case behavior on some real datasets, with encouraging results.}, keywords = {display advertising, modeling, online algorithms, OPTIMIZATION, simulation}, isbn = {978-1-60558-671-7}, doi = {10.1145/1592748.1592758}, url = {http://doi.acm.org/10.1145/1592748.1592758}, author = {Alaei,Saeed and Arcaute,Esteban and Khuller, Samir and Ma,Wenjing and Malekian,Azarakhsh and Tomlin,John} } @conference {13619, title = {Page Rule-Line Removal using Linear Subspaces in Monochromatic Handwritten Arabic Documents}, booktitle = {Intl. Conf. on Document Analysis and Recognition (ICDAR 09)}, year = {2009}, month = {2009///}, pages = {768 - 772}, abstract = {In this paper we present a novel method for removing page rule lines in monochromatic handwritten Arabic documents using subspace methods with minimal effect on the quality of the foreground text. We use moment and histogram properties to extract features that represent the characteristics of the underlying rule lines. A linear subspace is incrementally built to obtain a line model that can be used to identify rule line pixels. We also introduce a novel scheme for evaluating noise removal algorithms in general and we use it to assess the quality of our rule line removal algorithm. Experimental results presented on a data set of 50 Arabic documents, handwritten by different writers, demonstrate the effectiveness of the proposed method.}, author = {Abd-Almageed, Wael and Kumar,Jayant and David Doermann} } @article {16827, title = {Path oracles for spatial networks}, journal = {Proc. VLDB Endow.}, volume = {2}, year = {2009}, month = {2009/08//}, pages = {1210 - 1221}, abstract = {The advent of location-based services has led to an increased demand for performing operations on spatial networks in real time. The challenge lies in being able to cast operations on spatial networks in terms of relational operators so that they can be performed in the context of a database. A linear-sized construct termed a path oracle is introduced that compactly encodes the n2 shortest paths between every pair of vertices in a spatial network having n vertices thereby reducing each of the paths to a single tuple in a relational database and enables finding shortest paths by repeated application of a single SQL SELECT operator. The construction of the path oracle is based on the observed coherence between the spatial positions of both source and destination vertices and the shortest paths between them which facilitates the aggregation of source and destination vertices into groups that share common vertices or edges on the shortest paths between them. With the aid of the Well-Separated Pair (WSP) technique, which has been applied to spatial networks using the network distance measure, a path oracle is proposed that takes O(sdn) space, where s is empirically estimated to be around 12 for road networks, but that can retrieve an intermediate link in a shortest path in O(logn) time using a B-tree. An additional construct termed the path-distance oracle of size O(n {\textperiodcentered} max(sd, 1/εd)) (empirically (n {\textperiodcentered} max(122, 2.5/ε2))) is proposed that can retrieve an intermediate vertex as well as an ε-approximation of the network distances in O(logn) time using a B-tree. Experimental results indicate that the proposed oracles are linear in n which means that they are scalable and can enable complicated query processing scenarios on massive spatial network datasets.}, isbn = {2150-8097}, url = {http://dl.acm.org/citation.cfm?id=1687627.1687763}, author = {Sankaranarayanan,Jagan and Samet, Hanan and Alborzi,Houman} } @inbook {18916, title = {Planning for Interactions among Autonomous Agents}, booktitle = {Programming Multi-Agent Systems}, series = {Lecture Notes in Computer Science}, volume = {5442}, year = {2009}, month = {2009///}, pages = {1 - 23}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {AI planning research has traditionally focused on offline pl- anning for static single-agent environments. In environments where an agent needs to plan its interactions with other autonomous agents, planning is much more complicated, because the actions of the other agents can induce a combinatorial explosion in the number of contingencies that the planner will need to consider. This paper discusses several ways to alleviate the combinatorial explosion, and illustrates their use in several different kinds of multi-agent planning domains.}, keywords = {Computer science}, isbn = {978-3-642-03277-6}, url = {http://www.springerlink.com/content/j258015ux2p38383/abstract/}, author = {Au,Tsz-Chiu and Kuter,Ugur and Nau, Dana S.}, editor = {Hindriks,Koen and Pokahr,Alexander and Sardina,Sebastian} } @article {14284, title = {Profile-guided static typing for dynamic scripting languages}, journal = {ACM SIGPLAN Notices}, volume = {44}, year = {2009}, month = {2009///}, pages = {283 - 300}, author = {Furr,M. and An,J.D. and Foster, Jeffrey S.} } @article {15081, title = {Proofs of storage from homomorphic identification protocols}, journal = {Advances in Cryptology{\textendash}ASIACRYPT 2009}, year = {2009}, month = {2009///}, pages = {319 - 333}, abstract = {Proofs of storage (PoS) are interactive protocols allowing a client to verify that a server faithfully stores a file. Previous work has shown that proofs of storage can be constructed from any homomorphic linear authenticator (HLA). The latter, roughly speaking, are signature/message authentication schemes where {\textquoteleft}tags{\textquoteright} on multiple messages can be homomorphically combined to yield a {\textquoteleft}tag{\textquoteright} on any linear combination of these messages.We provide a framework for building public-key HLAs from any identification protocol satisfying certain homomorphic properties. We then show how to turn any public-key HLA into a publicly-verifiable PoS with communication complexity independent of the file length and supporting an unbounded number of verifications. We illustrate the use of our transformations by applying them to a variant of an identification protocol by Shoup, thus obtaining the first unbounded-use PoS based on factoring (in the random oracle model). }, doi = {10.1007/978-3-642-10366-7_19}, author = {Ateniese,G. and Kamara,S. and Katz, Jonathan} } @conference {17636, title = {On random sampling auctions for digital goods}, booktitle = {Proceedings of the 10th ACM conference on Electronic commerce}, series = {EC {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {187 - 196}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {In the context of auctions for digital goods, an interesting Random Sampling Optimal Price auction (RSOP) has been proposed by Goldberg, Hartline and Wright; this leads to a truthful mechanism. Since random sampling is a popular approach for auctions that aims to maximize the seller{\textquoteright}s revenue, this method has been analyzed further by Feige, Flaxman, Hartline and Kleinberg, who have shown that it is 15-competitive in the worst case -- which is substantially better than the previously proved bounds but still far from the conjectured competitive ratio of 4. In this paper, we prove that RSOP is indeed 4-competitive for a large class of instances in which the number λ of bidders receiving the item at the optimal uniform price, is at least 6. We also show that it is 4.68 competitive for the small class of remaining instances thus leaving a negligible gap between the lower and upper bound. Furthermore, we develop a robust version of RSOP -- one in which the seller{\textquoteright}s revenue is, with high probability, not much below its mean -- when the above parameter λ grows large. We employ a mix of probabilistic techniques and dynamic programming to compute these bounds.}, keywords = {auction, mechanism design, random sampling}, isbn = {978-1-60558-458-4}, doi = {10.1145/1566374.1566402}, url = {http://doi.acm.org/10.1145/1566374.1566402}, author = {Alaei,Saeed and Malekian,Azarakhsh and Srinivasan, Aravind} } @conference {14251, title = {Real-time shape retrieval for robotics using skip Tri-Grams}, booktitle = {IEEE/RSJ International Conference on Intelligent Robots and Systems, 2009. IROS 2009}, year = {2009}, month = {2009/10/10/15}, pages = {4731 - 4738}, publisher = {IEEE}, organization = {IEEE}, abstract = {The real time requirement is an additional constraint on many intelligent applications in robotics, such as shape recognition and retrieval using a mobile robot platform. In this paper, we present a scalable approach for efficiently retrieving closed contour shapes. The contour of an object is represented by piecewise linear segments. A skip Tri-Gram is obtained by selecting three segments in the clockwise order while allowing a constant number of segments to be {\~A}‚{\^A}{\textquestiondown}skipped{\~A}‚{\^A}{\textquestiondown} in between. The main idea is to use skip Tri-Grams of the segments to implicitly encode the distant dependency of the shape. All skip Tri-Grams are used for efficiently retrieving closed contour shapes without pairwise matching feature points from two shapes. The retrieval is at least an order of magnitude faster than other state-of-the-art algorithms. We score 80\% in the Bullseye retrieval test on the whole MPEG 7 shape dataset. We further test the algorithm using a mobile robot platform in an indoor environment. 8 objects are used for testing from different viewing directions, and we achieve 82\% accuracy.}, keywords = {Bullseye retrieval test, Clocks, closed contour shape retrieval, Image retrieval, Image segmentation, Indexing, Information retrieval, Intelligent robots, Jacobian matrices, mobile robot, Mobile robots, MPEG 7 shape dataset, piecewise linear segments, Piecewise linear techniques, Real time systems, real-time shape retrieval, robot vision, SHAPE, shape recognition, shape representation, skip Tri-Grams, Testing}, isbn = {978-1-4244-3803-7}, doi = {10.1109/IROS.2009.5354738}, author = {Li,Yi and Bitsakos,K. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {12773, title = {Recovering Views of Inter-System Interaction Behaviors}, booktitle = {Reverse Engineering, Working Conference on}, year = {2009}, month = {2009///}, pages = {53 - 61}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {This paper presents an approach for recovering application-level views of the interaction behaviors between systems that communicate via networks. Rather than illustrating a single behavior, a sequence diagram is constructed that describes the characteristics of multiple combined behaviors. The approach has several properties that make it particularly suitable for analyzing heterogeneous systems. First, since the interactions are retrieved from observing the network communication, our technique can be applied to systems that are implemented in different languages and run on different platforms. Second, it does not require the availability or modification of source code. After the behaviors are extracted, we employ methods to merge multiple observed behaviors to a single sequence diagram that illustrates the overall behavior.The contributions of this paper are a technique for observing and processing the network communication to derive a model of the behavior. Furthermore, it describes a series of model transformations to construct a sequence diagram view of all observed behaviors.}, keywords = {distributed systems, program understanding, reliability}, doi = {http://doi.ieeecomputersociety.org/10.1109/WCRE.2009.34}, author = {Ackermann,Christopher and Lindvall,Mikael and Cleaveland, Rance} } @article {12884, title = {RNA Colony Blot Hybridization Method for Enumeration of Culturable Vibrio Cholerae and Vibrio Mimicus Bacteria}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {75}, year = {2009}, month = {2009/09/01/}, pages = {5439 - 5444}, abstract = {A species-specific RNA colony blot hybridization protocol was developed for enumeration of culturable Vibrio cholerae and Vibrio mimicus bacteria in environmental water samples. Bacterial colonies on selective or nonselective plates were lysed by sodium dodecyl sulfate, and the lysates were immobilized on nylon membranes. A fluorescently labeled oligonucleotide probe targeting a phylogenetic signature sequence of 16S rRNA of V. cholerae and V. mimicus was hybridized to rRNA molecules immobilized on the nylon colony lift blots. The protocol produced strong positive signals for all colonies of the 15 diverse V. cholerae-V. mimicus strains tested, indicating 100\% sensitivity of the probe for the targeted species. For visible colonies of 10 nontarget species, the specificity of the probe was calculated to be 90\% because of a weak positive signal produced by Grimontia (Vibrio) hollisae, a marine bacterium. When both the sensitivity and specificity of the assay were evaluated using lake water samples amended with a bioluminescent V. cholerae strain, no false-negative or false-positive results were found, indicating 100\% sensitivity and specificity for culturable bacterial populations in freshwater samples when G. hollisae was not present. When the protocol was applied to laboratory microcosms containing V. cholerae attached to live copepods, copepods were found to carry approximately 10,000 to 50,000 CFU of V. cholerae per copepod. The protocol was also used to analyze pond water samples collected in an area of cholera endemicity in Bangladesh over a 9-month period. Water samples collected from six ponds demonstrated a peak in abundance of total culturable V. cholerae bacteria 1 to 2 months prior to observed increases in pathogenic V. cholerae and in clinical cases recorded by the area health clinic. The method provides a highly specific and sensitive tool for monitoring the dynamics of V. cholerae in the environment. The RNA blot hybridization protocol can also be applied to detection of other gram-negative bacteria for taxon-specific enumeration.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.02007-08}, url = {http://aem.asm.org/content/75/17/5439}, author = {Grim,Christopher J. and Zo,Young-Gun and Hasan,Nur A. and Ali,Afsar and Chowdhury,Wasimul B. and Islam,Atiqul and Rashid,Mohammed H. and Alam,Munirul and Morris,J. Glenn and Huq,Anwar and Rita R Colwell} } @article {12520, title = {Robust Estimation of Albedo for Illumination-Invariant Matching and Shape Recovery}, journal = {Pattern Analysis and Machine Intelligence, IEEE Transactions on}, volume = {31}, year = {2009}, month = {2009/05//}, pages = {884 - 899}, abstract = {We present a nonstationary stochastic filtering framework for the task of albedo estimation from a single image. There are several approaches in the literature for albedo estimation, but few include the errors in estimates of surface normals and light source direction to improve the albedo estimate. The proposed approach effectively utilizes the error statistics of surface normals and illumination direction for robust estimation of albedo, for images illuminated by single and multiple light sources. The albedo estimate obtained is subsequently used to generate albedo-free normalized images for recovering the shape of an object. Traditional shape-from-shading (SFS) approaches often assume constant/piecewise constant albedo and known light source direction to recover the underlying shape. Using the estimated albedo, the general problem of estimating the shape of an object with varying albedo map and unknown illumination source is reduced to one that can be handled by traditional SFS approaches. Experimental results are provided to show the effectiveness of the approach and its application to illumination-invariant matching and shape recovery. The estimated albedo maps are compared with the ground truth. The maps are used as illumination-invariant signatures for the task of face recognition across illumination variations. The recognition results obtained compare well with the current state-of-the-art approaches. Impressive shape recovery results are obtained using images downloaded from the Web with little control over imaging conditions. The recovered shapes are also used to synthesize novel views under novel illumination conditions.}, keywords = {albedo estimation;error statistics;face recognition;illumination-invariant matching;nonstationary stochastic filtering;robust estimation;shape recovery;shape-from-shading approach;error statistics;face recognition;filtering theory;image matching;Algorithm, Automated;Photometry;Reproducibility of Results;Sensitivity and Specificity;, Computer-Assisted;Imaging, Three-Dimensional;Lighting;Pattern Recognition}, isbn = {0162-8828}, doi = {10.1109/TPAMI.2008.135}, author = {Biswas,S. and Aggarwal,G. and Chellapa, Rama} } @article {14743, title = {The ruby intermediate language}, journal = {SIGPLAN Not.}, volume = {44}, year = {2009}, month = {2009/10//}, pages = {89 - 98}, abstract = {Ruby is a popular, dynamic scripting language that aims to "feel natural to programmers" and give users the "freedom to choose" among many different ways of doing the same thing. While this arguably makes programming in Ruby easier, it makes it hard to build analysis and transformation tools that operate on Ruby source code. In this paper, we present the Ruby Intermediate Language (RIL), a Ruby front-end and intermediate representation that addresses these. RIL includes an extensible GLR parser for Ruby, and an automatic translation into an easy-to-analyze intermediate form. This translation eliminates redundant language constructs, unravels the often subtle ordering among side effecting operations, and makes implicit interpreter operations explicit. We also describe several additional useful features of RIL, such as a dynamic instrumentation library for profiling source code and a dataflow analysis engine. We demonstrate the usefulness of RIL by presenting a static and dynamic analysis to eliminate null pointer errors in Ruby programs. We hope that RIL{\textquoteright}s features will enable others to more easily build analysis tools for Ruby, and that our design will inspire the of similar frameworks for other dynamic languages.}, keywords = {intermediate language, profile guided anlaysis, ril, ruby}, isbn = {0362-1340}, doi = {10.1145/1837513.1640148}, url = {http://doi.acm.org/10.1145/1837513.1640148}, author = {Furr,Michael and An,Jong-hoon (David) and Foster, Jeffrey S. and Hicks, Michael W.} } @article {17655, title = {Scheduling on Unrelated Machines under Tree-Like Precedence Constraints}, journal = {Algorithmica}, volume = {55}, year = {2009}, month = {2009///}, pages = {205 - 226}, abstract = {We present polylogarithmic approximations for the R | prec | C max and R | prec |∑ j w j C j problems, when the precedence constraints are {\textquotedblleft}treelike{\textquotedblright}{\textemdash}i.e., when the undirected graph underlying the precedences is a forest. These are the first non-trivial generalizations of the job shop scheduling problem to scheduling with precedence constraints that are not just chains. These are also the first non-trivial results for the weighted completion time objective on unrelated machines with precedence constraints of any kind . We obtain improved bounds for the weighted completion time and flow time for the case of chains with restricted assignment{\textemdash}this generalizes the job shop problem to these objective functions. We use the same lower bound of {\textquotedblleft}congestion + dilation{\textquotedblright}, as in other job shop scheduling approaches (e.g. Shmoys, Stein and Wein, SIAM J. Comput. 23, 617{\textendash}632, 1994 ). The first step in our algorithm for the R | prec | C max problem with treelike precedences involves using the algorithm of Lenstra, Shmoys and Tardos to obtain a processor assignment with the congestion + dilation value within a constant factor of the optimal. We then show how to generalize the random-delays technique of Leighton, Maggs and Rao to the case of trees. For the special case of chains, we show a dependent rounding technique which leads to a bicriteria approximation algorithm for minimizing the flow time, a notoriously hard objective function.}, isbn = {0178-4617}, url = {http://dx.doi.org/10.1007/s00453-007-9004-y}, author = {Anil Kumar,V. and Marathe,Madhav and Parthasarathy,Srinivasan and Srinivasan, Aravind} } @article {11958, title = {Sensory grammars for sensor networks}, journal = {Journal of Ambient Intelligence and Smart Environments}, volume = {1}, year = {2009}, month = {2009///}, pages = {15 - 21}, author = {Aloimonos, J.} } @article {12882, title = {Serogroup, Virulence, and Genetic Traits of Vibrio Parahaemolyticus in the Estuarine Ecosystem of Bangladesh}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {75}, year = {2009}, month = {2009/10/01/}, pages = {6268 - 6274}, abstract = {Forty-two strains of Vibrio parahaemolyticus were isolated from Bay of Bengal estuaries and, with two clinical strains, analyzed for virulence, phenotypic, and molecular traits. Serological analysis indicated O8, O3, O1, and K21 to be the major O and K serogroups, respectively, and O8:K21, O1:KUT, and O3:KUT to be predominant. The K antigen(s) was untypeable, and pandemic serogroup O3:K6 was not detected. The presence of genes toxR and tlh were confirmed by PCR in all but two strains, which also lacked toxR. A total of 18 (41\%) strains possessed the virulence gene encoding thermostable direct hemolysin (TDH), and one had the TDH-related hemolysin (trh) gene, but not tdh. Ten (23\%) strains exhibited Kanagawa phenomenon that surrogates virulence, of which six, including the two clinical strains, possessed tdh. Of the 18 tdh-positive strains, 17 (94\%), including the two clinical strains, had the seromarker O8:K21, one was O9:KUT, and the single trh-positive strain was O1:KUT. None had the group-specific or ORF8 pandemic marker gene. DNA fingerprinting employing pulsed-field gel electrophoresis (PFGE) of SfiI-digested DNA and cluster analysis showed divergence among the strains. Dendrograms constructed using PFGE (SfiI) images from a soft database, including those of pandemic and nonpandemic strains of diverse geographic origin, however, showed that local strains formed a cluster, i.e., {\textquotedblleft}clonal cluster,{\textquotedblright} as did pandemic strains of diverse origin. The demonstrated prevalence of tdh-positive and diarrheagenic serogroup O8:K21 strains in coastal villages of Bangladesh indicates a significant human health risk for inhabitants.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.00266-09}, url = {http://aem.asm.org/content/75/19/6268}, author = {Alam,Munirul and Chowdhury,Wasimul B. and Bhuiyan,N. A. and Islam,Atiqul and Hasan,Nur A. and Nair,G. Balakrish and Watanabe,H. and Siddique,A. K. and Huq,Anwar and Sack,R. Bradley and Akhter,M. Z. and Grim,Christopher J. and Kam,K.-M. and Luey,C. K. Y. and Endtz,Hubert P. and Cravioto,Alejandro and Rita R Colwell} } @article {15610, title = {Space-time tradeoffs for approximate nearest neighbor searching}, journal = {Journal of the ACM (JACM)}, volume = {57}, year = {2009}, month = {2009/11//}, pages = {1:1{\textendash}1:54 - 1:1{\textendash}1:54}, abstract = {Nearest neighbor searching is the problem of preprocessing a set of n point points in d-dimensional space so that, given any query point q, it is possible to report the closest point to q rapidly. In approximate nearest neighbor searching, a parameter ϵ > 0 is given, and a multiplicative error of (1 + ϵ) is allowed. We assume that the dimension d is a constant and treat n and ϵ as asymptotic quantities. Numerous solutions have been proposed, ranging from low-space solutions having space O(n) and query time O(log n + 1/ϵd-1) to high-space solutions having space roughly O((n log n)/ϵd) and query time O(log (n/ϵ)). We show that there is a single approach to this fundamental problem, which both improves upon existing results and spans the spectrum of space-time tradeoffs. Given a tradeoff parameter γ, where 2 <= γ <= 1/ϵ, we show that there exists a data structure of space O(nγd-1 log(1/ϵ)) that can answer queries in time O(log(nγ) + 1/(ϵγ)(d-1)/2. When γ = 2, this yields a data structure of space O(n log (1/ϵ)) that can answer queries in time O(log n + 1/ϵ(d-1)/2). When γ = 1/ϵ, it provides a data structure of space O((n/ϵd-1)log(1/ϵ)) that can answer queries in time O(log(n/ϵ)). Our results are based on a data structure called a (t,ϵ)-AVD, which is a hierarchical quadtree-based subdivision of space into cells. Each cell stores up to t representative points of the set, such that for any query point q in the cell at least one of these points is an approximate nearest neighbor of q. We provide new algorithms for constructing AVDs and tools for analyzing their total space requirements. We also establish lower bounds on the space complexity of AVDs, and show that, up to a factor of O(log (1/ϵ)), our space bounds are asymptotically tight in the two extremes, γ = 2 and γ = 1/ϵ.}, keywords = {nearest neighbor searching, space-time tradeoffs}, isbn = {0004-5411}, doi = {10.1145/1613676.1613677}, url = {http://doi.acm.org/10.1145/1613676.1613677}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @conference {14738, title = {Static type inference for Ruby}, booktitle = {Proceedings of the 2009 ACM symposium on Applied Computing}, series = {SAC {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {1859 - 1866}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Many general-purpose, object-oriented scripting languages are dynamically typed, which provides flexibility but leaves the programmer without the benefits of static typing, including early error detection and the documentation provided by type annotations. This paper describes Diamondback Ruby (DRuby), a tool that blends Ruby{\textquoteright}s dynamic type system with a static typing discipline. DRuby provides a type language that is rich enough to precisely type Ruby code we have encountered, without unneeded complexity. When possible, DRuby infers static types to discover type errors in Ruby programs. When necessary, the programmer can provide DRuby with annotations that assign static types to dynamic code. These annotations are checked at run time, isolating type errors to unverified code. We applied DRuby to a suite of benchmarks and found several bugs that would cause run-time type errors. DRuby also reported a number of warnings that reveal questionable programming practices in the benchmarks. We believe that DRuby takes a major step toward bringing the benefits of combined static and dynamic typing to Ruby and other object-oriented languages.}, keywords = {Contracts, dynamic typing, ruby, Type inference}, isbn = {978-1-60558-166-8}, doi = {10.1145/1529282.1529700}, url = {http://doi.acm.org/10.1145/1529282.1529700}, author = {Furr,Michael and An,Jong-hoon (David) and Foster, Jeffrey S. and Hicks, Michael W.} } @article {12539, title = {Surface reconstruction from gradient fields via gradient transformations}, journal = {International Journal of Computer Vision}, year = {2009}, month = {2009///}, author = {Agrawal,A. and Raskar, R. and Chellapa, Rama} } @article {11909, title = {System and Method for Spatio-Temporal-Context Aware Interaction of Users ...}, volume = {12/267,921}, year = {2009}, month = {2009/05/14/}, abstract = {A multifunctional interaction system which is capable of spatio-temporal context localization of users and of communication of audio/video streams to an entity of interest defined by the user, includes a communication domain supporting a predefined localization service, a server associated with the communication domain, client devices, and a dynamically changing context database which is customized in accord with the dynamics of interaction sessions of client devices with the server. The client communicates with the system to either request services therefrom or to send a message to the entity of interest. The system is provided with a panic alert mechanism which, upon actuation, transmits an audio/video data stream along with the client location tag, time stamp, and client ID, to a police precinct for prompt action.}, url = {http://www.google.com/patents?id=nZbGAAAAEBAJ}, author = {Agrawala, Ashok K. and Varshney, Amitabh and Almazan,Christian B.}, editor = {University of Maryland} } @conference {14740, title = {Tests to the left of me, types to the right: how not to get stuck in the middle of a ruby execution}, booktitle = {Proceedings for the 1st workshop on Script to Program Evolution}, series = {STOP {\textquoteright}09}, year = {2009}, month = {2009///}, pages = {14 - 16}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Ruby is a popular dynamic scripting language that permits terse, expressive code, but provides no static checks to detect errors before running the program. To address this, we have developed Diamond-back Ruby (DRuby), a tool that blends the benefits of static and dynamic typing. This paper briefly describes the main features of DRuby, which we will present in a tool demonstration. The presentation will concentrate on the development of a small, statically typed Ruby program, illustrating how DRuby might be used in practice. The audience will learn about some of the practical design decisions that went into DRuby, and how to use it to develop a type-safe Ruby program.}, isbn = {978-1-60558-543-7}, doi = {10.1145/1570506.1570508}, url = {http://doi.acm.org/10.1145/1570506.1570508}, author = {Furr,Michael and An,Jong-hoon (David) and Foster, Jeffrey S. and Hicks, Michael W.} } @conference {18884, title = {Translating HTNs to PDDL: A small amount of domain knowledge can go a long way}, year = {2009}, month = {2009///}, pages = {1629 - 1634}, abstract = {We show how to translate HTN domain descrip- tions (if they satisfy certain restrictions) into PDDL so that they can be used by classical planners. We provide correctness results for our translation al- gorithm, and show that it runs in linear time and space. We also show that even small and incom- plete amounts of HTN knowledge, when translated into PDDL using our algorithm, can greatly im- prove a classical planner{\textquoteright}s performance. In ex- periments on several thousand randomly gener- ated problems in three different planning domains, such knowledge speeded up the well-known Fast- Forward planner by several orders of magnitude, and enabled it to solve much larger problems than it could otherwise solve.}, url = {https://www.aaai.org/ocs/index.php/IJCAI/IJCAI-09/paper/download/573/853}, author = {Alford,R. and Kuter,U. and Nau, Dana S.} } @conference {12806, title = {Validating Automotive Control Software Using Instrumentation-Based Verification}, booktitle = {Automated Software Engineering, International Conference on}, year = {2009}, month = {2009///}, pages = {15 - 25}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {This paper discusses the results of an application of a formally based verification technique, called Instrumentation-Based Verification (IBV), to a production automotive lighting controller. The goal of the study is to assess, from both a tools as well as a methodological perspective, the performance of IBV in an industrial setting. The insights obtained as a result of the project include a refinement of a previously developed architecture for requirements specifications; observations about changes to model-based design workflows; insights into the role of requirements during development; and the capability of automated verification to detect inconsistencies among requirements as well as between requirements and design models.}, doi = {http://doi.ieeecomputersociety.org/10.1109/ASE.2009.98}, author = {Ray,Arnab and Morschhaeuser,Iris and Ackermann,Chris and Cleaveland, Rance and Shelton,Charles and Martin,Chris} } @article {17491, title = {Visual overviews for discovering key papers and influences across research fronts}, journal = {Journal of the American Society for Information Science and Technology}, volume = {60}, year = {2009}, month = {2009/11/01/}, pages = {2219 - 2228}, abstract = {Gaining a rapid overview of an emerging scientific topic, sometimes called research fronts, is an increasingly common task due to the growing amount of interdisciplinary collaboration. Visual overviews that show temporal patterns of paper publication and citation links among papers can help researchers and analysts to see the rate of growth of topics, identify key papers, and understand influences across subdisciplines. This article applies a novel network-visualization tool based on meaningful layouts of nodes to present research fronts and show citation links that indicate influences across research fronts. To demonstrate the value of two-dimensional layouts with multiple regions and user control of link visibility, we conducted a design-oriented, preliminary case study with 6 domain experts over a 4-month period. The main benefits were being able (a) to easily identify key papers and see the increasing number of papers within a research front, and (b) to quickly see the strength and direction of influence across related research fronts.}, isbn = {1532-2890}, doi = {10.1002/asi.21160}, url = {http://onlinelibrary.wiley.com/doi/10.1002/asi.21160/abstract}, author = {Aris,Aleks and Shneiderman, Ben and Qazvinian,Vahed and Radev,Dragomir} } @conference {13697, title = {Voronoi++: ADynamic Page Segmentation approach based on Voronoi and Docstrum features}, booktitle = {International Conference on Document Analysis and Recognition (ICDAR {\textquoteright}09)}, year = {2009}, month = {2009///}, pages = {1011 - 1015}, abstract = {This paper presents a dynamic approach to document page segmentation. Current page segmentation algorithms lack the ability to dynamically adapt local variations in the size, orientation and distance of components within a page. Our approach builds upon one of the best algorithms, Kise et. al. work based on Area Voronoi Diagrams, which adapts globally to page content to determine algorithm parameters. In our approach, local thresholds are determined dynamically based on parabolic relations between components, and Docstrum based angular and neighborhood features are integrated to improve accuracy. Zone-based evaluation was performed on four sets of printed and handwritten documents in English and Arabic scripts and an increase of 33\% in accuracy is reported.}, author = {Agrawal,Mudit and David Doermann} } @article {15998, title = {What puts the {\textquotedblleft}meta{\textquotedblright} in metacognition?}, journal = {Behavioral and Brain Sciences}, volume = {32}, year = {2009}, month = {2009///}, pages = {138 - 139}, author = {Anderson,M. L and Perlis, Don} } @article {18544, title = {Accountable internet protocol (aip)}, journal = {SIGCOMM Comput. Commun. Rev.}, volume = {38}, year = {2008}, month = {2008/08//}, pages = {339 - 350}, abstract = {This paper presents AIP (Accountable Internet Protocol), a network architecture that provides accountability as a first-order property. AIP uses a hierarchy of self-certifying addresses, in which each component is derived from the public key of the corresponding entity. We discuss how AIP enables simple solutions to source spoofing, denial-of-service, route hijacking, and route forgery. We also discuss how AIP{\textquoteright}s design meets the challenges of scaling, key management, and traffic engineering.}, keywords = {accountability, address, internet architecture, scalability, Security}, isbn = {0146-4833}, doi = {10.1145/1402946.1402997}, url = {http://doi.acm.org/10.1145/1402946.1402997}, author = {Andersen,David G. and Balakrishnan,Hari and Feamster, Nick and Koponen,Teemu and Moon,Daekyeong and Shenker,Scott} } @article {15954, title = {Active logic semantics for a single agent in a static world}, journal = {Artificial Intelligence}, volume = {172}, year = {2008}, month = {2008/05//}, pages = {1045 - 1063}, abstract = {For some time we have been developing, and have had significant practical success with, a time-sensitive, contradiction-tolerant logical reasoning engine called the active logic machine (ALMA). The current paper details a semantics for a general version of the underlying logical formalism, active logic. Central to active logic are special rules controlling the inheritance of beliefs in general (and of beliefs about the current time in particular), very tight controls on what can be derived from direct contradictions (P\&{\textlnot}P), and mechanisms allowing an agent to represent and reason about its own beliefs and past reasoning. Furthermore, inspired by the notion that until an agent notices that a set of beliefs is contradictory, that set seems consistent (and the agent therefore reasons with it as if it were consistent), we introduce an "apperception function" that represents an agent{\textquoteright}s limited awareness of its own beliefs, and serves to modify inconsistent belief sets so as to yield consistent sets. Using these ideas, we introduce a new definition of logical consequence in the context of active logic, as well as a new definition of soundness such that, when reasoning with consistent premises, all classically sound rules remain sound in our new sense. However, not everything that is classically sound remains sound in our sense, for by classical definitions, all rules with contradictory premises are vacuously sound, whereas in active logic not everything follows from a contradiction.}, keywords = {Active logic, Autonomous agents, Brittleness, Logic, Nonmonotonic logic, Paraconsistent logic, semantics, Soundness, TIME}, isbn = {0004-3702}, doi = {16/j.artint.2007.11.005}, url = {http://www.sciencedirect.com/science/article/pii/S0004370207001993}, author = {Anderson,Michael L. and Gomaa,Walid and Grant,John and Perlis, Don} } @article {12562, title = {Activity representation using 3D shape models}, journal = {J. Image Video Process.}, volume = {2008}, year = {2008}, month = {2008/01//}, pages = {5:1{\textendash}5:16 - 5:1{\textendash}5:16}, abstract = {We present a method for characterizing human activities using 3D deformable shape models. The motion trajectories of points extracted from objects involved in the activity are used to build models for each activity, and these models are used for classification and detection of unusual activities. The deformable models are learnt using the factorization theorem for nonrigid 3D models. We present a theory for characterizing the degree of deformation in the 3D models from a sequence of tracked observations. This degree, termed as deformation index (DI), is used as an input to the 3D model estimation process. We study the special case of ground plane activities in detail because of its importance in video surveillance applications. We present results of our activity modeling approach using videos of both high-resolution single individual activities and ground plane surveillance activities.}, isbn = {1687-5176}, doi = {10.1155/2008/347050}, url = {http://dx.doi.org/10.1155/2008/347050}, author = {Abdelkader,Mohamed F. and Roy-Chowdhury,Amit K. and Chellapa, Rama and Akdemir,Umut} } @conference {18546, title = {Answering what-if deployment and configuration questions with wise}, booktitle = {Proceedings of the ACM SIGCOMM 2008 conference on Data communication}, series = {SIGCOMM {\textquoteright}08}, year = {2008}, month = {2008///}, pages = {99 - 110}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Designers of content distribution networks often need to determine how changes to infrastructure deployment and configuration affect service response times when they deploy a new data center, change ISP peering, or change the mapping of clients to servers. Today, the designers use coarse, back-of-the-envelope calculations, or costly field deployments; they need better ways to evaluate the effects of such hypothetical "what-if" questions before the actual deployments. This paper presents What-If Scenario Evaluator (WISE), a tool that predicts the effects of possible configuration and deployment changes in content distribution networks. WISE makes three contributions: (1) an algorithm that uses traces from existing deployments to learn causality among factors that affect service response-time distributions; (2) an algorithm that uses the learned causal structure to estimate a dataset that is representative of the hypothetical scenario that a designer may wish to evaluate, and uses these datasets to predict future response-time distributions; (3) a scenario specification language that allows a network designer to easily express hypothetical deployment scenarios without being cognizant of the dependencies between variables that affect service response times. Our evaluation, both in a controlled setting and in a real-world field deployment at a large, global CDN, shows that WISE can quickly and accurately predict service response-time distributions for many practical What-If scenarios.}, keywords = {content distribution networks, performance modeling, what-if scenario evaluation}, isbn = {978-1-60558-175-0}, doi = {10.1145/1402958.1402971}, url = {http://doi.acm.org/10.1145/1402958.1402971}, author = {Tariq,Mukarram and Zeitoun,Amgad and Valancius,Vytautas and Feamster, Nick and Ammar,Mostafa} } @article {12905, title = {Biofilms in water, its role and impact in human disease transmission}, journal = {Current Opinion in Biotechnology}, volume = {19}, year = {2008}, month = {2008/06//}, pages = {244 - 247}, abstract = {Understanding the mechanism of biofilm formation is the first step in determining its function and, thereby, its impact and role in the environment. Extensive studies accomplished during the past few years have elucidated the genetics and biochemistry of biofilm formation. Cell-to-cell communication, that is, quorum sensing, is a key factor in the initiation of biofilm. Occurrence of viable but nonculturable bacteria, including Vibrio cholerae in biofilms has been reported and most likely such cells were overlooked previously because appropriate methods of detection were not employed. For this reason discovery and investigation of this important bacterial ecological niche in the environment were impeded.}, isbn = {0958-1669}, doi = {10.1016/j.copbio.2008.04.005}, url = {http://www.sciencedirect.com/science/article/pii/S0958166908000505}, author = {Huq,Anwar and Whitehouse,Chris A. and Grim,Christopher J. and Alam,Munirul and Rita R Colwell} } @article {12549, title = {A Constrained Probabilistic Petri Net Framework for Human Activity Detection in Video}, journal = {Multimedia, IEEE Transactions on}, volume = {10}, year = {2008}, month = {2008/10//}, pages = {982 - 996}, abstract = {Recognition of human activities in restricted settings such as airports, parking lots and banks is of significant interest in security and automated surveillance systems. In such settings, data is usually in the form of surveillance videos with wide variation in quality and granularity. Interpretation and identification of human activities requires an activity model that a) is rich enough to handle complex multi-agent interactions, b) is robust to uncertainty in low-level processing and c) can handle ambiguities in the unfolding of activities. We present a computational framework for human activity representation based on Petri nets. We propose an extension-Probabilistic Petri Nets (PPN)-and show how this model is well suited to address each of the above requirements in a wide variety of settings. We then focus on answering two types of questions: (i) what are the minimal sub-videos in which a given activity is identified with a probability above a certain threshold and (ii) for a given video, which activity from a given set occurred with the highest probability? We provide the PPN-MPS algorithm for the first problem, as well as two different algorithms (naive PPN-MPA and PPN-MPA) to solve the second. Our experimental results on a dataset consisting of bank surveillance videos and an unconstrained TSA tarmac surveillance dataset show that our algorithms are both fast and provide high quality results.}, keywords = {activity, dataset;automated, detection;human, interactions;security, net;human, nets;image, Petri, probabilistic, processing;multiagent, processing;video, representation;low-level, representation;video, signal, Surveillance, surveillance;, systems;constrained, systems;surveillance, tarmac, TSA, videos;Petri}, isbn = {1520-9210}, doi = {10.1109/TMM.2008.2001369}, author = {Albanese, M. and Chellapa, Rama and Moscato, V. and Picariello, A. and V.S. Subrahmanian and Turaga,P. and Udrea,O.} } @article {12906, title = {Covariability of Vibrio Cholerae Microdiversity and Environmental Parameters}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {74}, year = {2008}, month = {2008/05/01/}, pages = {2915 - 2920}, abstract = {Fine-scale diversity of natural bacterial assemblages has been attributed to neutral radiation because correspondence between bacterial phylogenetic signals in the natural environment and environmental parameters had not been detected. Evidence that such correspondence occurs is provided for Vibrio cholerae, establishing a critical role for environmental parameters in bacterial diversity.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.02139-07}, url = {http://aem.asm.org/content/74/9/2915}, author = {Zo,Young-Gun and Chokesajjawatee,Nipa and Arakawa,Eiji and Watanabe,Haruo and Huq,Anwar and Rita R Colwell} } @article {12127, title = {CrossTalk: The Journal of Defense Software Engineering. Volume 21, Number 10, October 2008}, year = {2008}, month = {2008///}, institution = {DTIC Document}, author = {Basili, Victor R. and Dangle,K. and Esker,L. and Marotta,F. and Rus,I. and Brosgol,B. M and Jamin,S. and Arthur,J. D and Ravichandar,R. and Wisnosky,D. E} } @inbook {19030, title = {Delegating Capabilities in Predicate Encryption Systems}, booktitle = {Automata, Languages and Programming}, series = {Lecture Notes in Computer Science}, volume = {5126}, year = {2008}, month = {2008}, pages = {560 - 578}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {In predicate encryption systems, given a capability, one can evaluate one or more predicates on the plaintext encrypted, while all other information about the plaintext remains hidden. We consider the role of delegation in such predicate encryption systems. Suppose Alice has a capability, and she wishes to delegate to Bob a more restrictive capability allowing the decryption of a subset of the information Alice can learn about the plaintext encrypted. We formally define delegation in predicate encryption systems, propose a new security definition for delegation, and give an efficient construction supporting conjunctive queries. The security of our construction can be reduced to the general 3-party Bilinear Diffie-Hellman assumption, and the Bilinear Decisional Diffie-Hellman assumption in composite order bilinear groups.}, keywords = {Computer science}, isbn = {978-3-540-70582-6}, url = {http://www.springerlink.com/content/w320422h15050004/abstract/}, author = {Elaine Shi and Waters,Brent}, editor = {Aceto,Luca and Damg{\r a}rd,Ivan and Goldberg,Leslie and Halld{\'o}rsson,Magn{\'u}s and Ing{\'o}lfsd{\'o}ttir,Anna and Walukiewicz,Igor} } @article {18768, title = {Development of a multi-piece multi-gate mold for manufacturing a flapping wing drive-mechanism}, journal = {Transactions of North American Manufacturing Research Institution of SME}, volume = {36}, year = {2008}, month = {2008///}, abstract = {Successful realization a flapping wing microair vehicle (MAV) requires development of a light weight drive-mechanism that can convert the continuous rotary motion to oscillatory flapping motion. Molded compliant drive-mechanisms are an attractive design option because of manufacturing scalability and reduction in the number of parts. The unique characteristics of this mechanism require development of a complex multi-piece multi-gate mold design. This paper describes a systematic approach for determining the part shape and size, optimizing the mold pieces, and placing the gates on the multi-piece mold. The novel aspects of this work include (1) selecting non-critical shape features to optimize the mold design and (2) the use of sacrificial structural elements to reduce the impact of the weld lines on the structural performance. The mold developed using our approach was utilized to realize a working flapping wing MAV. }, url = {http://web.mit.edu/arvinda/www/NAMRI_2008_Draft.pdf}, author = {Ananthanarayanan,A. and Bejgerowski,W. and Mueller,D. and Gupta,S.K.} } @conference {13550, title = {Document Zone Classification Using Partial Least Squares and Hybrid Classifiers}, booktitle = {ICPR 2008. 19th International Conference on Pattern Recognition, 2008.}, year = {2008}, month = {2008///}, pages = {1 - 4}, abstract = {This paper introduces a novel document-zone classification algorithm. Low level image features are first extracted from document zones and partial least squares is used on pairs of classes to compute discriminating pairwise features. Rather than using the popular one-against-all and one-against-one voting schemes, we introduce a novel hybrid method which combines the benefits of the two schemes. The algorithm is applied on the University of Washington dataset and 97.3\% classification accuracy is obtained.}, author = {Abd-Almageed, Wael and Agrawal,Mudit and Seo,W. and David Doermann} } @conference {15585, title = {Enhancements in robust feature matching}, booktitle = {Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, 2008. ECTI-CON 2008. 5th International Conference on}, volume = {1}, year = {2008}, month = {2008/05//}, pages = {505 - 508}, abstract = {We present in this paper a number of new enhancements to a branch-and-bound algorithm given by Mount, Netanyahu, and Le Moigne [8] for feature matching. We introduce a new distance measure, which is more robust to the presence of outliers than the previously used measure, as well as a new variant of the search algorithm and a new search strategy. We also present experimental results, which show that these enhancements offer significant tangible improvements in performance.}, keywords = {branch-and-bound algorithm, distance measure, Image Enhancement, Image matching, robust feature matching, tree searching}, doi = {10.1109/ECTICON.2008.4600481}, author = {Ratanasanya,S. and Mount, Dave and Netanyahu,N. S and Achalakul,T.} } @article {16396, title = {Exurbia from the bottom-up: Confronting empirical challenges to characterizing a complex system}, journal = {Geoforum}, volume = {39}, year = {2008}, month = {2008/03//}, pages = {805 - 818}, abstract = {We describe empirical results from a multi-disciplinary project that support modeling complex processes of land-use and land-cover change in exurban parts of Southeastern Michigan. Based on two different conceptual models, one describing the evolution of urban form as a consequence of residential preferences and the other describing land-cover changes in an exurban township as a consequence of residential preferences, local policies, and a diversity of development types, we describe a variety of empirical data collected to support the mechanisms that we encoded in computational agent-based models. We used multiple methods, including social surveys, remote sensing, and statistical analysis of spatial data, to collect data that could be used to validate the structure of our models, calibrate their specific parameters, and evaluate their output. The data were used to investigate this system in the context of several themes from complexity science, including have (a) macro-level patterns; (b) autonomous decision making entities (i.e., agents); (c) heterogeneity among those entities; (d) social and spatial interactions that operate across multiple scales and (e) nonlinear feedback mechanisms. The results point to the importance of collecting data on agents and their interactions when producing agent-based models, the general validity of our conceptual models, and some changes that we needed to make to these models following data analysis. The calibrated models have been and are being used to evaluate landscape dynamics and the effects of various policy interventions on urban land-cover patterns.}, keywords = {Ecological effects, Land-cover change, Land-use change, spatial modeling, Urban sprawl}, isbn = {0016-7185}, doi = {10.1016/j.geoforum.2007.02.010}, url = {http://www.sciencedirect.com/science/article/pii/S0016718507000371}, author = {Brown,Daniel G. and Robinson,Derek T. and An,Li and Nassauer,Joan I. and Zellner,Moira and Rand, William and Riolo,Rick and Page,Scott E. and Low,Bobbi and Wang,Zhifang} } @inbook {19586, title = {Flow Analysis, Linearity, and PTIME}, booktitle = {Static Analysis}, series = {Lecture Notes in Computer Science}, year = {2008}, month = {2008/01/01/}, pages = {255 - 269}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, abstract = {Flow analysis is a ubiquitous and much-studied component of compiler technology{\textemdash}and its variations abound. Amongst the most well known is Shivers{\textquoteright} 0CFA; however, the best known algorithm for 0CFA requires time cubic in the size of the analyzed program and is unlikely to be improved. Consequently, several analyses have been designed to approximate 0CFA by trading precision for faster computation. Henglein{\textquoteright}s simple closure analysis, for example, forfeits the notion of directionality in flows and enjoys an {\textquotedblleft}almost linear{\textquotedblright} time algorithm. But in making trade-offs between precision and complexity, what has been given up and what has been gained? Where do these analyses differ and where do they coincide? We identify a core language{\textemdash}the linear λ-calculus{\textemdash}where 0CFA, simple closure analysis, and many other known approximations or restrictions to 0CFA are rendered identical. Moreover, for this core language, analysis corresponds with (instrumented) evaluation. Because analysis faithfully captures evaluation, and because the linear λ-calculus is complete for ptime, we derive ptime-completeness results for all of these analyses.}, keywords = {Logics and Meanings of Programs, Mathematical Logic and Formal Languages, Programming Languages, Compilers, Interpreters, Programming Techniques, software engineering}, isbn = {978-3-540-69163-1, 978-3-540-69166-2}, url = {http://link.springer.com/chapter/10.1007/978-3-540-69166-2_17}, author = {David Van Horn and Mairson, Harry G.}, editor = {Alpuente, Mar{\'\i}a and Vidal, Germ{\'a}n} } @conference {14331, title = {Flowers or a robot army?: encouraging awareness \& activity with personal, mobile displays}, booktitle = {Proceedings of the 10th international conference on Ubiquitous computing}, series = {UbiComp {\textquoteright}08}, year = {2008}, month = {2008///}, pages = {54 - 63}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Personal, mobile displays, such as those on mobile phones, are ubiquitous, yet for the most part, underutilized. We present results from a field experiment that investigated the effectiveness of these displays as a means for improving awareness of daily life (in our case, self-monitoring of physical activity). Twenty-eight participants in three experimental conditions used our UbiFit system for a period of three months in their day-to-day lives over the winter holiday season. Our results show, for example, that participants who had an awareness display were able to maintain their physical activity level (even during the holidays), while the level of physical activity for participants who did not have an awareness display dropped significantly. We discuss our results and their general implications for the use of everyday mobile devices as awareness displays.}, keywords = {ambient display, awareness, calm technology, field experiment, fitness, glanceable display, mobile phone, personal mobile display, persuasive technology, physical activity}, isbn = {978-1-60558-136-1}, doi = {10.1145/1409635.1409644}, url = {http://doi.acm.org/10.1145/1409635.1409644}, author = {Consolvo,Sunny and Klasnja,Predrag and McDonald,David W. and Avrahami,Daniel and Jon Froehlich and LeGrand,Louis and Libby,Ryan and Mosher,Keith and Landay,James A.} } @article {16271, title = {Genome-Wide Analysis of Repetitive Elements in Papaya}, journal = {Tropical Plant Biology}, volume = {1}, year = {2008}, month = {2008///}, pages = {191 - 201}, abstract = {Papaya ( Carica papaya L.) is an important fruit crop cultivated in tropical and subtropical regions worldwide. A first draft of its genome sequence has been recently released. Together with Arabidopsis , rice, poplar, grapevine and other genomes in the pipeline, it represents a good opportunity to gain insight into the organization of plant genomes. Here we report a detailed analysis of repetitive elements in the papaya genome, including transposable elements (TEs), tandemly-arrayed sequences, and high copy number genes. These repetitive sequences account for \~{}56\% of the papaya genome with TEs being the most abundant at 52\%, tandem repeats at 1.3\% and high copy number genes at 3\%. Most common types of TEs are represented in the papaya genome with retrotransposons being the dominant class, accounting for 40\% of the genome. The most prevalent retrotransposons are Ty3-gypsy (27.8\%) and Ty1-copia (5.5\%). Among the tandem repeats, microsatellites are the most abundant in number, but represent only 0.19\% of the genome. Minisatellites and satellites are less abundant, but represent 0.68\% and 0.43\% of the genome, respectively, due to greater repeat length. Despite an overall smaller gene repertoire in papaya than many other angiosperms, a significant fraction of genes (\>2\%) are present in large gene families with copy number greater than 20. This repeat database clarified a major part of the papaya genome organization and partly explained the lower gene repertoire in papaya than in Arabidopsis .}, isbn = {1935-9756}, url = {http://dx.doi.org/10.1007/s12042-008-9015-0}, author = {Nagarajan,Niranjan and Navajas-P{\'e}rez,Rafael and Pop, Mihai and Alam,Maqsudul and Ming,Ray and Paterson,Andrew and Salzberg,Steven} } @conference {19450, title = {Hci for community and international development}, booktitle = {SIGCHI EA {\textquoteright}08}, series = {CHI EA {\textquoteright}08}, year = {2008}, month = {2008///}, pages = {3909 - 3912}, publisher = {ACM}, organization = {ACM}, abstract = {This workshop explores the challenges in applying, extending and inventing appropriate methods and contributions of Human Computer Interaction (HCI) to International economic and community Development. We address interaction design for parts of the world that are often marginalized by the Global North as well as people in the Global North who are themselves similarly marginalized by poverty or other barriers. We hope to extend the boundaries of the field of Human Computer Interaction by spurring a discussion on how existing methods and practices can be adapted and modified, and how new practices can be developed, to deal with the unique challenges posed by these contexts.}, keywords = {community design, ict4d, information and communication technology, international development, participatory design, ucd4id, User centered design}, isbn = {978-1-60558-012-8}, url = {http://doi.acm.org/10.1145/1358628.1358954}, author = {Thomas, John and Dearden, Andy and Dray, Susan and Light, Ann and Best, Michael and Arkin, Nuray and Maunder, Andrew and Kam, Mathew and Marshini Chetty and Sambasivan, Nithya and Buckhalter, Celeste and Krishnan, Gaurishankar} } @conference {13124, title = {Human detection using iterative feature selection and logistic principal component analysis}, booktitle = {Robotics and Automation, 2008. ICRA 2008. IEEE International Conference on}, year = {2008}, month = {2008/05//}, pages = {1691 - 1697}, abstract = {We present a fast feature selection algorithm suitable for object detection applications where the image being tested must be scanned repeatedly to detected the object of interest at different locations and scales. The algorithm iteratively estimates the belongness probability of image pixels to foreground of the image. To prove the validity of the algorithm, we apply it to a human detection problem. The edge map is filtered using a feature selection algorithm. The filtered edge map is then projected onto an eigen space of human shapes to determine if the image contains a human. Since the edge maps are binary in nature, Logistic Principal Component Analysis is used to obtain the eigen human shape space. Experimental results illustrate the accuracy of the human detector.}, keywords = {algorithm;logistic, analysis;edge, analysis;probability;, applications;principal, belongness, component, DETECTION, detection;feature, detection;iterative, detection;principal, extraction;filtering, feature, filtering;human, MAP, methods;object, PCA;object, probability;edge, selection, theory;iterative}, doi = {10.1109/ROBOT.2008.4543444}, author = {Abd-Almageed, Wael and Davis, Larry S.} } @article {18532, title = {Learning To Predict Bad Behavior}, journal = {NIPS 2007 Workshop on Machine Learning in Adversarial Environments for Computer Security}, year = {2008}, month = {2008///}, author = {Ahmed Syed,N. and Feamster, Nick and Gray,A.} } @conference {13118, title = {A Logic Framework for Sports Video Summarization Using Text-Based Semantic Annotation}, booktitle = {Semantic Media Adaptation and Personalization, 2008. SMAP {\textquoteright}08. Third International Workshop on}, year = {2008}, month = {2008/12//}, pages = {69 - 75}, abstract = {Detection of semantic events in sports videos is an essential step towards video summarization. A large volume of research has been conducted for automatic semantic event detection and summarization of sports videos. In this paper we present a novel sports video summarization framework using a combination of text, video and logic analysis. Parse trees are used to analyze structured and free-style text webcasting of sports games and extract the game{\^A}{\textquestiondown}s semantic events, such as goals and penalties in soccer games. Semantic events are then hierarchically arranged before being passed to a logic processing engine. The logic engine receives the summary preferences from the user and subsequently parses the event hierarchy to generate the game{\^A}{\textquestiondown}s summary according to the user{\^A}{\textquestiondown}s preferences. The proposed framework was applied to both soccer and basketball videos. We achieved an average accuracy of 98.6\% and 100\% on soccer and basketball videos, respectively.}, keywords = {(mathematics);video, analysis;trees, annotation;Internet;broadcasting;sport;text, AUTOMATIC, detection;logic, engine;parse, event, PROCESSING, processing;, semantic, signal, summarization;text, trees;sports, video, Webcasting;text-based}, doi = {10.1109/SMAP.2008.25}, author = {Refaey,M.A. and Abd-Almageed, Wael and Davis, Larry S.} } @conference {14229, title = {Measuring 1st order stretchwith a single filter}, booktitle = {IEEE International Conference on Acoustics, Speech and Signal Processing, 2008. ICASSP 2008}, year = {2008}, month = {2008/04/31/March}, pages = {909 - 912}, publisher = {IEEE}, organization = {IEEE}, abstract = {We analytically develop a filter that is able to measure the linear stretch of the transformation around a point, and present results of applying it to real signals. We show that this method is a real-time alternative solution for measuring local signal transformations. Experimentally, this method can accurately measure stretch, however, it is sensitive to shift.}, keywords = {Cepstral analysis, Educational institutions, filter, filtering theory, Fourier transforms, Frequency domain analysis, Frequency estimation, Gabor filters, Image analysis, IMAGE PROCESSING, linear stretch measurement, local signal transformation measurement, Nonlinear filters, Phase estimation, Signal analysis, Speech processing}, isbn = {978-1-4244-1483-3}, doi = {10.1109/ICASSP.2008.4517758}, author = {Bitsakos,K. and Domke, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {12851, title = {Model Based Design Verification: A Monitor Based Approach}, volume = {2008-01-0741}, year = {2008}, month = {2008/04//}, institution = {SAE International}, address = {Warrendale, PA}, abstract = {This paper assesses the utility of an automated functional verification methodology in the design of a body electronics application. The use of design models as detailed specifications for software is gaining steady acceptance in the automotive industry because of the advantages it offers in terms of rapid system prototyping and automatic code generation. This paper argues that such modeling effort can also be used to provide support for design verification, so that errors may be pinpointed before they find their way into code, when they become much harder and more expensive to isolate, diagnose and repair.}, url = {http://papers.sae.org/2008-01-0741}, author = {Ackermann,Chris and Ray,Arnab and Cleaveland, Rance and Heit,J{\"u}rgen and Martin,Christopher and Shelton,Charles} } @conference {12556, title = {Multi-biometric cohort analysis for biometric fusion}, booktitle = {Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on}, year = {2008}, month = {2008/04/31/4}, pages = {5224 - 5227}, abstract = {Biometric matching decisions have traditionally been made based solely on a score that represents the similarity of the query biometric to the enrolled biometric(s) of the claimed identity. Fusion schemes have been proposed to benefit from the availability of multiple biometric samples (e.g., multiple samples of the same fingerprint) or multiple different biometrics (e.g., face and fingerprint). These commonly adopted fusion approaches rarely make use of the large number of non-matching biometric samples available in the database in the form of other enrolled identities or training data. In this paper, we study the impact of combining this information with the existing fusion methodologies in a cohort analysis framework. Experimental results are provided to show the usefulness of such a cohort-based fusion of face and fingerprint biometrics.}, keywords = {(access, analysis;biometrics, biometric, biometrics;fingerprint, biometrics;multi-biometric, cohort, control);face, data;, decisions;face, fusion;biometric, identification;security, MATCHING, of, recognition;fingerprint}, doi = {10.1109/ICASSP.2008.4518837}, author = {Aggarwal,G. and Ratha,N. K and Bolle, R.M. and Chellapa, Rama} } @article {12900, title = {New records of phytoplankton for Bangladesh. 2. Cryptophyceae and Synurophyceae}, journal = {Bangladesh Journal of Botany}, volume = {36}, year = {2008}, month = {2008/11/05/}, abstract = {This study presents two species of Rhodomonas, four species of Chroomonas, six species of Cryptomonas and Cryptochrysis minor, Cyanomonas coeruleus, Chrysodidymus synuroideus and Mallomonas akrokomos. These species have been reported from some ponds of Mathbaria in Pirojpur and Bakerganj of Barisal district in Bangladesh.}, isbn = {0253-5416}, doi = {10.3329/bjb.v36i1.1549}, url = {http://www.banglajol.info/bd/index.php/BJB/article/viewArticle/1549}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasmin,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @article {12902, title = {New records of phytoplankton for Bangladesh. 5. Euglena, Euglenocapsa}, journal = {Bangladesh Journal of Plant Taxonomy}, volume = {15}, year = {2008}, month = {2008/06/16/}, abstract = {This study presents 20 taxa of the genus Euglena and one species of the rare euglenoid genus Euglenocapsa. All these taxa are reported for the first time from some pond ecosystems of Mathbaria in Pirojpur and Bakerganj of Barisal districts of Bangladesh.}, isbn = {1028-2092}, doi = {10.3329/bjpt.v15i1.910}, url = {http://www.banglajol.info/index.php/BJPT/article/viewArticle/910}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasmin,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @article {12899, title = {New records of phytoplankton for Bangladesh. 7. Phacus spp.}, journal = {Bangladesh Journal of Botany}, volume = {37}, year = {2008}, month = {2008/11/05/}, abstract = {Thirteen species of Phacus hitherto not reported from Bangladesh have been described and illustrated. Freshwater ponds at southern districts of Pirojpur and Barisal revealed these presence of the species.}, isbn = {0253-5416}, doi = {10.3329/bjb.v37i1.1564}, url = {http://www.banglajol.info/bd/index.php/BJB/article/viewArticle/1564}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasmin,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @article {12897, title = {New records of phytoplankton for Bangladesh. 8. Trachelomonas Ehr. (Euglenophyceae)}, journal = {Bangladesh Journal of Botany}, volume = {37}, year = {2008}, month = {2008/12/19/}, abstract = {Investigation of pelagic plankton communities from some freshwater ponds of Pirojpur and Barisal districts revealed the presence of 17 species under the genus Trachelomonas Ehr. for the first time in Bangladesh.}, isbn = {0253-5416}, doi = {10.3329/bjb.v37i2.1719}, url = {http://www.banglajol.info/bd/index.php/BJB/article/viewArticle/1719}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasmin,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @article {16937, title = {A Node Aggregation Strategy to Reduce Complexity of Network Visualization using Semantic Substrates}, year = {2008}, month = {2008///}, institution = {Human-Computer Interaction Lab, University of Maryland}, abstract = {Semantic substrates are spatial templates for networks, wherenodes are grouped into regions and laid out within each region according to one or more node attributes. Analysts{\textquoteright} ability to design their own substrates leads to a different approach than other more automatic approaches to layout nodes (force-directed, circular, etc.). While the semantic substrate approach provides more interpretable node locations, sometimes a set of nodes is compressed into a small space on the display, leading to node overlap. In this paper, we significantly improve this situation by introducing the node aggregation mechanism in the context of semantic substrates. We illustrate this functionality in a document citation network and provide pros and cons of the approach. We conclude with guidelines and future directions for this research. Throughout the paper, examples are illustrated with NVSS 3.0, the network visualization tool developed to explore the semantic substrate idea. }, author = {Aris,A. and Shneiderman, Ben} } @conference {12548, title = {A Non-generative Approach for Face Recognition Across Aging}, booktitle = {Biometrics: Theory, Applications and Systems, 2008. BTAS 2008. 2nd IEEE International Conference on}, year = {2008}, month = {2008/10//undefin}, pages = {1 - 6}, abstract = {Human faces undergo a lot of change in appearance as they age. Though facial aging has been studied for decades, it is only recently that attempts have been made to address the problem from a computational point of view. Most of these early efforts follow a simulation approach in which matching is performed by synthesizing face images at the target age. Given the innumerable different ways in which a face can potentially age, the synthesized aged image may not be similar to the actual aged image. In this paper, we bypass the synthesis step and directly analyze the drifts of facial features with aging from a purely matching perspective. Our analysis is based on the observation that facial appearance changes in a coherent manner as people age. We provide measures to capture this coherency in feature drifts. Illustrations and experimental results show the efficacy of such an approach for matching faces across age progression.}, keywords = {appearance;nongenerative, approach;face, Face, image, matching;, recognition;facial, recognition;image, synthesis;face}, doi = {10.1109/BTAS.2008.4699331}, author = {Biswas,S. and Aggarwal,G. and Ramanathan,N. and Chellapa, Rama} } @article {12910, title = {Occurrence and Expression of Luminescence in Vibrio Cholerae}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {74}, year = {2008}, month = {2008/02/01/}, pages = {708 - 715}, abstract = {Several species of the genus Vibrio, including Vibrio cholerae, are bioluminescent or contain bioluminescent strains. Previous studies have reported that only 10\% of V. cholerae strains are luminescent. Analysis of 224 isolates of non-O1/non-O139 V. cholerae collected from Chesapeake Bay, MD, revealed that 52\% (116/224) were luminescent when an improved assay method was employed and 58\% (130/224) of isolates harbored the luxA gene. In contrast, 334 non-O1/non-O139 V. cholerae strains isolated from two rural provinces in Bangladesh yielded only 21 (6.3\%) luminescent and 35 (10.5\%) luxA+ isolates. An additional 270 clinical and environmental isolates of V. cholerae serogroups O1 and O139 were tested, and none were luminescent or harbored luxA. These results indicate that bioluminescence may be a trait specific for non-O1/non-O139 V. cholerae strains that frequently occur in certain environments. Luminescence expression patterns of V. cholerae were also investigated, and isolates could be grouped based on expression level. Several strains with defective expression of the lux operon, including natural K variants, were identified.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.01537-07}, url = {http://aem.asm.org/content/74/3/708}, author = {Grim,Christopher J. and Taviani,Elisa and Alam,Munirul and Huq,Anwar and Sack,R. Bradley and Rita R Colwell} } @conference {12570, title = {An ontology based approach for activity recognition from video}, booktitle = {Proceedings of the 16th ACM international conference on Multimedia}, series = {MM {\textquoteright}08}, year = {2008}, month = {2008///}, pages = {709 - 712}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Representation and recognition of human activities is an important problem for video surveillance and security applications. Considering the wide variety of settings in which surveillance systems are being deployed, it is necessary to create a common knowledge-base or ontology of human activities. Most current attempts at ontology design in computer vision for human activities have been empirical in nature. In this paper, we present a more systematic approach to address the problem of designing ontologies for visual activity recognition. We draw on general ontology design principles and adapt them to the specific domain of human activity ontologies. Then, we discuss qualitative evaluation principles and provide several examples from existing ontologies and how they can be improved upon. Finally, we demonstrate quantitatively in terms of recognition performance, the efficacy and validity of our approach for bank and airport tarmac surveillance domains.}, keywords = {activity ontologies, visual surveillance}, isbn = {978-1-60558-303-7}, doi = {10.1145/1459359.1459466}, url = {http://doi.acm.org/10.1145/1459359.1459466}, author = {Akdemir,Umut and Turaga,Pavan and Chellapa, Rama} } @inbook {19642, title = {Optimal Cryptographic Hardness of Learning Monotone Functions}, booktitle = {Automata, Languages and Programming}, series = {Lecture Notes in Computer Science}, year = {2008}, month = {2008/01/01/}, pages = {36 - 47}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, abstract = {A wide range of positive and negative results have been established for learning different classes of Boolean functions from uniformly distributed random examples. However, polynomial-time algorithms have thus far been obtained almost exclusively for various classes of monotone functions, while the computational hardness results obtained to date have all been for various classes of general (nonmonotone) functions. Motivated by this disparity between known positive results (for monotone functions) and negative results (for nonmonotone functions), we establish strong computational limitations on the efficient learnability of various classes of monotone functions. We give several such hardness results which are provably almost optimal since they nearly match known positive results. Some of our results show cryptographic hardness of learning polynomial-size monotone circuits to accuracy only slightly greater than 1/2+1/n--√1/2 + 1/\sqrt{n} ; this accuracy bound is close to optimal by known positive results (Blum et al., FOCS {\textquoteright}98). Other results show that under a plausible cryptographic hardness assumption, a class of constant-depth, sub-polynomial-size circuits computing monotone functions is hard to learn; this result is close to optimal in terms of the circuit size parameter by known positive results as well (Servedio, Information and Computation {\textquoteright}04). Our main tool is a complexity-theoretic approach to hardness amplification via noise sensitivity of monotone functions that was pioneered by O{\textquoteright}Donnell (JCSS {\textquoteright}04).}, keywords = {Data structures, Data Structures, Cryptology and Information Theory, Discrete Mathematics in Computer Science, Numeric Computing, Software Engineering/Programming and Operating Systems, Theory of computation}, isbn = {978-3-540-70574-1, 978-3-540-70575-8}, url = {http://link.springer.com/chapter/10.1007/978-3-540-70575-8_4}, author = {Dana Dachman-Soled and Lee, Homin K. and Malkin, Tal and Servedio, Rocco A. and Wan, Andrew and Wee, Hoeteck}, editor = {Aceto, Luca and Damg{\r a}rd, Ivan and Goldberg, Leslie Ann and Halld{\'o}rsson, Magn{\'u}s M. and Ing{\'o}lfsd{\'o}ttir, Anna and Walukiewicz, Igor} } @article {12568, title = {Pose and Illumination Issues in Face-and Gait-Based Identification}, journal = {Advances in Biometrics}, year = {2008}, month = {2008///}, pages = {307 - 322}, abstract = {Although significant work has been done in the field of face- and gait- based recognition, the performance of the state-of-the-art recognition algorithms is not good enough to be effective in operational systems. Most algorithms do reasonably well for controlled images but are susceptible to changes in illumination conditions and pose. This has shifted the focus of research to more challenging tasks of obtaining better performance for uncontrolled realistic scenarios. In this chapter, we discuss several recent advances made to achieve this goal.}, author = {Chellapa, Rama and Aggarwal,G.} } @inbook {17668, title = {The Randomized Coloring Procedure with Symmetry-Breaking}, booktitle = {Automata, Languages and ProgrammingAutomata, Languages and Programming}, series = {Lecture Notes in Computer Science}, volume = {5125}, year = {2008}, month = {2008///}, pages = {306 - 319}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {A basic randomized coloring procedure has been used in probabilistic proofs to obtain remarkably strong results on graph coloring. These results include the asymptotic version of the List Coloring Conjecture due to Kahn, the extensions of Brooks{\textquoteright} Theorem to sparse graphs due to Kim and Johansson, and Luby{\textquoteright}s fast parallel and distributed algorithms for graph coloring. The most challenging aspect of a typical probabilistic proof is showing adequate concentration bounds for key random variables. In this paper, we present a simple symmetry-breaking augmentation to the randomized coloring procedure that works well in conjunction with Azuma{\textquoteright}s Martingale Inequality to easily yield the requisite concentration bounds. We use this approach to obtain a number of results in two areas: frugal coloring and weighted equitable coloring . A β-frugal coloring of a graph G is a proper vertex-coloring of G in which no color appears more than β times in any neighborhood. Let G = ( V , E ) be a vertex-weighted graph with weight function w : V {\textrightarrow}[0, 1] and let W = ∑ v ∈ V w ( v ). A weighted equitable coloring of G is a proper k -coloring such that the total weight of every color class is {\textquotedblleft}large{\textquotedblright}, i.e., {\textquotedblleft}not much smaller{\textquotedblright} than W / k ; this notion is useful in obtaining tail bounds for sums of dependent random variables.}, isbn = {978-3-540-70574-1}, url = {http://dx.doi.org/10.1007/978-3-540-70575-8_26}, author = {Pemmaraju,Sriram and Srinivasan, Aravind}, editor = {Aceto,Luca and Damg{\r a}rd,Ivan and Goldberg,Leslie and Halld{\'o}rsson,Magn{\'u}s and Ing{\'o}lfsd{\'o}ttir,Anna and Walukiewicz,Igor} } @conference {12173, title = {Readability of scanned books in digital libraries}, booktitle = {Proceeding of the twenty-sixth annual SIGCHI conference on Human factors in computing systems}, year = {2008}, month = {2008///}, pages = {705 - 714}, author = {Quinn,A.J. and Hu,C. and Arisaka,T. and Rose,A. and Bederson, Benjamin B.} } @conference {13628, title = {Re-Targetable OCR with Intelligent Character Segmentation}, booktitle = {DAS {\textquoteright}08: Proceedings of the 2008 The Eighth IAPRInternational Workshop on Document Analysis Systems}, year = {2008}, month = {2008/09//}, pages = {183 - 190}, abstract = {We have developed a font-model based intelligent character segmentation and recognition system. Using characteristics of structurally similar TrueType fonts, our system automatically builds a model to be used for the segmentation and recognition of the new script, independent of glyph composition. The key is a reliance on known font attributes. In our system three feature extraction methods are used to demonstrate the importance of appropriate features for classification. The methods are tested on both Latin (English) and non-Latin (Khmer) scripts. Results show that the character-level recognition accuracy exceeds 92\% for Khmer and 96\% for English on degraded documents. This work is a step toward the recognition of scripts of low-density languages which typically do not warrant the development of commercial OCR, yet often have complete TrueType font descriptions.}, author = {Agrawal,Mudit and David Doermann} } @conference {16010, title = {The role of metacognition in robust AI systems}, booktitle = {AAAI-08 Workshop on Metareasoning,(Chicago, IL)}, year = {2008}, month = {2008///}, author = {Schmill,M. D and Oates,T. and Anderson,M. and Fults,S. and Josyula,D. and Perlis, Don and Wilson,S.} } @conference {16840, title = {Scalable network distance browsing in spatial databases}, booktitle = {Proceedings of the 2008 ACM SIGMOD international conference on Management of data}, series = {SIGMOD {\textquoteright}08}, year = {2008}, month = {2008///}, pages = {43 - 54}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {An algorithm is presented for finding the k nearest neighbors in a spatial network in a best-first manner using network distance. The algorithm is based on precomputing the shortest paths between all possible vertices in the network and then making use of an encoding that takes advantage of the fact that the shortest paths from vertex u to all of the remaining vertices can be decomposed into subsets based on the first edges on the shortest paths to them from u. Thus, in the worst case, the amount of work depends on the number of objects that are examined and the number of links on the shortest paths to them from q, rather than depending on the number of vertices in the network. The amount of storage required to keep track of the subsets is reduced by taking advantage of their spatial coherence which is captured by the aid of a shortest path quadtree. In particular, experiments on a number of large road networks as well as a theoretical analysis have shown that the storage has been reduced from O(N3) to O(N1.5) (i.e., by an order of magnitude equal to the square root). The precomputation of the shortest paths along the network essentially decouples the process of computing shortest paths along the network from that of finding the neighbors, and thereby also decouples the domain S of the query objects and that of the objects from which the neighbors are drawn from the domain V of the vertices of the spatial network. This means that as long as the spatial network is unchanged, the algorithm and underlying representation of the shortest paths in the spatial network can be used with different sets of objects.}, keywords = {decoupling, nearest neighbor, scalability, shortest path quadtree, spatial networks}, isbn = {978-1-60558-102-6}, doi = {10.1145/1376616.1376623}, url = {http://doi.acm.org/10.1145/1376616.1376623}, author = {Samet, Hanan and Sankaranarayanan,Jagan and Alborzi,Houman} } @article {12907, title = {Seasonal Cholera from Multiple Small Outbreaks, Rural Bangladesh}, journal = {Emerging Infectious DiseasesEmerg Infect Dis}, volume = {14}, year = {2008}, month = {2008/05//}, pages = {831 - 833}, abstract = {Clinical and environmental Vibrio cholerae organisms collected from February 2004 through April 2005 were systematically isolated from 2 rural Bangladeshi locales. Their genetic relatedness was evaluated at 5 loci that contained a variable number of tandem repeats (VNTR). The observed minimal overlap in VNTR patterns between the 2 communities was consistent with sequential, small outbreaks from local sources.}, isbn = {1080-6040}, doi = {10.3201/eid1405.071116}, author = {Stine,O. Colin and Alam,Munirul and Tang,Li and Nair,G. Balakrish and Siddique,A. Kasem and Faruque,Shah M. and Huq,Anwar and Rita R Colwell and Sack,R. Bradley and Morris,J. Glenn} } @article {15965, title = {A self-help guide for autonomous systems}, journal = {AI Magazine}, volume = {29}, year = {2008}, month = {2008///}, pages = {67 - 67}, author = {Anderson,M. L and Fults,S. and Josyula,D. P and Oates,T. and Perlis, Don and Wilson,S. and Wright,D.} } @article {15557, title = {Space-time tradeoffs for proximity searching in doubling spaces}, journal = {Algorithms-ESA 2008}, year = {2008}, month = {2008///}, pages = {112 - 123}, abstract = {We consider approximate nearest neighbor searching in metric spaces of constant doubling dimension. More formally, we are given a set S of n points and an error bound ε> 0. The objective is to build a data structure so that given any query point q in the space, it is possible to efficiently determine a point of S whose distance from q is within a factor of (1 + ε) of the distance between q and its nearest neighbor in S. In this paper we obtain the following space-time tradeoffs. Given a parameter γ ∈ [2,1/ε], we show how to construct a data structure of space nO(dim)log(1) space that can answer queries in time O(log(n))+(1())O(dim) . This is the first result that offers space-time tradeoffs for approximate nearest neighbor queries in doubling spaces. At one extreme it nearly matches the best result currently known for doubling spaces, and at the other extreme it results in a data structure that can answer queries in time O(log(n/ε)), which matches the best query times in Euclidean space. Our approach involves a novel generalization of the AVD data structure from Euclidean space to doubling space.}, doi = {10.1007/978-3-540-87744-8_10}, author = {Arya,S. and Mount, Dave and Vigneron,A. and Xia,J.} } @conference {18777, title = {A Survey of the Virtual Environments-based Assembly Training Applications}, year = {2008}, month = {2008///}, abstract = {The advent of virtual environments is presenting new ways of trainingtomorrow{\textquoteright}s workforce. Virtual environments offer numerous benefits in training applications. First, virtual environments allow extensive user interactions in a very convenient and natural manner. This interaction is greatly beneficial for increasing the user{\textquoteright}s retention of spatial information compared to text-based or video-based instructions that are non-interactive in nature. Second, virtual environments provide users with a 3D immersive experience. This feature helps users gain a better understanding of spatial relationships compared to 2D displays. Third, virtual environments support multi-media instructions. One can watch standard videos, view 3D animations, view text instructions, listen to audio instructions, and interact with 3D objects in the scene. This paper describes representative research and associated systems that use of virtual environments in assembly training applications. }, url = {http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.151.2892\&rep=rep1\&type=pdf}, author = {Gupta,S.K. and Anand,D. K. and Brough,J. E. and Kavetsky,R. A. and Schwartz,M. and Thakur,A.} } @conference {18920, title = {Synthesis of strategies from interaction traces}, series = {AAMAS {\textquoteright}08}, year = {2008}, month = {2008///}, pages = {855 - 862}, publisher = {International Foundation for Autonomous Agents and Multiagent Systems}, organization = {International Foundation for Autonomous Agents and Multiagent Systems}, address = {Richland, SC}, abstract = {We describe how to take a set of interaction traces produced by different pairs of players in a two-player repeated game, and combine them into a composite strategy. We provide an algorithm that, in polynomial time, can generate the best such composite strategy. We describe how to incorporate the composite strategy into an existing agent, as an enhancement of the agent{\textquoteright}s original strategy. We provide experimental results using interaction traces from 126 agents (most of them written by students as class projects) for the Iterated Prisoner{\textquoteright}s Dilemma, Iterated Chicken Game, and Iterated Battle of the Sexes. We compared each agent with the enhanced version of that agent produced by our algorithm. The enhancements improved the agents{\textquoteright} scores by about 5\% in the IPD, 11\% in the ICG, and 26\% in the IBS, and improved their rank by about 12\% in the IPD, 38\% in the ICG, and 33\% in the IBS.}, keywords = {agents, interaction, learning, multi-agent systems, prisoner{\textquoteright}s dilemma, repeated games}, isbn = {978-0-9817381-1-6}, url = {http://dl.acm.org/citation.cfm?id=1402298.1402343}, author = {Au,Tsz-Chiu and Kraus,Sarit and Nau, Dana S.} } @article {15282, title = {Toward automatic facet analysis and need negotiation: Lessons from mediated search}, journal = {ACM Transactions on Information Systems (TOIS)}, volume = {27}, year = {2008}, month = {2008/12//}, pages = {6:1{\textendash}6:42 - 6:1{\textendash}6:42}, abstract = {This work explores the hypothesis that interactions between a trained human search intermediary and an information seeker can inform the design of interactive IR systems. We discuss results from a controlled Wizard-of-Oz case study, set in the context of the TREC 2005 HARD track evaluation, in which a trained intermediary executed an integrated search and interaction strategy based on conceptual facet analysis and informed by need negotiation techniques common in reference interviews. Having a human {\textquotedblleft}in the loop{\textquotedblright} yielded large improvements over fully automated systems as measured by standard ranked-retrieval metrics, demonstrating the value of mediated search. We present a detailed analysis of the intermediary{\textquoteright}s actions to gain a deeper understanding of what worked and why. One contribution is a taxonomy of clarification types informed both by empirical results and existing theories in library and information science. We discuss how these findings can guide the development of future systems. Overall, this work illustrates how studying human information-seeking processes can lead to better information retrieval applications.}, keywords = {interactive information retrieval, Reference interview}, isbn = {1046-8188}, doi = {10.1145/1416950.1416956}, url = {http://doi.acm.org/10.1145/1416950.1416956}, author = {Jimmy Lin and Wu,Philip and Abels,Eileen} } @conference {15573, title = {Tradeoffs in Approximate Range Searching Made Simpler}, booktitle = {Computer Graphics and Image Processing, 2008. SIBGRAPI {\textquoteright}08. XXI Brazilian Symposium on}, year = {2008}, month = {2008/10//}, pages = {237 - 244}, abstract = {Range searching is a fundamental problem in computational geometry. The problem involves preprocessing a set of n points in R^d into a data structure, so that it is possible to determine the subset of points lying within a given query range. In approximate range searching, a parameter eps epsiv gt; 0 is given, and for a given query range R the points lying within distance eps diam(R) of the range{\textquoteright}s boundary may be counted or not. In this paper we present three results related to the issue of tradeoffs in approximate range searching. First, we introduce the range sketching problem. Next, we present a space-time tradeoff for smooth convex ranges, which generalize spherical ranges. Finally, we show how to modify the previous data structure to obtain a space-time tradeoff for simplex ranges. In contrast to existing results, which are based on relatively complex data structures, all three of our results are based on simple, practical data structures.}, keywords = {approximate range searching, computational geometry, data structure, Data structures, query range}, doi = {10.1109/SIBGRAPI.2008.24}, author = {Arya,S. and Fonseca,G. and Mount, Dave} } @conference {12038, title = {Who killed the directed model?}, booktitle = {Computer Vision and Pattern Recognition, IEEE Computer Society Conference on}, year = {2008}, month = {2008///}, pages = {1 - 8}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {Prior distributions are useful for robust low-level vision, and undirected models (e.g. Markov Random Fields) have become a central tool for this purpose. Though sometimes these priors can be specified by hand, this becomes difficult in large models, which has motivated learning these models from data. However, maximum likelihood learning of undirected models is extremely difficult- essentially all known methods require approximations and/or high computational cost. Conversely, directed models are essentially trivial to learn from data, but have not received much attention for low-level vision. We compare the two formalisms of directed and undirected models, and conclude that there is no a priori reason to believe one better represents low-level vision quantities. We formulate two simple directed priors, for natural images and stereo disparity, to empirically test if the undirected formalism is superior. We find in both cases that a simple directed model can achieve results similar to the best learnt undirected models with significant speedups in training time, suggesting that directed models are an attractive choice for tractable learning.}, isbn = {978-1-4244-2242-5}, doi = {http://doi.ieeecomputersociety.org/10.1109/CVPR.2008.4587817}, author = {Domke, Justin and Karapurkar, Alap and Aloimonos, J.} } @article {17870, title = {Active semantic caching to optimize multidimensional data analysis in parallel and distributed environments}, journal = {Parallel Computing}, volume = {33}, year = {2007}, month = {2007/08//}, pages = {497 - 520}, abstract = {In this paper, we present a multi-query optimization framework based on the concept of active semantic caching. The framework permits the identification and transparent reuse of data and computation in the presence of multiple queries (or query batches) that specify user-defined operators and aggregations originating from scientific data-analysis applications. We show how query scheduling techniques, coupled with intelligent cache replacement policies, can further improve the performance of query processing by leveraging the active semantic caching operators. We also propose a methodology for functionally decomposing complex queries in terms of primitives so that multiple reuse sites are exposed to the query optimizer, to increase the amount of reuse. The optimization framework and the database system implemented with it are designed to be efficient irrespective of the underlying parallel and/or distributed machine configuration. We present experimental results highlighting the performance improvements obtained by our methods using real scientific data-analysis applications on multiple parallel and distributed processing configurations (e.g., single symmetric multiprocessor (SMP) machine, cluster of SMP nodes, and a Grid computing configuration).}, keywords = {Active semantic caching, Parallel databases, Query optimization, Scientific data analysis}, isbn = {0167-8191}, doi = {10.1016/j.parco.2007.03.001}, url = {http://www.sciencedirect.com/science/article/pii/S0167819107000506}, author = {Andrade,Henrique and Kurc,Tahsin and Sussman, Alan and Saltz,Joel} } @article {14894, title = {Appearance Characterization of Linear Lambertian Objects, Generalized Photometric Stereo, and Illumination-Invariant Face Recognition}, journal = {Pattern Analysis and Machine Intelligence, IEEE Transactions on}, volume = {29}, year = {2007}, month = {2007/02//}, pages = {230 - 245}, abstract = {Traditional photometric stereo algorithms employ a Lambertian reflectance model with a varying albedo field and involve the appearance of only one object. In this paper, we generalize photometric stereo algorithms to handle all appearances of all objects in a class, in particular the human face class, by making use of the linear Lambertian property. A linear Lambertian object is one which is linearly spanned by a set of basis objects and has a Lambertian surface. The linear property leads to a rank constraint and, consequently, a factorization of an observation matrix that consists of exemplar images of different objects (e.g., faces of different subjects) under different, unknown illuminations. Integrability and symmetry constraints are used to fully recover the subspace bases using a novel linearized algorithm that takes the varying albedo field into account. The effectiveness of the linear Lambertian property is further investigated by using it for the problem of illumination-invariant face recognition using just one image. Attached shadows are incorporated in the model by a careful treatment of the inherent nonlinearity in Lambert{\textquoteright}s law. This enables us to extend our algorithm to perform face recognition in the presence of multiple illumination sources. Experimental results using standard data sets are presented}, keywords = {albedo field;appearance characterization;generalized photometric stereo algorithms;illumination-invariant face recognition;linear Lambertian objects;observation matrix factorization;face recognition;matrix decomposition;Algorithms;Artificial Intelligence;, Automated;Photogrammetry;Reproducibility of Results;Sensitivity and Specificity;, Computer-Assisted;Information Storage and Retrieval;Lighting;Linear Models;Pattern Recognition}, isbn = {0162-8828}, doi = {10.1109/TPAMI.2007.25}, author = {Zhou,S. K and Aggarwal,G. and Chellapa, Rama and Jacobs, David W.} } @conference {16019, title = {Application of MCL in a dialog agent}, booktitle = {Third Language and Technology Conference}, year = {2007}, month = {2007///}, author = {Josyula,D. P and Fults,S. and Anderson,M. L and Wilson,S. and Perlis, Don} } @conference {16448, title = {Bid based scheduler with backfilling for a multiprocessor system}, booktitle = {Proceedings of the ninth international conference on Electronic commerce}, year = {2007}, month = {2007///}, pages = {459 - 468}, author = {Yahav,I. and Raschid, Louiqa and Andrade,H.} } @article {18897, title = {CARA: A Cultural-Reasoning Architecture}, journal = {IEEE Intelligent Systems}, volume = {22}, year = {2007}, month = {2007///}, pages = {12 - 16}, abstract = {The Cognitive Architecture for Reasoning about Adversaries supports methods to gather data about different cultural groups and learn the intensity of those groups{\textquoteright} opinions on various topics. It also lets users build and extract models of those groups{\textquoteright} behavior and continuously refine those models through shared multiperson learning experiences. Researchers have applied the architecture to develop cultural-reasoning applications for example, a game that teaches soldiers about cultural norms in an Afghan village.}, keywords = {cultural reasoning, opinion extraction, opponent modeling}, isbn = {1541-1672}, author = {V.S. Subrahmanian and Albanese,Massimiliano and Martinez,Mar?a Vanina and Nau, Dana S. and Reforgiato,Diego and Simari,Gerardo I. and Sliva,Amy and Udrea,Octavian and Wilkenfeld,Jonathan} } @article {13740, title = {Combining outputs from multiple machine translation systems}, journal = {Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Proceedings of the Main Conference}, year = {2007}, month = {2007///}, pages = {228 - 235}, abstract = {Currently there are several approaches tomachine translation (MT) based on differ- ent paradigms; e.g., phrasal, hierarchical and syntax-based. These three approaches yield similar translation accuracy despite using fairly different levels of linguistic knowledge. The availability of such a variety of systems has led to a growing interest toward finding better translations by combining outputs from multiple sys- tems. This paper describes three differ- ent approaches to MT system combina- tion. These combination methods oper- ate on sentence, phrase and word level exploiting information from N -best lists, system scores and target-to-source phrase alignments. The word-level combination provides the most robust gains but the best results on the development test sets (NIST MT05 and the newsgroup portion of GALE 2006 dry-run) were achieved by combining all three methods. }, author = {Rosti,A.V.I. and Ayan,N.F. and Xiang,B. and Matsoukas,S. and Schwartz,R. and Dorr, Bonnie J} } @conference {13430, title = {Data Clustering with a Relational Push-Pull Model}, booktitle = {Data Mining Workshops, 2007. ICDM Workshops 2007. Seventh IEEE International Conference on}, year = {2007}, month = {2007/10//}, pages = {189 - 194}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present a new generative model for relational data in which relations between objects can have ei- ther a binding or a separating effect. For example, in a group of students separated into gender clusters, a "dating" relation would appear most frequently between the clusters, but a "roommate" relation would appear more often within clusters. In visualizing these rela- tions, one can imagine that the "dating" relation effec- tively pushes clusters apart, while the "roommate" re- lation pulls clusters into tighter formations. A unique aspect of the model is that an edge{\textquoteright}s existence is depen- dent on both the clusters to which the two connected objects belong and the features of the connected objects. We use simulated annealing to search for optimal val- ues of the unknown model parameters, where the ob- jective function is a Bayesian score derived from the generative model. Results describing the performance of the model are shown with artificial data as well as a subset of the Internet Movie Database. The results show that discovering a relation{\textquoteright}s tendency to either push or pull is critical to discovering a consistent clustering.}, isbn = {0-7695-3033-8, 978-0-7695-3019-2}, doi = {10.1109/ICDMW.2007.61}, url = {http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=4476666\&tag=1}, author = {Anthony,Adam and desJardins, Marie} } @conference {16425, title = {The defined cliffs variant in dynamic environments: a case study using the shaky ladder hyperplane-defined functions}, booktitle = {Proceedings of the 9th annual conference on Genetic and evolutionary computation}, series = {GECCO {\textquoteright}07}, year = {2007}, month = {2007///}, pages = {1158 - 1164}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {The shaky ladder hyperplane-defined functions (sl-hdfs) are a test suite utilized for exploring the behavior of the genetic algorithm (GA) in dynamic environments. This test suite can generate arbitrary problems with similar levels of difficulty and it provides a platform for systematic controlled observations of the GA in dynamic environments. Previous work has found two factors that contribute to the GA{\textquoteright}s success on sl-hdfs: (1) short initial building blocks and (2) significantly changing the reward structure during fitness landscape changes. Therefore a test function that combines these two features should facilitate even better GA performance. This has led to the construction of a new sl-hdf variant, "Defined Cliffs," in which we combine short elementary building blocks with sharp transitions in the environment. We examine this variant with two different levels of dynamics, static and regularly changing, using four different metrics. The results show superior GA performance on the Defined Cliffs over all previous variants (Cliffs, Weight, and Smooth). Our observations and conclusions in this variant further the understanding of the GA in dynamic environments.}, keywords = {building blocks, dynamic environments, Genetic algorithms, shaky ladder hyperplane-defined functions}, isbn = {978-1-59593-697-4}, doi = {10.1145/1276958.1277186}, url = {http://doi.acm.org/10.1145/1276958.1277186}, author = {Alharbi,Abir and Rand, William and Riolo,Rick} } @article {17072, title = {Designing Semantic Substrates for Visual Network Exploration}, journal = {Information VisualizationInformation Visualization}, volume = {6}, year = {2007}, month = {2007/12/21/}, pages = {281 - 300}, abstract = {A semantic substrate is a spatial template for a network, where nodes are grouped into regions and laid out within each region according to one or more node attributes. This paper shows how users can be given control in designing their own substrates and how this ability leads to a different approach to network data exploration. Users can create a semantic substrate, enter their data, get feedback from domain experts, edit the semantic substrate, and iteratively continue this procedure until the domain experts are satisfied with the insights they have gained. We illustrate this process in two case studies with domain experts working with legal precedents and food webs. Guidelines for designing substrates are provided, including how to locate, size, and align regions in a substrate, which attributes to choose for grouping nodes into regions, how to select placement methods and which attributes to set as parameters of the selected placement method. Throughout the paper, examples are illustrated with NVSS 2.0, the network visualization tool developed to explore the semantic substrate idea.}, keywords = {data exploration and analysis, Graphical user interfaces, Information Visualization, network visualization, semantic substrate design}, isbn = {1473-8716, 1473-8724}, doi = {10.1057/palgrave.ivs.9500162}, url = {http://ivi.sagepub.com/content/6/4/281}, author = {Aris,Aleks and Shneiderman, Ben} } @conference {17782, title = {Detecting stochastically scheduled activities in video}, booktitle = {International Joint Conference on Artificial Intelligence}, year = {2007}, month = {2007///}, pages = {1802 - 1807}, abstract = {The ability to automatically detect activities invideo is of increasing importance in applications such as bank security, airport tarmac security, bag- gage area security and building site surveillance. We present a stochastic activity model composed of atomic actions which are directly observable through image understanding primitives. We focus on answering two types of questions: (i) what are the minimal sub-videos in which a given action is identified with probability above a certain thresh- old and (ii) for a given video, can we decide which activity from a given set most likely occurred? We provide the MPS algorithm for the first problem, as well as two different algorithms (naiveMPA and MPA) to solve the second. Our experimental re- sults on a dataset consisting of staged bank robbery videos (described in [Vu et al., 2003]) show that our algorithms are both fast and provide high qual- ity results when compared to human reviewers. }, author = {Albanese, M. and Moscato, V. and Picariello, A. and V.S. Subrahmanian and Udrea,O.} } @article {18756, title = {Development of in-mold assembly process for realizing mesoscale revolute joints}, journal = {Transactions of North American Manufacturing Research Institution of SME}, volume = {35}, year = {2007}, month = {2007///}, abstract = {In-mold Assembly process at the mesoscalepresents several manufacturing challenges. Results reported in this paper demonstrate the technical feasibility of creating rigid body mesoscale revolute joints using In-Mold Assembly process. The following new results are reported in this paper. First, we describe a mold design with varying cavity shape to perform In-Mold Assembly. This mold design uses an accurate mold piece positioning method to avoid damage to delicate mesoscale parts during the cavity change step. Second, we describe a mold insert fabrication process for making mold inserts with the desired surface characteristics for mesoscale molding. Finally, we describe methods to limit the adhesion at the interfaces and hence create articulated revolute joint. Using the advances reported in this paper we have successfully molded a mesoscale revolute joint. To the best of our knowledge, this is the first demonstration of In-Mold Assembly process using a varying cavity shape mold to create a mesoscale revolute joint. }, url = {ftp://ftp.eng.umd.edu/:/home/glue/s/k/skgupta/pub/Publication/NAMRC07_Ananthanarayanan_draft.pdf}, author = {Ananthanarayanan,A. and Gupta,S.K. and Bruck,H. A. and Yu,Z. and Rajurkar,K. P.} } @conference {16079, title = {Discovering interesting usage patterns in text collections: integrating text mining with visualization}, booktitle = {Proceedings of the sixteenth ACM conference on Conference on information and knowledge management}, series = {CIKM {\textquoteright}07}, year = {2007}, month = {2007///}, pages = {213 - 222}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {This paper addresses the problem of making text mining results more comprehensible to humanities scholars, journalists, intelligence analysts, and other researchers, in order to support the analysis of text collections. Our system, FeatureLens1, visualizes a text collection at several levels of granularity and enables users to explore interesting text patterns. The current implementation focuses on frequent itemsets of n-grams, as they capture the repetition of exact or similar expressions in the collection. Users can find meaningful co-occurrences of text patterns by visualizing them within and across documents in the collection. This also permits users to identify the temporal evolution of usage such as increasing, decreasing or sudden appearance of text patterns. The interface could be used to explore other text features as well. Initial studies suggest that FeatureLens helped a literary scholar and 8 users generate new hypotheses and interesting insights using 2 text collections.}, keywords = {digital humanities, frequent closed itemsets, n-grams, text mining, user interface}, isbn = {978-1-59593-803-9}, doi = {10.1145/1321440.1321473}, url = {http://doi.acm.org/10.1145/1321440.1321473}, author = {Don,Anthony and Zheleva,Elena and Gregory,Machon and Tarkan,Sureyya and Auvil,Loretta and Clement,Tanya and Shneiderman, Ben and Plaisant, Catherine} } @inbook {17584, title = {Distributed Ranked Search}, booktitle = {High Performance Computing {\textendash} HiPC 2007High Performance Computing {\textendash} HiPC 2007}, series = {Lecture Notes in Computer Science}, volume = {4873}, year = {2007}, month = {2007///}, pages = {7 - 20}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {P2P deployments are a natural infrastructure for building distributed search networks. Proposed systems support locating and retrieving all results, but lack the information necessary to rank them. Users, however, are primarily interested in the most relevant results, not necessarily all possible results. Using random sampling, we extend a class of well-known information retrieval ranking algorithms such that they can be applied in this decentralized setting. We analyze the overhead of our approach, and quantify how our system scales with increasing number of documents, system size, document to node mapping (uniform versus non-uniform), and types of queries (rare versus popular terms). Our analysis and simulations show that a) these extensions are efficient, and scale with little overhead to large systems, and b) the accuracy of the results obtained using distributed ranking is comparable to that of a centralized implementation.}, isbn = {978-3-540-77219-4}, url = {http://dx.doi.org/10.1007/978-3-540-77220-0_6}, author = {Gopalakrishnan,Vijay and Morselli,Ruggero and Bhattacharjee, Bobby and Keleher,Pete and Srinivasan, Aravind}, editor = {Aluru,Srinivas and Parashar,Manish and Badrinath,Ramamurthy and Prasanna,Viktor} } @article {16257, title = {Draft Genome of the Filarial Nematode Parasite Brugia Malayi}, journal = {Science}, volume = {317}, year = {2007}, month = {2007/09/21/}, pages = {1756 - 1760}, abstract = {Parasitic nematodes that cause elephantiasis and river blindness threaten hundreds of millions of people in the developing world. We have sequenced the \~{}90 megabase (Mb) genome of the human filarial parasite Brugia malayi and predict \~{}11,500 protein coding genes in 71 Mb of robustly assembled sequence. Comparative analysis with the free-living, model nematode Caenorhabditis elegans revealed that, despite these genes having maintained little conservation of local synteny during \~{}350 million years of evolution, they largely remain in linkage on chromosomal units. More than 100 conserved operons were identified. Analysis of the predicted proteome provides evidence for adaptations of B. malayi to niches in its human and vector hosts and insights into the molecular basis of a mutualistic relationship with its Wolbachia endosymbiont. These findings offer a foundation for rational drug design.}, isbn = {0036-8075, 1095-9203}, doi = {10.1126/science.1145406}, url = {http://www.sciencemag.org/content/317/5845/1756}, author = {Ghedin,Elodie and Wang,Shiliang and Spiro,David and Caler,Elisabet and Zhao,Qi and Crabtree,Jonathan and Allen,Jonathan E and Delcher,Arthur L. and Guiliano,David B and Miranda-Saavedra,Diego and Angiuoli,Samuel V and Creasy,Todd and Amedeo,Paolo and Haas,Brian and El-Sayed, Najib M. and Wortman,Jennifer R. and Feldblyum,Tamara and Tallon,Luke and Schatz,Michael and Shumway,Martin and Koo,Hean and Salzberg,Steven L. and Schobel,Seth and Pertea,Mihaela and Pop, Mihai and White,Owen and Barton,Geoffrey J and Carlow,Clotilde K. S and Crawford,Michael J and Daub,Jennifer and Dimmic,Matthew W and Estes,Chris F and Foster,Jeremy M and Ganatra,Mehul and Gregory,William F and Johnson,Nicholas M and Jin,Jinming and Komuniecki,Richard and Korf,Ian and Kumar,Sanjay and Laney,Sandra and Li,Ben-Wen and Li,Wen and Lindblom,Tim H and Lustigman,Sara and Ma,Dong and Maina,Claude V and Martin,David M. A and McCarter,James P and McReynolds,Larry and Mitreva,Makedonka and Nutman,Thomas B and Parkinson,John and Peregr{\'\i}n-Alvarez,Jos{\'e} M and Poole,Catherine and Ren,Qinghu and Saunders,Lori and Sluder,Ann E and Smith,Katherine and Stanke,Mario and Unnasch,Thomas R and Ware,Jenna and Wei,Aguan D and Weil,Gary and Williams,Deryck J and Zhang,Yinhua and Williams,Steven A and Fraser-Liggett,Claire and Slatko,Barton and Blaxter,Mark L and Scott,Alan L} } @conference {12588, title = {Efficient Indexing For Articulation Invariant Shape Matching And Retrieval}, booktitle = {Computer Vision and Pattern Recognition, 2007. CVPR {\textquoteright}07. IEEE Conference on}, year = {2007}, month = {2007/06//}, pages = {1 - 8}, abstract = {Most shape matching methods are either fast but too simplistic to give the desired performance or promising as far as performance is concerned but computationally demanding. In this paper, we present a very simple and efficient approach that not only performs almost as good as many state-of-the-art techniques but also scales up to large databases. In the proposed approach, each shape is indexed based on a variety of simple and easily computable features which are invariant to articulations and rigid transformations. The features characterize pairwise geometric relationships between interest points on the shape, thereby providing robustness to the approach. Shapes are retrieved using an efficient scheme which does not involve costly operations like shape-wise alignment or establishing correspondences. Even for a moderate size database of 1000 shapes, the retrieval process is several times faster than most techniques with similar performance. Extensive experimental results are presented to illustrate the advantages of our approach as compared to the best in the field.}, keywords = {alignment;image, articulation, geometric, invariant, matching;image, matching;indexing;invariant, relationships;shape-wise, retrieval;indexing;, retrieval;pairwise, SHAPE}, doi = {10.1109/CVPR.2007.383227}, author = {Biswas,S. and Aggarwal,G. and Chellapa, Rama} } @article {16670, title = {Evaluating a cross-cultural children{\textquoteright}s online book community: Lessons learned for sociability, usability, and cultural exchange}, journal = {Interacting with Computers}, volume = {19}, year = {2007}, month = {2007/07//}, pages = {494 - 511}, abstract = {The use of computers for human-to-human communication among adults has been studied for many years, but using computer technology to enable children from all over the world to talk to each other has rarely been discussed by researchers. The goal of our research is to fill this gap and explore the design and evaluation of children{\textquoteright}s cross-language online communities via a case study of the International Children{\textquoteright}s Digital Library Communities (ICDLCommunities). This project supports the development of communities for children (ages 7{\textendash}11) that form around the International Digital Children{\textquoteright}s Library (ICDL) book collection. In this community the children can learn about each others{\textquoteright} cultures and make friends even if they do not speak the same language. They can also read and create stories and ask and answer questions about these. From this evaluation study we learned that: (i) children are very interested in their counterparts in other countries and a remarkable amount of communication takes place even when they do not share a common language; (ii) representing their identity online in many different forms is particularly important to children when communicating in an online community; (iii) children enjoy drawing but representing stories in a sequence of diagrams is challenging and needs support; and (iv) asking and answering questions without language is possible using graphical templates. In this paper we present our findings and make recommendations for designing children{\textquoteright}s cross-cultural online communities.}, keywords = {children, Cross-cultural, Evaluation and design, International Children{\textquoteright}s Digital Library, Online communities}, isbn = {0953-5438}, doi = {10.1016/j.intcom.2007.03.001}, url = {http://www.sciencedirect.com/science/article/pii/S0953543807000240}, author = {Komlodi,Anita and Hou,Weimin and Preece,Jenny and Druin, Allison and Golub,Evan and Alburo,Jade and Liao,Sabrina and Elkiss,Aaron and Resnik, Philip} } @article {16261, title = {Evolution of genes and genomes on the Drosophila phylogeny}, journal = {Nature}, volume = {450}, year = {2007}, month = {2007/11/08/}, pages = {203 - 218}, abstract = {Comparative analysis of multiple genomes in a phylogenetic framework dramatically improves the precision and sensitivity of evolutionary inference, producing more robust results than single-genome analyses can provide. The genomes of 12 Drosophila species, ten of which are presented here for the first time (sechellia, simulans, yakuba, erecta, ananassae, persimilis, willistoni, mojavensis, virilis and grimshawi), illustrate how rates and patterns of sequence divergence across taxa can illuminate evolutionary processes on a genomic scale. These genome sequences augment the formidable genetic tools that have made Drosophila melanogaster a pre-eminent model for animal genetics, and will further catalyse fundamental research on mechanisms of development, cell biology, genetics, disease, neurobiology, behaviour, physiology and evolution. Despite remarkable similarities among these Drosophila species, we identified many putatively non-neutral changes in protein-coding genes, non-coding RNA genes, and cis-regulatory regions. These may prove to underlie differences in the ecology and behaviour of these diverse species.}, isbn = {0028-0836}, doi = {10.1038/nature06341}, url = {http://www.nature.com/nature/journal/v450/n7167/full/nature06341.html}, author = {Clark,Andrew G. and Eisen,Michael B. and Smith,Douglas R. and Bergman,Casey M. and Oliver,Brian and Markow,Therese A. and Kaufman,Thomas C. and Kellis,Manolis and Gelbart,William and Iyer,Venky N. and Pollard,Daniel A. and Sackton,Timothy B. and Larracuente,Amanda M. and Singh,Nadia D. and Abad,Jose P. and Abt,Dawn N. and Adryan,Boris and Aguade,Montserrat and Akashi,Hiroshi and Anderson,Wyatt W. and Aquadro,Charles F. and Ardell,David H. and Arguello,Roman and Artieri,Carlo G. and Barbash,Daniel A. and Barker,Daniel and Barsanti,Paolo and Batterham,Phil and Batzoglou,Serafim and Begun,Dave and Bhutkar,Arjun and Blanco,Enrico and Bosak,Stephanie A. and Bradley,Robert K. and Brand,Adrianne D. and Brent,Michael R. and Brooks,Angela N. and Brown,Randall H. and Butlin,Roger K. and Caggese,Corrado and Calvi,Brian R. and Carvalho,A. Bernardo de and Caspi,Anat and Castrezana,Sergio and Celniker,Susan E. and Chang,Jean L. and Chapple,Charles and Chatterji,Sourav and Chinwalla,Asif and Civetta,Alberto and Clifton,Sandra W. and Comeron,Josep M. and Costello,James C. and Coyne,Jerry A. and Daub,Jennifer and David,Robert G. and Delcher,Arthur L. and Delehaunty,Kim and Do,Chuong B. and Ebling,Heather and Edwards,Kevin and Eickbush,Thomas and Evans,Jay D. and Filipski,Alan and Findei|[szlig]|,Sven and Freyhult,Eva and Fulton,Lucinda and Fulton,Robert and Garcia,Ana C. L. and Gardiner,Anastasia and Garfield,David A. and Garvin,Barry E. and Gibson,Greg and Gilbert,Don and Gnerre,Sante and Godfrey,Jennifer and Good,Robert and Gotea,Valer and Gravely,Brenton and Greenberg,Anthony J. and Griffiths-Jones,Sam and Gross,Samuel and Guigo,Roderic and Gustafson,Erik A. and Haerty,Wilfried and Hahn,Matthew W. and Halligan,Daniel L. and Halpern,Aaron L. and Halter,Gillian M. and Han,Mira V. and Heger,Andreas and Hillier,LaDeana and Hinrichs,Angie S. and Holmes,Ian and Hoskins,Roger A. and Hubisz,Melissa J. and Hultmark,Dan and Huntley,Melanie A. and Jaffe,David B. and Jagadeeshan,Santosh and Jeck,William R. and Johnson,Justin and Jones,Corbin D. and Jordan,William C. and Karpen,Gary H. and Kataoka,Eiko and Keightley,Peter D. and Kheradpour,Pouya and Kirkness,Ewen F. and Koerich,Leonardo B. and Kristiansen,Karsten and Kudrna,Dave and Kulathinal,Rob J. and Kumar,Sudhir and Kwok,Roberta and Lander,Eric and Langley,Charles H. and Lapoint,Richard and Lazzaro,Brian P. and Lee,So-Jeong and Levesque,Lisa and Li,Ruiqiang and Lin,Chiao-Feng and Lin,Michael F. and Lindblad-Toh,Kerstin and Llopart,Ana and Long,Manyuan and Low,Lloyd and Lozovsky,Elena and Lu,Jian and Luo,Meizhong and Machado,Carlos A. and Makalowski,Wojciech and Marzo,Mar and Matsuda,Muneo and Matzkin,Luciano and McAllister,Bryant and McBride,Carolyn S. and McKernan,Brendan and McKernan,Kevin and Mendez-Lago,Maria and Minx,Patrick and Mollenhauer,Michael U. and Montooth,Kristi and Mount, Stephen M. and Mu,Xu and Myers,Eugene and Negre,Barbara and Newfeld,Stuart and Nielsen,Rasmus and Noor,Mohamed A. F. and O{\textquoteright}Grady,Patrick and Pachter,Lior and Papaceit,Montserrat and Parisi,Matthew J. and Parisi,Michael and Parts,Leopold and Pedersen,Jakob S. and Pesole,Graziano and Phillippy,Adam M and Ponting,Chris P. and Pop, Mihai and Porcelli,Damiano and Powell,Jeffrey R. and Prohaska,Sonja and Pruitt,Kim and Puig,Marta and Quesneville,Hadi and Ram,Kristipati Ravi and Rand,David and Rasmussen,Matthew D. and Reed,Laura K. and Reenan,Robert and Reily,Amy and Remington,Karin A. and Rieger,Tania T. and Ritchie,Michael G. and Robin,Charles and Rogers,Yu-Hui and Rohde,Claudia and Rozas,Julio and Rubenfield,Marc J. and Ruiz,Alfredo and Russo,Susan and Salzberg,Steven L. and Sanchez-Gracia,Alejandro and Saranga,David J. and Sato,Hajime and Schaeffer,Stephen W. and Schatz,Michael C and Schlenke,Todd and Schwartz,Russell and Segarra,Carmen and Singh,Rama S. and Sirot,Laura and Sirota,Marina and Sisneros,Nicholas B. and Smith,Chris D. and Smith,Temple F. and Spieth,John and Stage,Deborah E. and Stark,Alexander and Stephan,Wolfgang and Strausberg,Robert L. and Strempel,Sebastian and Sturgill,David and Sutton,Granger and Sutton,Granger G. and Tao,Wei and Teichmann,Sarah and Tobari,Yoshiko N. and Tomimura,Yoshihiko and Tsolas,Jason M. and Valente,Vera L. S. and Venter,Eli and Venter,J. Craig and Vicario,Saverio and Vieira,Filipe G. and Vilella,Albert J. and Villasante,Alfredo and Walenz,Brian and Wang,Jun and Wasserman,Marvin and Watts,Thomas and Wilson,Derek and Wilson,Richard K. and Wing,Rod A. and Wolfner,Mariana F. and Wong,Alex and Wong,Gane Ka-Shu and Wu,Chung-I and Wu,Gabriel and Yamamoto,Daisuke and Yang,Hsiao-Pei and Yang,Shiaw-Pyng and Yorke,James A. and Yoshida,Kiyohito and Zdobnov,Evgeny and Zhang,Peili and Zhang,Yu and Zimin,Aleksey V. and Baldwin,Jennifer and Abdouelleil,Amr and Abdulkadir,Jamal and Abebe,Adal and Abera,Brikti and Abreu,Justin and Acer,St Christophe and Aftuck,Lynne and Alexander,Allen and An,Peter and Anderson,Erica and Anderson,Scott and Arachi,Harindra and Azer,Marc and Bachantsang,Pasang and Barry,Andrew and Bayul,Tashi and Berlin,Aaron and Bessette,Daniel and Bloom,Toby and Blye,Jason and Boguslavskiy,Leonid and Bonnet,Claude and Boukhgalter,Boris and Bourzgui,Imane and Brown,Adam and Cahill,Patrick and Channer,Sheridon and Cheshatsang,Yama and Chuda,Lisa and Citroen,Mieke and Collymore,Alville and Cooke,Patrick and Costello,Maura and D{\textquoteright}Aco,Katie and Daza,Riza and Haan,Georgius De and DeGray,Stuart and DeMaso,Christina and Dhargay,Norbu and Dooley,Kimberly and Dooley,Erin and Doricent,Missole and Dorje,Passang and Dorjee,Kunsang and Dupes,Alan and Elong,Richard and Falk,Jill and Farina,Abderrahim and Faro,Susan and Ferguson,Diallo and Fisher,Sheila and Foley,Chelsea D. and Franke,Alicia and Friedrich,Dennis and Gadbois,Loryn and Gearin,Gary and Gearin,Christina R. and Giannoukos,Georgia and Goode,Tina and Graham,Joseph and Grandbois,Edward and Grewal,Sharleen and Gyaltsen,Kunsang and Hafez,Nabil and Hagos,Birhane and Hall,Jennifer and Henson,Charlotte and Hollinger,Andrew and Honan,Tracey and Huard,Monika D. and Hughes,Leanne and Hurhula,Brian and Husby,M Erii and Kamat,Asha and Kanga,Ben and Kashin,Seva and Khazanovich,Dmitry and Kisner,Peter and Lance,Krista and Lara,Marcia and Lee,William and Lennon,Niall and Letendre,Frances and LeVine,Rosie and Lipovsky,Alex and Liu,Xiaohong and Liu,Jinlei and Liu,Shangtao and Lokyitsang,Tashi and Lokyitsang,Yeshi and Lubonja,Rakela and Lui,Annie and MacDonald,Pen and Magnisalis,Vasilia and Maru,Kebede and Matthews,Charles and McCusker,William and McDonough,Susan and Mehta,Teena and Meldrim,James and Meneus,Louis and Mihai,Oana and Mihalev,Atanas and Mihova,Tanya and Mittelman,Rachel and Mlenga,Valentine and Montmayeur,Anna and Mulrain,Leonidas and Navidi,Adam and Naylor,Jerome and Negash,Tamrat and Nguyen,Thu and Nguyen,Nga and Nicol,Robert and Norbu,Choe and Norbu,Nyima and Novod,Nathaniel and O{\textquoteright}Neill,Barry and Osman,Sahal and Markiewicz,Eva and Oyono,Otero L. and Patti,Christopher and Phunkhang,Pema and Pierre,Fritz and Priest,Margaret and Raghuraman,Sujaa and Rege,Filip and Reyes,Rebecca and Rise,Cecil and Rogov,Peter and Ross,Keenan and Ryan,Elizabeth and Settipalli,Sampath and Shea,Terry and Sherpa,Ngawang and Shi,Lu and Shih,Diana and Sparrow,Todd and Spaulding,Jessica and Stalker,John and Stange-Thomann,Nicole and Stavropoulos,Sharon and Stone,Catherine and Strader,Christopher and Tesfaye,Senait and Thomson,Talene and Thoulutsang,Yama and Thoulutsang,Dawa and Topham,Kerri and Topping,Ira and Tsamla,Tsamla and Vassiliev,Helen and Vo,Andy and Wangchuk,Tsering and Wangdi,Tsering and Weiand,Michael and Wilkinson,Jane and Wilson,Adam and Yadav,Shailendra and Young,Geneva and Yu,Qing and Zembek,Lisa and Zhong,Danni and Zimmer,Andrew and Zwirko,Zac and Jaffe,David B. and Alvarez,Pablo and Brockman,Will and Butler,Jonathan and Chin,CheeWhye and Gnerre,Sante and Grabherr,Manfred and Kleber,Michael and Mauceli,Evan and MacCallum,Iain} } @article {16850, title = {Execution time analysis of a top-down R-tree construction algorithm}, journal = {Information Processing Letters}, volume = {101}, year = {2007}, month = {2007/01/16/}, pages = {6 - 12}, abstract = {A detailed CPU execution-time analysis and implementation are given for a bulk loading algorithm to construct R-trees due to Garc{\'\i}a et al. [Y.J. Garc{\'\i}a, M.A. L{\'o}pez, S.T. Leutenegger, A greedy algorithm for bulk loading R-trees, in: GIS{\textquoteright}98: Proc. of the 6th ACM Intl. Symp. on Advances in Geographic Information Systems, Washington, DC, 1998, pp. 163{\textendash}164] which is known as the top-down greedy split (TGS) bulk loading algorithm. The TGS algorithm makes use of a classical bottom-up packing approach. In addition, an alternative packing approach termed top-down packing is introduced which may lead to improved query performance, and it is shown how to incorporate it into the TGS algorithm. A discussion is also presented of the tradeoffs of using the bottom-up and top-down packing approaches.}, keywords = {Bulk loading, Data structures, Packing, R-trees, Spatial databases}, isbn = {0020-0190}, doi = {10.1016/j.ipl.2006.07.010}, url = {http://www.sciencedirect.com/science/article/pii/S002001900600233X}, author = {Alborzi,Houman and Samet, Hanan} } @article {12134, title = {Experimenting with software testbeds for evaluating new technologies}, journal = {Empirical Software Engineering}, volume = {12}, year = {2007}, month = {2007///}, pages = {417 - 444}, abstract = {The evolution of a new technology depends upon a good theoretical basis for developing the technology, as well as upon its experimental validation. In order to provide for this experimentation, we have investigated the creation of a software testbed and the feasibility of using the same testbed for experimenting with a broad set of technologies. The testbed is a set of programs, data, and supporting documentation that allows researchers to test their new technology on a standard software platform. An important component of this testbed is the Unified Model of Dependability (UMD), which was used to elicit dependability requirements for the testbed software. With a collection of seeded faults and known issues of the target system, we are able to determine if a new technology is adept at uncovering defects or providing other aids proposed by its developers. In this paper, we present the Tactical Separation Assisted Flight Environment (TSAFE) testbed environment for which we modeled and evaluated dependability requirements and defined faults to be seeded for experimentation. We describe two completed experiments that we conducted on the testbed. The first experiment studies a technology that identifies architectural violations and evaluates its ability to detect the violations. The second experiment studies model checking as part of design for verification. We conclude by describing ongoing experimental work studying testing, using the same testbed. Our conclusion is that even though these three experiments are very different in terms of the studied technology, using and re-using the same testbed is beneficial and cost effective.}, doi = {10.1007/s10664-006-9034-0}, author = {Lindvall,M. and Rus,I. and Donzelli,P. and Memon, Atif M. and Zelkowitz, Marvin V and Betin-Can,A. and Bultan,T. and Ackermann,C. and Anders,B. and Asgari, S. and Basili, Victor R.} } @article {12599, title = {Face and Gesture Recognition-Appearance Characterization of Linear Lambertian Objects, Generalized Photometric Stereo, and Illumination-Invariant Face Recognition}, journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence}, volume = {29}, year = {2007}, month = {2007///}, pages = {230 - 245}, author = {Zhou,S. K and Aggarwal,G. and Chellapa, Rama and Jacobs, David W.} } @article {14581, title = {GATA and Nkx factors synergistically regulate tissue-specific gene expression and development in vivo}, journal = {Development}, volume = {134}, year = {2007}, month = {2007/01/01/}, pages = {189 - 198}, abstract = {In vitro studies have suggested that members of the GATA and Nkx transcription factor families physically interact, and synergistically activate pulmonary epithelial- and cardiac-gene promoters. However, the relevance of this synergy has not been demonstrated in vivo. We show that Gata6-Titf1 (Gata6-Nkx2.1) double heterozygous (G6-Nkx DH) embryos and mice have severe defects in pulmonary epithelial differentiation and distal airway development, as well as reduced phospholipid production. The defects in G6-Nkx DH embryos and mice are similar to those observed in human neonates with respiratory distress syndromes, including bronchopulmonary dysplasia, and differential gene expression analysis reveals essential developmental pathways requiring synergistic regulation by both Gata6 and Titf1 (Nkx2.1). These studies indicate that Gata6 and Nkx2.1 act in a synergistic manner to direct pulmonary epithelial differentiation and development in vivo, providing direct evidence that interactions between these two transcription factor families are crucial for the development of the tissues in which they are co-expressed.}, doi = {10.1242/dev.02720}, url = {http://dev.biologists.org/content/134/1/189.abstract}, author = {Zhang,Yuzhen and Rath,Nibedita and Hannenhalli, Sridhar and Wang,Zhishan and Cappola,Thomas and Kimura,Shioko and Atochina-Vasserman,Elena and Lu,Min Min and Beers,Michael F. and Morrisey,Edward E.} } @conference {17774, title = {A general framework for reasoning about inconsistency}, booktitle = {Proceedings of the International Joint Conference on Artificial Intelligence}, year = {2007}, month = {2007///}, pages = {599 - 604}, abstract = {Numerous logics have been developed for reason-ing about inconsistency which differ in (i) the logic to which they apply, and (ii) the criteria used to draw inferences. In this paper, we propose a gen- eral framework for reasoning about inconsistency in a wide variety of logics including ones for which inconsistency resolution methods have not yet been studied (e.g. various temporal and epistemic log- ics). We start with Tarski and Scott{\textquoteright}s axiomatiza- tion of logics, but drop their monotonicity require- ments that we believe are too strong for AI. For such a logic L, we define the concept of an option. Options are sets of formulas in L that are closed and consistent according to the notion of consequence and consistency in L. We show that by defining an appropriate preference relation on options, we can capture several existing works such as Brewka{\textquoteright}s subtheories. We also provide algorithms to com- pute most preferred options. }, author = {V.S. Subrahmanian and Amgoud,L.} } @article {15279, title = {Genome Analysis Linking Recent European and African Influenza (H5N1) Viruses}, journal = {Emerging Infectious DiseasesEmerg Infect Dis}, volume = {13}, year = {2007}, month = {2007/05//}, pages = {713 - 718}, abstract = {Although linked, these viruses are distinct from earlier outbreak strains., To better understand the ecology and epidemiology of the highly pathogenic avian influenza virus in its transcontinental spread, we sequenced and analyzed the complete genomes of 36 recent influenza A (H5N1) viruses collected from birds in Europe, northern Africa, and southeastern Asia. These sequences, among the first complete genomes of influenza (H5N1) viruses outside Asia, clearly depict the lineages now infecting wild and domestic birds in Europe and Africa and show the relationships among these isolates and other strains affecting both birds and humans. The isolates fall into 3 distinct lineages, 1 of which contains all known non-Asian isolates. This new Euro-African lineage, which was the cause of several recent (2006) fatal human infections in Egypt and Iraq, has been introduced at least 3 times into the European-African region and has split into 3 distinct, independently evolving sublineages. One isolate provides evidence that 2 of these sublineages have recently reassorted.}, isbn = {1080-6040}, doi = {10.3201/eid1305.070013}, author = {Salzberg,Steven L. and Kingsford, Carl and Cattoli,Giovanni and Spiro,David J. and Janies,Daniel A. and Aly,Mona Mehrez and Brown,Ian H. and Couacy-Hymann,Emmanuel and De Mia,Gian Mario and Dung,Do Huu and Guercio,Annalisa and Joannis,Tony and Ali,Ali Safar Maken and Osmani,Azizullah and Padalino,Iolanda and Saad,Magdi D. and Savi{\'c},Vladimir and Sengamalay,Naomi A. and Yingst,Samuel and Zaborsky,Jennifer and Zorman-Rojs,Olga and Ghedin,Elodie and Capua,Ilaria} } @conference {19449, title = {Grow and Know: Understanding Record-keeping Needs for Tracking the Development of Young Children}, booktitle = {SIGCHI {\textquoteright}07}, series = {CHI {\textquoteright}07}, year = {2007}, month = {2007///}, pages = {1351 - 1360}, publisher = {ACM}, organization = {ACM}, abstract = {From birth through age five, children undergo rapid development and learn skills that will influence them their entire lives. Regular visits to the pediatrician and detailed record-keeping can ensure that children are progressing and can identify early warning signs of developmental delay or disability. However, new parents are often overwhelmed with new responsibilities, and we believe there is an opportunity for computing technology to assist in this process. In this paper, we present a qualitative study aimed at uncovering some specific needs for record-keeping and analysis for new parents and their network of caregivers. Through interviews and focus groups, we have confirmed assumptions about the rationales parents have and the functions required for using technology for record-keeping. We also identify new themes, potential prototypes, and design guidelines for this domain.}, keywords = {children, design requirements, developmental delay, healthcare, qualitative study}, isbn = {978-1-59593-593-9}, url = {http://doi.acm.org/10.1145/1240624.1240830}, author = {Kientz, Julie A. and Arriaga, Rosa I. and Marshini Chetty and Hayes, Gillian R. and Richardson, Jahmeilah and Patel, Shwetak N. and Abowd, Gregory D.} } @conference {18573, title = {Holding the Internet Accountable}, booktitle = {Proc. 6th ACM Workshop on Hot Topics in Networks (Hotnets-VI)}, year = {2007}, month = {2007/11/01/}, abstract = {Today{\textquoteright}s IP network layer provides little to no protection against misconfiguration or malice. Despite some progress in improving the robustness and security of the IP layer, misconfigurations and attacks still occur frequently. We show how a network layer that provides accountability, i.e., the ability to associate each action with the responsible entity, provides a firm foundation for defenses against misconfiguration and malice. We present the design of a network layer that incorporates accountability called AIP (Accountable Internet Protocol) and show how its features{\textemdash}notably, its use of self-certifying addresses{\textemdash} can improve both source accountability (the ability to trace actions to a particular end host and stop that host from misbehaving) and control-plane accountability (the ability to pinpoint and prevent attacks on routing).}, url = {http://repository.cmu.edu/compsci/66}, author = {Andersen,David and Balakrishnan,Hari and Feamster, Nick and Koponen,Teemu and Moon,Daekyong and Shenker,Scott} } @conference {15986, title = {Hood College, Master of Business Administration, 2005 Hood College, Master of Science (Computer Science), 2001 Hood College, Bachelor of Science (Computer Science), 1998 Frederick Community College, Associate in Arts (Business Administration), 1993}, booktitle = {Proceedings of the Workshop on Metareasoning in Agent-Based Systems}, year = {2007}, month = {2007///}, author = {Anderson,M. L and Schmill,M. and Oates,T. and Perlis, Don and Josyula,D. and Wright,D. and Human,S. W.T.D.N and Metacognition,L. and Fults,S. and Josyula,D. P} } @inbook {19453, title = {How Smart Homes Learn: The Evolution of the Networked Home and Household}, booktitle = {UbiComp 2007: Ubiquitous Computing}, series = {Lecture Notes in Computer Science}, year = {2007}, month = {2007/01/01/}, pages = {127 - 144}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, abstract = {Despite a growing desire to create smart homes, we know little about how networked technologies interact with a house{\textquoteright}s infrastructure. In this paper, we begin to close this gap by presenting findings from a study that examined the relationship between home networking and the house itself{\textemdash}and the work that results for householders as a consequence of this interaction. We discuss four themes that emerged: an ambiguity in understanding the virtual boundaries created by wireless networks, the home network control paradox, a new home network access paradox, and the relationship between increased responsibilities and the possibilities of wireless networking.}, keywords = {Computer Communication Networks, computers and society, home networking, Information Systems Applications (incl.Internet), infrastructure, smart home, software engineering, Systems and Data Security, User Interfaces and Human Computer Interaction}, isbn = {978-3-540-74852-6, 978-3-540-74853-3}, url = {http://link.springer.com/chapter/10.1007/978-3-540-74853-3_8}, author = {Marshini Chetty and Sung, Ja-Young and Grinter, Rebecca E.}, editor = {Krumm, John and Abowd, Gregory D. and Seneviratne, Aruna and Strang, Thomas} } @inbook {17208, title = {Human Values for Shaping the Made World}, booktitle = {Human-Computer Interaction {\textendash} INTERACT 2007Human-Computer Interaction {\textendash} INTERACT 2007}, series = {Lecture Notes in Computer Science}, volume = {4662}, year = {2007}, month = {2007///}, pages = {1 - 1}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Interface design principles have been effective in shaping new desktop applications, web-based resources, and mobile devices. Usability and sociability promote successful online communities and social network services. The contributions of human-computer interaction researchers have been effective in raising the quality of design of many products and services. As our influence grows, we can play an even more profound role in guaranteeing that enduring human values are embedded in the next generation of technology. This talk identifies which goals are realistic, such as universality, responsibility, trust, empathy, and privacy, and how we might ensure that they become part of future services and systems.}, isbn = {978-3-540-74794-9}, url = {http://dx.doi.org/10.1007/978-3-540-74796-3_1}, author = {Shneiderman, Ben}, editor = {Baranauskas,C{\'e}cilia and Palanque,Philippe and Abascal,Julio and Barbosa,Simone} } @conference {18855, title = {Investigation of Revolute Joint Clearances Created by an In-Mold Assembly Process}, year = {2007}, month = {2007/07//}, pages = {112 - 117}, abstract = {Revolute joints are frequently used in articulated structures. Traditionally, such a joint is formed by assembling two components. As an alternative, revolute joints can be created inside the mold using an in-mold assembly process. This process eliminates the need for post-molding assembly, thus significantly reducing the cycle time and part count. The functional performance of a revolute joint depends on the clearance in the joint. The clearance in turn depends on the part shrinkage and the mold deformation during the molding process. The presence of a polymer part during the second molding stage makes an in-mold assembly process significantly different from the traditional molding process due to the difference in heat transfer and deformation characteristics. This paper presents experimental data and a preliminary model to explain the differences in clearance produced by an aluminum mold and an Aluminum mold with an Acrylonitrile butadiene styrene (ABS) insert. Our data indicates that there is a significant difference between the clearances observed from these two different types of molds. We believe that clearances produced depend strongly on the thermal history of the parts.}, keywords = {acrylonitrile butadiene styrene insert, aluminum mold, assembling, couplings, cycle time reduction, heat transfer, injection moulding, inmold assembly process, mold deformation, molding process, moulding equipment, part count reduction, plastics industry, revolute joint clearances}, doi = {10.1109/ISAM.2007.4288458}, author = {Ananthanarayanan,A. and Thamire,C. and Gupta,S.K.} } @article {12036, title = {A Language for Human Action}, journal = {Computer}, volume = {40}, year = {2007}, month = {2007/05//}, pages = {42 - 51}, abstract = {Human-centered computing (HCC) involves conforming computer technology to humans while naturally achieving human-machine interaction. In a human-centered system, the interaction focuses on human requirements, capabilities, and limitations. These anthropocentric systems also focus on the consideration of human sensory-motor skills in a wide range of activities. This ensures that the interface between artificial agents and human users accounts for perception and action in a novel interaction paradigm. In turn, this leads to behavior understanding through cognitive models that allow content description and, ultimately, the integration of real and virtual worlds. Our work focuses on building a language that maps to the lower-level sensory and motor languages and to the higher-level natural language. An empirically demonstrated human activity language provides sensory-motor-grounded representations for understanding human actions. A linguistic framework allows the analysis and synthesis of these actions.}, keywords = {anthropocentric system, artificial agent, Artificial intelligence, cognitive model, Concrete, Databases, human action, human activity language, Human computer interaction, human factors, human sensory-motor skill, human-centered computing, human-machine interaction, HUMANS, Intelligent sensors, linguistic framework, linguistics, Mirrors, Morphology, natural language, natural languages, Neurons, Power system modeling, user interface, User interfaces}, isbn = {0018-9162}, doi = {10.1109/MC.2007.154}, author = {Guerra-Filho,G. and Aloimonos, J.} } @conference {17769, title = {MAGIC: A Multi-Activity Graph Index for Activity Detection}, booktitle = {Information Reuse and Integration, 2007. IRI 2007. IEEE International Conference on}, year = {2007}, month = {2007/08//}, pages = {267 - 272}, abstract = {Suppose we are given a set A of activities of interest, a set O of observations, and a probability threshold p. We are interested in finding the set of all pairs (a, O{\textquoteright}), where a epsi A and O{\textquoteright} sube O, that minimally validate the fact that an instance of activity a occurs in O with probability p or more. The novel contribution of this paper is the notion of the multi-activity graph index (MAGIC), which can index very large numbers of observations from interleaved activities and quickly retrieve completed instances of the monitored activities. We introduce two complexity reducing restrictions of the problem (which takes exponential time) and develop algorithms for each. We experimentally evaluate our exponential algorithm as well as the restricted algorithms on both synthetic data and a real (depersonalized) travel data set consisting of 5.5 million observations. Our experiments show that MAGIC consumes reasonable amounts of memory and can retrieve completed instances of activities in just a few seconds. We also report appropriate statistical significance results validating our experimental hypotheses.}, keywords = {complexity;graph, detection;complexity, graph, index;probability, MAGIC;activity, reducing, restrictions;exponential, theory;probability;, threshold;computational, time;multiactivity}, doi = {10.1109/IRI.2007.4296632}, author = {Albanese, M. and Pugliese, A. and V.S. Subrahmanian and Udrea,O.} } @conference {13828, title = {Measuring variability in sentence ordering for news summarization}, booktitle = {Proceedings of the Eleventh European Workshop on Natural Language Generation}, series = {ENLG {\textquoteright}07}, year = {2007}, month = {2007///}, pages = {81 - 88}, publisher = {Association for Computational Linguistics}, organization = {Association for Computational Linguistics}, address = {Stroudsburg, PA, USA}, abstract = {The issue of sentence ordering is an important one for natural language tasks such as multi-document summarization, yet there has not been a quantitative exploration of the range of acceptable sentence orderings for short texts. We present results of a sentence reordering experiment with three experimental conditions. Our findings indicate a very high degree of variability in the orderings that the eighteen subjects produce. In addition, the variability of reorderings is significantly greater when the initial ordering seen by subjects is different from the original summary. We conclude that evaluation of sentence ordering should use multiple reference orderings. Our evaluation presents several metrics that might prove useful in assessing against multiple references. We conclude with a deeper set of questions: (a) what sorts of independent assessments of quality of the different reference orderings could be made and (b) whether a large enough test set would obviate the need for such independent means of quality assessment.}, url = {http://dl.acm.org/citation.cfm?id=1610163.1610177}, author = {Madnani,Nitin and Passonneau,Rebecca and Ayan,Necip Fazil and Conroy,John M. and Dorr, Bonnie J and Klavans,Judith L. and O{\textquoteright}Leary, Dianne P. and Schlesinger,Judith D.} } @conference {18586, title = {MobCast: Overlay Architecture for Seamless IP Mobility using Scalable Anycast Proxies}, booktitle = {Wireless Communications and Networking Conference, 2007.WCNC 2007. IEEE}, year = {2007}, month = {2007/03//}, pages = {3872 - 3876}, abstract = {We propose a routing overlay system, MobCast, for simple and efficient routing to mobile hosts. Mobcast nodes advertise the same address space at each proxy location, and each mobile host is assigned a "universal" IP address from this address space, so packets sent to a mobile host{\textquoteright}s universal IP address automatically go to the nearest proxy on the overlay. The overlay then delivers the packets to the mobile host. Our architecture enables seamless mobility for both micro and macro mobility. While our initial design is not as mature as Mobile IP, it shows great promise to solve the traditional problems of ingress routing, firewalls, NATs, and rapid mobility with much lower complexity. We present our design as a scalable and deployable alternative to mobile IP. In this paper, we focus on describing the MobCast system architecture. We form our arguments for scalability, handoff-speed, and simplicity, and give our initial results for scalability. We postpone a detailed discussion of MobCast{\textquoteright}s security model for future work.}, keywords = {address space, handoff-speed, IP networks, MobCast, mobile hosts, mobile radio, overlay architecture, proxy location, routing overlay system, scalable anycast proxies, seamless IP mobility, telecommunication network routing, universal IP address}, doi = {10.1109/WCNC.2007.708}, author = {Lee,C.P. and Attrey,K. and Caballero,C. and Feamster, Nick and Mihail,M. and Copeland,J.A.} } @article {13834, title = {Multiple alternative sentence compressions for automatic text summarization}, journal = {Proceedings of DUC}, year = {2007}, month = {2007///}, abstract = {We perform multi-document summariza-tion by generating compressed versions of source sentences as summary candi- dates and using weighted features of these candidates to construct summaries. We combine a parse-and-trim approach with a novel technique for producing multiple alternative compressions for source sen- tences. In addition, we use a novel method for tuning the feature weights that maxi- mizes the change in the ROUGE-2 score (∆ROUGE) between the already existing summary state and the new state that re- sults from the addition of the candidate under consideration. We also describe ex- periments using a new paraphrase-based feature for redundancy checking. Finally, we present the results of our DUC2007 submissions and some ideas for future work. }, author = {Madnani,N. and Zajic, David and Dorr, Bonnie J and Ayan,N.F. and Jimmy Lin} } @conference {11985, title = {Multiple View Image Reconstruction: A Harmonic Approach}, booktitle = {IEEE Conference on Computer Vision and Pattern Recognition, 2007. CVPR {\textquoteright}07}, year = {2007}, month = {2007/06/17/22}, pages = {1 - 8}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper presents a new constraint connecting the signals in multiple views of a surface. The constraint arises from a harmonic analysis of the geometry of the imaging process and it gives rise to a new technique for multiple view image reconstruction. Given several views of a surface from different positions, fundamentally different information is present in each image, owing to the fact that cameras measure the incoming light only after the application of a low-pass filter. Our analysis shows how the geometry of the imaging is connected to this filtering. This leads to a technique for constructing a single output image containing all the information present in the input images.}, keywords = {CAMERAS, filtering theory, Geometry, Harmonic analysis, Image reconstruction, imaging process geometry, Information filtering, Joining processes, Low pass filters, low-pass filter, low-pass filters, multiple view image reconstruction, Position measurement, Power harmonic filters, surface reconstruction}, isbn = {1-4244-1180-7}, doi = {10.1109/CVPR.2007.383285}, author = {Domke, J. and Aloimonos, J.} } @article {12918, title = {New records of phytoplankton for Bangladesh. 3. Volvocales}, journal = {Bangladesh Journal of Plant Taxonomy}, volume = {14}, year = {2007}, month = {2007/12/01/}, abstract = {This study presents 21 species of Chlamydomonas, four species of Carteria, two species of each of Nephroselmis, Pyramidomonas and Scherffelia, and Collodictyon triciliatum, Polytoma minus, Tetrachloridium ? allorgei and Tetraselmis cordiformis. These species have been reported from some ponds of Mathbaria of Pirojpur and Bakerganj of Barisal districts in Bangladesh.}, isbn = {1028-2092}, doi = {10.3329/bjpt.v14i1.518}, url = {http://www.banglajol.info/bd/index.php/BJPT/article/viewArticle/518}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasmin,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @article {12917, title = {New records of phytoplankton for Bangladesh. 4. Chlorococcales}, journal = {Bangladesh Journal of Plant Taxonomy}, volume = {14}, year = {2007}, month = {2007/12/02/}, abstract = {This study presents three species from each of Schroederia, Monoraphidium and Ankistrodesmus, two species and one variety of Dictyosphaerium, two varieties of Pediastrum, and Tetraedron arthrodesmiforme var. contorta, Chlorotetraedron polymorphum, Myrmecia aquatica, Oocystis tainoensis, Nephrocytium spirale, Kirchneriella irregularis, Coelastrum indicum and Scenedesmus similagineus. These taxa have been reported from some ponds of Mathbaria of Pirojpur and Bakerganj of Barisal Districts in Bangladesh.}, isbn = {1028-2092}, doi = {10.3329/bjpt.v14i2.528}, url = {http://www.banglajol.info/bd/index.php/BJPT/article/viewArticle/528}, author = {Khondker,Moniruzzaman and Bhuiyan,Rauf Ahmed and Yeasim,Jenat and Alam,Munirul and Sack,R. Bradley and Huq,Anwar and Rita R Colwell} } @conference {15997, title = {Ontologies for reasoning about failures in AI systems}, booktitle = {Proceedings from the Workshop on Metareasoning in Agent Based Systems at the Sixth International Joint Conference on Autonomous Agents and Multiagent Sytems}, year = {2007}, month = {2007///}, author = {Schmill,M. and Josyula,D. and Anderson,M. L and Wilson,S. and Oates,T. and Perlis, Don and Fults,S.} } @article {15652, title = {Optimal expected-case planar point location}, journal = {SIAM Journal on Computing}, volume = {37}, year = {2007}, month = {2007///}, pages = {584 - 584}, abstract = {Point location is the problem of preprocessing a planar polygonal subdivision S of size ninto a data structure in order to determine efficiently the cell of the subdivision that contains a given query point. We consider this problem from the perspective of expected query time. We are given the probabilities pz that the query point lies within each cell z ∈ S. The entropy H of the resulting discrete probability distribution is the dominant term in the lower bound on the expected-case query time. We show that it is possible to achieve query time H + O( √ H + 1) with space O(n), which is optimal up to lower order terms in the query time. We extend this result to subdivisions with convex cells, assuming a uniform query distribution within each cell. In order to achieve space efficiency, we introduce the concept of entropy-preserving cuttings. }, author = {Arya,S. and Malamatos,T. and Mount, Dave and Wong,K.C.} } @conference {14803, title = {Performance Measurement of Novice HPC Programmers Code}, booktitle = {Proceedings of the 3rd International Workshop on Software Engineering for High Performance Computing Applications}, series = {SE-HPC {\textquoteright}07}, year = {2007}, month = {2007///}, pages = {3{\textendash} - 3{\textendash}}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, abstract = {Performance is one of the key factors of improving productivity in High Performance Computing (HPC). In this paper we discuss current studies in the field of performance measurement of codes captured in classroom experiments for the High Productivity Computing Project (HPCS). We give two examples of measurements introducing two new hypotheses: spending more effort doesn{\textquoteright}t always result in improvement of performance for novices; the use of higher level MPI functions promises better performance for novices. We also present a tool - the Automated Performance Measurement System (APMS). APMS helps to partially automate the measurement of the performance of a set of parallel programs with several inputs. The design and implementation of the tool is flexible enough to allow other researchers to conduct similar studies.}, keywords = {measurement, performance, performance measures, product metrics, program analysis}, isbn = {0-7695-2969-0}, doi = {http://dx.doi.org/10.1109/SE-HPC.2007.4}, url = {http://dx.doi.org/10.1109/SE-HPC.2007.4}, author = {Alameh, Rola and Zazworka, Nico and Hollingsworth, Jeffrey K} } @article {15602, title = {Pointerless implementation of hierarchical simplicial meshes and efficient neighbor finding in arbitrary dimensions}, journal = {International Journal of Computational Geometry and Applications}, volume = {17}, year = {2007}, month = {2007///}, pages = {595 - 631}, abstract = {We describe a pointerless representation of hierarchical regular simplicial meshes, basedon a bisection approach proposed by Maubach. We introduce a new labeling scheme, called an LPT code, which uniquely encodes the geometry of each simplex of the hi- erarchy, and we present rules to compute the neighbors of a given simplex efficiently through the use of these codes. In addition, we show how to traverse the associated tree and how to answer point location and interpolation queries. Our system works in arbitrary dimensions. }, author = {Atalay,F. B and Mount, Dave and Mitchell,J.} } @article {17850, title = {Principles for designing data-/compute-intensive distributed applications and middleware systems for heterogeneous environments}, journal = {Journal of Parallel and Distributed Computing}, volume = {67}, year = {2007}, month = {2007/07//}, pages = {755 - 771}, abstract = {The nature of distributed systems is constantly and steadily changing as the hardware and software landscape evolves. Porting applications and adapting existing middleware systems to ever changing computational platforms has become increasingly complex and expensive. Therefore, the design of applications, as well as the design of next generation middleware systems, must follow a set of guiding principles in order to insure long-term {\textquotedblleft}survivability{\textquotedblright} without costly re-engineering. From our practical experience, the key determinants to success in this endeavor are adherence to the following principles: (1) Design for change; (2) Provide for storage subsystem I/O coordination; (3) Employ workload partitioning and load balancing techniques; (4) Employ caching; (5) Schedule the workload; and (6) Understand the workload. In order to support these principles, we have collected extensive experimental results comparing three middleware systems targeted at data- and compute-intensive applications implemented by our research group during the course of the last decade, on a single data- and compute-intensive application. The main contribution of this work is the analysis of a level playing field, where we discuss and quantify how adherence to these guiding principles impacts overall system throughput and response time.}, keywords = {Computational science applications, Data-/compute-intensive applications, Heterogeneous environments, Middleware systems}, isbn = {0743-7315}, doi = {10.1016/j.jpdc.2007.04.006}, url = {http://www.sciencedirect.com/science/article/pii/S0743731507000603}, author = {Kim,Jik-Soo and Andrade,Henrique and Sussman, Alan} } @article {12063, title = {A probabilistic framework for correspondence and egomotion}, journal = {Dynamical Vision}, year = {2007}, month = {2007///}, pages = {232 - 242}, author = {Domke, J. and Aloimonos, J.} } @article {12043, title = {A probabilistic notion of camera geometry: Calibrated vs. uncalibrated}, journal = {PHOTOGRAMMETRIE FERNERKUNDUNG GEOINFORMATION}, volume = {2007}, year = {2007}, month = {2007///}, pages = {25 - 25}, author = {Domke, J. and Aloimonos, J.} } @conference {18919, title = {Reactive Query Policies: A Formalism for Planning with Volatile External Information}, year = {2007}, month = {2007/04/01/5}, pages = {243 - 250}, abstract = {To generate plans for collecting data for data mining, an important problem is information volatility during planning: the information needed by the planning system may change or expire during the planning process, as changes occur in the data being collected. In such situations, the planning system faces two challenges: how to generate plans despite these changes, and how to guarantee that a plan returned by the planner will remain valid for some period of time after the planning ends. The focus of our work is to address both of the above challenges. In particular, we provide: 1) A formalism for reactive query policies, a class of strategies for deciding when to reissue queries for information that has changed during the planning process. This class includes all query management strategies that have yet been developed. 2) A new reactive query policy called the presumptive strategy. In our experiments, the presumptive strategy ran exponentially faster than the lazy strategy, the best previously known query management strategy. In the hardest set of problems we tested, the presumptive strategy took 4.7\% as much time and generated 6.9\% as many queries as the lazy strategy}, keywords = {data collection, data mining, information volatility, planning (artificial intelligence), planning system, presumptive strategy, query management, Query processing, reactive query policies, volatile external information}, doi = {10.1109/CIDM.2007.368880}, author = {Au,Tsz-Chiu and Nau, Dana S.} } @inbook {12601, title = {Recognizing Faces Across Age Progression}, booktitle = {Face Biometrics for Personal IdentificationFace Biometrics for Personal Identification}, series = {Signals and Communication Technology}, year = {2007}, month = {2007///}, pages = {27 - 42}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, isbn = {978-3-540-49346-4}, url = {http://dx.doi.org/10.1007/978-3-540-49346-4_3}, author = {Ramanathan,Narayanan and Chellapa, Rama}, editor = {Hammoud,Riad and Abidi,Besma and Abidi,Mongi} } @article {12008, title = {A roadmap to the integration of early visual modules}, journal = {International Journal of Computer Vision}, volume = {72}, year = {2007}, month = {2007///}, pages = {9 - 25}, author = {Ogale, A. S and Aloimonos, J.} } @article {13158, title = {Robust appearance modeling for pedestrian and vehicle tracking}, journal = {Multimodal Technologies for Perception of Humans}, year = {2007}, month = {2007///}, pages = {209 - 215}, abstract = {This paper describes a system for tracking people and vehicles for stationary-camera visual surveillance. The appearance of objects being tracked is modeled using mixtures of mixtures of Gaussians. Particles filters are used to track the states of object. Results show the robustness of the system to various lighting and object conditions.}, author = {Abd-Almageed, Wael and Davis, Larry S.} } @conference {17360, title = {Similarity-Based Forecasting with Simultaneous Previews: A River Plot Interface for Time Series Forecasting}, booktitle = {Information Visualization, 2007. IV {\textquoteright}07. 11th International Conference}, year = {2007}, month = {2007/07/04/6}, pages = {191 - 196}, publisher = {IEEE}, organization = {IEEE}, abstract = {Time-series forecasting has a large number of applications. Users with a partial time series for auctions, new stock offerings, or industrial processes desire estimates of the future behavior. We present a data driven forecasting method and interface called similarity-based forecasting (SBF). A pattern matching search in an historical time series dataset produces a subset of curves similar to the partial time series. The forecast is displayed graphically as a river plot showing statistical information about the SBF subset. A forecasting preview interface allows users to interactively explore alternative pattern matching parameters and see multiple forecasts simultaneously. User testing with 8 users demonstrated advantages and led to improvements.}, keywords = {data driven forecasting method, data visualisation, Data visualization, Economic forecasting, forecasting preview interface, Graphical user interfaces, historical time series dataset, Laboratories, new stock offerings, partial time series, pattern matching, pattern matching search, Predictive models, river plot interface, Rivers, similarity-based forecasting, Smoothing methods, Technological innovation, Testing, time series, time series forecasting, Weather forecasting}, isbn = {0-7695-2900-3}, doi = {10.1109/IV.2007.101}, author = {Buono,P. and Plaisant, Catherine and Simeone,A. and Aris,A. and Shneiderman, Ben and Shmueli,G. and Jank,W.} } @article {15542, title = {A simple entropy-based algorithm for planar point location}, journal = {ACM Transactions on Algorithms (TALG)}, volume = {3}, year = {2007}, month = {2007/05//}, abstract = {Given a planar polygonal subdivision S, point location involves preprocessing this subdivision into a data structure so that given any query point q, the cell of the subdivision containing q can be determined efficiently. Suppose that for each cell z in the subdivision, the probability pz that a query point lies within this cell is also given. The goal is to design the data structure to minimize the average search time. This problem has been considered before, but existing data structures are all quite complicated. It has long been known that the entropy H of the probability distribution is the dominant term in the lower bound on the average-case search time. In this article, we show that a very simple modification of a well-known randomized incremental algorithm can be applied to produce a data structure of expected linear size that can answer point-location queries in O(H) average time. We also present empirical evidence for the practical efficiency of this approach.}, keywords = {Entropy, expected-case complexity, Point location, polygonal subdivision, randomized algorithms, trapezoidal maps}, isbn = {1549-6325}, doi = {10.1145/1240233.1240240}, url = {http://doi.acm.org/10.1145/1240233.1240240}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @article {16080, title = {Something that is interesting is interesting then: Using text minig and visualizations to aid interpreting repetition in Gertrude Stein{\textquoteright}s The Making of Americans}, journal = {Proceedings of the Digital Humanities Conference}, year = {2007}, month = {2007///}, pages = {40 - 44}, author = {Clement,T. and Don,A. and Plaisant, Catherine and Auvil,L. and Pape,G. and Goren,V.} } @article {17779, title = {Story creation from heterogeneous data sources}, journal = {Multimedia Tools and Applications}, volume = {33}, year = {2007}, month = {2007///}, pages = {351 - 377}, abstract = {There are numerous applications where there is a need to rapidly infer a story about a given subject from a given set of potentially heterogeneous data sources. In this paper, we formally define a story to be a set of facts about a given subject that satisfies a {\textquotedblleft}story length{\textquotedblright} constraint. An optimal story is a story that maximizes the value of an objective function measuring the goodness of a story. We present algorithms to extract stories from text and other data sources. We also develop an algorithm to compute an optimal story, as well as three heuristic algorithms to rapidly compute a suboptimal story. We run experiments to show that constructing stories can be efficiently performed and that the stories constructed by these heuristic algorithms are high quality stories. We have built a prototype STORY system based on our model{\textemdash}we briefly describe the prototype as well as one application in this paper.}, doi = {10.1007/s11042-007-0100-4}, author = {Fayzullin,M. and V.S. Subrahmanian and Albanese, M. and Cesarano,C. and Picariello, A.} } @conference {12589, title = {Symmetric Objects are Hardly Ambiguous}, booktitle = {Computer Vision and Pattern Recognition, 2007. CVPR {\textquoteright}07. IEEE Conference on}, year = {2007}, month = {2007/06//}, pages = {1 - 7}, abstract = {Given any two images taken under different illumination conditions, there always exist a physically realizable object which is consistent with both the images even if the lighting in each scene is constrained to be a known point light source at infinity. In this paper, we show that images are much less ambiguous for the class of bilaterally symmetric Lambertian objects. In fact, the set of such objects can be partitioned into equivalence classes such that it is always possible to distinguish between two objects belonging to different equivalence classes using just one image per object. The conditions required for two objects to belong to the same equivalence class are very restrictive, thereby leading to the conclusion that images of symmetric objects are hardly ambiguous. The observation leads to an illumination-invariant matching algorithm to compare images of bilaterally symmetric Lambertian objects. Experiments on real data are performed to show the implications of the theoretical result even when the symmetry and Lambertian assumptions are not strictly satisfied.}, keywords = {algorithm;symmetric, bilaterally, images;image, Lambertian, MATCHING, matching;, objects, objects;illumination-invariant, objects;symmetric, symmetric}, doi = {10.1109/CVPR.2007.383289}, author = {Aggarwal,G. and Biswas,S. and Chellapa, Rama} } @conference {16008, title = {Toward domain-neutral human-level metacognition}, booktitle = {AAAI Spring Symposium on Logical Formalizations of Commonsense Reasoning}, year = {2007}, month = {2007///}, author = {Anderson,M. L and Schmill,M. and Oates,T. and Perlis, Don and Josyula,D. and Wright,D. and Wilson,S.} } @article {18869, title = {Towards the development of a virtual environment-based training system for mechanical assembly operations}, journal = {Virtual Reality}, volume = {11}, year = {2007}, month = {2007///}, pages = {189 - 206}, abstract = {In this paper, we discuss the development of Virtual Training Studio (VTS), a virtual environment-based training system that allows training supervisors to create training instructions and allows trainees to learn assembly operations in a virtual environment. Our system is mainly focused on the cognitive side of training so that trainees can learn to recognize parts, remember assembly sequences, and correctly orient the parts during assembly operations. Our system enables users to train using the following three training modes: (1) Interactive Simulation, (2) 3D Animation, and (3) Video. Implementing these training modes required us to develop several new system features. This paper presents an overview of the VTS system and describes a few main features of the system. We also report user test results that show how people train using our system. The user test results indicate that the system is able to support a wide variety of training preferences and works well to support training for assembly operations.}, keywords = {Computer science}, isbn = {1359-4338}, doi = {10.1007/s10055-007-0076-4}, url = {http://www.springerlink.com/content/g7760422m13l83x1/abstract/}, author = {Brough,John and Schwartz,Maxim and Gupta, Satyandra K. and Anand,Davinder and Kavetsky,Robert and Pettersen,Ralph} } @conference {17781, title = {T-rex: A domain-independent system for automated cultural information extraction}, booktitle = {Proceedings of the First International Conference on Computational Cultural Dynamics (ICCCD 2007)}, year = {2007}, month = {2007///}, abstract = {RDF (Resource Description Framework) is a web standarddefined by the World Wide Web Consortium. In RDF, we can define schemas of interest. For example, we can define a schema about tribes on the Pakistan-Afghanistan borderland, or a schema about violent events. An RDF instance is a set of facts that are compatible with the schema. The principal contribution of this paper is the development of a scalable system called T-REX (short for {\textquotedblleft}The RDF EXtractor{\textquotedblright}) that allows us to extract instances associated with a user-specified schema, independently of the domain about which we wish to extract data. Using T-REX, we have successfully extracted information about various aspects of about 20 tribes living in the Pakistan-Afghanistan border. Moreover, we have used T-REX to successfully extract occurrences of violent events from a set of 80 news sites in approximately 50 countries. T-REX scales well {\textendash} it has processed approximately 45,000 web pages per day for the last 6 months. }, author = {Albanese, M. and V.S. Subrahmanian} } @article {16386, title = {Understanding the Semantics of the Genetic Algorithm in Dynamic Environments}, journal = {Applications of Evolutionary Computing}, year = {2007}, month = {2007///}, pages = {657 - 667}, abstract = {Researchers examining genetic algorithms (GAs) in applied settings rarely have access to anything other than fitness values of the best individuals to observe the behavior of the GA. In particular, researchers do not know what schemata are present in the population. Even when researchers look beyond best fitness values, they concentrate on either performance related measures like average fitness and robustness, or low-level descriptions like bit-level diversity measures. To understand the behavior of the GA on dynamic problems, it would be useful to track what is occurring on the {\textquotedblleft}semantic{\textquotedblright} level of schemata. Thus in this paper we examine the evolving {\textquotedblleft}content{\textquotedblright} in terms of schemata, as the GA solves dynamic problems. This allows us to better understand the behavior of the GA in dynamic environments. We finish by summarizing this knowledge and speculate about future work to address some of the new problems that we discovered during these experiments.}, doi = {10.1007/978-3-540-71805-5_72}, author = {Alharbi,A. and Rand, William and Riolo,R} } @conference {16662, title = {Using paraphrases for parameter tuning in statistical machine translation}, booktitle = {Proceedings of the Second Workshop on Statistical Machine Translation}, year = {2007}, month = {2007///}, pages = {120 - 127}, author = {Madnani,N. and Ayan,N.F. and Resnik, Philip and Dorr, Bonnie J} } @conference {13889, title = {Using paraphrases for parameter tuning in statistical machine translation}, booktitle = {Proceedings of the Second Workshop on Statistical Machine Translation}, series = {StatMT {\textquoteright}07}, year = {2007}, month = {2007///}, pages = {120 - 127}, publisher = {Association for Computational Linguistics}, organization = {Association for Computational Linguistics}, address = {Stroudsburg, PA, USA}, abstract = {Most state-of-the-art statistical machine translation systems use log-linear models, which are defined in terms of hypothesis features and weights for those features. It is standard to tune the feature weights in order to maximize a translation quality metric, using held-out test sentences and their corresponding reference translations. However, obtaining reference translations is expensive. In this paper, we introduce a new full-sentence paraphrase technique, based on English-to-English decoding with an MT system, and we demonstrate that the resulting paraphrases can be used to drastically reduce the number of human reference translations needed for parameter tuning, without a significant decrease in translation quality.}, url = {http://dl.acm.org/citation.cfm?id=1626355.1626371}, author = {Madnani,Nitin and Ayan,Necip Fazil and Resnik, Philip and Dorr, Bonnie J} } @conference {18766, title = {Using virtual demonstrations for creating multi-media training instructions}, year = {2007}, month = {2007///}, abstract = {Often generating training instructions for virtual environments is a long and tedious process. In thispaper, we discuss the development of a virtual environment (VE) instruction generating tool called Virtual Author which is the main component of the Virtual Training Studio (VTS). VTS is a virtual environment-based training system that provides instructors with a tool to create training instructions and allows trainees to learn assembly operations in a personal virtual environment. The Virtual Author tool is designed to allow an instructor to perform virtual demonstrations using CAD models in the virtual environment in order to quickly generate VE-based training instructions for use in VTS. This paper describes the algorithms used to carry out motion smoothening of instructor{\textquoteright}s actions, automated text instruction generation based on part and assembly motions, and extraction of alignment constraints from 3D CAD models to support instruction generation. We also present examples to illustrate how the use of the Virtual Author tool leads to a significant reduction in the training instruction generation time. }, url = {http://www.cadanda.com/CAD_4_1-4__11.PDF}, author = {Schwartz,M. and Gupta,S.K. and Anand,D. K. and Brough,J. E. and Kavetsky,R.} } @article {12923, title = {Viable but nonculturable Vibrio cholerae O1 in biofilms in the aquatic environment and their role in cholera transmission}, journal = {Proceedings of the National Academy of Sciences}, volume = {104}, year = {2007}, month = {2007///}, pages = {17801 - 17801}, abstract = {Vibrio cholerae persists in aquatic environments predominantly in a nonculturable state. In this study coccoid, nonculturable V. cholerae O1 in biofilms maintained for 495 days in Mathbaria, Bangladesh, pond water became culturable upon animal passage. Culturability, biofilm formation, and the wbe, ctxA, and rstR2 genes were monitored by culture, direct fluorescent antibody (DFA), and multiplex PCR. DFA counts were not possible after formation of biofilm. Furthermore, wbe, but not ctxA, were amplifiable, even after incubation for 54 and 68 days at room temperature (≈25{\textdegree}C) and 4{\textdegree}C, respectively, when no growth was detectable. Slower biofilm formation and extended culturability were observed for cultures incubated at 4{\textdegree}C, compared with ≈25{\textdegree}C, suggesting biofilm production to be temperature dependent and linked to loss of culturability. Small colonies appearing after incubation in microcosms for 54 and 68 days at 25{\textdegree}C and 4{\textdegree}C, respectively, were wbe positive and ctxA and rstR2 negative, indicating loss of bacteriophage CTXΦ. The coccoid V. cholerae O1 observed as free cells in microcosms incubated for 495 days could not be cultured, but biofilms in the same microcosms yielded culturable cells. It is concluded that biofilms can act as a reservoir for V. cholerae O1 between epidemics because of its long-term viability in biofilms. In contrast to biofilms produced in Mathbaria pond water, V. cholerae O1 in biofilms present in cholera stools and incubated under identical conditions as the Mathbaria pond water biofilms could not be cultured after 2 months, indicating that those V. cholerae cells freshly discharged into the environment are significantly less robust than cells adapted to environmental conditions.Bangladesh bacteriophage CTXΦ DFA multiplex-PCR ctxA }, isbn = {0027-8424, 1091-6490}, doi = {10.1073/pnas.0705599104}, url = {http://www.pnas.org/content/104/45/17801}, author = {Alam,M. and Sultana,M. and Nair,G. B. and Siddique,A. K. and Hasan,N. A. and Sack,R. B. and Sack,D. A. and Ahmed,K. U. and Sadique,A. and Watanabe,H. and Rita R Colwell} } @conference {12578, title = {Video Biometrics}, booktitle = {Image Analysis and Processing, 2007. ICIAP 2007. 14th International Conference on}, year = {2007}, month = {2007/09//}, pages = {363 - 370}, abstract = {A strong requirement to come up with secure and user- friendly ways to authenticate and identify people, to safeguard their rights and interests, has probably been the main guiding force behind biometrics research. Though a vast amount of research has been done to recognize humans based on still images, the problem is still far from solved for unconstrained scenarios. This has led to an increased interest in using video for the task of biometric recognition. Not only does video provide more information, but also is more suitable for recognizing humans in general surveillance scenarios. Other than the multitude of still frames, video makes it possible to characterize biometrics based on inherent dynamics like gait which is not possible with still images. In this paper, we describe several recent algorithms to illustrate the usefulness of videos to identify humans. A brief discussion on remaining challenges is also included.}, keywords = {(access, analysis;video, biometrics;biometrics, control);face, dynamics;ofbiometric, images;surveillance, inherent, MOTION, processing;, recognition;image, recognition;still, scenarios;unconstrained, scenarios;video, signal}, doi = {10.1109/ICIAP.2007.4362805}, author = {Chellapa, Rama and Aggarwal,G.} } @article {11970, title = {View-invariant modeling and recognition of human actions using grammars}, journal = {Dynamical Vision}, year = {2007}, month = {2007///}, pages = {115 - 126}, author = {Ogale, A. and Karapurkar, A. and Aloimonos, J.} } @conference {18880, title = {Accident or intention: That is the question (in the noisy iterated prisoner{\textquoteright}s dilemma)}, year = {2006}, month = {2006///}, pages = {561 - 568}, abstract = {This paper focuses on the Noisy Iterated Prisoner{\textquoteright}sDilemma, a version of the Iterated Prisoner{\textquoteright}s Dilemma (IPD) in which there is a nonzero probability that a {\textquotedblleft}co- operate{\textquotedblright} action will accidentally be changed into a {\textquotedblleft}defect{\textquotedblright} action and vice versa. Tit-For-Tat and other strategies that do quite well in the ordinary (non-noisy) IPD can do quite badly in the Noisy IPD. This paper presents a technique called symbolic noise de- tection, for detecting whether anomalies in player{\textquoteright}s behavior are deliberate or accidental. The key idea is to construct a model of the other agent{\textquoteright}s behavior, and watch for any de- viation from this model. If the other agent{\textquoteright}s next action is inconsistent with this model, the inconsistency can be due either to noise or to a genuine change in their behavior; and we can often distinguish between two cases by waiting to see whether this inconsistency persists in next few moves. We entered several different versions of our strategy in the 2005 Iterated Prisoner{\textquoteright}s Dilemma competition, in Cat- egory 2 (noisy environments). Out of the 165 contestants in this category, our programs consistently ranked among top ten. The best of our programs ranked third, and it was beaten only by two {\textquotedblleft}master-slave strategy{\textquotedblright} programs that each had a large number of {\textquotedblleft}slave{\textquotedblright} programs feeding points to them. }, url = {http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.120.7640\&rep=rep1\&type=pdf}, author = {Au,T. C and Nau, Dana S.} } @conference {15244, title = {Achieving anonymity via clustering}, booktitle = {Proceedings of the twenty-fifth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems}, series = {PODS {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {153 - 162}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Publishing data for analysis from a table containing personal records, while maintaining individual privacy, is a problem of increasing importance today. The traditional approach of de-identifying records is to remove identifying fields such as social security number, name etc. However, recent research has shown that a large fraction of the US population can be identified using non-key attributes (called quasi-identifiers) such as date of birth, gender, and zip code [15]. Sweeney [16] proposed the k-anonymity model for privacy where non-key attributes that leak information are suppressed or generalized so that, for every record in the modified table, there are at least k-1 other records having exactly the same values for quasi-identifiers. We propose a new method for anonymizing data records, where quasi-identifiers of data records are first clustered and then cluster centers are published. To ensure privacy of the data records, we impose the constraint that each cluster must contain no fewer than a pre-specified number of data records. This technique is more general since we have a much larger choice for cluster centers than k-Anonymity. In many cases, it lets us release a lot more information without compromising privacy. We also provide constant-factor approximation algorithms to come up with such a clustering. This is the first set of algorithms for the anonymization problem where the performance is independent of the anonymity parameter k. We further observe that a few outlier points can significantly increase the cost of anonymization. Hence, we extend our algorithms to allow an ε fraction of points to remain unclustered, i.e., deleted from the anonymized publication. Thus, by not releasing a small fraction of the database records, we can ensure that the data published for analysis has less distortion and hence is more useful. Our approximation algorithms for new clustering objectives are of independent interest and could be applicable in other clustering scenarios as well.}, keywords = {anonymity, Approximation algorithms, clustering, privacy}, isbn = {1-59593-318-2}, doi = {10.1145/1142351.1142374}, url = {http://doi.acm.org/10.1145/1142351.1142374}, author = {Aggarwal,Gagan and Feder,Tom{\'a}s and Kenthapadi,Krishnaram and Khuller, Samir and Panigrahy,Rina and Thomas,Dilys and Zhu,An} } @conference {12093, title = {Applying flow-sensitive CQUAL to verify MINIX authorization check placement}, booktitle = {Proceedings of the 2006 workshop on Programming languages and analysis for security}, series = {PLAS {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {3 - 6}, publisher = {ACM}, organization = {ACM}, address = {Ottawa, Ontario, Canada}, abstract = {We present the first use of flow-sensitive CQUAL to verify the placement of operating system authorization checks. Our analysis of MINIX 3 system servers and discovery of a non-exploitable Time-Of-Check/Time-Of-Use bug demonstrate the effectiveness of flow sensitive CQUAL and its advantage over earlier flow-insensitive versions. We also identify and suggest alternatives to current CQUAL usability features that encourage analysts to make omissions that cause the otherwise sound tool to produce false-negative results.}, keywords = {access controls, cqual, minix}, isbn = {1-59593-374-3}, doi = {10.1145/1134744.1134747}, author = {Fraser,Timothy and Petroni,Jr. and Arbaugh, William A.} } @article {18653, title = {An architecture for adaptive intrusion-tolerant applications}, journal = {Software: Practice and Experience}, volume = {36}, year = {2006}, month = {2006///}, pages = {1331 - 1354}, abstract = {Applications that are part of a mission-critical information system need to maintain a usable level of key services through ongoing cyber-attacks. In addition to the well-publicized denial of service (DoS) attacks, these networked and distributed applications are increasingly threatened by sophisticated attacks that attempt to corrupt system components and violate service integrity. While various approaches have been explored to deal with DoS attacks, corruption-inducing attacks remain largely unaddressed. We have developed a collection of mechanisms based on redundancy, Byzantine fault tolerance, and adaptive middleware that help distributed, object-based applications tolerate corruption-inducing attacks. In this paper, we present the ITUA architecture, which integrates these mechanisms in a framework for auto-adaptive intrusion-tolerant systems, and we describe our experience in using the technology to defend a critical application that is part of a larger avionics system as an example. We also motivate the adaptive responses that are key to intrusion tolerance, and explain the use of the ITUA architecture to support them in an architectural framework. Copyright {\textcopyright} 2006 John Wiley \& Sons, Ltd.}, keywords = {adaptive defense, adaptive middleware, Byzantine fault tolerance, intrusion tolerance, redundancy, survivability architecture}, isbn = {1097-024X}, doi = {10.1002/spe.747}, url = {http://onlinelibrary.wiley.com/doi/10.1002/spe.747/abstract}, author = {Pal,Partha and Rubel,Paul and Atighetchi,Michael and Webber,Franklin and Sanders,William H. and Seri,Mouna and Ramasamy,HariGovind and Lyons,James and Courtney,Tod and Agbaria,Adnan and Michel Cukier and Gossett,Jeanna and Keidar,Idit} } @conference {12085, title = {An architecture for specification-based detection of semantic integrity violations in kernel dynamic data}, booktitle = {Proceedings of the 15th conference on USENIX Security Symposium}, year = {2006}, month = {2006///}, pages = {20 - 20}, author = {Petroni Jr,N. L and Fraser,T. and Walters,A. A and Arbaugh, William A.} } @article {18007, title = {A bootstrapping model for directional wireless networks}, journal = {Communications Letters, IEEE}, volume = {10}, year = {2006}, month = {2006/12//}, pages = {840 - 842}, abstract = {Initially configuring or bootstrapping a connected topology in directional wireless networks is a challenging problem, especially when nodes only have local connectivity information and a limited number of transceivers. This paper presents a scalable bootstrapping model which integrates: 1) a distributed bottom-up algorithm that constructs a spanning tree with degree at most one larger than the optimal 2) a resource discovery algorithm for efficient dissemination of local connectivity information, and 3) synchronization protocols to guarantee the efficient emergence of overall network connectivity from local interactions. We investigate the feasibility and scalability of the proposed model. Results are presented for different network systems, with varying size and signaling data rates.}, keywords = {(mathematics);, algorithm;resource, algorithm;spanning, bootstrapping, bottom-up, discovery, model;directional, network;distributed, networks;trees, protocols;protocols;radio, tree;synchronization, wireless}, isbn = {1089-7798}, doi = {10.1109/LCOMM.2006.060808}, author = {Milner,S. and Llorca,J. and Anibha,A. and Vishkin, Uzi} } @article {18531, title = {Challenges and Opportunities in Internet Data Mining}, volume = {CMU-PDL-06-102}, year = {2006}, month = {2006///}, pages = {44 - 44}, institution = {Parallel Data Laboratory, Carnegie Mellon University}, abstract = {Internet measurement data provides the foundation for the operation and planning of the networks that comprise the Internet, and is a necessary component in research for analysis, simulation, and emulation. Despite its critical role, however, the management of this data{\textemdash}from collection and transmission to storage and its use within applications{\textemdash}remains primarily ad hoc, using techniques created and re-created by each corporation or researcher that uses the data. This paper examines several of the challenges faced when attempting to collect and archive large volumes of network measurement data, and outlines an architecture for an Internet data repository{\textemdash}the datapository designed to create a framework for collaboratively addressing these challenges.}, author = {Andersen,D.G. and Feamster, Nick} } @article {17544, title = {A client-driven approach for channel management in wireless LANs}, journal = {IEEE Infocom}, volume = {6}, year = {2006}, month = {2006///}, abstract = {We propose an efficient client-based approach for channel man-agement (channel assignment and load balancing) in 802.11-based WLANs that lead to better usage of the wireless spectrum. This approach is based on a {\textquotedblleft}conflict set coloring{\textquotedblright} formulation that jointly performs load balancing along with channel assignment. Such a formulation has a number of advantages. First, it explicitly captures interference effects at clients. Next, it intrinsically ex- poses opportunities for better channel re-use. Finally, algorithms based on this formulation do not depend on specific physical RF models and hence can be applied efficiently to a wide-range of in- building as well as outdoor scenarios. We have performed extensive packet-level simulations and mea- surements on a deployed wireless testbed of 70 APs to validate the performance of our proposed algorithms. We show that in addi- tion to single network scenarios, the conflict set coloring formu- lation is well suited for channel assignment where multiple wire- less networks share and contend for spectrum in the same physical space. Our results over a wide range of both simulated topologies and in-building testbed experiments indicate that our approach improves application level performance at the clients by upto three times (and atleast 50\%) in comparison to current best-known tech- niques. }, author = {Mishra,A. and Brik,V. and Banerjee,S. and Srinivasan, Aravind and Arbaugh, William A.} } @article {17574, title = {Client-driven channel management for wireless LANs}, journal = {SIGMOBILE Mob. Comput. Commun. Rev.}, volume = {10}, year = {2006}, month = {2006/10//}, pages = {8 - 10}, abstract = {With the explosive growth in the density of 802.11 access points (APs) in the form of hotspots and public/home access networks, coordinating the shared use of spectrum has become an important problem. The irregular coverage topologies present in WLANs due to the vagaries of the indoor RF environment make the channel assignment algorithms in cellular networks inapplicable [1, 2].}, isbn = {1559-1662}, doi = {10.1145/1215976.1215981}, url = {http://doi.acm.org/10.1145/1215976.1215981}, author = {Mishra,Arunesh and Brik,Vladimir and Banerjee,Suman and Srinivasan, Aravind and Arbaugh, William A.} } @article {15619, title = {The cost of compatible refinement of simplex decomposition trees}, journal = {Proceedings of the 15th International Meshing Roundtable}, year = {2006}, month = {2006///}, pages = {57 - 69}, abstract = {A hierarchical simplicial mesh is a recursive decomposition of space into cells that are simplices. Such a mesh is compatible if pairs of neighboring cells meet along a single common face. Compatibility condition is important in many applications where the mesh serves as a discretization of a function. Enforcing compatibility involves refining the simplices of the mesh further, thus generates a larger mesh. We show that the size of a simplicial mesh grows by no more than a constant factor when compatibly refined. We prove a tight upper bound on the expansion factor for 2-dimensional meshes, and we sketch upper bounds for d-dimensional meshes.}, doi = {10.1007/978-3-540-34958-7_4}, author = {Atalay,F. and Mount, Dave} } @conference {11973, title = {Deformation and viewpoint invariant color histograms}, booktitle = {British Machine Vision Conference}, volume = {2}, year = {2006}, month = {2006///}, pages = {509 - 518}, author = {Domke, J. and Aloimonos, J.} } @article {13185, title = {Density estimation using mixtures of mixtures of Gaussians}, journal = {Computer Vision{\textendash}ECCV 2006}, year = {2006}, month = {2006///}, pages = {410 - 422}, abstract = {In this paper we present a new density estimation algorithm using mixtures of mixtures of Gaussians. The new algorithm overcomes the limitations of the popular Expectation Maximization algorithm. The paper first introduces a new model selection criterion called the Penalty-less Information Criterion, which is based on the Jensen-Shannon divergence. Mean-shift is used to automatically initialize the means and covariances of the Expectation Maximization in order to obtain better structure inference. Finally, a locally linear search is performed using the Penalty-less Information Criterion in order to infer the underlying density of the data. The validity of the algorithm is verified using real color images.}, author = {Abd-Almageed, Wael and Davis, Larry S.} } @article {14605, title = {Differential Transcriptional Response to Nonassociative and Associative Components of Classical Fear Conditioning in the Amygdala and Hippocampus}, journal = {Learning \& MemoryLearn. Mem.}, volume = {13}, year = {2006}, month = {2006/03/01/}, pages = {135 - 142}, abstract = {Classical fear conditioning requires the recognition of conditioned stimuli (CS) and the association of the CS with an aversive stimulus. We used Affymetrix oligonucleotide microarrays to characterize changes in gene expression compared to naive mice in both the amygdala and the hippocampus 30 min after classical fear conditioning and 30 min after exposure to the CS in the absence of an aversive stimulus. We found that in the hippocampus, levels of gene regulation induced by classical fear conditioning were not significantly greater than those induced by CS alone, whereas in the amygdala, classical fear conditioning did induce significantly greater levels of gene regulation compared to the CS. Computational studies suggest that transcriptional changes in the hippocampus and amygdala are mediated by large and overlapping but distinct combinations of molecular events. Our results demonstrate that an increase in gene regulation in the amygdala was partially correlated to associative learning and partially correlated to nonassociative components of the task, while gene regulation in the hippocampus was correlated to nonassociative components of classical fear conditioning, including configural learning.}, isbn = {1072-0502, 1549-5485}, doi = {10.1101/lm.86906}, url = {http://learnmem.cshlp.org/content/13/2/135}, author = {Keeley,Michael B and Wood,Marcelo A and Isiegas,Carolina and Stein,Joel and Hellman,Kevin and Hannenhalli, Sridhar and Abel,Ted} } @conference {16870, title = {Distance join queries on spatial networks}, booktitle = {Proceedings of the 14th annual ACM international symposium on Advances in geographic information systems}, series = {GIS {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {211 - 218}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {The result of a distance join operation on two sets of objects R, S on a spatial network G is a set P of object pairs pq, p {\'E} R, q {\'E} S such that the distance of an object pair pq is the shortest distance from p to q in G. Several variations to the distance join operation such as UnOrdered, Incremental, topk, Semi-Join impose additional constraints on the distance between the object pairs in P, the ordering of object pairs in P, and on the cardinality of P. A distance join algorithm on spatial networks is proposed that works in conjunction with the SILC framework, which is a new approach to query processing on spatial networks. Experimental results demonstrate up to an order of magnitude speed up when compared with a prominent existing technique.}, keywords = {location-based services, path coherence, Query processing, SILC framework, Spatial databases, spatial networks}, isbn = {1-59593-529-0}, doi = {10.1145/1183471.1183506}, url = {http://doi.acm.org/10.1145/1183471.1183506}, author = {Sankaranarayanan,Jagan and Alborzi,Houman and Samet, Hanan} } @conference {12643, title = {Edge Suppression by Gradient Field Transformation Using Cross-Projection Tensors}, booktitle = {Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on}, volume = {2}, year = {2006}, month = {2006///}, pages = {2301 - 2308}, abstract = {We propose a new technique for edge-suppressing operations on images. We introduce cross projection tensors to achieve affine transformations of gradient fields. We use these tensors, for example, to remove edges in one image based on the edge-information in a second image. Traditionally, edge suppression is achieved by setting image gradients to zero based on thresholds. A common application is in the Retinex problem, where the illumination map is recovered by suppressing the reflectance edges, assuming it is slowly varying. We present a class of problems where edge-suppression can be a useful tool. These problems involve analyzing images of the same scene under variable illumination. Instead of resetting gradients, the key idea in our approach is to derive local tensors using one image and to transform the gradient field of another image using them. Reconstructed image from the modified gradient field shows suppressed edges or textures at the corresponding locations. All operations are local and our approach does not require any global analysis. We demonstrate the algorithm in the context of several applications such as (a) recovering the foreground layer under varying illumination, (b) estimating intrinsic images in non-Lambertian scenes, (c) removing shadows from color images and obtaining the illumination map, and (d) removing glass reflections.}, doi = {10.1109/CVPR.2006.106}, author = {Agrawal,A. and Raskar, R. and Chellapa, Rama} } @article {12930, title = {Effect of transport at ambient temperature on detection and isolation of Vibrio cholerae from environmental samples}, journal = {Applied and environmental microbiology}, volume = {72}, year = {2006}, month = {2006///}, pages = {2185 - 2190}, abstract = {It has long been assumed that prolonged holding of environmental samples at the ambient air temperature prior to bacteriological analysis is detrimental to isolation and detection of Vibrio cholerae, the causative agent of pandemic cholera. The present study was aimed at understanding the effect of transporting environmental samples at the ambient air temperature on isolation and enumeration of V. cholerae. For water and plankton samples held at ambient temperatures ranging from 31{\textdegree}C to 35{\textdegree}C for 20 h, the total counts did not increase significantly but the number of culturable V. cholerae increased significantly compared to samples processed within 1 h of collection, as measured by culture, acridine orange direct count, direct fluorescent-antibody-direct viable count (DFA-DVC), and multiplex PCR analyses. For total coliform counts, total bacterial counts, and DFA-DVC counts, the numbers did not increase significantly, but the culturable plate counts for V. cholerae increased significantly after samples were held at the ambient temperature during transport to the laboratory for analysis. An increase in the recovery of V. cholerae O1 and improved detection of V. cholerae O1 rfb and ctxA also occurred when samples were enriched after they were kept for 20 h at the ambient temperature during transport. Improved detection and isolation of toxigenic V. cholerae from freshwater ecosystems can be achieved by holding samples at the ambient temperature, an observation that has significant implications for tracking this pathogen in diverse aquatic environments.}, author = {Alam,M. and Sadique,A. and Bhuiyan,N. A. and Nair,G. B. and Siddique,A. K. and Sack,D. A. and Ahsan,S. and Huq,A. and Sack,R. B. and Rita R Colwell and others} } @conference {16865, title = {Enabling Query Processing on Spatial Networks}, booktitle = {Data Engineering, 2006. ICDE {\textquoteright}06. Proceedings of the 22nd International Conference on}, year = {2006}, month = {2006/04//}, pages = {163 - 163}, abstract = {A system that enables real time query processing on large spatial networks is demonstrated. The system provides functionality for processing a wide range of spatial queries such as nearest neighbor searches and spatial joins on spatial networks of sufficiently large sizes.}, doi = {10.1109/ICDE.2006.60}, author = {Sankaranarayanan,J. and Alborzi,H. and Samet, Hanan} } @article {15932, title = {Enhancing reinforcement learning with metacognitive monitoring and control for improved perturbation tolerance}, journal = {Journal of Experimental and Theoretical Artificial Intelligence}, volume = {18}, year = {2006}, month = {2006///}, pages = {387 - 411}, author = {Anderson,M. L and Oates,T. and Chong,W. and Perlis, Don} } @article {12135, title = {Experiments to understand HPC time to development}, journal = {CTWatch Quarterly}, year = {2006}, month = {2006///}, abstract = {In order to understand how high performance computing (HPC) programs are developed, a series of experiments, using students in graduate level HPC classes, have been conducted at many universities in the US. In this paper we discuss the general process of conducting those experiments, give some of the early results of those experiments, and describe a web-based process we are developing that will allow us to run additional experiments at other universities and laboratories that will be easier to conduct and generate results that more accurately reflect the process of building HPC programs.}, keywords = {hackystat, HPC, publications-journals}, url = {http://csdl.ics.hawaii.edu/techreports/06-08/06-08.pdf}, author = {Hochstein, Lorin and Nakamura,Taiga and Basili, Victor R. and Asgari, Sima and Zelkowitz, Marvin V and Hollingsworth, Jeffrey K and Shull, Forrest and Carver,Jeffrey and Voelp,Martin and Zazworka, Nico and Johnson,Philip} } @article {14435, title = {Exploiting independence for branch operations in Bayesian learning of C\&RTs}, journal = {Probabilistic, Logical and Relational Learning-Towards a Synthesis}, year = {2006}, month = {2006///}, abstract = {In this paper we extend a methodology for Bayesian learning via MCMC, with the ability to grow arbitrarily long branches in C\&RT models. We are able to do so by exploiting independence in the model construction process. The ability to grow branches rather than single nodes has been noted as desirable in the literature. The most singular feature of the underline methodology used here in comparison to other approaches is the coupling of the prior and the proposal. The main contribution of this paper is to show how taking advantage of independence in the coupled process, can allow branch growing and swapping for proposal models.}, author = {Angelopoulos,N. and Cussens,J. and De Raedt,L. and Dietterich,T. and Getoor, Lise and Muggleton,S. H} } @article {16081, title = {Exploring auction databases through interactive visualization}, journal = {Decision Support Systems}, volume = {42}, year = {2006}, month = {2006/12//}, pages = {1521 - 1538}, abstract = {We introduce AuctionExplorer, a suite of tools for exploring databases of online auctions. The suite combines tools for collecting, processing, and interactively exploring auction attributes (e.g., seller rating), and the bid history (price evolution represented as a time series). Part of AuctionExplorer{\textquoteright}s power comes from its coupling of the two information structures, thereby allowing exploration of relationships between them. Exploration can be directed by hypothesis testing or exploratory data analysis. We propose a process for visual data analysis and illustrate AuctionExplorer{\textquoteright}s operations with a dataset of eBay auctions. Insights may improve seller, bidder, auction house, and other vendors{\textquoteright} understanding of the market, thereby assisting their decision making process.}, keywords = {Auction dynamics, Bid history, Online auctions, time series, user interface}, isbn = {0167-9236}, doi = {10.1016/j.dss.2006.01.001}, url = {http://www.sciencedirect.com/science/article/pii/S0167923606000042}, author = {Shmueli,Galit and Jank,Wolfgang and Aris,Aleks and Plaisant, Catherine and Shneiderman, Ben} } @conference {16087, title = {Exploring erotics in Emily Dickinson{\textquoteright}s correspondence with text mining and visual interfaces}, booktitle = {Proceedings of the 6th ACM/IEEE-CS joint conference on Digital libraries}, series = {JCDL {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {141 - 150}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {This paper describes a system to support humanities scholars in their interpretation of literary work. It presents a user interface and web architecture that integrates text mining, a graphical user interface and visualization, while attempting to remain easy to use by non specialists. Users can interactively read and rate documents found in a digital libraries collection, prepare training sets, review results of classification algorithms and explore possible indicators and explanations. Initial evaluation steps suggest that there is a rationale for "provocational" text mining in literary interpretation.}, keywords = {case studies, humanities, literary criticism, text mining, user interface, Visualization}, isbn = {1-59593-354-9}, doi = {10.1145/1141753.1141781}, url = {http://doi.acm.org/10.1145/1141753.1141781}, author = {Plaisant, Catherine and Rose,James and Yu,Bei and Auvil,Loretta and Kirschenbaum,Matthew G. and Smith,Martha Nell and Clement,Tanya and Lord,Greg} } @conference {15306, title = {Exploring the limits of single-iteration clarification dialogs}, booktitle = {Proceedings of the 29th annual international ACM SIGIR conference on Research and development in information retrieval}, series = {SIGIR {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {469 - 476}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Single-iteration clarification dialogs, as implemented in the TREC HARD track, represent an attempt to introduce interaction into ad hoc retrieval, while preserving the many benefits of large-scale evaluations. Although previous experiments have not conclusively demonstrated performance gains resulting from such interactions, it is unclear whether these findings speak to the nature of clarification dialogs, or simply the limitations of current systems. To probe the limits of such interactions, we employed a human intermediary to formulate clarification questions and exploit user responses. In addition to establishing a plausible upper bound on performance, we were also able to induce an "ontology of clarifications" to characterize human behavior. This ontology, in turn, serves as the input to a regression model that attempts to determine which types of clarification questions are most helpful. Our work can serve to inform the design of interactive systems that initiate user dialogs.}, keywords = {interactive retrieval, intermediated search, TREC HARD}, isbn = {1-59593-369-7}, doi = {10.1145/1148170.1148251}, url = {http://doi.acm.org/10.1145/1148170.1148251}, author = {Jimmy Lin and Wu,Philip and Demner-Fushman,Dina and Abels,Eileen} } @article {14909, title = {First Steps Toward an Electronic Field Guide for Plants}, journal = {Taxon}, volume = {55}, year = {2006}, month = {2006///}, pages = {597 - 610}, abstract = {this paper, we will describe our progress towards building a digital collection of the Smithsonian{\textquoteright}s type specimens, developing recognition algorithms that can match an image of a leaf to the species of plant from which it comes, and designing user interfaces for interacting with an electronic field guide. To start, we are developing a prototype electronic field guide for the flora of Plummers Island, a small, well-studied island in the Potomac River. This prototype system contains multiple images for each of about 130 species of plants on the island, and should soon grow to cover all 200+ species currently recorded (Shetler et al., 2005). Images of full specimens are available, as well as images of isolated leaves of each species. A zoomable user interface allows a user to browse these images, zooming in on ones of interest. Visual recognition algorithms assist a botanist in locating the specimens that are most relevant to identify the species of a plant. The system currently runs on a small hand-held computer. We will describe the components of this prototype, and also describe some of the future challenges we anticipate if we are to provide botanists in the field with all the resources that are now currently available in the world{\textquoteright}s museums and herbaria. Type Specimen Digital Collection The first challenge in producing our electronic field guide is to create a digital collection covering all of the Smithsonian{\textquoteright}s 85,000 vascular plant type specimens. For each type specimen, the database should eventually include systematically acquired high-resolution digital images of the specimen, textual descriptions, links to decision trees, images of live plants, and 3D models. Figure 1: On the left, our set-up at the Smithsonian for digitally photographing type specimens. On the...}, author = {Haibin,Gaurav Agarwal and Agarwal,Gaurav and Ling,Haibin and Jacobs, David W. and Shirdhonkar,Sameer and Kress,W. John and Russell,Rusty and Belhumeur,Peter and Dixit,An and Feiner,Steve and Mahajan,Dhruv and Sunkavalli,Kalyan and Ramamoorthi,Ravi and White,Sean} } @article {14282, title = {Flow-insensitive type qualifiers}, journal = {ACM Transactions on Programming Languages and Systems (TOPLAS)}, volume = {28}, year = {2006}, month = {2006///}, pages = {1035 - 1087}, author = {Foster, Jeffrey S. and Johnson,R. and Kodumal,J. and Aiken,A.} } @conference {13781, title = {Going beyond AER: an extensive analysis of word alignments and their impact on MT}, booktitle = {Proceedings of the 21st International Conference on Computational Linguistics and the 44th annual meeting of the Association for Computational Linguistics}, series = {ACL-44}, year = {2006}, month = {2006///}, pages = {9 - 16}, publisher = {Association for Computational Linguistics}, organization = {Association for Computational Linguistics}, address = {Stroudsburg, PA, USA}, abstract = {This paper presents an extensive evaluation of five different alignments and investigates their impact on the corresponding MT system output. We introduce new measures for intrinsic evaluations and examine the distribution of phrases and untranslated words during decoding to identify which characteristics of different alignments affect translation. We show that precision-oriented alignments yield better MT output (translating more words and using longer phrases) than recall-oriented alignments.}, doi = {10.3115/1220175.1220177}, url = {http://dx.doi.org/10.3115/1220175.1220177}, author = {Ayan,Necip Fazil and Dorr, Bonnie J} } @article {11976, title = {Human activity language: Grounding concepts with a linguistic framework}, journal = {Semantic Multimedia}, year = {2006}, month = {2006///}, pages = {86 - 100}, author = {Guerra-Filho,G. and Aloimonos, J.} } @conference {15642, title = {On the importance of idempotence}, booktitle = {Proceedings of the thirty-eighth annual ACM symposium on Theory of computing}, series = {STOC {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {564 - 573}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Range searching is among the most fundamental problems in computational geometry. An n-element point set in Rd is given along with an assignment of weights to these points from some commutative semigroup. Subject to a fixed space of possible range shapes, the problem is to preprocess the points so that the total semigroup sum of the points lying within a given query range η can be determined quickly. In the approximate version of the problem we assume that η is bounded, and we are given an approximation parameter ε > 0. We are to determine the semigroup sum of all the points contained within η and may additionally include any of the points lying within distance ε {\textbullet} diam(η) of η{\textquoteright}s boundar.In this paper we contrast the complexity of range searching based on semigroup properties. A semigroup (S,+) is idempotent if x + x = x for all x ∈ S, and it is integral if for all k >= 2, the k-fold sum x + ... + x is not equal to x. For example, (R, min) and (0,1, ∨) are both idempotent, and (N, +) is integral. To date, all upper and lower bounds hold irrespective of the semigroup. We show that semigroup properties do indeed make a difference for both exact and approximate range searching, and in the case of approximate range searching the differences are dramatic.First, we consider exact halfspace range searching. The assumption that the semigroup is integral allows us to improve the best lower bounds in the semigroup arithmetic model. For example, assuming O(n) storage in the plane and ignoring polylog factors, we provide an Ω*(n2/5) lower bound for integral semigroups, improving upon the best lower bound of Ω*(n1/3), thus closing the gap with the O(n1/2) upper bound.We also consider approximate range searching for Euclidean ball ranges. We present lower bounds and nearly matching upper bounds for idempotent semigroups. We also present lower bounds for range searching for integral semigroups, which nearly match existing upper bounds. These bounds show that the advantages afforded by idempotency can result in major improvements. In particular, assuming roughly linear space, the exponent in the ε-dependencies is smaller by a factor of nearly 1/2. All our results are presented in terms of space-time tradeoffs, and our lower and upper bounds match closely throughout the entire spectrum.To our knowledge, our results provide the first proof that semigroup properties affect the computational complexity of range searching in the semigroup arithmetic model. These are the first lower bound results for any approximate geometric retrieval problems. The existence of nearly matching upper bounds, throughout the range of space-time tradeoffs, suggests that we are close to resolving the computational complexity of both idempotent and integral approximate spherical range searching in the semigroup arithmetic model.}, keywords = {Approximation algorithms, Idempotence, Range searching}, isbn = {1-59593-134-1}, doi = {10.1145/1132516.1132598}, url = {http://doi.acm.org/10.1145/1132516.1132598}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @article {18885, title = {The incompleteness of planning with volatile external information}, journal = {FRONTIERS IN ARTIFICIAL INTELLIGENCE AND APPLICATIONS}, volume = {141}, year = {2006}, month = {2006///}, pages = {839 - 839}, abstract = {In many real-world planning environments, someof the information about the world is both external (the planner must request it from external information sources) and volatile (it changes before the planning process completes). In such en- vironments, a planner faces two challenges: how to generate plans despite changes in the external information during plan- ning, and how to guarantee that a plan returned by the plan- ner will remain valid for some period of time after the plan- ning ends. Previous works on planning with volatile informa- tion have addressed the first challenge, but not the second one. This paper provides a general model for planning with volatile external information in which the planner offers a guar- antee of how long the solution will remain valid after it is re- turned, and an incompleteness theorem showing that there is no planner that can succeed in solving all solvable planning prob- lems in which there is volatile external information. }, url = {http://www.cs.umd.edu/~nau/papers/au2006incompleteness.pdf}, author = {Au,T. and Nau, Dana S.} } @conference {12620, title = {Integrated Motion Detection and Tracking for Visual Surveillance}, booktitle = {Computer Vision Systems, 2006 ICVS {\textquoteright}06. IEEE International Conference on}, year = {2006}, month = {2006/01//}, pages = {28 - 28}, abstract = {Visual surveillance systems have gained a lot of interest in the last few years. In this paper, we present a visual surveillance system that is based on the integration of motion detection and visual tracking to achieve better performance. Motion detection is achieved using an algorithm that combines temporal variance with background modeling methods. The tracking algorithm combines motion and appearance information into an appearance model and uses a particle filter framework for tracking the object in subsequent frames. The systems was tested on a large ground-truthed data set containing hundreds of color and FLIR image sequences. A performance evaluation for the system was performed and the average evaluation results are reported in this paper.}, doi = {10.1109/ICVS.2006.35}, author = {Abdelkader, M.F. and Chellapa, Rama and Qinfen Zheng and Chan, A.L.} } @conference {12014, title = {Integration of visual and inertial information for egomotion: a stochastic approach}, booktitle = {Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006}, year = {2006}, month = {2006/05/15/19}, pages = {2053 - 2059}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present a probabilistic framework for visual correspondence, inertial measurements and egomotion. First, we describe a simple method based on Gabor filters to produce correspondence probability distributions. Next, we generate a noise model for inertial measurements. Probability distributions over the motions are then computed directly from the correspondence distributions and the inertial measurements. We investigate combining the inertial and visual information for a single distribution over the motions. We find that with smaller amounts of correspondence information, fusion of the visual data with the inertial sensor results in much better egomotion estimation. This is essentially because inertial measurements decrease the "translation-rotation" ambiguity. However, when more correspondence information is used, this ambiguity is reduced to such a degree that the inertial measurements provide negligible improvement in accuracy. This suggests that inertial and visual information are more closely integrated in a compositional sense}, keywords = {Computer vision, data mining, Distributed computing, egomotion estimation, Gabor filters, Gravity, inertial information, inertial sensor, Laboratories, Motion estimation, Noise measurement, Probability distribution, probability distributions, Rotation measurement, stochastic approach, Stochastic processes, visual information}, isbn = {0-7803-9505-0}, doi = {10.1109/ROBOT.2006.1642007}, author = {Domke, J. and Aloimonos, J.} } @conference {12611, title = {Invariant Geometric Representation of 3D Point Clouds for Registration and Matching}, booktitle = {Image Processing, 2006 IEEE International Conference on}, year = {2006}, month = {2006/10//}, pages = {1209 - 1212}, abstract = {Though implicit representations of surfaces have often been used for various computer graphics tasks like modeling and morphing of objects, it has rarely been used for registration and matching of 3D point clouds. Unlike in graphics, where the goal is precise reconstruction, we use isosurfaces to derive a smooth and approximate representation of the underlying point cloud which helps in generalization. Implicit surfaces are generated using a variational interpolation technique. Implicit function values on a set of concentric spheres around the 3D point cloud of object are used as features for matching. Geometric-invariance is achieved by decomposing implicit values based feature set into various spherical harmonics. The decomposition provides a compact representation of 3D point clouds while achieving rotation invariance}, keywords = {3D, cloud;computer, function, geometric, graphics;geophysical, graphics;image, Interpolation, matching;image, point, processing;image, reconstruction;image, registration;image, registration;implicit, representation;interpolation;, representation;variational, signal, technique;clouds;computer, value;invariant}, doi = {10.1109/ICIP.2006.312542}, author = {Biswas,S. and Aggarwal,G. and Chellapa, Rama} } @inbook {17620, title = {Lower Bounds on the Deterministic and Quantum Communication Complexities of Hamming-Distance Problems}, booktitle = {Algorithms and ComputationAlgorithms and Computation}, series = {Lecture Notes in Computer Science}, volume = {4288}, year = {2006}, month = {2006///}, pages = {628 - 637}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {A sub-area of discrepancy theory that has received much attention in computer science re-cently, is that of explicit constructions of low-discrepancy point sets for various types of rectangle families in high dimension. This research has led to interesting applications in error-control cod- ing, distributed protocols, Web document filtering, derandomization, and other areas. We give a short survey of this area here. }, isbn = {978-3-540-49694-6}, url = {http://dx.doi.org/10.1007/11940128_63}, author = {Ambainis,Andris and Gasarch,William and Srinivasan, Aravind and Utis,Andrey}, editor = {Asano,Tetsuo} } @conference {18889, title = {Maintaining cooperation in noisy environments}, volume = {21}, year = {2006}, month = {2006///}, pages = {1561 - 1561}, abstract = {To prevent or alleviate conflicts in multi-agent environ- ments, it is important to distinguish between situations where another agent has misbehaved intentionally and situations where the misbehavior was accidental. One situation where this problem arises is the Noisy Iterated Prisoner{\textquoteright}s Dilemma, a version of the Iterated Prisoner{\textquoteright}s Dilemma (IPD) in which there is a nonzero probability that a {\textquotedblleft}cooperate{\textquotedblright} action will accidentally be changed into a {\textquotedblleft}defect{\textquotedblright} action and vice versa. Tit-For-Tat and other strategies that do quite well in the ordi- nary (non-noisy) IPD can do quite badly in the Noisy IPD.This paper presents a technique called symbolic noise detec- tion, for detecting whether anomalies in player{\textquoteright}s behavior are deliberate or accidental. This idea to use player{\textquoteright}s determin- istic behavior to tell whether an action has been affected by noise. We also present DBS, an algorithm that uses symbolic noise detection in the Noisy IPD. DBS constructs a model of the other agent{\textquoteright}s deterministic behavior, and watches for any deviation from this model. If the other agent{\textquoteright}s next ac- tion is inconsistent with this model, the inconsistency can be due either to noise or to a genuine change in their behavior; and DBS can often distinguish between two cases by waiting to see whether this inconsistency persists in next few moves. This technique is effective because many IPD players often have clear deterministic patterns of behavior. We entered several different implementations of DBS in the 2005 Iterated Prisoner{\textquoteright}s Dilemma competition, in Category 2 (noisy environments). Out of the 165 contestants in this cate- gory, most of DBS implementations ranked among top ten. The best one ranked third, and it was beaten only by two {\textquotedblleft}master-and-slaves strategy{\textquotedblright} programs that each had a large number of {\textquotedblleft}slave{\textquotedblright} programs feeding points to them. }, url = {https://www.aaai.org/Papers/AAAI/2006/AAAI06-250.pdf}, author = {Au,T. C and Nau, Dana S.} } @conference {13706, title = {A maximum entropy approach to combining word alignments}, booktitle = {Proceedings of the main conference on Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics}, series = {HLT-NAACL {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {96 - 103}, publisher = {Association for Computational Linguistics}, organization = {Association for Computational Linguistics}, address = {Stroudsburg, PA, USA}, abstract = {This paper presents a new approach to combining outputs of existing word alignment systems. Each alignment link is represented with a set of feature functions extracted from linguistic features and input alignments. These features are used as the basis of alignment decisions made by a maximum entropy approach. The learning method has been evaluated on three language pairs, yielding significant improvements over input alignments and three heuristic combination methods. The impact of word alignment on MT quality is investigated, using a phrase-based MT system.}, doi = {10.3115/1220835.1220848}, url = {http://dx.doi.org/10.3115/1220835.1220848}, author = {Ayan,Necip Fazil and Dorr, Bonnie J} } @article {12636, title = {Measuring human movement for biomechanical applications using markerless motion capture}, journal = {Proceedings of SPIE}, volume = {6056}, year = {2006}, month = {2006///}, pages = {60560R - 60560R}, abstract = {Modern biomechanical and clinical applications require the accurate capture of normal and pathological humanmovement without the artifacts associated with standard marker-based motion capture techniques such as soft tissue artifacts and the risk of artificial stimulus of taped-on or strapped-on markers. In this study, the need for new markerless human motion capture methods is discussed in view of biomechanical applications. Three different approaches for estimating human movement from multiple image sequences were explored. The first two approaches tracked a 3D articulated model in 3D representations constructed from the image sequences, while the third approach tracked a 3D articulated model in multiple 2D image planes. The three methods are systematically evaluated and results for real data are presented. The role of choosing appropriate technical equipment and algorithms for accurate markerless motion capture is critical. The implementation of this new methodology offers the promise for simple, time-efficient, and potentially more meaningful assessments of human movement in research and clinical practice. }, author = {M{\"u}ndermann, L. and Corazza, S. and Chaudhari, A.M. and Andriacchi, T.P. and Sundaresan, A. and Chellapa, Rama} } @article {16006, title = {The metacognitive loop I: Enhancing reinforcement learning with metacognitive monitoring and control for improved perturbation tolerance}, journal = {Journal of Experimental \& Theoretical Artificial Intelligence}, volume = {18}, year = {2006}, month = {2006///}, pages = {387 - 411}, author = {Anderson,M. L and Oates,T. and Chong,W. and Perlis, Don} } @conference {17854, title = {Model-based OpenMP implementation of a 3D facial pose tracking system}, booktitle = {2006 International Conference on Parallel Processing Workshops, 2006. ICPP 2006 Workshops}, year = {2006}, month = {2006///}, pages = {8 pp.-73 - 8 pp.-73}, publisher = {IEEE}, organization = {IEEE}, abstract = {Most image processing applications are characterized by computation-intensive operations, and high memory and performance requirements. Parallelized implementation on shared-memory systems offer an attractive solution to this class of applications. However, we cannot thoroughly exploit the advantages of such architectures without proper modeling and analysis of the application. In this paper, we describe our implementation of a 3D facial pose tracking system using the OpenMP platform. Our implementation is based on a design methodology that uses coarse-grain dataflow graphs to model and schedule the application. We present our modeling approach, details of the implementation that we derived based on this modeling approach, and associated performance results. The parallelized implementation achieves significant speedup, and meets or exceeds the target frame rate under various configurations}, keywords = {3D facial pose tracking system, application modeling, application program interfaces, application scheduling, coarse-grain dataflow graphs, Concurrent computing, data flow graphs, Educational institutions, face recognition, IMAGE PROCESSING, image processing applications, Inference algorithms, Message passing, OpenMP platform, parallel implementation, PARALLEL PROCESSING, parallel programming, Particle tracking, Processor scheduling, SHAPE, shared memory systems, shared-memory systems, Solid modeling, tracking}, isbn = {0-7695-2637-3}, doi = {10.1109/ICPPW.2006.55}, author = {Saha,S. and Chung-Ching Shen and Chia-Jui Hsu and Aggarwal,G. and Veeraraghavan,A. and Sussman, Alan and Bhattacharyya, Shuvra S.} } @article {14081, title = {Molecular Characterization of Serine-, Alanine-, and Proline-Rich Proteins of Trypanosoma cruzi and Their Possible Role in Host Cell Infection}, journal = {Infect. Immun.}, volume = {74}, year = {2006}, month = {2006/03/01/}, pages = {1537 - 1546}, abstract = {We previously reported the isolation of a novel protein gene family, termed SAP (serine-, alanine-, and proline-rich protein), from Trypanosoma cruzi. Aided by the availability of the completed genome sequence of T. cruzi, we have now identified 39 full-length sequences of SAP, six pseudogenes and four partial genes. SAPs share a central domain of about 55 amino acids and can be divided into four groups based on their amino (N)- and carboxy (C)-terminal sequences. Some SAPs have conserved N- and C-terminal domains encoding a signal peptide and a glycosylphosphatidylinositol anchor addition site, respectively. Analysis of the expression of SAPs in metacyclic trypomastigotes by two-dimensional electrophoresis and immunoblotting revealed that they are likely to be posttranslationally modified in vivo. We have also demonstrated that some SAPs are shed into the extracellular medium. The recombinant SAP exhibited an adhesive capacity toward mammalian cells, where binding was dose dependent and saturable, indicating a possible ligand-receptor interaction. SAP triggered the host cell Ca2+ response required for parasite internalization. A cell invasion assay performed in the presence of SAP showed inhibition of internalization of the metacyclic forms of the CL strain. Taken together, these results show that SAP is involved in the invasion of mammalian cells by metacyclic trypomastigotes, and they confirm the hypothesis that infective trypomastigotes exploit an arsenal of surface glycoproteins and shed proteins to induce signaling events required for their internalization.}, doi = {

10.1128/IAI.74.3.1537-1546.2006

}, url = {http://iai.asm.org/cgi/content/abstract/74/3/1537}, author = {Baida,Renata C. P. and Santos,Marcia R. M. and Carmo,Mirian S. and Yoshida,Nobuko and Ferreira,Danielle and Ferreira,Alice Teixeira and El Sayed,Najib M. and Andersson,Bjorn and da Silveira,Jose Franco} } @conference {17852, title = {Multiple range query optimization with distributed cache indexing}, booktitle = {Proceedings of the 2006 ACM/IEEE conference on Supercomputing}, year = {2006}, month = {2006///}, pages = {100{\textendash}es - 100{\textendash}es}, author = {Nam,B. and Andrade,H. and Sussman, Alan} } @article {13184, title = {Multi-view Reconstruction of Static and Dynamic Scenes}, journal = {Handbook of Mathematical Models in Computer Vision}, year = {2006}, month = {2006///}, pages = {405 - 422}, abstract = {We explore the reconstruction of a three-dimensional scene from multiple images captured from far away viewpoints (wide-baseline camera arrangement). Such an arrangement is required for complex scenes where the visibility flora any one viewpoint is not sufficient to adequately reconstruct the entire scene. Also, such an arrangement reduces the error in triangulation of the features, thereby improving the accuracy of reconstruction. Our emphasis is on algorithms that recover a volumetric model of the scene from calibrated cameras by explicitly modeling and detecting occlusions. We present a brief overview of the state of the art in such methods for multi-view reconstruction. In particular, algorithms based on a probabilistic framework have become quite popular and produce very accurate models. Several such probabilistic volume reconstruction methods will be described. For the dynamic parts of the scene, where an online reconstruction is needed, simpler methods are required. An important case of such scenes is that of walking people in a surveillance scenario. For this case, we present fast online algorithms that recover approximate shape and appearance models and 3D trajectories of people as they walk in the scene. Finally, we address the problem of optimal placement of cameras in order to acquire the best possible image data for the reconstruction of a given scene according to the particular task specification.}, author = {Agrawal,M. and Mittal,A. and Davis, Larry S.} } @article {17307, title = {Network Visualization by Semantic Substrates}, journal = {IEEE Transactions on Visualization and Computer Graphics}, volume = {12}, year = {2006}, month = {2006/10//Sept}, pages = {733 - 740}, abstract = {Networks have remained a challenge for information visualization designers because of the complex issues of node and link layout coupled with the rich set of tasks that users present. This paper offers a strategy based on two principles: (1) layouts are based on user-defined semantic substrates, which are non-overlapping regions in which node placement is based on node attributes, (2) users interactively adjust sliders to control link visibility to limit clutter and thus ensure comprehensibility of source and destination. Scalability is further facilitated by user control of which nodes are visible. We illustrate our semantic substrates approach as implemented in NVSS 1.0 with legal precedent data for up to 1122 court cases in three regions with 7645 legal citations}, keywords = {Automatic control, data visualisation, Data visualization, Displays, Filters, Graphical user interfaces, Information Visualization, information visualization designers, Law, legal citations, Legal factors, legal precedent data, network visualization, NVSS 1.0, scalability, semantic substrate, Terminology, Tunneling, user-defined semantic substrates}, isbn = {1077-2626}, doi = {10.1109/TVCG.2006.166}, author = {Shneiderman, Ben and Aris,A.} } @article {17309, title = {Nuclear Envelope Dystrophies Show a Transcriptional Fingerprint Suggesting Disruption of Rb{\textendash}MyoD Pathways in Muscle Regeneration}, journal = {BrainBrain}, volume = {129}, year = {2006}, month = {2006/04/01/}, pages = {996 - 1013}, abstract = {Mutations of lamin A/C (LMNA) cause a wide range of human disorders, including progeria, lipodystrophy, neuropathies and autosomal dominant Emery{\textendash}Dreifuss muscular dystrophy (EDMD). EDMD is also caused by X-linked recessive loss-of-function mutations of emerin, another component of the inner nuclear lamina that directly interacts with LMNA. One model for disease pathogenesis of LMNA and emerin mutations is cell-specific perturbations of the mRNA transcriptome in terminally differentiated cells. To test this model, we studied 125 human muscle biopsies from 13 diagnostic groups (125 U133A, 125 U133B microarrays), including EDMD patients with LMNA and emerin mutations. A Visual and Statistical Data Analyzer (VISDA) algorithm was used to statistically model cluster hierarchy, resulting in a tree of phenotypic classifications. Validations of the diagnostic tree included permutations of U133A and U133B arrays, and use of two probe set algorithms (MAS5.0 and MBEI). This showed that the two nuclear envelope defects (EDMD LMNA, EDMD emerin) were highly related disorders and were also related to fascioscapulohumeral muscular dystrophy (FSHD). FSHD has recently been hypothesized to involve abnormal interactions of chromatin with the nuclear envelope. To identify disease-specific transcripts for EDMD, we applied a leave-one-out (LOO) cross-validation approach using LMNA patient muscle as a test data set, with reverse transcription{\textendash}polymerase chain reaction (RT{\textendash}PCR) validations in both LMNA and emerin patient muscle. A high proportion of top-ranked and validated transcripts were components of the same transcriptional regulatory pathway involving Rb1 and MyoD during muscle regeneration (CRI-1, CREBBP, Nap1L1, ECREBBP/p300), where each was specifically upregulated in EDMD. Using a muscle regeneration time series (27 time points) we develop a transcriptional model for downstream consequences of LMNA and emerin mutations. We propose that key interactions between the nuclear envelope and Rb and MyoD fail in EDMD at the point of myoblast exit from the cell cycle, leading to poorly coordinated phosphorylation and acetylation steps. Our data is consistent with mutations of nuclear lamina components leading to destabilization of the transcriptome in differentiated cells.}, keywords = {EDMD = Emery{\textendash}Dreifuss muscular dystrophy, emerin, Emery-Dreifuss muscular dystrophy, FSHD = fascioscapulohumeral muscular dystrophy, IDG = individual discriminatory genes, JDG = jointly discriminatory genes, lamin A/C, LGMD = limb-girdle muscular dystrophy, LOO = leave-one-out, RT{\textendash}PCR = reverse transcription{\textendash}polymerase chain reaction; VISDA = Visual and Statistical Data Analyzer, Skeletal muscle, wFC = weighted Fisher criterion}, isbn = {0006-8950, 1460-2156}, doi = {10.1093/brain/awl023}, url = {http://brain.oxfordjournals.org/content/129/4/996}, author = {Bakay,Marina and Wang,Zuyi and Melcon,Gisela and Schiltz,Louis and Xuan,Jianhua and Zhao,Po and Sartorelli,Vittorio and Seo,Jinwook and Pegoraro,Elena and Angelini,Corrado and Shneiderman, Ben and Escolar,Diana and Chen,Yi-Wen and Winokur,Sara T and Pachman,Lauren M and Fan,Chenguang and Mandler,Raul and Nevo,Yoram and Gordon,Erynn and Zhu,Yitan and Dong,Yibin and Wang,Yue and Hoffman,Eric P} } @article {13457, title = {Open problems in relational data clustering}, journal = {Proceedings of the ICML Workshop on Open Problems in Stastistical Relational Learning}, year = {2006}, month = {2006///}, abstract = {Data clustering is the task of detecting pat-terns in a set of data. Most algorithms take non-relational data as input and are sometimes unable to find significant patterns. Many data sets can include relational infor- mation, as well as independent object at- tributes. We believe that clustering with re- lational data will help find significant pat- terns where non-relational algorithms fail. This paper discusses two open problems in relational data clustering: clustering hetero- geneous data, and relation selection or ex- traction. Potential methods for addressing the problems are presented. }, author = {Anthony,A. and desJardins, Marie} } @article {19018, title = {OverDoSe: A generic DDoS protection service using an overlay network}, year = {2006}, month = {2006}, institution = {School of Computer Science, Carnegie Mellon University}, abstract = {We present the design and implementation of OverDoSe, an overlay network offering generic DDoS protection for targeted sites. OverDoSe clients and servers are isolated at the IP level. Overlay nodes route packets between a client and a server, and regulate traffic according to the server{\textquoteright}s instructions. Through the use of light-weight security primitives, OverDoSe achieves resilience against compromised overlay nodes with a minimal performance overhead. OverDoSe can be deployed by a single ISP who wishes to offer DDoS protection as a value-adding service to its customers.}, isbn = {CMU-CS-06-114}, doi = {Technical Report}, url = {http://repository.cmu.edu/cgi/viewcontent.cgi?article=1073\&context=compsci}, author = {Elaine Shi and Stoica,I. and Andersen,D.G. and Perrig, A.} } @conference {12083, title = {Partially overlapped channels not considered harmful}, booktitle = {Proceedings of the joint international conference on Measurement and modeling of computer systems}, year = {2006}, month = {2006///}, pages = {63 - 74}, author = {Mishra,A. and Shrivastava,V. and Banerjee,S. and Arbaugh, William A.} } @article {17783, title = {The priority curve algorithm for video summarization}, journal = {Information Systems}, volume = {31}, year = {2006}, month = {2006/11//}, pages = {679 - 695}, abstract = {In this paper, we introduce the concept of a priority curve associated with a video. We then provide an algorithm that can use the priority curve to create a summary (of a desired length) of any video. The summary thus created exhibits nice continuity properties and also avoids repetition. We have implemented the priority curve algorithm (PriCA) and compared it with other summarization algorithms in the literature with respect to both performance and the output quality. The quality of summaries was evaluated by a group of 200 students in Naples, Italy, who watched soccer videos. We show that PriCA is faster than existing algorithms and also produces better quality summaries. We also briefly describe a soccer video summarization system we have built on using the PriCA architecture and various (classical) image processing algorithms.}, keywords = {Content based retrieval, Video databases, Video summarization}, isbn = {0306-4379}, doi = {10.1016/j.is.2005.12.003}, url = {http://www.sciencedirect.com/science/article/pii/S0306437905001250}, author = {Albanese, M. and Fayzullin,M. and Picariello, A. and V.S. Subrahmanian} } @conference {11955, title = {A probabilistic notion of correspondence and the epipolar constraint}, booktitle = {Proceedings of the Third International Symposium on 3D Data Processing, Visualization, and Transmission (3DPVT{\textquoteright}06)}, year = {2006}, month = {2006///}, pages = {41 - 48}, author = {Domke, J. and Aloimonos, J.} } @article {17645, title = {Provable algorithms for parallel generalized sweep scheduling}, journal = {Journal of Parallel and Distributed Computing}, volume = {66}, year = {2006}, month = {2006///}, pages = {807 - 821}, author = {Anil Kumar,V. S and Marathe,M. V and Parthasarathy,S. and Srinivasan, Aravind and Zust,S.} } @book {13182, title = {Real-Time Distributed Algorithms for Visual and Battlefield Reasoning}, year = {2006}, month = {2006///}, publisher = {MARYLAND UNIV COLLEGE PARK OFFICE OF RESEARCH ADMINISTRATION AND ADVANCEMENT}, organization = {MARYLAND UNIV COLLEGE PARK OFFICE OF RESEARCH ADMINISTRATION AND ADVANCEMENT}, abstract = {Information is key to the success of the next generation battlefield. There is a critical need to determine, in real-time, what the enemy is doing, and to interpret that information in the context of past related events. In this project we examined two aspects of this issue: development of a high-level task definition language for tasking a network of sensors to carry out given objectives, and interpreting recounted events so that past related scenarios could be automatically identified from a case database.}, author = {Davis, Larry S. and Basili, Victor R. and V.S. Subrahmanian and Reggia, James A. and Aloimonos, J.} } @conference {13169, title = {Real-Time Human Detection, Tracking, and Verification in Uncontrolled Camera Motion Environments}, booktitle = {Computer Vision Systems, 2006 ICVS {\textquoteright}06. IEEE International Conference on}, year = {2006}, month = {2006/01//}, pages = {41 - 41}, abstract = {In environments where a camera is installed on a freely moving platform, e.g. a vehicle or a robot, object detection and tracking becomes much more difficult. In this paper, we presents a real time system for human detection, tracking, and verification in such challenging environments. To deliver a robust performance, the system integrates several computer vision algorithms to perform its function: a human detection algorithm, an object tracking algorithm, and a motion analysis algorithm. To utilize the available computing resources to the maximum possible extent, each of the system components is designed to work in a separate thread that communicates with the other threads through shared data structures. The focus of this paper is more on the implementation issues than on the algorithmic issues of the system. Object oriented design was adopted to abstract algorithmic details away from the system structure.}, doi = {10.1109/ICVS.2006.52}, author = {Hussein,M. and Abd-Almageed, Wael and Yang Ran and Davis, Larry S.} } @article {12618, title = {Robust ego-motion estimation and 3-D model refinement using surface parallax}, journal = {Image Processing, IEEE Transactions on}, volume = {15}, year = {2006}, month = {2006/05//}, pages = {1215 - 1225}, abstract = {We present an iterative algorithm for robustly estimating the ego-motion and refining and updating a coarse depth map using parametric surface parallax models and brightness derivatives extracted from an image pair. Given a coarse depth map acquired by a range-finder or extracted from a digital elevation map (DEM), ego-motion is estimated by combining a global ego-motion constraint and a local brightness constancy constraint. Using the estimated camera motion and the available depth estimate, motion of the three-dimensional (3-D) points is compensated. We utilize the fact that the resulting surface parallax field is an epipolar field, and knowing its direction from the previous motion estimates, estimate its magnitude and use it to refine the depth map estimate. The parallax magnitude is estimated using a constant parallax model (CPM) which assumes a smooth parallax field and a depth based parallax model (DBPM), which models the parallax magnitude using the given depth map. We obtain confidence measures for determining the accuracy of the estimated depth values which are used to remove regions with potentially incorrect depth estimates for robustly estimating ego-motion in subsequent iterations. Experimental results using both synthetic and real data (both indoor and outdoor sequences) illustrate the effectiveness of the proposed algorithm.}, keywords = {3D model refinement, algorithms, Artificial intelligence, Automated;Subtraction Technique;, Computer-Assisted;Imaging, constant parallax model, depth based parallax model, digital elevation map, epipolar field, Image Enhancement, Image Interpretation, iterative algorithm, iterative methods, Motion estimation, robust ego-motion estimation, smooth parallax field, surface parallax, Three-Dimensional;Information Storage and Retrieval;Pattern Recognition}, isbn = {1057-7149}, doi = {10.1109/TIP.2005.864167}, author = {Agrawal,A. and Chellapa, Rama} } @article {12926, title = {Seasonal Cholera Caused by Vibrio Cholerae Serogroups O1 and O139 in the Coastal Aquatic Environment of Bangladesh}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {72}, year = {2006}, month = {2006/06/01/}, pages = {4096 - 4104}, abstract = {Since Vibrio cholerae O139 first appeared in 1992, both O1 El Tor and O139 have been recognized as the epidemic serogroups, although their geographic distribution, endemicity, and reservoir are not fully understood. To address this lack of information, a study of the epidemiology and ecology of V. cholerae O1 and O139 was carried out in two coastal areas, Bakerganj and Mathbaria, Bangladesh, where cholera occurs seasonally. The results of a biweekly clinical study (January 2004 to May 2005), employing culture methods, and of an ecological study (monthly in Bakerganj and biweekly in Mathbaria from March 2004 to May 2005), employing direct and enrichment culture, colony blot hybridization, and direct fluorescent-antibody methods, showed that cholera is endemic in both Bakerganj and Mathbaria and that V. cholerae O1, O139, and non-O1/non-O139 are autochthonous to the aquatic environment. Although V. cholerae O1 and O139 were isolated from both areas, most noteworthy was the isolation of V. cholerae O139 in March, July, and September 2004 in Mathbaria, where seasonal cholera was clinically linked only to V. cholerae O1. In Mathbaria, V. cholerae O139 emerged as the sole cause of a significant outbreak of cholera in March 2005. V. cholerae O1 reemerged clinically in April 2005 and established dominance over V. cholerae O139, continuing to cause cholera in Mathbaria. In conclusion, the epidemic potential and coastal aquatic reservoir for V. cholerae O139 have been demonstrated. Based on the results of this study, the coastal ecosystem of the Bay of Bengal is concluded to be a significant reservoir for the epidemic serogroups of V. cholerae.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.00066-06}, url = {http://aem.asm.org/content/72/6/4096}, author = {Alam,Munirul and Hasan,Nur A. and Sadique,Abdus and Bhuiyan,N. A. and Ahmed,Kabir U. and Nusrin,Suraia and Nair,G. Balakrish and Siddique,A. K. and Sack,R. Bradley and Sack,David A. and Huq,Anwar and Rita R Colwell} } @conference {11979, title = {A sensory grammar for inferring behaviors in sensor networks}, booktitle = {Proceedings of the 5th international conference on Information processing in sensor networks}, series = {IPSN {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {251 - 259}, publisher = {ACM}, organization = {ACM}, address = {Nashville, Tennessee, USA}, abstract = {The ability of a sensor network to parse out observable activities into a set of distinguishable actions is a powerful feature that can potentially enable many applications of sensor networks to everyday life situations. In this paper we introduce a framework that uses a hierarchy of Probabilistic Context Free Grammars (PCFGs) to perform such parsing. The power of the framework comes from the hierarchical organization of grammars that allows the use of simple local sensor measurements for reasoning about more macroscopic behaviors. Our presentation describes how to use a set of phonemes to construct grammars and how to achieve distributed operation using a messaging model. The proposed framework is flexible. It can be mapped to a network hierarchy or can be applied sequentially and across the network to infer behaviors as they unfold in space and time. We demonstrate this functionality by inferring simple motion patterns using a sequence of simple direction vectors obtained from our camera sensor network testbed.}, keywords = {algorithms, behavior identification, human activity, human factors, pcfg, real-time and embedded systems, sensor grammars, sensor networks}, isbn = {1-59593-334-4}, doi = {10.1145/1127777.1127817}, author = {Lymberopoulos, Dimitrios and Ogale, Abhijit S. and Savvides, Andreas and Aloimonos, J.} } @conference {12026, title = {A Sensory-Motor Language for Human Activity Understanding}, booktitle = {2006 6th IEEE-RAS International Conference on Humanoid Robots}, year = {2006}, month = {2006/12/04/6}, pages = {69 - 75}, publisher = {IEEE}, organization = {IEEE}, abstract = {We have empirically discovered that the space of human actions has a linguistic framework. This is a sensory-motor space consisting of the evolution of the joint angles of the human body in movement. The space of human activity has its own phonemes, morphemes, and sentences. We present a human activity language (HAL) for symbolic non-arbitrary representation of visual and motor information. In phonology, we define atomic segments (kinetemes) that are used to compose human activity. We introduce the concept of a kinetological system and propose five basic properties for such a system: compactness, view-invariance, reproducibility, selectivity, and reconstructivity. In morphology, we extend sequential language learning to incorporate associative learning with our parallel learning approach. Parallel learning is effective in identifying the kinetemes and active joints in a particular action. In syntax, we suggest four lexical categories for our human activity language (noun, verb, adjective, and adverb). These categories are combined into sentences through syntax for human movement}, keywords = {Actuators, associative learning, atomic segments, computational linguistics, Computer science, Computer vision, Educational institutions, grammars, human activity language, human activity understanding, human movement syntax, Humanoid robots, HUMANS, joint angles, kinetemes, kinetological system, Laboratories, learning (artificial intelligence), List key index terms here, Morphology, motor information, No mare than 5, parallel learning, Reproducibility of results, Robot kinematics, Robot programming, robot vision, sensory-motor language, sequential language learning, symbolic nonarbitrary representation, visual information}, isbn = {1-4244-0200-X}, doi = {10.1109/ICHR.2006.321365}, author = {Guerra-Filho,G. and Aloimonos, J.} } @article {17658, title = {Session 8A-Combinatorial Optimization and Quantum Computing-Lower Bounds on the Deterministic and Quantum Communication Complexities of Hamming-Distance Problems}, journal = {Lecture Notes in Computer Science}, volume = {4288}, year = {2006}, month = {2006///}, pages = {628 - 637}, author = {Ambainis,A. and Gasarch,W. and Srinivasan, Aravind and Utis,A.} } @conference {17538, title = {A structural approach to latency prediction}, booktitle = {Proceedings of the 6th ACM SIGCOMM conference on Internet measurement}, series = {IMC {\textquoteright}06}, year = {2006}, month = {2006///}, pages = {99 - 104}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Several models have been recently proposed for predicting the latency of end to end Internet paths. These models treat the Internet as a black-box, ignoring its internal structure. While these models are simple, they can often fail systematically; for example, the most widely used models use metric embeddings that predict no benefit to detour routes even though half of all Internet routes can benefit from detours.In this paper, we adopt a structural approach that predicts path latency based on measurements of the Internet{\textquoteright}s routing topology, PoP connectivity, and routing policy. We find that our approach outperforms Vivaldi, the most widely used black-box model. Furthermore, unlike metric embeddings, our approach successfully predicts 65\% of detour routes in the Internet. The number of measurements used in our approach is comparable with that required by black box techniques, but using traceroutes instead of pings.}, keywords = {internet topology, latency prediction, route measurements}, isbn = {1-59593-561-4}, doi = {10.1145/1177080.1177092}, url = {http://doi.acm.org/10.1145/1177080.1177092}, author = {Madhyastha,Harsha V. and Anderson,Thomas and Krishnamurthy,Arvind and Spring, Neil and Venkataramani,Arun} } @article {17663, title = {System and method for locating a closest server in response to a client ...}, volume = {09/726,192}, year = {2006}, month = {2006/03/28/}, abstract = {A scalable system and method for locating a closest server in response to a client request via an interactive distribution network, such as the Internet, are provided. A closest content server is defined as having the least round trip time for responding to a client request. The system including a plurality of content servers; and a local server in communication with a plurality of clients, the local server acting as a proxy for communicating client requests from clients to a redirection server. Client network distance and load information is periodically collected at each content server in the network from clients communicating with each of the respective content servers. The redirection server periodically aggregates the network distance and load information from each content server to create client clusters from both current and previously aggregated network distance and load information. Each client cluster represents a division or partition of the total IP address space. Each...}, url = {http://www.google.com/patents?id=cE54AAAAEBAJ}, author = {Andrews,Matthew and Hofmann,Markus and Shepherd,Bruce and Srinivasan, Aravind and Winkler,Peter and Zane,Francis}, editor = {Lucent Technologies Inc.} } @conference {11996, title = {Towards a sensorimotor WordNet SM: Closing the semantic gap}, booktitle = {Proc. of the International WordNet Conference (GWC)}, year = {2006}, month = {2006///}, author = {Guerra-Filho,G. and Aloimonos, J.} } @article {12927, title = {Toxigenic Vibrio Cholerae in the Aquatic Environment of Mathbaria, Bangladesh}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {72}, year = {2006}, month = {2006/04/01/}, pages = {2849 - 2855}, abstract = {Toxigenic Vibrio cholerae, rarely isolated from the aquatic environment between cholera epidemics, can be detected in what is now understood to be a dormant stage, i.e., viable but nonculturable when standard bacteriological methods are used. In the research reported here, biofilms have proved to be a source of culturable V. cholerae, even in nonepidemic periods. Biweekly environmental surveillance for V. cholerae was carried out in Mathbaria, an area of cholera endemicity adjacent to the Bay of Bengal, with the focus on V. cholerae O1 and O139 Bengal. A total of 297 samples of water, phytoplankton, and zooplankton were collected between March and December 2004, yielding eight V. cholerae O1 and four O139 Bengal isolates. A combination of culture methods, multiplex-PCR, and direct fluorescent antibody (DFA) counting revealed the Mathbaria aquatic environment to be a reservoir for V. cholerae O1 and O139 Bengal. DFA results showed significant clumping of the bacteria during the interepidemic period for cholera, and the fluorescent micrographs revealed large numbers of V. cholerae O1 in thin films of exopolysaccharides (biofilm). A similar clumping of V. cholerae O1 was also observed in samples collected from Matlab, Bangladesh, where cholera also is endemic. Thus, the results of the study provided in situ evidence for V. cholerae O1 and O139 in the aquatic environment, predominantly as viable but nonculturable cells and culturable cells in biofilm consortia. The biofilm community is concluded to be an additional reservoir of cholera bacteria in the aquatic environment between seasonal epidemics of cholera in Bangladesh.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.72.4.2849-2855.2006}, url = {http://aem.asm.org/content/72/4/2849}, author = {Alam,Munirul and Sultana,Marzia and Nair,G. Balakrish and Sack,R. Bradley and Sack,David A. and Siddique,A. K. and Ali,Afsar and Huq,Anwar and Rita R Colwell} } @conference {13161, title = {Tracking Articulating Objects from Ground Vehicles using Mixtures of Mixtures}, booktitle = {Intelligent Robots and Systems, 2006 IEEE/RSJ International Conference on}, year = {2006}, month = {2006/10//}, pages = {1230 - 1236}, abstract = {An algorithm for tracking articulating objects from moving camera platforms is presented. Mixtures of mixtures are used to model the appearance of the object and the background. The state of the object is tracked using a particle filter. Egomotion information are estimated and used to set the state variance of the particle filter. Results of tracking human objects from an unmanned ground vehicle are used to evaluate the tracking algorithm}, keywords = {articulating, filter;unmanned, ground, information;mixtures, mixtures;particle, objects, of, operated, tracking;egomotion, vehicle;remotely, vehicles;robot, vision;}, doi = {10.1109/IROS.2006.281881}, author = {Abd-Almageed, Wael and Hussein,M. and Davis, Larry S.} } @article {11981, title = {Understanding visuo-motor primitives for motion synthesis and analysis}, journal = {Computer Animation and Virtual Worlds}, volume = {17}, year = {2006}, month = {2006/07/01/}, pages = {207 - 217}, abstract = {The problem addressed in this paper concerns the representation of human movement in terms of atomic visuo-motor primitives considering both generation and perception of movement. We introduce the concept of kinetology, the phonology of human movement, and five principles on which such a system should be based: compactness, view-invariance, reproducibility, selectivity, and reconstructivity. We propose visuo-motor primitives and demonstrate their kinetological properties. Further evaluation is accomplished with experiments on compression and decompression. Our long-term goal is to demonstrate that action has a space characterized by a visuo-motor language. Copyright {\textcopyright} 2006 John Wiley \& Sons, Ltd.}, keywords = {compression, motion synthesis, visuo-motor primitives}, isbn = {1546-427X}, doi = {10.1002/cav.124}, url = {http://onlinelibrary.wiley.com/doi/10.1002/cav.124/abstract}, author = {Guerra-Filho,Gutemberg and Aloimonos, J.} } @inbook {12642, title = {What Is the Range of Surface Reconstructions from a Gradient Field?}, booktitle = {Computer Vision {\textendash} ECCV 2006Computer Vision {\textendash} ECCV 2006}, series = {Lecture Notes in Computer Science}, volume = {3951}, year = {2006}, month = {2006///}, pages = {578 - 591}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We propose a generalized equation to represent a continuum of surface reconstruction solutions of a given non-integrable gradient field. We show that common approaches such as Poisson solver and Frankot-Chellappa algorithm are special cases of this generalized equation. For a N {\texttimes} N pixel grid, the subspace of all integrable gradient fields is of dimension N 2 {\textendash} 1. Our framework can be applied to derive a range of meaningful surface reconstructions from this high dimensional space. The key observation is that the range of solutions is related to the degree of anisotropy in applying weights to the gradients in the integration process. While common approaches use isotropic weights, we show that by using a progression of spatially varying anisotropic weights, we can achieve significant improvement in reconstructions. We propose (a) α-surfaces using binary weights, where the parameter α allows trade off between smoothness and robustness, (b) M-estimators and edge preserving regularization using continuous weights and (c) Diffusion using affine transformation of gradients. We provide results on photometric stereo, compare with previous approaches and show that anisotropic treatment discounts noise while recovering salient features in reconstructions.}, isbn = {978-3-540-33832-1}, url = {http://dx.doi.org/10.1007/11744023_45}, author = {Agrawal,Amit and Raskar,Ramesh and Chellapa, Rama}, editor = {Leonardis,Ale{\v s} and Bischof,Horst and Pinz,Axel} } @article {12094, title = {Wireless Network Security and Interworking}, journal = {Proceedings of the IEEE}, volume = {94}, year = {2006}, month = {2006/02//}, pages = {455 - 466}, abstract = {A variety of wireless technologies have been standardized and commercialized, but no single technology is considered the best because of different coverage and bandwidth limitations. Thus, interworking between heterogeneous wireless networks is extremely important for ubiquitous and high-performance wireless communications. Security in interworking is a major challenge due to the vastly different security architectures used within each network. The goal of this paper is twofold. First, we provide a comprehensive discussion of security problems and current technologies in 3G and WLAN systems. Second, we provide introductory discussions about the security problems in interworking, the state-of-the-art solutions, and open problems.}, keywords = {3G mobile communication, 3G systems, Authentication, Bandwidth, Communication system security, computer network security, computer security, Data security, internetworking, Land mobile radio cellular systems, Paper technology, security architectures, security of data, telecommunication security, wireless communication, wireless communications, Wireless LAN, wireless network security, Wireless networks, wireless technologies, WLAN systems}, isbn = {0018-9219}, doi = {10.1109/JPROC.2005.862322}, author = {Shin,M. and Ma,J. and Mishra,A. and Arbaugh, William A.} } @inbook {12679, title = {3D Facial Pose Tracking in Uncalibrated Videos}, booktitle = {Pattern Recognition and Machine IntelligencePattern Recognition and Machine Intelligence}, series = {Lecture Notes in Computer Science}, volume = {3776}, year = {2005}, month = {2005///}, pages = {515 - 520}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {This paper presents a method to recover the 3D configuration of a face in each frame of a video. The 3D configuration consists of the 3 translational parameters and the 3 orientation parameters which correspond to the yaw, pitch and roll of the face, which is important for applications like face modeling, recognition, expression analysis, etc. The approach combines the structural advantages of geometric modeling with the statistical advantages of a particle-filter based inference. The face is modeled as the curved surface of a cylinder which is free to translate and rotate arbitrarily. The geometric modeling takes care of pose and self-occlusion while the statistical modeling handles moderate occlusion and illumination variations. Experimental results on multiple datasets are provided to show the efficacy of the approach. The insensitivity of our approach to calibration parameters (focal length) is also shown.}, isbn = {978-3-540-30506-4}, url = {http://dx.doi.org/10.1007/11590316_81}, author = {Aggarwal,Gaurav and Veeraraghavan,Ashok and Chellapa, Rama}, editor = {Pal,Sankar and Bandyopadhyay,Sanghamitra and Biswas,Sambhunath} } @conference {12650, title = {An algebraic approach to surface reconstruction from gradient fields}, booktitle = {Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on}, volume = {1}, year = {2005}, month = {2005/10//}, pages = {174 - 181 Vol. 1 - 174 - 181 Vol. 1}, abstract = {Several important problems in computer vision such as shape from shading (SFS) and photometric stereo (PS) require reconstructing a surface from an estimated gradient field, which is usually non-integrable, i.e. have non-zero curl. We propose a purely algebraic approach to enforce integrability in discrete domain. We first show that enforcing integrability can be formulated as solving a single linear system Ax =b over the image. In general, this system is under-determined. We show conditions under which the system can be solved and a method to get to those conditions based on graph theory. The proposed approach is non-iterative, has the important property of local error confinement and can be applied to several problems. Results on SFS and PS demonstrate the applicability of our method.}, keywords = {algebra;, algebraic, approach;, Computer, confinement;, discrete, domain, error, field;, from, gradient, graph, image, integrability;, linear, local, methods;, photometric, reconstruction;, shading;, SHAPE, stereo;, surface, system;, theory;, vision;}, doi = {10.1109/ICCV.2005.31}, author = {Agrawal,A. and Chellapa, Rama and Raskar, R.} } @conference {13722, title = {Alignment link projection using transformation-based learning}, booktitle = {Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing}, series = {HLT {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {185 - 192}, publisher = {Association for Computational Linguistics}, organization = {Association for Computational Linguistics}, address = {Stroudsburg, PA, USA}, abstract = {We present a new word-alignment approach that learns errors made by existing word alignment systems and corrects them. By adapting transformation-based learning to the problem of word alignment, we project new alignment links from already existing links, using features such as POS tags. We show that our alignment link projection approach yields a significantly lower alignment error rate than that of the best performing alignment system (22.6\% relative reduction on English-Spanish data and 23.2\% relative reduction on English-Chinese data).}, doi = {10.3115/1220575.1220599}, url = {http://dx.doi.org/10.3115/1220575.1220599}, author = {Ayan,Necip Fazil and Dorr, Bonnie J and Monz,Christof} } @article {18893, title = {Applications of SHOP and SHOP2}, journal = {Intelligent Systems, IEEE}, volume = {20}, year = {2005}, month = {2005/04//march}, pages = {34 - 41}, abstract = {We design the simple hierarchical ordered planner (SHOP) and its successor, SHOP2, with two goals in mind: to investigate research issues in automated planning and to provide some simple, practical planning tools. SHOP and SHOP2 are based on a planning formalism called hierarchical task network planning. SHOP and SHOP2 use a search-control strategy called ordered task decomposition, which breaks tasks into subtasks and generates the plan{\textquoteright}s actions in the same order that the plan executor executes them. So, throughout the planning process, the planner can tell what the state of the world at each step of the plan.}, keywords = {automated planning, hierarchical task network planning, ordered task decomposition, planning (artificial intelligence), problem solving, search-control strategy, simple hierarchical ordered planner, trees (mathematics), uncertainty handling}, isbn = {1541-1672}, doi = {10.1109/MIS.2005.20}, author = {Nau, Dana S. and Au,T.-C. and Ilghami,O. and Kuter,U. and Wu,D. and Yaman,F. and Munoz-Avila,H. and Murdock,J. W.} } @conference {12673, title = {An architectural level design methodology for embedded face detection}, booktitle = {Proceedings of the 3rd IEEE/ACM/IFIP international conference on Hardware/software codesign and system synthesis}, series = {CODES+ISSS {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {136 - 141}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Face detection and recognition research has attracted great attention in recent years. Automatic face detection has great potential in a large array of application areas, including banking and security system access control, video surveillance, and multimedia information retrieval. In this paper, we discuss an architectural level design methodology for implementation of an embedded face detection system on a reconfigurable system on chip. We present models for performance estimation and validate these models with experimental values obtained from implementing our system on an FPGA platform. This modeling approach is shown to be efficient, accurate, and intuitive for designers to work with. Using this approach, we present several design options that trade-off various architectural features.}, keywords = {design space exploration, Face detection, platforms, reconfigurable, system-level models}, isbn = {1-59593-161-9}, doi = {10.1145/1084834.1084872}, url = {http://doi.acm.org/10.1145/1084834.1084872}, author = {Kianzad,V. and Saha,S. and Schlessman,J. and Aggarwal,G. and Bhattacharyya, Shuvra S. and Wolf,W. and Chellapa, Rama} } @article {14206, title = {Chromatic Induction and Perspective Distortion}, journal = {Journal of VisionJ Vis}, volume = {5}, year = {2005}, month = {2005/09/23/}, pages = {1026 - 1026}, abstract = {A pattern presented by Robertson (1996), consisting of yellow and blue square waves with red squares superimposed on either the blue or yellow bands give rise to assimilation effects similar as in the White illusion. The red squares on the yellow bands appear more blue, and the red squares on the blue bands appear more yellow. Varying the position and orientation of the pattern in space changes the effect. For example, slanting the pattern or increasing the distance to the pattern in space increases the effect. But rotating the pattern by 90 degrees and slanting it, nearly eliminates the effect. These changes in color appearance can be explained as the result of averaging in receptive field of extended size.Due to perspective distortion, the area of the pattern in space mapping to an image patch changes with the position and orientation of the pattern. A computational simulation, implementing the distortion from 3D space to image space, followed by a color segmentation (k-means clustering) gave the same color distortion and segmentation as perceived. Thus the effect seems to be simple an artifact of sampling which causes averaging or smoothing with some filters. We propose the hypothesis that chromatic induction, which takes the two forms of chromatic contrast and assimilation, results, because the system samples with large fields and following in the segmentation stage attempts to compensate for the averaging effects in neighboring regions }, isbn = {, 1534-7362}, doi = {10.1167/5.8.1026}, url = {http://www.journalofvision.org/content/5/8/1026}, author = {Aloimonos, J. and Ferm{\"u}ller, Cornelia} } @article {14078, title = {Comparative Genomics of Trypanosomatid Parasitic Protozoa}, journal = {Science}, volume = {309}, year = {2005}, month = {2005/07/15/}, pages = {404 - 409}, abstract = {A comparison of gene content and genome architecture of Trypanosoma brucei, Trypanosoma cruzi, and Leishmania major, three related pathogens with different life cycles and disease pathology, revealed a conserved core proteome of about 6200 genes in large syntenic polycistronic gene clusters. Many species-specific genes, especially large surface antigen families, occur at nonsyntenic chromosome-internal and subtelomeric regions. Retroelements, structural RNAs, and gene family expansion are often associated with syntenic discontinuities that{\textemdash}along with gene divergence, acquisition and loss, and rearrangement within the syntenic regions{\textemdash}have shaped the genomes of each parasite. Contrary to recent reports, our analyses reveal no evidence that these species are descended from an ancestor that contained a photosynthetic endosymbiont.}, doi = {10.1126/science.1112181}, url = {http://www.sciencemag.org/content/309/5733/404.abstract}, author = {El-Sayed, Najib M. and Myler,Peter J. and Blandin,Ga{\"e}lle and Berriman,Matthew and Crabtree,Jonathan and Aggarwal,Gautam and Caler,Elisabet and Renauld,Hubert and Worthey,Elizabeth A. and Hertz-Fowler,Christiane and Ghedin,Elodie and Peacock,Christopher and Bartholomeu,Daniella C. and Haas,Brian J. and Tran,Anh-Nhi and Wortman,Jennifer R. and Alsmark,U. Cecilia M. and Angiuoli,Samuel and Anupama,Atashi and Badger,Jonathan and Bringaud,Frederic and Cadag,Eithon and Carlton,Jane M. and Cerqueira,Gustavo C. and Creasy,Todd and Delcher,Arthur L. and Djikeng,Appolinaire and Embley,T. Martin and Hauser,Christopher and Ivens,Alasdair C. and Kummerfeld,Sarah K. and Pereira-Leal,Jose B. and Nilsson,Daniel and Peterson,Jeremy and Salzberg,Steven L. and Shallom,Joshua and Silva,Joana C. and Sundaram,Jaideep and Westenberger,Scott and White,Owen and Melville,Sara E. and Donelson,John E. and Andersson,Bj{\"o}rn and Stuart,Kenneth D. and Hall,Neil} } @conference {17887, title = {Comparing the Performance of High-Level Middleware Systems in Shared and Distributed Memory Parallel Environments}, booktitle = {Parallel and Distributed Processing Symposium, 2005. Proceedings. 19th IEEE International}, year = {2005}, month = {2005/04//}, pages = {30 - 30}, publisher = {IEEE}, organization = {IEEE}, abstract = {The utilization of toolkits for writing parallel and/or distributed applications has been shown to greatly enhance developer{\textquoteright}s productivity. Such an approach hides many of the complexities associated with writing these applications, rather than relying solely on programming language aids and parallel library support, such as MPI or PVM. In this work, we evaluate three different middleware systems that have been used to implement a computation and I/O-intensive data analysis application from the domain of computer vision. This study shows the benefits and overheads associated with each of the middleware systems, in different homogeneous computational environments and with different workloads. Our results lead the way toward being able to make better decisions for tuning the application environment, for selecting the appropriate middleware, and also for designing more powerful middleware systems to efficiently build and run highly complex applications in both parallel and distributed computing environments.}, keywords = {Application software, Computer science, Computer vision, Data analysis, Distributed computing, distributed computing environment, distributed memory parallel environment, distributed shared memory systems, Educational institutions, high-level middleware system, I/O-intensive data analysis application, Libraries, Middleware, parallel computing environment, parallel library support, parallel memories, programming language, programming languages, Runtime environment, shared memory parallel environment, Writing}, isbn = {0-7695-2312-9}, doi = {10.1109/IPDPS.2005.144}, author = {Kim,Jik-Soo and Andrade,H. and Sussman, Alan} } @article {12931, title = {Critical Factors Influencing the Occurrence of Vibrio Cholerae in the Environment of Bangladesh}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {71}, year = {2005}, month = {2005/08/01/}, pages = {4645 - 4654}, abstract = {The occurrence of outbreaks of cholera in Africa in 1970 and in Latin America in 1991, mainly in coastal communities, and the appearance of the new serotype Vibrio cholerae O139 in India and subsequently in Bangladesh have stimulated efforts to understand environmental factors influencing the growth and geographic distribution of epidemic Vibrio cholerae serotypes. Because of the severity of recent epidemics, cholera is now being considered by some infectious disease investigators as a {\textquotedblleft}reemerging{\textquotedblright} disease, prompting new work on the ecology of vibrios. Epidemiological and ecological surveillance for cholera has been under way in four rural, geographically separated locations in Bangladesh for the past 4 years, during which both clinical and environmental samples were collected at biweekly intervals. The clinical epidemiology portion of the research has been published (Sack et al., J. Infect. Dis. 187:96-101, 2003). The results of environmental sampling and analysis of the environmental and clinical data have revealed significant correlations of water temperature, water depth, rainfall, conductivity, and copepod counts with the occurrence of cholera toxin-producing bacteria (presumably V. cholerae). The lag periods between increases or decreases in units of factors, such as temperature and salinity, and occurrence of cholera correlate with biological parameters, e.g., plankton population blooms. The new information on the ecology of V. cholerae is proving useful in developing environmental models for the prediction of cholera epidemics.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.71.8.4645-4654.2005}, url = {http://aem.asm.org/content/71/8/4645}, author = {Huq,Anwar and Sack,R. Bradley and Nizam,Azhar and Longini,Ira M. and Nair,G. Balakrish and Ali,Afsar and Morris,J. Glenn and Khan,M. N. Huda and Siddique,A. Kasem and Yunus,Mohammed and Albert,M. John and Sack,David A. and Rita R Colwell} } @inbook {14211, title = {Detecting Independent 3D Movement}, booktitle = {Handbook of Geometric ComputingHandbook of Geometric Computing}, year = {2005}, month = {2005///}, pages = {383 - 401}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, isbn = {978-3-540-28247-1}, url = {http://dx.doi.org/10.1007/3-540-28247-5_12}, author = {Corrochano,Eduardo Bayro and Ogale, Abhijit S. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {14175, title = {Discovering a language for human activity}, booktitle = {Proceedings of the AAAI 2005 Fall Symposium on Anticipatory Cognitive Embodied Systems, Washington, DC}, year = {2005}, month = {2005///}, author = {Guerra-Filho,G. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {16879, title = {Efficient query processing on spatial networks}, booktitle = {Proceedings of the 13th annual ACM international workshop on Geographic information systems}, series = {GIS {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {200 - 209}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {A framework for determining the shortest path and the distance between every pair of vertices on a spatial network is presented. The framework, termed SILC, uses path coherence between the shortest path and the spatial positions of vertices on the spatial network, thereby, resulting in an encoding that is compact in representation and fast in path and distance retrievals. Using this framework, a wide variety of spatial queries such as incremental nearest neighbor searches and spatial distance joins can be shown to work on datasets of locations residing on a spatial network of sufficiently large size. The suggested framework is suitable for both main memory and disk-resident datasets.}, keywords = {location-based services, path coherence, Query processing, SILC framework, Spatial databases, spatial networks}, isbn = {1-59593-146-5}, doi = {10.1145/1097064.1097093}, url = {http://doi.acm.org/10.1145/1097064.1097093}, author = {Sankaranarayanan,Jagan and Alborzi,Houman and Samet, Hanan} } @article {17592, title = {Efficient strategies for channel management in wireless lans}, journal = {UMD, CS Tech. Rep. CS-TR}, volume = {4729}, year = {2005}, month = {2005///}, abstract = {We define efficient algorithms for channel management (channelassignment and load balancing among APs) in 802.11-based WLANs that lead to better usage of the wireless spectrum. These algorithms (called CFAssign) are based on a {\textquotedblleft}conflict-free set coloring{\textquotedblright} for- mulation that jointly perform load balancing along with channel assignment. Such a formulation has a number of advantages. First, it explicitly captures interference effects at clients. Next, it intrinsi- cally exposes opportunities for better channel re-use. Finally, algo- rithms based on this formulation do not depend on specific physical RF models and hence can be applied efficiently to a wide-range of in-building as well as outdoor scenarios. We have performed extensive packet-level simulations and mea- surements on a deployed wireless testbed of 70 APs to validate the performance of our proposed algorithms. We show that in ad- dition to single network scenarios, CFAssign algorithms are well suited for channel assignment in scenarios where multiple wireless networks share the same physical space and contend for the same frequency spectrum. Our results over a wide range of scenarios indicate that CFAssign reduces the interference at clients by about 50-70\% in comparison to current best-known techniques. }, author = {Mishra,A. and Brik,V. and Banerjee,S. and Srinivasan, Aravind and Arbaugh, William A.} } @conference {12664, title = {Ego-Motion Estimation and 3D Model Refinement in Scenes with Varying Illumination}, booktitle = {Application of Computer Vision, 2005. WACV/MOTIONS {\textquoteright}05 Volume 1. Seventh IEEE Workshops on}, volume = {2}, year = {2005}, month = {2005/01//}, pages = {140 - 146}, abstract = {We present an iterative algorithm for robustly estimating the ego-motion and refining and updating a coarse depth map using surface parallax and a generalized dynamic image (GDI) model. Given a coarse depth map acquired by a range-finder or extracted from a Digital Elevation Map (DEM), we first estimate the ego-motion by combining a global ego-motion constraint and a local GDI model. Using the estimated camera motion and the available depth estimate, motion of the 3D points is compensated. We utilize the fact that the resulting surface parallax field is an epipolar field and constrain its direction using the previous motion estimates. We then estimate the magnitude of the parallax field and the GDI model parameters locally and use them to refine the depth map estimates. We use a tensor based approach to formulate the depth refinement procedure as an eigen-value problem and obtain confidence measures for determining the accuracy of the estimated depth values. These confidence measures are used to remove regions with potentially incorrect depth estimates for robustly estimating ego-motion in the next iteration. Experimental results using both synthetic and real data are presented. Comparisons with results obtained using a brightness constancy (BC) model show that the proposed algorithm works significantly better when time-varying illumination changes are present in the scene.}, doi = {10.1109/ACVMOT.2005.45}, author = {Agrawal,Amit K and Chellapa, Rama} } @article {16389, title = {EvoSTOC Contributions}, journal = {Applications of evolutionary computing: EvoWorkshops 2005, EvoBIO, EvoCOMNET, EvoHOT, EvoIASP, EvoMUSART, and EvoSTOC}, year = {2005}, month = {2005///}, author = {Merkle,D. and Middendorf,M. and Scheidler,A. and Avigad,G. and Moshaiov,A. and Brauner,N. and Parsopoulos,K.E. and Vrahatis,M.N. and Rand, William and Riolo,R} } @conference {12071, title = {Exploiting partially overlapping channels in wireless networks: Turning a peril into an advantage}, booktitle = {Proceedings of the 5th ACM SIGCOMM conference on Internet Measurement}, year = {2005}, month = {2005///}, pages = {29 - 29}, author = {Mishra,A. and Rozner,E. and Banerjee,S. and Arbaugh, William A.} } @conference {12651, title = {Face recognition in the presence of multiple illumination sources}, booktitle = {Computer Vision, 2005. ICCV 2005. Tenth IEEE International Conference on}, volume = {2}, year = {2005}, month = {2005/10//}, pages = {1169 -1176 Vol. 2 - 1169 -1176 Vol. 2}, abstract = {Most existing face recognition algorithms work well for controlled images but are quite susceptible to changes in illumination and pose. This has led to the rise of analysis-by-synthesis approaches due to their inherent potential to handle these external factors. Though these approaches work quite well, most of them assume that the face is illuminated by a single light source which is usually not true in realistic conditions. In this paper, we propose an algorithm to recognize faces illuminated by arbitrarily placed, multiple light sources. The algorithm does not need to know the number of light sources and works extremely well even while recognizing faces illuminated by different number of light sources. Results using this algorithm are reported on multiple-illumination datasets generated from PIE by T. Sim, et al. (2003) and Yale Face Database B by W. Zhao, et al. (2003). We also highlight the importance of the hard non-linearity in the Lambert{\textquoteright}s law which is often ignored, probably to linearize the estimation process}, keywords = {approach;estimation, illumination, Lambert, law;analysis-by-synthesis, nonlinearity;multiple, process;face, recognition;hard, recognition;light, source;face, sources;}, doi = {10.1109/ICCV.2005.92}, author = {Aggarwal,G. and Chellapa, Rama} } @inbook {19494, title = {Fault-Tolerant Middleware and the Magical 1\%}, booktitle = {Middleware 2005}, series = {Lecture Notes in Computer Science}, year = {2005}, month = {2005/01/01/}, pages = {431 - 441}, publisher = {Springer Berlin Heidelberg}, organization = {Springer Berlin Heidelberg}, abstract = {Through an extensive experimental analysis of over 900 possible configurations of a fault-tolerant middleware system, we present empirical evidence that the unpredictability inherent in such systems arises from merely 1\% of the remote invocations. The occurrence of very high latencies cannot be regulated through parameters such as the number of clients, the replication style and degree or the request rates. However, by selectively filtering out a {\textquotedblleft}magical 1\%{\textquotedblright} of the raw observations of various metrics, we show that performance, in terms of measured end-to-end latency and throughput, can be bounded, easy to understand and control. This simple statistical technique enables us to guarantee, with some level of confidence, bounds for percentile-based quality of service (QoS) metrics, which dramatically increase our ability to tune and control a middleware system in a predictable manner.}, keywords = {Computer Communication Networks, Information Systems Applications (incl.Internet), Operating systems, Programming Languages, Compilers, Interpreters, Programming Techniques, software engineering}, isbn = {978-3-540-30323-7, 978-3-540-32269-6}, url = {http://link.springer.com/chapter/10.1007/11587552_24}, author = {Tudor Dumitras and Narasimhan, Priya}, editor = {Alonso, Gustavo} } @inbook {12678, title = {Fusing Depth and Video Using Rao-Blackwellized Particle Filter}, booktitle = {Pattern Recognition and Machine IntelligencePattern Recognition and Machine Intelligence}, series = {Lecture Notes in Computer Science}, volume = {3776}, year = {2005}, month = {2005///}, pages = {521 - 526}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We address the problem of fusing sparse and noisy depth data obtained from a range finder with features obtained from intensity images to estimate ego-motion and refine 3D structure of a scene using a Rao-Blackwellized particle filter. For scenes with low depth variability, the algorithm shows an alternate way of performing Structure from Motion (SfM) starting with a flat depth map. Instead of using 3D depths, we formulate the problem using 2D image domain parallax and show that conditioned on non-linear motion parameters, the parallax magnitude with respect to the projection of the vanishing point forms a linear subsystem independent of camera motion and their distributions can be analytically integrated. Thus, the structure is obtained by estimating parallax with respect to the given depths using a Kalman filter and only the ego-motion is estimated using a particle filter. Hence, the required number of particles becomes independent of the number of feature points which is an improvement over previous algorithms. Experimental results on both synthetic and real data show the effectiveness of our approach.}, isbn = {978-3-540-30506-4}, url = {http://dx.doi.org/10.1007/11590316_82}, author = {Agrawal,Amit and Chellapa, Rama}, editor = {Pal,Sankar and Bandyopadhyay,Sanghamitra and Biswas,Sambhunath} } @article {17177, title = {Generating and Querying Semantic Web Environments for Photo Libraries (2005)}, journal = {Institute for Systems Research Technical Reports}, year = {2005}, month = {2005///}, abstract = {Online photo libraries require a method to efficiently search a collection of photographs, and retrieve photos with similar attributes. Our motivation was to incorporate an existing collection of over 250 photographs of over 200 faculty members and events spanning 7 decades into a library called CS PhotoHistory that is available in hypertext and on the Semantic Web. In this paper, we identify challenges related to making this repository available on the Semantic Web, including issues of automation, modeling, and expressivity. Using CS PhotoHistory as a case study, we describe the process of creating an ontology and a querying interface for interacting with a digital photo library on the Semantic Web.}, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/6538}, author = {Axelrod,Adam and Golbeck,Jennifer and Shneiderman, Ben} } @conference {18394, title = {Generating testable hypotheses from tacit knowledge for high productivity computing}, booktitle = {Proceedings of the second international workshop on Software engineering for high performance computing system applications}, series = {SE-HPCS {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {17 - 21}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {In this research, we are developing our understanding of how the high performance computing community develops effective parallel implementations of programs by collecting the folklore within the community. We use this folklore as the basis for a series of experiments, which we expect, will validate or negate these assumptions.}, keywords = {folklore elicitation, high productivity development time experimental studies, hypothesis generation, tacit knowledge solicitation, testable hypotheses}, isbn = {1-59593-117-1}, doi = {10.1145/1145319.1145325}, url = {http://doi.acm.org/10.1145/1145319.1145325}, author = {Asgari, Sima and Hochstein, Lorin and Basili, Victor R. and Zelkowitz, Marvin V and Hollingsworth, Jeffrey K and Carver, Jeff and Shull, Forrest} } @article {19077, title = {The genome of the protist parasite Entamoeba histolytica}, journal = {Nature}, volume = {433}, year = {2005}, month = {2005}, pages = {865 - 868}, author = {Loftus,B. and Anderson,I. and Davies,R. and Alsmark,U. C.M and Samuelson,J. and Amedeo,P. and Roncaglia,P. and Berriman,M. and Hirt,R. P and Mann,B. J and others} } @article {14071, title = {The Genome Sequence of Trypanosoma cruzi, Etiologic Agent of Chagas Disease}, journal = {Science}, volume = {309}, year = {2005}, month = {2005/07/15/}, pages = {409 - 415}, abstract = {Whole-genome sequencing of the protozoan pathogen Trypanosoma cruzi revealed that the diploid genome contains a predicted 22,570 proteins encoded by genes, of which 12,570 represent allelic pairs. Over 50\% of the genome consists of repeated sequences, such as retrotransposons and genes for large families of surface molecules, which include trans-sialidases, mucins, gp63s, and a large novel family (>1300 copies) of mucin-associated surface protein (MASP) genes. Analyses of the T. cruzi, T. brucei, and Leishmania major (Tritryp) genomes imply differences from other eukaryotes in DNA repair and initiation of replication and reflect their unusual mitochondrial DNA. Although the Tritryp lack several classes of signaling molecules, their kinomes contain a large and diverse set of protein kinases and phosphatases; their size and diversity imply previously unknown interactions and regulatory processes, which may be targets for intervention.}, doi = {10.1126/science.1112631}, url = {http://www.sciencemag.org/content/309/5733/409.abstract}, author = {El-Sayed, Najib M. and Myler,Peter J. and Bartholomeu,Daniella C. and Nilsson,Daniel and Aggarwal,Gautam and Tran,Anh-Nhi and Ghedin,Elodie and Worthey,Elizabeth A. and Delcher,Arthur L. and Blandin,Ga{\"e}lle and Westenberger,Scott J. and Caler,Elisabet and Cerqueira,Gustavo C. and Branche,Carole and Haas,Brian and Anupama,Atashi and Arner,Erik and {\r A}slund,Lena and Attipoe,Philip and Bontempi,Esteban and Bringaud,Fr{\'e}d{\'e}ric and Burton,Peter and Cadag,Eithon and Campbell,David A. and Carrington,Mark and Crabtree,Jonathan and Darban,Hamid and da Silveira,Jose Franco and de Jong,Pieter and Edwards,Kimberly and Englund,Paul T. and Fazelina,Gholam and Feldblyum,Tamara and Ferella,Marcela and Frasch,Alberto Carlos and Gull,Keith and Horn,David and Hou,Lihua and Huang,Yiting and Kindlund,Ellen and Klingbeil,Michele and Kluge,Sindy and Koo,Hean and Lacerda,Daniela and Levin,Mariano J. and Lorenzi,Hernan and Louie,Tin and Machado,Carlos Renato and McCulloch,Richard and McKenna,Alan and Mizuno,Yumi and Mottram,Jeremy C. and Nelson,Siri and Ochaya,Stephen and Osoegawa,Kazutoyo and Pai,Grace and Parsons,Marilyn and Pentony,Martin and Pettersson,Ulf and Pop, Mihai and Ramirez,Jose Luis and Rinta,Joel and Robertson,Laura and Salzberg,Steven L. and Sanchez,Daniel O. and Seyler,Amber and Sharma,Reuben and Shetty,Jyoti and Simpson,Anjana J. and Sisk,Ellen and Tammi,Martti T. and Tarleton,Rick and Teixeira,Santuza and Van Aken,Susan and Vogt,Christy and Ward,Pauline N. and Wickstead,Bill and Wortman,Jennifer and White,Owen and Fraser,Claire M. and Stuart,Kenneth D. and Andersson,Bj{\"o}rn} } @article {16291, title = {The Genome Sequence of Trypanosoma Cruzi, Etiologic Agent of Chagas Disease}, journal = {ScienceScience}, volume = {309}, year = {2005}, month = {2005/07/15/}, pages = {409 - 415}, abstract = {Whole-genome sequencing of the protozoan pathogen Trypanosoma cruzi revealed that the diploid genome contains a predicted 22,570 proteins encoded by genes, of which 12,570 represent allelic pairs. Over 50\% of the genome consists of repeated sequences, such as retrotransposons and genes for large families of surface molecules, which include trans-sialidases, mucins, gp63s, and a large novel family (>1300 copies) of mucin-associated surface protein (MASP) genes. Analyses of the T. cruzi, T. brucei, and Leishmania major (Tritryp) genomes imply differences from other eukaryotes in DNA repair and initiation of replication and reflect their unusual mitochondrial DNA. Although the Tritryp lack several classes of signaling molecules, their kinomes contain a large and diverse set of protein kinases and phosphatases; their size and diversity imply previously unknown interactions and regulatory processes, which may be targets for intervention.}, isbn = {0036-8075, 1095-9203}, doi = {10.1126/science.1112631}, url = {http://www.sciencemag.org/content/309/5733/409}, author = {El-Sayed, Najib M. and Myler,Peter J. and Bartholomeu,Daniella C. and Nilsson,Daniel and Aggarwal,Gautam and Tran,Anh-Nhi and Ghedin,Elodie and Worthey,Elizabeth A. and Delcher,Arthur L. and Blandin,Ga{\"e}lle and Westenberger,Scott J. and Caler,Elisabet and Cerqueira,Gustavo C. and Branche,Carole and Haas,Brian and Anupama,Atashi and Arner,Erik and {\r A}slund,Lena and Attipoe,Philip and Bontempi,Esteban and Bringaud,Fr{\'e}d{\'e}ric and Burton,Peter and Cadag,Eithon and Campbell,David A. and Carrington,Mark and Crabtree,Jonathan and Darban,Hamid and da Silveira,Jose Franco and de Jong,Pieter and Edwards,Kimberly and Englund,Paul T. and Fazelina,Gholam and Feldblyum,Tamara and Ferella,Marcela and Frasch,Alberto Carlos and Gull,Keith and Horn,David and Hou,Lihua and Huang,Yiting and Kindlund,Ellen and Klingbeil,Michele and Kluge,Sindy and Koo,Hean and Lacerda,Daniela and Levin,Mariano J. and Lorenzi,Hernan and Louie,Tin and Machado,Carlos Renato and McCulloch,Richard and McKenna,Alan and Mizuno,Yumi and Mottram,Jeremy C. and Nelson,Siri and Ochaya,Stephen and Osoegawa,Kazutoyo and Pai,Grace and Parsons,Marilyn and Pentony,Martin and Pettersson,Ulf and Pop, Mihai and Ramirez,Jose Luis and Rinta,Joel and Robertson,Laura and Salzberg,Steven L. and Sanchez,Daniel O. and Seyler,Amber and Sharma,Reuben and Shetty,Jyoti and Simpson,Anjana J. and Sisk,Ellen and Tammi,Martti T. and Tarleton,Rick and Teixeira,Santuza and Van Aken,Susan and Vogt,Christy and Ward,Pauline N. and Wickstead,Bill and Wortman,Jennifer and White,Owen and Fraser,Claire M. and Stuart,Kenneth D. and Andersson,Bj{\"o}rn} } @article {19073, title = {The genome sequence of Trypanosoma cruzi, etiologic agent of Chagas disease}, journal = {Science}, volume = {309}, year = {2005}, month = {2005}, author = {El-Sayed, N.M. and Myler,P. J and Bartholomeu,D. C and Nilsson,D. and Aggarwal,G. and Tran,A. N and Ghedin,E. and Worthey,E. A and Delcher,A. L and Blandin,G. and others} } @conference {11897, title = {Identifying and segmenting human-motion for mobile robot navigation using alignment errors}, booktitle = {12th International Conference on Advanced Robotics, 2005. ICAR {\textquoteright}05. Proceedings}, year = {2005}, month = {2005/07//}, pages = {398 - 403}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper presents a new human-motion identification and segmentation algorithm, for mobile robot platforms. The algorithm is based on computing the alignment error between pairs of object images acquired from a moving platform. Pairs of images generating relatively small alignment errors are used to estimate the fundamental frequency of the object{\textquoteright}s motion. A decision criterion is then used to test the significance of the estimated frequency and to classify the object{\textquoteright}s motion. To verify the validity of the proposed approach, experimental results are shown on different classes of objects}, keywords = {Computer errors, Educational institutions, Frequency estimation, human-motion identification, human-motion segmentation, HUMANS, Image motion analysis, Image segmentation, mobile robot navigation, Mobile robots, Motion estimation, Navigation, Object detection, robot vision, SHAPE}, isbn = {0-7803-9178-0}, doi = {10.1109/ICAR.2005.1507441}, author = {Abd-Almageed, Wael and Burns,B. J and Davis, Larry S.} } @inbook {12677, title = {Image Sequence Stabilization, Mosaicking, and Superresolution}, booktitle = {Handbook of Image and Video Processing (Second Edition)Handbook of Image and Video Processing (Second Edition)}, year = {2005}, month = {2005///}, pages = {309-VII - 309-VII}, publisher = {Academic Press}, organization = {Academic Press}, address = {Burlington}, abstract = {A sequence of temporal images gathered from a single sensor adds a whole new dimension to two-dimensional (2D) image data. Availability of an image sequence permits the measurement of quantities such as subpixel intensities, camera motion and depth, and detection and tracking of moving objects. In turn, the processing of image sequences necessitates the development of sophisticated techniques to extract this information. With the recent availability of powerful yet inexpensive computers, data storage systems, and image acquisition devices, image sequence analysis has transitioned from an esoteric research domain to a practical area with significant commercial interest.}, isbn = {978-0-12-119792-6}, url = {http://www.sciencedirect.com/science/article/pii/B9780121197926500826}, author = {Chellapa, Rama and Srinivasan, S. and Aggarwal,G. and Veeraraghavan,A.}, editor = {Al Bovik} } @inbook {16103, title = {Information Visualization and the Challenge of Universal Usability}, booktitle = {Exploring GeovisualizationExploring Geovisualization}, year = {2005}, month = {2005///}, pages = {53 - 82}, publisher = {Elsevier}, organization = {Elsevier}, address = {Oxford}, abstract = {Information Visualization aims to provide compact graphical presentations and user interfaces for interactively manipulating large numbers of items. We present a simple {\textquotedblleft}data by tasks taxonomy{\textquotedblright} then discuss the challenges of providing universal usability, with example applications using geo-referenced data. Information Visualization has been shown to be a powerful visual thinking or decision tool but it is becoming important for services to reach and empower every citizen. Technological advances are needed to deal with user diversity (age, language, disabilities, etc.) but also with the variety of technology used (screen size, network speed, etc.) and the gaps in user{\textquoteright}s knowledge (general knowledge, knowledge of the application domain, of the interface syntax or semantic). We present examples that illustrate how those challenges can be addressed.}, isbn = {978-0-08-044531-1}, url = {http://www.sciencedirect.com/science/article/pii/B9780080445311504218}, author = {Plaisant, Catherine}, editor = {Jason Dykes and Alan M. MacEachren and Menno-Jan KraakA2 - Jason Dykes,Alan M. MacEachren and Menno-Jan Kraak} } @article {17255, title = {Interactive pattern search in time series}, journal = {Proceedings of SPIE}, volume = {5669}, year = {2005}, month = {2005/03/11/}, pages = {175 - 186}, abstract = {The need for pattern discovery in long time series data led researchers to develop algorithms for similarity search. Most of the literature about time series focuses on algorithms that index time series and bring the data into the main storage, thus providing fast information retrieval on large time series. This paper reviews the state of the art in visualizing time series, and focuses on techniques that enable users to visually and interactively query time series. Then, it presents TimeSearcher 2, a tool that enables users to explore multidimensional data using synchronized tables and graphs with overview+detail, filter the time series data to reduce the scope of the search, select an existing pattern to find similar occurrences, and interactively adjust similarity parameters to narrow the result set. This tool is an extension of previous work, TimeSearcher 1, which uses graphical timeboxes to interactively query time series data.}, isbn = {0277786X}, doi = {doi:10.1117/12.587537}, url = {http://spiedigitallibrary.org/proceedings/resource/2/psisdg/5669/1/175_1?isAuthorized=no}, author = {Buono,Paolo and Aris,Aleks and Plaisant, Catherine and Khella,Amir and Shneiderman, Ben} } @article {17256, title = {Interactive Pattern Search in Time Series (2004)}, year = {2005}, month = {2005///}, abstract = {The need for pattern discovery in long time series data led researchers to develop algorithms for similarity search. Most of the literature about time series focuses on algorithms that index time series and bring the data into the main storage, thus providing fast information retrieval on large time series. This paper reviews the state of the art in visualizing time series, and focuses on techniques that enable users to interactively query time series. Then it presents TimeSearcher 2, a tool that enables users to explore multidimensional data using coordinated tables and graphs with overview+detail, filter the time series data to reduce the scope of the search, select an existing pattern to find similar occurrences, and interactively adjust similarity parameters to narrow the result set. This tool is an extension of previous work, TimeSearcher 1, which uses graphical timeboxes to interactively query time series data.}, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/6519}, author = {Buono,Paolo and Aris,Aleks and Plaisant, Catherine and Khella,Amir and Shneiderman, Ben} } @conference {16384, title = {The Java memory model}, booktitle = {Proceedings of the 32nd ACM SIGPLAN-SIGACT symposium on Principles of programming languages}, series = {POPL {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {378 - 391}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {This paper describes the new Java memory model, which has been revised as part of Java 5.0. The model specifies the legal behaviors for a multithreaded program; it defines the semantics of multithreaded Java programs and partially determines legal implementations of Java virtual machines and compilers.The new Java model provides a simple interface for correctly synchronized programs -- it guarantees sequential consistency to data-race-free programs. Its novel contribution is requiring that the behavior of incorrectly synchronized programs be bounded by a well defined notion of causality. The causality requirement is strong enough to respect the safety and security properties of Java and weak enough to allow standard compiler and hardware optimizations. To our knowledge, other models are either too weak because they do not provide for sufficient safety/security, or are too strong because they rely on a strong notion of data and control dependences that precludes some standard compiler transformations.Although the majority of what is currently done in compilers is legal, the new model introduces significant differences, and clearly defines the boundaries of legal transformations. For example, the commonly accepted definition for control dependence is incorrect for Java, and transformations based on it may be invalid.In addition to providing the official memory model for Java, we believe the model described here could prove to be a useful basis for other programming languages that currently lack well-defined models, such as C++ and C$\#$.}, keywords = {concurrency, java, memory model, Multithreading}, isbn = {1-58113-830-X}, doi = {10.1145/1040305.1040336}, url = {http://doi.acm.org/10.1145/1040305.1040336}, author = {Manson,Jeremy and Pugh, William and Adve,Sarita V.} } @conference {18911, title = {Learning approximate preconditions for methods in hierarchical plans}, series = {ICML {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {337 - 344}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {A significant challenge in developing planning systems for practical applications is the difficulty of acquiring the domain knowledge needed by such systems. One method for acquiring this knowledge is to learn it from plan traces, but this method typically requires a huge number of plan traces to converge. In this paper, we show that the problem with slow convergence can be circumvented by having the learner generate solution plans even before the planning domain is completely learned. Our empirical results show that these improvements reduce the size of the training set that is needed to find correct answers to a large percentage of planning problems in the test set.}, isbn = {1-59593-180-5}, doi = {10.1145/1102351.1102394}, url = {http://doi.acm.org/10.1145/1102351.1102394}, author = {Ilghami,Okhtay and Mu{\~n}oz-Avila,H{\'e}ctor and Nau, Dana S. and Aha,David W.} } @conference {11900, title = {A learning automata based power management for ad-hoc networks}, booktitle = {2005 IEEE International Conference on Systems, Man and Cybernetics}, volume = {4}, year = {2005}, month = {2005/10//}, pages = {3569- 3573 Vol. 4 - 3569- 3573 Vol. 4}, publisher = {IEEE}, organization = {IEEE}, abstract = {Power management is a very important aspect of ad-hoc networks. It directly impacts the network throughput among other network metrics. On the other hand, transmission power management may result in disconnected networks and increased level of collisions. In this paper, we introduce a transmission power control based on stochastic learning automata (SLA) to modify the transmission power. Based on the level of successful transmissions and the level of packet retransmissions, the SLA will modify the transmission power level either by increasing it or decreasing it. The probabilistic nature of SLA makes it a useful choice for ad-hoc networks. Using the network simulator NS, we show that using SLA for transmission power will result in an increased system bandwidth and a decrease in the collision levels.}, keywords = {ad hoc networks, ad-hoc networks, Computer network management, Computer networks, Energy management, Engineering management, learning automata, network metrics, network simulator, packet retransmissions, power control, power system management, power transmission control, stochastic learning automata, stochastic learning automta, Stochastic processes, system bandwidth, Technology management, Throughput, transmission power control, transmission power level, transmission power management}, isbn = {0-7803-9298-1}, doi = {10.1109/ICSMC.2005.1571701}, author = {El-Osery,A. I and Baird,D. and Abd-Almageed, Wael} } @article {18912, title = {Learning preconditions for planning from plan traces and HTN structure}, journal = {Computational Intelligence}, volume = {21}, year = {2005}, month = {2005///}, pages = {388 - 413}, abstract = {A great challenge in developing planning systems for practical applications is the difficulty of acquiring the domain information needed to guide such systems. This paper describes a way to learn some of that knowledge. More specifically, the following points are discussed. (1) We introduce a theoretical basis for formally defining algorithms that learn preconditions for Hierarchical Task Network (HTN) methods. (2) We describe Candidate Elimination Method Learner (CaMeL), a supervised, eager, and incremental learning process for preconditions of HTN methods. We state and prove theorems about CaMeL{\textquoteright}s soundness, completeness, and convergence properties. (3) We present empirical results about CaMeL{\textquoteright}s convergence under various conditions. Among other things, CaMeL converges the fastest on the preconditions of the HTN methods that are needed the most often. Thus CaMeL{\textquoteright}s output can be useful even before it has fully converged.}, keywords = {candidate elimination, HTN planning, learning, version spaces}, isbn = {1467-8640}, doi = {10.1111/j.1467-8640.2005.00279.x}, url = {http://onlinelibrary.wiley.com/doi/10.1111/j.1467-8640.2005.00279.x/abstract}, author = {Ilghami,Okhtay and Nau, Dana S. and Mu{\~n}oz-Avila,H{\'e}ctor and Aha,David W.} } @article {15990, title = {Logic, self-awareness and self-improvement: The metacognitive loop and the problem of brittleness}, journal = {Journal of Logic and Computation}, volume = {15}, year = {2005}, month = {2005///}, pages = {21 - 21}, author = {Anderson,M. L and Perlis, Don} } @article {12645, title = {Matching shape sequences in video with applications in human movement analysis}, journal = {Pattern Analysis and Machine Intelligence, IEEE Transactions on}, volume = {27}, year = {2005}, month = {2005/12//}, pages = {1896 - 1909}, abstract = {We present an approach for comparing two sequences of deforming shapes using both parametric models and nonparametric methods. In our approach, Kendall{\textquoteright}s definition of shape is used for feature extraction. Since the shape feature rests on a non-Euclidean manifold, we propose parametric models like the autoregressive model and autoregressive moving average model on the tangent space and demonstrate the ability of these models to capture the nature of shape deformations using experiments on gait-based human recognition. The nonparametric model is based on dynamic time-warping. We suggest a modification of the dynamic time-warping algorithm to include the nature of the non-Euclidean space in which the shape deformations take place. We also show the efficacy of this algorithm by its application to gait-based human recognition. We exploit the shape deformations of a person{\textquoteright}s silhouette as a discriminating feature and provide recognition results using the nonparametric model. Our analysis leads to some interesting observations on the role of shape and kinematics in automated gait-based person authentication.}, keywords = {Automated;Photography;Reproducibility of Results;Sensitivity and Specificity;Subtraction Technique;Video Recording;, autoregressive moving average model;deforming shapes;dynamic time-warping;feature extraction;gait-based human recognition;human movement analysis;matching shape sequences;video;autoregressive moving average processes;feature extraction;image matching;imag, Biological;Models, Computer-Assisted;Imaging, Statistical;Movement;Pattern Recognition, Three-Dimensional;Leg;Models}, isbn = {0162-8828}, doi = {10.1109/TPAMI.2005.246}, author = {A,Veeraraghavan and Roy-Chowdhury, A.K. and Chellapa, Rama} } @conference {18402, title = {Measuring productivity on high performance computers}, booktitle = {11th IEEE International Symposium on Software Metric}, year = {2005}, month = {2005///}, abstract = {In the high performance computing domain, the speed ofexecution of a program has typically been the primary performance metric. But productivity is also of concern to high performance computing developers. In this paper we will discuss the problems of defining and measuring productivity for these machines and we develop a model of productivity that includes both a performance component and a component that measures the development time of the program. We ran several experiments using students in high performance courses at several universities, and we report on those results with respect to our model of productivity. }, author = {Zelkowitz, Marvin V and Basili, Victor R. and Asgari, S. and Hochstein, L. and Hollingsworth, Jeffrey K and Nakamura, T.} } @conference {15924, title = {Metacognition for dropping and reconsidering intentions}, booktitle = {AAAI Spring Symposium on Metacognition in Computation}, year = {2005}, month = {2005///}, author = {Josyula,D. P and Anderson,M. L and Perlis, Don} } @article {11910, title = {Method and system for providing physical security in an area of interest}, volume = {10/916,545}, year = {2005}, month = {2005/03/10/}, abstract = {A system for detecting the presence of an intruder in a protected area utilizes a received signal strength indicator (RSSI) value of signals broadcast from transmitting stations deployed in the protected area. The system includes monitoring points for receiving broadcast signals, measuring the RSSI values of the received signals, and transmitting the measured RSSI values to a security system server. The security system server analyzes the RSSI values, and initiates security measures when the physical security of the protected area is violated which is detected when the measured RSSI values deviate from a predetermined strength of the broadcast signals. The security system also has the ability to track objects in the protected area and keep track of their movement in real time and report such movement. The system may be based on a Wi-Fi infrastructure in the protected area.}, url = {http://www.google.com/patents?id=VIaWAAAAEBAJ}, author = {Agrawala, Ashok K. and Youssef,Moustafa Amin and Shahamat,Layla} } @article {14230, title = {Motion segmentation using occlusions}, journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence}, volume = {27}, year = {2005}, month = {2005/06//}, pages = {988 - 992}, abstract = {We examine the key role of occlusions in finding independently moving objects instantaneously in a video obtained by a moving camera with a restricted field of view. In this problem, the image motion is caused by the combined effect of camera motion (egomotion), structure (depth), and the independent motion of scene entities. For a camera with a restricted field of view undergoing a small motion between frames, there exists, in general, a set of 3D camera motions compatible with the observed flow field even if only a small amount of noise is present, leading to ambiguous 3D motion estimates. If separable sets of solutions exist, motion-based clustering can detect one category of moving objects. Even if a single inseparable set of solutions is found, we show that occlusion information can be used to find ordinal depth, which is critical in identifying a new class of moving objects. In order to find ordinal depth, occlusions must not only be known, but they must also be filled (grouped) with optical flow from neighboring regions. We present a novel algorithm for filling occlusions and deducing ordinal depth under general circumstances. Finally, we describe another category of moving objects which is detected using cardinal comparisons between structure from motion and structure estimates from another source (e.g., stereo).}, keywords = {3D motion estimation, algorithms, Artificial intelligence, CAMERAS, Computer vision, Filling, hidden feature removal, Image Enhancement, Image Interpretation, Computer-Assisted, image motion, Image motion analysis, Image segmentation, Layout, MOTION, Motion detection, Motion estimation, motion segmentation, Movement, Object detection, occlusion, occlusions, optical flow, ordinal depth, Pattern Recognition, Automated, Photography, Reproducibility of results, segmentation, Semiconductor device modeling, Sensitivity and Specificity, video analysis., Video Recording}, isbn = {0162-8828}, doi = {10.1109/TPAMI.2005.123}, author = {Ogale, A. S and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @inbook {12670, title = {Moving Object Detection and Compression in IR Sequences}, booktitle = {Computer Vision Beyond the Visible SpectrumComputer Vision Beyond the Visible Spectrum}, series = {Advances in Pattern Recognition}, year = {2005}, month = {2005///}, pages = {141 - 165}, publisher = {Springer London}, organization = {Springer London}, abstract = {We consider the problem of remote surveillance using infrared (IR) sensors. The aim is to use IR image sequences to detect moving objects (humans or vehicles), and to transmit a few {\textquotedblleft}best-view images{\textquotedblright} of every new object that is detected. Since the available bandwidth is usually low, if the object chip is big, it needs to be compressed before being transmitted. Due to low computational power of computing devices attached to the sensor, the algorithms should be computationally simple. We present two approaches for object detection {\textemdash} one which specifically solves the more difficult long-range object detection problem, and the other for objects at short range. For objects at short range, we also present techniques for selecting a single best-view object chip and computationally simple techniques for compressing it to very low bit rates due to the channel bandwidth constraint. A fast image chip compression scheme implemented in the wavelet domain by combining a non-iterative zerotree coding method with 2D-DPCM for both low-and high-frequency subbands is presented. Comparisons with some existing schemes are also included. The object detection and compression algorithms have been implemented in C/C++ and their performance has been evaluated using the Hitachi{\textquoteright}s SH4 platform with software simulation.}, isbn = {978-1-84628-065-8}, url = {http://dx.doi.org/10.1007/1-84628-065-6_5}, author = {Vaswani,Namrata and Agrawal,Amit and Qinfen Zheng and Chellapa, Rama}, editor = {Bhanu,Bir and Pavlidis,Ioannis} } @conference {12665, title = {Moving Object Segmentation and Dynamic Scene Reconstruction Using Two Frames}, booktitle = {Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP {\textquoteright}05). IEEE International Conference on}, volume = {2}, year = {2005}, month = {2005//18/23}, pages = {705 - 708}, keywords = {3D, analysis;, constraints;, dynamic, ego-motion, estimation;, flow, image, images;, independent, INTENSITY, least, mean, median, method;, methods;, model;, MOTION, motion;, moving, object, of, parallax, parallax;, parametric, processing;, reconstruction;, scene, segmentation;, signal, squares, squares;, static, structure;, subspace, surface, translational, two-frame, unconstrained, video}, doi = {10.1109/ICASSP.2005.1415502}, author = {Agrawala, Ashok K. and Chellapa, Rama} } @conference {13835, title = {NeurAlign: combining word alignments using neural networks}, booktitle = {Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing}, series = {HLT {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {65 - 72}, publisher = {Association for Computational Linguistics}, organization = {Association for Computational Linguistics}, address = {Stroudsburg, PA, USA}, abstract = {This paper presents a novel approach to combining different word alignments. We view word alignment as a pattern classification problem, where alignment combination is treated as a classifier ensemble, and alignment links are adorned with linguistic features. A neural network model is used to learn word alignments from the individual alignment systems. We show that our alignment combination approach yields a significant 20--34\% relative error reduction over the best-known alignment combination technique on English-Spanish and English-Chinese data.}, doi = {10.3115/1220575.1220584}, url = {http://dx.doi.org/10.3115/1220575.1220584}, author = {Ayan,Necip Fazil and Dorr, Bonnie J and Monz,Christof} } @inbook {12676, title = {Pattern Recognition in Video}, booktitle = {Pattern Recognition and Machine IntelligencePattern Recognition and Machine Intelligence}, series = {Lecture Notes in Computer Science}, volume = {3776}, year = {2005}, month = {2005///}, pages = {11 - 20}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Images constitute data that live in a very high dimensional space, typically of the order of hundred thousand dimensions. Drawing inferences from correlated data of such high dimensions often becomes intractable. Therefore traditionally several of these problems like face recognition, object recognition, scene understanding etc. have been approached using techniques in pattern recognition. Such methods in conjunction with methods for dimensionality reduction have been highly popular and successful in tackling several image processing tasks. Of late, the advent of cheap, high quality video cameras has generated new interests in extending still image-based recognition methodologies to video sequences. The added temporal dimension in these videos makes problems like face and gait-based human recognition, event detection, activity recognition addressable. Our research has focussed on solving several of these problems through a pattern recognition approach. Of course, in video streams patterns refer to both patterns in the spatial structure of image intensities around interest points and temporal patterns that arise either due to camera motion or object motion. In this paper, we discuss the applications of pattern recognition in video to problems like face and gait-based human recognition, behavior classification, activity recognition and activity based person identification.}, isbn = {978-3-540-30506-4}, url = {http://dx.doi.org/10.1007/11590316_2}, author = {Chellapa, Rama and Veeraraghavan,Ashok and Aggarwal,Gaurav}, editor = {Pal,Sankar and Bandyopadhyay,Sanghamitra and Biswas,Sambhunath} } @conference {11908, title = {Pedestrian classification from moving platforms using cyclic motion pattern}, booktitle = {IEEE International Conference on Image Processing, 2005. ICIP 2005}, volume = {2}, year = {2005}, month = {2005/09//}, pages = {II- 854-7 - II- 854-7}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper describes an efficient pedestrian detection system for videos acquired from moving platforms. Given a detected and tracked object as a sequence of images within a bounding box, we describe the periodic signature of its motion pattern using a twin-pendulum model. Then a principle gait angle is extracted in every frame providing gait phase information. By estimating the periodicity from the phase data using a digital phase locked loop (dPLL), we quantify the cyclic pattern of the object, which helps us to continuously classify it as a pedestrian. Past approaches have used shape detectors applied to a single image or classifiers based on human body pixel oscillations, but ours is the first to integrate a global cyclic motion model and periodicity analysis. Novel contributions of this paper include: i) development of a compact shape representation of cyclic motion as a signature for a pedestrian, ii) estimation of gait period via a feedback loop module, and iii) implementation of a fast online pedestrian classification system which operates on videos acquired from moving platforms.}, keywords = {compact shape representation, cyclic motion pattern, data mining, Detectors, digital phase locked loop, digital phase locked loops, feedback loop module, gait analysis, gait phase information, human body pixel oscillations, HUMANS, image classification, Image motion analysis, image representation, image sequence, Image sequences, Motion detection, Object detection, pedestrian classification, pedestrian detection system, Phase estimation, Phase locked loops, principle gait angle, SHAPE, tracking, Videos}, isbn = {0-7803-9134-9}, doi = {10.1109/ICIP.2005.1530190}, author = {Yang Ran and Qinfen Zheng and Weiss, I. and Davis, Larry S. and Abd-Almageed, Wael and Liang Zhao} } @conference {13192, title = {Pedestrian classification from moving platforms using cyclic motion pattern}, booktitle = {Image Processing, 2005. ICIP 2005. IEEE International Conference on}, volume = {2}, year = {2005}, month = {2005/09//}, pages = {II - 854-7 - II - 854-7}, abstract = {This paper describes an efficient pedestrian detection system for videos acquired from moving platforms. Given a detected and tracked object as a sequence of images within a bounding box, we describe the periodic signature of its motion pattern using a twin-pendulum model. Then a principle gait angle is extracted in every frame providing gait phase information. By estimating the periodicity from the phase data using a digital phase locked loop (dPLL), we quantify the cyclic pattern of the object, which helps us to continuously classify it as a pedestrian. Past approaches have used shape detectors applied to a single image or classifiers based on human body pixel oscillations, but ours is the first to integrate a global cyclic motion model and periodicity analysis. Novel contributions of this paper include: i) development of a compact shape representation of cyclic motion as a signature for a pedestrian, ii) estimation of gait period via a feedback loop module, and iii) implementation of a fast online pedestrian classification system which operates on videos acquired from moving platforms.}, keywords = {analysis;, angle;, body, classification;, compact, cyclic, DETECTION, detection;, digital, Feedback, Gait, human, image, information;, locked, loop, loop;, loops;, module;, MOTION, object, oscillations;, pattern;, pedestrian, phase, Pixel, principle, representation;, sequence;, sequences;, SHAPE, system;}, doi = {10.1109/ICIP.2005.1530190}, author = {Yang Ran and Qinfen Zheng and Weiss, I. and Davis, Larry S. and Abd-Almageed, Wael and Liang Zhao} } @conference {17890, title = {Query planning for the grid: adapting to dynamic resource availability}, booktitle = {Cluster Computing and the Grid, IEEE International Symposium on}, volume = {2}, year = {2005}, month = {2005///}, pages = {751 - 758}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {The availability of massive datasets, comprising sensor measurements or the results of scientific simulations, has had a significant impact on the methodology of scientific reasoning. Scientists require storage, bandwidth and computational capacity to query and analyze these datasets, to understand physical phenomena or to test hypotheses. This paper addresses the challenge of identifying and selecting resources to develop an evaluation plan for large scale data analysis queries when data processing capabilities and datasets are dispersed across nodes in one or more computing and storage clusters. We show that generating an optimal plan is hard and we propose heuristic techniques to find a good choice of resources. We also consider heuristics to cope with dynamic resource availability; in this situation we have stale information about reusable cached results (datasets) and the load on various nodes.}, isbn = {0-7803-9074-1}, doi = {http://doi.ieeecomputersociety.org/10.1109/CCGRID.2005.1558638}, author = {Zhang,K. and Andrade,H. and Raschid, Louiqa and Sussman, Alan} } @conference {15994, title = {On the reasoning of real-world agents: Toward a semantics for active logic}, booktitle = {7-th Annual Symposium on the Logical Formalization of Commonsense Reasoning}, year = {2005}, month = {2005///}, author = {Anderson,M. L and Gomaa,W. and Grant,J. and Perlis, Don} } @inbook {16114, title = {Representing Unevenly-Spaced Time Series Data for Visualization and Interactive Exploration}, booktitle = {Human-Computer Interaction - INTERACT 2005Human-Computer Interaction - INTERACT 2005}, series = {Lecture Notes in Computer Science}, volume = {3585}, year = {2005}, month = {2005///}, pages = {835 - 846}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Visualizing time series is useful to support discovery of relations and patterns in financial, genomic, medical and other applications. Often, measurements are equally spaced over time. We discuss the challenges of unevenly-spaced time series and present fourrepresentationmethods: sampled events, aggregated sampled events, event index and interleaved event index. We developed these methods while studying eBay auction data with TimeSearcher. We describe the advantages, disadvantages, choices for algorithms and parameters, and compare the different methods for different tasks. Interaction issues such as screen resolution, response time for dynamic queries, and learnability are governed by these decisions.}, isbn = {978-3-540-28943-2}, url = {http://dx.doi.org/10.1007/11555261_66}, author = {Aris,Aleks and Shneiderman, Ben and Plaisant, Catherine and Shmueli,Galit and Jank,Wolfgang}, editor = {Costabile,Maria and Patern{\`o},Fabio} } @article {16113, title = {Representing Unevenly-Spaced Time Series Data for Visualization and Interactive Exploration (2005)}, journal = {Institute for Systems Research Technical Reports}, year = {2005}, month = {2005///}, abstract = {Visualizing time series data is useful to support discovery of relations and patterns in financial, genomic, medical and other applications. In most time series, measurements are equally spaced over time. This paper discusses the challenges for unevenly-spaced time series data and presents four methods to represent them: sampled events, aggregated sampled events, event index and interleaved event index. We developed these methods while studying eBay auction data with TimeSearcher. We describe the advantages, disadvantages, choices for algorithms and parameters, and compare the different methods. Since each method has its advantages, this paper provides guidance for choosing the right combination of methods, algorithms, and parameters to solve a given problem for unevenly-spaced time series. Interaction issues such as screen resolution, response time for dynamic queries, and meaning of the visual display are governed by these decisions.}, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/6537}, author = {Aris,Aleks and Shneiderman, Ben and Plaisant, Catherine and Shmueli,Galit and Jank,Wolfgang} } @conference {11975, title = {Robust Contrast Invariant Stereo Correspondence}, booktitle = {Proceedings of the 2005 IEEE International Conference on Robotics and Automation, 2005. ICRA 2005}, year = {2005}, month = {2005/04/18/22}, pages = {819 - 824}, publisher = {IEEE}, organization = {IEEE}, abstract = {A stereo pair of cameras attached to a robot will inevitably yield images with different contrast. Even if we assume that the camera hardware is identical, due to slightly different points of view, the amount of light entering the two cameras is also different, causing dynamically adjusted internal parameters such as aperture, exposure and gain to be different. Due to the difficulty of obtaining and maintaining precise intensity or color calibration between the two cameras, contrast invariance becomes an extremely desirable property of stereo correspondence algorithms. The problem of achieving point correspondence between a stereo pair of images is often addressed by using the intensity or color differences as a local matching metric, which is sensitive to contrast changes. We present an algorithm for contrast invariant stereo matching which relies on multiple spatial frequency channels for local matching. A fast global framework uses the local matching to compute the correspondences and find the occlusions. We demonstrate that the use of multiple frequency channels allows the algorithm to yield good results even in the presence of significant amounts of noise.}, keywords = {Apertures, Calibration, CAMERAS, Computer science, contrast invariance, diffusion, Educational institutions, Frequency, gabor, Hardware, occlusions, Robot vision systems, Robotics and automation, Robustness, stereo}, isbn = {0-7803-8914-X}, doi = {10.1109/ROBOT.2005.1570218}, author = {Ogale, A. S and Aloimonos, J.} } @article {12082, title = {Robust routing in malicious environment for ad hoc networks}, journal = {Information Security Practice and Experience}, year = {2005}, month = {2005///}, pages = {36 - 47}, author = {Yu,Z. and Seng,C. Y and Jiang,T. and Wu,X. and Arbaugh, William A.} } @article {12104, title = {ROMER: resilient opportunistic mesh routing for wireless mesh networks}, journal = {Proc. of IEEE WiMesh}, volume = {166}, year = {2005}, month = {2005///}, author = {Yuan,Y. and Yang,H. and Wong,S. H.Y and Lu,S. and Arbaugh, William A.} } @article {15939, title = {The roots of self-awareness}, journal = {Phenomenology and the Cognitive Sciences}, volume = {4}, year = {2005}, month = {2005///}, pages = {297 - 333}, author = {Anderson,M. L and Perlis, Don} } @article {12057, title = {Shape and the stereo correspondence problem}, journal = {International Journal of Computer Vision}, volume = {65}, year = {2005}, month = {2005///}, pages = {147 - 162}, author = {Ogale, A. S and Aloimonos, J.} } @conference {15638, title = {Space-time tradeoffs for approximate spherical range counting}, booktitle = {Proceedings of the sixteenth annual ACM-SIAM symposium on Discrete algorithms}, series = {SODA {\textquoteright}05}, year = {2005}, month = {2005///}, pages = {535 - 544}, publisher = {Society for Industrial and Applied Mathematics}, organization = {Society for Industrial and Applied Mathematics}, address = {Philadelphia, PA, USA}, abstract = {We present space-time tradeoffs for approximate spherical range counting queries. Given a set S of n data points in Rd along with a positive approximation factor ε, the goal is to preprocess the points so that, given any Euclidean ball B, we can return the number of points of any subset of S that contains all the points within a (1 - ε)-factor contraction of B, but contains no points that lie outside a (1 + ε)-factor expansion of B.In many applications of range searching it is desirable to offer a tradeoff between space and query time. We present here the first such tradeoffs for approximate range counting queries. Given 0 < ε <= 1/2 and a parameter γ, where 2 <= γ <= 1/ε, we show how to construct a data structure of space O(nγd log (1/ε)) that allows us to answer ε-approximate spherical range counting queries in time O(log(nγ) + 1/(εγd-1). The data structure can be built in time O(nγd log (n/ε)) log (1/ε)). Here n, ε, and γ are asymptotic quantities, and the dimension d is assumed to be a fixed constant.At one extreme (low space), this yields a data structure of space O(n log (1/e)) that can answer approximate range queries in time O(logn + 1/(ed-1) which, up to a factor of O(n log (1/e) in space, matches the best known result for approximate spherical range counting queries. At the other extreme (high space), it yields a data structure of space O((n/ed) log(1/ε)) that can answer queries in time O(logn + 1/ε). This is the fastest known query time for this problem.We also show how to adapt these data structures to the problem of computing an ε-approximation to the kth nearest neighbor, where k is any integer from 1 to n given at query time. The space bounds are identical to the range searching results, and the query time is larger only by a factor of O(1/(εγ)).Our approach is broadly based on methods developed for approximate Voronoi diagrams (AVDs), but it involves a number of significant extensions from the context of nearest neighbor searching to range searching. These include generalizing AVD node-separation properties from leaves to internal nodes of the tree and constructing efficient generator sets through a radial decomposition of space. We have also developed new arguments to analyze the time and space requirements in this more general setting.}, isbn = {0-89871-585-7}, url = {http://dl.acm.org/citation.cfm?id=1070432.1070505}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @article {18718, title = {Structural Determinants for Selective Recognition of a Lys48-Linked Polyubiquitin Chain by a UBA Domain}, journal = {Molecular Cell}, volume = {18}, year = {2005}, month = {2005/06/10/}, pages = {687 - 698}, abstract = {SummaryAlthough functional diversity in polyubiquitin chain signaling has been ascribed to the ability of differently linked chains to bind in a distinctive manner to effector proteins, structural models of such interactions have been lacking. Here, we use NMR to unveil the structural basis of selective recognition of Lys48-linked di- and tetraubiquitin chains by the UBA2 domain of hHR23A. Although the interaction of UBA2 with Lys48-linked diubiquitin involves the same hydrophobic surface on each ubiquitin unit as that utilized in monoubiquitin:UBA complexes, our results show how the {\textquotedblleft}closed{\textquotedblright} conformation of Lys48-linked diubiquitin is crucial for high-affinity binding. Moreover, recognition of Lys48-linked diubiquitin involves a unique epitope on UBA, which allows the formation of a sandwich-like diubiqutin:UBA complex. Studies of the UBA-tetraubiquitin interaction suggest that this mode of UBA binding to diubiquitin is relevant for longer chains. }, isbn = {1097-2765}, doi = {10.1016/j.molcel.2005.05.013}, url = {http://www.sciencedirect.com/science/article/pii/S1097276505013195}, author = {Varadan,Ranjani and Assfalg,Michael and Raasi,Shahri and Pickart,Cecile and Fushman, David} } @conference {12099, title = {Toward resilient security in wireless sensor networks}, booktitle = {Proceedings of the 6th ACM international symposium on Mobile ad hoc networking and computing}, year = {2005}, month = {2005///}, pages = {34 - 45}, author = {Yang,H. and Ye,F. and Yuan,Y. and Lu,S. and Arbaugh, William A.} } @conference {17444, title = {Turning information visualization innovations into commercial products: lessons to guide the next success}, booktitle = {IEEE Symposium on Information Visualization, 2005. INFOVIS 2005}, year = {2005}, month = {2005/10/23/25}, pages = {241 - 244}, publisher = {IEEE}, organization = {IEEE}, abstract = {As information visualization matures as an academic research field, commercial spinoffs are proliferating, but success stories are harder to find. This is the normal process of emergence for new technologies, but the panel organizers believe that there are certain strategies that facilitate success. To teach these lessons, we have invited several key figures who are seeking to commercialize information visualization tools. The panelists make short presentations, engage in a moderated discussion, and respond to audience questions.}, keywords = {Books, commercial development, commercial product, Computer interfaces, Computer science, data visualisation, Data visualization, Educational institutions, exploratory data analysis, information visualization innovation, information visualization tool, innovation management, Laboratories, Management training, new technology emergence, Technological innovation, technology transfer, Turning, User interfaces}, isbn = {0-7803-9464-X}, doi = {10.1109/INFVIS.2005.1532153}, author = {Shneiderman, Ben and Rao,R. and Andrews,K. and Ahlberg,C. and Brodbeck,D. and Jewitt,T. and Mackinlay,J.} } @conference {12657, title = {UMD Experiments with FRGC Data}, booktitle = {Computer Vision and Pattern Recognition - Workshops, 2005. CVPR Workshops. IEEE Computer Society Conference on}, year = {2005}, month = {2005/06//}, pages = {172 - 172}, abstract = {Although significant work has been done in the field of face recognition, the performance of state-of-the art face recognition algorithms is not good enough to be effective in operational systems. Though most algorithms work well for controlled images, they are quite susceptible to changes in illumination and pose. Face Recognition Grand Challenge (FRGC) is an effort to examine such issues to suitably guide future research in the area. This paper describes the efforts made at UMD in this direction. We present our results on several experiments suggested in FRGC. We believe that though pattern classification techniques play an extremely significant role in automatic face recognition under controlled conditions, physical modeling is required to generalize across varying situations. Accordingly, we describe a generative approach to recognize faces across varying illumination. Unlike most current methods, our method does not ignore shadows. Instead we use them to our benefit by modeling attached shadows in our formulation.}, doi = {10.1109/CVPR.2005.586}, author = {Aggarwal,G. and Biswas,S. and Chellapa, Rama} } @inbook {18726, title = {Using NMR Spectroscopy to Monitor Ubiquitin Chain Conformation and Interactions with Ubiquitin-Binding Domains}, booktitle = {Ubiquitin and Protein Degradation, Part B}, volume = {Volume 399}, year = {2005}, month = {2005///}, pages = {177 - 192}, publisher = {Academic Press}, organization = {Academic Press}, abstract = {Polyubiquitin (polyUb) chains function as signaling molecules that mediate a diverse set of cellular events. The outcome of polyubiquitination depends on the specific linkage between Ub moieties in the chain, and differently linked chains function as distinct intracellular signals. Although an increasing number of Ub-binding proteins that transmit the regulatory information conferred by (poly)ubiquitination have been identified, the molecular mechanisms of linkage-specific signaling and recognition still remain to be understood. Knowledge of the chain structure is expected to provide insights into the basis of diversity in polyUb signaling. Here we describe several NMR approaches aimed at determining the physiological conformation of polyUb and characterization of the chains{\textquoteright} interactions with ubiquitin-binding proteins.}, isbn = {0076-6879}, url = {http://www.sciencedirect.com/science/article/pii/S0076687905990125}, author = {Varadan,Ranjani and Assfalg,Michael and Fushman, David}, editor = {Deshaies,Raymond J.} } @inbook {18924, title = {Web Service Composition with Volatile Information}, booktitle = {The Semantic Web {\textendash} ISWC 2005}, series = {Lecture Notes in Computer Science}, volume = {3729}, year = {2005}, month = {2005///}, pages = {52 - 66}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {In many Web service composition problems, information may be needed from Web services during the composition process. Existing research on Web service composition (WSC) procedures has generally assumed that this information will not change. We describe two ways to take such WSC procedures and systematically modify them to deal with volatile information. The black-box approach requires no knowledge of the WSC procedure{\textquoteright}s internals: it places a wrapper around the WSC procedure to deal with volatile information. The gray-box approach requires partial information of those internals, in order to insert coding to perform certain bookkeeping operations. We show theoretically that both approaches work correctly. We present experimental results showing that the WSC procedures produced by the gray-box approach can run much faster than the ones produced by the black-box approach.}, keywords = {Computer science}, isbn = {978-3-540-29754-3}, url = {http://www.springerlink.com/content/y105x8464k54l760/abstract/}, author = {Au,Tsz-Chiu and Kuter,Ugur and Nau, Dana S.}, editor = {Gil,Yolanda and Motta,Enrico and Benjamins,V. and Musen,Mark} } @article {12097, title = {Weighted coloring based channel assignment for WLANs}, journal = {ACM SIGMOBILE Mobile Computing and Communications Review}, volume = {9}, year = {2005}, month = {2005///}, pages = {19 - 31}, author = {Mishra,A. and Banerjee,S. and Arbaugh, William A.} } @conference {12706, title = {3D model refinement using surface-parallax}, booktitle = {Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP {\textquoteright}04). IEEE International Conference on}, volume = {3}, year = {2004}, month = {2004/05//}, pages = {iii - 285-8 vol.3 - iii - 285-8 vol.3}, abstract = {We present an approach to update and refine coarse 3D models of urban environments from a sequence of intensity images using surface parallax. This generalizes the plane-parallax recovery methods to surface-parallax using arbitrary surfaces. A coarse and potentially incomplete depth map of the scene obtained from a digital elevation map (DEM) is used as a reference surface which is refined and updated using this approach. The reference depth map is used to estimate the camera motion and the motion of the 3D points on the reference surface is compensated. The resulting parallax, which is an epipolar field, is estimated using an adaptive windowing technique and used to obtain the refined depth map.}, keywords = {3D, adaptive, arbitrary, camera, coarse, compensation;, Computer, DEM;, depth, digital, ELEVATION, environments;, epipolar, estimation;, field;, image, incomplete, INTENSITY, map;, model, MOTION, parallax;, plane-parallax, reconstruction;, recovery;, refinement;, sequence;, sequences;, surface, surfaces;, urban, vision;, windowing;}, doi = {10.1109/ICASSP.2004.1326537}, author = {Agrawala, Ashok K. and Chellapa, Rama} } @conference {12086, title = {Achieving packet-level quality of service through scheduling in multirate WLANs}, booktitle = {Vehicular Technology Conference, 2004. VTC2004-Fall. 2004 IEEE 60th}, volume = {4}, year = {2004}, month = {2004/09/26/29}, pages = {2730- 2734 Vol. 4 - 2730- 2734 Vol. 4}, publisher = {IEEE}, organization = {IEEE}, abstract = {Wireless packet scheduling has been a popular paradigm to achieve packet-level QoS in terms of fairness and throughput in the presence of channel errors. However, the current design does not anticipate the multi-rate capability offered by the IEEE 802.11a/b/g physical layer, thus suffering significant performance degradation in 802.11 WLANs. In this paper, we propose multirate wireless fair scheduling (WMFS). In MWFS, each flow is granted a temporal fair share of the channel, in contrast to the throughput fair share adopted by existing algorithms. Therefore, each flow receives services in proportion to its perceived transmission rate, and high-rate flows are able to opportunistically exploit their good channel conditions and receive more services. MWFS also renovates the compensation model in order to allow for error-prone flows to catch up, thus ensuring fairness for all flows over error-prone channels. We demonstrate the effectiveness of MWFS through both simulations and analysis. Especially, WMFS achieves system throughput 159\% of state-of-the-art scheduling algorithms in simulated scenarios.}, keywords = {Analytical models, channel conditions, channel errors, channel temporal fair share, compensation, Computer science, Delay, error-prone flow compensation, IEEE 802.11a/b/g physical layer, multirate wireless fair scheduling, multirate WLAN, packet radio networks, packet-level quality of service, Physical layer, Processor scheduling, QoS, quality of service, radio access networks, Scheduling algorithm, Throughput, throughput fairness, USA Councils, Wireless LAN, wireless packet scheduling, WMFS}, isbn = {0-7803-8521-7}, doi = {10.1109/VETECF.2004.1400554}, author = {Yuan Yuan and Daqing Gu and Arbaugh, William A. and Jinyun Zhang} } @article {11898, title = {Active deformable models using density estimation}, journal = {International Journal of Image and Graphics}, volume = {4}, year = {2004}, month = {2004///}, pages = {343 - 362}, author = {Abd-Almageed, Wael and Smith,C.E.} } @conference {16003, title = {Active logic for more effective human-computer interaction and other commonsense applications}, booktitle = {Proceedings of the Workshop Empirically Successful First-Order Reasoning, International Joint Conference on Automated Reasoning}, year = {2004}, month = {2004///}, author = {Anderson,M. L and Josyula,D. and Perlis, Don and Purang,K.} } @article {14167, title = {The Argus eye, a new tool for robotics}, journal = {IEEE Robotics and Automation Magazine}, volume = {11}, year = {2004}, month = {2004///}, pages = {31 - 38}, author = {Baker, P. and Ogale, A. S and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {14774, title = {Challenges in Measuring HPCS Learner Productivity in an Age of Ubiquitous Computing: The HPCS Program}, booktitle = {In Proceedings of ICSE Workshop on High Productivity Computing. May 2004}, year = {2004}, month = {2004///}, author = {Asgari, S. and Basili, Victor R. and Carver, J. and Hochstein, L. and Hollingsworth, Jeffrey K and Zelkowitz, Marvin V} } @conference {14289, title = {A comparison of bug finding tools for Java}, booktitle = {Software Reliability Engineering, 2004. ISSRE 2004. 15th International Symposium on}, year = {2004}, month = {2004///}, pages = {245 - 256}, author = {Rutar,N. and Almazan,C.B. and Foster, Jeffrey S.} } @conference {14209, title = {Compound eye sensor for 3D ego motion estimation}, booktitle = {2004 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2004. (IROS 2004). Proceedings}, volume = {4}, year = {2004}, month = {2004/10/28/Sept.}, pages = {3712- 3717 vol.4 - 3712- 3717 vol.4}, publisher = {IEEE}, organization = {IEEE}, abstract = {We describe a compound eye vision sensor for 3D ego motion computation. Inspired by eyes of insects, we show that the compound eye sampling geometry is optimal for 3D camera motion estimation. This optimality allows us to estimate the 3D camera motion in a scene-independent and robust manner by utilizing linear equations. The mathematical model of the new sensor can be implemented in analog networks resulting in a compact computational sensor for instantaneous 3D ego motion measurements in full six degrees of freedom.}, keywords = {3D camera motion estimation, CAMERAS, compound eye vision sensor, Computer vision, Equations, Eyes, Geometry, Image sensors, Insects, linear equations, Motion estimation, robot vision, Robustness, sampling geometry, Sampling methods, Sensor phenomena and characterization}, isbn = {0-7803-8463-6}, doi = {10.1109/IROS.2004.1389992}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J. and Brajovic,V.} } @conference {12069, title = {Copilot - a coprocessor-based kernel runtime integrity monitor}, booktitle = {Proceedings of the 13th conference on USENIX Security Symposium - Volume 13}, series = {SSYM{\textquoteright}04}, year = {2004}, month = {2004///}, pages = {13 - 13}, publisher = {USENIX Association}, organization = {USENIX Association}, address = {San Diego, CA}, abstract = {Copilot is a coprocessor-based kernel integrity monitor for commodity systems. Copilot is designed to detect malicious modifications to a host{\textquoteright}s kernel and has correctly detected the presence of 12 real-world rootkits, each within 30 seconds of their installation with less than a 1\% penalty to the host{\textquoteright}s performance. Copilot requires no modifications to the protected host{\textquoteright}s software and can be expected to operate correctly even when the host kernel is thoroughly compromised - an advantage over traditional monitors designed to run on the host itself.}, keywords = {design, management, MONITORS, Security, security and protection}, url = {http://portal.acm.org/citation.cfm?id=1251375.1251388}, author = {Petroni,Jr. and Fraser,Timothy and Molina,Jesus and Arbaugh, William A.} } @conference {13757, title = {Divergence Unraveling for Word Alignment...}, booktitle = {Natural Language Engineering}, year = {2004}, month = {2004///}, abstract = {We describe the use of parallel text for divergence unraveling in word-level alignment. DUSTer (Divergence Unraveling for Statistical Translation) is a system that combines linguistic and statistical knowledge to resolve structural di$\#$erences between languages, i.e., translation divergences, during the process of alignment. Our immediate goal is to induce word-level alignments that are more accurate than those produced by an existing state-ofthe -art statistical system. The long-term goal is to improve the output quality of statistical machine translation and lexical acquisition systems by using DUSTer as one possible input to a framework that accommodates multiple alignments. We show that a systematic characterization of alignment errors made by a statistical system validates the use of linguistically-motivated universal rules for identifying and handling divergences. These rules relate one or more linguistically-motivated categories associated with the (English) input words to those of another language (foreign language); the resulting match sets are used to infer corrected alignments. Using a human-aligned corpus as our gold standard, we demonstrate an improvement in alignments over an existing state-of-the-art alignment algorithm.}, author = {Dorr, Bonnie J and Ayan,N.F. and Habash,N.} } @conference {15989, title = {Domain-independent reason-enhanced controller for task-oriented systems-DIRECTOR}, booktitle = {PROCEEDINGS OF THE NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE}, year = {2004}, month = {2004///}, pages = {1014 - 1015}, author = {Josyula,D. P and Anderson,M. L and Perlis, Don} } @article {17885, title = {Efficient Execution of Multi-Query Data Analysis Batches Using Compiler Optimization Strategies}, journal = {Languages and Compilers for Parallel Computing}, year = {2004}, month = {2004///}, abstract = {This work investigates the leverage that can be obtained from compiler optimization techniques for efficient execution of multi-query workloads in data analysis applications. Our approach is to address multi-query optimization at the algorithmic level, by transforming a declarative specification of scientific data analysis queries into a high-level imperative program that can be made more efficient by applying compiler optimization techniques. These techniques {\textendash} including loop fusion, common subexpression elimination and dead code elimination {\textendash} are employed to allow data and computation reuse across queries. We describe a preliminary experimental analysis on a real remote sensing application that analyzes very large quantities of satellite data. The results show our techniques achieve sizable reductions in the amount of computation and I/O necessary for executing query batches and in average execution times for the individual queries in a given batch.}, doi = {10.1007/978-3-540-24644-2_33}, author = {Andrade,H. and Aryangat,S. and Kurc, T. and Saltz, J. and Sussman, Alan} } @conference {12131, title = {Empirical-based estimation of the effect on software dependability of a technique for architecture conformance verification}, booktitle = {Proceedings of the ICSE 2004 Workshop on Architecting Dependable Systems, Edinburgh, UK}, year = {2004}, month = {2004///}, abstract = {The High Dependability Computing Program (HDCP)project is a NASA initiative for increasing dependability of software-based systems. It researches achieving high dependability by introducing new technologies. We focus on the evaluation of the effectiveness of technologies with respect to dependability. We employ empirical evaluation methods along with evaluation testbeds. In this paper, our technology evaluation approach is described. A testbed representative of air traffic control applications is used. An experiment to evaluate a technology to for identification of architectural violations is presented. }, author = {Asgari, S. and Basili, Victor R. and Costa,P. and Donzelli,P. and Hochstein, L. and Lindvall,M. and Rus,I. and Shull, F. and Tvedt,R. and Zelkowitz, Marvin V} } @conference {18673, title = {Ferret: a host vulnerability checking tool}, year = {2004}, month = {2004/03//}, pages = {389 - 394}, abstract = {Evaluation of computing system security requires knowledge of the vulnerabilities present in the system and of potential attacks against the system. Vulnerabilities can be classified based on their location as application vulnerabilities, network vulnerabilities, or host vulnerabilities. We describe Ferret, a new software tool for checking host vulnerabilities. Ferret helps system administrators by quickly finding vulnerabilities that are present on a host. It is designed and implemented in a modular way: a different plug-in module is used for each vulnerability checked, and each possible output format is specified by a plug-in module. As a result, Ferret is extensible, and can easily be kept up-to-date through addition of checks for new vulnerabilities as they are discovered; the modular approach also makes it easy to provide specific configurations of Ferret tailored to specific operating systems or use environments. Ferret is a freely available open-source software implemented in Perl.}, keywords = {Ferret software tool, host vulnerability checking tool, open-source software, Perl, plug-in module, program verification, security auditing tool, security evaluation, security of data, software tools}, doi = {10.1109/PRDC.2004.1276595}, author = {Sharma,Anil and Martin,J.R. and Anand,N. and Michel Cukier and Sanders,W. H.} } @article {14184, title = {A hierarchy of cameras for 3D photography}, journal = {Computer Vision and Image Understanding}, volume = {96}, year = {2004}, month = {2004/12//}, pages = {274 - 293}, abstract = {The view-independent visualization of 3D scenes is most often based on rendering accurate 3D models or utilizes image-based rendering techniques. To compute the 3D structure of a scene from a moving vision sensor or to use image-based rendering approaches, we need to be able to estimate the motion of the sensor from the recorded image information with high accuracy, a problem that has been well-studied. In this work, we investigate the relationship between camera design and our ability to perform accurate 3D photography, by examining the influence of camera design on the estimation of the motion and structure of a scene from video data. By relating the differential structure of the time varying plenoptic function to different known and new camera designs, we can establish a hierarchy of cameras based upon the stability and complexity of the computations necessary to estimate structure and motion. At the low end of this hierarchy is the standard planar pinhole camera for which the structure from motion problem is non-linear and ill-posed. At the high end is a camera, which we call the full field of view polydioptric camera, for which the motion estimation problem can be solved independently of the depth of the scene which leads to fast and robust algorithms for 3D Photography. In between are multiple view cameras with a large field of view which we have built, as well as omni-directional sensors.}, keywords = {Camera design, Multi-view geometry, Polydioptric cameras, Spatio-temporal image analysis, structure from motion}, isbn = {1077-3142}, doi = {10.1016/j.cviu.2004.03.013}, url = {http://www.sciencedirect.com/science/article/pii/S1077314204000505}, author = {Neumann, Jan and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {12100, title = {High-performance MAC for high-capacity wireless LANs}, booktitle = {13th International Conference on Computer Communications and Networks, 2004. ICCCN 2004. Proceedings}, year = {2004}, month = {2004/10/11/13}, pages = {167 - 172}, publisher = {IEEE}, organization = {IEEE}, abstract = {The next-generation wireless technologies, e.g., 802.11n and 802.15.3a, offer a physical-layer speed at least an-order-of-magnitude higher than the current standards. However, direct application of current MACs leads to high protocol overhead and significant throughput degradation. In this paper, we propose ADCA, a high-performance MAC that works with high-capacity physical layer. ADCA exploits two ideas of adaptive batch transmission and opportunistic selection of high-rate hosts to simultaneously reduce the overhead and improve the aggregate throughput. It opportunistically favors high-rate hosts by providing higher access probability and more access time, while ensuring each low-rate host certain minimum amount of channel access time. Simulations show that the ADCA design increases the throughput by 112\% and reduces the average delay by 55\% compared with the legacy DCF. It delivers more than 100 Mbps MAC-layer throughput as compared with 35 Mbps offered by the legacy MAC}, keywords = {35 Mbit/s, access protocols, Aggregates, Bandwidth, batch transmission, Computer science, Educational institutions, high-capacity wireless LAN, high-performance MAC, Laboratories, Local area networks, Media Access Protocol, opportunistic selection, Physical layer, probability, Throughput, Wireless LAN}, isbn = {0-7803-8814-3}, doi = {10.1109/ICCCN.2004.1401615}, author = {Yuan Yuan and Daqing Gu and Arbaugh, William A. and Jinyun Zhang} } @conference {12034, title = {The influence of shape on image correspondence}, booktitle = {2nd International Symposium on 3D Data Processing, Visualization and Transmission, 2004. 3DPVT 2004. Proceedings}, year = {2004}, month = {2004/09/06/9}, pages = {945 - 952}, publisher = {IEEE}, organization = {IEEE}, abstract = {We examine the implications of shape on the process of finding dense correspondence and half-occlusions for a stereo pair of images. The desired property of the depth map is that it should be a piecewise continuous function which is consistent with the images and which has the minimum number of discontinuities. To zeroeth order, piecewise continuity becomes piecewise constancy. Using this approximation, we first discuss an approach for dealing with such a fronto-parallel shapeless world, and the problems involved therein. We then introduce horizontal and vertical slant to create a first order approximation to piecewise continuity. We highlight the fact that a horizontally slanted surface (ie. having depth variation in the direction of the separation of the two cameras) appears horizontally stretched in one image as compared to the other image. Thus, while corresponding two images, N pixels on a scanline in one image may correspond to a different number of pixels M in the other image, which has consequences with regard to sampling and occlusion detection. We also discuss the asymmetry between vertical and horizontal slant, and the central role of nonhorizontal edges in the context of vertical slant. Using experiments, we discuss cases where existing algorithms fail, and how the incorporation of new constraints provides correct results.}, keywords = {Automation, CAMERAS, Computational modeling, first order approximation, Geometrical optics, hidden feature removal, image sampling, Image segmentation, Layout, occlusion detection, piecewise continuous function, Pixel, SHAPE, Simulated annealing, stereo image processing, surface fitting}, isbn = {0-7695-2223-8}, doi = {10.1109/TDPVT.2004.1335418}, author = {Ogale, A. S and Aloimonos, J.} } @article {17624, title = {Method of designing signaling networks for internet telephony}, volume = {:~09/517,658}, year = {2004}, month = {2004/02/03/}, abstract = {A method is disclosed for designing a signaling network of call coordinators (CCs) for internet telephony. The new method can be used to design a CC network of arbitrary size that satisfies, with high probability, limitations on the maximum number of sockets per CC and on the maximum number of hops between an arbitrary pair of switches in the network. According to the disclosed method, the network of CCs is treated initially as a collection of isolated points, one point for each CC. Then, links are added between pairs of CCs, excluding pairs lying within the same switch. The links are added randomly, but with a particular probability p, which may be different for different pairs. Thus, whether a given link is added depends, in effect, upon the outcome of a loaded coin toss in which the probability of a positive outcome is p.}, url = {http://www.google.com/patents?id=VTsSAAAAEBAJ}, author = {Aravamudan,Murali and Kumaran,Krishnan and Ramakrishnan,Kajamalai Gopalaswamy and Srinivasan, Aravind}, editor = {Lucent Technologies Inc.} } @inbook {13830, title = {Multi-align: Combining Linguistic and Statistical Techniques to Improve Alignments for Adaptable MT}, booktitle = {Machine Translation: From Real Users to ResearchMachine Translation: From Real Users to Research}, series = {Lecture Notes in Computer Science}, volume = {3265}, year = {2004}, month = {2004///}, pages = {17 - 26}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {An adaptable statistical or hybrid MT system relies heavily on the quality of word-level alignments of real-world data. Statistical alignment approaches provide a reasonable initial estimate for word alignment. However, they cannot handle certain types of linguistic phenomena such as long-distance dependencies and structural differences between languages. We address this issue in Multi-Align, a new framework for incremental testing of different alignment algorithms and their combinations. Our design allows users to tune their systems to the properties of a particular genre/domain while still benefiting from general linguistic knowledge associated with a language pair. We demonstrate that a combination of statistical and linguistically-informed alignments can resolve translation divergences during the alignment process.}, isbn = {978-3-540-23300-8}, url = {http://dx.doi.org/10.1007/978-3-540-30194-3_3}, author = {Ayan,Necip and Dorr, Bonnie J and Habash,Nizar}, editor = {Frederking,Robert and Taylor,Kathryn} } @inbook {18915, title = {Plan Databases: Model and Algebra}, booktitle = {Foundations of Information and Knowledge Systems}, series = {Lecture Notes in Computer Science}, volume = {2942}, year = {2004}, month = {2004///}, pages = {302 - 319}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Despite the fact that thousands of applications manipulate plans, there has been no work to date on managing large databases of plans. In this paper, we first propose a formal model of plan databases. We describe important notions of consistency and coherence for such databases. We then propose a set of operators similar to the relational algebra to query such databases of plans.}, keywords = {Computer science}, isbn = {978-3-540-20965-2}, url = {http://www.springerlink.com/content/yqhlqtu4te18q2e1/abstract/}, author = {Yaman,Fusun and Adali,Sibel and Nau, Dana S. and Sapino,Maria and Subrahmanian,V.}, editor = {Seipel,Dietmar and Turull-Torres,Jos{\'e}} } @conference {17802, title = {The priority curve algorithm for video summarization}, booktitle = {Proceedings of the 2nd ACM international workshop on Multimedia databases}, series = {MMDB {\textquoteright}04}, year = {2004}, month = {2004///}, pages = {28 - 35}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {In this paper, we introduce the concept of a priority curve associated with a video. We then provide an algorithm that can use the priority curve to create a summary (of a desired length) of any video. The summary thus created exhibits nice continuity properties and also avoids repetition. We have implemented the priority curve algorithm (PCA) and compared it with other summarization algorithms in the literature. We show that PCA is faster than existing algorithms and also produces better quality summaries. The quality of summaries was evaluated by a group of 200 students in Naples, Italy, who watched soccer videos. We also briefly describe a soccer video summarization system we have built on using the PCA architecture and various (classical) image processing algorithms.}, keywords = {probabilistic, Summarization, system, video}, isbn = {1-58113-975-6}, doi = {10.1145/1032604.1032611}, url = {http://doi.acm.org/10.1145/1032604.1032611}, author = {Fayzullin,M. and V.S. Subrahmanian and Albanese, M. and Picariello, A.} } @article {12080, title = {Proactive key distribution using neighbor graphs}, journal = {IEEE Wireless Communications}, volume = {11}, year = {2004}, month = {2004/02//}, pages = {26 - 36}, abstract = {User mobility in wireless data networks is increasing because of technological advances, and the desire for voice and multimedia applications. These applications, however, require that handoffs between base stations (or access points) be fast to maintain the quality of the connections. In this article we introduce a novel data structure, the neighbor graph, that dynamically captures the mobility topology of a wireless network. We show how neighbor graphs can be utilized to obtain a 99 percent reduction in the authentication time of an IEEE 802.11 handoff (full EAP-TLS) by proactively distributing necessary key material one hop ahead of the mobile user. We also present a reactive method for fast authentication that requires only firmware changes to access points and hence can easily be deployed on existing wireless networks.}, keywords = {access points, Authentication, authentication time, Base stations, Communication system security, Delay, graph theory, GSM, IEEE 802.11 handoff, Land mobile radio cellular systems, Message authentication, mobile radio, Multiaccess communication, neighbor graph, Network topology, Roaming, telecommunication security, Telephone sets, user mobility, Wi-Fi networks, wireless data networks, Wireless LAN, Wireless networks}, isbn = {1536-1284}, doi = {10.1109/MWC.2004.1269714}, author = {Mishra,A. and Min Ho Shin and Petroni,N. L. and Clancy,T. C and Arbaugh, William A.} } @article {14253, title = {Robotics \& Automation Magazine Vol. 11}, journal = {IEEE Robotics \& Automation Magazine}, year = {2004}, month = {2004///}, pages = {132 - 132}, author = {Aliaga, D.G. and Allen,PK and Archibald,JK and Argyros,AA and Arkin,RC and Baker,C. and Baker, P. and Beard,RW and Bicchi,A. and Birgmajer,B. and others} } @conference {12687, title = {Robust ego-motion estimation and 3D model refinement using depth based parallax model}, booktitle = {Image Processing, 2004. ICIP {\textquoteright}04. 2004 International Conference on}, volume = {4}, year = {2004}, month = {2004/10//}, pages = {2483 - 2486 Vol. 4 - 2483 - 2486 Vol. 4}, abstract = {We present an iterative algorithm for robustly estimating the ego-motion and refining and updating a coarse, noisy and partial depth map using a depth based parallax model and brightness derivatives extracted from an image pair. Given a coarse, noisy and partial depth map acquired by a range-finder or obtained from a Digital Elevation Map (DFM), we first estimate the ego-motion by combining a global ego-motion constraint and a local brightness constancy constraint. Using the estimated camera motion and the available depth map estimate, motion of the 3D points is compensated. We utilize the fact that the resulting surface parallax field is an epipolar field and knowing its direction from the previous motion estimates, estimate its magnitude and use it to refine the depth map estimate. Instead of assuming a smooth parallax field or locally smooth depth models, we locally model the parallax magnitude using the depth map, formulate the problem as a generalized eigen-value analysis and obtain better results. In addition, confidence measures for depth estimates are provided which can be used to remove regions with potentially incorrect (and outliers in) depth estimates for robustly estimating ego-motion in the next iteration. Results on both synthetic and real examples are presented.}, keywords = {3D, algorithm;, analysis;, and, based, camera;, coarse, compensation;, DEM;, depth, digital, ego-motion, eigen-value, eigenfunctions;, eigenvalues, ELEVATION, epipolar, estimation;, extraction;, feature, field;, iteration, iterative, map;, method;, methods;, model, model;, MOTION, parallax, partial, range-finding;, refinement;, refining;, surface}, doi = {10.1109/ICIP.2004.1421606}, author = {Agrawala, Ashok K. and Chellapa, Rama} } @article {12075, title = {Security issues in IEEE 802.11 wireless local area networks: a survey}, journal = {Wireless Communications and Mobile Computing}, volume = {4}, year = {2004}, month = {2004///}, pages = {821 - 833}, author = {Mishra,A. and Petroni Jr,N. L and Arbaugh, William A. and Fraser,T.} } @article {18713, title = {Solution Conformation of Lys63-linked Di-ubiquitin Chain Provides Clues to Functional Diversity of Polyubiquitin Signaling}, journal = {Journal of Biological ChemistryJ. Biol. Chem.}, volume = {279}, year = {2004}, month = {2004/02/20/}, pages = {7055 - 7063}, abstract = {Diverse cellular events are regulated by post-translational modification of substrate proteins via covalent attachment of one or a chain of ubiquitin molecules. The outcome of (poly)ubiquitination depends upon the specific lysine residues involved in the formation of polyubiquitin chains. Lys48-linked chains act as a universal signal for proteasomal degradation, whereas Lys63-linked chains act as a specific signal in several non-degradative processes. Although it has been anticipated that functional diversity between alternatively linked polyubiquitin chains relies on linkage-dependent differences in chain conformation/topology, direct structural evidence in support of this model has been lacking. Here we use NMR methods to determine the structure of a Lys63-linked di-ubiquitin chain. The structure is characterized by an extended conformation, with no direct contact between the hydrophobic residues Leu8, Ile44, and Val70 on the ubiquitin units. This structure contrasts with the closed conformation observed for Lys48-linked di-ubiquitin wherein these residues form the interdomain interface (Cook, W. J., Jeffrey, L. C., Carson, M., Zhijian, C., and Pickart, C. M. (1992) J. Biol. Chem. 267, 16467-16471; Varadan, R., Walker, O., Pickart, C., and Fushman, D. (2002) J. Mol. Biol. 324, 637-647). Consistent with the open conformation of the Lys63-linked di-ubiquitin, our binding studies show that both ubiquitin domains in this chain can bind a ubiquitin-associated domain from HHR23A independently and in a mode similar to that for mono-ubiquitin. In contrast, Lys48-linked di-ubiquitin binds in a different, higher affinity mode that has yet to be determined. This is the first experimental evidence that alternatively linked polyubiquitin chains adopt distinct conformations.}, isbn = {0021-9258, 1083-351X}, doi = {10.1074/jbc.M309184200}, url = {http://www.jbc.org/content/279/8/7055}, author = {Varadan,Ranjani and Assfalg,Michael and Haririnia,Aydin and Raasi,Shahri and Pickart,Cecile and Fushman, David} } @conference {11951, title = {Stereo Correspondence with Slanted Surfaces: Critical Implications of Horizontal Slant}, booktitle = {Computer Vision and Pattern Recognition, IEEE Computer Society Conference on}, volume = {1}, year = {2004}, month = {2004///}, pages = {568 - 573}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {We examine the stereo correspondence problem in the presence of slanted scene surfaces. In particular, we highlight a previously overlooked geometric fact: a horizontally slanted surface (i.e. having depth variation in the direction of the separation of the two cameras) will appear horizontally stretched in one image as compared to the other image. Thus, while corresponding two images, N pixels on a scanline in one image may correspond to a different number of pixels M in the other image. This leads to three important modifications to existing stereo algorithms: (a) due to unequal sampling, intensity matching metrics such as the popular Birchfield-Tomasi procedure must be modified, (b) unequal numbers of pixels in the two images must be allowed to correspond to each other, and (c) the uniqueness constraint, which is often used for detecting occlusions, must be changed to a 3D uniqueness constraint. This paper discusses these new constraints and provides a simple scanline based matching algorithm for illustration. We experimentally demonstrate test cases where existing algorithms fail, and how the incorporation of these new constraints provides correct results. Experimental comparisons of the scanline based algorithm with standard data sets are also provided.}, doi = {http://doi.ieeecomputersociety.org/10.1109/CVPR.2004.241}, author = {Ogale, Abhijit S. and Aloimonos, J.} } @article {12068, title = {Structure from motion of parallel lines}, journal = {Computer Vision-ECCV 2004}, year = {2004}, month = {2004///}, pages = {229 - 240}, author = {Baker, P. and Aloimonos, J.} } @conference {14761, title = {Studying code development for high performance computing: the HPCS program}, booktitle = {First International Workshop On Software Engineering for High Performance Computing System Applications}, year = {2004}, month = {2004///}, pages = {32 - 32}, author = {Carver, J. and Asgari, S. and Basili, Victor R. and Hochstein, L. and Hollingsworth, Jeffrey K and Shull, F. and Zelkowitz, Marvin V} } @conference {12694, title = {A system identification approach for video-based face recognition}, booktitle = {Pattern Recognition, 2004. ICPR 2004. Proceedings of the 17th International Conference on}, volume = {4}, year = {2004}, month = {2004/08//}, pages = {175 - 178 Vol.4 - 175 - 178 Vol.4}, abstract = {The paper poses video-to-video face recognition as a dynamical system identification and classification problem. We model a moving face as a linear dynamical system whose appearance changes with pose. An autoregressive and moving average (ARMA) model is used to represent such a system. The choice of ARMA model is based on its ability to take care of the change in appearance while modeling the dynamics of pose, expression etc. Recognition is performed using the concept of sub space angles to compute distances between probe and gallery video sequences. The results obtained are very promising given the extent of pose, expression and illumination variation in the video data used for experiments.}, keywords = {and, autoregressive, average, dynamical, Face, gallery, identification;, image, linear, model;, moving, processes;, processing;, recognition;, sequences;, signal, system, system;, video, video-based}, doi = {10.1109/ICPR.2004.1333732}, author = {Aggarwal,G. and Chowdhury, A.K.R. and Chellapa, Rama} } @conference {17891, title = {Time and space optimization for processing groups of multi-dimensional scientific queries}, booktitle = {Proceedings of the 18th annual international conference on Supercomputing}, series = {ICS {\textquoteright}04}, year = {2004}, month = {2004///}, pages = {95 - 105}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Data analysis applications in areas as diverse as remote sensing and telepathology require operating on and processing very large datasets. For such applications to execute efficiently, careful attention must be paid to the storage, retrieval, and manipulation of the datasets. This paper addresses the optimizations performed by a high performance database system that processes groups of data analysis requests for these applications, which we call queries. The system performs end-to-end processing of the requests, formulated as PostgreSQL declarative queries. The queries are converted into imperative descriptions, multiple imperative descriptions are merged into a single execution plan, the plan is optimized to decrease execution time via common compiler optimization techniques, and, finally, the plan is optimized to decrease memory consumption. The last two steps are experimentally shown to effectively reduc the amount of time required while conserving memory space as a group of queries is processed by the database.}, keywords = {multi-query optimization, Scientific databases}, isbn = {1-58113-839-3}, doi = {10.1145/1006209.1006224}, url = {http://doi.acm.org/10.1145/1006209.1006224}, author = {Aryangat,Suresh and Andrade,Henrique and Sussman, Alan} } @conference {17807, title = {Utilizing volatile external information during planning}, booktitle = {Proceedings of the European Conference on Artificial Intelligence (ECAI)}, year = {2004}, month = {2004///}, pages = {647 - 651}, author = {Au,T. C and Nau, Dana S. and V.S. Subrahmanian} } @article {12936, title = {Viable but Nonculturable Vibrio Cholerae O1 in the Aquatic Environment of Argentina}, journal = {Applied and Environmental MicrobiologyAppl. Environ. Microbiol.}, volume = {70}, year = {2004}, month = {2004/12/01/}, pages = {7481 - 7486}, abstract = {In Argentina, as in other countries of Latin America, cholera has occurred in an epidemic pattern. Vibrio cholerae O1 is native to the aquatic environment, and it occurs in both culturable and viable but nonculturable (VNC) forms, the latter during interepidemic periods. This is the first report of the presence of VNC V. cholerae O1 in the estuarine and marine waters of the R{\'\i}o de la Plata and the Argentine shelf of the Atlantic Ocean, respectively. Employing immunofluorescence and PCR methods, we were able to detect reservoirs of V. cholerae O1 carrying the virulence-associated genes ctxA and tcpA. The VNC forms of V. cholerae O1 were identified in samples of water, phytoplankton, and zooplankton; the latter organisms were mainly the copepods Acartia tonsa, Diaptomus sp., Paracalanus crassirostris, and Paracalanus parvus. We found that under favorable conditions, the VNC form of V. cholerae can revert to the pathogenic, transmissible state. We concluded that V. cholerae O1 is a resident of Argentinean waters, as has been shown to be the case in other geographic regions of the world.}, isbn = {0099-2240, 1098-5336}, doi = {10.1128/AEM.70.12.7481-7486.2004}, url = {http://aem.asm.org/content/70/12/7481}, author = {Binsztein,Norma and Costagliola,Marcela C. and Pichel,Mariana and Jurquiza,Ver{\'o}nica and Ram{\'\i}rez,Fernando C. and Akselman,Rut and Vacchino,Marta and Huq,Anwarul and Rita R Colwell} } @conference {13217, title = {Window-based, discontinuity preserving stereo}, booktitle = {Computer Vision and Pattern Recognition, 2004. CVPR 2004. Proceedings of the 2004 IEEE Computer Society Conference on}, volume = {1}, year = {2004}, month = {2004/07/02/june}, pages = {I-66 - I-73 Vol.1 - I-66 - I-73 Vol.1}, abstract = {Traditionally, the problem of stereo matching has been addressed either by a local window-based approach or a dense pixel-based approach using global optimization. In this paper we present an algorithm which combines window-based local matching into a global optimization framework. Our local matching algorithm assumes that local windows can have at most two disparities. Under this assumption, the local matching can be performed very efficiently using graph cuts. The global matching is formulated as minimization of an energy term that takes into account the matching constraints induced by the local stereo algorithm. Fast, approximate minimization of this energy is achieved through graph cuts. The key feature of our algorithm is that it preserves discontinuities both during the local as well as global matching phase.}, keywords = {algorithm;, approach;, based, cuts;, dense, discontinuity, global, graph, image, local, MATCHING, matching;, minimisation;, optimization;, Pixel, preserving, processing;, stereo, theory;, window}, doi = {10.1109/CVPR.2004.1315015}, author = {Agrawal,M. and Davis, Larry S.} } @article {14066, title = {2003 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 19-22, 2003, New Paltz. NY}, journal = {2003 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics: October 19-22, 2003, Mohonk Mountain House, New Paltz, NY, USA}, year = {2003}, month = {2003///}, pages = {238 - 238}, author = {Adams,N. and EssI,G. and Makino,S. and Martens,A.A.C.F.W.L. and Algazi,R. and Fellers,M. and Martin,R. and Avendano,C. and Fielder,L. and Mellody,M. and others} } @article {12955, title = {A 4-Year Study of the Epidemiology of Vibrio Cholerae in Four Rural Areas of Bangladesh}, journal = {Journal of Infectious DiseasesJ Infect Dis.}, volume = {187}, year = {2003}, month = {2003/01/01/}, pages = {96 - 101}, abstract = {How Vibrio cholerae spreads around the world and what determines its seasonal peaks in endemic areas are not known. These features of cholera have been hypothesized to be primarily the result of environmental factors associated with aquatic habitats that can now be identified. Since 1997, fortnightly surveillance in 4 widely separated geographic locations in Bangladesh has been performed to identify patients with cholera and to collect environmental data. A total of 5670 patients (53\% <5 years of age) have been studied; 14.3\% had cholera (10.4\% due to V. cholerae O1 El Tor, 3.8\% due to O139). Both serogroups were found in all locations; outbreaks were seasonal and often occurred simultaneously. Water-use patterns showed that bathing and washing clothes in tube-well water was significantly protective in two of the sites. These data will be correlated with environmental factors, to develop a model for prediction of cholera outbreaks}, isbn = {0022-1899, 1537-6613}, doi = {10.1086/345865}, url = {http://jid.oxfordjournals.org/content/187/1/96}, author = {Sack,R. Bradley and Siddique,A. Kasem and Longini,Ira M. and Nizam,Azhar and Yunus,Md and M. Sirajul Islam and Morris,Jr and Ali,Afsar and Huq,Anwar and Nair,G. Balakrish and Qadri,Firdausi and Faruque,Shah M. and Sack,David A. and Rita R Colwell} } @article {12091, title = {An Analysis of the Layer 2 Handoff costs in Wireless Local Area Networks}, journal = {ACM Computer Communications Review}, volume = {33}, year = {2003}, month = {2003///}, author = {Mishra,A. and Shin,M. H. and Arbaugh, William A.} } @inbook {17558, title = {Approximation Algorithms for Channel Allocation Problems in Broadcast Networks}, booktitle = {Approximation, Randomization, and Combinatorial Optimization: Algorithms and Techniques}, series = {Lecture Notes in Computer Science}, volume = {2764}, year = {2003}, month = {2003///}, pages = {821 - 826}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We study two packing problems that arise in the area of dissemination-based information systems; a second theme is the study of distributed approximation algorithms. The problems considered have the property that the space occupied by a collection of objects together could be significantly less than the sum of the sizes of the individual objects. In the Channel Allocation Problem , there are users who request subsets of items. There are a fixed number of channels that can carry an arbitrary amount of information. Each user must get all of the requested items from one channel, i.e., all the data items of each request must be broadcast on some channel. The load on any channel is the number of items that are broadcast on that channel; the objective is to minimize the maximum load on any channel. We present approximation algorithms for this problem and also show that the problem is MAX-SNP hard. The second problem is the Edge Partitioning Problem addressed by Goldschmidt, Hochbaum, Levin, and Olinick ( Networks, 41:13-23, 2003 ). Each channel here can deliver information to at most k users, and we aim to minimize the total load on all channels. We present an O ( n 1/3 ){\textendash}approximation algorithm and also show that the algorithm can be made fully distributed with the same approximation guarantee; we also generalize to the case of hypergraphs.}, isbn = {978-3-540-40770-6}, url = {http://dx.doi.org/10.1007/978-3-540-45198-3_5}, author = {Gandhi,Rajiv and Khuller, Samir and Srinivasan, Aravind and Wang,Nan}, editor = {Arora,Sanjeev and Jansen,Klaus and Rolim,Jos{\'e} and Sahai,Amit} } @conference {12076, title = {Bootstrapping security associations for routing in mobile ad-hoc networks}, booktitle = {IEEE Global Telecommunications Conference, 2003. GLOBECOM {\textquoteright}03}, volume = {3}, year = {2003}, month = {2003/12/01/5}, pages = {1511- 1515 vol.3 - 1511- 1515 vol.3}, publisher = {IEEE}, organization = {IEEE}, abstract = {To date, most solutions proposed for secure routing in mobile ad-hoc networks (MANETs), assume that secure associations between pairs of nodes can be established on-line; e.g., by a trusted third party, by distributed trust establishment. However, establishing such security associations, with or without trusted third parties, requires reliance on routing layer security. In this paper, we eliminate this apparent cyclic dependency between security services and secure routing in MANETs and show how to bootstrap security for the routing layer. We use the notion of statistically unique and cryptographically verifiable (SUCV) identifiers to implement a secure binding between IP addresses and keys that is independent of any trusted security service. We illustrate our solution with the dynamic source routing (DSR) protocol and compare it with other solutions for secure routing.}, keywords = {ad hoc networks, bootstrapping security association, Cryptographic protocols, dynamic source routing protocol, Fabrics, Intelligent networks, IP address, IP key, IP networks, Message authentication, mobile ad-hoc network, mobile radio, Protection, Public key, public key cryptography, routing layer security reliability, routing protocols, secure routing, Security, security service, statistically unique cryptographically verification, telecommunication security}, isbn = {0-7803-7974-8}, doi = {10.1109/GLOCOM.2003.1258490}, author = {Bobba,R. B and Eschenauer,L. and Gligor,V. and Arbaugh, William A.} } @conference {13238, title = {Camera calibration using spheres: a semi-definite programming approach}, booktitle = {Computer Vision, 2003. Proceedings. Ninth IEEE International Conference on}, year = {2003}, month = {2003/10//}, pages = {782 -789 vol.2 - 782 -789 vol.2}, abstract = {Vision algorithms utilizing camera networks with a common field of view are becoming increasingly feasible and important. Calibration of such camera networks is a challenging and cumbersome task. The current approaches for calibration using planes or a known 3D target may not be feasible as these objects may not be simultaneously visible in all the cameras. In this paper, we present a new algorithm to calibrate cameras using occluding contours of spheres. In general, an occluding contour of a sphere projects to an ellipse in the image. Our algorithm uses the projection of the occluding contours of three spheres and solves for the intrinsic parameters and the locations of the spheres. The problem is formulated in the dual space and the parameters are solved for optimally and efficiently using semidefinite programming. The technique is flexible, accurate and easy to use. In addition, since the contour of a sphere is simultaneously visible in all the cameras, our approach can greatly simplify calibration of multiple cameras with a common field of view. Experimental results from computer simulated data and real world data, both for a single camera and multiple cameras, are presented.}, keywords = {3D, algorithms;calibration;cameras;computer, approach;sphere, calibration;camera, contours;semidefinite, extraction;, field;ellipse;intrinsic, location;spheres;vision, networks;common, parameters;occluding, Programming, target;camera, view, vision;feature}, doi = {10.1109/ICCV.2003.1238428}, author = {Agrawal,M. and Davis, Larry S.} } @conference {14290, title = {Checking and inferring local non-aliasing}, booktitle = {ACM SIGPLAN Notices}, volume = {38}, year = {2003}, month = {2003///}, pages = {129 - 140}, author = {Aiken,A. and Foster, Jeffrey S. and Kodumal,J. and Terauchi,T.} } @article {13255, title = {Complete camera calibration using spheres: Dual space approach}, journal = {IEEE}, volume = {206}, year = {2003}, month = {2003///}, pages = {782 - 789}, abstract = {Camera calibration is a fundamental problem in computer vision and photogrammetry. We present a new algorithmto calibrate cameras using spheres. In general, an occluding contour of a sphere projects to an ellipse in the image. Our algorithm uses the projection of the occluding contours of three spheres and solves for the intrinsic parameters and the locations of the spheres. The problem is formulated in the dual space and the parameters are solved for optimally and efficiently using semi-definite programming. The technique is flexible, accurate and easy to use. In addition, it can be used to simultaneously calibrate multiple cameras with a common field of view. Experimen- tal results from computer simulated data and real world data, both for a single camera and multiple cameras, are presented. }, author = {Agrawal,M. and Davis, Larry S.} } @article {11956, title = {Computational video}, journal = {The Visual Computer}, volume = {19}, year = {2003}, month = {2003///}, pages = {355 - 359}, author = {Aloimonos, J.} } @article {12088, title = {The dangers of mitigating security design flaws: a wireless case study}, journal = {IEEE Security \& Privacy}, volume = {1}, year = {2003}, month = {2003/02//Jan}, pages = {28 - 36}, abstract = {Mitigating design flaws often provides the only means to protect legacy equipment, particularly in wireless local area networks. A synchronous active attack against the wired equivalent privacy protocol demonstrates how mitigating one flaw or attack can facilitate another.}, keywords = {Communication system security, computer security, cryptography, design flaw mitigation, Dictionaries, legacy equipment, privacy, Protection, Protocols, security design flaws, security of data, synchronous active attack, telecommunication security, Telecommunication traffic, wired equivalent privacy protocol, Wireless LAN, wireless local area networks, Wireless networks}, isbn = {1540-7993}, doi = {10.1109/MSECP.2003.1176993}, author = {Petroni,N. L. and Arbaugh, William A.} } @article {13758, title = {Domain Tuning of Bilingual Lexicons for MT}, year = {2003}, month = {2003/02//}, institution = {Instititue for Advanced Computer Studies, Univ of Maryland, College Park}, abstract = {Our overall objective is to translate a domain-specific document in a foreign language (in this case, Chinese) to English. Using automatically induced domain-specific, comparable documents and language-independent clustering, we apply domain-tuning techniques to a bilingual lexicon for downstream translation of the input document to English. We will describe our domain-tuning technique and demonstrate its effectiveness by comparing our results to manually constructed domain-specific vocabulary. Our coverage/accuracy experiments indicate that domain-tuned lexicons achieve 88/\% precision and 66/\% recall. We also ran a Bleu experiment to compare our domain-tuned version to its un-tuned counterpart in an IR Ni-style NIT system. Our domain-tuned lexicons brought about an improvement in the Blen scores: 9.4/\% higher than a system trained on a uniformly- weighted dictionary and 275/\% higher than a system trained on no dictionary at all.}, keywords = {*DICTIONARIES, *FOREIGN LANGUAGES, accuracy, BILINGUAL LEXICONS, DOCUMENTS, linguistics, Vocabulary}, url = {http://stinet.dtic.mil/oai/oai?\&verb=getRecord\&metadataPrefix=html\&identifier=ADA455197}, author = {Ayan,Necip F and Dorr, Bonnie J and Kolak,Okan} } @conference {11911, title = {Energy-efficient broadcast and multicast trees for reliable wireless communication}, booktitle = {Wireless Communications and Networking, 2003. WCNC 2003. 2003 IEEE}, volume = {1}, year = {2003}, month = {2003/03//}, pages = {660 -667 vol.1 - 660 -667 vol.1}, abstract = {We define energy-efficient broadband and multicast schemes for reliable communication in multi-hop wireless networks. Unlike previous techniques, the choice of neighbors in the broadband and multicast trees in these schemes, are based not only on the link distance, but also on the error rates associated with the link. Our schemes can be implemented using both positive and negative acknowledgement based reliable broadcast techniques in the link layer. Through simulations, we show that our scheme achieves up to 45\% improvement over previous schemes on realistic 100-node network topologies. A positive acknowledgment based implementation is preferred. Our simulations show that the additional benefits of a positive acknowledgement based implementation is marginal (1-2\%). Therefore a negative acknowledgement based implementation of our schemes is equally applicable in constructing energy-efficient reliable and multicast data delivery paths.}, keywords = {broadcast, channels;multicast, communication;broadcast, communication;network, data, delivery, energy-efficient, layer;multicast, network, networks;network, networks;telecommunication, paths;multicast, rates;link, reliability;, topologies;wireless, topology;radio, trees;error, trees;multihop, wireless}, doi = {10.1109/WCNC.2003.1200429}, author = {Banerjee,S. and Misra,A. and Yeo,Jihwang and Agrawala, Ashok K.} } @article {13768, title = {Evaluation Techniques Applied to Domain Tuning of MT Lexicons}, journal = {Proceedings of the Workshop {\textquotedblleft}Towards Systematizing MT Evaluation}, volume = {27}, year = {2003}, month = {2003///}, pages = {3 - 11}, abstract = {We describe a set of evaluation techniques applied to domain tuning of bilingual lexicons for machine translation.Our overall objective is to translate a domain-specific document in a foreign language (in this case, Chinese) to En- glish. First, we perform an intrinsic evaluation of the effectiveness of our domain-tuning techniques by comparing our domain-tuned lexicon to a manually constructed domain-specific bilingual termlist. Our results indicate that we achieve 66\% recall and 95\% precision with respect to a human-derived gold standard. Next, an extrinsic evaluation demonstrates that our domain-tuned lexicon improves the Bleu scores 50\% over a statistical system{\textemdash}with a smaller improvement when the system is trained on a uniformly-weighted dictionary. }, author = {Ayan,N.F. and Dorr, Bonnie J and Kolak,O.} } @conference {17834, title = {Exploiting Functional Decomposition for Efficient Parallel Processing of Multiple Data Analysis Queries}, booktitle = {Parallel and Distributed Processing Symposium, International}, year = {2003}, month = {2003///}, pages = {81a - 81a}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {Reuse is a powerful method for increasing system performance. In this paper, we examine functional decomposition for improving data and computation reuse and, therefore, overall query execution performance in the context of data analysis applications. Additionally, we look at the performance effects of using various projection primitives that make it possible to transform intermediate results generated by a query so that they can be reused by a new query. A satellite data analysis application is used to experimentally show the performance benefits achieved using functional decomposition and projection primitives.}, doi = {http://doi.ieeecomputersociety.org/10.1109/IPDPS.2003.1213184}, author = {Andrade,Henrique and Kurc,Tahsin and Sussman, Alan and Saltz,Joel} } @conference {12000, title = {Eye design in the plenoptic space of light rays}, booktitle = {Ninth IEEE International Conference on Computer Vision, 2003. Proceedings}, year = {2003}, month = {2003/10/13/16}, pages = {1160-1167 vol.2 - 1160-1167 vol.2}, publisher = {IEEE}, organization = {IEEE}, abstract = {Natural eye designs are optimized with regard to the tasks the eye-carrying organism has to perform for survival. This optimization has been performed by the process of natural evolution over many millions of years. Every eye captures a subset of the space of light rays. The information contained in this subset and the accuracy to which the eye can extract the necessary information determines an upper limit on how well an organism can perform a given task. In this work we propose a new methodology for camera design. By interpreting eyes as sample patterns in light ray space we can phrase the problem of eye design in a signal processing framework. This allows us to develop mathematical criteria for optimal eye design, which in turn enables us to build the best eye for a given task without the trial and error phase of natural evolution. The principle is evaluated on the task of 3D ego-motion estimation.}, keywords = {3D ego-motion estimation, Assembly, B-splines, Camera design, CAMERAS, captured image, compound eyes, Computer vision, data mining, eye, eye-carrying organism, Eyes, filter optimization, image representation, image resolution, Information geometry, Laboratories, light field reconstruction, light gathering power, light rays, mixed spherical-Cartesian coordinate system, Motion estimation, natural evolution process, natural eye designs, natural image statistics, optical nanotechnology, Optical signal processing, optimal eye design mathematical criteria, Organisms, plenoptic image formation, plenoptic space, plenoptic video geometry, sampling operators, sensory ecology, Signal design, Signal processing, signal processing framework, signal processing tool, square-summable sequences, visual acuity}, isbn = {0-7695-1950-4}, doi = {10.1109/ICCV.2003.1238623}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {15632, title = {Interpolation over light fields with applications in computer graphics}, booktitle = {Proceedings of the Fifth Workshop on Algorithm Engineering and Experiments}, year = {2003}, month = {2003///}, pages = {56 - 68}, author = {Atalay,F. B and Mount, Dave} } @conference {11896, title = {Kernel snakes: non-parametric active contour models}, booktitle = {IEEE International Conference on Systems, Man and Cybernetics, 2003}, volume = {1}, year = {2003}, month = {2003/10//}, pages = {240- 244 vol.1 - 240- 244 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {In this paper, a new non-parametric generalized formulation to statistical pressure snakes is presented. We discuss the shortcomings of the traditional pressure snakes. We then introduce a new generic pressure model that alleviates these shortcomings, based on the Bayesian decision theory. Non-parametric techniques are used to obtain the statistical models that drive the snake. We discuss the advantages of using the proposed non-parametric model compared to other parametric techniques. Multi-colored-target tracking is used to demonstrate the performance of the proposed approach. Experimental results show enhanced, real-time performance.}, keywords = {Active contours, Artificial intelligence, Bayes methods, Bayesian decision theory, Bayesian methods, decision theory, Deformable models, Image edge detection, Image segmentation, Intelligent robots, Kernel, kernel snakes, Laboratories, multicolored target tracking, nonparametric active contour models, nonparametric generalized formulation, nonparametric model, nonparametric statistics, nonparametric techniques, real time performance, Robot vision systems, statistical pressure snakes, target tracking}, isbn = {0-7803-7952-7}, doi = {10.1109/ICSMC.2003.1243822}, author = {Abd-Almageed, Wael and Smith,C.E. and Ramadan,S.} } @article {17706, title = {Matrix Algorithms, Volume II: Eigensystems}, journal = {Applied Mechanics Reviews}, volume = {56}, year = {2003}, month = {2003///}, pages = {B2-B2 - B2-B2}, keywords = {algorithm theory, Eigenvalues and eigenfunctions, linear algebra, matrix algebra, reviews}, doi = {10.1115/1.1523352}, url = {http://link.aip.org/link/?AMR/56/B2/1}, author = {Stewart, G.W. and Author and Mahajan,A. and Reviewer} } @article {18583, title = {Measuring the effects of internet path faults on reactive routing}, journal = {SIGMETRICS Perform. Eval. Rev.}, volume = {31}, year = {2003}, month = {2003/06//}, pages = {126 - 137}, abstract = {Empirical evidence suggests that reactive routing systems improve resilience to Internet path failures. They detect and route around faulty paths based on measurements of path performance. This paper seeks to understand why and under what circumstances these techniques are effective.To do so, this paper correlates end-to-end active probing experiments, loss-triggered traceroutes of Internet paths, and BGP routing messages. These correlations shed light on three questions about Internet path failures: (1) Where do failures appear? (2) How long do they last? (3) How do they correlate with BGP routing instability?Data collected over 13 months from an Internet testbed of 31 topologically diverse hosts suggests that most path failures last less than fifteen minutes. Failures that appear in the network core correlate better with BGP instability than failures that appear close to end hosts. On average, most failures precede BGP messages by about four minutes, but there is often increased BGP traffic both before and after failures. Our findings suggest that reactive routing is most effective between hosts that have multiple connections to the Internet. The data set also suggests that passive observations of BGP routing messages could be used to predict about 20\% of impending failures, allowing re-routing systems to react more quickly to failures.}, isbn = {0163-5999}, doi = {10.1145/885651.781043}, url = {http://doi.acm.org/10.1145/885651.781043}, author = {Feamster, Nick and Andersen,David G. and Balakrishnan,Hari and Kaashoek,M. Frans} } @conference {14235, title = {New eyes for robotics}, booktitle = {2003 IEEE/RSJ International Conference on Intelligent Robots and Systems, 2003. (IROS 2003). Proceedings}, volume = {1}, year = {2003}, month = {2003/10/27/31}, pages = {1018- 1023 vol.1 - 1018- 1023 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper describes an imaging system that has been designed to facilitate robotic tasks of motion. The system consists of a number of cameras in a network arranged so that they sample different parts of the visual sphere. This geometric configuration has provable advantages compared to small field of view cameras for the estimation of the system{\textquoteright}s own motion and consequently the estimation of shape models from the individual cameras. The reason is that inherent ambiguities of confusion between translation and rotation disappear. Pairs of cameras may also be arranged in multiple stereo configurations which provide additional advantages for segmentation. Algorithms for the calibration of the system and the 3D motion estimation are provided.}, keywords = {3D motion estimation, Argus eye, array signal processing, Birds, Calibration, CAMERAS, Control systems, Eyes, geometric configuration, imaging, imaging system, Layout, Motion estimation, multiple stereo configurations, Robot kinematics, robot vision, Robot vision systems, ROBOTICS, Robotics and automation, SHAPE, shape models}, isbn = {0-7803-7860-1}, doi = {10.1109/IROS.2003.1250761}, author = {Baker, P. and Ogale, A. S and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {11905, title = {Non-parametric expectation maximization: a learning automata approach}, booktitle = {IEEE International Conference on Systems, Man and Cybernetics, 2003}, volume = {3}, year = {2003}, month = {2003/10//}, pages = {2996- 3001 vol.3 - 2996- 3001 vol.3}, publisher = {IEEE}, organization = {IEEE}, abstract = {The famous expectation maximization technique suffers two major drawbacks. First, the number of components has to be specified apriori. Also, the expectation maximization is sensitive to initialization. In this paper, we present a new stochastic technique for estimating the mixture parameters. Parzen Window is used to estimate a discrete estimate of the PDF of the given data. Stochastic learning automata is then used to select the mixture parameters that minimize the distance between the discrete estimate of the PDF and the estimate of the expectation maximization. The validity of the proposed approach is verified using bivariate simulation data.}, keywords = {Automatic speech recognition, bivariate simulation data, Computational modeling, Computer vision, Density functional theory, expectation maximization technique, learning automata, mixture parameters estimation, nonparametric expectation maximization, nonparametric statistics, optimisation, parameter estimation, Parzen Window, PDF, probability, probability density function, Speech processing, Speech recognition, stochastic learning automata, Stochastic processes, stochastic technique}, isbn = {0-7803-7952-7}, doi = {10.1109/ICSMC.2003.1244347}, author = {Abd-Almageed, Wael and El-Osery,A. and Smith,C.E.} } @conference {18931, title = {On-line computation of two types of structural relations in Japanese}, year = {2003}, month = {2003///}, author = {Aoshima,S. and Phillips,C. and Weinberg, Amy} } @article {16143, title = {Overlaying graph links on treemaps}, journal = {IEEE Symposium on Information Visualization Conference Compendium (demonstration)}, year = {2003}, month = {2003///}, abstract = {Every graph can be decomposed into a tree structure plus a set ofremaining edges. We describe a visualization technique that displays the tree structure as a Treemap and the remaining edges as curved links overlaid on the Treemap. Link curves are designed to show where the link starts and where it ends without requiring an explicit arrow that would clutter the already dense visualization. This technique is effective for visualizing structures where the underlying tree has some meaning, such as Web sites or XML documents with cross-references. Graphic attributes of the links {\textendash} such as color or thickness {\textendash} can be used to represent attributes of the edges. Users can choose to see all links at once or only the links to and from the node or branch under the cursor. }, author = {Fekete,J. D and Wang,D. and Dang,N. and Aris,A. and Plaisant, Catherine} } @conference {12055, title = {Polydioptric camera design and 3D motion estimation}, booktitle = {2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings}, volume = {2}, year = {2003}, month = {2003/06/18/20}, pages = {II- 294-301 vol.2 - II- 294-301 vol.2}, publisher = {IEEE}, organization = {IEEE}, abstract = {Most cameras used in computer vision applications are still based on the pinhole principle inspired by our own eyes. It has been found though that this is not necessarily the optimal image formation principle for processing visual information using a machine. We describe how to find the optimal camera for 3D motion estimation by analyzing the structure of the space formed by the light rays passing through a volume of space. Every camera corresponds to a sampling pattern in light ray space, thus the question of camera design can be rephrased as finding the optimal sampling pattern with regard to a given task. This framework suggests that large field-of-view multi-perspective (polydioptric) cameras are the optimal image sensors for 3D motion estimation. We conclude by proposing design principles for polydioptric cameras and describe an algorithm for such a camera that estimates its 3D motion in a scene independent and robust manner.}, keywords = {3D motion estimation, Algorithm design and analysis, Application software, CAMERAS, Computer vision, Eyes, field-of-view camera, Image motion analysis, image sampling, image sensor, Image sensors, Layout, light ray, Motion estimation, multiperspective camera, optimal camera, optimal image formation, optimal sampling pattern, pinhole principle, polydioptric camera design, ray space, scene independent estimation, space structure analysis, stereo image processing, visual information processing}, isbn = {0-7695-1900-8}, doi = {10.1109/CVPR.2003.1211483}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {17522, title = {Quantifying the causes of path inflation}, journal = {Topology}, volume = {100}, year = {2003}, month = {2003///}, pages = {2 - 1}, author = {Spring, Neil and Mahajan,R. and Anderson,T.} } @article {15400, title = {Random doping-induced fluctuations of subthreshold characteristics in MOSFET devices}, journal = {Solid-State Electronics}, volume = {47}, year = {2003}, month = {2003/11//}, pages = {2055 - 2061}, abstract = {The random doping-induced fluctuations of subthreshold characteristics in MOSFET devices are analyzed. A technique for the computations of sensitivity coefficients and variances of subthreshold parameters is presented and applied to the computation of fluctuations of subthreshold current and gate-voltage swing. This technique is based on the linearization of transport equations with respect to the fluctuating quantities. It is computationally much more efficient than purely {\textquotedblleft}statistical{\textquotedblright} methods (Monte-Carlo methods) that are based on the simulations of a large number of devices with different doping realizations. The numerical implementation of this technique is discussed and numerous computational results are presented.}, keywords = {Fluctuations, Mismatch, MOSFET, Sensitivity analysis, Statistics, Submicron devices}, isbn = {0038-1101}, doi = {10.1016/S0038-1101(03)00236-3}, url = {http://www.sciencedirect.com/science/article/pii/S0038110103002363}, author = {Andrei,Petru and Mayergoyz, Issak D} } @article {13846, title = {Rapid porting of DUSTer to Hindi}, journal = {ACM Transactions on Asian Language Information Processing (TALIP)}, volume = {2}, year = {2003}, month = {2003/06//}, pages = {118 - 123}, abstract = {The frequent occurrence of divergences{\textemdash}structural differences between languages---presents a great challenge for statistical word-level alignment and machine translation. This paper describes the adaptation of DUSTer, a divergence unraveling package, to Hindi during the DARPA TIDES-2003 Surprise Language Exercise. We show that it is possible to port DUSTer to Hindi in under 3 days.}, keywords = {Divergences, Machine translation}, isbn = {1530-0226}, doi = {10.1145/974740.974744}, url = {http://doi.acm.org/10.1145/974740.974744}, author = {Dorr, Bonnie J and Ayan,Necip Fazil and Habash,Nizar and Madnani,Nitin and Hwa,Rebecca} } @book {12096, title = {Real 802.11 Security: Wi-Fi Protected Access and 802.11 i, 480 pages}, year = {2003}, month = {2003///}, publisher = {Addison Wesley}, organization = {Addison Wesley}, author = {Edney,J. and Arbaugh, William A.} } @article {12953, title = {Reduction of Cholera in Bangladeshi Villages by Simple Filtration}, journal = {Proceedings of the National Academy of SciencesPNAS}, volume = {100}, year = {2003}, month = {2003/02/04/}, pages = {1051 - 1055}, abstract = {Based on results of ecological studies demonstrating that Vibrio cholerae, the etiological agent of epidemic cholera, is commensal to zooplankton, notably copepods, a simple filtration procedure was developed whereby zooplankton, most phytoplankton, and particulates >20 μm were removed from water before use. Effective deployment of this filtration procedure, from September 1999 through July 2002 in 65 villages of rural Bangladesh, of which the total population for the entire study comprised ≈133,000 individuals, yielded a 48\% reduction in cholera (P < 0.005) compared with the control.}, isbn = {0027-8424, 1091-6490}, doi = {10.1073/pnas.0237386100}, url = {http://www.pnas.org/content/100/3/1051}, author = {Rita R Colwell and Huq,Anwar and M. Sirajul Islam and K. M. A. Aziz and Yunus,M. and N. Huda Khan and A. Mahmud and Sack,R. Bradley and Nair,G. B. and J. Chakraborty and Sack,David A. and E. Russek-Cohen} } @conference {17530, title = {Scriptroute: A public Internet measurement facility}, booktitle = {Proceedings of the 4th conference on USENIX Symposium on Internet Technologies and Systems-Volume 4}, year = {2003}, month = {2003///}, pages = {17 - 17}, author = {Spring, Neil and Wetherall,D. and Anderson,T.} } @conference {12078, title = {A secure service discovery protocol for MANET}, booktitle = {14th IEEE Proceedings on Personal, Indoor and Mobile Radio Communications, 2003. PIMRC 2003}, volume = {1}, year = {2003}, month = {2003/09/07/10}, pages = {502- 506 Vol.1 - 502- 506 Vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {Service discovery technologies are exploited to enable services to advertise their existence in a dynamic way, and can be discovered, configured and used by other devices with minimum manual efforts. It plays an essential role in future network scenarios especially with development of mobile ad hoc network (MANET) and emergence of pervasive computing. Because MANET allows these devices to communicate dynamically without fixed infrastructure and centralized administration, it gives rise to the challenges of the service discovery techniques. In this paper, we present a dynamic service discovery infrastructure that uses XML to describe services and match using the semantic content of service descriptions for MANET. We believe that the architecture we have designed is a necessary component of service discovery in non-infrastructure network by further exploring the secure and performance issues of this infrastructure.}, keywords = {ad hoc networks, centralized administration, Computer architecture, Computer science, dynamic service discovery infrastructure, Educational institutions, MANET, Manuals, mobile ad hoc network, Mobile ad hoc networks, Mobile computing, mobile radio, noninfrastructure network, Pervasive computing, Protocols, routing protocols, secure service discovery protocol, Security, service discovery techniques, service discovery technologies, telecommunication computing, telecommunication services, XML}, isbn = {0-7803-7822-9}, doi = {10.1109/PIMRC.2003.1264322}, author = {Yuan Yuan and Arbaugh, William A.} } @article {18870, title = {SHOP2: An HTN planning system}, journal = {J. Artif. Intell. Res. (JAIR)}, volume = {20}, year = {2003}, month = {2003///}, pages = {379 - 404}, abstract = {The SHOP2 planning system received one of the awards for distinguished performance in the 2002 International Planning Competition. This paper describes the features of SHOP2 which enabled it to excel in the competition, especially those aspects of SHOP2 that deal with temporal and metric planning domains.}, url = {https://www.aaai.org/Papers/JAIR/Vol20/JAIR-2013.pdf}, author = {Nau, Dana S. and Au,T. C and Ilghami,O. and Kuter,U. and Murdock,J. W. and Wu,D. and Yaman,F.} } @book {14916, title = {Special section on perceptual organization in computer vision}, year = {2003}, month = {2003///}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, author = {Jacobs, David W. and Lindenbaum,M. and August,J. and Zucker,SW and Ben-Shahar,O. and Zucker,SW and Tuytelaars,T. and Turina,A. and Van Gool,L. and Mahamud,S.} } @conference {15096, title = {Toward secure key distribution in truly ad-hoc networks}, booktitle = {Applications and the Internet Workshops, 2003. Proceedings. 2003 Symposium on}, year = {2003}, month = {2003/01//}, pages = {342 - 346}, abstract = {Ad-hoc networks - and in particular wireless mobile ad-hoc networks $have unique characteristics and constraints that make traditional cryptographic mechanisms and assumptions inappropriate. In particular it may not be warranted to assume pre-existing shared secrets between members of the network or the presence of a common PKI. Thus, the issue of key distribution in ad-hoc networks represents an important problem. Unfortunately, this issue has been largely ignored; as an example, most protocols for secure ad-hoc routing assume that key distribution has already taken place. Traditional key distribution schemes either do not apply in an ad-hoc scenario or are not efficient enough for small, resource-constrained devices. We propose to combine efficient techniques from identity-based (ID-based) and threshold cryptography to provide a mechanism that enables flexible and efficient key distribution while respecting the constraints of ad-hoc networks. We also discuss the available mechanisms and their suitability for the proposed task.}, keywords = {ad, ad-hoc, cryptography;, distribution;, hoc, identity-based, key, mobile, network, networks;, routing;, secure, security;, Telecommunication, threshold, wireless}, doi = {10.1109/SAINTW.2003.1210183}, author = {Khalili,A. and Katz, Jonathan and Arbaugh, William A.} } @conference {15930, title = {Towards domain-independent, task-oriented, conversational adequacy}, booktitle = {INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE}, volume = {18}, year = {2003}, month = {2003///}, pages = {1637 - 1638}, author = {Josyula,D. P and Anderson,M. L and Perlis, Don} } @article {16898, title = {Use of the SAND spatial browser for digital government applications}, journal = {Commun. ACM}, volume = {46}, year = {2003}, month = {2003/01//}, pages = {61 - 64}, abstract = {Numerous federal agencies produce official statistics made accessible to ordinary citizens for searching and data retrieval. This is frequently done via the Internet through a Web browser interface. If this data is presented in textual format, it can often be searched and retrieved by such attributes as topic, responsible agency, keywords, or press release. However, if the data is of spatial nature, for example, in the form of a map, then using text-based queries is often too cumbersome for the intended audience. We describe the use of the SAND Spatial Browser to provide more power to users of these databases by enabling them to define and explore the specific spatial region of interest graphically. The SAND Spatial Browser allows users to form either purely spatial or mixed spatial/nonspatial queries intuitively, which can present information to users that might have been missed if only a textual interface was available.}, keywords = {GIS}, isbn = {0001-0782}, doi = {10.1145/602421.602453}, url = {http://doi.acm.org/10.1145/602421.602453}, author = {Samet, Hanan and Alborzi,Houman and Brabec,Franti{\v s}ek and Esperan{\c c}a,Claudio and Hjaltason,G{\'\i}sli R. and Morgan,Frank and Tanin,Egemen} } @conference {17524, title = {User-level internet path diagnosis}, booktitle = {ACM SIGOPS Operating Systems Review}, volume = {37}, year = {2003}, month = {2003///}, pages = {106 - 119}, author = {Mahajan,R. and Spring, Neil and Wetherall,D. and Anderson,T.} } @article {12084, title = {WLAN problems and solutions}, journal = {Communications of the ACM}, volume = {46}, year = {2003}, month = {2003///}, pages = {31 - 34}, author = {Housley,R. and Arbaugh, William A.} } @article {13414, title = {AAAI 2002 Workshops}, journal = {AI Magazine}, volume = {23}, year = {2002}, month = {2002/12/15/}, pages = {113 - 113}, isbn = {0738-4602}, doi = {10.1609/aimag.v23i4.1678}, url = {http://www.aaai.org/ojs/index.php/aimagazine/article/viewArticle/1678}, author = {Blake,Brian and Haigh,Karen and Hexmoor,Henry and Falcone,Rino and Soh,Leen-Kiat and Baral,Chitta and McIlraith,Sheila and Gmytrasiewicz,Piotr and Parsons,Simon and Malaka,Rainer and Krueger,Antonio and Bouquet,Paolo and Smart,Bill and Kurumantani,Koichi and Pease,Adam and Brenner,Michael and desJardins, Marie and Junker,Ulrich and Delgrande,Jim and Doyle,Jon and Rossi,Francesca and Schaub,Torsten and Gomes,Carla and Walsh,Toby and Guo,Haipeng and Horvitz,Eric J and Ide,Nancy and Welty,Chris and Anger,Frank D and Guegen,Hans W and Ligozat,Gerald} } @conference {17883, title = {Active Proxy-G: Optimizing the query execution process in the Grid}, booktitle = {Proceedings of the 2002 ACM/IEEE conference on Supercomputing}, year = {2002}, month = {2002///}, pages = {1 - 15}, author = {Andrade,H. and Kurc, T. and Sussman, Alan and Saltz, J.} } @article {11915, title = {AD (Attacker Defender) Game}, journal = {Technical Reports from UMIACS, UMIACS-TR-2001-45}, year = {2002}, month = {2002/01/31/}, abstract = {Information Dynamics is a framework for agent-based systems that givesa central position to the role of information, time, and the value of information. We illustrate system design in the Information Dynamics Framework by developing an intelligence game called AD involving attackers, defenders and targets operating in some space of locations. The goal of the attackers is to destroy all targets. Target destruction takes place when the number of attackers in the target{\textquoteright}s neighborhood exceeds the number of defenders in this neighborhood by a value WINNING_DIFFERENCE. The goal of defenders is to prevent attackers from achieving their goal. (Also UMIACS-TR-2001-45) }, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/1142}, author = {Kochut,Andrzej and Agrawala, Ashok K. and Larsen,Ronald L and A. Udaya Shankar} } @article {15541, title = {Algorithmic issues in modeling motion}, journal = {ACM Comput. Surv.}, volume = {34}, year = {2002}, month = {2002/12//}, pages = {550 - 572}, abstract = {This article is a survey of research areas in which motion plays a pivotal role. The aim of the article is to review current approaches to modeling motion together with related data structures and algorithms, and to summarize the challenges that lie ahead in producing a more unified theory of motion representation that would be useful across several disciplines.}, keywords = {computational geometry, Computer vision, mobile networks, modeling, molecular biology, motion modeling, physical simulation, robotoics, spatio-temporal databases}, isbn = {0360-0300}, doi = {10.1145/592642.592647}, url = {http://doi.acm.org/10.1145/592642.592647}, author = {Agarwal,Pankaj K. and Guibas,Leonidas J. and Edelsbrunner,Herbert and Erickson,Jeff and Isard,Michael and Har-Peled,Sariel and Hershberger,John and Jensen,Christian and Kavraki,Lydia and Koehl,Patrice and Lin,Ming and Manocha,Dinesh and Metaxas,Dimitris and Mirtich,Brian and Mount, Dave and Muthukrishnan,S. and Pai,Dinesh and Sacks,Elisha and Snoeyink,Jack and Suri,Subhash and Wolefson,Ouri} } @conference {16002, title = {ALLI: An Information Integration System Based on Active Logic Framework}, booktitle = {Proceedings of the Third International Conference on Management Information Systems, Greece}, year = {2002}, month = {2002///}, pages = {339 - 348}, author = {Barfourosh,A. A. and Nezhad,H. R.M and Anderson,M. and Perlis, Don} } @article {14538, title = {Approximating the head-related transfer function using simple geometric models of the head and torso}, journal = {The Journal of the Acoustical Society of America}, volume = {112}, year = {2002}, month = {2002///}, pages = {2053 - 2053}, author = {Algazi,V. R and Duda,R. O and Duraiswami, Ramani and Gumerov, Nail A. and Tang,Z.} } @conference {18933, title = {Argument structure saturation in a constraint based model}, year = {2002}, month = {2002///}, url = {http://cognet.mit.edu/library/conferences/paper?paper_id=55048}, author = {Drury,J. and Resnik, Philip and Weinberg, Amy and Gennari,S. and Aoshima,S.} } @inbook {16960, title = {An Augmented Visual Query Mechanism for Finding Patterns in Time Series Data}, booktitle = {Flexible Query Answering SystemsFlexible Query Answering Systems}, series = {Lecture Notes in Computer Science}, volume = {2522}, year = {2002}, month = {2002///}, pages = {240 - 250}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Relatively few query tools exist for data exploration and pattern identification in time series data sets. In previous work we introduced Timeboxes . Timeboxes are rectangular, direct-manipulation queries for studying time-series datasets. We demonstrated how Timeboxes can be used to support interactive exploration via dynamic queries, along with overviews of query results and drag-and-drop support for query-by-example. In this paper, we extend our work by introducing Variable Time Timeboxes (VTT). VTTs are a natural generalization of Timeboxes, which permit the specification of queries that allow a degree of uncertainty in the time axis. We carefully motivate the need for these more expressive queries, and demonstrate the utility of our approach on several data sets.}, isbn = {978-3-540-00074-7}, url = {http://dx.doi.org/10.1007/3-540-36109-X_19}, author = {Keogh,Eamonn and Hochheiser,Harry and Shneiderman, Ben}, editor = {Carbonell,Jaime and Siekmann,J{\"o}rg and Andreasen,Troels and Christiansen,Henning and Motro,Amihai and Legind Larsen,Henrik} } @article {14204, title = {Bias in visual motion processes: A theory predicting illusions}, journal = {Statistical Methods in Video Processing.(in conjunction with European Conference on Computer Vision)}, year = {2002}, month = {2002///}, abstract = {The first step in the interpretation of video sequences is theestimation of spatio-temporal image representations, that is the image motion or the correspondence of points or lines. Noise in the data, however, causes a serious problem for their estimation; in particular, it causes bias. As a result the location of image points, the orientation of image lines and the length and orientation of image motion vectors are esti- mated erroneously. It is suggested that machine vision sys- tems processing video should include statistical procedures to alleviate the bias. However, complete correction in gen- eral does not appear to be feasible, and this provides an ex- planation for most well-known geometrical optical illusions, such as the caf{\'e} wall, the Z{\"o}llner, the Poggendorff illusion and other recently discovered illusions of movement. }, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J. and Malm,H.} } @conference {17868, title = {On cache replacement policies for servicing mixed data intensive query workloads}, booktitle = {Proceedings of the 2nd Workshop on Caching, Coherence, and Consistency, held in conjunction with the 16th ACM International Conference on Supercomputing}, year = {2002}, month = {2002///}, abstract = {When data analysis applications are employed in a multi-client environment, a data server must service multiple si- multaneous queries, each of which may employ complex user-defined data structures and operations on the data. It is then necessary to harness inter- and intra-query com- monalities and system resources to improve the performance of the data server. We have developed a framework and customizable middleware to enable reuse of intermediate and final results among queries, through an in-memory ac- tive semantic cache and user-defined transformation func- tions. Since resources such as processing power and mem- ory space are limited on the machine hosting the server, ef- fective scheduling of incoming queries and efficient cache replacement policies are challenging issues that must be ad- dressed. We have worked on the scheduling problem in ear- lier work, and in this paper we describe and evaluate several cache replacement policies. We present experimental eval- uation of the policies on a shared-memory parallel system using two applications from different application domains. }, author = {Andrade,H. and Kurc, T. and Sussman, Alan and Borovikov,E. and Saltz, J.} } @conference {18873, title = {CaMeL: Learning method preconditions for HTN planning}, year = {2002}, month = {2002///}, pages = {131 - 14}, abstract = {A great challenge in using any planning system to solve real-world problems is the difficulty of acquiring the do- main knowledge that the system will need. We present a way to address part of this problem, in the context of Hierarchical Task Network (HTN) planning, by having the planning system incrementally learn conditions for HTN methods under expert supervision. We present a general formal framework for learning HTN methods, and a supervised learning algorithm, named CaMeL, based on this formalism. We present theoretical results about CaMeL{\textquoteright}s soundness, completeness, and conver- gence properties. We also report experimental results about its speed of convergence under different condi- tions. The experimental results suggest that CaMeL has the potential to be useful in real-world applications.}, url = {https://www.aaai.org/Papers/AIPS/2002/AIPS02-014.pdf}, author = {Ilghami,O. and Nau, Dana S. and Munoz-Avila,H. and Aha,D. W.} } @conference {17575, title = {Clustering and server selection using passive monitoring}, booktitle = {IEEE INFOCOM 2002. Twenty-First Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings}, volume = {3}, year = {2002}, month = {2002///}, pages = {1717- 1725 vol.3 - 1717- 1725 vol.3}, publisher = {IEEE}, organization = {IEEE}, abstract = {We consider the problem of client assignment in a distributed system of content servers. We present a system called Webmapper for clustering IP addresses and assigning each cluster to an optimal content server. The system is passive in that the only information it uses comes from monitoring the TCP connections between the clients and the servers. It is also flexible in that it makes no a priori assumptions about network topology and server placement and it can react quickly to changing network conditions. We present experimental results to evaluate the performance of Webmapper.}, keywords = {client assignment, client-server systems, clustering, content servers, Delay, distributed system, Educational institutions, Internet, IP addresses, Monitoring, network conditions, Network servers, Network topology, optimal content server, passive monitoring, server selection, Space technology, TCPIP, Transport protocols, Web pages, Web server, Webmapper}, isbn = {0-7803-7476-2}, doi = {10.1109/INFCOM.2002.1019425}, author = {Andrews,M. and Shepherd,B. and Srinivasan, Aravind and Winkler,P. and Zane,F.} } @inbook {18914, title = {On the Complexity of Plan Adaptation by Derivational Analogy in a Universal Classical Planning Framework}, booktitle = {Advances in Case-Based Reasoning}, series = {Lecture Notes in Computer Science}, volume = {2416}, year = {2002}, month = {2002///}, pages = {199 - 206}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {In this paper we present an algorithm called DerUCP, which can be regarded as a general model for plan adaptation using Derivational Analogy. Using DerUCP, we show that previous results on the complexity of plan adaptation do not apply to Derivational Analogy. We also show that Derivational Analogy can potentially produce exponential reductions in the size of the search space generated by a planning system.}, keywords = {Computer science}, isbn = {978-3-540-44109-0}, url = {http://www.springerlink.com/content/db2mr13j90mfrcaw/abstract/}, author = {Au,Tsz-Chiu and Mu{\~n}oz-Avila,H{\'e}ctor and Nau, Dana S.}, editor = {Craw,Susan and Preece,Alun} } @conference {11904, title = {Contour migration: solving object ambiguity with shape-space visual guidance}, booktitle = {IEEE/RSJ International Conference on Intelligent Robots and Systems, 2002}, volume = {1}, year = {2002}, month = {2002///}, pages = {330- 335 vol.1 - 330- 335 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {A fundamental problem in computer vision is the issue of shape ambiguity. Simply stated, a silhouette cannot uniquely identify an object or an object{\textquoteright}s classification since many unique objects can present identical occluding contours. This problem has no solution in the general case for a monocular vision system. This paper presents a method for disambiguating objects during silhouette matching using a visual servoing system. This method identifies the camera motion(s) that gives disambiguating views of the objects. These motions are identified through a new technique called contour migration. The occluding contour{\textquoteright}s shape is used to identify objects or object classes that are potential matches for that shape. A contour migration is then determined that disambiguates the possible matches by purposive viewpoint adjustment. The technique is demonstrated using an example set of objects.}, keywords = {Artificial intelligence, camera motion, CAMERAS, Computer vision, contour migration, Databases, edge detection, Intelligent robots, Laboratories, Machine vision, object ambiguity, Object recognition, pattern matching, Robot vision systems, servomechanisms, SHAPE, shape matching, shape-space visual guidance, silhouette matching, visual servoing}, isbn = {0-7803-7398-7}, doi = {10.1109/IRDS.2002.1041410}, author = {Abd-Almageed, Wael and Smith,C.E.} } @article {11912, title = {Efficient Data Processing using Cross Layer Hints}, journal = {Technical Reports from UMIACS, UMIACS-TR-2002-59}, year = {2002}, month = {2002/12/19/}, abstract = {Conventional network stacks define a layered architecture, whereeach layer implements a set of services and exports a well-defined interface to be used by its immediate upper layer. A key design choice of the layered architecture has been to provide isolation between the functional modules of distinct layers. While such an architecture provides an useful abstraction for system development, the strict isolation of this layered architecture limits the flexibility of tailoring the behavior of the lower layers of the stack to the needs of the application. In this paper we define a new architecture, called X-Tags, which allows flexible interaction between layers for cooperative data processing without impacting the isolation property. In this architecture, applications use special tags to provide semantic hints for data processing to lower layers. We motivate the usefulness of this architecture by describing ts applicability to some emerging applications. UMIACS-TR-2002-59 }, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/1211}, author = {Banerjee,Suman and Agrawala, Ashok K. and Kramer,Michael J} } @conference {12763, title = {An experimental evaluation of linear and kernel-based methods for face recognition}, booktitle = {Applications of Computer Vision, 2002. (WACV 2002). Proceedings. Sixth IEEE Workshop on}, year = {2002}, month = {2002///}, pages = {13 - 18}, abstract = {In this paper we present the results of a comparative study of linear and kernel-based methods for face recognition. The methods used for dimensionality reduction are Principal Component Analysis (PCA), Kernel Principal Component Analysis (KPCA), Linear Discriminant Analysis (LDA) and Kernel Discriminant Analysis (KDA). The methods used for classification are Nearest Neighbor (NN) and Support Vector Machine (SVM). In addition, these classification methods are applied on raw images to gauge the performance of these dimensionality reduction techniques. All experiments have been performed on images from UMIST Face Database.}, keywords = {analysis;, classification;, component, discriminant, Face, image, Kernel, linear, Machine;, nearest, neighbor;, principal, recognition;, Support, vector}, doi = {10.1109/ACV.2002.1182137}, author = {Gupta, H. and Agrawala, Ashok K. and Pruthi, T. and Shekhar, C. and Chellapa, Rama} } @conference {11893, title = {Eye tracking using active deformable models}, booktitle = {Proc. of the IIIrd Indian Conference on Computer Vision, Graphics and Image Processing}, year = {2002}, month = {2002///}, author = {Ramadan,S. and Abd-Almageed, Wael and Smith,C.E.} } @conference {12037, title = {Eyes form eyes: New cameras for structure from motion}, booktitle = {Proceedings Workshop on Omnidirectional Vision (OMNIVIS)}, year = {2002}, month = {2002///}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {14219, title = {Eyes from eyes: new cameras for structure from motion}, booktitle = {Third Workshop on Omnidirectional Vision, 2002. Proceedings}, year = {2002}, month = {2002///}, pages = {19 - 26}, publisher = {IEEE}, organization = {IEEE}, abstract = {We investigate the relationship between camera design and the problem of recovering the motion and structure of a scene from video data. The visual information that could possibly be obtained is described by the plenoptic function. A camera can be viewed as a device that captures a subset of this function, that is, it measures some of the light rays in some part of the space. The information contained in the subset determines how difficult it is to solve subsequent interpretation processes. By examining the differential structure of the time varying plenoptic function we relate different known and new camera models to the spatiotemporal structure of the observed scene. This allows us to define a hierarchy of camera designs, where the order is determined by the stability and complexity of the computations necessary to estimate structure and motion. At the low end of this hierarchy is the standard planar pinhole camera for which the structure from motion problem is non-linear and ill-posed. At the high end is a new camera, which we call the full field of view polydioptric camera, for which the problem is linear and stable. In between are multiple-view cameras with large fields of view which we have built, as well as catadioptric panoramic sensors and other omni-directional cameras. We develop design suggestions for the polydioptric camera, and based upon this new design we propose a linear algorithm for ego-motion estimation, which in essence combines differential motion estimation with differential stereo.}, keywords = {Algorithm design and analysis, Camera design, CAMERAS, Design automation, differential stereo, Educational institutions, ego-motion estimation, Eyes, Geometrical optics, IMAGE PROCESSING, Layout, Motion estimation, observed scene, Optical films, polydioptric camera, RETINA}, isbn = {0-7695-1629-7}, doi = {10.1109/OMNVIS.2002.1044486}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14312, title = {Flow-sensitive type qualifiers}, journal = {ACM SIGPLAN Notices}, volume = {37}, year = {2002}, month = {2002///}, pages = {1 - 12}, author = {Foster, Jeffrey S. and Terauchi,T. and Aiken,A.} } @article {13776, title = {Generating A Parsing Lexicon from an LCS-Based Lexicon}, journal = {LREC 2002 Workshop Proceedings: Linguistic Knowledge Acquisition and Representation: Bootstrapping Annotated Language Data}, year = {2002}, month = {2002///}, abstract = {This paper describes a technique for generating parsing lexicons for a principle-based parser (Minipar). Our approach mapslexical entries in a large LCS-based repository of semantically classified verbs to their corresponding syntactic patterns. A by-product of this mapping is a lexicon that is directly usable in the Minipar system. We evaluate the accuracy and coverage of this lexicon using LDOCE syntactic codes as a gold standard. We show that this lexicon is comparable to the hand-generated Minipar lexicon (i.e., similar recall and precision values). In a later experiment, we automate the process of mapping between the LCS-based repository and syntactic patterns. The advantage of automating the process is that the same technique can be applied directly to lexicons we have for other languages, for example, Arabic, Chinese, and Spanish. }, author = {Ayan,N.F. and Dorr, Bonnie J} } @article {13777, title = {Generating A Parsing Lexicon From Lexical-Conceptual Structure}, year = {2002}, month = {2002///}, institution = {Instititue for Advanced Computer Studies, Univ of Maryland, College Park}, abstract = {This paper describes the generation of a lexicon for a principle-based parser (Minipar[5,6]) using descriptions of verbs based on Lexical-Conceptual Structure (LCS [1,2]).1 We aim to determine how much syntactic information we can obtain from a semantic- based lexicon. More specifically, we aim to provide a general approach to projection of syntactic entries from semantic (language-independent) lexicons-and to test the effect of such lexicons on parser performance. Verbs are grouped together into classes-each denoted by an LCS representation adn the thematic grid. These are mapped systemat- ically into syntactic categories associated with entries in the Minipar parser. The main advantage of this LCS-to-syntax projection is language protability: We currently have LCS lexicons for English, Arabic, Spanish, and Chinese; thus, our LCS-projection ap- proach allows us to produce syntactic lexicons for parsing in each of these languages. For comparing the performance of the projection from the LCS to Minpar coes, we also generated the mappings for the codes of Longman{\textquoteright}s Directionary of Contemporary En- glish (LDOCE [8])-the most comprehensive online dictionary for syntactic categorization. Preliminary experiments indicate that our approach yields a categorization of verbs with 58\% precision and 65\% recall as measured against LDOCE-with an improved precision of 74\% when redundancy is removed. The next section presents a brief description of each code set we use. In Section 3, we explain how we generated Minipar codes from LCS representation. Finally, Section 4, discusses our experiments and results. }, author = {Ayan,N.F. and Dorr, Bonnie J} } @article {16270, title = {Genome sequence and comparative analysis of the model rodent malaria parasite Plasmodium yoelii yoelii}, journal = {Nature}, volume = {419}, year = {2002}, month = {2002/10/03/}, pages = {512 - 519}, abstract = {Species of malaria parasite that infect rodents have long been used as models for malaria disease research. Here we report the whole-genome shotgun sequence of one species, Plasmodium yoelii yoelii, and comparative studies with the genome of the human malaria parasite Plasmodium falciparum clone 3D7. A synteny map of 2,212 P. y. yoelii contiguous DNA sequences (contigs) aligned to 14 P. falciparum chromosomes reveals marked conservation of gene synteny within the body of each chromosome. Of about 5,300 P. falciparum genes, more than 3,300 P. y. yoelii orthologues of predominantly metabolic function were identified. Over 800 copies of a variant antigen gene located in subtelomeric regions were found. This is the first genome sequence of a model eukaryotic parasite, and it provides insight into the use of such systems in the modelling of Plasmodium biology and disease.}, isbn = {0028-0836}, doi = {10.1038/nature01099}, url = {http://www.nature.com/nature/journal/v419/n6906/full/nature01099.html}, author = {Carlton,Jane M. and Angiuoli,Samuel V and Suh,Bernard B. and Kooij,Taco W. and Pertea,Mihaela and Silva,Joana C. and Ermolaeva,Maria D. and Allen,Jonathan E and Jeremy D Selengut and Koo,Hean L. and Peterson,Jeremy D. and Pop, Mihai and Kosack,Daniel S. and Shumway,Martin F. and Bidwell,Shelby L. and Shallom,Shamira J. and Aken,Susan E. van and Riedmuller,Steven B. and Feldblyum,Tamara V. and Cho,Jennifer K. and Quackenbush,John and Sedegah,Martha and Shoaibi,Azadeh and Cummings,Leda M. and Florens,Laurence and Yates,John R. and Raine,J. Dale and Sinden,Robert E. and Harris,Michael A. and Cunningham,Deirdre A. and Preiser,Peter R. and Bergman,Lawrence W. and Vaidya,Akhil B. and Lin,Leo H. van and Janse,Chris J. and Waters,Andrew P. and Smith,Hamilton O. and White,Owen R. and Salzberg,Steven L. and Venter,J. Craig and Fraser,Claire M. and Hoffman,Stephen L. and Gardner,Malcolm J. and Carucci,Daniel J.} } @conference {11894, title = {Hidden Markov models for silhouette classification}, booktitle = {Automation Congress, 2002 Proceedings of the 5th Biannual World}, volume = {13}, year = {2002}, month = {2002///}, pages = {395 - 402}, publisher = {IEEE}, organization = {IEEE}, abstract = {In this paper, a new technique for object classification from silhouettes is presented. Hidden Markov models are used as a classification mechanism. Through a set of experiments, we show the validity of our approach and show its invariance under severe rotation conditions. Also, a comparison with other techniques that use hidden Markov models for object classification from silhouettes is presented.}, keywords = {Computer vision, Feature extraction, Fourier transforms, hidden Markov models, HMM, image classification, Neural networks, object classification, Object recognition, parameter estimation, pattern recognition, Probability distribution, Shape measurement, silhouette classification, Wavelet transforms}, isbn = {1-889335-18-5}, doi = {10.1109/WAC.2002.1049575}, author = {Abd-Almageed, Wael and Smith,C.} } @article {14091, title = {Identification of non-autonomous non-LTR retrotransposons in the genome of Trypanosoma cruzi}, journal = {Molecular and Biochemical Parasitology}, volume = {124}, year = {2002}, month = {2002/09/10/}, pages = {73 - 78}, abstract = {As observed for most eukaryotic cells, trypanosomatids contains non-LTR retrotransposons randomly inserted in the nuclear genome. Autonomous retroelements which, code for their own transposition, have been characterized in Trypanosoma brucei (ingi) and Trypanosoma cruzi (L1Tc), whereas non-autonomous retroelements have only been characterized in T. brucei (RIME). Here, we have characterized in the genome of Trypanosoma cruzi four complete copies of a non-autonomous non-LTR retrotransposon, called NARTc. This 0.26 kb NARTc element has the characteristics of non-LTR retrotransposons: the presence a poly(dA) tail and of a short flanking duplicated motif. Analysis of the Genome Survey Sequence databases indicated that the Trypanosoma cruzi haploid genome contains about 140 NARTc copies and about twice as many L1Tc copies. Interestingly, the NARTc and L1Tc retroelements share, with the Trypanosoma brucei ingi and RIME retrotransposons, a common sequence (the first 45 bp with 91\% identity), whereas the remaining sequences are very divergent. This suggests that these four trypanosome non-LTR retrotransposons were derived from the same common ancester and the sequence of their 5{\textquoteright}-extremity may have a functional role. In addition, the genome of Leishmania major contains the same conserved motif present in the trypanosome retroelements, whicle no transposable elements have been detected so far in Leishmania sp.}, keywords = {Ingi, L1Tc, Non-LTR retrotransposon, RIME, Trypanosoma brucei, Trypanosoma cruzi}, isbn = {0166-6851}, doi = {16/S0166-6851(02)00167-6}, url = {http://www.sciencedirect.com/science/article/pii/S0166685102001676}, author = {Bringaud,Fr{\'e}d{\'e}ric and Garc{\'\i}a-P{\'e}rez,Jos{\'e} Luis and Heras,Sara R. and Ghedin,Elodie and El-Sayed, Najib M. and Andersson,Bj{\"o}rn and Baltz,Th{\'e}o and Lopez,Manuel C.} } @conference {17539, title = {Inferring link weights using end-to-end measurements}, booktitle = {Proceedings of the 2nd ACM SIGCOMM Workshop on Internet measurment}, series = {IMW {\textquoteright}02}, year = {2002}, month = {2002///}, pages = {231 - 236}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {We describe a novel constraint-based approach to approximate ISP link weights using only end-to-end measurements. Common routing protocols such as OSPF and IS-IS choose least-cost paths using link weights, so inferred weights provide a simple, concise, and useful model of intradomain routing. Our approach extends router-level ISP maps, which include only connectivity, with link weights that are consistent with routing. Our inferred weights agree well with observed routing: while our inferred weights fully characterize the set of shortest paths between 84--99\% of the router-pairs, alternative models based on hop count and latency do so for only 47--81\% of the pairs.}, isbn = {1-58113-603-X}, doi = {10.1145/637201.637237}, url = {http://doi.acm.org/10.1145/637201.637237}, author = {Mahajan,Ratul and Spring, Neil and Wetherall,David and Anderson,Tom} } @article {16028, title = {Metareasoning for More Effective Human-Computer Dialogue}, year = {2002}, month = {2002/12/31/}, institution = {University of Maryland, College Park}, abstract = {The research project explores specific meta-dialogue behaviors in terms of both how a system could be made to perform them, and to what extent they can increase overall system performance. We focus on two types of meta-dialogue capabilities: ability to detect and recover from anomalous dialogue patterns in simple exchanges, and on-line extensions or changes to working vocabulary. Our main method involves detailed representation of the dialogue context, separating domain, language, and dialogue specific aspects, and significant amounts of meta-reasoning about the system{\textquoteright}s processing of these representations. An existing logical inference system, ALMA/CARNE, developed as part of a pilot study, is being used in an implementation phase of this work. We are also engaged in a study of existing dialogue corpora to investigate the range and frequency of meta-dialogue expressions in different task domains.}, keywords = {*LEARNING MACHINES, COMPUTER LOGIC, COMPUTER PROGRAMMING AND SOFTWARE, CYBERNETICS, MAN COMPUTER INTERFACE, METAREASONING, NATURAL LANGUAGE., PROCESSING}, url = {http://stinet.dtic.mil/oai/oai?\&verb=getRecord\&metadataPrefix=html\&identifier=ADA414209}, author = {Perlis, Don and Anderson,Mike} } @conference {11899, title = {Mixture models for dynamic statistical pressure snakes}, booktitle = {16th International Conference on Pattern Recognition, 2002. Proceedings}, volume = {2}, year = {2002}, month = {2002///}, pages = {721- 724 vol.2 - 721- 724 vol.2}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper introduces a new approach to statistical pressure snakes. It uses statistical modeling for both object and background to obtain a more robust pressure model. The Expectation Maximization (EM) algorithm is used to model the data into a Mixture of Gaussians (MoG). Bayesian theory is then employed as a decision making mechanism. Experimental results using the traditional pressure model and the new mixture pressure model demonstrate the effectiveness of the new models.}, keywords = {active contour models, Active contours, Artificial intelligence, Bayes methods, Bayesian methods, Bayesian theory, complex colored object, Computer vision, decision making, decision making mechanism, dynamic statistical pressure snakes, Equations, expectation maximization algorithm, Gaussian distribution, image colour analysis, Image edge detection, Image segmentation, Intelligent robots, mixture models, mixture of Gaussians, mixture pressure model, Robot vision systems, robust pressure model, Robustness, segmentation results, statistical analysis, statistical modeling}, isbn = {0-7695-1695-X}, doi = {10.1109/ICPR.2002.1048404}, author = {Abd-Almageed, Wael and Smith,C.E.} } @conference {17835, title = {Multiple Query Optimization for Data Analysis Applications on Clusters of SMPs}, booktitle = {2nd IEEE/ACM International Symposium on Cluster Computing and the Grid, 2002}, year = {2002}, month = {2002/05/21/24}, pages = {154 - 154}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper is concerned with the efficient execution of multiple query workloads on a cluster of SMPs. We target applications that access and manipulate large scientific datasets. Queries in these applications involve user-defined processing operations and distributed data structures to hold intermediate and final results. Our goal is to implement system components to leverage previously computed query results and to effectively utilize processing power and aggregated I/O bandwidth on SMP nodes so that both single queries and multi-query batches can be efficiently executed.}, keywords = {Aggregates, Application software, Bandwidth, Data analysis, Data structures, Delay, Query processing, scheduling, Subcontracting, Switched-mode power supply}, isbn = {0-7695-1582-7}, doi = {10.1109/CCGRID.2002.1017123}, author = {Andrade,H. and Kurc, T. and Sussman, Alan and Saltz, J.} } @conference {11901, title = {A non-intrusive Kalman filter-based tracker for pursuit eye movement}, booktitle = {American Control Conference, 2002. Proceedings of the 2002}, volume = {2}, year = {2002}, month = {2002///}, pages = {1443- 1447 vol.2 - 1443- 1447 vol.2}, publisher = {IEEE}, organization = {IEEE}, abstract = {In this paper, we introduce a new non-intrusive approach to estimating the eye position during pursuit motion of the eye. We introduce a new characterization for the pursuit eye movement. Our characterization is based on the decomposition of the pursuit eye motion into a deterministic component and random component. We use a discrete Kalman filter to estimate the random component and calculate the deterministic component. We add the two components to obtain an estimate of the eye position. Simulation results are provided to illustrate the eye position estimation.}, keywords = {Application software, characterization, Computer vision, Current measurement, deterministic component, Electric variables measurement, eye position estimation, eye tracking, gaze tracking, Human computer interaction, Kalman filter, Kalman filters, Lenses, Motion estimation, Optical reflection, pursuit eye movement, pursuit motion, random component, Skin, tracking}, isbn = {0-7803-7298-0}, doi = {10.1109/ACC.2002.1023224}, url = {http://ieeexplore.ieee.org/ielx5/7965/22015/01023224.pdf?tp=\&arnumber=1023224\&isnumber=22015}, author = {Abd-Almageed, Wael and Fadali,M. S and Bebis,G.} } @conference {17886, title = {Persistent caching in a multiple query optimization framework}, booktitle = {Proceedings of the 6th Workshop on Languages, Compilers, and Run-Time Systems for Scalable Computers}, year = {2002}, month = {2002///}, abstract = {This paper focuses on persistent caching in multi-client en-vironments, which aims to improve the performance of a data server by caching on disk query results that can be expensive to produce. We present and evaluate extensions to an existing multi-query optimization framework, called MQO, to incorporate persistent caching capabilities. }, author = {Andrade,H. and Kurc, T. and Catalyurek,U. and Sussman, Alan and Saltz, J.} } @inbook {14248, title = {Polydioptric Cameras: New Eyes for Structure from Motion}, booktitle = {Pattern RecognitionPattern Recognition}, series = {Lecture Notes in Computer Science}, volume = {2449}, year = {2002}, month = {2002///}, pages = {618 - 625}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We examine the influence of camera design on the estimation of the motion and structure of a scene from video data. Every camera captures a subset of the light rays passing though some volume in space. By relating the differential structure of the time varying space of light rays to different known and new camera designs, we can establish a hierarchy of cameras. This hierarchy is based upon the stability and complexity of the computations necessary to estimate structure and motion. At the low end of this hierarchy is the standard planar pinhole camera for which the structure from motion problem is non-linear and ill-posed. At the high end is a camera, which we call the full field of view polydioptric camera, for which the problem is linear and stable. We develop design suggestions for the polydioptric camera, and based upon this new design we propose a linear algorithm for structure-from-motion estimation, which combines differential motion estimation with differential stereo.}, isbn = {978-3-540-44209-7}, url = {http://dx.doi.org/10.1007/3-540-45783-6_74}, author = {Neumann, Jan and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Van Gool,Luc} } @article {12045, title = {Polydioptric cameras: New eyes for structure from motion}, journal = {Pattern Recognition}, year = {2002}, month = {2002///}, pages = {618 - 625}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {11914, title = {A Probabilistic Clustering-Based Indoor Location Determination System}, journal = {Technical Reports from UMIACS, UMIACS-TR-2002-30}, year = {2002}, month = {2002/04/04/}, abstract = {We present an indoor location determination system based on signalstrength probability distributions for tackling the noisy wireless channel and clustering to reduce computation requirements. We provide two implementation techniques, namely, Joint Clustering and Incremental Triangulation and describe their tradeoffs in terms of location determination accuracy and computation requirement. Both techniques have been incorporated in two implemented context-aware systems: User Positioning System and the Rover System, both running on Compaq iPAQ Pocket PC{\textquoteright}s with Familiar distribution of Linux for PDA{\textquoteright}s. The results obtained show that both techniques give the user location with over 90\% accuracy to within 7 feet with very low computation requirements, hence enabling a set of context-aware applications. Also UMIACS-TR-2002-30 }, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/1192}, author = {Youssef,Moustafa A and Agrawala, Ashok K. and A. Udaya Shankar and Noh,Sam H} } @article {17865, title = {Processing large-scale multi-dimensional data in parallel and distributed environments}, journal = {Parallel Computing}, volume = {28}, year = {2002}, month = {2002/05//}, pages = {827 - 859}, abstract = {Analysis of data is an important step in understanding and solving a scientific problem. Analysis involves extracting the data of interest from all the available raw data in a dataset and processing it into a data product. However, in many areas of science and engineering, a scientist{\textquoteright}s ability to analyze information is increasingly becoming hindered by dataset sizes. The vast amount of data in scientific datasets makes it a difficult task to efficiently access the data of interest, and manage potentially heterogeneous system resources to process the data. Subsetting and aggregation are common operations executed in a wide range of data-intensive applications. We argue that common runtime and programming support can be developed for applications that query and manipulate large datasets. This paper presents a compendium of frameworks and methods we have developed to support efficient execution of subsetting and aggregation operations in applications that query and manipulate large, multi-dimensional datasets in parallel and distributed computing environments.}, keywords = {Data-intensive applications, Distributed computing, Multi-dimensional datasets, PARALLEL PROCESSING, Runtime systems}, isbn = {0167-8191}, doi = {10.1016/S0167-8191(02)00097-2}, url = {http://www.sciencedirect.com/science/article/pii/S0167819102000972}, author = {Beynon,Michael and Chang,Chialin and Catalyurek,Umit and Kurc,Tahsin and Sussman, Alan and Andrade,Henrique and Ferreira,Renato and Saltz,Joel} } @article {15612, title = {Ray interpolants for fast raytracing reflections and refractions}, journal = {Journal of WSCG (Proc. International Conf. in Central Europe on Comp. Graph., Visualization and Comp. Vision)}, volume = {10}, year = {2002}, month = {2002///}, pages = {1 - 8}, abstract = {To render an object by ray tracing, one or more rays are shot from the viewpoint through every pixel ofthe image plane. For reflective and refractive objects, especially for multiple levels of reflections and/or refractions, this requires many expensive intersection calculations. This paper presents a new method for accelerating ray-tracing of reflective and refractive objects by substituting accurate-but-slow intersection calculations with approximate-but-fast interpolation computations. Our approach is based on modeling the reflective/refractive object as a function that maps input rays entering the object to output rays exiting the object. We are interested in computing the output ray without actually tracing the input ray through the object. This is achieved by adaptively sampling rays from multiple viewpoints in various directions, as a preprocessing phase, and then interpolating the collection of nearby samples to compute an approximate output ray for any input ray. In most cases, object boundaries and other discontinuities are handled by ap- plying various heuristics. In cases where we cannot find sufficient evidence to interpolate, we perform ray tracing as a last resort. We provide performance studies to demonstrate the efficiency of this method. }, author = {Atalay,F. B and Mount, Dave} } @article {11913, title = {Rover: scalable location-aware computing}, journal = {Computer}, volume = {35}, year = {2002}, month = {2002/10//}, pages = {46 - 53}, abstract = {All the components necessary for realizing location-aware computing are available in the marketplace today. What has hindered the widespread deployment of location-based systems is the lack of an integration architecture that scales with user populations. The authors have completed the initial implementation of Rover, a system designed to achieve this sort of integration and to automatically tailor information and services to a mobile user{\textquoteright}s location. Their studies have validated Rover{\textquoteright}s underlying software architecture, which achieves system scalability through high-resolution, application-specific resource scheduling at the servers and network. The authors believe that this technology will greatly enhance the user experience in many places, including museums, amusement and theme parks, shopping malls, game fields, offices, and business centers. They designed the system specifically to scale to large user populations and expect its benefits to increase with them.}, keywords = {amusement, application-specific, architecture;, automation;, business, business;, computing;, data, entertainment;, handheld, humanities;, integration, LAN;, location-aware, malls;, mobile, museums;, office, parks;, processing;, resource, Rover;, scalability;, scalable, scheduling;, shopping, software, system, theme, units;, user;, wireless}, isbn = {0018-9162}, doi = {10.1109/MC.2002.1039517}, author = {Banerjee,S. and Agarwal,S. and Kamel,K. and Kochut, A. and Kommareddy,C. and Nadeem,T. and Thakkar,P. and Trinh,Bao and Youssef,A. and Youssef, M. and Larsen,R.L. and Udaya Shankar,A. and Agrawala, Ashok K.} } @conference {17859, title = {Scheduling multiple data visualization query workloads on a shared memory machine}, booktitle = {Parallel and Distributed Processing Symposium., Proceedings International, IPDPS 2002, Abstracts and CD-ROM}, year = {2002}, month = {2002///}, pages = {11 - 18}, publisher = {IEEE}, organization = {IEEE}, abstract = {Query scheduling plays an important role when systems are faced with limited resources and high workloads. It becomes even more relevant for servers applying multiple query optimization techniques to batches of queries, in which portions of datasets as well as intermediate results are maintained in memory to speed up query evaluation. We present a dynamic query scheduling model based on a priority queue implementation using a directed graph and a strategy for ranking queries. We examine the relative performance of several ranking strategies on a shared-memory machine using two different versions of an application, called the Virtual Microscope, for browsing digitized microscopy images}, keywords = {Atomic force microscopy, Biomedical informatics, Computer science, Data analysis, data visualisation, Data visualization, datasets, deductive databases, digitized microscopy image browsing, directed graph, directed graphs, dynamic query scheduling model, Educational institutions, high workloads, image database, limited resources, multiple data visualization query workloads, multiple query optimization, performance, priority queue, Processor scheduling, Query processing, query ranking, Relational databases, scheduling, shared memory machine, shared memory systems, Virtual Microscope, visual databases}, isbn = {0-7695-1573-8}, doi = {10.1109/IPDPS.2002.1015482}, author = {Andrade,H. and Kurc, T. and Sussman, Alan and Saltz, J.} } @conference {15643, title = {Space-efficient approximate Voronoi diagrams}, booktitle = {Proceedings of the thiry-fourth annual ACM symposium on Theory of computing}, series = {STOC {\textquoteright}02}, year = {2002}, month = {2002///}, pages = {721 - 730}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {(MATH) Given a set $S$ of $n$ points in $\IR^d$, a {\em $(t,\epsilon)$-approximate Voronoi diagram (AVD)} is a partition of space into constant complexity cells, where each cell $c$ is associated with $t$ representative points of $S$, such that for any point in $c$, one of the associated representatives approximates the nearest neighbor to within a factor of $(1+\epsilon)$. Like the Voronoi diagram, this structure defines a spatial subdivision. It also has the desirable properties of being easy to construct and providing a simple and practical data structure for answering approximate nearest neighbor queries. The goal is to minimize the number and complexity of the cells in the AVD.(MATH) We assume that the dimension $d$ is fixed. Given a real parameter $\gamma$, where $2 \le \gamma \le 1/\epsilon$, we show that it is possible to construct a $(t,\epsilon)$-AVD consisting of \[O(n \epsilon^{\frac{d-1}{2}} \gamma^{\frac{3(d-1)}{2}} \log \gamma) \] cells for $t = O(1/(\epsilon \gamma)^{(d-1)/2})$. This yields a data structure of $O(n \gamma^{d-1} \log \gamma)$ space (including the space for representatives) that can answer $\epsilon$-NN queries in time $O(\log(n \gamma) + 1/(\epsilon \gamma)^{(d-1)/2})$. (Hidden constants may depend exponentially on $d$, but do not depend on $\epsilon$ or $\gamma$).(MATH) In the case $\gamma = 1/\epsilon$, we show that the additional $\log \gamma$ factor in space can be avoided, and so we have a data structure that answers $\epsilon$-approximate nearest neighbor queries in time $O(\log (n/\epsilon))$ with space $O(n/\epsilon^{d-1})$, improving upon the best known space bounds for this query time. In the case $\gamma = 2$, we have a data structure that can answer approximate nearest neighbor queries in $O(\log n + 1/\epsilon^{(d-1)/2})$ time using optimal $O(n)$ space. This dramatically improves the previous best space bound for this query time by a factor of $O(1/\epsilon^{(d-1)/2})$.(MATH) We also provide lower bounds on the worst-case number of cells assuming that cells are axis-aligned rectangles of bounded aspect ratio. In the important extreme cases $\gamma \in \{2, 1/\epsilon\}$, our lower bounds match our upper bounds asymptotically. For intermediate values of $\gamma$ we show that our upper bounds are within a factor of $O((1/\epsilon)^{(d-1)/2}\log \gamma)$ of the lower bound.}, isbn = {1-58113-495-9}, doi = {10.1145/509907.510011}, url = {http://doi.acm.org/10.1145/509907.510011}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @article {12032, title = {Spatio-temporal stereo using multi-resolution subdivision surfaces}, journal = {International Journal of Computer Vision}, volume = {47}, year = {2002}, month = {2002///}, pages = {181 - 193}, author = {Neumann, J. and Aloimonos, J.} } @article {15935, title = {Symbol systems}, journal = {Encyclopedia of Cognitive Science}, year = {2002}, month = {2002///}, author = {Anderson,M. L and Perlis, Don} } @conference {15972, title = {Time-situated agency: Active logic and intention formation}, booktitle = {in: Workshop on Cognitive Agents, 25th German Conference on Artificial Intelligence}, year = {2002}, month = {2002///}, author = {Anderson,M. L and Josyula,D. P and Okamoto,Y. A and Perlis, Don} } @conference {18604, title = {Topology inference from BGP routing dynamics}, booktitle = {Proceedings of the 2nd ACM SIGCOMM Workshop on Internet measurment}, series = {IMW {\textquoteright}02}, year = {2002}, month = {2002///}, pages = {243 - 248}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {This paper describes a method of inferring logical relationships between network prefixes within an Autonomous System (AS) using only passive monitoring of BGP messages. By clustering these prefixes based upon similarities between their update times, we create a hierarchy linking the prefixes within the larger AS. We can frequently identify groups of prefixes routed to the same ISP Point of Presence (POP), despite the lack of identifying information in the BGP messages. Similarly, we observe disparate prefixes under common organizational control, or with long shared network paths. In addition to discovering interesting network characteristics, our passive method facilitates topology discovery by potentially reducing the number of active probes required in traditional traceroute-based Internet mapping mechanisms.}, isbn = {1-58113-603-X}, doi = {10.1145/637201.637239}, url = {http://doi.acm.org/10.1145/637201.637239}, author = {Andersen,David G. and Feamster, Nick and Bauer,Steve and Balakrishnan,Hari} } @conference {15955, title = {The use-mention distinction and its importance to HCI}, booktitle = {Proceedings of the Sixth Workshop on the Semantics and Pragmatics of Dialog}, year = {2002}, month = {2002///}, pages = {21 - 28}, author = {Anderson,M. L and Okamoto,Y. and Josyula,D. and Perlis, Don} } @article {12073, title = {Using independent auditors as intrusion detection systems}, journal = {Information and Communications Security}, year = {2002}, month = {2002///}, pages = {291 - 302}, author = {Molina,J. and Arbaugh, William A.} } @article {12060, title = {Visual space-time geometry - A tool for perception and the imagination}, journal = {Proceedings of the IEEE}, volume = {90}, year = {2002}, month = {2002/07//}, pages = {1113 - 1135}, abstract = {Although the fundamental ideas underlying research efforts in the field of computer vision have not radically changed in the past two decades, there has been a transformation in the way work in this field is conducted. This is primarily due to the emergence of a number of tools, of both a practical and a theoretical nature. One such tool, celebrated throughout the nineties, is the geometry of visual space-time. It is known under a variety of headings, such as multiple view geometry, structure from motion, and model building. It is a mathematical theory relating multiple views (images) of a scene taken at different viewpoints to three-dimensional models of the (possibly dynamic) scene. This mathematical theory gave rise to algorithms that take as input images (or video) and provide as output a model of the scene. Such algorithms are one of the biggest successes of the field and they have many applications in other disciplines, such as graphics (image-based rendering, motion capture) and robotics (navigation). One of the difficulties, however is that the current tools cannot yet be fully automated, and they do not provide very accurate results. More research is required for automation and high precision. During the past few years we have investigated a number of basic questions underlying the structure from motion problem. Our investigations resulted in a small number of principles that characterize the problem. These principles, which give rise to automatic procedures and point to new avenues for studying the next level of the structure from motion problem, are the subject of this paper.}, keywords = {3-D motion estimation, Buildings, Computer vision, Geometry, Graphics, Image sequences, Layout, Mathematical model, mathematical theory, model building, Motion estimation, multiple view geometry, multiple views, Navigation, optical flow, optical illusions, patch correspondence, Rendering (computer graphics), Robotics and automation, Solid modeling, structure from motion, three-dimensional models, visual space-time}, isbn = {0018-9219}, doi = {10.1109/JPROC.2002.801440}, author = {Ferm{\"u}ller, Cornelia and Baker, P. and Aloimonos, J.} } @article {12087, title = {Your 802.11 network has no clothes}, journal = {IEEE Communications Magazine}, volume = {9}, year = {2002}, month = {2002///}, pages = {44 - 51}, author = {Mishra,A. and Shin,M. and Arbaugh, William A.} } @article {14198, title = {Animated heads: From 3d motion fields to action descriptions}, journal = {Proceedings of the IFIP TC5/WG5}, volume = {10}, year = {2001}, month = {2001///}, pages = {1 - 11}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {18815, title = {Automated design of sheet metal punches for bending multiple parts in a single setup}, journal = {Robotics and Computer-Integrated Manufacturing}, volume = {17}, year = {2001}, month = {2001/02//}, pages = {33 - 47}, abstract = {Sheet metal bending is a process in which bends are formed using a combination of a punch and a die. A very large number of mechanical products such as furniture panels, shelves, cabinets, housing for electro-mechanical devices, etc. are created by using the sheet metal bending process. Bending tools need to satisfy the following two criteria: (1) tools should be able to withstand bending forces, and (2) tool shapes should be such that there is no tool-part interference. In this paper, we describe a methodology for automatically designing shapes of bending punches for bending multiple parts in a single setup. We create parametric geometric models of punches. These parametric models describe the family of possible punch shapes. Using the part geometry and parametric punch shape models, we automatically generated constraints on tool parameters that eliminate the possibility of part-tool interference. We use mixed-integer techniques to identify parameters of punch shapes that result in the maximum punch strength. Finally, we perform strength analysis of the designed punch shape using finite element analysis methods, to verify that the designed punch shape is capable of withstanding the bending forces.}, keywords = {Praess planning, Punch design, Sheet metal bending}, isbn = {0736-5845}, doi = {10.1016/S0736-5845(00)00035-1}, url = {http://www.sciencedirect.com/science/article/pii/S0736584500000351}, author = {Alva,Ujval and Gupta, Satyandra K.} } @book {16985, title = {Becoming a Scholar in the Digital Age}, year = {2001}, month = {2001///}, publisher = {National Research Council}, organization = {National Research Council}, keywords = {bibtex-import, digital-library, dlbook}, author = {Duderstadt,James and Arms,William and Messina,Paul and Ellisman,Mark and Atkins,Daniel and Fox,Edward and Shneiderman, Ben and Nissenbaum,Helen and Lederberg,Joshua} } @conference {15645, title = {Entropy-preserving cuttings and space-efficient planar point location}, booktitle = {Proceedings of the twelfth annual ACM-SIAM symposium on Discrete algorithms}, series = {SODA {\textquoteright}01}, year = {2001}, month = {2001///}, pages = {256 - 261}, publisher = {Society for Industrial and Applied Mathematics}, organization = {Society for Industrial and Applied Mathematics}, address = {Philadelphia, PA, USA}, abstract = {Point location is the problem of preprocessing a planar polygonal subdivision S into a data structure in order to determine efficiently the cell of the subdivision that contains a given query point. Given the probabilities pz that the query point lies within each cell z ∈ S, a natural question is how to design such a structure so as to minimize the expected-case query time. The entropy H of the probability distribution is the dominant term in the lower bound on the expected-case search time. Clearly the number of edges n of the subdivision is a lower bound on the space required. There is no known approach that simultaneously achieves the goals of H + \&Ogr;(H) query time and \&Ogr;(n) space. In this paper we introduce entropy-preserving cuttings and show how to use them to achieve query time H + \&Ogr;(H), using only \&Ogr;(n log* n) space.}, isbn = {0-89871-490-7}, url = {http://dl.acm.org/citation.cfm?id=365411.365456}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @conference {16647, title = {Evolving columnar circuitry for lateral cortical inhibition}, booktitle = {Neural Networks, 2001. Proceedings. IJCNN{\textquoteright}01. International Joint Conference on}, volume = {1}, year = {2001}, month = {2001///}, pages = {278 - 283}, author = {Ayers,D. and Reggia, James A.} } @inbook {14217, title = {Eyes from Eyes}, booktitle = {3D Structure from Images {\textemdash} SMILE 20003D Structure from Images {\textemdash} SMILE 2000}, series = {Lecture Notes in Computer Science}, volume = {2018}, year = {2001}, month = {2001///}, pages = {204 - 217}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We describe a family of new imaging systems, called Argus eyes, that consist of common video cameras arranged in some network. The system we built consists of six cameras arranged so that they sample different parts of the visual sphere. This system has the capability of very accurately estimating its own 3D motion and consequently estimating shape models from the individual videos. The reason is that inherent ambiguities of confusion between translation and rotation disappear in this case. We provide an algorithm and several experiments using real outdoor or indoor images demonstrating the superiority of the new sensor with regard to 3D motion estimation.}, isbn = {978-3-540-41845-0}, url = {http://dx.doi.org/10.1007/3-540-45296-6_14}, author = {Baker,Patrick and Pless,Robert and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Pollefeys,Marc and Van Gool,Luc and Zisserman,Andrew and Fitzgibbon,Andrew} } @book {14218, title = {Eyes from Eyes: Analysis of Camera Design Using Plenoptic Video Geometry}, year = {2001}, month = {2001///}, publisher = {Computer Vision Laboratory, Center for Automation Research, University of Maryland}, organization = {Computer Vision Laboratory, Center for Automation Research, University of Maryland}, abstract = {We investigate the relationship between camera design and the problem of recovering the motion andstructure of a scene from video data. The visual information that could possibly be obtained is described by the plenoptic function. A camera can be viewed as a device that captures a subset of this function, that is, it measures some of the light rays in some part of the space. The information contained in the subset determines how difficult it is to solve subsequent interpretation processes. By examining the differential structure of the time varying plenoptic function we relate different known and new camera models to the spatio-temporal structure of the observed scene. This allows us to define a hierarchy of camera designs, where the order is determined by the stability and complexity of the computations necessary to estimate structure and motion. At the low end of this hierarchy is the standard planar pinhole camera for which the structure from motion problem is non-linear and ill-posed. At the high end is a camera, which we call the full field of view polydioptric camera, for which the problem is linear and stable. In between are multiple-view cameras with large fields of view which we have built, as well as catadioptric panoramic sensors and other omni-directional cameras. We develop design suggestions for the polydioptric camera, and based upon this new design we propose a linear algorithm for ego- motion estimation, which in essence combines differential motion estimation with differential stereo. }, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J. and University of Maryland (College Park, Md.). Center for Automation Research} } @inbook {14221, title = {Geometry of Eye Design: Biology and Technology}, booktitle = {Multi-Image Analysis}, series = {Lecture Notes in Computer Science}, volume = {2032}, year = {2001}, month = {2001///}, pages = {22 - 38}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Natural or artificial vision systems process the images that they collect with their eyes or cameras in order to derive information for performing tasks related to navigation and recognition. Since the way images are acquired determines how dificult it is to perform a visual task, and since systems have to cope with limited resources, the eyes used by a specific system should be designed to optimize subsequent image processing as it relates to particular tasks. Different ways of sampling light, i.e., different eyes, may be less or more powerful with respect to particular competences. This seems intuitively evident in view of the variety of eye designs in the biological world. It is shown here that a spherical eye (an eye or system of eyes providing panoramic vision) is superior to a camera-type eye (an eye with restricted field of view) as regards the competence of three-dimensional motion estimation. This result is derived from a statistical analysis of all the possible computational models that can be used for estimating 3D motion from an image sequence. The findings explain biological design in a mathematical manner, by showing that systems that fly and thus need good estimates of 3D motion gain advantages from panoramic vision. Also, insights obtained from this study point to new ways of constructing powerful imaging devices that suit particular tasks in robotics, visualization and virtual reality better than conventional cameras, thus leading to a new camera technology.}, isbn = {978-3-540-42122-1}, url = {http://dx.doi.org/10.1007/3-540-45134-X_2}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Klette,Reinhard and Gimel{\textquoteright}farb,Georgy and Huang,Thomas} } @article {11916, title = {Operating Systems}, journal = {Real-Time Systems}, year = {2001}, month = {2001///}, author = {Agrawala, Ashok K.} } @conference {12074, title = {Personal secure booting}, booktitle = {Information Security and Privacy}, year = {2001}, month = {2001///}, pages = {130 - 144}, author = {Itoi,N. and Arbaugh, William A. and Pollack,S. and Reeves,D.} } @article {12070, title = {The price of safety in an active network}, journal = {Journal of Communications and Networks}, volume = {3}, year = {2001}, month = {2001///}, pages = {4 - 18}, author = {Alexander,D. S and Menage,P. B and Keromytis,A. D and Arbaugh, William A. and Anagnostakis,K. G and Smith,J. M} } @article {17828, title = {Probabilistic temporal databases}, journal = {ACM Transaction on Database Systems}, volume = {26}, year = {2001}, month = {2001///}, pages = {41 - 95}, author = {Alex,D. and Robert,R. and V.S. Subrahmanian} } @article {11917, title = {Revolutionary Advances in Ubiquitious, Real-Time Multicomputers and Runtime Environments}, volume = {A340293}, year = {2001}, month = {2001///}, pages = {225 - 225}, institution = {MISSISSIPPI STATE UNIVERSITY}, abstract = {This work was a grant to enhance the Maruti operating system in several ways, in order to provide Mississippi State with a platform upon which their work on the Real-Time Message Passing Interface could be developed. Key technical achievements: (1) Developed predictable Myrinet communications for use in a real-time NOW; (2) Developed the MSU-Kernel to provide a POSIX OS for real- time NOWs; (3) Developed and implemented an algorithm for deploying a globally synchronized clock in a real-time NOW; (4) Developed an improved real-time scheduler for the Maruti hard real-time operating system at University of Maryland (UMD); and (5) Introduced a new parametric approach in Maruti for dynamic scheduling at UMD. Details of the results of the work are presented in papers, thesis and project reports.}, author = {Agrawala, Ashok K.} } @conference {15644, title = {A simple entropy-based algorithm for planar point location}, booktitle = {Proceedings of the twelfth annual ACM-SIAM symposium on Discrete algorithms}, series = {SODA {\textquoteright}01}, year = {2001}, month = {2001///}, pages = {262 - 268}, publisher = {Society for Industrial and Applied Mathematics}, organization = {Society for Industrial and Applied Mathematics}, address = {Philadelphia, PA, USA}, abstract = {Given a planar polygonal subdivision S, point location involves preprocessing this subdivision into a data structure so that given any query point q, the cell of the subdivision containing q can be determined efficiently. Suppose that for each cell z in the subdivision, the probability pz that a query point lies within this cell is also given. The goal is to design the data structure to minimize the average search time. It has long been known that the entropy H of the probability distribution is the dominant term in the lower bound on the average-case search time. This problem has been considered before, but existing data structures are all quite complicated. In this paper, we show that a very simple modification of a well-known randomized incremental algorithm can be applied to produce a data structure of expected linear size that can answer point location queries in \&Ogr;(H) average time. We also present empirical evidence for the practical efficiency of this approach.}, isbn = {0-89871-490-7}, url = {http://dl.acm.org/citation.cfm?id=365411.365457}, author = {Arya,Sunil and Malamatos,Theocharis and Mount, Dave} } @conference {14188, title = {A spherical eye from multiple cameras (makes better models of the world)}, booktitle = {Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2001. CVPR 2001}, volume = {1}, year = {2001}, month = {2001///}, pages = {I-576- I-583 vol.1 - I-576- I-583 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {The paper describes an imaging system that has been designed specifically for the purpose of recovering egomotion and structure from video. The system consists of six cameras in a network arranged so that they sample different parts of the visual sphere. This geometric configuration has provable advantages compared to small field of view cameras for the estimation of the system{\textquoteright}s own motion and consequently the estimation of shape models from the individual cameras. The reason is that inherent ambiguities of confusion between translation and rotation disappear. We provide algorithms for the calibration of the system and 3D motion estimation. The calibration is based on a new geometric constraint that relates the images of lines parallel in space to the rotation between the cameras. The 3D motion estimation uses a constraint relating structure directly to image gradients.}, keywords = {3D motion estimation, Calibration, camera network, CAMERAS, Computer vision, egomotion recovery, geometric configuration, geometric constraint, image gradients, image sampling, imaging system, Laboratories, Layout, Motion estimation, multiple cameras, Pixel, Robot vision systems, SHAPE, shape models, Space technology, spherical eye, system calibration, video, video cameras, video signal processing, visual sphere sampling}, isbn = {0-7695-1272-0}, doi = {10.1109/CVPR.2001.990525}, author = {Baker, P. and Ferm{\"u}ller, Cornelia and Aloimonos, J. and Pless, R.} } @inbook {14261, title = {Statistics Explains Geometrical Optical Illusions}, booktitle = {Foundations of Image UnderstandingFoundations of Image Understanding}, series = {The Kluwer International Series in Engineering and Computer Science}, volume = {628}, year = {2001}, month = {2001///}, pages = {409 - 445}, publisher = {Springer US}, organization = {Springer US}, abstract = {Azriel Rosenfeld has been our mentor for the last decade. In everyday conversation he stressed the importance of intuition, the unreasonable effectiveness of mathematics in understanding the world, and the power of simplicity inherent in deep ideas. Because we mostly worked on 3D vision, he often argued that 2D vision has been, is, and will continue to be a large source of problems. Strangely enough, we arrived at this study, which is our first on 2D vision, through problems we encountered in our work in 3D motion and shape. Azriel Rosenfeld was also one of the first to apply statistics [ 27 , 28 ] to image analysis and understanding, and he always reminded us of the uncertainties involved in visual computations. This paper shows that statistics cannot be ignored, not even in the interpretation of two simple straight intersecting lines. It demonstrates that uncertainty in the visual data causes problems for the early visual processes. Because of noise, the estimation of features, such as lines, intersections of lines, and local image motion, is biased. The inevitability of this bias provides an explanation for many well-known geometrical optical illusions.}, isbn = {978-1-4615-1529-6}, url = {http://dx.doi.org/10.1007/978-1-4615-1529-6_14}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Davis, Larry S.} } @article {14170, title = {The Statistics of Optical Flow}, journal = {Computer Vision and Image Understanding}, volume = {82}, year = {2001}, month = {2001/04//}, pages = {1 - 32}, abstract = {When processing image sequences some representation of image motion must be derived as a first stage. The most often used representation is the optical flow field, which is a set of velocity measurements of image patterns. It is well known that it is very difficult to estimate accurate optical flow at locations in an image which correspond to scene discontinuities. What is less well known, however, is that even at the locations corresponding to smooth scene surfaces, the optical flow field often cannot be estimated accurately.Noise in the data causes many optical flow estimation techniques to give biased flow estimates. Very often there is consistent bias: the estimate tends to be an underestimate in length and to be in a direction closer to the majority of the gradients in the patch. This paper studies all three major categories of flow estimation methods{\textemdash}gradient-based, energy-based, and correlation methods, and it analyzes different ways of compounding one-dimensional motion estimates (image gradients, spatiotemporal frequency triplets, local correlation estimates) into two-dimensional velocity estimates, including linear and nonlinear methods. Correcting for the bias would require knowledge of the noise parameters. In many situations, however, these are difficult to estimate accurately, as they change with the dynamic imagery in unpredictable and complex ways. Thus, the bias really is a problem inherent to optical flow estimation. We argue that the bias is also integral to the human visual system. It is the cause of the illusory perception of motion in the Ouchi pattern and also explains various psychophysical studies of the perception of moving plaids. }, isbn = {1077-3142}, doi = {10.1006/cviu.2000.0900}, url = {http://www.sciencedirect.com/science/article/pii/S1077314200909007}, author = {Ferm{\"u}ller, Cornelia and Shulman,David and Aloimonos, J.} } @conference {12021, title = {Towards the ultimate motion capture technology}, booktitle = {Deformable avatars: IFIP TC5/WG5. 10 DEFORM{\textquoteright}2000 Workshop, November 29-30, 2000, Geneva, Switzerland and AVATARS{\textquoteright}2000 Workshop, November 30-December 1, 2000, Lausanne, Switzerland}, volume = {68}, year = {2001}, month = {2001///}, pages = {143 - 143}, author = {Stuart, B. and Baker, P. and Aloimonos, J.} } @conference {12098, title = {A trend analysis of exploitations}, booktitle = {2001 IEEE Symposium on Security and Privacy, 2001. S\&P 2001. Proceedings}, year = {2001}, month = {2001///}, pages = {214 - 229}, publisher = {IEEE}, organization = {IEEE}, abstract = {We have conducted an empirical study of a number of computer security exploits and determined that the rates at which incidents involving the exploit are reported to CERT can be modeled using a common mathematical framework. Data associated with three significant exploits involving vulnerabilities in phf, imap, and bind can all be modeled using the formula C=I+S{\texttimes}√M where C is the cumulative count of reported incidents, M is the time since the start of the exploit cycle, and I and S are the regression coefficients determined by analysis of the incident report data. Further analysis of two additional exploits involving vulnerabilities in mountd and statd confirm the model. We believe that the models will aid in predicting the severity of subsequent vulnerability exploitations, based on the rate of early incident reports}, keywords = {Computer science, computer security exploits, Data analysis, data mining, Educational institutions, exploitations, Performance analysis, Predictive models, Regression analysis, Risk management, security of data, software engineering, system intrusions, System software, trend analysis, vulnerabilities, vulnerability exploitation}, isbn = {0-7695-1046-9}, doi = {10.1109/SECPRI.2001.924300}, author = {Browne,H. K and Arbaugh, William A. and McHugh,J. and Fithen,W. L} } @inbook {14197, title = {Analyzing Action Representations}, booktitle = {Algebraic Frames for the Perception-Action CycleAlgebraic Frames for the Perception-Action Cycle}, series = {Lecture Notes in Computer Science}, volume = {1888}, year = {2000}, month = {2000///}, pages = {1 - 21}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We argue that actions represent the basic seed of intelligence underlying perception of the environment, and the representations encoding actions should be the starting point upon which further studies of cognition are built. In this paper we make a first effort in characterizing these action representations. In particular, from the study of simple actions related to 3D rigid motion interpretation, we deduce a number of principles for the possible computations responsible for the interpretation of space-time geometry. Using these principles, we then discuss possible avenues on how to proceed in analyzing the representations of more complex human actions.}, isbn = {978-3-540-41013-3}, url = {http://dx.doi.org/10.1007/10722492_1}, author = {Aloimonos, J. and Ferm{\"u}ller, Cornelia}, editor = {Sommer,Gerald and Zeevi,Yehoshua} } @article {15563, title = {Approximate range searching}, journal = {Computational Geometry}, volume = {17}, year = {2000}, month = {2000/12//}, pages = {135 - 152}, abstract = {The range searching problem is a fundamental problem in computational geometry, with numerous important applications. Most research has focused on solving this problem exactly, but lower bounds show that if linear space is assumed, the problem cannot be solved in polylogarithmic time, except for the case of orthogonal ranges. In this paper we show that if one is willing to allow approximate ranges, then it is possible to do much better. In particular, given a bounded range Q of diameter w and ε\>0, an approximate range query treats the range as a fuzzy object, meaning that points lying within distance εw of the boundary of Q either may or may not be counted. We show that in any fixed dimension d, a set of n points in Rd can be preprocessed in O(n+logn) time and O(n) space, such that approximate queries can be answered in O(logn(1/ε)d) time. The only assumption we make about ranges is that the intersection of a range and a d-dimensional cube can be answered in constant time (depending on dimension). For convex ranges, we tighten this to O(logn+(1/ε)d-1) time. We also present a lower bound for approximate range searching based on partition trees of Ω(logn+(1/ε)d-1), which implies optimality for convex ranges (assuming fixed dimensions). Finally, we give empirical evidence showing that allowing small relative errors can significantly improve query execution times.}, keywords = {Approximation algorithms, Box-decomposition trees, Partition trees, Range searching}, isbn = {0925-7721}, doi = {10.1016/S0925-7721(00)00022-5}, url = {http://www.sciencedirect.com/science/article/pii/S0925772100000225}, author = {Arya,Sunil and Mount, Dave} } @conference {18789, title = {Automated design of sheet metal bending tools}, year = {2000}, month = {2000///}, abstract = {Sheet metal bending is a process in which bends are formed using a combination of apunch and a die. Bending tools need to satisfy the following two criteria: (1) tools should be able to withstand bending forces, and (2) tool shapes should be such that there is no tool-part interference. In this paper, we describe a systematic methodology for automatically synthesizing a punch shape that can be used to produce more than one type of part. We create a parametric geometric model of possible punches. This parametric model describes a family of possible punch shapes. Using the geometric models of various given part types and a parametric punch model, we automatically generate constraints on punch parameters that eliminate the possibility of part-tool interference. We use mixed integer programming techniques to solve these constraints and identify parameters of a punch shape that can work for the given set of parts. Finally, we perform strength analysis of the designed punch to verify that the designed punch is capable of withstanding the bending forces. }, url = {http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.34.5031\&rep=rep1\&type=pdf}, author = {Alva,U. and Gupta,S.K. and Gupta,R. K.} } @conference {16617, title = {Cortical inhibition as explained by the competitive distribution hypothesis}, booktitle = {Network models for control and processing}, year = {2000}, month = {2000///}, pages = {31 - 62}, author = {Reggia, James A. and Sutton III,G. G. and Lynne,C. and D{\textquoteright}Autrechy,S. C and Armentrout,S. L} } @conference {13893, title = {Designing StoryRooms: interactive storytelling spaces for children}, booktitle = {Proceedings of the 3rd conference on Designing interactive systems: processes, practices, methods, and techniques}, year = {2000}, month = {2000///}, pages = {95 - 104}, author = {Alborzi,H. and Druin, Allison and Montemayor,J. and Platner,M. and Porteous,J. and Sherman,L. and Boltman,A. and Tax{\'e}n,G. and Best,J. and Hammer,J. and others} } @article {12027, title = {Detecting independent motion: The statistics of temporal continuity}, journal = {Pattern Analysis and Machine Intelligence, IEEE Transactions on}, volume = {22}, year = {2000}, month = {2000///}, pages = {768 - 773}, author = {Pless, R. and Brodsky, T. and Aloimonos, J.} } @conference {11918, title = {A dual interpretation of {\textquotedblleft}standard constraints{\textquotedblright} in parametric scheduling}, booktitle = {Formal Techniques in Real-Time and Fault-Tolerant Systems}, year = {2000}, month = {2000///}, pages = {205 - 223}, abstract = {Parametric scheduling in real-time systems, in the presence of linear relative constraints between the start and execution times of tasks, is a well-studied problem. Prior research established the existence of polynomial time algorithms for the case when the constraints are restricted to be standard and the execution time vectors belong to an axis-parallel hyper-rectangle. In this paper we present a polynomial time algorithm for the case when the execution time vectors belong to arbitrary convex domains. Our insights into the problem occur primarily as a result of studying the dual polytope of the constraint system.}, author = {Subramani,K. and Agrawala, Ashok K.} } @article {15547, title = {Efficient expected-case algorithms for planar point location}, journal = {Algorithm Theory-SWAT 2000}, year = {2000}, month = {2000///}, pages = {537 - 543}, doi = {10.1007/3-540-44985-X_31}, author = {Arya,S. and Cheng,S. W and Mount, Dave and Ramesh,H.} } @article {17907, title = {Efficiently computing and updating triangle strips for real-time rendering}, journal = {Computer Aided Design}, volume = {32}, year = {2000}, month = {2000///}, pages = {753 - 772}, abstract = {Triangle strips are a widely used hardware-supported data-structure to compactly represent and efficiently render polygonal meshes. In thispaper we survey the efficient generation of triangle strips as well as their variants. We present efficient algorithms for partitioning polygonal meshes into triangle strips. Triangle strips have traditionally used a buffer size of two vertices. In this paper we also study the impact of larger buffer sizes and various queuing disciplines on the effectiveness of triangle strips. View-dependent simplification has emerged as a powerful tool for graphics acceleration in visualization of complex environments. However, in a view-dependent framework the triangle mesh connectivity changes at every frame making it difficult to use triangle strips. In this paper we present a novel data-structure, Skip Strip, that efficiently maintains triangle strips during such view-dependent changes. A Skip Strip stores the vertex hierarchy nodes in a skip-list-like manner with path compression. We anticipate that Skip Strips will provide a road-map to combine rendering acceleration techniques for static datasets, typical of retained-mode graphics applications, with those for dynamic datasets found in immediate-mode applications. }, author = {El-Sana,J. and Evans,F. and Kalaiah,A. and Varshney, Amitabh and Skiena,S. and Azanli,E.} } @conference {14287, title = {A first step towards automated detection of buffer overrun vulnerabilities}, booktitle = {The 2000 Network and Distributed Systems Security Conference. San Diego, CA}, volume = {14}, year = {2000}, month = {2000///}, author = {Wagner,D. and Foster, Jeffrey S. and Brewer,E.A. and Aiken,A.} } @article {18960, title = {The genome sequence of Drosophila melanogaster}, journal = {Science}, volume = {287}, year = {2000}, month = {2000///}, pages = {2185 - 2185}, abstract = {The fly Drosophila melanogaster is one of the most intensively studied organisms in biology and serves as a model system for the investigation of many developmental and cellular processes common to higher eukaryotes, including humans. We have determined the nucleotide sequence of nearly all of the \~{}120-megabase euchromatic portion of theDrosophila genome using a whole-genome shotgun sequencing strategy supported by extensive clone-based sequence and a high-quality bacterial artificial chromosome physical map. Efforts are under way to close the remaining gaps; however, the sequence is of sufficient accuracy and contiguity to be declared substantially complete and to support an initial analysis of genome structure and preliminary gene annotation and interpretation. The genome encodes \~{}13,600 genes, somewhat fewer than the smaller Caenorhabditis elegansgenome, but with comparable functional diversity.}, isbn = {0036-8075, 1095-9203}, doi = {10.1126/science.287.5461.2185}, url = {http://www.sciencemag.org/content/287/5461/2185}, author = {Adams,M.D. and Celniker,S.E. and Holt,R.A. and Evans,C.A. and Gocayne,J.D. and Amanatides,P.G. and Scherer,S.E. and Li,P.W. and Hoskins,R.A. and Galle,R.F. and others} } @article {15370, title = {A model for magnetic aftereffect in the presence of time varying demagnetizing fields}, journal = {Magnetics, IEEE Transactions on}, volume = {36}, year = {2000}, month = {2000/09//}, pages = {3182 - 3184}, abstract = {Preisach models driven by stochastic inputs are employed to model magnetic aftereffect in the presence of demagnetizing fields. As a result of thermal relaxation, the magnetization and, consequently, the demagnetizing field vary in time. The new model generalizes earlier work and self-consistently accounts for the effect of time varying demagnetizing fields on the relaxation process. Three practical numerical techniques are proposed to compute the time variation in the magnetic field and the state of the Preisach plane}, keywords = {aftereffect;, aftereffect;magnetization;numerical, demagnetizing, field;demagnetisation;magnetic, model;magnetic, Preisach, relaxation;time, simulation;thermal, varying}, isbn = {0018-9464}, doi = {10.1109/20.908729}, author = {Korman,C.E. and Adly,A.A. and Mayergoyz, Issak D and Rugkwamsook,P.} } @conference {14232, title = {Multi-camera networks: eyes from eyes}, booktitle = {IEEE Workshop on Omnidirectional Vision, 2000. Proceedings}, year = {2000}, month = {2000///}, pages = {11 - 18}, publisher = {IEEE}, organization = {IEEE}, abstract = {Autonomous or semi-autonomous intelligent systems, in order to function appropriately, need to create models of their environment, i.e., models of space time. These are descriptions of objects and scenes and descriptions of changes of space over time, that is, events and actions. Despite the large amount of research on this problem, as a community we are still far from developing robust descriptions of a system{\textquoteright}s spatiotemporal environment using video input (image sequences). Undoubtedly, some progress has been made regarding the understanding of estimating the structure of visual space, but it has not led to solutions to specific applications. There is, however, an alternative approach which is in line with today{\textquoteright}s {\textquotedblleft}zeitgeist.{\textquotedblright} The vision of artificial systems can be enhanced by providing them with new eyes. If conventional video cameras are put together in various configurations, new sensors can be constructed that have much more power and the way they {\textquotedblleft}see{\textquotedblright} the world makes it much easier to solve problems of vision. This research is motivated by examining the wide variety of eye design in the biological world and obtaining inspiration for an ensemble of computational studies that relate how a system sees to what that system does (i.e. relating perception to action). This, coupled with the geometry of multiple views that has flourished in terms of theoretical results in the past few years, points to new ways of constructing powerful imaging devices which suit particular tasks in robotics, visualization, video processing, virtual reality and various computer vision applications, better than conventional cameras. This paper presents a number of new sensors that we built using common video cameras and shows their superiority with regard to developing models of space and motion}, keywords = {Biosensors, CAMERAS, Computer vision, Eyes, Image sequences, intelligent systems, Layout, Machine vision, Robot vision systems, Robustness, Spatiotemporal phenomena, video cameras, Virtual reality}, isbn = {0-7695-0704-2}, doi = {10.1109/OMNVIS.2000.853797}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J. and Baker, P. and Pless, R. and Neumann, J. and Stuart, B.} } @conference {15589, title = {Nearly optimal expected-case planar point location}, booktitle = {Foundations of Computer Science, 2000. Proceedings. 41st Annual Symposium on}, year = {2000}, month = {2000///}, pages = {208 - 218}, abstract = {We consider the planar point location problem from the perspective of expected search time. We are given a planar polygonal subdivision S and for each polygon of the subdivision the probability that a query point lies within this polygon. The goal is to compute a search structure to determine which cell of the subdivision contains a given query point, so as to minimize the expected search time. This is a generalization of the classical problem of computing an optimal binary search tree for one-dimensional keys. In the one-dimensional case it has long been known that the entropy H of the distribution is the dominant term in the lower bound on the expected-case search time, and further there exist search trees achieving expected search times of at most H+2. Prior to this work, there has been no known structure for planar point location with an expected search time better than 2H, and this result required strong assumptions on the nature of the query point distribution. Here we present a data structure whose expected search time is nearly equal to the entropy lower bound, namely H+o(H). The result holds for any polygonal subdivision in which the number of sides of each of the polygonal cells is bounded, and there are no assumptions on the query distribution within each cell. We extend these results to subdivisions with convex cells, assuming a uniform query distribution within each cell}, keywords = {computational geometry, convex cells, data structure, expected search time, nearly optimal expected-case planar point location, optimal binary search tree, planar point location, planar polygonal subdivision, polygonal cells, polygonal subdivision, probability, search problems, search structure, subdivision, trees (mathematics)}, doi = {10.1109/SFCS.2000.892108}, author = {Arya,S. and Malamatos,T. and Mount, Dave} } @article {14234, title = {New eyes for building models from video}, journal = {Computational Geometry}, volume = {15}, year = {2000}, month = {2000/02//}, pages = {3 - 23}, abstract = {Models of real-world objects and actions for use in graphics, virtual and augmented reality and related fields can only be obtained through the use of visual data and particularly video. This paper examines the question of recovering shape models from video information. Given video of an object or a scene captured by a moving camera, a prerequisite for model building is to recover the three-dimensional (3D) motion of the camera which consists of a rotation and a translation at each instant. It is shown here that a spherical eye (an eye or system of eyes providing panoramic vision) is superior to a camera-type eye (an eye with restricted field of view such as a common video camera) as regards the competence of 3D motion estimation. This result is derived from a geometric/statistical analysis of all the possible computational models that can be used for estimating 3D motion from an image sequence. Regardless of the estimation procedure for a camera-type eye, the parameters of the 3D rigid motion (translation and rotation) contain errors satisfying specific geometric constraints. Thus, translation is always confused with rotation, resulting in inaccurate results. This confusion does not happen for the case of panoramic vision. Insights obtained from this study point to new ways of constructing powerful imaging devices that suit particular tasks in visualization and virtual reality better than conventional cameras, thus leading to a new camera technology. Such new eyes are constructed by putting together multiple existing video cameras in specific ways, thus obtaining eyes from eyes. For a new eye of this kind we describe an implementation for deriving models of scenes from video data, while avoiding the correspondence problem in the video sequence.}, keywords = {model building, shape reconstruction, structure from motion, video analysis}, isbn = {0925-7721}, doi = {10.1016/S0925-7721(99)00044-9}, url = {http://www.sciencedirect.com/science/article/pii/S0925772199000449}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J. and Brodsk{\'y},Tom{\'a}{\v s}} } @inbook {14236, title = {New Eyes for Shape and Motion Estimation}, booktitle = {Biologically Motivated Computer VisionBiologically Motivated Computer Vision}, series = {Lecture Notes in Computer Science}, volume = {1811}, year = {2000}, month = {2000///}, pages = {23 - 47}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Motivated by the full field of view of insect eyes and their fast and accurate estimation of egomotion, we constructed a system of cameras to take advantage of the full field of view (FOV) constraints that insects use. In this paper, we develop a new ego-motion algorithm for a rigidly mounted set of cameras undergoing arbitrary rigid motion. This egomotion algorithm combines the unambiguous components of the motion computed by each separate camera. We prove that the cyclotorsion is resistant to errors and show this empirically. We show how to calibrate the system with two novel algorithms, one using secondary cameras and one using self calibration. Given this system calibration, the new 3D motion algorithm first computes the rotation and then the 3D translation. We apply this algorithm to a camera system constructed with four rigidly mounted synchronized cameras pointing in various directions and present motion estimation results at www.cfar.umd.edu/ pbaker/argus.html.}, isbn = {978-3-540-67560-0}, url = {http://dx.doi.org/10.1007/3-540-45482-9_12}, author = {Baker,Patrick and Pless,Robert and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Lee,Seong-Whan and B{\"u}lthoff,Heinrich and Poggio,Tomaso} } @conference {14186, title = {A New Framework for Multi-camera Structure from Motion}, booktitle = {Mustererkennung 2000, 22. DAGM-Symposium}, year = {2000}, month = {2000///}, pages = {75 - 82}, author = {Neumann, J. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14239, title = {Observability of 3D Motion}, journal = {International Journal of Computer Vision}, volume = {37}, year = {2000}, month = {2000///}, pages = {43 - 63}, abstract = {This paper examines the inherent difficulties in observing 3D rigid motion from image sequences. It does so without considering a particular estimator. Instead, it presents a statistical analysis of all the possible computational models which can be used for estimating 3D motion from an image sequence. These computational models are classified according to the mathematical constraints that they employ and the characteristics of the imaging sensor (restricted field of view and full field of view). Regarding the mathematical constraints, there exist two principles relating a sequence of images taken by a moving camera. One is the {\textquotedblleft}epipolar constraint,{\textquotedblright} applied to motion fields, and the other the {\textquotedblleft}positive depth{\textquotedblright} constraint, applied to normal flow fields. 3D motion estimation amounts to optimizing these constraints over the image. A statistical modeling of these constraints leads to functions which are studied with regard to their topographic structure, specifically as regards the errors in the 3D motion parameters at the places representing the minima of the functions. For conventional video cameras possessing a restricted field of view, the analysis shows that for algorithms in both classes which estimate all motion parameters simultaneously, the obtained solution has an error such that the projections of the translational and rotational errors on the image plane are perpendicular to each other. Furthermore, the estimated projection of the translation on the image lies on a line through the origin and the projection of the real translation. The situation is different for a camera with a full (360 degree) field of view (achieved by a panoramic sensor or by a system of conventional cameras). In this case, at the locations of the minima of the above two functions, either the translational or the rotational error becomes zero, while in the case of a restricted field of view both errors are non-zero. Although some ambiguities still remain in the full field of view case, the implication is that visual navigation tasks, such as visual servoing, involving 3D motion estimation are easier to solve by employing panoramic vision. Also, the analysis makes it possible to compare properties of algorithms that first estimate the translation and on the basis of the translational result estimate the rotation, algorithms that do the opposite, and algorithms that estimate all motion parameters simultaneously, thus providing a sound framework for the observability of 3D motion. Finally, the introduced framework points to new avenues for studying the stability of image-based servoing schemes.}, isbn = {0920-5691}, url = {http://dx.doi.org/10.1023/A:1008177429387}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {17640, title = {Optimal design of signaling networks for Internet telephony}, booktitle = {IEEE INFOCOM 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings}, volume = {2}, year = {2000}, month = {2000///}, pages = {707-716 vol.2 - 707-716 vol.2}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present an approach for efficient design of a signaling network for a network of software switches supporting Internet telephony. While one may take an integer programming approach to solve this problem, it quickly becomes intractable even for modest-sized networks. Instead, our topology design uses random graphs that we show to be nearly optimal in cost, highly connected, and computationally efficient even for large networks. We then formulate a quadratic assignment problem (QAP) to map the abstract topology into the physical network to achieve optimal load balancing for given demand forecasts, which we solve using randomized heuristics. Numerical results on several example networks illustrate the performance and computational efficiency of our method. A graphical design tool has been developed based on our algorithms}, keywords = {bandwidth allocation, Computational efficiency, Computer networks, Cost function, Demand forecasting, demand forecasts, graph theory, graphical design tool, Internet telephony, Linear programming, Load forecasting, Load management, Network topology, optimal design, optimal load balancing, optimisation, performance, quadratic assignment problem, random graphs, randomised algorithms, randomized heuristics, Signal design, signaling networks, Switches, telecommunication signalling, topology design}, isbn = {0-7803-5880-5}, doi = {10.1109/INFCOM.2000.832245}, author = {Srinivasan, Aravind and Ramakrishnan,K. G and Kumaran,K. and Aravamudan,M. and Naqvi,S.} } @article {14166, title = {The Ouchi illusion as an artifact of biased flow estimation}, journal = {Vision Research}, volume = {40}, year = {2000}, month = {2000/01//}, pages = {77 - 95}, abstract = {A pattern by Ouchi has the surprising property that small motions can cause illusory relative motion between the inset and background regions. The effect can be attained with small retinal motions or a slight jiggling of the paper and is robust over large changes in the patterns, frequencies and boundary shapes. In this paper, we explain that the cause of the illusion lies in the statistical difficulty of integrating local one-dimensional motion signals into two-dimensional image velocity measurements. The estimation of image velocity generally is biased, and for the particular spatial gradient distributions of the Ouchi pattern the bias is highly pronounced, giving rise to a large difference in the velocity estimates in the two regions. The computational model introduced to describe the statistical estimation of image velocity also accounts for the findings of psychophysical studies with variations of the Ouchi pattern and for various findings on the perception of moving plaids. The insight gained from this computational study challenges the current models used to explain biological vision systems and to construct robotic vision systems. Considering the statistical difficulties in image velocity estimation in conjunction with the problem of discontinuity detection in motion fields suggests that theoretically the process of optical flow computations should not be carried out in isolation but in conjunction with the higher level processes of 3D motion estimation, segmentation and shape computation.}, keywords = {Bias, MOTION, optical flow, Plaid, Statistics}, isbn = {0042-6989}, doi = {10.1016/S0042-6989(99)00162-5}, url = {http://www.sciencedirect.com/science/article/pii/S0042698999001625}, author = {Ferm{\"u}ller, Cornelia and Pless,Robert and Aloimonos, J.} } @conference {14297, title = {Polymorphic versus monomorphic points-to analysis}, booktitle = {Proceedings of the 7th International Static Analysis Symposium, Lecture Notes in Computer Science. Springer Verlag}, year = {2000}, month = {2000///}, author = {Foster, Jeffrey S. and Fahndrich,M. and Aiken,A.} } @conference {17532, title = {Receiver based management of low bandwidth access links}, booktitle = {IEEE INFOCOM 2000. Nineteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings}, volume = {1}, year = {2000}, month = {2000///}, pages = {245-254 vol.1 - 245-254 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {In this paper, we describe a receiver-based congestion control policy that leverages TCP flow control mechanisms to prioritize mixed traffic loads across access links. We manage queueing at the access link to: (1) improve the response time of interactive network applications; (2) reduce congestion-related packet losses; while (3) maintaining high throughput for bulk-transfer applications. Our policy controls queue length by manipulating receive socket buffer sizes. We have implemented this solution in a dynamically loadable Linux kernel module, and tested it over low-bandwidth links. Our approach yields a 7-fold improvement in packet latency over an unmodified system while maintaining 94\% link utilization. In the common case, congestion-related packet losses at the access link can be eliminated. Finally, by prioritizing short flows, we show that our system reduces the time to download a complex Web page during a large background transfer by a factor of two}, keywords = {Bandwidth, buffer storage, bulk-transfer applications, complex Web page, congestion control policy, Delay, dynamically loadable Linux kernel module, information resources, interactive network, Internet, Kernel, link utilization, Linux, low-bandwidth access links, mixed traffic load, packet latency, queue length, queueing theory, receive socket buffer sizes, receiver-based management, response time, short flow prioritizing, Size control, Sockets, subscriber loops, TCP flow control, telecommunication congestion control, telecommunication network management, Telecommunication traffic, Testing, Throughput, Transport protocols, Unix, Web pages}, isbn = {0-7803-5880-5}, doi = {10.1109/INFCOM.2000.832194}, author = {Spring, Neil and Chesire,M. and Berryman,M. and Sahasranaman,V. and Anderson,T. and Bershad,B.} } @article {14652, title = {A Review of Current Routing Protocols for Ad Hoc}, journal = {IEEE Personal Communications}, volume = {29}, year = {2000}, month = {2000///}, pages = {156 - 71}, author = {Royer,E.M. and Toh,C.K. and Hicks, Michael W. and Kakkar,P. and Moore,J. T and Hicks, Michael W. and Moore,J. T and Alexander,D. S and Gunter,C. A and Nettles,S. and others} } @inbook {14732, title = {Scalable Resource Control in Active Networks}, booktitle = {Active NetworksActive Networks}, series = {Lecture Notes in Computer Science}, volume = {1942}, year = {2000}, month = {2000///}, pages = {343 - 357}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {The increased complexity of the service model relative to store-and-forward routers has made resource management one of the paramount concerns in active networking research and engineering. In this paper,we address two major challenges in scaling resource management-to-many-node active networks. The first is the use of market mechanisms and trading amongst nodes and programs with varying degrees of competition and cooperation to provide a scalable approach to managing active network resources. The second is the use of a trust-management architecture to ensure that the participants in the resource management marketplace have a policy-driven {\textquotedblleft}rule of law{\textquotedblright} in which marketplace decisions can be made and relied upon. We have used lottery scheduling and the Keynote trust-management system for our implementation, for which we provide some initial performance indications.}, isbn = {978-3-540-41179-6}, url = {http://dx.doi.org/10.1007/3-540-40057-5_25}, author = {Anagnostakis,Kostas and Hicks, Michael W. and Ioannidis,Sotiris and Keromytis,Angelos and Smith,Jonathan}, editor = {Yasuda,Hiroshi} } @article {12089, title = {Secure quality of service handling: SQoSH}, journal = {IEEE Communications Magazine}, volume = {38}, year = {2000}, month = {2000/04//}, pages = {106 - 112}, abstract = {Proposals for programmable network infrastructures, such as active networks and open signaling, provide programmers with access to network resources and data structures. The motivation for providing these interfaces is accelerated introduction of new services, but exposure of the interfaces introduces many new security risks. We describe some of the security issues raised by active networks. We then describe our secure active network environment (SANE) architecture. SANE was designed as a security infrastructure for active networks, and was implemented in the SwitchWare architecture. SANE restricts the actions that loaded modules can perform by restricting the resources that can be named; this is further extended to remote invocation by means of cryptographic credentials. SANE can be extended to support restricted control of quality of service in a programmable network element. The Piglet lightweight device kernel provides a {\textquotedblleft}virtual clock{\textquotedblright} type of scheduling discipline for network traffic, and exports several tuning knobs with which the clock can be adjusted. The ALIEN active loader provides safe access to these knobs to modules that operate on the network element. Thus, the proposed SQoSH architecture is able to provide safe, secure access to network resources, while allowing these resources to be managed by end users needing customized networking services. A desirable consequence of SQoSH{\textquoteright}s integration of access control and resource control is that a large class of denial-of-service attacks, unaddressed solely with access control and cryptographic protocols, can now be prevented}, keywords = {Acceleration, Access control, active networks, ALIEN active loader, Clocks, Computer network management, cryptographic credentials, cryptography, customized networking services, Data security, Data structures, denial-of-service attacks, interfaces, Kernel, loaded modules, network resources, network traffic, open signaling, packet switching, Piglet lightweight device kernel, programmable network element, programmable network infrastructures, Programming profession, Proposals, quality of service, remote invocation, resource control, restricted control of quality of service, SANE, scheduling, scheduling discipline, secure active network environment architecture, secure quality of service handling, security infrastructure, security risks, SQoSH, SwitchWare architecture, telecommunication security, tuning knobs, virtual clock}, isbn = {0163-6804}, doi = {10.1109/35.833566}, author = {Alexander,D. S and Arbaugh, William A. and Keromytis,A. D and Muir,S. and Smith,J. M} } @conference {14267, title = {The statistics of optical flow: implications for the process of correspondence in vision}, booktitle = {15th International Conference on Pattern Recognition, 2000. Proceedings}, volume = {1}, year = {2000}, month = {2000///}, pages = {119-126 vol.1 - 119-126 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper studies the three major categories of flow estimation methods: gradient-based, energy-based, and correlation methods; it analyzes different ways of compounding 1D motion estimates (image gradients, spatio-temporal frequency triplets, local correlation estimates) into 2D velocity estimates, including linear and nonlinear methods. Correcting for the bias would require knowledge of the noise parameters. In many situations, however, these are difficult to estimate accurately, as they change with the dynamic imagery in unpredictable and complex ways. Thus, the bias really is a problem inherent to optical flow estimation. We argue that the bias is also integral to the human visual system. It is the cause of the illusory perception of motion in the Ouchi pattern and also explains various psychophysical studies of the perception of moving plaids. Finally, the implication of the analysis is that flow or correspondence can be estimated very accurately only when feedback is utilized}, keywords = {Bias, Computer vision, correlation, correlation methods, energy-based method, flow estimation, Frequency estimation, gradient method, gradient methods, Image analysis, Image motion analysis, Image sequences, least squares, least squares approximations, Motion estimation, Nonlinear optics, Optical feedback, optical flow, Optical harmonic generation, Optical noise, Statistics, Visual perception}, isbn = {0-7695-0750-6}, doi = {10.1109/ICPR.2000.905288}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {11968, title = {Structure from motion: Beyond the epipolar constraint}, journal = {International Journal of Computer Vision}, volume = {37}, year = {2000}, month = {2000///}, pages = {231 - 258}, author = {Brodsky, T. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {12081, title = {Windows of Vulnerability: A Case Study Analysis}, volume = {33}, year = {2000}, month = {2000///}, pages = {52 - 59}, abstract = {The authors propose a life-cycle model for system vulnerabilities, applying to three case studies to show how systems remain vulnerable long after security fixes are available. Complex information and communication systems give rise to design, implementation, and management errors, leading to a vulnerability in an information technology product that can allow security policy violations. Using their vulnerability life-cycle model, the authors present a case study analysis of specific computer vulnerabilities. For each case, the authors provide background information about the vulnerability, such as how attackers exploited it and which systems were affected. They tie the case to the life-cycle model by identifying the dates for each state within the model. Finally, they use a histogram of reported intrusions to show the life of the vulnerability and conclude with an analysis specific to the particular vulnerability.}, isbn = {0018-9162}, author = {Arbaugh, William A. and Fithen,William L. and McHugh,John} } @inbook {14191, title = {Active Perception}, booktitle = {Wiley Encyclopedia of Electrical and Electronics EngineeringWiley Encyclopedia of Electrical and Electronics Engineering}, year = {1999}, month = {1999/12/27/}, publisher = {John Wiley \& Sons, Inc.}, organization = {John Wiley \& Sons, Inc.}, isbn = {9780471346081}, url = {http://onlinelibrary.wiley.com/doi/10.1002/047134608X.W5515/abstract?systemMessage=Wiley+Online+Library+will+be+disrupted+17+March+from+10-14+GMT+\%2806-10+EDT\%29+for+essential+maintenance\&userIsAuthenticated=false\&deniedAccessCustomisedMessage=}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {15376, title = {Advances in spherical harmonic device modeling: calibration and nanoscale electron dynamics}, booktitle = {Simulation of Semiconductor Processes and Devices, 1999. SISPAD {\textquoteright}99. 1999 International Conference on}, year = {1999}, month = {1999///}, pages = {167 - 170}, abstract = {Improvements in the Spherical Harmonic (SH) method for solving Boltzmann Transport Equation (BTE) are presented. The simulation results provide the same physical detail as analytical band Monte Carlo (MC) calculations, and are obtained approximately a thousand times faster. A new physical model for surface scattering has also been developed. As a result, the SHBTE model achieves calibration for a complete process of I-V characteristics and substrate current consistently for the first time}, keywords = {Boltzmann, characteristics;SHBTE, current;surface, device, dynamics;spherical, electron, equation;calibration;semiconductor, equation;I-V, harmonic, model;substrate, models;surface, scattering;, scattering;Boltzmann, simulation;calibration;nanoscale, transport}, doi = {10.1109/SISPAD.1999.799287}, author = {Lin,Chung-Kai and Goldsman,N. and Mayergoyz, Issak D and Aronowitz,S. and Belova,N.} } @conference {18776, title = {Automated punch shape synthesis for sheet metal bending operations}, year = {1999}, month = {1999///}, author = {Alva,U. and Gupta,S.K.} } @article {14781, title = {Binary version management for computational grids}, journal = {Parallel processing letters}, volume = {9}, year = {1999}, month = {1999///}, pages = {215 - 225}, author = {Hollingsworth, Jeffrey K and Miller, E. L and Akala, K.} } @article {18668, title = {Coverage estimation methods for stratified fault-injection}, journal = {Computers, IEEE Transactions on}, volume = {48}, year = {1999}, month = {1999/07//}, pages = {707 - 723}, abstract = {This paper addresses the problem of estimating fault tolerance coverage through statistical processing of observations collected in fault-injection experiments. In an earlier paper, venous estimators based on simple sampling in the complete fault/activity input space and stratified sampling in a partitioned space were studied; frequentist confidence limits were derived based on a normal approximation. In this paper, the validity of this approximation is analyzed. The theory of confidence regions is introduced to estimate coverage without approximation when stratification is used. Three statistics are considered for defining confidence regions. It is shown that one-a vectorial statistic-is often more conservative than the other two. However, only the vectorial statistic is computationally tractable. We then consider Bayesian estimation methods for stratified sampling. Two methods are presented to obtain an approximation of the posterior distribution of the coverage by calculating its moments. The moments are then used to identify the type of the distribution in the Pearson distribution system, to estimate its parameters, and to obtain the coverage confidence limit. Three hypothetical example systems are used to compare the validity and the conservatism of the frequentist and Bayesian estimations}, keywords = {Bayes methods, Bayesian estimations, confidence regions, coverage estimation methods, fault tolerance coverage, fault tolerant computing, frequentist confidence limits, parameter estimation, parameters estimation, Pearson distribution system, statistical processing, stratified fault-injection, stratified sampling, vectorial statistic}, isbn = {0018-9340}, doi = {10.1109/12.780878}, author = {Michel Cukier and Powell,D. and Ariat,J.} } @conference {18568, title = {Field-to-frame transcoding with spatial and temporal downsampling}, booktitle = {Image Processing, 1999. ICIP 99. Proceedings. 1999 International Conference on}, volume = {4}, year = {1999}, month = {1999///}, pages = {271 -275 vol.4 - 271 -275 vol.4}, abstract = {We present an algorithm for transcoding high-rate compressed bitstreams containing field-coded interlaced video to lower-rate compressed bitstreams containing frame-coded progressive video. We focus on MPEG-2 to H.263 transcoding, however these results can be extended to other lower-rate video compression standards including MPEG-4 simple profile and MPEG-1. A conventional approach to the transcoding problem involves decoding the input bitstream, spatially and temporally downsampling the decoded frames, and re-encoding the result. The proposed transcoder achieves improved performance by exploiting the details of the MPEG-2 and H.263 compression standards when performing interlaced to progressive (or field to frame) conversion with spatial downsampling and frame-rate reduction. The transcoder reduces the MPEG-2 decoding requirements by temporally downsampling the data at the bitstream level and reduces the H.263 encoding requirements by largely bypassing H.263 motion estimation by reusing the motion vectors and coding modes given in the input bitstream. In software implementations, the proposed approach achieved a 5 times; speedup over the conventional approach with only a 0.3 and 0.5 dB loss in PSNR for the Carousel and Bus sequences}, keywords = {data compression, decoding, Encoding, field-coded interlaced video, field-to-frame transcoding, frame-coded progressive video, H.263 motion estimation, H.263 transcoding, high-rate compressed bitstreams, lower-rate compressed bitstreams, Motion estimation, MPEG-1, MPEG-2, MPEG-4, spatial downsampling, standards, temporal downsampling, transcoder, video coding, video compression standards}, doi = {10.1109/ICIP.1999.819593}, author = {Wee,S.J. and Apostolopoulos,J.G. and Feamster, Nick} } @conference {15923, title = {Fkpurang, darsana, traum, cfa, perlisg@ cs. umd. edu}, booktitle = {In Proceedings of the IJCAI{\textquoteright}99 Workshop on Practical Reasoning and Rationality}, year = {1999}, month = {1999///}, author = {Edu,C. U and Purang,K. and Purushothaman,D. and Traum,D. and Andersen,C. and Perlis, Don} } @conference {11986, title = {Independent motion: the importance of history}, booktitle = {Computer Vision and Pattern Recognition, 1999. IEEE Computer Society Conference on.}, volume = {2}, year = {1999}, month = {1999///}, pages = {97 Vol. 2 - 97 Vol. 2}, publisher = {IEEE}, organization = {IEEE}, abstract = {We consider a problem central in aerial visual surveillance applications-detection and tracking of small, independently moving objects in long and noisy video sequences. We directly use spatiotemporal image intensity gradient measurements to compute an exact model of background motion. This allows the creation of accurate mosaics over many frames and the definition of a constraint violation function which acts as an indication of independent motion. A novel temporal integration method maintains confidence measures over long subsequences without computing the optic flow, requiring object models, or using a Kalman filler. The mosaic acts as a stable feature frame, allowing precise localization of the independently moving objects. We present a statistical analysis of the effects of image noise on the constraint violation measure and find a good match between the predicted probability distribution function and the measured sample frequencies in a test sequence}, keywords = {aerial visual surveillance, background image, Fluid flow measurement, Frequency measurement, History, Motion detection, Motion estimation, Motion measurement, Noise measurement, Optical computing, Optical noise, spatiotemporal image intensity gradient measurements, Spatiotemporal phenomena, Surveillance, Video sequences}, isbn = {0-7695-0149-4}, doi = {10.1109/CVPR.1999.784614}, author = {Pless, R. and Brodsky, T. and Aloimonos, J.} } @conference {15984, title = {Mixed initiative dialogue and intelligence via active logic}, booktitle = {proceedings of the AAAI99 Workshop on Mixed-Initiative Intelligence}, year = {1999}, month = {1999///}, pages = {60 - 67}, author = {Andersen,C. and Traum,D. and Purang,K. and Purushothaman,D. and Perlis, Don} } @conference {12041, title = {Motion Segmentation: A Synergistic Approach}, booktitle = {Computer Vision and Pattern Recognition, IEEE Computer Society Conference on}, volume = {2}, year = {1999}, month = {1999///}, pages = {2226 - 2226}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {Since estimation of camera motion requires knowledge of independent motion, and moving object detection and localization requires knowledge about the camera motion, the two problems of motion estimation and segmentation need to be solved together in a synergistic manner. This paper provides an approach to treating both these problems simultaneously. The technique introduced here is based on a novel concept, {\textquoteleft}{\textquoteleft}scene ruggedness,{\textquoteright}{\textquoteright} which parameterizes the variation in estimated scene depth with the error in the underlying three-dimensional (3D) motion. The idea is that incorrect 3D motion estimates cause distortions in the estimated depth map, and as a result smooth scene patches are computed as rugged surfaces. The correct 3D motion can be distinguished, as it does not cause any distortion and thus gives rise to the background patches with the least depth variation between depth discontinuities, with the locations corresponding to independent motion being rugged. The algorithm presented employs a binocular observer whose nature is exploited in the extraction of depth discontinuities, a step that facilitates the overall procedure, but the technique can be extended to a monocular observer in a variety of ways.}, keywords = {epipolar minimization, independent motion detection, Motion analysis}, doi = {http://doi.ieeecomputersociety.org/10.1109/CVPR.1999.784633}, author = {Ferm{\"u}ller, Cornelia and Brodsky, Tomas and Aloimonos, J.} } @conference {14723, title = {PLANet: an active internetwork}, booktitle = {IEEE INFOCOM {\textquoteright}99. Eighteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings}, volume = {3}, year = {1999}, month = {1999/03/21/25}, pages = {1124-1133 vol.3 - 1124-1133 vol.3}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present PLANet: an active network architecture and implementation. In addition to a standard suite of Internet-like services, PLANet has two key programmability features: (1) all packets contain programs; and (2) router functionality may be extended dynamically. Packet programs are written in our special purpose programming language PLAN, the Packet Language for Active Networks, while dynamic router extensions are written in OCaml, a dialect of ML. Currently, PLANet routers run as byte-code-interpreted Linux user-space applications, and support Ethernet and IP as link layers. PLANet achieves respectable performance on standard networking operations: on 300 MHz Pentium-II{\textquoteright}s attached to 100 Mbps Ethernet, PLANet can route 48 Mbps and switch over 5000 packets per second. We demonstrate the utility of PLANet{\textquoteright}s activeness by showing experimentally how it can nontrivially improve application and aggregate network performance in congested conditions}, keywords = {100 Mbit/s, 300 MHz, 48 Mbit/s, active internetwork, active network architecture, active network implementation, byte-code-interpreted applications, Computer architecture, Computer languages, Computer networks, congested conditions, dynamic programming, dynamic router extensions, Ethernet, Ethernet networks, INFORMATION SCIENCE, Internet, Internet-like services, internetworking, IP, IP networks, link layers, Linux user-space applications, Local area networks, ML dialect, Network performance, networking operations, OCaml, Packet Language for Active Networks, packet programs, packet switching, Pentium-II, performance, performance evaluation, PLAN, PLANet, Planets, programmability features, programming languages, router functionality, special purpose programming language, Switches, telecommunication network routing, Transport protocols, Web and internet services}, isbn = {0-7803-5417-6}, doi = {10.1109/INFCOM.1999.751668}, author = {Hicks, Michael W. and Moore,J. T and Alexander,D. S and Gunter,C. A and Nettles,S. M} } @conference {16023, title = {Practical reasoning and plan execution with active logic}, booktitle = {Proceedings of the IJCAI-99 Workshop on Practical Reasoning and Rationality}, year = {1999}, month = {1999///}, pages = {30 - 38}, author = {Purang,K. and Purushothaman,D. and Traum,D. and Andersen,C. and Perlis, Don} } @article {13465, title = {Reports on the AAAI Fall Symposia}, journal = {AI Magazine}, volume = {20}, year = {1999}, month = {1999/09/15/}, pages = {87 - 87}, abstract = {The Association for the Advancement of Artificial Intelligence (AAAI) held its 1998 Fall Symposium Series on 23 to 25 October at the Omni Rosen Hotel in Orlando, Florida. This article contains summaries of seven of the symposia that were conducted: (1) Cognitive Robotics; (2) Distributed, Continual Planning; (3) Emotional and Intelligent: The Tangled Knot of Cognition; (4) Integrated Planning for Autonomous Agent Architectures; (5) Planning with Partially Observable Markov Decision Processes; (6) Reasoning with Visual and Diagrammatic Representations; and (7) Robotics and Biology: Developing Connections.}, isbn = {0738-4602}, doi = {10.1609/aimag.v20i3.1470}, url = {http://www.aaai.org/ojs/index.php/aimagazine/article/viewArticle/1470}, author = {De Giacomo,Giuseppe and desJardins, Marie and Canamero,Dolores and Wasson,Glenn and Littman,Michael and Allwein,Gerard and Marriott,Kim and Meyer,Bernd and Webb,Barbara and Consi,Tom} } @book {19005, title = {The RNA World}, year = {1999}, month = {1999///}, publisher = {Cold Spring Harbor Laboratory Press}, organization = {Cold Spring Harbor Laboratory Press}, address = {Cold Spring Harbor, New York}, author = {Woodson,Sarah A. and Mount, Stephen M.}, editor = {Gesteland,Raymond F. and Cech,Thomas R. and Atkins,John F.} } @article {13908, title = {The Role of Children in the Design of New Technology}, journal = {Saatavilla www-muodossa: ftp://ftp. cs. umd. edu/pub/hcil/Reports-Abstracts-Bibliography/99-23html/99-23. pdf (Luettu 17.1. 2007)}, year = {1999}, month = {1999///}, author = {Allison,D.} } @conference {12054, title = {Shape from Video}, booktitle = {Computer Vision and Pattern Recognition, IEEE Computer Society Conference on}, volume = {2}, year = {1999}, month = {1999///}, pages = {2146 - 2146}, publisher = {IEEE Computer Society}, organization = {IEEE Computer Society}, address = {Los Alamitos, CA, USA}, abstract = {This paper presents a novel technique for recovering the shape of a static scene from a video sequence due to a rigidly moving camera. The solution procedure consists of two stages. In the first stage, the rigid motion of the camera at each instant in time is recovered. This provides the transformation between successive viewing positions. The solution is achieved through new constraints which relate 3D motion and shape directly to the image derivatives. These constraints allow the combination of the processes of 3D motion estimation and segmentation by exploiting the geometry and statistics inherent in the data. In the second stage the scene surfaces are reconstructed through an optimization procedure which utilizes data from all the frames of a short video sequence. A number of experimental results demonstrate the potential of the approach.}, keywords = {epipolar minimization, model extraction, Motion analysis, vision and graphics}, doi = {http://doi.ieeecomputersociety.org/10.1109/CVPR.1999.784622}, author = {Brodsky, Tomas and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {17962, title = {Skip Strips: maintaining triangle strips for view-dependent rendering}, booktitle = {Visualization {\textquoteright}99. Proceedings}, year = {1999}, month = {1999/10//}, pages = {131 - 518}, abstract = {View-dependent simplification has emerged as a powerful tool for graphics acceleration in visualization of complex environments. However, view-dependent simplification techniques have not been able to take full advantage of the underlying graphics hardware. Specifically, triangle strips are a widely used hardware-supported mechanism to compactly represent and efficiently render static triangle meshes. However, in a view-dependent framework, the triangle mesh connectivity changes at every frame, making it difficult to use triangle strips. We present a novel data structure, Skip Strip, that efficiently maintains triangle strips during such view-dependent changes. A Skip Strip stores the vertex hierarchy nodes in a skip-list-like manner with path compression. We anticipate that Skip Strips will provide a road map to combine rendering acceleration techniques for static datasets, typical of retained-mode graphics applications, with those for dynamic datasets found in immediate-mode applications.}, keywords = {(computer, Acceleration, acceleration;graphics, applications;path, applications;skip-list-like, changes;view-dependent, compression;rendering, connectivity;triangle, data, datasets;graphics, datasets;static, environments;data, equipment;data, graphic, Graphics, graphics);spatial, hardware;hardware-supported, hierarchy, manner;static, mechanism;immediate-mode, mesh, meshes;triangle, nodes;view-dependent, rendering;view-dependent, simplification;visualization;computer, Skip, Strips;complex, strips;vertex, structure;dynamic, structures;, techniques;retained-mode, triangle, visualisation;rendering}, doi = {10.1109/VISUAL.1999.809877}, author = {El-Sana,J. and Azanli,E. and Varshney, Amitabh} } @conference {11965, title = {Statistical biases in optic flow}, booktitle = {Computer Vision and Pattern Recognition, 1999. IEEE Computer Society Conference on.}, volume = {1}, year = {1999}, month = {1999///}, pages = {566 Vol. 1 - 566 Vol. 1}, publisher = {IEEE}, organization = {IEEE}, abstract = {The computation of optical flow from image derivatives is biased in regions of non uniform gradient distributions. A least-squares or total least squares approach to computing optic flow from image derivatives even in regions of consistent flow can lead to a systematic bias dependent upon the direction of the optic flow, the distribution of the gradient directions, and the distribution of the image noise. The bias a consistent underestimation of length and a directional error. Similar results hold for various methods of computing optical flow in the spatiotemporal frequency domain. The predicted bias in the optical flow is consistent with psychophysical evidence of human judgment of the velocity of moving plaids, and provides an explanation of the Ouchi illusion. Correction of the bias requires accurate estimates of the noise distribution; the failure of the human visual system to make these corrections illustrates both the difficulty of the task and the feasibility of using this distorted optic flow or undistorted normal flow in tasks requiring higher lever processing}, keywords = {Distributed computing, Frequency domain analysis, HUMANS, image derivatives, Image motion analysis, Image sequences, Least squares methods, Motion estimation, Optical computing, Optical distortion, optical flow, Optical noise, Ouchi illusion, perception of motion, Psychology, Spatiotemporal phenomena, statistical analysis, systematic bias, total least squares}, isbn = {0-7695-0149-4}, doi = {10.1109/CVPR.1999.786994}, author = {Ferm{\"u}ller, Cornelia and Pless, R. and Aloimonos, J.} } @article {16171, title = {Temporal visualization for legal case histories}, journal = {PROCEEDINGS OF THE ANNUAL MEETING-AMERICAN SOCIETY FOR INFORMATION SCIENCE}, volume = {36}, year = {1999}, month = {1999///}, pages = {271 - 279}, abstract = {This paper discusses visualization of legal information using a tool for temporal information called LifeLines. The direct and indirect histories of cases can become very complex. We explored ways that LifeLines could aid in viewing the links between the original case and the direct and indirect histories. The Apple Computer, Inc. v. Microsoft Corporation and Hewlett Packard Company case is used to illustrate the prototype. For example, if users want to find out how the rulings or statutes changed throughout this case, they could retrieve this information within a single display. Using the timeline, users could also choose at which point in time they would like to begin viewing the case. LifeLines support various views of a case{\textquoteright}s history. For instance, users can view the trial history of a case, the references involved in a case, and citations made to a case. The paper describes improvements to LifeLines that could help in providing a more useful visualization of case history.}, author = {Harris,C. and Allen,R.B. and Plaisant, Catherine and Shneiderman, Ben} } @conference {14303, title = {A theory of type qualifiers}, booktitle = {ACM SIGPLAN Notices}, volume = {34}, year = {1999}, month = {1999///}, pages = {192 - 203}, author = {Foster, Jeffrey S. and F{\"a}hndrich,M. and Aiken,A.} } @article {14181, title = {3D motion and shape representations in visual servo control}, journal = {The International Journal of Robotics Research}, volume = {17}, year = {1998}, month = {1998///}, pages = {4 - 4}, author = {Ferm{\"u}ller, Cornelia and Cheong, L. F and Aloimonos, J.} } @article {14195, title = {Ambiguity in Structure from Motion: Sphere versus Plane}, journal = {International Journal of Computer Vision}, volume = {28}, year = {1998}, month = {1998///}, pages = {137 - 154}, abstract = {If 3D rigid motion can be correctly estimated from image sequences, the structure of the scene can be correctly derived using the equations for image formation. However, an error in the estimation of 3D motion will result in the computation of a distorted version of the scene structure. Of computational interest are these regions in space where the distortions are such that the depths become negative, because in order for the scene to be visible it has to lie in front of the image, and thus the corresponding depth estimates have to be positive. The stability analysis for the structure from motion problem presented in this paper investigates the optimal relationship between the errors in the estimated translational and rotational parameters of a rigid motion that results in the estimation of a minimum number of negative depth values. The input used is the value of the flow along some direction, which is more general than optic flow or correspondence. For a planar retina it is shown that the optimal configuration is achieved when the projections of the translational and rotational errors on the image plane are perpendicular. Furthermore, the projection of the actual and the estimated translation lie on a line through the center. For a spherical retina, given a rotational error, the optimal translation is the correct one; given a translational error, the optimal rotational negative deptherror depends both in direction and value on the actual and estimated translation as well as the scene in view. The proofs, besides illuminating the confounding of translation and rotation in structure from motion, have an important application to ecological optics. The same analysis provides a computational explanation of why it is easier to estimate self-motion in the case of a spherical retina and why shape can be estimated easily in the case of a planar retina, thus suggesting that nature{\textquoteright}s design of compound eyes (or panoramic vision) for flying systems and camera-type eyes for primates (and other systems that perform manipulation) is optimal.}, isbn = {0920-5691}, url = {http://dx.doi.org/10.1023/A:1008063000586}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {15634, title = {Approximation algorithms for multiple-tool miling}, booktitle = {Proceedings of the fourteenth annual symposium on Computational geometry}, series = {SCG {\textquoteright}98}, year = {1998}, month = {1998///}, pages = {297 - 306}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-973-4}, doi = {10.1145/276884.276918}, url = {http://doi.acm.org/10.1145/276884.276918}, author = {Arya,Sunil and Cheng,Siu-Wing and Mount, Dave} } @conference {12102, title = {Automated recovery in a secure bootstrap process}, booktitle = {Proceedings of the Symposium on Network and Distributed Systems Security (NDSS{\textquoteright}98)}, year = {1998}, month = {1998///}, pages = {155 - 167}, author = {Arbaugh, William A. and Keromytis,A. D and Farber,D. J and Smith,J. M} } @conference {14762, title = {Benchmarking a network of PCs running parallel applications}, booktitle = {Performance, Computing and Communications, 1998. IPCCC {\textquoteright}98., IEEE International}, year = {1998}, month = {1998/02/16/18}, pages = {1 - 7}, publisher = {IEEE}, organization = {IEEE}, abstract = {Presents a benchmarking study that compares the performance of a network of four PCs connected by a 100 Mbit/s fast Ethernet running three different system software configurations: TCP/IP on Windows NT, TCP/IP on Linux and a lightweight message-passing protocol (U-Net active messages) on Linux. For each configuration, we report results for communication micro-benchmarks and the NAS (Numerical Aerodynamics Simulation) parallel benchmarks. For the NAS benchmarks, the overall running time using Linux TCP/IP was 12-500\% less than the Windows NT TCP/IP configuration. Likewise, the Linux U-Net based message-passing protocol outperformed the Linux TCP/IP version by 5-200\%+. We also show that, by using Linux U-Net, we are able to achieve 125 μs latency between two processes using PVM. Finally, we report that the default mathematical libraries supplied with NT (for both gcc and Visual C++) are substantially slower than the one supplied with Linux}, keywords = {100 Mbit/s, 125 mus, Aerodynamics, Application software, communication micro-benchmarks, default mathematical libraries, Delay, Ethernet, Ethernet networks, gcc, latency, lightweight message-passing protocol, Linux, Local area networks, mathematics computing, Message passing, microcomputer applications, Microsoft Windows NT, NAS parallel benchmarks, network operating systems, Numerical simulation, parallel applications, PARALLEL PROCESSING, PC network benchmarking, performance comparison, performance evaluation, Personal communication networks, Protocols, PVM, running time, software libraries, System software, system software configurations, TCP/IP, TCPIP, Transport protocols, U-Net active messages, Visual C++}, isbn = {0-7803-4468-5}, doi = {10.1109/PCCC.1998.659876}, author = {Hollingsworth, Jeffrey K and Guven, E. and Akinlar, C.} } @inbook {14202, title = {Beyond the Epipolar Constraint: Integrating 3D Motion and Structure Estimation}, booktitle = {3D Structure from Multiple Images of Large-Scale Environments3D Structure from Multiple Images of Large-Scale Environments}, series = {Lecture Notes in Computer Science}, volume = {1506}, year = {1998}, month = {1998///}, pages = {109 - 123}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {This paper develops a novel solution to the problem of recovering the structure of a scene given an uncalibrated video sequence depicting the scene. The essence of the technique lies in a method for recovering the rigid transformation between the different views in the image sequence. Knowledge of this 3D motion allows for self-calibration and for subsequent recovery of 3D structure. The introduced method breaks away from applying only the traditionally used epipolar constraint and introduces a new constraint based on the interaction between 3D motion and shape. Up to now, structure from motion algorithms proceeded in two well defined steps, where the first and most important step is recovering the rigid transformation between two views, and the subsequent step is using this transformation to compute the structure of the scene in view. Here both aforementioned steps are accomplished in a synergistic manner. Existing approaches to 3D motion estimation are mostly based on the use of optic flow which however poses a problem at the locations of depth discontinuities. If we knew where depth discontinuities were, we could (using a multitude of approaches based on smoothness constraints) estimate accurately flow values for image patches corresponding to smooth scene patches; but to know the discontinuities requires solving the structure from motion problem first. In the past this dilemma has been addressed by improving the estimation of flow through sophisticated optimization techniques, whose performance often depends on the scene in view. In this paper the main idea is based on the interaction between 3D motion and shape which allows us to estimate the 3D motion while at the same time segmenting the scene. If we use a wrong 3D motion estimate to compute depth, then we obtain a distorted version of the depth function. The distortion, however, is such that the worse the motion estimate, the more likely we are to obtain depth estimates that are locally unsmooth, i.e., they vary more than the correct ones. Since local variability of depth is due either to the existence of a discontinuity or to a wrong 3D motion estimate, being able to differentiate between these two cases provides the correct motion, which yields the {\textquotedblleft}smoothest{\textquotedblright} estimated depth as well as the image locations of scene discontinuities. Although no optic flow values are computed, we show that our algorithm is very much related to minimizing the epipolar constraint when the scene in view is smooth. When however the imaged scene is not smooth, the introduced constraint has in general different properties from the epipolar constraint and we present experimental results with real sequences where it performs better.}, isbn = {978-3-540-65310-3}, url = {http://dx.doi.org/10.1007/3-540-49437-5_8}, author = {Brodsk{\'y},Tom{\'a}{\v s} and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Koch,Reinhard and Van Gool,Luc} } @article {17864, title = {The design and evaluation of a high-performance earth science database}, journal = {Parallel Computing}, volume = {24}, year = {1998}, month = {1998/01//}, pages = {65 - 89}, abstract = {Earth scientists have encountered two major obstacles in their attempts to use remotely sensed imagery to analyze the earth{\textquoteright}s land cover dynamics. First, the volume of data involved is very large and second, significant preprocessing is needed before the data can be used. This is particularly so for studies that analyze global trends using data sets that cover multiple years. In this paper, we present the design of an earth science database as well as our early experiences with it. The primary design goal of this database is to facilitate efficient access to and preprocessing of large volumes of satellite data. Our initial design assumed that the main bottleneck in the system would be retrieving data from the disks. However, experimental results show that precise identification of all the data values corresponding to a query can take a significant amount of time. The problem is even more pronounced in designing the system to attempt to minimize time spent performing I/O. We therefore discuss a major redesign of the system that includes a reworking of the indexing scheme and a reorganization of the data on disks. Experimental results show that the redesigned system performs significantly better than the original system, providing interactive response times for local queries.}, keywords = {Communication, High-performance I/O, load balancing, scalability, Scientific databases}, isbn = {0167-8191}, doi = {10.1016/S0167-8191(97)00117-8}, url = {http://www.sciencedirect.com/science/article/pii/S0167819197001178}, author = {Shock,Carter T. and Chang,Chialin and Moon,Bongki and Acharya,Anurag and Davis, Larry S. and Saltz,Joel and Sussman, Alan} } @article {12090, title = {DHCP++: Applying an efficient implementation method for fail-stop cryptographic protocols}, journal = {Proceedings of Global Internet (GlobeCom){\textquoteright}98}, year = {1998}, month = {1998///}, pages = {59 - 65}, author = {Arbaugh, William A. and Keromytis,A. D and Smith,J. M} } @conference {17832, title = {Digital dynamic telepathology{\textendash}the Virtual Microscope.}, booktitle = {Proceedings of the AMIA Symposium}, year = {1998}, month = {1998///}, pages = {912 - 912}, author = {Afework,A. and Beynon,M. D and Bustamante,F. and Cho,S. and Demarzo,A. and Ferreira,R. and Miller,R. and Silberman,M. and Saltz, J. and Sussman, Alan and others} } @article {14214, title = {Directions of Motion Fields are Hardly Ever Ambiguous}, journal = {International Journal of Computer Vision}, volume = {26}, year = {1998}, month = {1998///}, pages = {5 - 24}, abstract = {If instead of the full motion field, we consider only the direction of the motion field due to a rigid motion, what can we say about the three-dimensional motion information contained in it? This paper provides a geometric analysis of this question based solely on the constraint that the depth of the surfaces in view is positive. The motivation behind this analysis is to provide a theoretical foundation for image constraints employing only the sign of flow in various directions and justify their utilization for addressing 3D dynamic vision problems.}, isbn = {0920-5691}, url = {http://dx.doi.org/10.1023/A:1007928406666}, author = {Brodsky, Tomas and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @inbook {16258, title = {Drawing of Two-Dimensional Irregular Meshes}, booktitle = {Graph DrawingGraph Drawing}, series = {Lecture Notes in Computer Science}, volume = {1547}, year = {1998}, month = {1998///}, pages = {1 - 14}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We present a method for transforming two-dimensional irregular meshes into square meshes with only a constant blow up in area. We also explore context invariant transformations of irregular meshes into square meshes and provide a lower bound for the transformation of down-staircases.}, isbn = {978-3-540-65473-5}, url = {http://dx.doi.org/10.1007/3-540-37623-2_1}, author = {Aggarwal,Alok and Rao Kosaraju,S. and Pop, Mihai}, editor = {Whitesides,Sue} } @article {14215, title = {Effects of Errors in the Viewing Geometry on Shape Estimation}, journal = {Computer Vision and Image Understanding}, volume = {71}, year = {1998}, month = {1998/09//}, pages = {356 - 372}, abstract = {A sequence of images acquired by a moving sensor contains information about the three-dimensional motion of the sensor and the shape of the imaged scene. Interesting research during the past few years has attempted to characterize the errors that arise in computing 3D motion (egomotion estimation) as well as the errors that result in the estimation of the scene{\textquoteright}s structure (structure from motion). Previous research is characterized by the use of optic flow or correspondence of features in the analysis as well as by the employment of particular algorithms and models of the scene in recovering expressions for the resulting errors. This paper presents a geometric framework that characterizes the relationship between 3D motion and shape in the presence of errors. We examine how the three-dimensional space recovered by a moving monocular observer, whose 3D motion is estimated with some error, is distorted. We characterize the space of distortions by its level sets, that is, we characterize the systematic distortion via a family of iso-distortion surfaces, which describes the locus over which the depths of points in the scene in view are distorted by the same multiplicative factor. The framework introduced in this way has a number of applications: Since the visible surfaces have positive depth (visibility constraint), by analyzing the geometry of the regions where the distortion factor is negative, that is, where the visibility constraint is violated, we make explicit situations which are likely to give rise to ambiguities in motion estimation, independent of the algorithm used. We provide a uniqueness analysis for 3D motion analysis from normal flow. We study the constraints on egomotion, object motion, and depth for an independently moving object to be detectable by a moving observer, and we offer a quantitative account of the precision needed in an inertial sensor for accurate estimation of 3D motion.}, isbn = {1077-3142}, doi = {10.1006/cviu.1997.0649}, url = {http://www.sciencedirect.com/science/article/pii/S1077314297906494}, author = {LoongFah Cheong and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {11952, title = {Effects of Errors in the Viewing Geometry on Shape Estimation* 1,* 2}, journal = {Computer Vision and Image Understanding}, volume = {71}, year = {1998}, month = {1998///}, pages = {356 - 372}, author = {Cheong, L. F and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {17130, title = {Emergent patterns of teaching/learning in electronic classrooms}, journal = {Educational Technology Research and Development}, volume = {46}, year = {1998}, month = {1998///}, pages = {23 - 42}, abstract = {Novel patterns of teaching/learning have emerged from faculty and students who use our three teaching/learning theaters at the University of Maryland, College Park. These fully-equipped electronic classrooms have been used by 74 faculty in 264 semester-long courses since the fall of 1991 with largely enthusiastic reception by both faculty and students. The designers of the teaching/learning theaters sought to provide a technologically rich environment and a support staff so that faculty could concentrate on changing the traditional lecture from its unidirectional information flow to a more collaborative activity. As faculty have evolved their personal styles in using the electronic classrooms, novel patterns of teaching/learning have emerged. In addition to enhanced lectures, we identified three common patterns: (a) active individual learning, (b) small-group collaborative learning, and (c) entire-class collaborative learning.}, isbn = {1042-1629}, url = {http://dx.doi.org/10.1007/BF02299671}, author = {Shneiderman, Ben and Borkowski,Ellen and Alavi,Maryam and Norman,Kent} } @article {15367, title = {Extension of Spherical Harmonic Method to RF Transient Regime}, journal = {Simulation of semiconductor processes and devices 1998: SISPAD 98}, year = {1998}, month = {1998///}, pages = {42 - 42}, abstract = {The space and time dependent electron Boltzmann transport equation (BTE)is solved sclf-consistently with the Poisson and transient hole current-continuity equation. A transient Spherical Harmonic expansion method is used to solve the BTE. By this method we can efficiently solve the BTE in the RF regime to observe how the complete distribution function responds to a rapid transient. Calculations on a BJT, which give the time dependent distribution function over a large energy range 0-3eV, throughout the device, as well as average quantities, require only 40 minutes CPU time on an Alpha workstation. }, author = {Lin,C. K. and Goldsman,N. and Chang,C. H. and Mayergoyz, Issak D and Aronowitz,S. and Dong,J. and Belova,N.} } @article {18640, title = {Frequentist and Bayesian Coverage Estimations for Stratified Fault-Injection}, journal = {DEPENDABLE COMPUTING AND FAULT TOLERANT SYSTEMS}, volume = {11}, year = {1998}, month = {1998///}, pages = {43 - 62}, abstract = {This paper addresses the problem of estimating the coverage of fault tolerancethrough statistical processing of observations collected in fault-injection experiments. In an earlier paper, we have studied various frequentist estimation methods based on simple sampling in the whole fault/activity input space and stratified sampling in a partitioned space. In this paper, Bayesian estimation methods are introduced for stratified sampling. Two methods are presented to obtain an approximation of the posterior distribution of the coverage by calculating its moments. The moments are then used to identify the type of the distribution in the Pearson distribution system, to estimate its parameters and to obtain the coverage confidence limit. Two hypothetical example systems are used to compare the validity and the conservatism of the Bayesian and frequentist estimations. It is shown that one Bayesian estimation method is valid for both examples and that its estimations are much less conservative than the frequentist ones. However, the Bayesian estimations for stratified sampling are still conservative compared to estimations for simple sampling. }, url = {http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.29.6784\&rep=rep1\&type=pdf}, author = {Michel Cukier and Arlat,J. and Powell,D.} } @article {14102, title = {Genetic nomenclature for Trypanosoma and Leishmania.}, journal = {Molecular and biochemical parasitology}, volume = {97}, year = {1998}, month = {1998///}, pages = {221 - 221}, author = {Clayton,C. and Adams,M. and Almeida,R. and Baltz,T. and Barrett,M. and Bastien,P. and Belli,S. and Beverley,S. and Biteau,N. and Blackwell,J. and others} } @article {20053, title = {Genetic nomenclature for Trypanosoma and Leishmania.}, journal = {Mol Biochem Parasitol}, volume = {97}, year = {1998}, month = {1998 Nov 30}, pages = {221-4}, keywords = {Animals, Leishmania, Terminology as Topic, Trypanosoma}, issn = {0166-6851}, author = {Clayton, C and Adams, M and Almeida, R and Baltz, T and Barrett, M and Bastien, P and Belli, S and Beverley, S and Biteau, N and Blackwell, J and Blaineau, C and Boshart, M and Bringaud, F and Cross, G and Cruz, A and Degrave, W and Donelson, J and El-Sayed, N and Fu, G and Ersfeld, K and Gibson, W and Gull, K and Ivens, A and Kelly, J and Vanhamme, L} } @article {14241, title = {On the geometry of visual correspondence}, year = {1998}, month = {1998/10/15/undef}, abstract = {Image displacement fieldsoptical flow fields, stereo disparity fields, normal flow fieldsdue to rigid motion possess a global geometric structure which is independent of the scene in view. Motion vectors of certain lengths and directions are constraine d to lie on the imaging surface at particular loci whose location and form depends solely on the 3D motion parameters. If optical flow fields or stereo disparity fields are considered, then equal vectors are shown to lie on conic sections. Similarly, for normal motion fields, equal vectors lie within regions whose boundaries also constitute conics. By studying various properties of these curves and regions and their relationships, a characterization of the structure of rigid motion fields is given. The go al of this paper is to introduce a concept underlying the global structure of image displacement fields. This concept gives rise to various constraints that could form the basis of algorithms for the recovery of visual information from multiple views.}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14944, title = {An On-line Variable Length Binary Encoding}, volume = {UMIACS-TR-95-39}, year = {1998}, month = {1998/10/15/}, institution = {Instititue for Advanced Computer Studies, Univ of Maryland, College Park}, abstract = {We present a methodology of an on-line variable-length binaryencoding of a set of integers. The basic principle of this methodology is to maintain the prefix property amongst the codes assigned on-line to a set of integers growing dynamically. The prefix property enables unique decoding of a string of elements from this set. To show the utility of this on-line variable length binary encoding, we apply this methodology to encode the LZW codes. Application of this encoding scheme significantly improves the compression achieved by the standard LZW scheme. This encoding can be applied in other compression schemes to encode the pointers using variable-length binary codes. (Also cross-referenced as UMIACS-TR-95-39) }, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/714}, author = {Acharya,Tinku and JaJa, Joseph F.} } @article {15611, title = {An optimal algorithm for approximate nearest neighbor searching fixed dimensions}, journal = {Journal of the ACM (JACM)}, volume = {45}, year = {1998}, month = {1998/11//}, pages = {891 - 923}, abstract = {Consider a set of S of n data points in real d-dimensional space, Rd, where distances are measured using any Minkowski metric. In nearest neighbor searching, we preprocess S into a data structure, so that given any query point q ∈ Rd, is the closest point of S to q can be reported quickly. Given any positive real \&egr;, data point p is a (1 +\&egr;)-approximate nearest neighbor of q if its distance from q is within a factor of (1 + \&egr;) of the distance to the true nearest neighbor. We show that it is possible to preprocess a set of n points in Rd in O(dn log n) time and O(dn) space, so that given a query point q ∈ Rd, and \&egr; > 0, a (1 + \&egr;)-approximate nearest neighbor of q can be computed in O(cd, \&egr; log n) time, where cd,\&egr;<=d 1 + 6d/e;d is a factor depending only on dimension and \&egr;. In general, we show that given an integer k >= 1, (1 + \&egr;)-approximations to the k nearest neighbors of q can be computed in additional O(kd log n) time.}, keywords = {Approximation algorithms, Box-decomposition trees, closet-point queries, nearest neighbor searching, post-office problem, priority search}, isbn = {0004-5411}, doi = {10.1145/293347.293348}, url = {http://doi.acm.org/10.1145/293347.293348}, author = {Arya,Sunil and Mount, Dave and Netanyahu,Nathan S. and Silverman,Ruth and Wu,Angela Y.} } @conference {14302, title = {Partial online cycle elimination in inclusion constraint graphs}, booktitle = {ACM SIGPLAN Conference on Programming Language Design and Implementation (PLDI)}, year = {1998}, month = {1998///}, author = {Aiken,A. and Fhndrich,M. and Foster, Jeffrey S. and Su,Z.} } @conference {14809, title = {Performance measurement using low perturbation and high precision hardware assists}, booktitle = {, The 19th IEEE Real-Time Systems Symposium, 1998. Proceedings}, year = {1998}, month = {1998/12/02/4}, pages = {379 - 388}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present the design and implementation of MultiKron PCI, a hardware performance monitor that can be plugged into any computer with a free PCI bus slot. The monitor provides a series of high-resolution timers, and the ability to monitor the utilization of the PCI bus. We also demonstrate how the monitor can be integrated with online performance monitoring tools such as the Paradyn parallel performance measurement tools to improve the overhead of key timer operations by a factor of 25. In addition, we present a series of case studies using the MultiKron hardware performance monitor to measure and tune high-performance parallel completing applications. By using the monitor, we were able to find and correct a performance bug in a popular implementation of the MPI message passing library that caused some communication primitives to run at one half their potential speed}, keywords = {Clocks, Computerized monitoring, Counting circuits, Debugging, Hardware, hardware performance monitor, high precision hardware assists, low perturbation, measurement, MPI message passing library, MultiKron hardware performance monitor, MultiKron PCI, NIST, online performance monitoring tools, Paradyn parallel performance measurement tools, PCI bus slot, performance bug, performance evaluation, performance measurement, program debugging, program testing, real-time systems, Runtime, Timing}, isbn = {0-8186-9212-X}, doi = {10.1109/REAL.1998.739771}, author = {Mink, A. and Salamon, W. and Hollingsworth, Jeffrey K and Arunachalam, R.} } @article {14742, title = {The PLAN system for building Active Networks}, journal = {University of Pennsylvania (February 27, 1998)}, year = {1998}, month = {1998///}, author = {Hicks, Michael W. and Kakkar,P. and Moore,J. T and Gunter,C. A and Alexander,D. S and Nettles,S.} } @article {12072, title = {Safety and security of programmable network infrastructures}, journal = {IEEE Communications Magazine}, volume = {36}, year = {1998}, month = {1998/10//}, pages = {84 - 92}, abstract = {Safety and security are two reliability properties of a system. A {\textquotedblleft}safe{\textquotedblright} system provides protection against errors of trusted users, while a {\textquotedblleft}secure{\textquotedblright} system protects against errors introduced by untrusted users. There is considerable overlap between mechanisms to support each property. Requirements for rapid service creation have stimulated the development of programmable network infrastructures, where end users or service providers can customize the properties of a network infrastructure while it continues to operate. A central concern of potential users of such systems is their reliability and, most specifically, their safety and security. In this article we explain the impact the network service model and architecture have on safety and security, and provide a model with which policies can be translated into restrictions of a general system. We illustrate these ideas with the Secure Active Network Environment (SANE) architecture, which provides a means of controlling access to the functions provided by any programmable infrastructure}, keywords = {Access control, error protection, IP networks, Multicast protocols, network architecture, network operating systems, network service model, operating system, Power system dynamics, Power system modeling, Power system reliability, programmable languages, programmable network infrastructures, programming languages, Proposals, Protection, reliability properties, Safety, Secure Active Network Environment, Security, security of data, service creation, service providers, Switches, telecommunication computing, telecommunication network reliability, Web and internet services}, isbn = {0163-6804}, doi = {10.1109/35.722141}, author = {Alexander,S. and Arbaugh, William A. and Keromytis,A. D and Smith,J. M} } @article {17951, title = {Salient Frame Detection for Molecular Dynamics Simulations}, journal = {Scientific Visualization: Interactions, Features, Metaphors}, volume = {2}, year = {1998}, month = {1998///}, pages = {160 - 175}, abstract = {Saliency-based analysis can be applied to time-varying 3D datasetsfor the purpose of summarization, abstraction, and motion analysis. As the sizes of time-varying datasets continue to grow, it becomes more and more difficult to comprehend vast amounts of data and information in a short period of time. Au- tomatically generated thumbnail images and previewing of time-varying datasets can help viewers explore and understand the datasets significantly faster as well as provide new insights. In this paper, we introduce a novel method for detect- ing salient frames for molecular dynamics simulations. Our method effectively detects crucial transitions in simulated mechanosensitive ion channel (MscS), in agreement with experimental data. }, author = {Kim,Y. and Patro,R. and Ip,C. Y and O{\textquoteright}Leary,D. P and Anishkin,A. and Sukharev,S. and Varshney, Amitabh} } @article {11919, title = {Scheduling Aperiodic and Sporadic Tasks in Hard Real-Time Systems}, journal = {Technical Reports from UMIACS, UMIACS-TR-97-44}, year = {1998}, month = {1998/10/15/}, abstract = {The stringent timing constraints as well as the functional correctnessare essential requirements of hard real-time systems. In such systems, scheduling plays a very important role in satisfying these constraints. The priority based scheduling schemes have been used commonly due to the simplicity of the scheduling algorithm. However, in the presence of task interdependencies and complex timing constraints, such scheduling schemes may not be appropriate due to the lack of an efficient mechanism to schedule them and to carry out the schedulability analysis. In contrast, the time based scheduling scheme may be used to schedule a set of tasks with greater degree of schedulability achieved at a cost of higher complexity of off-line scheduling. One of the drawbacks of currently available scheduling schemes, however, is known to be their inflexibility in dynamic environments where dynamic processes exist, such as aperiodic and sporadic processes. We develop and analyze scheduling schemes which efficiently provide the flexibility required in real-time systems for scheduling processes arriving dynamically. This enables static hard periodic processes and dynamic processes(aperiodic or sporadic) to be jointly scheduled. (Also cross-referenced as UMIACS-TR-97-44) }, keywords = {Technical Report}, url = {http://drum.lib.umd.edu/handle/1903/900}, author = {Choi,Seonho and Agrawala, Ashok K.} } @article {12101, title = {A secure active network environment architecture: realization in SwitchWare}, journal = {IEEE Network}, volume = {12}, year = {1998}, month = {1998/06//May}, pages = {37 - 45}, abstract = {An active network is a network infrastructure which is programmable on a per-user or even per-packet basis. Increasing the flexibility of such network infrastructures invites new security risks. Coping with these security risks represents the most fundamental contribution of active network research. The security concerns can be divided into those which affect the network as a whole and those which affect individual elements. It is clear that the element problems must be solved first, since the integrity of network-level solutions will be based on trust in the network elements. In this article we describe the architecture and implementation of a secure active network environment (SANE), which we believe provides a basis for implementing secure network-level solutions. We guarantee that a node begins operation in a trusted state with the AEGIS secure bootstrap architecture. We guarantee that the system remains in a trusted state by applying dynamic integrity checks in the network element{\textquoteright}s runtime system, using a novel naming system, and applying node-to-node authentication when needed. The construction of an extended LAN is discussed}, keywords = {access protocols, AEGIS secure bootstrap architecture, architecture, Authentication, Collaboration, Communication switching, dynamic integrity checks, extended LAN, Functional programming, implementation, integrity, Intelligent networks, IP networks, Local area networks, network infrastructure, network infrastructures, network operating systems, network-level solutions, node, node-to-node authentication, packet switching, Proposals, ramming system, SANE, secure active network environment architecture, security of data, Switches, SwitchWare, trusted state, Web and internet services}, isbn = {0890-8044}, doi = {10.1109/65.690960}, author = {Alexander,D. S and Arbaugh, William A. and Keromytis,A. D and Smith,J. M} } @article {12103, title = {Security for virtual private intranets}, journal = {Computer}, volume = {31}, year = {1998}, month = {1998/09//}, pages = {48 - 55}, abstract = {As telecommuting grows, businesses must consider security when extending their network environment to employees{\textquoteright} homes. Researchers at the University of Pennsylvania have addressed the problem with smart cards, operating system modifications, and network authentication. We note the distinction between trust and integrity: trust is determined through the verification of components and the dependencies among them, while integrity demonstrates that components haven{\textquoteright}t been modified. Thus integrity checking in a trustworthy system is about preserving an established trust or trust relationship. Our solution to the challenge of isolating functional roles that may share a single hardware platform is called secure identity based lending (SIBL). SIBL provides multiple personalities by partitioning the hard drive into n+1 partitions, where n is the number of supported personalities. All personalities use the system partition for core operating system components and shared applications. Each of the personalities is also associated with one of the remaining partitions, which are encrypted using a symmetric algorithm}, keywords = {businesses, Clouds, Companies, core operating system components, cryptography, Data security, employee homes, encryption, functional roles, hard drive, Home computing, home working, integrity checking, Internet, Local area networks, multiple personalities, network authentication, network environment, operating system modifications, Operating systems, Roads, secure identity based lending, security management, security of data, shared applications, SIBL, single hardware platform, smart cards, symmetric algorithm, system partition, telecommuting, Teleworking, trust relationship, trustworthy system, virtual private intranets}, isbn = {0018-9162}, doi = {10.1109/2.708450}, author = {Arbaugh, William A. and Davin,J. R and Farber,D. J and Smith,J. M} } @conference {11993, title = {Self-calibration from image derivatives}, booktitle = {Sixth International Conference on Computer Vision, 1998}, year = {1998}, month = {1998/01/04/7}, pages = {83 - 89}, publisher = {IEEE}, organization = {IEEE}, abstract = {This study investigates the problem of estimating the calibration parameters from image motion fields induced by a rigidly moving camera with unknown calibration parameters, where the image formation is modeled with a linear pinhole-camera model. The equations obtained show the flow to be clearly separated into a component due to the translation and the calibration parameters and a component due to the rotation and the calibration parameters. A set of parameters encoding the latter component are linearly related to the flow, and from these parameters the calibration can be determined. However, as for discrete motion, in the general case it is not possible, to decouple image measurements from two frames only into their translational and rotational component. Geometrically, the ambiguity takes the form of a part of the rotational component being parallel to the translational component, and thus the scene can be reconstructed only up to a projective transformation. In general, for a full calibration at least four successive image frames are necessary with the 3D-rotation changing between the measurements. The geometric analysis gives rise to a direct self-calibration method that avoids computation of optical flow or point correspondences and uses only normal flow measurements. In this technique the direction of translation is estimated employing in a novel way smoothness constraints. Then the calibration parameters are estimated from the rotational components of several flow fields using Levenberg-Marquardt parameter estimation, iterative in the calibration parameters only. The technique proposed does not require calibration objects in the scene or special camera motions and it also avoids the computation of exact correspondence. This makes it suitable for the calibration of active vision systems which have to acquire knowledge about their intrinsic parameters while they perform other tasks, or as a tool for analyzing image sequences in large video databases}, keywords = {3D-rotation, active vision, Calibration, CAMERAS, discrete motion, Encoding, Equations, image derivatives, image formation, image measurements, Image motion analysis, image motion fields, Image reconstruction, Image sequences, large video databases, Layout, Levenberg-Marquardt parameter estimation, linear pinhole-camera model, Motion estimation, Motion measurement, Optical computing, parameter estimation, projective transformation, rigidly moving camera, self-calibration, smoothness constraints, unknown calibration parameters}, isbn = {81-7319-221-9}, doi = {10.1109/ICCV.1998.710704}, author = {Brodsky, T. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14257, title = {Shape from video: Beyond the epipolar constraint}, journal = {Proceedings of the DARPA Image Understanding Workshop}, year = {1998}, month = {1998///}, author = {Brodsky, T. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @inbook {14258, title = {Simultaneous estimation of viewing geometry and structure}, booktitle = {Computer Vision {\textemdash} ECCV{\textquoteright}98Computer Vision {\textemdash} ECCV{\textquoteright}98}, series = {Lecture Notes in Computer Science}, volume = {1406}, year = {1998}, month = {1998///}, pages = {342 - 358}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Up to now, structure from motion algorithms proceeded in two well defined steps, where the first and most important step is recovering the rigid transformation between two views, and the subsequent step is using this transformation to compute the structure of the scene in view. This paper introduces a novel approach to structure from motion in which both aforementioned steps are accomplished in a synergistic manner. Existing approaches to 3D motion estimation are mostly based on the use of optic flow which however poses a problem at the locations of depth discontinuities. If we knew where depth discontinuities were, we could (using a multitude of approaches based on smoothness constraints) estimate accurately flow values for image patches corresponding to smooth scene patches; but to know the discontinuities requires solving the structure from motion problem first. In the past this dilemma has been addressed by improving the estimation of flow through sophisticated optimization techniques, whose performance often depends on the scene in view. In this paper we follow a different approach. The main idea is based on the interaction between 3D motion and shape which allows us to estimate the 3D motion while at the same time segmenting the scene. If we use a wrong 3D motion estimate to compute depth, then we obtain a distorted version of the depth function. The distortion, however, is such that the worse the motion estimate, the more likely we are to obtain depth estimates that are locally unsmooth, i.e., they vary more than the correct ones. Since local variability of depth is due either to the existence of a discontinuity or to a wrong 3D motion estimate, being able to differentiate between these two cases provides the correct motion, which yields the {\textquotedblleft}smoothest{\textquotedblright} estimated depth as well as the image location of scene discontinuities. Although no optic flow values are computed, we show that our algorithm is very much related to minimizing the epipolar constraint and we present a number of experimental results with real image sequences indicating the robustness of the method.}, isbn = {978-3-540-64569-6}, url = {http://dx.doi.org/10.1007/BFb0055677}, author = {Brodsk{\'y},Tom{\'a}{\v s} and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Burkhardt,Hans and Neumann,Bernd} } @article {12079, title = {The SwitchWare active network architecture}, journal = {IEEE Network}, volume = {12}, year = {1998}, month = {1998/06//May}, pages = {29 - 36}, abstract = {Active networks must balance the flexibility of a programmable network infrastructure against the safety and security requirements inherent in sharing that infrastructure. Furthermore, this balance must be achieved while maintaining the usability of the network. The SwitchWare active network architecture is a novel approach to achieving this balance using three layers: active packets, which contain mobile programs that replace traditional packets; active extensions, which provide services on the network elements and can be dynamically loaded; and a secure active router infrastructure, which forms a high-integrity base on which the security of the other layers depends. In addition to integrity checking and cryptography-based authentication, security in our architecture depends heavily on verification techniques from programming languages, such as strong type checking}, keywords = {active extensions, active packets, Authentication, Computer languages, Computer networks, cryptography, cryptography-based authentication, high-integrity base, integrity checking, IP networks, LAN interconnection, mobile programs, network operating systems, packet switching, programmable network infrastructure, programming languages, Protocols, Safety, safety requirements, scalability, secure active router infrastructure, Security, security requirements, services, strong type checking, Switches, SwitchWare active network architecture, telecommunication network routing, Tin, usability, verification techniques}, isbn = {0890-8044}, doi = {10.1109/65.690959}, author = {Alexander,D. S and Arbaugh, William A. and Hicks, Michael W. and Kakkar,P. and Keromytis,A. D and Moore,J. T and Gunter,C. A and Nettles,S. M and Smith,J. M} } @article {14745, title = {The switchware active network implementation}, journal = {The ML Workshop, International Conference on Functional Programming (ICFP)}, year = {1998}, month = {1998///}, author = {Alexander,D. S and Hicks, Michael W. and Kakkar,P. and Keromytis,A. D and Shaw,M. and Moore,J. T and Gunter,C. A and Jim,T. and Nettles,S. M and Smith,J. M} } @article {17833, title = {T2}, journal = {ACM SIGMOD Record}, volume = {27}, year = {1998}, month = {1998/03/01/}, pages = {58 - 66}, isbn = {01635808}, doi = {10.1145/273244.273264}, url = {http://dl.acm.org/citation.cfm?id=273264}, author = {Chang,Chialin and Acharya,Anurag and Sussman, Alan and Saltz,Joel} } @conference {11920, title = {Temporally determinate disk access (extended abstract): an experimental approach}, booktitle = {Proceedings of the 1998 ACM SIGMETRICS joint international conference on Measurement and modeling of computer systems}, series = {SIGMETRICS {\textquoteright}98/PERFORMANCE {\textquoteright}98}, year = {1998}, month = {1998///}, pages = {280 - 281}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-982-3}, doi = {10.1145/277851.277950}, url = {http://doi.acm.org/10.1145/277851.277950}, author = {Aboutabl,Mohamed and Agrawala, Ashok K. and Decotignie,Jean-Dominique} } @book {13028, title = {Trends in the Early Careers of Life Scientists}, year = {1998}, month = {1998///}, publisher = {National Academy Press}, organization = {National Academy Press}, address = {Washington, DC}, author = {Tilghman,S and Astin,HS and Brinkley,W and Chilton,MD and Cummings, Michael P. and Ehrenberg,RG and Fox,MF and Glenn,K and Green,PJ and Hans,S and Kelman,A and LaPidus,J and Levin,B and McIntosh,JR and Riecken,H and Stephen,PE} } @article {13029, title = {Trends in the early careers of life scientists - Preface and executive summary}, journal = {Mol Biol Cell}, volume = {9}, year = {1998}, month = {1998/11//}, pages = {3007 - 3015}, author = {Tilghman,S and Astin,HS and Brinkley,W and Chilton,MD and Cummings, Michael P. and Ehrenberg,RG and Fox,MF and Glenn,K and Green,PJ and Hans,S and Kelman,A and LaPidus,J and Levin,B and McIntosh,JR and Riecken,H and Stephen,PE} } @article {14870, title = {Uncertainty propagation in model-based recognition}, journal = {International Journal of Computer Vision}, volume = {27}, year = {1998}, month = {1998///}, pages = {127 - 159}, abstract = {Robust recognition systems require a careful understanding of the effects of error in sensed features. In model-based recognition, matches between model features and sensed image features typically are used to compute a model pose and then project the unmatched model features into the image. The error in the image features results in uncertainty in the projected model features. We first show how error propagates when poses are based on three pairs of 3D model and 2D image points. In particular, we show how to simply and efficiently compute the distributed region in the image where an unmatched model point might appear, for both Gaussian and bounded error in the detection of image points, and for both scaled-orthographic and perspective projection models. Next, we provide geometric and experimental analyses to indicate when this linear approximation will succeed and when it will fail. Then, based on the linear approximation, we show how we can utilize Linear Programming to compute bounded propagated error regions for any number of initial matches. Finally, we use these results to extend, from two-dimensional to three-dimensional objects, robust implementations of alignment, interpretation-tree search, and transformation clustering.}, doi = {10.1023/A:1007989016491}, author = {Alter,T. D. and Jacobs, David W.} } @inbook {14268, title = {The Video Yardstick}, booktitle = {Modelling and Motion Capture Techniques for Virtual EnvironmentsModelling and Motion Capture Techniques for Virtual Environments}, series = {Lecture Notes in Computer Science}, volume = {1537}, year = {1998}, month = {1998///}, pages = {144 - 158}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Given uncalibrated video sequences, how can we recover rich descriptions of the scene content, beyond two-dimensional (2D) measurements such as color/texture or motion fields {\textemdash} descriptions of shape and three-dimensional (3D) motion? This is the well known structure from motion (SFM) problem. Up to now, SFM algorithms proceeded in two well defined steps, where the first and most important step is recovering the rigid transformation between two views, and the subsequent step is using this transformation to compute the structure of the scene in view. This paper introduces a novel approach to structure from motion in which both steps are accomplished in a synergistic manner. It deals with the classical structure from motion problem considering a calibrated camera as well as the extension to an uncalibrated optical device. Existing approaches to estimation of the viewing geometry are mostly based on the use of optic flow, which, however, poses a problem at the locations of depth discontinuities. If we knew where depth discontinuities were, we could (using a multitude of approaches based on smoothness constraints) accurately estimate flow values for image patches corresponding to smooth scene patches; but to know the discontinuities requires solving the structure from motion problem first. In the past this dilemma has been addressed by improving the estimation of flow through sophisticated optimization techniques, whose performance often depends on the scene in view. In this paper we follow a different approach. We directly utilize the image derivatives and employ constraints which involve the 3D motion and shape of the scene, leading to a geometric and statistical estimation problem. The main idea is based on the interaction between 3D motion and shape which allows us to estimate the 3D motion while at the same time segmenting the scene. If we use a wrong 3D motion estimate to compute depth, we obtain a distorted version of the depth function. The distortion, however, is such that the worse the motion estimate, the more likely we are to obtain depth estimates that are locally unsmooth, i.e., they vary more than the correct ones. Since local variability of depth is due either to the existence of a discontinuity or to a wrong 3D motion estimate, being able to differentiate between these two cases provides the correct motion, which yields the {\textquotedblleft}smoothest{\textquotedblright} estimated depth as well as the image locations of scene discontinuities. We analyze the new constraints introduced by our approach and show their relationship to the minimization of the epipolar constraint, which becomes a special case of our theory. Finally, we present a number of experimental results with real image sequences indicating the robustness of our method and the improvement over traditional methods. The resulting system is a video yardstick that can be applied to any video sequence to recover first the calibration parameters of the camera that captured the video and, subsequently, the structure of the scene.}, isbn = {978-3-540-65353-0}, url = {http://dx.doi.org/10.1007/3-540-49384-0_12}, author = {Brodsk{\'y},Tom{\'a}{\v s} and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Magnenat-Thalmann,Nadia and Thalmann,Daniel} } @article {16192, title = {Viewing personal history records: A comparison of tabular format and graphical presentation using LifeLines}, journal = {Behaviour \& Information Technology}, volume = {17}, year = {1998}, month = {1998///}, pages = {249 - 262}, abstract = {Thirty-six participants used a static version of either LifeLines, a graphical interface, or a tabular representation to answer questions about a database of temporal personal history information. Results suggest that overall the LifeLines representation led to much faster response times, primarily for questions which involved interval comparisons and making intercategorical connections. A {\textquoteright}first impression{\textquoteright} test showed that LifeLines can reduce some of the biases of the tabular record summary. A post-experimental memory test led to significantly (p< 0.004) higher recall for LifeLines. Finally, simple interaction techniques are proposed to compensate for the problems of the static LifeLines display{\textquoteright}s ability to deal with precise dates, attribute coding and overlaps.Thirty-six participants used a static version of either LifeLines, a graphical interface, or a tabular representation to answer questions about a database of temporal personal history information. Results suggest that overall the LifeLines representation led to much faster response times, primarily for questions which involved interval comparisons and making intercategorical connections. A {\textquoteright}first impression{\textquoteright} test showed that LifeLines can reduce some of the biases of the tabular record summary. A post-experimental memory test led to significantly (p< 0.004) higher recall for LifeLines. Finally, simple interaction techniques are proposed to compensate for the problems of the static LifeLines display{\textquoteright}s ability to deal with precise dates, attribute coding and overlaps. }, isbn = {0144-929X}, doi = {10.1080/014492998119328}, url = {http://www.tandfonline.com/doi/abs/10.1080/014492998119328}, author = {Alonso,Diane Lindwarm and Rose,Anne and Plaisant, Catherine and Norman,Kent L} } @inbook {14277, title = {What is computed by structure from motion algorithms?}, booktitle = {Computer Vision {\textemdash} ECCV{\textquoteright}98}, series = {Lecture Notes in Computer Science}, volume = {1406}, year = {1998}, month = {1998///}, pages = {359 - 375}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {In the literature we find two classes of algorithms which, on the basis of two views of a scene, recover the rigid transformation between the views and subsequently the structure of the scene. The first class contains techniques which require knowledge of the correspondence or the motion field between the images and are based on the epipolar constraint. The second class contains so-called direct algorithms which require knowledge about the value of the flow in one direction only and are based on the positive depth constraint. Algorithms in the first class achieve the solution by minimizing a function representing deviation from the epipolar constraint while direct algorithms find the 3D motion that, when used to estimate depth, produces a minimum number of negative depth values. This paper presents a stability analysis of both classes of algorithms. The formulation is such that it allows comparison of the robustness of algorithms in the two classes as well as within each class. Specifically, a general statistical model is employed to express the functions which measure the deviation from the epipolar constraint and the number of negative depth values, and these functions are studied with regard to their topographic structure, specifically as regards the errors in the 3D motion parameters at the places representing the minima of the functions. The analysis shows that for algorithms in both classes which estimate all motion parameters simultaneously, the obtained solution has an error such that the projections of the translational and rotational errors on the image plane are perpendicular to each other. Furthermore, the estimated projection of the translation on the image lies on a line through the origin and the projection of the real translation.}, isbn = {978-3-540-64569-6}, url = {http://dx.doi.org/10.1007/BFb0055678}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Burkhardt,Hans and Neumann,Bernd} } @conference {12047, title = {Which shape from motion?}, booktitle = {Sixth International Conference on Computer Vision, 1998}, year = {1998}, month = {1998/01/04/7}, pages = {689 - 695}, publisher = {IEEE}, organization = {IEEE}, abstract = {In a practical situation, the rigid transformation relating different views is recovered with errors. In such a case, the recovered depth of the scene contains errors, and consequently a distorted version of visual space is computed. What then are meaningful shape representations that can be computed from the images? The result presented in this paper states that if the rigid transformation between different views is estimated in a way that gives rise to a minimum number of negative depth values, then at the center of the image affine shape can be correctly computed. This result is obtained by exploiting properties of the distortion function. The distortion model turns out to be a very powerful tool in the analysis and design of 3D motion and shape estimation algorithms, and as a byproduct of our analysis we present a computational explanation of psychophysical results demonstrating human visual space distortion from motion information}, keywords = {3D motion estimation, affine shape, Algorithm design and analysis, Computer vision, distorted version, distortion function, human visual space distortion, HUMANS, Image motion analysis, image representation, Information analysis, Layout, Motion analysis, Motion estimation, motion information, Psychology, rigid transformation, SHAPE, shape estimation, shape representations, State estimation, visual space}, isbn = {81-7319-221-9}, doi = {10.1109/ICCV.1998.710792}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {15561, title = {ANN: A library for approximate nearest neighbor searching}, journal = {CGC 2nd Annual Fall Workshop on Computational Geometry}, year = {1997}, month = {1997///}, author = {Mount, Dave and Arya,S.} } @conference {17554, title = {Approximating hyper-rectangles: learning and pseudo-random sets}, booktitle = {Proceedings of the twenty-ninth annual ACM symposium on Theory of computing}, series = {STOC {\textquoteright}97}, year = {1997}, month = {1997///}, pages = {314 - 323}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, keywords = {approximations of distributions, derandomization, explicit constructions, machine learning, multiple-instance learning, PAC learning, pseudorandomness, Ramsey graphs, random graphs, rectangles, sample complexity}, isbn = {0-89791-888-6}, doi = {10.1145/258533.258611}, url = {http://doi.acm.org/10.1145/258533.258611}, author = {Auer,Peter and Long,Philip M. and Srinivasan, Aravind} } @conference {14264, title = {The confounding of translation and rotation in reconstruction from multiple views}, booktitle = {, 1997 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1997. Proceedings}, year = {1997}, month = {1997/06/17/19}, pages = {250 - 256}, publisher = {IEEE}, organization = {IEEE}, abstract = {If 3D rigid motion is estimated with some error a distorted version of the scene structure will in turn be computed. Of computational interest are these regions in space where the distortions are such that the depths become negative, because in order to be visible the scene has to lie in front of the image. The stability analysis for the structure-from-motion problem presented in this paper investigates the optimal relationship between the errors in the estimated translational and rotational parameters of a rigid motion, that results in the estimation of a minimum number of negative depth values. The input used is the value of the flow along some direction, which is more general than optic flow or correspondence. For a planar retina it is shown that the optimal configuration is achieved when the projections of the translational and rotational errors on the image plane are perpendicular. Furthermore, the projection of the actual and the estimated translation lie on a line passing through the image center. For a spherical retina given a rotational error, the optimal translation is the correct one, while given a translational error. The optimal rotational error is normal to the translational one at an equal distance from the real and estimated translations. The proofs, besides illuminating the confounding of translation and rotation in structure from motion, have an important application to ecological optics, explaining differences of planar and spherical eye or camera designs in motion and shape estimation}, keywords = {3D rigid motion, CAMERAS, ecological optics, Error correction, Image motion analysis, Image reconstruction, Layout, Motion analysis, Motion estimation, multiple views, Optical distortion, optimal configuration, optimal rotational error, planar retina, RETINA, rigid motion, ROTATION, scene structure, shape estimation, spherical retina, Stability analysis, structure-from-motion problem, translation}, isbn = {0-8186-7822-4}, doi = {10.1109/CVPR.1997.609328}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {12017, title = {DIRECT MOTION PERCEPTION}, journal = {Visual navigation: from biological systems to unmanned ground vehicles}, year = {1997}, month = {1997///}, pages = {135 - 135}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {16461, title = {The distributed information search component (disco) and the world wide web}, booktitle = {Proc. of ACM SIGMOD Conf. on Management of Data}, year = {1997}, month = {1997///}, author = {Naacke,H. and Kapitskaia,O. and Tomasic,A. and Bonnet,P. and Raschid, Louiqa and Amouroux,R.} } @article {18382, title = {Editorial: Evaluation and assessment in software engineering}, journal = {Information and Software Technology}, volume = {39}, year = {1997}, month = {1997///}, author = {Kitchenham,B. and Brereton,P. and Budgen,D. and Linkman,S. and Almstrum,V. L and Pfleeger,S. L and Zelkowitz, Marvin V and Wallace,D.} } @article {14220, title = {Families of Stationary Patterns Producing Illusory Movement: Insights into the Visual System}, journal = {Proceedings of the Royal Society of London. Series B: Biological SciencesProc. R. Soc. Lond. B}, volume = {264}, year = {1997}, month = {1997/06/22/}, pages = {795 - 806}, abstract = {A computational explanation of the illusory movement experienced upon extended viewing of Enigma, a static figure painted by Leviant, is presented. The explanation relies on a model for the interpretation of three{\textendash}dimensional motion information contained in retinal motion measurements. This model shows that the Enigma figure is a special case of a larger class of figures exhibiting the same illusory movement and these figures are introduced here. Our explanation suggests that eye movements and/or accommodation changes cause weak retinal motion signals, which are interpreted by higher{\textendash}level processes in a way that gives rise to these illusions, and proposes a number of new experiments to unravel the functional structure of the motion pathway.}, isbn = {0962-8452, 1471-2954}, doi = {10.1098/rspb.1997.0112}, url = {http://rspb.royalsocietypublishing.org/content/264/1383/795}, author = {Ferm{\"u}ller, Cornelia and Pless,Robert and Aloimonos, J.} } @article {14242, title = {On the Geometry of Visual Correspondence}, journal = {International Journal of Computer Vision}, volume = {21}, year = {1997}, month = {1997///}, pages = {223 - 247}, abstract = {Image displacement fields{\textemdash}optical flow fields, stereo disparity fields, normal flow fields{\textemdash}due to rigid motion possess a global geometric structure which is independent of the scene in view. Motion vectors of certain lengths and directions are constrained to lie on the imaging surface at particular loci whose location and form depends solely on the 3D motion parameters. If optical flow fields or stereo disparity fields are considered, then equal vectors are shown to lie on conic sections. Similarly, for normal motion fields, equal vectors lie within regions whose boundaries also constitute conics. By studying various properties of these curves and regions and their relationships, a characterization of the structure of rigid motion fields is given. The goal of this paper is to introduce a concept underlying the global structure of image displacement fields. This concept gives rise to various constraints that could form the basis of algorithms for the recovery of visual information from multiple views.}, isbn = {0920-5691}, url = {http://dx.doi.org/10.1023/A:1007951901001}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @inbook {14265, title = {The geometry of visual space distortion}, booktitle = {Algebraic Frames for the Perception-Action CycleAlgebraic Frames for the Perception-Action Cycle}, series = {Lecture Notes in Computer Science}, volume = {1315}, year = {1997}, month = {1997///}, pages = {249 - 277}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {The encounter of perception and action happens at the intermediate representations of space-time. In many of the computational models employed in the past, it has been assumed that a metric representation of physical space can be derived by visual means. Psychophysical experiments, as well as computational considerations, can convince us that the perception of space and shape has a much more complicated nature, and that only a distorted version of actual, physical space can be computed. This paper develops a computational geometric model that explains why such distortion might take place. The basic idea is that, both in stereo and motion, we perceive the world from multiple views. Given the rigid transformation between the views and the properties of the image correspondence, the depth of the scene can be obtained. Even a slight error in the rigid transformation parameters causes distortion of the computed depth of the scene. The unified framework introduced here describes this distortion in computational terms. We characterize the space of distortions by its level sets, that is, we characterize the systematic distortion via a family of iso-distortion surfaces which describes the locus over which depths are distorted by some multiplicative factor. Clearly, functions of the distorted space exhibiting some sort of invariance, produce desirable representations for biological and artificial systems [13]. Given that humans{\textquoteright} estimation of egomotion or estimation of the extrinsic parameters of the stereo apparatus is likely to be imprecise, the framework is used to explain a number of psychophysical experiments on the perception of depth from motion or stereo.}, isbn = {978-3-540-63517-8}, url = {http://dx.doi.org/10.1007/BFb0017872}, author = {Ferm{\"u}ller, Cornelia and LoongFah Cheong and Aloimonos, J.}, editor = {Sommer,Gerald and Koenderink,Jan} } @article {17612, title = {Improved parallel approximation of a class of integer programming problems}, journal = {Algorithmica}, volume = {17}, year = {1997}, month = {1997///}, pages = {449 - 462}, abstract = {We present a method to derandomize RNC algorithms, converting them to NC algorithms. Using it, we show how to approximate a class of NP-hard integer programming problems in NC , to within factors better than the current-best NC algorithms (of Berger and Rompel and Motwani et al. ); in some cases, the approximation factors are as good as the best-known sequential algorithms, due to Raghavan. This class includes problems such as global wire-routing in VLSI gate arrays and a generalization of telephone network planning in SONET rings. Also for a subfamily of the {\textquotedblleft}packing{\textquotedblright} integer programs, we provide the first NC approximation algorithms; this includes problems such as maximum matchings in hypergraphs, and generalizations. The key to the utility of our method is that it involves sums of superpolynomially many terms, which can however be computed in NC ; this superpolynomiality is the bottleneck for some earlier approaches, due to Berger and Rompel and Motwani et al.}, isbn = {0178-4617}, url = {http://dx.doi.org/10.1007/BF02523683}, author = {Alon,N. and Srinivasan, Aravind} } @article {16698, title = {A Language Identification Application Built on the Java Client/Server Platform}, journal = {From Research to Commercial Applications: Making NLP Work in Practice}, year = {1997}, month = {1997///}, pages = {43 - 47}, author = {Adams,G. and Resnik, Philip} } @article {14627, title = {Local Rules for Protein Folding on a Triangular Lattice and Generalized Hydrophobicity in the HP Model}, journal = {Journal of Computational Biology}, volume = {4}, year = {1997}, month = {1997/01//}, pages = {275 - 296}, abstract = {We consider the problem of determining the three-dimensional folding of a protein given its one-dimensional amino acid sequence. We use the HP model for protein folding proposed by Dill (1985), which models protein as a chain of amino acid residues that are either hydrophobic or polar, and hydrophobic interactions are the dominant initial driving force for the protein folding. Hart and Istrail (1996a) gave approximation algorithms for folding proteins on the cubic lattice under the HP model. In this paper, we examine the choice of a lattice by considering its algorithmic and geometric implications and argue that the triangular lattice is a more reasonable choice. We present a set of folding rules for a triangular lattice and analyze the approximation ratio they achieve. In addition, we introduce a generalization of the HP model to account for residues having different levels of hydrophobicity. After describing the biological foundation for this generalization, we show that in the new model we are able to achieve similar constant factor approximation guarantees on the triangular lattice as were achieved in the standard HP model. While the structures derived from our folding rules are probably still far from biological reality, we hope that having a set of folding rules with different properties will yield more interesting folds when combined.}, isbn = {1066-5277, 1557-8666}, doi = {10.1089/cmb.1997.4.275}, url = {http://www.liebertonline.com/doi/abs/10.1089/cmb.1997.4.275}, author = {Agarwala,Richa and Batzoglou,Serafim and Dan{\v c}{\'\i}K,Vlado and Decatur,Scott E. and Hannenhalli, Sridhar and Farach,Martin and Muthukrishnan,S. and Skiena,Steven} } @article {12324, title = {Optimized software synthesis for synchronous dataflow}, journal = {Proceedings of the ASAP97}, year = {1997}, month = {1997///}, author = {Bhattacharyya, Shuvra S. and America,H.} } @inbook {11921, title = {Regulation of Cable Television}, booktitle = {The Froehlich/Kent Encyclopedia of Telecommunications: Volume 15-Radio Astronomy to Submarine Cable SystemsThe Froehlich/Kent Encyclopedia of Telecommunications: Volume 15-Radio Astronomy to Submarine Cable Systems}, volume = {15}, year = {1997}, month = {1997///}, pages = {84 - 84}, author = {Agrawala, Ashok K. and CILINGIROGLU,A. and Lee,S.} } @article {17879, title = {Run-time and compiler support for programming in adaptive parallel environments}, journal = {Scientific Programming}, volume = {6}, year = {1997}, month = {1997///}, pages = {215 - 227}, author = {Edjlali,G. and Agrawal,G. and Sussman, Alan and Humphries,J. and Saltz, J.} } @conference {12092, title = {A secure and reliable bootstrap architecture}, booktitle = {, 1997 IEEE Symposium on Security and Privacy, 1997. Proceedings}, year = {1997}, month = {1997/05/04/7}, pages = {65 - 71}, publisher = {IEEE}, organization = {IEEE}, abstract = {In a computer system, the integrity of lower layers is typically treated as axiomatic by higher layers. Under the presumption that the hardware comprising the machine (the lowest layer) is valid, the integrity of a layer can be guaranteed if and only if: (1) the integrity of the lower layers is checked and (2) transitions to higher layers occur only after integrity checks on them are complete. The resulting integrity {\textquotedblleft}chain{\textquotedblright} inductively guarantees system integrity. When these conditions are not met, as they typically are not in the bootstrapping (initialization) of a computer system, no integrity guarantees can be made, yet these guarantees are increasingly important to diverse applications such as Internet commerce, security systems and {\textquotedblleft}active networks{\textquotedblright}. In this paper, we describe the AEGIS architecture for initializing a computer system. It validates integrity at each layer transition in the bootstrap process. AEGIS also includes a recovery process for integrity check failures, and we show how this results in robust systems}, keywords = {active networks, AEGIS architecture, bootstrap architecture, Computer architecture, computer bootstrapping, data integrity, Distributed computing, Hardware, hardware validity, initialization, integrity chain, integrity check failures, Internet, Internet commerce, IP networks, Laboratories, lower-layer integrity, Microprogramming, Operating systems, recovery process, reliability, robust systems, Robustness, Security, security of data, software reliability, system integrity guarantees, system recovery, transitions, Virtual machining}, isbn = {0-8186-7828-3}, doi = {10.1109/SECPRI.1997.601317}, author = {Arbaugh, William A. and Farber,D. J and Smith,J. M} } @article {15562, title = {Testing simple polygons}, journal = {Computational Geometry}, volume = {8}, year = {1997}, month = {1997/07//}, pages = {97 - 114}, abstract = {We consider the problem of verifying a simple polygon in the plane using {\textquotedblleft}test points{\textquotedblright}. A test point is a geometric probe that takes as input a point in Euclidean space, and returns {\textquotedblleft}+{\textquotedblright} if the point is inside the object being probed or {\textquotedblleft}-{\textquotedblright} if it is outside. A verification procedure takes as input a description of a target object, including its location and orientation, and it produces a set of test points that are used to verify whether a test object matches the description. We give a procedure for verifying an n-sided, non-degenerate, simple target polygon using 5n test points. This testing strategy works even if the test polygon has n + 1 vertices, and we show a lower bound of 3n + 1 test points for this case. We also give algorithms using O(n) test points for simple polygons that may be degenerate and for test polygons that may have up to n + 2 vertices. All of these algorithms work for polygons with holes. We also discuss extensions of our results to higher dimensions.}, keywords = {probing, Testing, Verifying}, isbn = {0925-7721}, doi = {10.1016/S0925-7721(96)00015-6}, url = {http://www.sciencedirect.com/science/article/pii/S0925772196000156}, author = {Arkin,Esther M. and Belleville,Patrice and Mitchell,Joseph S.B. and Mount, Dave and Romanik,Kathleen and Salzberg,Steven and Souvaine,Diane} } @conference {17856, title = {Titan: a High-Performance Remote-sensing Database}, booktitle = {In Proceedings of the 1997 International Conference on Data Engineering}, year = {1997}, month = {1997///}, author = {Bongki,C. C and Chang,C. and Moon,B. and Acharya, A. and Shock,C. and Sussman, Alan and Saltz, J.} } @inbook {14269, title = {Toward motion picture grammars}, booktitle = {Computer Vision {\textemdash} ACCV{\textquoteright}98Computer Vision {\textemdash} ACCV{\textquoteright}98}, series = {Lecture Notes in Computer Science}, volume = {1352}, year = {1997}, month = {1997///}, pages = {283 - 290}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {We are interested in processing video data for the purpose of solving a variety of problems in video search, analysis, indexing, browsing and compression. Instead of concentrating on a particular problem, in this paper we present a framework for developing video applications. Our basic thesis is that video data can be represented at a higher level of abstraction as a string generated by a grammar, termed motion picture grammar. The rules of that grammar relate different spatiotemporal representations of the video content and, in particular, representations of action.}, isbn = {978-3-540-63931-2}, url = {http://dx.doi.org/10.1007/3-540-63931-4_228}, author = {Bolle,Ruud and Aloimonos, J. and Ferm{\"u}ller, Cornelia}, editor = {Chin,Roland and Pong,Ting-Chuen} } @article {12016, title = {Visual Navigation: Flies, Bees, and UGV{\textquoteright}s}, journal = {Visual navigation: from biological systems to unmanned ground vehicles}, year = {1997}, month = {1997///}, author = {Aloimonos, J.} } @book {12050, title = {Visual navigation: from biological systems to unmanned ground vehicles}, year = {1997}, month = {1997///}, publisher = {Lawrence Erlbaum}, organization = {Lawrence Erlbaum}, author = {Aloimonos, J.} } @article {14275, title = {Visual space distortion}, journal = {Biological Cybernetics}, volume = {77}, year = {1997}, month = {1997///}, pages = {323 - 337}, abstract = {We are surrounded by surfaces that we perceive by visual means. Understanding the basic principles behind this perceptual process is a central theme in visual psychology, psychophysics, and computational vision. In many of the computational models employed in the past, it has been assumed that a metric representation of physical space can be derived by visual means. Psychophysical experiments, as well as computational considerations, can convince us that the perception of space and shape has a much more complicated nature, and that only a distorted version of actual, physical space can be computed. This paper develops a computational geometric model that explains why such distortion might take place. The basic idea is that, both in stereo and motion, we perceive the world from multiple views. Given the rigid transformation between the views and the properties of the image correspondence, the depth of the scene can be obtained. Even a slight error in the rigid transformation parameters causes distortion of the computed depth of the scene. The unified framework introduced here describes this distortion in computational terms. We characterize the space of distortions by its level sets, that is, we characterize the systematic distortion via a family of iso-distortion surfaces which describes the locus over which depths are distorted by some multiplicative factor. Given that humans{\textquoteright} estimation of egomotion or estimation of the extrinsic parameters of the stereo apparatus is likely to be imprecise, the framework is used to explain a number of psychophysical experiments on the perception of depth from motion or stereo.}, isbn = {0340-1200}, url = {http://dx.doi.org/10.1007/s004220050393}, author = {Ferm{\"u}ller, Cornelia and LoongFah Cheong and Aloimonos, J.} } @article {15580, title = {Accounting for boundary effects in nearest-neighbor searching}, journal = {Discrete \& Computational Geometry}, volume = {16}, year = {1996}, month = {1996///}, pages = {155 - 176}, abstract = {Given n data points ind-dimensional space, nearest-neighbor searching involves determining the nearest of these data points to a given query point. Most averagecase analyses of nearest-neighbor searching algorithms are made under the simplifying assumption thatd is fixed and thatn is so large relative tod thatboundary effects can be ignored. This means that for any query point the statistical distribution of the data points surrounding it is independent of the location of the query point. However, in many applications of nearest-neighbor searching (such as data compression by vector quantization) this assumption is not met, since the number of data pointsn grows roughly as 2 d .Largely for this reason, the actual performances of many nearest-neighbor algorithms tend to be much better than their theoretical analyses would suggest. We present evidence of why this is the case. We provide an accurate analysis of the number of cells visited in nearest-neighbor searching by the bucketing andk-d tree algorithms. We assumem dpoints uniformly distributed in dimensiond, wherem is a fixed integer >=2. Further, we assume that distances are measured in theL $\infty$ metric. Our analysis is tight in the limit asd approaches infinity. Empirical evidence is presented showing that the analysis applies even in low dimensions.}, doi = {10.1007/BF02716805}, author = {Arya,S. and Mount, Dave and Narayan,O.} } @inbook {15735, title = {Conjugate Gradients and Related KMP Algorithms: The Beginnings}, booktitle = {Linear and Nonlinear Conjugate Gradient-Related MethodsLinear and Nonlinear Conjugate Gradient-Related Methods}, year = {1996}, month = {1996///}, pages = {1 - 8}, publisher = {SIAM}, organization = {SIAM}, address = {Philadelphia}, author = {O{\textquoteright}Leary, Dianne P.}, editor = {Adams,Loyce and Nazareth,J. L.} } @inbook {14213, title = {Directions of motion fields are hardly ever ambiguous}, booktitle = {Computer Vision {\textemdash} ECCV {\textquoteright}96Computer Vision {\textemdash} ECCV {\textquoteright}96}, series = {Lecture Notes in Computer Science}, volume = {1065}, year = {1996}, month = {1996///}, pages = {119 - 128}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {Recent literature [7, 10, 11, 9, 13, 17] provides a number of results regarding uniqueness aspects of motion fields and exact image displacements due to 3-D rigid motion. Here instead of the full motion field we consider only the direction of the motion field due to a rigid motion and ask what can we say about the three-dimensional motion information contained in it. This paper provides a geometric analysis of this question based solely on the fact that the depth of the surfaces in view is positive (i.e. that the surface in view is in front of the camera). With this analysis we thus offer a theoretical foundation for image constraints employing only the sign of flow in various directions and provide a solid basis for their utilization in addressing 3D dynamic vision problems.}, isbn = {978-3-540-61123-3}, url = {http://dx.doi.org/10.1007/3-540-61123-1_132}, author = {Brodsky, Tomas and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Buxton,Bernard and Cipolla,Roberto} } @article {12066, title = {Early detection of independent motion from active control of normal image flow patterns}, journal = {Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on}, volume = {26}, year = {1996}, month = {1996///}, pages = {42 - 52}, author = {Sharma, R. and Aloimonos, J.} } @conference {15355, title = {Effect of chemical reactions on the decay of isotropic homogeneous turbulence}, booktitle = {AIAA Fluid Dynamics Conference}, volume = {96}, year = {1996}, month = {1996///}, pages = {131 - 131}, address = {New Orleans, LA}, abstract = {Direct numerical simulations, (DNS), are usedto simulate the decay of an isotopic, turbulent, chemically-reacting flow at high temperatures. The independent parameters that govern the physical pro- cess are introduced. The different effects from each of the parameters in the flow are explained by using the results from the DNS. It is found that there is a feedback mechanism between the chemical reactions and the turbulent motion. This feedback is positive for exothermic reactions and negative for endother- mic reactions. }, author = {Aguirre,M and Candler,G. V and Martin, M.P} } @article {18644, title = {Estimation of time-dependent coverage}, volume = {96466}, year = {1996}, month = {1996///}, institution = {LAAS-CNRS: Laboratory for analysis and architecture of systems}, address = {Toulouse, France}, abstract = {It is well-known that the dependability that can be achieved by a fault-tolerant systemis particularly sensitive to both the asymptotic value of coverage and to the time distribution of coverage. However, most previous work on coverage evaluation by statistical processing of the results of fault-injection experiments has only been concerned with estimating asymptotic coverage. In this paper, we tackle the problem of estimating the parameters of models that also account for coverage latency. After discussing some data sets resulting from fault-injection experiments on practical systems, we propose a series of coverage latency models that might be considered to account for the observed phenomena in a system dependability evaluation. We consider both exponential and non-exponential models, and assess their pertinence by means of a sensitivity study. We confirm previous results that latency can have an extremely important effect on the achievable dependability. We also show that the shape of the latency distribution has only a minor impact in the practical case of systems with high asymptotic coverage. A simple action model based on an exponential latency distribution is therefore proposed. We show how worst-case confidence limits can be obtained for the parameters of this action model and study the effects of data truncation that are unavoidable in any practical measurements of latency. We conclude by a critical assessment of the proposed estimation technique and a demonstration of its application to practical data sets. }, url = {http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.48.8253\&rep=rep1\&type=pdf}, author = {Arlat,D. P. M. C. J. and Crouzet,Y.} } @book {14225, title = {Interaction between 3D Shape and Motion: Theory and Applications}, year = {1996}, month = {1996/06//}, publisher = {Computer Vision Laboratory, University of Maryland}, organization = {Computer Vision Laboratory, University of Maryland}, abstract = {Research during the past few years has attempted to characterize the errors that arise in computing 3D motion (egomotion estimation) and in a scene{\textquoteright}s structure (structure from motion) from a sequence of images acquired by a moving sensor. This paper presents a new geometric framework that characterizes how the three-dimensional space recovered by a moving monocular observer, whose 3D motion is estimated with some error, is distorted. We characterize the space of distortions by its level sets, that is, by a family of iso-distortion surfaces, each of which describes the locus over which the depths of points in the scene are distorted by the same multiplicative factor. By analyzing the geometry of the regions where the distortion factor is negative, that is, where the visibility constraint is violated, we make explicit situations which are likely to give rise to ambiguities in motion estimation. We also apply our approach to a uniqueness analysis for 3D motion analysis from normal flow; we study the constraints on egomotion, object motion and depth for an independently moving object to be detectable by a moving observer; and we offer a quantitative account of the precision needed in an inertial sensor for accurate estimation of 3D motion.}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J. and Cheong,L.} } @article {11922, title = {MARUTI at ARDEC.}, year = {1996}, month = {1996/07/11/}, institution = {University of Maryland, College Park}, abstract = {This is the final report of the effort undertaken at the University of Maryland, sponsored by DARPA under the DSSA (Domain Specific Software Architectures) program. The activities reported here required an active cooperation and collaboration of the US Army ARDEC (Automation and Robotics Laboratory) and the University of Maryland (Computer Science Department). The principal investigators have been developing a hard real time operating system, MARUTI, at the University of Maryland. The goal of this effort was to demonstrate the applicability and usefulness of this operating system for exercising PID control at high frequency on the ATB1000 testbed which has been used to simulate a gun turret. A MARUTI based PID controller which operates at 200 HZ and 400 HZ was successfully implemented to demonstrate this feasibility.}, keywords = {*OPERATING SYSTEMS(COMPUTERS), Computer architecture, COMPUTER PROGRAMMING AND SOFTWARE, CONTROL, Cooperation, DSSA(DOMAIN SPECIFIC SOFTWARE ARCHITECTURE), GUN TURRETS., HIGH FREQUENCY, Laboratories, REAL TIME, ROBOTICS, TEST BEDS}, url = {http://stinet.dtic.mil/oai/oai?\&verb=getRecord\&metadataPrefix=html\&identifier=ADA315086}, author = {Agrawala, Ashok K. and Tripathi,Satish K} } @article {14945, title = {An on-line variable-length binary encoding of text}, journal = {Information Sciences}, volume = {94}, year = {1996}, month = {1996/10//}, pages = {1 - 22}, abstract = {We present a methodology for on-line variable-length binary encoding of a dynamically growing set of integers. Our encoding maintains the prefix property that enables unique decoding of a string of integers from the set. In order to develop the formalism of this on-line binary encoding, we define a unique binary tree data structure called the {\textquotedblleft}phase in binary tree.{\textquotedblright} To show the utility of this on-line variable-length binary encoding, we apply this methodology to encode the pointers generated by the LZW algorithm. The experimental results obtained illustrate the superior performance of our algorithm compared to the most widely used algorithms. This on-line variable-length binary encoding can be applied in other dictionary-based text compression schemes as well to effectively encode the output pointers to enhance the compression ratio.}, isbn = {0020-0255}, doi = {10.1016/0020-0255(96)00089-8}, url = {http://www.sciencedirect.com/science/article/pii/0020025596000898}, author = {Acharya,Tinku and JaJa, Joseph F.} } @article {14243, title = {Ordinal representations of visual space}, journal = {Proc. Image Understanding Workshop}, year = {1996}, month = {1996///}, pages = {897 - 904}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {17872, title = {Runtime coupling of data-parallel programs}, booktitle = {Proceedings of the 10th international conference on Supercomputing}, year = {1996}, month = {1996///}, pages = {229 - 236}, author = {Ranganathan,M. and Acharya, A. and Edjlali,G. and Sussman, Alan and Saltz, J.} } @conference {17940, title = {Simplification envelopes}, booktitle = {Proceedings of the 23rd annual conference on Computer graphics and interactive techniques}, series = {SIGGRAPH {\textquoteright}96}, year = {1996}, month = {1996///}, pages = {119 - 128}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, keywords = {geometric modeling, hierarchical approximation, levels-of-detail generation, model simplification, offsets, shape approximation}, isbn = {0-89791-746-4}, doi = {10.1145/237170.237220}, url = {http://doi.acm.org/10.1145/237170.237220}, author = {Cohen,Jonathan and Varshney, Amitabh and Manocha,Dinesh and Turk,Greg and Weber,Hans and Agarwal,Pankaj and Brooks,Frederick and Wright,William} } @inbook {14259, title = {Spatiotemporal representations for visual navigation}, booktitle = {Computer Vision {\textemdash} ECCV {\textquoteright}96Computer Vision {\textemdash} ECCV {\textquoteright}96}, series = {Lecture Notes in Computer Science}, volume = {1064}, year = {1996}, month = {1996///}, pages = {671 - 684}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {The study of visual navigation problems requires the integration of visual processes with motor control. Most essential in approaching this integration is the study of appropriate spatio-temporal representations which the system computes from the imagery and which serve as interfaces to all motor activities. Since representations resulting from exact quantitative reconstruction have turned out to be very hard to obtain, we argue here for the necessity of representations which can be computed easily, reliably and in real time and which recover only the information about the 3D world which is really needed in order to solve the navigational problems at hand. In this paper we introduce a number of such representations capturing aspects of 3D motion and scene structure which are used for the solution of navigational problems implemented in visual servo systems. In particular, the following three problems are addressed: (a) to change the robot{\textquoteright}s direction of motion towards a fixed direction, (b) to pursue a moving target while keeping a certain distance from the target, and (c) to follow a wall-like perimeter. The importance of the introduced representations lies in the following: \&$\#$x96; They can be extracted using minimal visual information, in particular the sign of flow measurements or the the first order spatiotemporal derivatives of the image intensity function. In that sense they are direct representations needing no intermediate level of computation such as correspondence. \&$\#$x96; They are global in the sense that they represent how three-dimensional information is globally encoded in them. Thus, they are robust representations since local errors do not affect them. \&$\#$x96; Usually, from sequences of images, three-dimensional quantities such as motion and shape are computed and used as input to control processes. The representations discussed here are given directly as input to the control procedures, thus resulting in a real time solution.}, isbn = {978-3-540-61122-6}, url = {http://dx.doi.org/10.1007/BFb0015577}, author = {LoongFah Cheong and Ferm{\"u}ller, Cornelia and Aloimonos, J.}, editor = {Buxton,Bernard and Cipolla,Roberto} } @inbook {18680, title = {On stratified sampling for high coverage estimations}, booktitle = {Dependable Computing {\textemdash} EDCC-2}, series = {Lecture Notes in Computer Science}, volume = {1150}, year = {1996}, month = {1996///}, pages = {35 - 54}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, abstract = {This paper addresses the problem of estimating the coverage of a fault tolerance mechanism through statistical processing of observations collected in faultinjection experiments. In an earlier paper, several techniques for sampling the fault/activity input space of a fault tolerance mechanism were presented. Various estimators based on simple sampling in the whole space and stratified sampling in a partitioned space were studied; confidence limits were derived based on a normal approximation. In this paper, the validity of this approximation is analyzed, especially for high coverage systems. The theory of confidence regions is then introduced to estimate the coverage without approximation when, for practical reasons, stratification is used. Three statistics are considered for defining confidence regions. It is shown that one of these statistics {\textemdash} a vectorial statistic {\textemdash} is often more conservative than the other two. However, only the vectorial statistic is computationally tractable. The results obtained are compared with those based on approximation by means of three hypothetical example systems.}, keywords = {Computer science}, isbn = {978-3-540-61772-3}, url = {http://www.springerlink.com/content/7t2w2u472601h730/abstract/}, author = {Powell,David and Michel Cukier and Arlat,Jean}, editor = {Hlawiczka,Andrzej and Silva,Jo{\~a}o and Simoncini,Luca} } @article {11997, title = {The synthesis of vision and action}, journal = {Exploratory vision: the active eye}, year = {1996}, month = {1996///}, pages = {205 - 205}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {14796, title = {Tuning the performance of I/O-intensive parallel applications}, booktitle = {Proceedings of the fourth workshop on I/O in parallel and distributed systems: part of the federated computing research conference}, year = {1996}, month = {1996///}, pages = {15 - 27}, author = {Acharya, A. and Uysal, M. and Bennett, R. and Mendelson, A. and Beynon, M. and Hollingsworth, Jeffrey K and Saltz, J. and Sussman, Alan} } @inbook {14272, title = {Video representations}, booktitle = {Recent Developments in Computer Vision}, series = {Lecture Notes in Computer Science}, volume = {1035}, year = {1996}, month = {1996///}, pages = {33 - 41}, publisher = {Springer Berlin / Heidelberg}, organization = {Springer Berlin / Heidelberg}, isbn = {978-3-540-60793-9}, url = {http://dx.doi.org/10.1007/3-540-60793-5_60}, author = {Bolle,Ruud and Aloimonos, J. and Ferm{\"u}ller, Cornelia}, editor = {Li,Stan and Mital,Dinesh and Teoh,Eam and Wang,Han} } @conference {16205, title = {Where is information visualization technology going?}, booktitle = {Proceedings of the 9th annual ACM symposium on User interface software and technology}, series = {UIST {\textquoteright}96}, year = {1996}, month = {1996///}, pages = {75 - 77}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-798-7}, doi = {10.1145/237091.237101}, url = {http://doi.acm.org/10.1145/237091.237101}, author = {Hasco{\"e}t-Zizi,Mountaz and Ahlberg,Chris and Korfhage,Robert and Plaisant, Catherine and Chalmers,Matthew and Rao,Ramana} } @conference {12033, title = {3D motion representations in visual servo control}, booktitle = {Proceedings of International Symposium on Computer Vision, 1995}, year = {1995}, month = {1995/11/21/23}, pages = {61 - 66}, publisher = {IEEE}, organization = {IEEE}, abstract = {A new approach to visual servoing and vision-guided robotics is introduced. This approach uses visual information for autonomous behavior. It amounts to using robust, global spatiotemporal representations easily extracted from the dynamic imagery. Specifically, the geometrical patterns of normal flows are used as the input to the servo mechanism. It is shown that the positions of these patterns are related to the three dimensional motion parameters. By locating the positions of these patterns, we can solve a variety of navigational problems with little computational effort}, keywords = {3D motion representations, autonomous behavior, CAMERAS, computational effort, Computer vision, dynamic imagery, global spatiotemporal representations, image representation, Mobile robots, motion control, Navigation, navigational problems, Optical computing, Robot kinematics, Robot sensing systems, robot vision, Robot vision systems, Robotics and automation, servomechanisms, Servosystems, Spatiotemporal phenomena, vision-guided robotics, visual information, visual servo control, visual servoing}, isbn = {0-8186-7190-4}, doi = {10.1109/ISCV.1995.476978}, author = {Ferm{\"u}ller, Cornelia and LoongFah Cheong and Aloimonos, J.} } @conference {15629, title = {Approximate range searching}, booktitle = {Proceedings of the eleventh annual symposium on Computational geometry}, series = {SCG {\textquoteright}95}, year = {1995}, month = {1995///}, pages = {172 - 181}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-724-3}, doi = {10.1145/220279.220298}, url = {http://doi.acm.org/10.1145/220279.220298}, author = {Arya,Sunil and Mount, Dave} } @book {19358, title = {Artificial Intelligence: Theory and Practice}, year = {1995}, month = {1995}, pages = {604}, publisher = {Benjamin/Cummings Pub.}, organization = {Benjamin/Cummings Pub.}, abstract = {This book provides a detailed understanding of the broad issues in artificial intelligence and a useful survey of current AI technology. The author delivers broad coverage of innovative representational techniques, including neural networks, image processing, and probabilistic reasoning, alongside the traditional methods of symbolic reasoning. AI algorithms are described in detailed prose in the text and fully implemented in LISP at the ends of chapters. A stand-alone LISP chapter makes an excellent reference and refresher. Each chapter includes a detailed description of an AI application.}, keywords = {Computers / Expert Systems, Computers / Intelligence (AI) \& Semantics}, isbn = {9780805325478}, author = {Dean, Thomas L. and ALLEN, JAMES AUTOR and Aloimonos, J.} } @article {17918, title = {Automatic generation of multiresolution for polygonal models}, journal = {First Workshop on Simulation and Interaction in Virtual Environments}, year = {1995}, month = {1995///}, author = {Varshney, Amitabh and Agarwal,P. and Brooks,F. and Wright,W. and Weber,H.} } @article {14073, title = {cDNA expressed sequence tags of Trypanosoma brucei rhodesiense provide new insights into the biology of the parasite}, journal = {Molecular and Biochemical Parasitology}, volume = {73}, year = {1995}, month = {1995/07//}, pages = {75 - 90}, abstract = {A total of 518 expressed sequence tags (ESTs) have been generated from clones randomly selected from a cDNA library and a spliced leader sub-library of a Trypanosoma brucei rhodesiense bloodstream clone. 205 (39\%) of the clones were identified based on matches to 113 unique genes in the public databases. Of these, 71 cDNAs display significant similarities to genes in unrelated organisms encoding metabolic enzymes, signal transduction proteins, transcription factors, ribosomal proteins, histones, a proliferation-associated protein and thimet oligopeptidase, among others. 313 of the cDNAs are not related to any other sequences in the databases. These cDNA ESTs provide new avenues of research for exploring both the novel trypanosome-specific genes and the genome organization of this parasite, as well as a resource for identifying trypanosome homologs to genes expressed in other organisms.}, keywords = {cDNA, Expressed sequence tag, Trypanosoma brucei rhodesiense}, isbn = {0166-6851}, doi = {16/0166-6851(95)00098-L}, url = {http://www.sciencedirect.com/science/article/pii/016668519500098L}, author = {El-Sayed, Najib M. and Alarcon,Clara M. and Beck,John C. and Sheffield,Val C. and Donelson,John E.} } @article {14094, title = {Crystallization and preliminary X-ray investigation of the recombinant Trypanosoma brucei rhodesiense calmodulin}, journal = {Proteins: Structure, Function, and Bioinformatics}, volume = {21}, year = {1995}, month = {1995///}, pages = {354 - 357}, author = {El-Sayed, Najib M. and Patton,C. L and Harkins,P. C and Fox,R. O and Anderson,K.} } @conference {17867, title = {Data parallel programming in an adaptive environment}, booktitle = {Parallel Processing Symposium, 1995. Proceedings., 9th International}, year = {1995}, month = {1995///}, pages = {827 - 832}, author = {Edjlali,G. and Agrawal,G. and Sussman, Alan and Saltz, J.} } @article {14212, title = {Direct Perception of Three-Dimensional Motion from Patterns of Visual Motion}, journal = {ScienceScience}, volume = {270}, year = {1995}, month = {1995/12/22/}, pages = {1973 - 1976}, abstract = {Measurements of retinal motion along a set of predetermined orientations on the retina of a moving system give rise to global patterns. Because the form and location of these patterns depend purely on three-dimensional (3D) motion, the effects of 3D motion and scene structure on image motion can be globally separated. The patterns are founded on easily derivable image measurements that depend only on the sign of image motion and do not require information about optical flow. The computational theory presented here explains how the self-motion of a system can be estimated by locating these patterns.}, isbn = {0036-8075, 1095-9203}, doi = {10.1126/science.270.5244.1973}, url = {http://www.sciencemag.org/content/270/5244/1973}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14963, title = {Enhancing LZW Coding Using a Variable-Length Binary Encoding}, volume = {ISR-TR-1995-70}, year = {1995}, month = {1995///}, institution = {Institute for Systems Research, University of Maryland, College Park}, abstract = {We present here a methodology to enhance the LZW coding for text compression using a variable-length binary encoding scheme. The basic principle of this encoding is based on allocating a set of prefix codes to a set of integers growing dynamically. The prefix property enables unique decoding of a string of elements from this set. We presented the experimental results to show the effectiveness of this variable-length binary encoding scheme.}, keywords = {algorithms, data compression, Systems Integration Methodology}, url = {http://drum.lib.umd.edu/handle/1903/5654}, author = {Acharya,Tinku and JaJa, Joseph F.} } @conference {15649, title = {Euclidean spanners: short, thin, and lanky}, booktitle = {Proceedings of the twenty-seventh annual ACM symposium on Theory of computing}, series = {STOC {\textquoteright}95}, year = {1995}, month = {1995///}, pages = {489 - 498}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-718-9}, doi = {10.1145/225058.225191}, url = {http://doi.acm.org/10.1145/225058.225191}, author = {Arya,Sunil and Das,Gautam and Mount, Dave and Salowe,Jeffrey S. and Smid,Michiel} } @article {17977, title = {Generating levels of detail for large-scale polygonal models}, volume = {CS-1995-20}, year = {1995}, month = {1995///}, institution = {Department of Computer Science, Duke University, North Carolina}, abstract = {We present an e cient algorithm for generating various levels-of-detail approximations for agiven polygonal model. Our algorithm guarantees that all points of an approximation are within a user-speci able distance from the original model and all points of the original model are within a distance from the approximation. Each approximation attempts to minimize the total number of polygons required to satisfy the previous constraint. We show how the problem of generating levels- of-detail approximations reduces to the classic set partition problem. The various approximations are guaranteed to be topologically consistent with the input polygonal model. The approximations can be constrained by the user to preserve any desired edges of the input model. We also propose a method to compute an estimate of the quality of the approximation generated by our algorithm with respect to the optimal approximation satisfying the same constraints. We have implemented our algorithm and have obtained experimental results of multiresolution hierarchy generation on over a thousand polygonal objects from a CAD model of a notional submarine. }, author = {Varshney, Amitabh and Agarwal,P. K and Brooks Jr,F. P and Wright,W. V and Weber,H.} } @conference {14168, title = {Global rigidity constraints in image displacement fields}, booktitle = {, Fifth International Conference on Computer Vision, 1995. Proceedings}, year = {1995}, month = {1995/06/20/23}, pages = {245 - 250}, publisher = {IEEE}, organization = {IEEE}, abstract = {Image displacement fields-optical flow fields, stereo disparity fields, normal flow fields-due to rigid motion possess a global geometric structure which is independent of the scene in view. Motion vectors of certain lengths and directions are constrained to lie on the imaging surface at particular loci whose location and form depends solely on the 3D motion parameters. If optical flow fields or stereo disparity fields are considered, then equal vectors are shown to lie on conic sections. Similarly, for normal motion fields, equal vectors lie within regions whose boundaries also constitute conics. By studying various properties of these curves and regions and their relationships, a characterization of the structure of rigid motion fields is given. The goal of this paper is to introduce a concept underlying the global structure of image displacement fields. This concept gives rise to various constraints that could form the basis of algorithms for the recovery of visual information from multiple views}, keywords = {3D motion parameters, algorithms, Automation, Computer science, Computer vision, conic sections, curves, equal vectors, Fluid flow measurement, global geometric structure, global rigidity constraints, image displacement fields, Image motion analysis, Image segmentation, Image sequences, imaging surface, Laboratories, Layout, Motion estimation, Motion measurement, motion vectors, multiple views, normal flow fields, optical flow fields, regions, rigid motion, stereo disparity fields, Stereo vision, vectors, visual information recovery}, isbn = {0-8186-7042-8}, doi = {10.1109/ICCV.1995.466779}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {11982, title = {Global rigidity constraints in image displacement fields}, booktitle = {Proceedings of Fifth International Conference on Computer Vision, 1995}, year = {1995}, month = {1995/06/20/23}, pages = {245 - 250}, publisher = {IEEE}, organization = {IEEE}, abstract = {Image displacement fields-optical flow fields, stereo disparity fields, normal flow fields-due to rigid motion possess a global geometric structure which is independent of the scene in view. Motion vectors of certain lengths and directions are constrained to lie on the imaging surface at particular loci whose location and form depends solely on the 3D motion parameters. If optical flow fields or stereo disparity fields are considered, then equal vectors are shown to lie on conic sections. Similarly, for normal motion fields, equal vectors lie within regions whose boundaries also constitute conics. By studying various properties of these curves and regions and their relationships, a characterization of the structure of rigid motion fields is given. The goal of this paper is to introduce a concept underlying the global structure of image displacement fields. This concept gives rise to various constraints that could form the basis of algorithms for the recovery of visual information from multiple views}, keywords = {3D motion parameters, algorithms, Automation, Computer science, Computer vision, conic sections, curves, equal vectors, Fluid flow measurement, global geometric structure, global rigidity constraints, image displacement fields, Image motion analysis, Image segmentation, Image sequences, imaging surface, Laboratories, Layout, Motion estimation, Motion measurement, motion vectors, multiple views, normal flow fields, optical flow fields, regions, rigid motion, stereo disparity fields, Stereo vision, vectors, visual information recovery}, isbn = {0-8186-7042-8}, doi = {10.1109/ICCV.1995.466779}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @conference {14171, title = {The information in the direction of image flow}, booktitle = {, International Symposium on Computer Vision, 1995. Proceedings}, year = {1995}, month = {1995/11/21/23}, pages = {461 - 466}, publisher = {IEEE}, organization = {IEEE}, abstract = {If instead of the full motion field, we consider only the direction of the motion field due to a rigid motion, what can we say about the information regarding the three-dimensional motion? In this paper it is shown that considering as the imaging surface the whole sphere, independently of the scene in view, two different rigid motions cannot give rise to the same directional motion field. If we restrict the image to half of a sphere (or an infinitely large image plane) two different rigid motions with instantaneous translational and rotational velocities (t1, ω1) and (t2, ω2) cannot give rise to the same directional motion field unless the plane through t1 and t2 is perpendicular to the plane through ω1 and ω2 (i.e., (t1{\texttimes}t2){\textperiodcentered}(ω1 {\texttimes}ω2)=0). In addition, in order to give a practical significance to these uniqueness results for the case of a limited field of view we also characterize the locations on the image where the motion vectors due to the different motions must have different directions. If (ω1{\texttimes}ω2){\textperiodcentered}(t1 {\texttimes}t2)=0 and certain additional constraints are met, then the two rigid motions could produce motion fields with the same direction. For this to happen the depth of each corresponding surface has to be within a certain range, defined by a second and a third order surface}, keywords = {Automation, CAMERAS, Computer vision, Educational institutions, image flow, Image motion analysis, Image sequences, imaging surface, Laboratories, Layout, Motion analysis, Motion estimation, motion field, motion vectors, Optical imaging, rigid motion, rigid motions, three-dimensional motion}, isbn = {0-8186-7190-4}, doi = {10.1109/ISCV.1995.477071}, author = {Brodsky, T. and Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {17839, title = {An integrated runtime and compile-time approach for parallelizing structured and block structured applications}, journal = {IEEE Transactions on Parallel and Distributed Systems}, volume = {6}, year = {1995}, month = {1995/07//}, pages = {747 - 754}, abstract = {In compiling applications for distributed memory machines, runtime analysis is required when data to be communicated cannot be determined at compile-time. One such class of applications requiring runtime analysis is block structured codes. These codes employ multiple structured meshes, which may be nested (for multigrid codes) and/or irregularly coupled (called multiblock or irregularly coupled regular mesh problems). In this paper, we present runtime and compile-time analysis for compiling such applications on distributed memory parallel machines in an efficient and machine-independent fashion. We have designed and implemented a runtime library which supports the runtime analysis required. The library is currently implemented on several different systems. We have also developed compiler analysis for determining data access patterns at compile time and inserting calls to the appropriate runtime routines. Our methods can be used by compilers for HPF-like parallel programming languages in compiling codes in which data distribution, loop bounds and/or strides are unknown at compile-time. To demonstrate the efficacy of our approach, we have implemented our compiler analysis in the Fortran 90D/HPF compiler developed at Syracuse University. We have experimented with a multi-bloc Navier-Stokes solver template and a multigrid code. Our experimental results show that our primitives have low runtime communication overheads and the compiler parallelized codes perform within 20\% of the codes parallelized by manually inserting calls to the runtime library}, keywords = {Bandwidth, block structured applications, block structured codes, compile-time approach, compiling applications, data access patterns, Data analysis, Delay, distributed memory machines, distributed memory systems, FORTRAN, Fortran 90D/HPF compiler, High performance computing, HPF-like parallel programming languages, integrated runtime approach, irregularly coupled regular mesh problems, multigrid code, Navier-Stokes solver template, Parallel machines, parallel programming, Pattern analysis, performance evaluation, program compilers, Program processors, Runtime library, Uninterruptible power systems}, isbn = {1045-9219}, doi = {10.1109/71.395403}, author = {Agrawal,G. and Sussman, Alan and Saltz, J.} } @conference {12031, title = {Iso-distortion contours and egomotion estimation}, booktitle = {Proceedings of International Symposium on Computer Vision, 1995}, year = {1995}, month = {1995/11/21/23}, pages = {55 - 60}, publisher = {IEEE}, organization = {IEEE}, abstract = {This paper introduces the framework of iso-distortion contour to deal with the problem of depth distortion due to erroneous motion estimates, and various related aspects such as the effectiveness of the visibility constraint. The framework can also be used to inquire the uniqueness aspect of normal flow. Future work will examine the implications of the iso-distortion contours for the problem of multiple frame integration}, keywords = {Automation, Computer vision, Degradation, depth distortion, Educational institutions, egomotion estimation, Equations, erroneous motion estimates, Error analysis, HUMANS, Image sequences, iso-distortion contours, Laboratories, Layout, Motion estimation, Robustness, visibility constraint}, isbn = {0-8186-7190-4}, doi = {10.1109/ISCV.1995.476977}, author = {LoongFah Cheong and Aloimonos, J.} } @conference {15599, title = {M.H.M: Euclidean spanners: short, thin, and lanky}, booktitle = {In: 27th ACM Symposium on Theory of Computing}, year = {1995}, month = {1995///}, pages = {489 - 498}, publisher = {ACM press}, organization = {ACM press}, abstract = {Euclidean spanners are important data structures in geometric algorithm design, because they provide a means of approximating the complete Euclidean graph with only O(n) edges, so that the shortest path length between each pair of points is not more than a constant factor longer than the Euclidean distance between the points. In many applications of spanners, it is important that the spanner possess a number of additional properties: low tot al edge weight, bounded degree, and low diameter. Existing research on spanners has considered one property or the other. We show that it is possible to build spanners in optimal O (n log n) time and O(n) space that achieve optimal or near optimal tradeoffs between all combinations of these *Max-Planck-Institut fiir Informatik, D-66123 Saarbrucken,}, author = {Arya,Sunil and Dast,Gautam and Mount, Dave and Salowe,Jeffrey S. and Smid,Michiel} } @article {16210, title = {Next generation network management technology}, journal = {AIP Conference Proceedings}, volume = {325}, year = {1995}, month = {1995/01/25/}, pages = {75 - 82}, abstract = {Today{\textquoteright}s telecommunications networks are becoming increasingly large, complex, mission critical and heterogeneous in several dimensions. For example, the underlying physical transmission facilities of a given network may be {\textquoteleft}{\textquoteleft}mixed media{\textquoteright}{\textquoteright} (copper, fiber-optic, radio, and satellite); the subnetworks may be acquired from different vendors due to economic, performance, or general availability reasons; the information being transmitted over the network may be {\textquoteleft}{\textquoteleft}multimedia{\textquoteright}{\textquoteright} (video, data, voice, and images) and, finally, varying performance criteria may be imposed e.g., data transfer may require high throughput while the others, whose concern is voice communications, may require low call blocking probability. For these reasons, future telecommunications networks are expected to be highly complex in their services and operations. Due to this growing complexity and the disparity among management systems for individual sub-networks, efficient network management systems have become critical to the current and future success of telecommunications companies. This paper addresses a research and development effort which focuses on prototyping configuration management, since that is the central process of network management and all other network management functions must be built upon it. Our prototype incorporates ergonomically designed graphical user interfaces tailored to the network configuration management subsystem and to the proposed advanced object-oriented database structure. The resulting design concept follows open standards such as Open Systems Interconnection (OSI) and incorporates object oriented programming methodology to associate data with functions, permit customization, and provide an open architecture environment. {\textcopyright} 1995 American Institute of Physics}, isbn = {0094243X}, doi = {doi:10.1063/1.47255}, url = {http://proceedings.aip.org/resource/2/apcpcs/325/1/75_1?isAuthorized=no}, author = {Baras,John S and Atallah,George C and Ball,Mike and Goli,Shravan and Karne,Ramesh K and Kelley,Steve and Kumar,Harsha and Plaisant, Catherine and Roussopoulos, Nick and Schneiderman,Ben and Srinivasarao,Mulugu and Stathatos,Kosta and Teittinen,Marko and Whitefield,David} } @article {14250, title = {Qualitative egomotion}, journal = {International Journal of Computer Vision}, volume = {15}, year = {1995}, month = {1995///}, pages = {7 - 29}, abstract = {Due to the aperture problem, the only motion measurement in images, whose computation does not require any assumptions about the scene in view, is normal flow{\textemdash}the projection of image motion on the gradient direction. In this paper we show how a monocular observer can estimate its 3D motion relative to the scene by using normal flow measurements in a global and qualitative way. The problem is addressed through a search technique. By checking constraints imposed by 3D motion parameters on the normal flow field, the possible space of solutions is gradually reduced. In the four modules that comprise the solution, constraints of increasing restriction are considered, culminating in testing every single normal flow value for its consistency with a set of motion parameters. The fact that motion is rigid defines geometric relations between certain values of the normal flow field. The selected values form patterns in the image plane that are dependent on only some of the motion parameters. These patterns, which are determined by the signs of the normal flow values, are searched for in order to find the axes of translation and rotation. The third rotational component is computed from normal flow vectors that are only due to rotational motion. Finally, by looking at the complete data set, all solutions that cannot give rise to the given normal flow field are discarded from the solution space.}, isbn = {0920-5691}, url = {http://dx.doi.org/10.1007/BF01450848}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {11972, title = {Qualitative vision}, journal = {International Journal of Computer Vision}, volume = {14}, year = {1995}, month = {1995///}, pages = {115 - 117}, author = {Aloimonos, J.} } @conference {14180, title = {Representations for active vision}, booktitle = {Proceedings of the 14th international joint conference on Artificial intelligence - Volume 1}, year = {1995}, month = {1995///}, pages = {20 - 26}, publisher = {Morgan Kaufmann Publishers Inc.}, organization = {Morgan Kaufmann Publishers Inc.}, address = {San Francisco, CA, USA}, isbn = {1-55860-363-8, 978-1-558-60363-9}, url = {http://dl.acm.org/citation.cfm?id=1625855.1625858}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14255, title = {Seeing and understanding: representing the visual world}, journal = {ACM Comput. Surv.}, volume = {27}, year = {1995}, month = {1995/09//}, pages = {307 - 309}, isbn = {0360-0300}, doi = {10.1145/212094.212101}, url = {http://doi.acm.org/10.1145/212094.212101}, author = {Aloimonos, J. and Ferm{\"u}ller, Cornelia and Rosenfeld, A.} } @article {18435, title = {Striving for correctness}, journal = {Computers \& Security}, volume = {14}, year = {1995}, month = {1995///}, pages = {719 - 738}, abstract = {In developing information technology, you want assurance that systems are secure and reliable, but you cannot have assurance or security without correctness. We discuss methods used to achieve correctness, focusing on weaknesses and approaches that management might take to increase belief in correctness. Formal methods, simulation, testing, and process modeling are addressed in detail. Structured programming, life-cycle modeling like the spiral model, use of CASE tools, use of formal methods, object-oriented design, reuse of existing code are also mentioned. Reliance on these methods involves some element of belief since no validated metrics on the effectiveness of these methods exist. Suggestions for using these methods as the basis for managerial decisions conclude the paper.}, keywords = {Assurance, belief, correctness, Formal Methods, MATHEMATICAL MODELS, metrics, Process models, Risk management, Security testing, Silver bullets, simulation, Trustworthiness}, isbn = {0167-4048}, doi = {10.1016/0167-4048(95)00022-4}, url = {http://www.sciencedirect.com/science/article/pii/0167404895000224}, author = {Abrams,Marshall D. and Zelkowitz, Marvin V} } @article {17481, title = {Using Treemaps to Visualize the Analytic Hierarchy Process}, journal = {Information Systems ResearchInformation Systems Research}, volume = {6}, year = {1995}, month = {1995/12/01/}, pages = {357 - 375}, abstract = {Treemaps, a visualization method for large hierarchical data spaces, are used to augment the capabilities of the Analytic Hierarchy Process (AHP) for decision-making. Two direct manipulation tools, presented metaphorically as a {\textquotedblleft}pump{\textquotedblright} and a {\textquotedblleft}hook,{\textquotedblright} were developed and applied to the treemap to support AHP sensitivity analysis. Users can change the importance of criteria dynamically on the two-dimensional treemap and immediately see the impact on the outcome of the decision. This fluid process dramatically speeds up exploration and provides a better understanding of the relative impact of the component criteria. A usability study with six subjects using a prototype AHP application showed that treemap representation was acceptable from a visualization and data operation standpoint.}, keywords = {AHP, analytic hierarchy process, decision support, treemap, User interfaces, Visualization}, isbn = {1047-7047, 1526-5536}, doi = {10.1287/isre.6.4.357}, url = {http://isr.journal.informs.org/content/6/4/357}, author = {Asahi,Toshiyuki and Turo,David and Shneiderman, Ben} } @article {14177, title = {Vision and action}, journal = {Image and Vision Computing}, volume = {13}, year = {1995}, month = {1995/12//}, pages = {725 - 744}, abstract = {Our work on active vision has recently focused on the computational modelling of navigational tasks, where our investigations were guided by the idea of approaching vision for behavioural systems in the form of modules that are directly related to perceptual tasks. These studies led us to branch in various directions and inquire into the problems that have to be addressed in order to obtain an overall understanding of perceptual systems. In this paper, we present our views about the architecture of vision systems, about how to tackle the design and analysis of perceptual systems, and promising future research directions. Our suggested approach for understanding behavioural vision to realize the relationships of perception and action builds on two earlier approaches, the Medusa philosophy1 and the Synthetic approach2. The resulting framework calls for synthesizing an artificial vision system by studying vision competences of increasing complexity and, at the same time, pursuing the integration of the perceptual components with action and learning modules. We expect that computer vision research in the future will progress in tight collaboration with many other disciplines that are concerned with empirical approaches to vision, i.e. the understanding of biological vision. Throughout the paper, we describe biological findings that motivate computational arguments which we believe will influence studies of computer vision in the near future.}, keywords = {active vision, perceptual system design and analysis, vision system architecture}, isbn = {0262-8856}, doi = {10.1016/0262-8856(95)98754-H}, url = {http://www.sciencedirect.com/science/article/pii/026288569598754H}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {12030, title = {Vision and action* 1}, journal = {Image and vision computing}, volume = {13}, year = {1995}, month = {1995///}, pages = {725 - 744}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {17483, title = {Vision, Graphic Design, and Visual Display}, journal = {Readings in human-computer interaction: toward the year 2000}, year = {1995}, month = {1995///}, pages = {411 - 411}, author = {Marcus,A. and Murch,G.M. and Baecker,R. and Small,I. and Mander,R. and Ahlberg,C. and Shneiderman, Ben} } @conference {17486, title = {Visual decision-making: using treemaps for the analytic hierarchy process}, booktitle = {Conference companion on Human factors in computing systems}, series = {CHI {\textquoteright}95}, year = {1995}, month = {1995///}, pages = {405 - 406}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-755-3}, doi = {10.1145/223355.223747}, url = {http://doi.acm.org/10.1145/223355.223747}, author = {Asahi,Toshiyuki and Turo,David and Shneiderman, Ben} } @article {17512, title = {Windows of opportunity in electronic classrooms}, journal = {Communications of the ACM}, volume = {38}, year = {1995}, month = {1995/11//}, pages = {19 - 24}, abstract = {Paradigm-shifting landmark buildings are cherished by their occupants and remembered because they reshape our expectations of schools, homes, or offices. Classic examples include Thomas Jefferson{\textquoteright}s communal design of the {\textquotedblleft}academical village{\textquotedblright} at the University of Virginia where faculty and students lived close to classrooms, Frank Lloyd Wright{\textquoteright}s organic harmony with nature in Fallingwater (in western Pennsylvania) where the waterfall sounds and leafy surroundings offered a stress-reducing getaway for an urban executive, or Kevin Roche{\textquoteright}s open glass-walled Ford Foundation (in New York City) that promoted new team-oriented management strategies.}, isbn = {0001-0782}, doi = {10.1145/219717.219725}, url = {http://doi.acm.org/10.1145/219717.219725}, author = {Shneiderman, Ben and Alavi,Maryann and Norman,Kent and Borkowski,Ellen Yu} } @conference {17404, title = {The alphaslider: a compact and rapid selector}, booktitle = {Proceedings of the SIGCHI conference on Human factors in computing systems: celebrating interdependence}, series = {CHI {\textquoteright}94}, year = {1994}, month = {1994///}, pages = {365 - 371}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, keywords = {Alphaslider, dynamic queries, menus, selection technology, widget}, isbn = {0-89791-650-6}, doi = {10.1145/191666.191790}, url = {http://doi.acm.org/10.1145/191666.191790}, author = {Ahlberg,Christopher and Shneiderman, Ben} } @article {16621, title = {Cortical map reorganization as a competitive process}, journal = {Neural Computation}, volume = {6}, year = {1994}, month = {1994///}, pages = {1 - 13}, author = {Sutton III,G. G. and Reggia, James A. and Armentrout,S. L and D{\textquoteright}Autrechy,C. L} } @article {11923, title = {Design and implementation of Maruti-II}, journal = {Principles of Real-Time Systems. Prentice Hall}, year = {1994}, month = {1994///}, author = {Saksena,M. and da Silva,J. and Agrawala, Ashok K.} } @article {15569, title = {Dynamic algorithms for geometric spanners of small diameter: Randomized solutions}, journal = {Computational Geometry: Theory and Applications}, volume = {13}, year = {1994}, month = {1994///}, pages = {13 - 91}, abstract = {Let S be a set of n points in IR d and let t ? 1 be a real number. A t-spanner for S is a directed graph having the points of S as its vertices, such that for any pair p and q of points there is a path from p to q of length at most t times the Euclidean distance between p and q. Such a path is called a t-spanner path. The spanner diameter of such a spanner is defined as the smallest integer D such that for any pair p and q of points there is a t-spanner path from p to q containing at most D edges. A randomized algorithm is given for constructing a t-spanner that, with high probability, contains O(n) edges and has spanner diameter O(log n). A data structure of size O(n log d n) is given that maintains this t-spanner in O(log d n log log n) expected amortized time per insertion and deletion, in the model of random updates, as introduced by Mulmuley. Keywords: Computational geometry, proximity problems, skip lists, randomization, dynamic data structures. Preprint submitted to Els...}, author = {Arya,Sunil and Mount, Dave and Smid,Michiel} } @conference {17896, title = {Efficient runtime support for parallelizing block structured applications}, booktitle = {Scalable High-Performance Computing Conference, 1994., Proceedings of the}, year = {1994}, month = {1994/05//}, pages = {158 - 167}, abstract = {Scientific and engineering applications often involve structured meshes. These meshes may be nested (for multigrid codes) and/or irregularly coupled (called multiblock or irregularly coupled regular mesh problems). We describe a runtime library for parallelizing these applications on distributed memory parallel machines in an efficient and machine-independent fashion. This runtime library is implemented on several different systems. This library can be used by application programmers to port applications by hand and can also be used by a compiler to handle communication for these applications. Our experimental results show that our primitives have low runtime communication overheads. We have used this library to port a multiblock template and a multigrid code. Effort is also underway to port a complete multiblock computational fluid dynamics code using our library}, keywords = {application programmers, block structured applications, distributed memory parallel machines, distributed memory systems, engineering applications, irregularly coupled regular mesh problems, machine-independent, multiblock, multiblock computational fluid dynamics code, multiblock template, multigrid codes, Parallel machines, parallel programming, Physics computing, runtime communication overhead, Runtime library, runtime support, scientific applications, software reusability, structured meshes}, doi = {10.1109/SHPCC.1994.296639}, author = {Agrawal,G. and Sussman, Alan and Saltz, J.} } @article {14907, title = {Error propagation in 3D-from-2D recognition: Scaled-orthographic and perspective projections}, journal = {Proceedings: ARPA Image Understanding Workshop, Monterey, California}, year = {1994}, month = {1994///}, abstract = {Robust recognition systems require a careful under-standing of the e ects of error in sensed features. Error in these image features results in uncertainty in the possible image location of each additional model feature. We present an accurate, analytic ap- proximation for this uncertainty when model poses are based on matching three image and model points. This result applies to objects that are fully three- dimensional, where past results considered only two- dimensional objects. Further, we introduce a lin- ear programming algorithm to compute this uncer- tainty when poses are based on any number of initial matches. }, author = {Alter,T. D. and Jacobs, David W.} } @conference {14839, title = {Error propagation in full 3D-from-2D object recognition}, booktitle = {Computer Vision and Pattern Recognition, 1994. Proceedings CVPR {\textquoteright}94., 1994 IEEE Computer Society Conference on}, year = {1994}, month = {1994/06//}, pages = {892 - 898}, abstract = {Robust recognition systems require a careful understanding of the effects of error in sensed features. Error in these image features results in uncertainty in the possible image location of each additional model feature. We present an accurate, analytic approximation for this uncertainty when model poses are based on matching three image and model points. This result applies to objects that are fully three-dimensional, where past results considered only two-dimensional objects. Further, we introduce a linear programming algorithm to compute this uncertainty when poses are based on any number of initial matches}, keywords = {3D-from-2D, algorithm;, error, extraction;, feature, features;, handling;, image, initial, linear, matches;, object, Programming, programming;, propagation;, recognition, recognition;, robust, sequences;, systems;, Uncertainty, uncertainty;}, doi = {10.1109/CVPR.1994.323920}, author = {Alter,T. D. and Jacobs, David W.} } @article {12061, title = {Estimating the heading direction using normal flow}, journal = {International Journal of Computer Vision}, volume = {13}, year = {1994}, month = {1994///}, pages = {33 - 56}, author = {Aloimonos, J. and Duric, Z.} } @book {19353, title = {Fault tolerant system design}, year = {1994}, month = {1994}, pages = {440}, publisher = {McGraw-Hill}, organization = {McGraw-Hill}, abstract = {This book presents a comprehensive exploration of the practical issues, tested techniques, and accepted theory for developing fault tolerant systems. It is a ready reference to work already done in the field, with new approaches devised by the authors.}, keywords = {Computers / Intelligence (AI) \& Semantics, Computers / Programming Languages / General, Computers / Software Development \& Engineering / Systems Analysis \& Design, Reference / General, Technology \& Engineering / Engineering (General)}, isbn = {9780070375154}, author = {Levi,Shem-Tov and Agrawala, Ashok K.} } @article {12019, title = {How normal flow constrains relative depth for an active observer}, journal = {Image and vision computing}, volume = {12}, year = {1994}, month = {1994///}, pages = {435 - 445}, author = {Huang, L. and Aloimonos, J.} } @conference {11924, title = {Integrated approaches for improving the effectiveness of Plan Reuse (A Progress Report)}, booktitle = {ARPA/Rome Laboratory knowledge-based planning and scheduling initiative: workshop proceedings: Tuscon, Arizona, February 21-24, 1994}, year = {1994}, month = {1994///}, pages = {325 - 325}, author = {Kambhampati,S. and Ihrig,L. and Katukam,S. and Chen,J. and Hendler,J. A and Agrawala, Ashok K.} } @article {11925, title = {Mission-oriented replication of periodic tasks in real-time distributed systems}, journal = {Computer Science Technical Report Series; Vol. CS-TR-3256}, year = {1994}, month = {1994///}, pages = {24 - 24}, author = {Cheng,S. T and Hwang,S. I and Agrawala, Ashok K.} } @article {16508, title = {A neural model of cortical map reorganization following a focal lesion}, journal = {Artificial Intelligence in Medicine}, volume = {6}, year = {1994}, month = {1994///}, pages = {383 - 400}, author = {Armentrout,S. L and Reggia, James A. and Weinrich,M.} } @article {16217, title = {Next Generation Network Management Technology}, year = {1994}, month = {1994///}, abstract = {Today{\textquoteright}s telecommunications networks are becoming increasingly large, complex, mission critical and heterogeneous in several dimensions. For example, the underlying physical transmission facilities of a given network may be ﲭixed media (copper, fiber- optic, radio, and satellite); the sub networks may be acquired from different vendors due to economic, performance, or general availability reasons; the information being transmitted over the network may be ﲭultimedia (video, data, voice, and images) and, finally, varying performance criteria may be imposed e.g. data transfer may require high throughput while the others, whose concern is voice communications, may require low call blocking probability. For these reasons, future telecommunications networks are expected to be highly complex in their services and operations. Due to this growing complexity and the disparity among management systems for individual sub networks, efficient network management systems have become critical to the current and future success of telecommunications companies. This paper addresses a research and development effort which focuses on prototyping configuration management, since that is the central process of network management and all other network management functions must be built upon it. Our prototype incorporates ergonomically designed graphical user interfaces tailored to the network configuration management subsystem and to the proposed advanced object-oriented database structure. The resulting design concept follows open standards such as Open Systems Interconnection (OSI) and incorporates object oriented programming methodology to associate data with functions, permit customization, and provide an open architecture environment. }, keywords = {Constraints for Network Management., Network Configuration Management, network management, Object Oriented Data Base Model for Network Management, Rules, Systems Integration, Visual Information Management for Network Configuration Management}, url = {http://drum.lib.umd.edu/handle/1903/5519}, author = {Atallah,George C and Ball,Michael O and Baras,John S and Goli,Shravan K and Karne,Ramesh K and Kelley,Stephen and Kumar,Harsha P. and Plaisant, Catherine and Roussopoulos, Nick and Shneiderman, Ben and Srinivasarao,Mulugu and Stathatos,Kostas and Teittinen,Marko and Whitefield,David} } @conference {11947, title = {Principles of computer vision}, booktitle = {Handbook of pattern recognition and image processing (vol. 2)}, year = {1994}, month = {1994///}, pages = {1 - 15}, author = {Aloimonos, J. and Rosenfeld, A.} } @conference {15588, title = {Randomized and deterministic algorithms for geometric spanners of small diameter}, booktitle = {Foundations of Computer Science, 1994 Proceedings., 35th Annual Symposium on}, year = {1994}, month = {1994/11//}, pages = {703 - 712}, abstract = {Let S be a set of n points in IRd and let t gt;1 be a real number. A t-spanner for S is a directed graph having the points of S as its vertices, such that for any pair p and q of points there is a path from p to q of length at most t times the Euclidean distance between p and p. Such a path is called a t-spanner path. The spanner diameter of such a spanner is defined as the smallest integer D such that for any pair p and q of points there is a t-spanner path from p to q containing at most D edges. Randomized and deterministic algorithms are given for constructing t-spanners consisting of O(n) edges and having O(log n) diameter. Also, it is shown how to maintain the randomized t-spanner under random insertions and deletions. Previously, no results were known for spanners with low spanner diameter and for maintaining spanners under insertions and deletions}, keywords = {computational geometry, deletions, deterministic algorithms, directed graph, directed graphs, geometric spanners, insertions, randomised algorithms, randomized algorithms}, doi = {10.1109/SFCS.1994.365722}, author = {Arya,S. and Mount, Dave and Smid,M.} } @article {14914, title = {Uncertainty Propagation in Model-Based Recognition.}, year = {1994}, month = {1994/12//}, institution = {MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LAB}, abstract = {Building robust recognition systems requires a careful understanding of the effects of error in sensed features. Error in these image features results in a region of uncertainty in the possible image location of each additional model feature. We present an accurate, analytic approximation for this uncertainty region when model poses are based on matching three image and model points, for both Gaussian and bounded error in the detection of image points, and for both scaled-orthographic and perspective projection models. This result applies to objects that are fully three-dimensional, where past results considered only two-dimensional objects. Further, we introduce a linear programming algorithm to compute the uncertainty region when poses are based on any number of initial matches. Finally, we use these results to extend, from two-dimensional to three-dimensional objects, robust implementations of alignment interpretation-tree search, and transformation clustering. (AN)}, keywords = {*IMAGE PROCESSING, *PATTERN RECOGNITION, algorithms, APPROXIMATION(MATHEMATICS), CYBERNETICS, ERROR CORRECTION CODES, image registration, Linear programming, MATCHING, MATHEMATICAL MODELS, PIXELS, PROJECTIVE TECHNIQUES., regions, THREE DIMENSIONAL, TWO DIMENSIONAL, Uncertainty}, url = {http://stinet.dtic.mil/oai/oai?\&verb=getRecord\&metadataPrefix=html\&identifier=ADA295642}, author = {Jacobs, David W. and Alter,T. D.} } @conference {17489, title = {Visual information seeking: tight coupling of dynamic query filters with starfield displays}, booktitle = {Proceedings of the SIGCHI conference on Human factors in computing systems: celebrating interdependence}, series = {CHI {\textquoteright}94}, year = {1994}, month = {1994///}, pages = {313 - 317}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, keywords = {database query, dynamic queries, information seeking, starfield displays, tight coupling}, isbn = {0-89791-650-6}, doi = {10.1145/191666.191775}, url = {http://doi.acm.org/10.1145/191666.191775}, author = {Ahlberg,Christopher and Shneiderman, Ben} } @article {17488, title = {Visual information seeking using the FilmFinder}, journal = {ACM CHI94 Conference, Boston}, year = {1994}, month = {1994///}, author = {Ahlberg,C. and Shneiderman, Ben} } @article {11987, title = {What I have learned}, journal = {CVGIP: Image Understanding}, volume = {60}, year = {1994}, month = {1994///}, pages = {74 - 85}, author = {Aloimonos, J.} } @book {12028, title = {Active perception.}, year = {1993}, month = {1993///}, publisher = {Lawrence Erlbaum Associates, Inc}, organization = {Lawrence Erlbaum Associates, Inc}, author = {Aloimonos, J.} } @book {11948, title = {Active Perception, Vol. I of Advances in Computer Vision series}, year = {1993}, month = {1993///}, publisher = {Lawrence Erlbaum Associates}, organization = {Lawrence Erlbaum Associates}, author = {Aloimonos, J.} } @article {11963, title = {Active vision revisited}, journal = {Active Perception}, year = {1993}, month = {1993///}, author = {Aloimonos, J.} } @conference {15577, title = {Algorithms for fast vector quantization}, booktitle = {Data Compression Conference, 1993. DCC {\textquoteright}93.}, year = {1993}, month = {1993///}, pages = {381 - 390}, abstract = {This paper shows that if one is willing to relax the requirement of finding the true nearest neighbor, it is possible to achieve significant improvements in running time and at only a very small loss in the performance of the vector quantizer. The authors present three algorithms for nearest neighbor searching: standard and priority k -d tree search algorithms and a neighborhood graph search algorithm in which a directed graph is constructed for the point set and edges join neighboring points}, keywords = {algorithm;, algorithms;, directed, fast, graph, graph;, graphs;, Neighborhood, performance;, problems;, quantisation;, quantization;, running, search, time;, tree, vector}, doi = {10.1109/DCC.1993.253111}, author = {Arya,S. and Mount, Dave} } @conference {15636, title = {Approximate nearest neighbor queries in fixed dimensions}, booktitle = {Proceedings of the fourth annual ACM-SIAM Symposium on Discrete algorithms}, series = {SODA {\textquoteright}93}, year = {1993}, month = {1993///}, pages = {271 - 280}, publisher = {Society for Industrial and Applied Mathematics}, organization = {Society for Industrial and Applied Mathematics}, address = {Philadelphia, PA, USA}, isbn = {0-89871-313-7}, url = {http://dl.acm.org/citation.cfm?id=313559.313768}, author = {Arya,Sunil and Mount, Dave} } @article {15617, title = {Approximate nearest neighbor searching}, journal = {Proceedings of 4th Annual ACMSIAM Symposium on Discrete Algorithms (SODA{\textquoteright}93)}, year = {1993}, month = {1993///}, pages = {271 - 280}, author = {Arya,S. and Mount, Dave} } @conference {15659, title = {Fast search algorithms with applications to split and multi-stage vector quantization of speech lsp parameters}, booktitle = {Speech Coding for Telecommunications, 1993. Proceedings., IEEE Workshop on}, year = {1993}, month = {1993///}, pages = {65 - 66}, doi = {10.1109/SCFT.1993.762341}, author = {Arya,S. and Phamdo,N. and Farvardin,N. and Mount, Dave} } @article {15208, title = {Geometric knapsack problems}, journal = {Algorithmica}, volume = {10}, year = {1993}, month = {1993///}, pages = {399 - 427}, abstract = {We study a variety of geometric versions of the classical knapsack problem. In particular, we consider the following ldquofence enclosurerdquo problem: given a setS ofn points in the plane with valuesv i ge 0, we wish to enclose a subset of the points with a fence (a simple closed curve) in order to maximize the ldquovaluerdquo of the enclosure. The value of the enclosure is defined to be the sum of the values of the enclosed points minus the cost of the fence. We consider various versions of the problem, such as allowingS to consist of points and/or simple polygons. Other versions of the problems are obtained by restricting the total amount of fence available and also allowing the enclosure to consist of at mostM connected components. When there is an upper bound on the length of fence available, we show that the problem is NP-complete. We also provide polynomial-time algorithms for many versions of the fence problem when an unrestricted amount of fence is available.}, doi = {10.1007/BF01769706}, author = {Arkin,E. M and Khuller, Samir and Mitchell,J. S.B} } @conference {11926, title = {Multiple Resource Allocation for Multiprocessor Distributed Real-Time Systems}, booktitle = {In Workshop on Parallel and Distributed Real-Time Systems (PDRTS), IEEE IPPS{\textquoteright}93}, year = {1993}, month = {1993///}, abstract = {this report are those of the author(s) and should not be interpreted as representing the official policies, either expressed or implied, of the Defense Advanced Research Projects Agency, ONR, the U.S. Government or Honeywell. Computer facilities were provided in part by NSF grant CCR-8811954. Workshop on Parallel and Dist Real-Time Syst, IEEE IPPS{\textquoteright}93 April 1993 in the queue and tries to schedule it. If the task is successfully scheduled, it is removed from the queue and executed. Otherwise, the task is unlocked so that other processors may examine it in the same fashion. This scheme is convenient for a shared memory architecture, but becomes too expensive and cumbersome when such facility is not available in hardware (for example, in distributed systems). The goal of this paper is twofold: first, to study the problem of allocating multiple resources to a single task; and second, to provide a heuristic allocation policy for selecting a resource among multiple instances of identical resources. We consider a distributed system in which each site is a uniform memory access (UMA) multiprocessor, but the scheme is easily extensible to non-UMA models. For the first problem, a three-tiered resource allocation scheme is proposed. A hierarchy of the resources is defined to reduce the probability of allocation failure as the allocation progresses. To reduce the complexity of multiple resource allocation, we group the resources into schedulable sets of resources. Since the objective in real-time systems is to meet deadlines, this approach is acceptable so long as it meets a justifiable minimum level of resource utilization. It is shown that our approach allows modularity, flexibility, and potential concurrency in scheduling. Another contribution of this paper is in explicitly ide...}, author = {Mosse,Daniel and Noh,Sam H and Trinh,Bao and Agrawala, Ashok K.} } @article {15554, title = {Point probe decision trees for geometric concept classes}, journal = {Algorithms and Data Structures}, year = {1993}, month = {1993///}, pages = {95 - 106}, abstract = {A fundamental problem in model-based computer vision is that of identifying to which of a given set of concept classes of geometric models an observed model belongs. Considering a ldquoproberdquo to be an oracle that tells whether or not the observed model is present at a given point in an image, we study the problem of computing efficient strategies (ldquodecision treesrdquo) for probing an image, with the goal to minimize the number of probes necessary (in the worst case) to determine in which class the observed model belongs. We prove a hardness result and give strategies that obtain decision trees whose height is within a log factor of optimal.These results grew out of discussions that began in a series of workshops on Geometric Probing in Computer Vision, sponsored by the Center for Night Vision and Electro-Optics, Fort Belvoir, Virginia, and monitored by the U.S. Army Research Office. The views, opinions, and/or findings contained in this report are those of the authors and should not be construed as an official Department of the Army position, policy, or decision, unless so designated by other documentation. }, doi = {10.1007/3-540-57155-8_239}, author = {Arkin,E. and Goodrich,M. and Mitchell,J. and Mount, Dave and Piatko,C. and Skiena,S.} } @article {12035, title = {Probabilistic analysis of some navigation strategies in a dynamic environment}, journal = {Systems, Man and Cybernetics, IEEE Transactions on}, volume = {23}, year = {1993}, month = {1993///}, pages = {1465 - 1474}, author = {Sharma, R. and Mount, Dave and Aloimonos, J.} } @article {14252, title = {Recognizing 3-D Motion}, journal = {INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE}, volume = {13}, year = {1993}, month = {1993///}, pages = {1624 - 1624}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {14266, title = {The role of fixation in visual motion analysis}, journal = {International Journal of Computer Vision}, volume = {11}, year = {1993}, month = {1993///}, pages = {165 - 186}, abstract = {How does the ability of humans and primates to fixate at environmental points in the presence of relative motion help their visual systems in solving various tasks? To state the question in a more formal setting, we investigate in this article the following problem: Suppose that we have an active vision system, that is, a camera resting on a platform and being controlled through motors by a computer that has access to the images sensed by the camera in real time. The platform can move freely in the environment. If this machine can fixate on targets being in relative motion with it, can it solve visual tasks in an efficient and robust manner? By restricting our attention to a set of navigational tasks, we find that such an active observer can solve the problems of 3-D motion estimation, egomotion recovery, and estimation of time-to-contact in a very efficient manner, using as input the spatiotemporal derivatives of the image-intensity function (or normal flow). Fixation over time changes the input (motion field) in a controlled way and from this change additional information is derived making the previously mentioned tasks easier to solve.}, isbn = {0920-5691}, url = {http://dx.doi.org/10.1007/BF01469227}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {16513, title = {Simple systems that exhibit template-directed replication}, journal = {Science}, volume = {259}, year = {1993}, month = {1993///}, pages = {1282 - 1287}, author = {Reggia, James A. and Armentrout,S. L and Chou,H. H. and Peng,Y.} } @conference {11978, title = {Active egomotion estimation: a qualitative approach}, booktitle = {Computer Vision{\textemdash}ECCV{\textquoteright}92}, year = {1992}, month = {1992///}, pages = {497 - 510}, author = {Aloimonos, J. and Duric, Z.} } @article {14199, title = {Behavioral Visual Motion Analysis,"}, journal = {Proceedings, DARPA Image Understanding Workshop}, year = {1992}, month = {1992///}, pages = {521 - 541}, author = {Aloimonos, J. and Duri{\c c},Z. and Ferm{\"u}ller, Cornelia and Huang, L. and Rivlin,E. and Sharma, R.} } @article {12044, title = {Coordinated motion planning: the warehouseman{\textquoteright}s problem with constraints on free space}, journal = {Systems, Man and Cybernetics, IEEE Transactions on}, volume = {22}, year = {1992}, month = {1992///}, pages = {130 - 141}, author = {Sharma, R. and Aloimonos, J.} } @article {12987, title = {copia-like retrotransposons are ubiquitous among plants}, journal = {Proc Natl Acad Sci USA}, volume = {89}, year = {1992}, month = {1992/08//}, pages = {7124 - 7128}, abstract = {Transposable genetic elements are assumed to be a feature of all eukaryotic genomes. Their identification, however, has largely been haphazard, limited principally to organisms subjected to molecular or genetic scrutiny. We assessed the phylogenetic distribution of copia-like retrotransposons, a class of transposable element that proliferates by reverse transcription, using a polymerase chain reaction assay designed to detect copia-like element reverse transcriptase sequences. copia-like retrotransposons were identified in 64 plant species as well as the photosynthetic protist Volvox carteri. The plant species included representatives from 9 of 10 plant divisions, including bryophytes, lycopods, ferns, gymnosperms, and angiosperms. DNA sequence analysis of 29 cloned PCR products and of a maize retrotransposon cDNA confirmed the identity of these sequences as copia-like reverse transcriptase sequences, thereby demonstrating that this class of retrotransposons is a ubiquitous component of plant genomes.}, author = {Voytas,D. F and Cummings, Michael P. and Koniczny,A. and Ausubel,F. M and Rodermel,S. R} } @conference {17106, title = {Dynamic queries: database searching by direct manipulation}, booktitle = {Proceedings of the SIGCHI conference on Human factors in computing systems}, series = {CHI {\textquoteright}92}, year = {1992}, month = {1992///}, pages = {669 - 670}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, isbn = {0-89791-513-5}, doi = {10.1145/142750.143082}, url = {http://doi.acm.org/10.1145/142750.143082}, author = {Shneiderman, Ben and Williamson,Christopher and Ahlberg,Christopher} } @conference {17103, title = {Dynamic queries for information exploration: an implementation and evaluation}, booktitle = {Proceedings of the SIGCHI conference on Human factors in computing systems}, series = {CHI {\textquoteright}92}, year = {1992}, month = {1992///}, pages = {619 - 626}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {We designed, implemented and evaluated a new concept for direct manipulation of databases, called dynamic queries, that allows users to formulate queries with graphical widgets, such as sliders. By providing a graphical visualization of the database and search results, users can find trends and exceptions easily. Eighteen undergraduate chemistry students performed statistically significantly faster using a dynamic queries interface compared to two interfaces both providing form fill-in as input method, one with graphical visualization output and one with all-textual output. The interfaces were used to explore the periodic table of elements and search on their properties.}, isbn = {0-89791-513-5}, doi = {10.1145/142750.143054}, url = {http://doi.acm.org/10.1145/142750.143054}, author = {Ahlberg,Christopher and Williamson,Christopher and Shneiderman, Ben} } @conference {15207, title = {Efficient minimum cost matching using quadrangle inequality}, booktitle = {Foundations of Computer Science, 1992. Proceedings., 33rd Annual Symposium on}, year = {1992}, month = {1992/10//}, pages = {583 - 592}, abstract = {The authors present efficient algorithms for finding a minimum cost perfect matching, and for solving the transportation problem in bipartite graphs, G = (Red cup; Blue, Red times; Blue), where |Red| = n, |Blue| = m, n les; m, and the cost function obeys the quadrangle inequality. The first results assume that all the red points and all the blue points lie on a curve that is homeomorphic to either a line or a circle and the cost function is given by the Euclidean distance along the curve. They present a linear time algorithm for the matching problem. They generalize the method to solve the corresponding transportation problem in O((m+n)log(m+n)) time. The next result is an O(n log m) algorithm for minimum cost matching when the cost array is a bitonic Monge array. An example of this is when the red points lie on one straight line and the blue points lie on another straight line (that is not necessarily parallel to the first one). Finally, they provide a weakly polynomial algorithm for the transportation problem in which the associated cost array is a bitonic Monge array}, keywords = {algorithm;, array;, bipartite, bitonic, blue, complexity;, computational, cost, distance;, Euclidean, function;, geometry;, graph, graphs;, inequality;, linear, MATCHING, matching;, minimisation;, minimum, Monge, perfect, points;, polynomial, problem;, quadrangle, red, theory;, TIME, transportation, transportation;, weakly}, doi = {10.1109/SFCS.1992.267793}, author = {Aggarwal,A. and Bar-Noy,A. and Khuller, Samir and Kravets,D. and Schieber,B.} } @article {18075, title = {Efficient pattern matching with scaling}, journal = {Journal of Algorithms}, volume = {13}, year = {1992}, month = {1992///}, pages = {2 - 32}, author = {Amir, A. and Landau,G. M and Vishkin, Uzi} } @conference {12020, title = {Exploratory active vision: theory}, booktitle = {Proceedings of 1992 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 1992}, year = {1992}, month = {1992/06/15/18}, pages = {10 - 15}, publisher = {IEEE}, organization = {IEEE}, abstract = {An active approach to the integration of shape from x modules-here shape from shading and shape from texture-is proposed. The question of what constitutes a good motion for the active observer is addressed. Generally, the role of the visual system is to provide depth information to an autonomous robot; a trajectory module will then interpret it to determine a motion for the robot, which in turn will affect the visual information received. It is suggested that the motion can also be chosen so as to improve the performance of the visual system}, keywords = {active observer, Active shape model, active vision, Automation, autonomous robot, CAMERAS, Computer science, Computer vision, depth information, Laboratories, Layout, Mobile robots, Motion analysis, optical flow, Robots, shape from shading, shape from texture, shape from x modules, trajectory module, Visual system}, isbn = {0-8186-2855-3}, doi = {10.1109/CVPR.1992.223234}, author = {Herve,J. -Y and Aloimonos, J.} } @conference {14246, title = {Perceptual computational advantages of tracking}, booktitle = {, 11th IAPR International Conference on Pattern Recognition, 1992. Vol.I. Conference A: Computer Vision and Applications, Proceedings}, year = {1992}, month = {1992/09/30/Aug-3}, pages = {599 - 602}, publisher = {IEEE}, organization = {IEEE}, abstract = {The paradigm of active vision advocates studying visual problems in the form of modules that are directly related to a visual task for observers that are active. It is argued that in many cases when an object is moving in an unrestricted manner (translation and rotation) in the 3D world only the motion{\textquoteright}s translational components are of interest. For a monocular observer, using only the normal flow-the spatiotemporal derivatives of the image intensity function-the authors solve the problem of computing the direction of translation. Their strategy uses fixation and tracking. Fixation simplifies much of the computation by placing the object at the center of the visual field, and the main advantage of tracking is the accumulation of information over time. The authors show how tracking is accomplished using normal flow measurements and use it for two different tasks in the solution process. First, it serves as a tool to compensate for the lack of existence of an optical flow field and thus to estimate the translation parallel to the image plane; and second, it gathers information about the motion component perpendicular to the image plane}, keywords = {active vision, Automation, Employment, fixation, image intensity function, Image motion analysis, IMAGE PROCESSING, Motion estimation, Nonlinear optics, Optical computing, Optical sensors, parameter estimation, pattern recognition, perceptual computational advantages, spatiotemporal derivatives, Spatiotemporal phenomena, tracking, unrestricted motion, visual flow measurements}, isbn = {0-8186-2910-X}, doi = {10.1109/ICPR.1992.201633}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {16650, title = {Simple systems exhibiting self-directed replication: annex of transition functions and software documentation}, year = {1992}, month = {1992///}, institution = {University of Maryland at College Park}, address = {College Park, MD, USA}, author = {Reggia, James A. and Chou,Hui-Hsien and Armentrout,Steven L. and Peng,Yun} } @article {17971, title = {Six Generations of Building Walkthrough: Final Technical Report to the National Science Foundation}, year = {1992}, month = {1992///}, institution = {University of North Carolina at Chapel Hill}, address = {Chapel Hill, NC, USA}, author = {Brooks,Frederick P. and Airey,John and Alspaugh,John and Bell,Andrew and Brown,Randolph and Hill,Curtis and Nimscheck,Uwe and Rheingans,Penny and Rohlf,John and Smith,Dana and Turner,Douglass and Varshney, Amitabh and Wang,Yulan and Weber,Hans and Yuan,Xialin} } @article {14270, title = {Tracking facilitates 3-D motion estimation}, journal = {Biological Cybernetics}, volume = {67}, year = {1992}, month = {1992///}, pages = {259 - 268}, abstract = {The recently emerging paradigm of Active Vision advocates studying visual problems in form of modules that are directly related to a visual task for observers that are active. Along these lines, we are arguing that in many cases when an object is moving in an unrestricted manner (translation and rotation) in the 3D world, we are just interested in the motion{\textquoteright}s translational components. For a monocular observer, using only the normal flow {\textemdash} the spatio-temporal derivatives of the image intensity function {\textemdash} we solve the problem of computing the direction of translation and the time to collision. We do not use optical flow since its computation is an ill-posed problem, and in the general case it is not the same as the motion field {\textemdash} the projection of 3D motion on the image plane. The basic idea of our motion parameter estimation strategy lies in the employment of fixation and tracking. Fixation simplifies much of the computation by placing the object at the center of the visual field, and the main advantage of tracking is the accumulation of information over time. We show how tracking is accomplished using normal flow measurements and use it for two different tasks in the solution process. First it serves as a tool to compensate for the lack of existence of an optical flow field and thus to estimate the translation parallel to the image plane; and second it gathers information about the motion component perpendicular to the image plane.}, isbn = {0340-1200}, url = {http://dx.doi.org/10.1007/BF00204399}, author = {Ferm{\"u}ller, Cornelia and Aloimonos, J.} } @article {11929, title = {Computer based systems engineering workshop}, journal = {Software Engineering Education}, year = {1991}, month = {1991///}, pages = {149 - 163}, abstract = {Modern computer based systems are complex multi-systems consisting of many connected individual subsystems; each one of them is typically also a multicomputer system. The subsystems in a multi-system can be either geographically distributed or locally connected systems. Typical examples of computer based systems are medical systems, process control systems, communications systems, weapon systems and large information systems.The development of these complex systems requires the establishment of a new engineering discipline in its own right, Computer Based Systems Engineering {\textemdash} CBSE. The definition of the discipline, its current and future practice and the ways to establish and promote it were discussed in an international IEEE workshop held in Neve-Ilan, Israel in May 1990. The major conclusion of the workshop was that CBSE should be established as a new field in its own right. To achieve this goal, the workshop participants recommended that the IEEE Computer Society shall set up a task force for the promotion of the field, the establishment of CBSE Institutes and the development of the educational framework of CBSE. The paper describes the major findings of the workshop that led to these conclusions and recommendations. }, author = {Lavi,J. and Agrawala, Ashok K. and Buhr,R. and Jackson,K. and Jackson,M. and Lang,B.} } @conference {11928, title = {DTP: An Efficient Transport Protocol}, booktitle = {In Proceedings of the IFIP TC6 Working Conference on ComACM Computing Surveys}, year = {1991}, month = {1991///}, pages = {91 - 133}, abstract = {We recently introduced a new flow control scheme, called send-time control, which is based on a deterministic model of virtual circuits in a computer network. In this scheme, the time at which a packet is sent by a source is computed from estimates of round-trip time, traffic in the network and bottleneck service time. In this paper, we describe a new transport protocol, called DTP, which uses send-time control as its flow control scheme. Preliminary measurements of coast-to-coast connections over the Internet show significant performance improvement over TCP, which is the most commonly used transport protocol in the Internet today. Keyword Codes: C.2.2 Keywords: Computer-Communication Networks, Network Protocols 1. Introduction Flow control is one of the most important function at the transport layer of the protocol hierarchy. The function of flow control schemes is to regulate the traffic entering into the network in order to avoid overloading of network resources, thus preventing c...}, author = {Sanghi,Dheeraj and Agrawala, Ashok K.} } @article {11930, title = {Modeling of cross traffic in conjunction with deterministic analysis of queues}, year = {1991}, month = {1991///}, institution = {University of Maryland at College Park}, address = {College Park, MD, USA}, author = {Agrawala, Ashok K. and Sanghi,Dheeraj} } @article {11989, title = {A multi-frame approach to visual motion perception}, journal = {International Journal of Computer Vision}, volume = {6}, year = {1991}, month = {1991///}, pages = {245 - 255}, author = {Spetsakis, M. and Aloimonos, J.} } @conference {11990, title = {Relative depth from motion using normal flow: an active and purposive solution}, booktitle = {Proceedings of the IEEE Workshop on Visual Motion, 1991}, year = {1991}, month = {1991/10/07/9}, pages = {196 - 204}, publisher = {IEEE}, organization = {IEEE}, abstract = {The authors show how an active observer can compute the relative depth of (stationary or moving) objects in the field of view using only the spatiotemporal derivatives of the time varying image intensity function. The solution they propose is purposive in the sense that it solves only the relative depth from motion problem and cannot be used for other problems related to motion; active in the sense that the activity of the observer is essential for the solution of the problem. Results indicate that exact computation of retinal motion does not appear to be a necessary first step for some problems related to visual motion. In addition, optic flow, whose computation is an ill-posed problem, is related to the motion of the scene only under very restrictive assumptions. As a result, the use of optic flow in some quantitative motion analysis studies is questionable}, keywords = {3D information, Automation, Computer vision, Educational institutions, Image motion analysis, image recognition, Image sequences, Laboratories, Layout, Motion analysis, Motion estimation, Nonlinear optics, normal flow, optic flow, Optical computing, Optical sensors, relative depth, retinal motion, spatiotemporal derivatives, time varying image intensity function, visual motion}, isbn = {0-8186-2153-2}, doi = {10.1109/WVM.1991.212807}, author = {Huang, Liuqing and Aloimonos, J.} } @article {12977, title = {A superfamily of {\i}t Arabidopsis thaliana retrotransposons}, journal = {Genetics}, volume = {127}, year = {1991}, month = {1991/04//}, pages = {801 - 809}, abstract = {We describe a superfamily of Arabidopsis thaliana retrotransposable elements that consists of at least ten related families designated Ta1-Ta10. The Ta1 family has been described previously. Two genomic clones representing the Ta2 and Ta3 elements were isolated from an A. thaliana (race Landsberg erecta) lambda library using sequences derived from the reverse transcriptase region of Ta1 as hybridization probes. Nucleotide sequence analysis showed that the Ta1, Ta2 and Ta3 families share greater than 75\% amino acid identity in pairwise comparisons of their reverse transcriptase and RNase H genes. In addition to Ta1, Ta2 and Ta3, we identified seven other related retrotransposon families in Landsberg erecta, Ta4-Ta10, using degenerate primers and the polymerase chain reaction to amplify a highly conserved region of retrotransposon-encoded reverse transcriptase. One to two copies of elements Ta2-Ta10 are present in the genomes of the A. thaliana races Landsberg erecta and Columbia indicating that the superfamily comprises at least 0.1\% of the A. thaliana genome. The nucleotide sequences of the reverse transcriptase regions of the ten element families place them in the category of copia-like retrotransposons and phylogenetic analysis of the amino acid sequences suggests that horizontal transfer may have played a role in their evolution.}, author = {Konieczny,A and Voytas,D. F and Cummings, Michael P. and Ausubel,F. M} } @article {11950, title = {On the visual mathematics of tracking}, journal = {Image and vision computing}, volume = {9}, year = {1991}, month = {1991///}, pages = {235 - 251}, author = {Aloimonos, J. and Tsakiris, D. P} } @book {11977, title = {Visual recovery}, year = {1991}, month = {1991///}, publisher = {University of Maryland, Center for Automation Research, Computer Vision Laboratory}, organization = {University of Maryland, Center for Automation Research, Computer Vision Laboratory}, author = {Aloimonos, J. and Rosenfeld, A.} } @conference {12062, title = {Catastrophe theory and computational vision}, booktitle = {Proc. AAAI-90 Workshop on Qualitative Vision}, year = {1990}, month = {1990///}, pages = {96 - 100}, author = {Herve, J. and Aloimonos, J.} } @article {11988, title = {Image motion estimation by clustering}, journal = {International Journal of Imaging Systems and Technology}, volume = {2}, year = {1990}, month = {1990/12/01/}, pages = {345 - 355}, abstract = {Image motion is estimated by matching feature {\textquotedblleft}interest{\textquotedblright} points in different frames of video image sequences. The matching is based on local similarity of the displacement vectors. Clustering in the displacement vector space is used to determine the set of plausible match vectors. Subsequently, a similarity-based algorithm performs the actual matching. The feature points are computed using a multiple-filter image decomposition operator. The algorithm has been tested on synthetic as well as real video images. The novelty of the approach is that it handles multiple motions and performs motion segmentation.}, isbn = {1098-1098}, doi = {10.1002/ima.1850020409}, url = {http://onlinelibrary.wiley.com/doi/10.1002/ima.1850020409/abstract}, author = {Bandopadhay, Amit and Aloimonos, J.} } @article {11971, title = {Motion--Boundary Illusions and their Regularization}, journal = {Proceedings of the Royal Society of London. Series B: Biological Sciences}, volume = {242}, year = {1990}, month = {1990/11/22/}, pages = {75 - 81}, abstract = {Humans use various cues to understand the structure of the world from images. One such cue is the contours of an object formed by occlusion or from surface discontinuities. It is known that contours in the image of an object provide various amounts of information about the shape of the object in view, depending on assumptions that the observer makes. Another powerful cue is motion. The ability of the human visual system to discern structure from a motion stimulus is well known and has a solid theoretical and experimental foundation. However, when humans interpret a visual scene they use various cues to understand what they observe, and the interpretation comes from combining the information acquired from the various modules devoted to specific cues. In such an integration of modules it seems that each cue carries a different weight and importance. We performed several experiments where we made sure that the only cues available to the observer were contour and motion. It turns out that when humans combine information from contour and motion to reconstruct the shape of an object in view, if the results of the two modules - shape from contour and structure from motion - are inconsistent, they experience a perceptual result which is due to the combination of the two modules, with the influence of the contour dominating, thus giving rise to the illusion. We describe here examples of such illusions and identify the conditions under which they happen. Finally, we introduce a computational theory for combining contour and motion using the theory of regularization. The theory explains such illusions and predicts many more. The same computational theory, when applied to retinal motion estimation, explains the effect of boundaries on the perception of motion that gives rise to a set of well-known illusions described by Wallach (1976).}, doi = {10.1098/rspb.1990.0106}, url = {http://rspb.royalsocietypublishing.org/content/242/1304/75.abstract}, author = {Aloimonos, J. and Huang, Liuqing} } @article {16800, title = {Optimal view caching}, journal = {Information Systems}, volume = {15}, year = {1990}, month = {1990///}, pages = {169 - 171}, abstract = {A view in a database is a subset of records selected from several files to satisfy some condition, e.g. the join of two relations.It has been shown that it is more efficient to store pointers to the satisfying records of the base relations rather than the complete records of the view. We are interested in efficient caching of such pointers (in effect constructing the view) when the database lies in secondary storage and only a limited number of buffers exists in memory. A view caching is optimal if it is done with the minimum number of reads from secondary storage. We prove that optimal view caching is NP -complete. }, isbn = {0306-4379}, doi = {16/0306-4379(90)90032-K}, url = {http://www.sciencedirect.com/science/article/pii/030643799090032K}, author = {Amir,Amihood and Roussopoulos, Nick} } @article {12059, title = {Perspective approximations}, journal = {Image and Vision Computing}, volume = {8}, year = {1990}, month = {1990/08//}, pages = {179 - 192}, abstract = {In recent years, researchers in computer vision working on problems such as object recognition, shape reconstruction, shape from texture, shape from contour, pose estimation, etc., have employed in their analyses approximations of the perspective projection as the image formation process. Depending on the task, these approximations often yield very good results and present the advantage of simplicity. Indeed when one relates lengths, angles or areas on the image with the respective units in the 3D world assuming perspective projection, the resulting expressions are very complex, and consequently they complicate the recovery process. However, if we assume that the image is formed with a projection which is a good approximation of the perspective, then the recovery process becomes easier. Two such approximations, are described, the paraperspective and the orthoperspective, and it is shown that for some tasks the error introduced by the use of such an approximation is negligible. Applications of these projections to the problems of shape from texture, shape from contour, and object recognition related problems (such as determining the view vector and pose estimation) are also described.}, keywords = {orthoperspective, paraperspective, perspective approximations}, isbn = {0262-8856}, doi = {16/0262-8856(90)90064-C}, url = {http://www.sciencedirect.com/science/article/pii/026288569090064C}, author = {Aloimonos, J.} } @conference {11960, title = {Purposive and qualitative active vision}, booktitle = {Proceedings of 10th International Conference on Pattern Recognition, 1990}, volume = {i}, year = {1990}, month = {1990/06/16/21}, pages = {346-360 vol.1 - 346-360 vol.1}, publisher = {IEEE}, organization = {IEEE}, abstract = {The traditional view of the problem of computer vision as a recovery problem is questioned, and the paradigm of purposive-qualitative vision is offered as an alternative. This paradigm considers vision as a general recognition problem (recognition of objects, patterns or situations). To demonstrate the usefulness of the framework, the design of the Medusa of CVL is described. It is noted that this machine can perform complex visual tasks without reconstructing the world. If it is provided with intentions, knowledge of the environment, and planning capabilities, it can perform highly sophisticated navigational tasks. It is explained why the traditional structure from motion problem cannot be solved in some cases and why there is reason to be pessimistic about the optimal performance of a structure from motion module. New directions for future research on this problem in the recovery paradigm, e.g., research on stability or robustness, are suggested}, keywords = {active vision, Automation, brain models, complex visual tasks, Computer vision, environmental knowledge, highly sophisticated navigational tasks, HUMANS, Image reconstruction, intentions, Kinetic theory, Laboratories, Medusa, Motion analysis, Navigation, planning, planning (artificial intelligence), purposive-qualitative vision, recovery problem, Robust stability, Robustness, SHAPE, stability}, isbn = {0-8186-2062-5}, doi = {10.1109/ICPR.1990.118128}, author = {Aloimonos, J.} } @article {13025, title = {The structure, distribution and evolution of the {\i}t Ta1 retrotransposable element family of {\i}t Arabidopsis thaliana}, journal = {Genetics}, volume = {126}, year = {1990}, month = {1990/11//}, pages = {713 - 721}, abstract = {The Ta1 elements are a low copy number, copia-like retrotransposable element family of Arabidopsis thaliana. Six Ta1 insertions comprise all of the Ta1 element copies found in three geographically diverse A. thaliana races. These six elements occupy three distinct target sites: Ta1-1 is located on chromosome 5 and is common to all three races (Col-0, Kas-1 and La-0). Ta1-2 is present in two races on chromosome 4 (Kas-1 and La-0), and Ta1-3, also located on chromosome 4, is present only in one race (La-0). The six Ta1 insertions share greater than 96\% nucleotide identity, yet are likely to be incapable of further transposition due to deletions or nucleotide changes that alter either the coding capacity of the elements or conserved protein domains required for retrotransposition. Nucleotide sequence comparisons of these elements and the distribution of Ta1 among 12 additional A. thaliana geographical races suggest that Ta1-1 predated the global dispersal of A. thaliana. As the species spread throughout the world, two additional transposition events occurred which gave rise first to Ta1-2 and finally to Ta1-3.}, author = {Voytas,D. F and Konieczny,A and Cummings, Michael P. and Ausubel,F. M} } @article {12009, title = {Structure from motion using line correspondences}, journal = {International Journal of Computer VisionInt J Comput Vision}, volume = {4}, year = {1990}, month = {1990/06//}, pages = {171 - 183}, isbn = {0920-5691}, doi = {10.1007/BF00054994}, url = {http://www.springerlink.com/content/r4h10053528335k2/}, author = {Spetsakis, Minas E. and Aloimonos, J.} } @conference {11991, title = {A Unified Theory Of Structure From Motion}, booktitle = {Proceedings of DARPA Image Understanding Workshop}, year = {1990}, month = {1990///}, pages = {271 - 283}, author = {Spetsakis, M. and Aloimonos, J.} } @article {11931, title = {Distributed Checkpointing for Globally Consistent States of Databases}, journal = {IEEE Trans. Softw. Eng.}, volume = {15}, year = {1989}, month = {1989/10//}, pages = {1157 - 1167}, keywords = {Availability, checkpoint, consistency, distributed database, noninterference, recovery, transaction}, isbn = {0098-5589}, doi = {10.1109/TSE.1989.559763}, url = {http://dx.doi.org/10.1109/TSE.1989.559763}, author = {Son,Sang Hyuk and Agrawala, Ashok K.} } @article {11932, title = {Dynamic bin packing problem}, journal = {SIAM Journal on Computing}, volume = {12}, year = {1989}, month = {1989///}, pages = {227 - 258}, abstract = {In a previous work we have defined a general architecture model for autonomous systems, which can be mapped easily to describe the functions of any automated system (SDAG-86-01). In this note, we use the model to describe the problem of thermal management in space stations. First we briefly review the architecture, then we present the environment of our application, and finally we detail the specific function for each functional block of the architecture for that environment.}, author = {Dikshit,Piyush and Guimaraes,Katia and Ramamurthy,Maya and Agrawala, Ashok K. and Larsen,Ronald L} } @article {12023, title = {On the kinetic depth effect}, journal = {Biological cybernetics}, volume = {60}, year = {1989}, month = {1989///}, pages = {445 - 455}, author = {Aloimonos, J. and Brown, C. M.} } @article {11954, title = {Learning early-vision computations}, journal = {JOSA A}, volume = {6}, year = {1989}, month = {1989///}, pages = {908 - 919}, author = {Aloimonos, J. and Shulman, D.} } @conference {11974, title = {Optimal motion estimation}, booktitle = {Workshop on Visual Motion, 1989.,Proceedings}, year = {1989}, month = {1989/03/20/22}, pages = {229 - 237}, publisher = {IEEE}, organization = {IEEE}, abstract = {The problem of using feature correspondences to recover the structure and 3D motion of a moving object from its successive images is analyzed. They formulate the problem as a quadratic minimization problem with a nonlinear constraint. Then they derive the condition for the solution to be optimal under the assumption of Gaussian noise in the input, in the maximum-likelihood-principle sense. The authors present two efficient ways to approximate it and discuss some inherent limitations of the structure-from-motion problem when two frames are used that should be taken into account in robotics applications that involve dynamic imagery. Finally, it is shown that some of the difficulties inherent in the two-frame approach disappear when redundancy in the data is introduced. This is concluded from experiments using a structure-from-motion algorithm that is based on multiple frames and uses only the rigidity assumption}, keywords = {3D motion interpretation, Automation, Computer vision, computerised pattern recognition, computerised picture processing, constraint minimization, dynamic imagery, Educational institutions, feature correspondences, Gaussian noise, Image motion analysis, Laboratories, maximum-likelihood-principle, Minimization methods, Motion analysis, Motion estimation, motion parameters, moving object, multiple frames, nonlinear constraint, Optical computing, optimal motion estimation, parameter estimation, quadratic minimization, quadratic programming, redundancy, rigidity assumption, robotics applications, structure-from-motion, successive images, two-frame}, isbn = {0-8186-1903-1}, doi = {10.1109/WVM.1989.47114}, author = {Spetsakis, M. E and Aloimonos, J.} } @article {14117, title = {Ordering techniques for the preconditioned conjugate gradient method on parallel computers}, journal = {Computer Physics Communications}, volume = {53}, year = {1989}, month = {1989/05//}, pages = {253 - 269}, abstract = {We consider the parallel implementation of the preconditioned conjugate gradient method using multicolor incomplete factorization as preconditioners. We discuss numerical experiments on sample problems arising from elliptic partial differential equations, together with an analytic study of the effects of communication and artihmetic costs on loosely coupled architectures. Our main conclusion is that multicolor orderings result in slower convergence of the preconditioned conjugate gradient method than natural orderings, but that the lower parallel costs of the multicolor techniques typically make their overall performance better.}, isbn = {0010-4655}, doi = {16/0010-4655(89)90164-1}, url = {http://www.sciencedirect.com/science/article/pii/0010465589901641}, author = {Elman, Howard and Agr{\'o}n, Elvira} } @conference {12002, title = {Unification and integration of visual modules}, booktitle = {Proceedings Image Understanding Workshop}, year = {1989}, month = {1989///}, pages = {507 - 551}, author = {Aloimonos, J.} } @article {12007, title = {Unifying Shading and Texture Through an Active Observer}, journal = {Proceedings of the Royal Society of London. B. Biological Sciences}, volume = {238}, year = {1989}, month = {1989/10/23/}, pages = {25 - 37}, abstract = {Shading (variations of image intensity) provides an important cue for understanding the shape of three-dimensional surfaces from monocular views. On the other hand. texture (distribution of discontinuities on the surface) is a strong cue for recovering surface orientation by using monocular images. But given the image of an object or scene, what technique should we use to recover the shape of what is imaged? Resolution of shape from shading requires knowledge of the reflectance of the imaged surface and, usually, the fact that it is smooth (i.e. it shows no discontinuities). Determination of shape from texture requires knowledge of the distribution of surface markings (i.e. discontinuities). One might expect that one method would work when the other does not. I present a theory on how an active observer can determine shape from the image of an object or scene regardless of whether the image is shaded, textured. or both, and without any knowledge of reflectance maps or the distribution of surface markings. The approach is successful because the active observer is able to manipulate the constraints behind the perceptual phenomenon at hand and thus derive a simple solution. Several experimental results are presented with real and synthetic images.}, doi = {10.1098/rspb.1989.0064}, url = {http://rspb.royalsocietypublishing.org/content/238/1290/25.abstract}, author = {Aloimonos, J.} } @article {12005, title = {Active vision}, journal = {International Journal of Computer Vision}, volume = {1}, year = {1988}, month = {1988///}, pages = {333 - 356}, author = {Aloimonos, J. and Weiss, I. and Bandyopadhyay, A.} } @book {11943, title = {Correspondenceless detection of depth and motion for a planar surface}, year = {1988}, month = {1988///}, publisher = {University of Maryland}, organization = {University of Maryland}, author = {Aloimonos, J. and Herv{\'e}, J. Y} } @conference {12010, title = {The Maryland approach to image understanding}, booktitle = {Science Applications International Corp, Proceedings: Image Understanding Workshop,}, volume = {1}, year = {1988}, month = {1988///}, author = {Aloimonos, J. and Davis, Larry S. and Rosenfeld, A.} } @conference {11957, title = {Optimal Computing Of Structure From Motion Using Point Correspondences In Two Frames}, booktitle = {Proceedings of Second International Conference on Computer Vision}, year = {1988}, month = {1988/12/05/8}, pages = {449 - 453}, publisher = {IEEE}, organization = {IEEE}, keywords = {Automation, Computer vision, Educational institutions, Gaussian noise, Image motion analysis, Laboratories, Least squares approximation, Least squares methods, Motion estimation, Optical computing}, isbn = {0-8186-0883-8}, doi = {10.1109/CCV.1988.590022}, author = {Spetsakis, M. E and Aloimonos, J.} } @article {18098, title = {Parallel construction of a suffix tree with applications}, journal = {Algorithmica}, volume = {3}, year = {1988}, month = {1988///}, pages = {347 - 365}, author = {Apostolico,A. and Iliopoulos,C. and Landau,G. M and Schieber,B. and Vishkin, Uzi} } @article {12011, title = {Robust computation of intrinsic images from multiple cues}, journal = {Advances in Computer Vision}, volume = {1}, year = {1988}, month = {1988///}, pages = {115 - 163}, author = {Aloimonos, J. and Brown, C. M.} } @article {12067, title = {Shape from patterns: Regularization}, journal = {International journal of computer vision}, volume = {2}, year = {1988}, month = {1988///}, pages = {171 - 187}, author = {Aloimonos, J. and Swain, M.} } @article {12018, title = {Shape from texture}, journal = {Biological Cybernetics}, volume = {58}, year = {1988}, month = {1988///}, pages = {345 - 360}, author = {Aloimonos, J.} } @conference {11962, title = {Using flow field divergence for obstacle avoidance in visual navigation}, booktitle = {Science Applications International Corp, Proceedings: Image Understanding Workshop,}, volume = {2}, year = {1988}, month = {1988///}, author = {Nelson, R. and Aloimonos, J.} } @article {11994, title = {Visual shape computation}, journal = {Proceedings of the IEEE}, volume = {76}, year = {1988}, month = {1988/08//}, pages = {899 - 916}, abstract = {Perceptual processes responsible for computing shape from several cues, including shading, texture, contour, and stereo, are examined. It is noted that these computational problems, as well as that of computing shaping from motion, are ill-posed in the sense of Hadamard. It is suggested that regularization theory can be used along with a priori knowledge to restrict the space of possible solutions, and thus restore the problem{\textquoteright}s well-prosedness. Some alternative methods are outlined, and the idea of active vision is explored briefly in connection with the problem}, keywords = {a priori knowledge, active vision, computational problems, Computer vision, computing shaping from motion, contour, cues, Focusing, HUMANS, ill posed problems, Machine vision, Psychology, regularization theory, RETINA, sense of Hadamard, shading, SHAPE, space of possible solutions, Stereo vision, Surface texture, TEXTURE, visual shape computation, Visual system}, isbn = {0018-9219}, doi = {10.1109/5.5964}, author = {Aloimonos, J.} } @conference {12024, title = {Determining three dimensional transformation parameters from images: Theory}, booktitle = {1987 IEEE International Conference on Robotics and Automation. Proceedings}, volume = {4}, year = {1987}, month = {1987/03//}, pages = {57 - 61}, publisher = {IEEE}, organization = {IEEE}, abstract = {We present a theory for the determination of the three dimensional transformation parameters of an object, from its images. The input to this process is the image intensity function and its temporal derivative. In particular, our results are: 1) If the structure of the transforming object in view is known, then the transformation parameters are determined from the solution of a linear system. Rigid motion is a special ease of our theory. 2)If the structure of the object in view is not known, then both the structure and transformation parameters may be computed through a hill climbing or simulated annealing algorithm.}, keywords = {Automation, Computational modeling, Educational institutions, Image motion analysis, Indium tin oxide, Linear systems, Motion estimation, Optical computing, Robots, Simulated annealing}, doi = {10.1109/ROBOT.1987.1088011}, author = {Ito, E. and Aloimonos, J.} } @article {12048, title = {Learning shape computations}, journal = {Proc. DARPA Image}, year = {1987}, month = {1987///}, author = {Aloimonos, J. and Shulman, D.} } @book {11944, title = {A robust algorithm for determining the translation of a rigidly moving surface without correspondence, for robotics applications}, year = {1987}, month = {1987///}, publisher = {University of Maryland}, organization = {University of Maryland}, author = {Basu, A. and Aloimonos, J.} } @article {11953, title = {Spatiotemporal blur paths for image flow estimation (A)}, journal = {Journal of the Optical Society of America A}, volume = {4}, year = {1987}, month = {1987///}, pages = {35 - 35}, author = {Spetsakis, M. E and Aloimonos, J.} } @book {12064, title = {Structure from Motion from Line Correspondencies: New Results}, year = {1987}, month = {1987///}, author = {Science, University of Maryland at College Park. Dept. of Computer and Spetsakis, M. E and Aloimonos, J. and Research, University of Maryland at College Park. Center for Automation} } @article {18104, title = {Tight comparison bounds on the complexity of parallel sorting}, journal = {SIAM J. Comput.}, volume = {16}, year = {1987}, month = {1987///}, pages = {458 - 464}, author = {Azar,Y. and Vishkin, Uzi} } @article {17428, title = {Time Stress Effects on Two Menu Selection Systems}, journal = {Proceedings of the Human Factors and Ergonomics Society Annual MeetingProceedings of the Human Factors and Ergonomics Society Annual Meeting}, volume = {31}, year = {1987}, month = {1987/09/01/}, pages = {727 - 731}, abstract = {The optimal number of menu items per display screen has been the topic of considerable debate and study. On the one hand, some designers have packed many items into each menu to conserve space and reduce the number of menus, whereas on the other hand there are designers who prefer a sparse display for menu structures and other videotex information. This study evaluated the effects of a broad/shallow menu compared to a narrow/deep menu structure under two conditions of time stress for inexperienced users. Results indicated that time stress both slowed performance, and increased errors. In addition, it was demonstrated that the broad/shallow menu was faster and resulted in fewer errors. Implications for menu design are discussed.}, isbn = {1071-1813,}, doi = {10.1177/154193128703100708}, url = {http://pro.sagepub.com/content/31/7/727}, author = {Wallace,Daniel F and Anderson,Nancy S and Shneiderman, Ben} } @article {11933, title = {Algorithm for concurrency control in replicated databases.}, journal = {COMP. SYST. SCI. ENG.}, volume = {1}, year = {1986}, month = {1986///}, pages = {75 - 81}, abstract = {A replicated database system is a distributed database system in which some data objects are stored redundantly at multiple sites to improve the reliability of the system. Without proper control mechanisms, the consistency of a replicated database system could be violated. A new scheme to increase the reliability as well as the degree of concurrency is described. It allows transactions to operate on a data object so long as one or more token copies are available. The scheme also exploits the fact that, for recovery reasons, there are two values for one data object. It is proved that transactions executed according to this scheme preserve consistency.}, author = {Son,S. H and Agrawala, Ashok K.} } @article {12051, title = {Computing Intrinsic Images.}, year = {1986}, month = {1986/08//}, abstract = {Low level modern computer vision is not domain dependent, but concentrates on problems that correspond to identifiable modules in the human visual system. Several theories have been proposed in the literature for the computation of shape from shading, shape from texture, retinal motion from spatiotemporal derivatives of the image intensity function and the like. The problems with the existing approach are basically the following: (1) The employed assumptions are very strong and so most of the algorithms fail when applied to real images. (2) Usually the constraints from the geometry and the physics of the problem are not enough to guarantee uniqueness of the computed parameters. (3) In most cases the resulting algorithms are not robust, in the sense that if there is a slight error in the input this results in a catastrophic error in the output. In this thesis the problem of machine vision is explored from its basics. A low level mathematical theory is presented for the unique robust computation of intrinsic parameters. The computational aspect of the theory envisages a cooperative highly parallel implementation, bringing in information from five different sources (shading, texture, motion, contour and stereo), to resolve ambiguities and ensure uniqueness and stability of the intrinsic parameters. The problems of shape from texture, shape from shading and motion, visual motion analysis and shape and motion from contour are analyzed in detail.}, keywords = {*ARTIFICIAL INTELLIGENCE, *COMPUTERS, *IMAGE PROCESSING, *VISION, algorithms, CAMERAS, CATASTROPHIC CONDITIONS, COMPUTATIONS, CYBERNETICS, ERRORS, HUMANS, IMAGES, INTENSITY, LOW LEVEL, MATHEMATICS, MOTION, RETINA, Robots, SHADOWS, SHAPE, TEXTURE, THEORY, THESES.}, url = {http://stinet.dtic.mil/oai/oai?\&verb=getRecord\&metadataPrefix=html\&identifier=ADA189440}, author = {Aloimonos, J.} } @article {11984, title = {Determining the Translation of a Rigidly Moving Surface, without Correspondence,}, year = {1986}, month = {1986/01//}, abstract = {A method is presented for the recovery of the three-dimensional translation of a rigidly moving textured object from its images. The novelty of the method consists of the fact that four cameras are used in order to avoid the solution of the correspondence problem. The method seems to be immune to small noise percentages and to have good behavior when the noise increases. Keywords: Computer vision.}, keywords = {*IMAGE PROCESSING, *OPTICAL DETECTION, *Translation(Image processing), CAMERAS, Computer vision, COMPUTERS, IMMUNITY., MOTION, MOVING TARGETS, NOISE, OPTICAL DETECTION AND DETECTORS, SURFACES, THREE DIMENSIONAL, VISION}, url = {http://stinet.dtic.mil/oai/oai?\&verb=getRecord\&metadataPrefix=html\&identifier=ADA179409}, author = {Aloimonos, J. and Basu, Anup} } @article {16626, title = {The parsimonious covering model for inexact abductive reasoning in diagnostic systems}, journal = {Recent Developments in the Theory and Applications of Fuzzy Sets. Proceedings of NAFIPS}, year = {1986}, month = {1986///}, pages = {86 - 1986}, author = {Ahuja,S. B and Reggia, James A.} } @conference {11934, title = {Practicality of non-interfering checkpoints in distributed database systems}, booktitle = {IEEE Real-Time Systems Symposium, New Orleans, Louisiana}, year = {1986}, month = {1986///}, pages = {234 - 241}, author = {Son,S. H and Agrawala, Ashok K.} } @conference {11966, title = {Shape and 3-d motion from contour without point to point correspondences: General principles}, booktitle = {CVPR86}, year = {1986}, month = {1986///}, pages = {518 - 527}, author = {Aloimonos, J. and Basu, A.} } @conference {18116, title = {Tight complexity bounds for parallel comparison sorting}, booktitle = {27th Annual Symposium on Foundations of Computer Science}, year = {1986}, month = {1986///}, pages = {502 - 510}, author = {Alon,N. and Azar,Y. and Vishkin, Uzi} } @conference {16509, title = {Using abductive inferencing to derive complex error classifications for discrete sequential processes}, booktitle = {Proceedings of the 19th annual symposium on Simulation}, year = {1986}, month = {1986///}, pages = {207 - 225}, author = {Ahuja,S. B and Reggia, James A.} } @conference {12058, title = {Contour, orientation and motion}, booktitle = {Proceedings: Image Understanding Workshop (Miami Beach, FL, December 9{\textendash}10, 1985)}, year = {1985}, month = {1985///}, pages = {129 - 136}, author = {Aloimonos, J. and Basu, A. and Brown, C. M.} } @book {12004, title = {Detection of Surface Orientation and Motion from Texture: The Case of Planes}, year = {1985}, month = {1985///}, publisher = {Dept. of Computer Science, University of Rochester}, organization = {Dept. of Computer Science, University of Rochester}, author = {Aloimonos, J. and Chou, P. B and Science, University of Rochester. Dept. of Computer} } @conference {12012, title = {Direct processing of curvilinear sensor motion from a sequence of perspective images}, booktitle = {Proc. Workshop on Computer Vision: Representation and Control}, volume = {72}, year = {1984}, month = {1984///}, pages = {77 - 77}, author = {Aloimonos, J. and Brown, C. M.} } @article {18131, title = {Finding Euler tours in parallel* 1}, journal = {Journal of Computer and System Sciences}, volume = {29}, year = {1984}, month = {1984///}, pages = {330 - 337}, author = {Atallah,M. and Vishkin, Uzi} } @article {12001, title = {The relationship between optical flow and surface orientation}, journal = {Proc. of the 7-th ICPR, Montreal-Canada}, year = {1984}, month = {1984///}, author = {Aloimonos, J. and Brown, C. M.} } @conference {16589, title = {Simulation of phonemic errors using artificial intelligence symbol processing techniques}, booktitle = {Proceedings of the 17th annual symposium on Simulation}, year = {1984}, month = {1984///}, pages = {49 - 64}, author = {Reggia, James A. and Ahuja,S. B} } @article {16582, title = {SLIPS}, journal = {Journal of Medical Systems}, volume = {8}, year = {1984}, month = {1984///}, pages = {197 - 204}, author = {Morris,M. S and Reggia, James A. and Ahuja,S. B and Hart,J. and Watson,V.} } @article {18139, title = {parallel MAX-FLOW algorithm}, journal = {Journal of Algorithms}, volume = {3}, year = {1982}, month = {1982///}, pages = {128 - 146}, author = {Shiloach,Y. and Vishkin, Uzi and An,O.} } @article {19007, title = {Transcription of cloned tRNA and 5S RNA genes in a Drosophila cell free extract}, journal = {Nucleic Acids ResearchNucl. Acids Res.}, volume = {9}, year = {1981}, month = {1981/08/25/}, pages = {3907 - 3918}, abstract = {We describe the preparation of a cell-free extract from Drosophila Kc cells which allows transcription of a variety of cloned eukaryotic RNA polymerase III genes. The extract has low RNA-processing nuclease activity and thus the major products obtained are primary transcripts.}, isbn = {0305-1048, 1362-4962}, doi = {10.1093/nar/9.16.3907}, url = {http://nar.oxfordjournals.org/content/9/16/3907}, author = {Dingermann,Theodor and Sharp,Stephen and Appel,Bernd and DeFranco,Donald and Mount, Stephen M. and Heiermann,Reinhard and Pongs,Olaf and S{\"o}ll,Dieter} } @article {11935, title = {An Algorithm for Mutual Exclusion in Computer Networks.}, year = {1980}, month = {1980///}, institution = {Department of Computer Science, University of Maryland, College Park}, abstract = {An algorithm is proposed which creates mutual exclusion in a computer network whose nodes can communicate only by messages and do not share memory. The algorithm sends only 2*(N-1) Messages, where N is the number of nodes in the network, per critical section invocation. This number of messages is a minimum if parallel, distributed, symmetric control is used; hence, the algorithm is optimal in this minimal under some general assumptions. Like Lamport{\textquoteright}s {\textquoteright}bakery algorithm,{\textquoteright} unbounded sequence numbers are used to provide first-come first-served priority into the critical section. It is shown that the number can be contained in a fixed amount of memory by storing it as the residue of a modulus. The number of messages required to implement the exclusion can be reduced by using sequential node-by-node processing, by using broadcast message techniques or by sending information through timing channels. The readers and writers problem is solved by a simple modification of the algorithm. The modifications necessary to make the algorithm robust are described.}, author = {Agrawala, Ashok K. and Ricart, G.} } @article {11945, title = {Correspondence from Correspondence}, journal = {Optical Society of America, Topical Meeting on Machine Vision}, year = {1980}, month = {1980///}, pages = {46 - 51}, abstract = {Image matching (or registration), is in general terms, aligning two identical or similar images or parts of images, and it is basic to a number of vision problems. These include optical nagivation (also known as motion analysis), stereo image interpretation, object analysis, change detection and others. The difficult part of all these problems, is the matching problem, i.e., the computation of the correspondence of elements in the different views. The computation of correspondence, in both the stereo and the motion problem is impossible without the use of some assumptions about the visual world and various researchers have proposed several theories for the computation of visual correspondence that worked well for the domains that fitted their assumptions. Studied here is the problem of the computation of visual correspondence, under the assumption that some values of the correspondence function are already known. The mathematical constraints that are needed for growing a solution for the correspondence function from points where its value is known are developed.}, keywords = {ALIGNMENT, COMPUTATION, Error analysis, IMAGE PROCESSING, IMAGES, STEREOSCOPIC VISION, VECTORS (MATHEMATICS)}, url = {http://adsabs.harvard.edu/abs/1980osa..meet...46A}, author = {Aloimonos, J. and Kamgar-Parsi,Behrooz} } @article {11936, title = {The Effect of the Future in Work Distribution.}, year = {1980}, month = {1980///}, institution = {COMPUTER SCIENCE CENTER, UNIVERSITY OF MARYLAND COLLEGE PARK}, abstract = {A controller is considered which routes arrivals among several servers of different speeds. A decision which sends work to the server which will complete it soonest does not optimize the average completion time (mean flow time) because it doesn{\textquoteright}t take into account the impact of the decision on future arrivals. This impact on future arrivals, the {\textquoteright}future effect{\textquoteright}, can be significant at high arrival rates. An estimate of the size of the future effect is derived and controllers which take it into account in routing decisions can reduce the average completion time to near optimum. The effect is most pronounced when the service requirements for arrivals are nearly constant, server speeds are markedly different, and the arrival rate is close to the system{\textquoteright}s capacity. A controller considering the future effect will more heavily weigh a potential server{\textquoteright}s backlog than the arrival{\textquoteright}s service time when making a routing decision}, author = {Agrawala, Ashok K. and Ricart, G.} } @book {11937, title = {A guide to major job accounting systems: the Logger system of the UNIVAC 1100 series operating system}, year = {1978}, month = {1978///}, publisher = {U.S. Dept. of Commerce, National Bureau of Standards : for sale by the Supt. of Docs., U.S. Govt. Print. Off.}, organization = {U.S. Dept. of Commerce, National Bureau of Standards : for sale by the Supt. of Docs., U.S. Govt. Print. Off.}, keywords = {Accounting, Business \& Economics / Accounting / General, Electronic digital computers, EXEC 8 (Electronic computer system), Reference / Questions \& Answers, Univac 1100 series (Computer), Univac computer}, author = {Mohr,Jeffrey Myron and Agrawala, Ashok K. and Flannagan,J. F.} } @book {11939, title = {Machine recognition of patterns}, year = {1977}, month = {1977///}, publisher = {John Wiley \& Sons Canada, Limited}, organization = {John Wiley \& Sons Canada, Limited}, author = {Agrawala, Ashok K.} } @conference {11938, title = {The relationship between the pattern recognition problem and the workload characterization problem}, booktitle = {1977 CMG Conference}, year = {1977}, month = {1977///}, author = {Agrawala, Ashok K. and Mohr,J. M} } @conference {17463, title = {Use of peer ratings in evaluating computer program quality}, booktitle = {Proceedings of the fifteenth annual SIGCPR conference}, series = {SIGCPR {\textquoteright}77}, year = {1977}, month = {1977///}, pages = {218 - 226}, publisher = {ACM}, organization = {ACM}, address = {New York, NY, USA}, abstract = {Peer review techniques can be useful tools for supplementing programmer education, improving cooperation and communication within programming teams, and providing programmer self-evaluation. This paper will explore the benefits of peer review for practicing professional programmers and describe a framework for administration of an annual or semi-annual peer review process.}, doi = {10.1145/800100.803247}, url = {http://doi.acm.org/10.1145/800100.803247}, author = {Anderson,Nancy and Shneiderman, Ben} } @book {11940, title = {The EXEC-8 Log System: Error Analysis}, year = {1976}, month = {1976///}, publisher = {University of Maryland, Dept. of Computer Science}, organization = {University of Maryland, Dept. of Computer Science}, author = {Mohr,J. M and Agrawala, Ashok K. and Flanagan,J. F. and Science,University of Maryland. Dept. of Computer} } @book {11941, title = {Foundations of microprogramming: architecture, software, and applications}, year = {1976}, month = {1976///}, publisher = {Academic Press}, organization = {Academic Press}, author = {Agrawala, Ashok K. and Rauscher,T. G} } @book {11942, title = {Pattern Recognition}, year = {1973}, month = {1973///}, publisher = {Computer Science Center, University of Maryland}, organization = {Computer Science Center, University of Maryland}, author = {Kanal,L. N and Agrawala, Ashok K.} }