%0 Conference Paper %B 2010 Conference Record of the Forty Fourth Asilomar Conference on Signals, Systems and Computers (ASILOMAR) %D 2010 %T Design and implementation of real-time signal processing applications on heterogeneous multiprocessor arrays %A Wu, Hsiang-Huang %A Chung-Ching Shen %A Bhattacharyya, Shuvra S. %A Compton, K. %A Schulte, M. %A Wolf, M. %A Zhang, Tong %K application specific integrated circuits %K application-specific integrated circuits %K computational elements %K Computer architecture %K decoding %K Field programmable gate arrays %K field programmable X arrays %K FPGA %K FPXA %K Integrated circuit modeling %K Logic Design %K microprocessor chips %K multicore processors %K multiprocessor arrays %K real-time signal processing %K reconfigurable architectures %K reconfigurable processors %K Routing %K Signal processing %K Signal processing systems %K systolic arrays %K Viterbi algorithm %X Processing structures based on arrays of computational elements form an important class of architectures, which includes field programmable gate arrays (FPGAs), systolic arrays, and various forms of multicore processors. A wide variety of design methods and tools have been targeted to regular processing arrays involving homogeneous processing elements. In this paper, we introduce the concept of field programmable X arrays (FPXAs) as an abstract model for design and implementation of heterogeneous multiprocessor arrays for signal processing systems. FPXAs are abstract structures that can be targeted for implementation on application-specific integrated circuits, FPGAs, or other kinds of reconfigurable processors. FPXAs can also be mapped onto multicore processors for flexible emulation. We discuss the use of dataflow models as an integrated application representation and intermediate representation for efficient specification and mapping of signal processing systems on FPXAs. We demonstrate our proposed models and techniques with a case study involving the embedding of an application-specific FPXA system on an off-the-shelf FPGA device. %B 2010 Conference Record of the Forty Fourth Asilomar Conference on Signals, Systems and Computers (ASILOMAR) %P 2121 - 2125 %8 2010 %G eng %0 Conference Paper %B 2010 21st IEEE International Conference on Application-specific Systems Architectures and Processors (ASAP) %D 2010 %T Loop transformations for interface-based hierarchies IN SDF graphs %A Piat, J. %A Bhattacharyya, Shuvra S. %A Raulet, M. %K Application software %K code generation %K Computer architecture %K Computer interfaces %K Data-Flow programming %K Digital signal processing %K Loop parallelization %K PARALLEL PROCESSING %K Power engineering computing %K Power system modeling %K Processor scheduling %K Programming profession %K scheduling %K SDF graph %K system recovery %X Data-flow has proven to be an attractive computation model for programming digital signal processing (DSP) applications. A restricted version of data-flow, termed synchronous data-flow (SDF), offers strong compile-time predictability properties, but has limited expressive power. A new type of hierarchy (Interface-based SDF) has been proposed allowing more expressivity while maintaining its predictability. One of the main problems with this hierarchical SDF model is the lack of trade-off between parallelism and network clustering. This paper presents a systematic method for applying an important class of loop transformation techniques in the context of interface-based SDF semantics. The resulting approach provides novel capabilities for integrating parallelism extraction properties of the targeted loop transformations with the useful modeling, analysis, and code reuse properties provided by SDF. %B 2010 21st IEEE International Conference on Application-specific Systems Architectures and Processors (ASAP) %P 341 - 344 %8 2010 %G eng %0 Conference Paper %B 2010 21st IEEE International Symposium on Rapid System Prototyping (RSP) %D 2010 %T Rapid prototyping for digital signal processing systems using Parameterized Synchronous Dataflow graphs %A Wu, Hsiang-Huang %A Kee, Hojin %A Sane, N. %A Plishker,W. %A Bhattacharyya, Shuvra S. %K abstract scheduling %K Computational modeling %K Computer architecture %K data flow graphs %K dataflow based design %K dataflow interchange format %K design flow %K design language %K Digital signal processing %K digital signal processing systems %K dynamic parameter reconfiguration %K Dynamic scheduling %K efficient hardware mapping %K efficient quasistatic scheduling %K Embedded software %K embedded systems %K Field programmable gate arrays %K flexible dynamic reconfiguration %K FPGA based systems %K FPGA implementations %K functional simulation %K Hardware %K parameterized synchronous dataflow graphs %K rapid prototyping %K Schedules %K scheduling %K semantics %K simulation tool %K software package %K systematic design methodology %X Parameterized Synchronous Dataflow (PSDF) has been used previously for abstract scheduling and as a model for architecting embedded software and FPGA implementations. PSDF has been shown to be attractive for these purposes due to its support for flexible dynamic reconfiguration, and efficient quasi-static scheduling. To apply PSDF techniques more deeply into the design flow, support for comprehensive functional simulation and efficient hardware mapping is important. By building on the DIF (Dataflow Interchange Format), which is a design language and associated software package for developing and experimenting with dataflow-based design techniques for signal processing systems, we have developed a tool for functional simulation of PSDF specifications. This simulation tool allows designers to model applications in PSDF and simulate their functionality, including use of the dynamic parameter reconfiguration capabilities offered by PSDF. Based on this simulation tool, we also present a systematic design methodology for applying PSDF to the design and implementation of digital signal processing systems, with emphasis on FPGA-based systems for signal processing. We demonstrate capabilities for rapid and accurate prototyping offered by our proposed design methodology, along with its novel support for PSDF-based FPGA system implementation. %B 2010 21st IEEE International Symposium on Rapid System Prototyping (RSP) %P 1 - 7 %8 2010 %G eng %0 Conference Paper %B Computer Software and Applications Conference, 2006. COMPSAC '06. 30th Annual International %D 2006 %T A Software Architectural Approach to Security by Design %A Ray,A. %A Cleaveland, Rance %K architecture description notation %K Clocks %K communication semantics %K Computer architecture %K computer crime %K computer security %K Connectors %K Costs %K Degradation %K Delay %K Educational institutions %K security design %K security of data %K Software architecture %K software engineering %X This paper shows how an architecture description notation that has support for timed events can be used to provide a meta-language for specifying exact communication semantics. The advantages of such an approach is that a designer is made fully aware of the ramifications of her design choices so that an attacker can no longer take advantage of hidden assumptions %B Computer Software and Applications Conference, 2006. COMPSAC '06. 30th Annual International %I IEEE %V 2 %P 83 - 86 %8 2006/09/17/21 %@ 0-7695-2655-1 %G eng %R 10.1109/COMPSAC.2006.102 %0 Journal Article %J IEEE Transactions on Circuits and Systems for Video Technology %D 2005 %T Class-based access control for distributed video-on-demand systems %A Mundur, Padma %A Sood,A. K %A Simon,R. %K Access control %K Admission control %K Analytical models %K blocking performance %K class-based access control %K Computational modeling %K Computer architecture %K Computer science %K Distributed control %K Distributed video-on-demand (VoD) system %K distributed video-on-demand system %K multimedia systems %K multirate service model %K Performance analysis %K QoS %K quality of service %K request handling policy %K resource allocation %K resource capacity %K telecommunication congestion control %K threshold-based admission control %K video on demand %X The focus of this paper is the analysis of threshold-based admission control policies for distributed video-on-demand (VoD) systems. Traditionally, admission control methods control access to a resource based on the resource capacity. We have extended that concept to include the significance of an arriving request to the VoD system by enforcing additional threshold restrictions in the admission control process on request classes deemed less significant. We present an analytical model for computing blocking performance of the VoD system under threshold-based admission control. Extending the same methodology to a distributed VoD architecture we show through simulation that the threshold performance conforms to the analytical model. We also show that threshold-based analysis can work in conjunction with other request handling policies and are useful for manipulating the VoD performance since we are able to distinguish between different request classes based on their merit. Enforcing threshold restrictions with the option of downgrading blocked requests in a multirate service environment results in improved performance at the same time providing different levels of quality of service (QoS). In fact, we show that the downgrade option combined with threshold restrictions is a powerful tool for manipulating an incoming request mix over which we have no control into a workload that the VoD system can handle. %B IEEE Transactions on Circuits and Systems for Video Technology %V 15 %P 844 - 853 %8 2005/07// %@ 1051-8215 %G eng %N 7 %R 10.1109/TCSVT.2005.848351 %0 Report %D 2005 %T Inferencing in Support of Active Templates %A Nau, Dana S. %K *ARTILLERY AMMUNITION %K *HIGH CAPACITY PROJECTILES %K AMMUNITION AND EXPLOSIVES %K BAYES THEOREM %K COALITION %K Collaboration %K Computer architecture %K COMPUTER HARDWARE %K Computer networks %K HICAP(HIGH CAPACITY ARTILLARY PROJECTILES) %K MILITARY COMMANDERS %K NEO(NONCOMBATANT EVACUATION OPERATIONS) %K NONCOMBATANT. %K PE63760E %K planning %K probability %K PROJECTILE TRAJECTORIES %K TASK SUPPORT %K WUAFRLATEMPO02 %X The primary accomplishments of this project include the following: (1) HICAP is a general purpose planning architecture to assist military commanders with planning NEOs (Noncombatant Evacuation Operations). HICAP integrates a hierarchical task editor with a planning tool. Military planners select a task to decompose in the task editor and then use the planning tool to interactively refine it into an operational plan. (2) SHOP and SHOP2 are some simple, practical planning tools based on HTN planning. Their practical utility is shown by the mergence of an active set of users, which include government laboratories, industrial R&D projects, and academic settings. SHOP2 received one of the top four awards in the 2002 International Planning Competition. (3) The Air Force Research Laboratory's Causal Analysis Tool (CAT) is a system for creating and analyzing causal models similar to Bayesian networks. We have enhanced CAT by developing an approach to quickly generate plans that have high probabilities of success. %I Department of Computer Science, University of Maryland, College Park %8 2005/02// %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA431023 %0 Conference Paper %B 14th IEEE Proceedings on Personal, Indoor and Mobile Radio Communications, 2003. PIMRC 2003 %D 2003 %T A secure service discovery protocol for MANET %A Yuan Yuan %A Arbaugh, William A. %K ad hoc networks %K centralized administration %K Computer architecture %K Computer science %K dynamic service discovery infrastructure %K Educational institutions %K MANET %K Manuals %K mobile ad hoc network %K Mobile ad hoc networks %K Mobile computing %K mobile radio %K noninfrastructure network %K Pervasive computing %K Protocols %K routing protocols %K secure service discovery protocol %K Security %K service discovery techniques %K service discovery technologies %K telecommunication computing %K telecommunication services %K XML %X Service discovery technologies are exploited to enable services to advertise their existence in a dynamic way, and can be discovered, configured and used by other devices with minimum manual efforts. It plays an essential role in future network scenarios especially with development of mobile ad hoc network (MANET) and emergence of pervasive computing. Because MANET allows these devices to communicate dynamically without fixed infrastructure and centralized administration, it gives rise to the challenges of the service discovery techniques. In this paper, we present a dynamic service discovery infrastructure that uses XML to describe services and match using the semantic content of service descriptions for MANET. We believe that the architecture we have designed is a necessary component of service discovery in non-infrastructure network by further exploring the secure and performance issues of this infrastructure. %B 14th IEEE Proceedings on Personal, Indoor and Mobile Radio Communications, 2003. PIMRC 2003 %I IEEE %V 1 %P 502- 506 Vol.1 - 502- 506 Vol.1 %8 2003/09/07/10 %@ 0-7803-7822-9 %G eng %R 10.1109/PIMRC.2003.1264322 %0 Journal Article %J IEEE Transactions on Information Technology in Biomedicine %D 2003 %T The virtual microscope %A Catalyurek,U. %A Beynon,M. D %A Chang,Chialin %A Kurc, T. %A Sussman, Alan %A Saltz, J. %K biomedical optical imaging %K client software %K client/server architecture %K Computer architecture %K Computer Graphics %K computer platforms %K Computer simulation %K Concurrent computing %K configured data server %K data server software %K database management systems %K database software %K Database systems %K digital slide images %K digital telepathology %K diseases %K emulation %K Environment %K Equipment Design %K Equipment Failure Analysis %K high power light microscope %K Image databases %K Image Enhancement %K Image Interpretation, Computer-Assisted %K Image retrieval %K Information retrieval %K Information Storage and Retrieval %K java %K local disks %K microscope image data %K Microscopy %K multiple clients %K optical microscopy %K PACS %K software %K Software design %K software system %K Software systems %K Systems Integration %K Telepathology %K User-Computer Interface %K virtual microscope design %K Virtual reality %K Workstations %X We present the design and implementation of the virtual microscope, a software system employing a client/server architecture to provide a realistic emulation of a high power light microscope. The system provides a form of completely digital telepathology, allowing simultaneous access to archived digital slide images by multiple clients. The main problem the system targets is storing and processing the extremely large quantities of data required to represent a collection of slides. The virtual microscope client software runs on the end user's PC or workstation, while database software for storing, retrieving and processing the microscope image data runs on a parallel computer or on a set of workstations at one or more potentially remote sites. We have designed and implemented two versions of the data server software. One implementation is a customization of a database system framework that is optimized for a tightly coupled parallel machine with attached local disks. The second implementation is component-based, and has been designed to accommodate access to and processing of data in a distributed, heterogeneous environment. We also have developed caching client software, implemented in Java, to achieve good response time and portability across different computer platforms. The performance results presented show that the Virtual Microscope systems scales well, so that many clients can be adequately serviced by an appropriately configured data server. %B IEEE Transactions on Information Technology in Biomedicine %V 7 %P 230 - 248 %8 2003/12// %@ 1089-7771 %G eng %N 4 %R 10.1109/TITB.2004.823952 %0 Conference Paper %B Geoscience and Remote Sensing Symposium, 1999. IGARSS '99 Proceedings. IEEE 1999 International %D 1999 %T Developing the next generation of Earth science data systems: the Global Land Cover Facility %A Lindsay,F.E. %A Townshend,J.R.G. %A JaJa, Joseph F. %A Humphries,J. %A Plaisant, Catherine %A Shneiderman, Ben %K Computer architecture %K data archiving %K data distribution system %K Data systems %K Distributed computing %K Earth science data products %K Earth science data system %K ESIP %K geographic information system %K geographic information systems %K Geography %K geophysical measurement technique %K geophysical signal processing %K geophysical techniques %K Geoscience %K GIS %K GLCF %K Global Land Cover Facility %K High performance computing %K Indexing %K information service %K Information services %K Institute for Advanced Computer Studies %K land cover %K NASA %K next generation %K PACS %K Remote sensing %K terrain mapping %K UMIACS %K University of Maryland %K User interfaces %K web-based interface %X A recent initiative by NASA has resulted in the formation of a federation of Earth science data partners. These Earth Science Information Partners (ESIPs) have been tasked with creating novel Earth science data products and services as well as distributing new and existing data sets to the Earth science community and the general public. The University of Maryland established its ESIP activities with the creation of the Global Land Cover Facility (GLCF). This joint effort of the Institute for Advanced Computer Studies (UMIACS) and the Department of Geography has developed an operational data archiving and distribution system aimed at advancing current land cover research efforts. The success of the GLCF is tied closely to assessing user needs as well. As the timely delivery of data products to the research community. This paper discusses the development and implementation of a web-based interface that allows users to query the authors' data holdings and perform user requested processing tasks on demand. The GLCF takes advantage of a scaleable, high performance computing architecture for the manipulation of very large remote sensing data sets and the rapid spatial indexing of multiple format data types. The user interface has been developed with the cooperation of the Human-Computer Interaction Laboratory (HCIL) and demonstrates advances in spatial and temporal querying tools as well as the ability to overlay multiple raster and vector data sets. Their work provides one perspective concerning how critical earth science data may be handled in the near future by a coalition of distributed data centers %B Geoscience and Remote Sensing Symposium, 1999. IGARSS '99 Proceedings. IEEE 1999 International %I IEEE %V 1 %P 616-618 vol.1 - 616-618 vol.1 %8 1999/// %@ 0-7803-5207-6 %G eng %R 10.1109/IGARSS.1999.773583 %0 Conference Paper %B IEEE INFOCOM '99. Eighteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %D 1999 %T PLANet: an active internetwork %A Hicks, Michael W. %A Moore,J. T %A Alexander,D. S %A Gunter,C. A %A Nettles,S. M %K 100 Mbit/s %K 300 MHz %K 48 Mbit/s %K active internetwork %K active network architecture %K active network implementation %K byte-code-interpreted applications %K Computer architecture %K Computer languages %K Computer networks %K congested conditions %K dynamic programming %K dynamic router extensions %K Ethernet %K Ethernet networks %K INFORMATION SCIENCE %K Internet %K Internet-like services %K internetworking %K IP %K IP networks %K link layers %K Linux user-space applications %K Local area networks %K ML dialect %K Network performance %K networking operations %K OCaml %K Packet Language for Active Networks %K packet programs %K packet switching %K Pentium-II %K performance %K performance evaluation %K PLAN %K PLANet %K Planets %K programmability features %K programming languages %K router functionality %K special purpose programming language %K Switches %K telecommunication network routing %K Transport protocols %K Web and internet services %X We present PLANet: an active network architecture and implementation. In addition to a standard suite of Internet-like services, PLANet has two key programmability features: (1) all packets contain programs; and (2) router functionality may be extended dynamically. Packet programs are written in our special purpose programming language PLAN, the Packet Language for Active Networks, while dynamic router extensions are written in OCaml, a dialect of ML. Currently, PLANet routers run as byte-code-interpreted Linux user-space applications, and support Ethernet and IP as link layers. PLANet achieves respectable performance on standard networking operations: on 300 MHz Pentium-II's attached to 100 Mbps Ethernet, PLANet can route 48 Mbps and switch over 5000 packets per second. We demonstrate the utility of PLANet's activeness by showing experimentally how it can nontrivially improve application and aggregate network performance in congested conditions %B IEEE INFOCOM '99. Eighteenth Annual Joint Conference of the IEEE Computer and Communications Societies. Proceedings %I IEEE %V 3 %P 1124-1133 vol.3 - 1124-1133 vol.3 %8 1999/03/21/25 %@ 0-7803-5417-6 %G eng %R 10.1109/INFCOM.1999.751668 %0 Journal Article %J IEEE Transactions on Knowledge and Data Engineering %D 1998 %T Techniques for update handling in the enhanced client-server DBMS %A Delis,A. %A Roussopoulos, Nick %K client disk managers %K client resources %K client-server computing paradigm %K client-server systems %K Computational modeling %K Computer architecture %K concurrency control %K data pages %K Database systems %K distributed databases %K enhanced client-server DBMS %K Hardware %K Local area networks %K long-term memory %K main-memory caches %K Network servers %K operational spaces %K Personal communication networks %K server update propagation techniques %K Transaction databases %K update handling %K Workstations %K Yarn %X The Client-Server computing paradigm has significantly influenced the way modern Database Management Systems are designed and built. In such systems, clients maintain data pages in their main-memory caches, originating from the server's database. The Enhanced Client-Server architecture takes advantage of all the available client resources, including their long-term memory. Clients can cache server data into their own disk units if these data are part of their operational spaces. However, when updates occur at the server, a number of clients may need to not only be notified about these changes, but also obtain portions of the updates as well. In this paper, we examine the problem of managing server imposed updates that affect data cached on client disk managers. We propose a number of server update propagation techniques in the context of the Enhanced Client-Server DBMS architecture, and examine the performance of these strategies through detailed simulation experiments. In addition, we study how the various settings of the network affect the performance of these policies %B IEEE Transactions on Knowledge and Data Engineering %V 10 %P 458 - 476 %8 1998/06//May %@ 1041-4347 %G eng %N 3 %R 10.1109/69.687978 %0 Conference Paper %B Performance, Computing and Communications, 1998. IPCCC '98., IEEE International %D 1998 %T Threshold-based admission control for multi-class video-on-demand systems %A Mundur, Padma %A Sood,A. %A Simon,R. %K Admission control %K admission control tests %K analytical model %K Analytical models %K blocking probabilities %K buffer storage %K Computer architecture %K distributed digital VoD systems %K File servers %K guaranteed service %K High-speed networks %K interactive video %K Motion pictures %K multi-class video-on-demand systems %K multimedia systems %K Network servers %K newly arriving requests %K numerical analysis %K Predictive models %K probability %K recursive algorithm %K system blocking rates %K telecommunication congestion control %K threshold-based admission control %K unified admission control %X The next generation of distributed digital video-on-demand (VoD) systems will use admission control tests to ensure that users receive predictable and guaranteed service. If the system cannot support a new request then that request is blocked This paper presents an analytical model to evaluate unified admission control strategies for distributed VoD systems with multiple video classes. We prove that there exists a computationally efficient technique to determine the blocking probabilities for newly arriving requests. Through numerical analysis we show the effect of different admission control policies on overall system blocking rates %B Performance, Computing and Communications, 1998. IPCCC '98., IEEE International %I IEEE %P 154 - 160 %8 1998/02/16/18 %@ 0-7803-4468-5 %G eng %R 10.1109/PCCC.1998.659937 %0 Conference Paper %B , 1997 IEEE Symposium on Security and Privacy, 1997. Proceedings %D 1997 %T A secure and reliable bootstrap architecture %A Arbaugh, William A. %A Farber,D. J %A Smith,J. M %K active networks %K AEGIS architecture %K bootstrap architecture %K Computer architecture %K computer bootstrapping %K data integrity %K Distributed computing %K Hardware %K hardware validity %K initialization %K integrity chain %K integrity check failures %K Internet %K Internet commerce %K IP networks %K Laboratories %K lower-layer integrity %K Microprogramming %K Operating systems %K recovery process %K reliability %K robust systems %K Robustness %K Security %K security of data %K software reliability %K system integrity guarantees %K system recovery %K transitions %K Virtual machining %X In a computer system, the integrity of lower layers is typically treated as axiomatic by higher layers. Under the presumption that the hardware comprising the machine (the lowest layer) is valid, the integrity of a layer can be guaranteed if and only if: (1) the integrity of the lower layers is checked and (2) transitions to higher layers occur only after integrity checks on them are complete. The resulting integrity “chain” inductively guarantees system integrity. When these conditions are not met, as they typically are not in the bootstrapping (initialization) of a computer system, no integrity guarantees can be made, yet these guarantees are increasingly important to diverse applications such as Internet commerce, security systems and “active networks”. In this paper, we describe the AEGIS architecture for initializing a computer system. It validates integrity at each layer transition in the bootstrap process. AEGIS also includes a recovery process for integrity check failures, and we show how this results in robust systems %B , 1997 IEEE Symposium on Security and Privacy, 1997. Proceedings %I IEEE %P 65 - 71 %8 1997/05/04/7 %@ 0-8186-7828-3 %G eng %R 10.1109/SECPRI.1997.601317 %0 Report %D 1996 %T MARUTI at ARDEC. %A Agrawala, Ashok K. %A Tripathi,Satish K %K *OPERATING SYSTEMS(COMPUTERS) %K Computer architecture %K COMPUTER PROGRAMMING AND SOFTWARE %K CONTROL %K Cooperation %K DSSA(DOMAIN SPECIFIC SOFTWARE ARCHITECTURE) %K GUN TURRETS. %K HIGH FREQUENCY %K Laboratories %K REAL TIME %K ROBOTICS %K TEST BEDS %X This is the final report of the effort undertaken at the University of Maryland, sponsored by DARPA under the DSSA (Domain Specific Software Architectures) program. The activities reported here required an active cooperation and collaboration of the US Army ARDEC (Automation and Robotics Laboratory) and the University of Maryland (Computer Science Department). The principal investigators have been developing a hard real time operating system, MARUTI, at the University of Maryland. The goal of this effort was to demonstrate the applicability and usefulness of this operating system for exercising PID control at high frequency on the ATB1000 testbed which has been used to simulate a gun turret. A MARUTI based PID controller which operates at 200 HZ and 400 HZ was successfully implemented to demonstrate this feasibility. %I University of Maryland, College Park %8 1996/07/11/ %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA315086 %0 Conference Paper %B Scalable High-Performance Computing Conference, 1994., Proceedings of the %D 1994 %T Dynamic program instrumentation for scalable performance tools %A Hollingsworth, Jeffrey K %A Miller, B. P %A Cargille, J. %K Application software %K binary image %K compiler writing %K Computer architecture %K Computer displays %K Computerized monitoring %K Concurrent computing %K data acquisition %K data collection %K data visualisation %K Data visualization %K dynamic program instrumentation %K efficient monitoring %K executing program %K Instruments %K large-scale parallel applications %K Large-scale systems %K operating system design %K Operating systems %K parallel programming %K program analysis %K program diagnostics %K program visualization %K Programming profession %K Sampling methods %K scalable performance tools %K software tools %X Presents a new technique called `dynamic instrumentation' that provides efficient, scalable, yet detailed data collection for large-scale parallel applications. Our approach is unique because it defers inserting any instrumentation until the application is in execution. We can insert or change instrumentation at any time during execution by modifying the application's binary image. Only the instrumentation required for the currently selected analysis or visualization is inserted. As a result, our technique collects several orders of magnitude less data than traditional data collection approaches. We have implemented a prototype of our dynamic instrumentation on the CM-5, and present results for several real applications. In addition, we include recommendations to operating system designers, compiler writers, and computer architects about the features necessary to permit efficient monitoring of large-scale parallel systems %B Scalable High-Performance Computing Conference, 1994., Proceedings of the %I IEEE %P 841 - 850 %8 1994/05// %@ 0-8186-5680-8 %G eng %R 10.1109/SHPCC.1994.296728 %0 Journal Article %J IEEE Transactions on Software Engineering %D 1993 %T Performance comparison of three modern DBMS architectures %A Delis,A. %A Roussopoulos, Nick %K client-server %K Computational modeling %K Computer architecture %K database management systems %K DBMS architectures %K design rationales %K functional components %K Indexes %K Local area networks %K Military computing %K Packaging %K Performance analysis %K performance evaluation %K RAD-UNIFY type %K simulation models %K simulation results %K Software architecture %K software architecture configurations %K software engineering %K Throughput %K Workstations %X The introduction of powerful workstations connected through local area networks (LANs) inspired new database management system (DBMS) architectures that offer high performance characteristics. The authors examine three such software architecture configurations: client-server (CS), the RAD-UNIFY type of DBMS (RU), and enhanced client-server (ECS). Their specific functional components and design rationales are discussed. Three simulation models are used to provide a performance comparison under different job workloads. Simulation results show that the RU almost always performs slightly better than the CS, especially under light workloads, and that ECS offers significant performance improvement over both CS and RU. Under reasonable update rates, the ECS over CS (or RU) performance ratio is almost proportional to the number of participating clients (for less than 32 clients). The authors also examine the impact of certain key parameters on the performance of the three architectures and show that ECS is more scalable that the other two %B IEEE Transactions on Software Engineering %V 19 %P 120 - 138 %8 1993/02// %@ 0098-5589 %G eng %N 2 %R 10.1109/32.214830 %0 Report %D 1988 %T Optimal Systolic Designs for the Computation of the Discrete Hartley and the Discrete Cosine Transforms %A Chakrabarti,Chaitali %A JaJa, Joseph F. %K *ALGORITHMS %K *BINARY ARITHMETIC %K Computer architecture %K DCT(DISCRETE COSINE TRANSFORM) %K DISCRETE FOURIER TRANSFORMS %K ITERATIONS %K NUMERICAL MATHEMATICS %K TWO DIMENSIONAL %X In this paper, we propose new algorithms for computing the Discrete Hartley and the Discrete Cosine Transform. The algorithms are based on iterative applications of the modified small n algorithms of DFT. The one dimensional transforms are mapped into two dimensions first and then implemented on two dimensional systolic arrays. Pipelined bit serial architectures operating on left to right LSB to MSB binary arithmetic is the basis of the hardware design. Different hardware schemes for implementing these transforms are studied. We show that our schemes achieve a substantial speed-up over existing schemes. %I Institute for Systems Research, University of Maryland, College Park %8 1988/// %G eng %U http://stinet.dtic.mil/oai/oai?&verb=getRecord&metadataPrefix=html&identifier=ADA452389 %0 Journal Article %J IEEE Transactions on Software Engineering %D 1985 %T SEES—A Software testing Environment Support System %A Roussopoulos, Nick %A Yeh,R. T %K Computer architecture %K Database systems %K Error correction %K Program processors %K Programming profession %K Relational databases %K Software testing %K software tools %K Workstations %K Writing %X SEES is a database system to support program testing. The program database is automatically created during the compilation of the program by a compiler built using the YACC compiler-compiler. %B IEEE Transactions on Software Engineering %V SE-11 %P 355 - 366 %8 1985/04// %@ 0098-5589 %G eng %N 4 %R 10.1109/TSE.1985.232225