WorldWideScience

Sample records for cluster file systems

  1. High-Performance, Multi-Node File Copies and Checksums for Clustered File Systems

    Science.gov (United States)

    Kolano, Paul Z.; Ciotti, Robert B.

    2012-01-01

    Modern parallel file systems achieve high performance using a variety of techniques, such as striping files across multiple disks to increase aggregate I/O bandwidth and spreading disks across multiple servers to increase aggregate interconnect bandwidth. To achieve peak performance from such systems, it is typically necessary to utilize multiple concurrent readers/writers from multiple systems to overcome various singlesystem limitations, such as number of processors and network bandwidth. The standard cp and md5sum tools of GNU coreutils found on every modern Unix/Linux system, however, utilize a single execution thread on a single CPU core of a single system, and hence cannot take full advantage of the increased performance of clustered file systems. Mcp and msum are drop-in replacements for the standard cp and md5sum programs that utilize multiple types of parallelism and other optimizations to achieve maximum copy and checksum performance on clustered file systems. Multi-threading is used to ensure that nodes are kept as busy as possible. Read/write parallelism allows individual operations of a single copy to be overlapped using asynchronous I/O. Multinode cooperation allows different nodes to take part in the same copy/checksum. Split-file processing allows multiple threads to operate concurrently on the same file. Finally, hash trees allow inherently serial checksums to be performed in parallel. Mcp and msum provide significant performance improvements over standard cp and md5sum using multiple types of parallelism and other optimizations. The total speed-ups from all improvements are significant. Mcp improves cp performance over 27x, msum improves md5sum performance almost 19x, and the combination of mcp and msum improves verified copies via cp and md5sum by almost 22x. These improvements come in the form of drop-in replacements for cp and md5sum, so are easily used and are available for download as open source software at http://mutil.sourceforge.net.

  2. Optimizing Instruction Scheduling and Register Allocation for Register-File-Connected Clustered VLIW Architectures

    Science.gov (United States)

    Tang, Haijing; Wang, Siye; Zhang, Yanjun

    2013-01-01

    Clustering has become a common trend in very long instruction words (VLIW) architecture to solve the problem of area, energy consumption, and design complexity. Register-file-connected clustered (RFCC) VLIW architecture uses the mechanism of global register file to accomplish the inter-cluster data communications, thus eliminating the performance and energy consumption penalty caused by explicit inter-cluster data move operations in traditional bus-connected clustered (BCC) VLIW architecture. However, the limit number of access ports to the global register file has become an issue which must be well addressed; otherwise the performance and energy consumption would be harmed. In this paper, we presented compiler optimization techniques for an RFCC VLIW architecture called Lily, which is designed for encryption systems. These techniques aim at optimizing performance and energy consumption for Lily architecture, through appropriate manipulation of the code generation process to maintain a better management of the accesses to the global register file. All the techniques have been implemented and evaluated. The result shows that our techniques can significantly reduce the penalty of performance and energy consumption due to access port limitation of global register file. PMID:23970841

  3. Optimizing Instruction Scheduling and Register Allocation for Register-File-Connected Clustered VLIW Architectures

    Directory of Open Access Journals (Sweden)

    Haijing Tang

    2013-01-01

    Full Text Available Clustering has become a common trend in very long instruction words (VLIW architecture to solve the problem of area, energy consumption, and design complexity. Register-file-connected clustered (RFCC VLIW architecture uses the mechanism of global register file to accomplish the inter-cluster data communications, thus eliminating the performance and energy consumption penalty caused by explicit inter-cluster data move operations in traditional bus-connected clustered (BCC VLIW architecture. However, the limit number of access ports to the global register file has become an issue which must be well addressed; otherwise the performance and energy consumption would be harmed. In this paper, we presented compiler optimization techniques for an RFCC VLIW architecture called Lily, which is designed for encryption systems. These techniques aim at optimizing performance and energy consumption for Lily architecture, through appropriate manipulation of the code generation process to maintain a better management of the accesses to the global register file. All the techniques have been implemented and evaluated. The result shows that our techniques can significantly reduce the penalty of performance and energy consumption due to access port limitation of global register file.

  4. Distributed metadata servers for cluster file systems using shared low latency persistent key-value metadata store

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Pedone, Jr., James M.; Tzelnic, Percy; Ting, Dennis P. J.; Ionkov, Latchesar A.; Grider, Gary

    2017-12-26

    A cluster file system is provided having a plurality of distributed metadata servers with shared access to one or more shared low latency persistent key-value metadata stores. A metadata server comprises an abstract storage interface comprising a software interface module that communicates with at least one shared persistent key-value metadata store providing a key-value interface for persistent storage of key-value metadata. The software interface module provides the key-value metadata to the at least one shared persistent key-value metadata store in a key-value format. The shared persistent key-value metadata store is accessed by a plurality of metadata servers. A metadata request can be processed by a given metadata server independently of other metadata servers in the cluster file system. A distributed metadata storage environment is also disclosed that comprises a plurality of metadata servers having an abstract storage interface to at least one shared persistent key-value metadata store.

  5. The Global File System

    Science.gov (United States)

    Soltis, Steven R.; Ruwart, Thomas M.; OKeefe, Matthew T.

    1996-01-01

    The global file system (GFS) is a prototype design for a distributed file system in which cluster nodes physically share storage devices connected via a network-like fiber channel. Networks and network-attached storage devices have advanced to a level of performance and extensibility so that the previous disadvantages of shared disk architectures are no longer valid. This shared storage architecture attempts to exploit the sophistication of storage device technologies whereas a server architecture diminishes a device's role to that of a simple component. GFS distributes the file system responsibilities across processing nodes, storage across the devices, and file system resources across the entire storage pool. GFS caches data on the storage devices instead of the main memories of the machines. Consistency is established by using a locking mechanism maintained by the storage devices to facilitate atomic read-modify-write operations. The locking mechanism is being prototyped in the Silicon Graphics IRIX operating system and is accessed using standard Unix commands and modules.

  6. Building a parallel file system simulator

    International Nuclear Information System (INIS)

    Molina-Estolano, E; Maltzahn, C; Brandt, S A; Bent, J

    2009-01-01

    Parallel file systems are gaining in popularity in high-end computing centers as well as commercial data centers. High-end computing systems are expected to scale exponentially and to pose new challenges to their storage scalability in terms of cost and power. To address these challenges scientists and file system designers will need a thorough understanding of the design space of parallel file systems. Yet there exist few systematic studies of parallel file system behavior at petabyte- and exabyte scale. An important reason is the significant cost of getting access to large-scale hardware to test parallel file systems. To contribute to this understanding we are building a parallel file system simulator that can simulate parallel file systems at very large scale. Our goal is to simulate petabyte-scale parallel file systems on a small cluster or even a single machine in reasonable time and fidelity. With this simulator, file system experts will be able to tune existing file systems for specific workloads, scientists and file system deployment engineers will be able to better communicate workload requirements, file system designers and researchers will be able to try out design alternatives and innovations at scale, and instructors will be able to study very large-scale parallel file system behavior in the class room. In this paper we describe our approach and provide preliminary results that are encouraging both in terms of fidelity and simulation scalability.

  7. Clustering of users of digital libraries through log file analysis

    Directory of Open Access Journals (Sweden)

    Juan Antonio Martínez-Comeche

    2017-09-01

    Full Text Available This study analyzes how users perform information retrieval tasks when introducing queries to the Hispanic Digital Library. Clusters of users are differentiated based on their distinct information behavior. The study used the log files collected by the server over a year and different possible clustering algorithms are compared. The k-means algorithm is found to be a suitable clustering method for the analysis of large log files from digital libraries. In the case of the Hispanic Digital Library the results show three clusters of users and the characteristic information behavior of each group is described.

  8. pcircle - A Suite of Scalable Parallel File System Tools

    Energy Technology Data Exchange (ETDEWEB)

    2015-10-01

    Most of the software related to file system are written for conventional local file system, they are serialized and can't take advantage of the benefit of a large scale parallel file system. "pcircle" software builds on top of ubiquitous MPI in cluster computing environment and "work-stealing" pattern to provide a scalable, high-performance suite of file system tools. In particular - it implemented parallel data copy and parallel data checksumming, with advanced features such as async progress report, checkpoint and restart, as well as integrity checking.

  9. Cut-and-Paste file-systems : integrating simulators and file systems

    NARCIS (Netherlands)

    Bosch, H.G.P.; Mullender, Sape J.

    1996-01-01

    We have implemented an integrated and configurable file system called the PFS and a trace-driven file-system simulator called Patsy. Patsy is used for off-line analysis of file-system algorithms, PFS is used for on-line file-system data storage. Algorithms are first analyzed in Patsy and when we are

  10. Cut-and-Paste file-systems: integrating simulators and file systems

    NARCIS (Netherlands)

    Bosch, H.G.P.; Mullender, Sape J.

    1995-01-01

    We have implemented an integrated and configurable file system called the Pegasus filesystem (PFS) and a trace-driven file-system simulator called Patsy. Patsy is used for off-line analysis of file-systemalgorithms, PFS is used for on-line file-systemdata storage. Algorithms are first analyzed in

  11. Adding Data Management Services to Parallel File Systems

    Energy Technology Data Exchange (ETDEWEB)

    Brandt, Scott [Univ. of California, Santa Cruz, CA (United States)

    2015-03-04

    The objective of this project, called DAMASC for “Data Management in Scientific Computing”, is to coalesce data management with parallel file system management to present a declarative interface to scientists for managing, querying, and analyzing extremely large data sets efficiently and predictably. Managing extremely large data sets is a key challenge of exascale computing. The overhead, energy, and cost of moving massive volumes of data demand designs where computation is close to storage. In current architectures, compute/analysis clusters access data in a physically separate parallel file system and largely leave it scientist to reduce data movement. Over the past decades the high-end computing community has adopted middleware with multiple layers of abstractions and specialized file formats such as NetCDF-4 and HDF5. These abstractions provide a limited set of high-level data processing functions, but have inherent functionality and performance limitations: middleware that provides access to the highly structured contents of scientific data files stored in the (unstructured) file systems can only optimize to the extent that file system interfaces permit; the highly structured formats of these files often impedes native file system performance optimizations. We are developing Damasc, an enhanced high-performance file system with native rich data management services. Damasc will enable efficient queries and updates over files stored in their native byte-stream format while retaining the inherent performance of file system data storage via declarative queries and updates over views of underlying files. Damasc has four key benefits for the development of data-intensive scientific code: (1) applications can use important data-management services, such as declarative queries, views, and provenance tracking, that are currently available only within database systems; (2) the use of these services becomes easier, as they are provided within a familiar file

  12. Experiences on File Systems: Which is the best file system for you?

    CERN Document Server

    Blomer, J

    2015-01-01

    The distributed file system landscape is scattered. Besides a plethora of research file systems, there is also a large number of production grade file systems with various strengths and weaknesses. The file system, as an abstraction of permanent storage, is appealing because it provides application portability and integration with legacy and third-party applications, including UNIX utilities. On the other hand, the general and simple file system interface makes it notoriously difficult for a distributed file system to perform well under a variety of different workloads. This contribution provides a taxonomy of commonly used distributed file systems and points out areas of research and development that are particularly important for high-energy physics.

  13. Accessing files in an Internet: The Jade file system

    Science.gov (United States)

    Peterson, Larry L.; Rao, Herman C.

    1991-01-01

    Jade is a new distribution file system that provides a uniform way to name and access files in an internet environment. It makes two important contributions. First, Jade is a logical system that integrates a heterogeneous collection of existing file systems, where heterogeneous means that the underlying file systems support different file access protocols. Jade is designed under the restriction that the underlying file system may not be modified. Second, rather than providing a global name space, Jade permits each user to define a private name space. These private name spaces support two novel features: they allow multiple file systems to be mounted under one directory, and they allow one logical name space to mount other logical name spaces. A prototype of the Jade File System was implemented on Sun Workstations running Unix. It consists of interfaces to the Unix file system, the Sun Network File System, the Andrew File System, and FTP. This paper motivates Jade's design, highlights several aspects of its implementation, and illustrates applications that can take advantage of its features.

  14. Accessing files in an internet - The Jade file system

    Science.gov (United States)

    Rao, Herman C.; Peterson, Larry L.

    1993-01-01

    Jade is a new distribution file system that provides a uniform way to name and access files in an internet environment. It makes two important contributions. First, Jade is a logical system that integrates a heterogeneous collection of existing file systems, where heterogeneous means that the underlying file systems support different file access protocols. Jade is designed under the restriction that the underlying file system may not be modified. Second, rather than providing a global name space, Jade permits each user to define a private name space. These private name spaces support two novel features: they allow multiple file systems to be mounted under one directory, and they allow one logical name space to mount other logical name spaces. A prototype of the Jade File System was implemented on Sun Workstations running Unix. It consists of interfaces to the Unix file system, the Sun Network File System, the Andrew File System, and FTP. This paper motivates Jade's design, highlights several aspects of its implementation, and illustrates applications that can take advantage of its features.

  15. Implementing Journaling in a Linux Shared Disk File System

    Science.gov (United States)

    Preslan, Kenneth W.; Barry, Andrew; Brassow, Jonathan; Cattelan, Russell; Manthei, Adam; Nygaard, Erling; VanOort, Seth; Teigland, David; Tilstra, Mike; O'Keefe, Matthew; hide

    2000-01-01

    In computer systems today, speed and responsiveness is often determined by network and storage subsystem performance. Faster, more scalable networking interfaces like Fibre Channel and Gigabit Ethernet provide the scaffolding from which higher performance computer systems implementations may be constructed, but new thinking is required about how machines interact with network-enabled storage devices. In this paper we describe how we implemented journaling in the Global File System (GFS), a shared-disk, cluster file system for Linux. Our previous three papers on GFS at the Mass Storage Symposium discussed our first three GFS implementations, their performance, and the lessons learned. Our fourth paper describes, appropriately enough, the evolution of GFS version 3 to version 4, which supports journaling and recovery from client failures. In addition, GFS scalability tests extending to 8 machines accessing 8 4-disk enclosures were conducted: these tests showed good scaling. We describe the GFS cluster infrastructure, which is necessary for proper recovery from machine and disk failures in a collection of machines sharing disks using GFS. Finally, we discuss the suitability of Linux for handling the big data requirements of supercomputing centers.

  16. Extending the POSIX I/O interface: a parallel file system perspective.

    Energy Technology Data Exchange (ETDEWEB)

    Vilayannur, M.; Lang, S.; Ross, R.; Klundt, R.; Ward, L.; Mathematics and Computer Science; VMWare, Inc.; SNL

    2008-12-11

    The POSIX interface does not lend itself well to enabling good performance for high-end applications. Extensions are needed in the POSIX I/O interface so that high-concurrency HPC applications running on top of parallel file systems perform well. This paper presents the rationale, design, and evaluation of a reference implementation of a subset of the POSIX I/O interfaces on a widely used parallel file system (PVFS) on clusters. Experimental results on a set of micro-benchmarks confirm that the extensions to the POSIX interface greatly improve scalability and performance.

  17. The global unified parallel file system (GUPFS) project: FY 2002 activities and results

    Energy Technology Data Exchange (ETDEWEB)

    Butler, Gregory F.; Lee, Rei Chi; Welcome, Michael L.

    2003-04-07

    The Global Unified Parallel File System (GUPFS) project is a multiple-phase, five-year project at the National Energy Research Scientific Computing (NERSC) Center to provide a scalable, high performance, high bandwidth, shared file system for all the NERSC production computing and support systems. The primary purpose of the GUPFS project is to make it easier to conduct advanced scientific research using the NERSC systems. This is to be accomplished through the use of a shared file system providing a unified file namespace, operating on consolidated shared storage that is directly accessed by all the NERSC production computing and support systems. During its first year, FY 2002, the GUPFS project focused on identifying, testing, and evaluating existing and emerging shared/cluster file system, SAN fabric, and storage technologies; identifying NERSC user input/output (I/O) requirements, methods, and mechanisms; and developing appropriate benchmarking methodologies and benchmark codes for a parallel environment. This report presents the activities and progress of the GUPFS project during its first year, the results of the evaluations conducted, and plans for near-term and longer-term investigations.

  18. A Metadata-Rich File System

    Energy Technology Data Exchange (ETDEWEB)

    Ames, S; Gokhale, M B; Maltzahn, C

    2009-01-07

    Despite continual improvements in the performance and reliability of large scale file systems, the management of file system metadata has changed little in the past decade. The mismatch between the size and complexity of large scale data stores and their ability to organize and query their metadata has led to a de facto standard in which raw data is stored in traditional file systems, while related, application-specific metadata is stored in relational databases. This separation of data and metadata requires considerable effort to maintain consistency and can result in complex, slow, and inflexible system operation. To address these problems, we have developed the Quasar File System (QFS), a metadata-rich file system in which files, metadata, and file relationships are all first class objects. In contrast to hierarchical file systems and relational databases, QFS defines a graph data model composed of files and their relationships. QFS includes Quasar, an XPATH-extended query language for searching the file system. Results from our QFS prototype show the effectiveness of this approach. Compared to the defacto standard, the QFS prototype shows superior ingest performance and comparable query performance on user metadata-intensive operations and superior performance on normal file metadata operations.

  19. A File Archival System

    Science.gov (United States)

    Fanselow, J. L.; Vavrus, J. L.

    1984-01-01

    ARCH, file archival system for DEC VAX, provides for easy offline storage and retrieval of arbitrary files on DEC VAX system. System designed to eliminate situations that tie up disk space and lead to confusion when different programers develop different versions of same programs and associated files.

  20. Zebra: A striped network file system

    Science.gov (United States)

    Hartman, John H.; Ousterhout, John K.

    1992-01-01

    The design of Zebra, a striped network file system, is presented. Zebra applies ideas from log-structured file system (LFS) and RAID research to network file systems, resulting in a network file system that has scalable performance, uses its servers efficiently even when its applications are using small files, and provides high availability. Zebra stripes file data across multiple servers, so that the file transfer rate is not limited by the performance of a single server. High availability is achieved by maintaining parity information for the file system. If a server fails its contents can be reconstructed using the contents of the remaining servers and the parity information. Zebra differs from existing striped file systems in the way it stripes file data: Zebra does not stripe on a per-file basis; instead it stripes the stream of bytes written by each client. Clients write to the servers in units called stripe fragments, which are analogous to segments in an LFS. Stripe fragments contain file blocks that were written recently, without regard to which file they belong. This method of striping has numerous advantages over per-file striping, including increased server efficiency, efficient parity computation, and elimination of parity update.

  1. Towards the Availability of the Distributed Cluster Rendering System: Automatic Modeling and Verification

    DEFF Research Database (Denmark)

    Wang, Kemin; Jiang, Zhengtao; Wang, Yongbin

    2012-01-01

    , whenever the number of node-n and related parameters vary, we can create the PRISM model file rapidly and then we can use PRISM model checker to verify ralated system properties. At the end of this study, we analyzed and verified the availability distributions of the Distributed Cluster Rendering System......In this study, we proposed a Continuous Time Markov Chain Model towards the availability of n-node clusters of Distributed Rendering System. It's an infinite one, we formalized it, based on the model, we implemented a software, which can automatically model with PRISM language. With the tool...

  2. Formalizing a hierarchical file system

    NARCIS (Netherlands)

    Hesselink, Wim H.; Lali, Muhammad Ikram

    An abstract file system is defined here as a partial function from (absolute) paths to data. Such a file system determines the set of valid paths. It allows the file system to be read and written at a valid path, and it allows the system to be modified by the Unix operations for creation, removal,

  3. The Galley Parallel File System

    Science.gov (United States)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    Most current multiprocessor file systems are designed to use multiple disks in parallel, using the high aggregate bandwidth to meet the growing I/0 requirements of parallel scientific applications. Many multiprocessor file systems provide applications with a conventional Unix-like interface, allowing the application to access multiple disks transparently. This interface conceals the parallelism within the file system, increasing the ease of programmability, but making it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. In addition to providing an insufficient interface, most current multiprocessor file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic scientific multiprocessor workloads. We discuss Galley's file structure and application interface, as well as the performance advantages offered by that interface.

  4. Formalizing a Hierarchical File System

    NARCIS (Netherlands)

    Hesselink, Wim H.; Lali, M.I.

    2009-01-01

    In this note, we define an abstract file system as a partial function from (absolute) paths to data. Such a file system determines the set of valid paths. It allows the file system to be read and written at a valid path, and it allows the system to be modified by the Unix operations for removal

  5. The Jade File System. Ph.D. Thesis

    Science.gov (United States)

    Rao, Herman Chung-Hwa

    1991-01-01

    File systems have long been the most important and most widely used form of shared permanent storage. File systems in traditional time-sharing systems, such as Unix, support a coherent sharing model for multiple users. Distributed file systems implement this sharing model in local area networks. However, most distributed file systems fail to scale from local area networks to an internet. Four characteristics of scalability were recognized: size, wide area, autonomy, and heterogeneity. Owing to size and wide area, techniques such as broadcasting, central control, and central resources, which are widely adopted by local area network file systems, are not adequate for an internet file system. An internet file system must also support the notion of autonomy because an internet is made up by a collection of independent organizations. Finally, heterogeneity is the nature of an internet file system, not only because of its size, but also because of the autonomy of the organizations in an internet. The Jade File System, which provides a uniform way to name and access files in the internet environment, is presented. Jade is a logical system that integrates a heterogeneous collection of existing file systems, where heterogeneous means that the underlying file systems support different file access protocols. Because of autonomy, Jade is designed under the restriction that the underlying file systems may not be modified. In order to avoid the complexity of maintaining an internet-wide, global name space, Jade permits each user to define a private name space. In Jade's design, we pay careful attention to avoiding unnecessary network messages between clients and file servers in order to achieve acceptable performance. Jade's name space supports two novel features: (1) it allows multiple file systems to be mounted under one direction; and (2) it permits one logical name space to mount other logical name spaces. A prototype of Jade was implemented to examine and validate its

  6. An analysis of file system and installation of the file management system for NOS operating system

    International Nuclear Information System (INIS)

    Lee, Young Jai; Park, Sun Hee; Hwang, In Ah; Kim, Hee Kyung

    1992-06-01

    In this technical report, we analyze NOS file structure for Cyber 170-875 and Cyber 960-31 computer system. We also describe functions, procedure and how-to-operate and how-to-use of VDS. VDS is used to manage large files effectively for Cyber computer system. The purpose of the VDS installation is to increase the virtual disk storage by utilizing magnetic tape, to assist the users of the computer system in managing their files, and to enhance the performance of KAERI Cyber computer system. (Author)

  7. Mixed-Media File Systems

    NARCIS (Netherlands)

    Bosch, H.G.P.

    1999-01-01

    This thesis addresses the problem of implementing mixed-media storage systems. In this work a mixed-media file system is defined to be a system that stores both conventional (best-effort) file data and real-time continuous-media data. Continuous-media data is usually bulky, and servers storing and

  8. PFS: a distributed and customizable file system

    OpenAIRE

    Bosch, H.G.P.; Mullender, Sape J.

    1996-01-01

    In this paper we present our ongoing work on the Pegasus File System (PFS), a distributed and customizable file system that can be used for off-line file system experiments and on-line file system storage. PFS is best described as an object-oriented component library from which either a true file system or a file-system simulator can be constructed. Each of the components in the library is easily replaced by another implementation to accommodate a wide range of applications.

  9. PFS: a distributed and customizable file system

    NARCIS (Netherlands)

    Bosch, H.G.P.; Mullender, Sape J.

    1996-01-01

    In this paper we present our ongoing work on the Pegasus File System (PFS), a distributed and customizable file system that can be used for off-line file system experiments and on-line file system storage. PFS is best described as an object-oriented component library from which either a true file

  10. Deceit: A flexible distributed file system

    Science.gov (United States)

    Siegel, Alex; Birman, Kenneth; Marzullo, Keith

    1989-01-01

    Deceit, a distributed file system (DFS) being developed at Cornell, focuses on flexible file semantics in relation to efficiency, scalability, and reliability. Deceit servers are interchangeable and collectively provide the illusion of a single, large server machine to any clients of the Deceit service. Non-volatile replicas of each file are stored on a subset of the file servers. The user is able to set parameters on a file to achieve different levels of availability, performance, and one-copy serializability. Deceit also supports a file version control mechanism. In contrast with many recent DFS efforts, Deceit can behave like a plain Sun Network File System (NFS) server and can be used by any NFS client without modifying any client software. The current Deceit prototype uses the ISIS Distributed Programming Environment for all communication and process group management, an approach that reduces system complexity and increases system robustness.

  11. The global unified parallel file system (GUPFS) project: FY 2003 activities and results

    Energy Technology Data Exchange (ETDEWEB)

    Butler, Gregory F.; Baird William P.; Lee, Rei C.; Tull, Craig E.; Welcome, Michael L.; Whitney Cary L.

    2004-04-30

    The Global Unified Parallel File System (GUPFS) project is a multiple-phase project at the National Energy Research Scientific Computing (NERSC) Center whose goal is to provide a scalable, high-performance, high-bandwidth, shared file system for all of the NERSC production computing and support systems. The primary purpose of the GUPFS project is to make the scientific users more productive as they conduct advanced scientific research at NERSC by simplifying the scientists' data management tasks and maximizing storage and data availability. This is to be accomplished through the use of a shared file system providing a unified file namespace, operating on consolidated shared storage that is accessible by all the NERSC production computing and support systems. In order to successfully deploy a scalable high-performance shared file system with consolidated disk storage, three major emerging technologies must be brought together: (1) shared/cluster file systems software, (2) cost-effective, high-performance storage area network (SAN) fabrics, and (3) high-performance storage devices. Although they are evolving rapidly, these emerging technologies individually are not targeted towards the needs of scientific high-performance computing (HPC). The GUPFS project is in the process of assessing these emerging technologies to determine the best combination of solutions for a center-wide shared file system, to encourage the development of these technologies in directions needed for HPC, particularly at NERSC, and to then put them into service. With the development of an evaluation methodology and benchmark suites, and with the updating of the GUPFS testbed system, the project did a substantial number of investigations and evaluations during FY 2003. The investigations and evaluations involved many vendors and products. From our evaluation of these products, we have found that most vendors and many of the products are more focused on the commercial market. Most vendors

  12. Virtual file system for PSDS

    Science.gov (United States)

    Runnels, Tyson D.

    1993-01-01

    This is a case study. It deals with the use of a 'virtual file system' (VFS) for Boeing's UNIX-based Product Standards Data System (PSDS). One of the objectives of PSDS is to store digital standards documents. The file-storage requirements are that the files must be rapidly accessible, stored for long periods of time - as though they were paper, protected from disaster, and accumulative to about 80 billion characters (80 gigabytes). This volume of data will be approached in the first two years of the project's operation. The approach chosen is to install a hierarchical file migration system using optical disk cartridges. Files are migrated from high-performance media to lower performance optical media based on a least-frequency-used algorithm. The optical media are less expensive per character stored and are removable. Vital statistics about the removable optical disk cartridges are maintained in a database. The assembly of hardware and software acts as a single virtual file system transparent to the PSDS user. The files are copied to 'backup-and-recover' media whose vital statistics are also stored in the database. Seventeen months into operation, PSDS is storing 49 gigabytes. A number of operational and performance problems were overcome. Costs are under control. New and/or alternative uses for the VFS are being considered.

  13. MOLA: a bootable, self-configuring system for virtual screening using AutoDock4/Vina on computer clusters

    Directory of Open Access Journals (Sweden)

    Abreu Rui MV

    2010-10-01

    Full Text Available Abstract Background Virtual screening of small molecules using molecular docking has become an important tool in drug discovery. However, large scale virtual screening is time demanding and usually requires dedicated computer clusters. There are a number of software tools that perform virtual screening using AutoDock4 but they require access to dedicated Linux computer clusters. Also no software is available for performing virtual screening with Vina using computer clusters. In this paper we present MOLA, an easy-to-use graphical user interface tool that automates parallel virtual screening using AutoDock4 and/or Vina in bootable non-dedicated computer clusters. Implementation MOLA automates several tasks including: ligand preparation, parallel AutoDock4/Vina jobs distribution and result analysis. When the virtual screening project finishes, an open-office spreadsheet file opens with the ligands ranked by binding energy and distance to the active site. All results files can automatically be recorded on an USB-flash drive or on the hard-disk drive using VirtualBox. MOLA works inside a customized Live CD GNU/Linux operating system, developed by us, that bypass the original operating system installed on the computers used in the cluster. This operating system boots from a CD on the master node and then clusters other computers as slave nodes via ethernet connections. Conclusion MOLA is an ideal virtual screening tool for non-experienced users, with a limited number of multi-platform heterogeneous computers available and no access to dedicated Linux computer clusters. When a virtual screening project finishes, the computers can just be restarted to their original operating system. The originality of MOLA lies on the fact that, any platform-independent computer available can he added to the cluster, without ever using the computer hard-disk drive and without interfering with the installed operating system. With a cluster of 10 processors, and a

  14. The File System Interface is an Anachronism

    OpenAIRE

    Ellard, Daniel

    2003-01-01

    Contemporary file systems implement a set of abstractions and semantics that are suboptimal for many (if not most) purposes. The philosophy of using the simple mechanisms of the file system as the basis for a vast array of higher-level mechanisms leads to inefficient and incorrect implementations. We propose several extensions to the canonical file system model, including explicit support for lock files, indexed files, and resource forks, and the benefit of session semantics for write updates...

  15. Apically extruded dentin debris by reciprocating single-file and multi-file rotary system.

    Science.gov (United States)

    De-Deus, Gustavo; Neves, Aline; Silva, Emmanuel João; Mendonça, Thais Accorsi; Lourenço, Caroline; Calixto, Camila; Lima, Edson Jorge Moreira

    2015-03-01

    This study aims to evaluate the apical extrusion of debris by the two reciprocating single-file systems: WaveOne and Reciproc. Conventional multi-file rotary system was used as a reference for comparison. The hypotheses tested were (i) the reciprocating single-file systems extrude more than conventional multi-file rotary system and (ii) the reciprocating single-file systems extrude similar amounts of dentin debris. After solid selection criteria, 80 mesial roots of lower molars were included in the present study. The use of four different instrumentation techniques resulted in four groups (n = 20): G1 (hand-file technique), G2 (ProTaper), G3 (WaveOne), and G4 (Reciproc). The apparatus used to evaluate the collection of apically extruded debris was typical double-chamber collector. Statistical analysis was performed for multiple comparisons. No significant difference was found in the amount of the debris extruded between the two reciprocating systems. In contrast, conventional multi-file rotary system group extruded significantly more debris than both reciprocating groups. Hand instrumentation group extruded significantly more debris than all other groups. The present results yielded favorable input for both reciprocation single-file systems, inasmuch as they showed an improved control of apically extruded debris. Apical extrusion of debris has been studied extensively because of its clinical relevance, particularly since it may cause flare-ups, originated by the introduction of bacteria, pulpal tissue, and irrigating solutions into the periapical tissues.

  16. Protecting your files on the DFS file system

    CERN Multimedia

    Computer Security Team

    2011-01-01

    The Windows Distributed File System (DFS) hosts user directories for all NICE users plus many more data.    Files can be accessed from anywhere, via a dedicated web portal (http://cern.ch/dfs). Due to the ease of access to DFS with in CERN it is of utmost importance to properly protect access to sensitive data. As the use of DFS access control mechanisms is not obvious to all users, passwords, certificates or sensitive files might get exposed. At least this happened in past to the Andrews File System (AFS) - the Linux equivalent to DFS) - and led to bad publicity due to a journalist accessing supposedly "private" AFS folders (SonntagsZeitung 2009/11/08). This problem does not only affect the individual user but also has a bad impact on CERN's reputation when it comes to IT security. Therefore, all departments and LHC experiments agreed recently to apply more stringent protections to all DFS user folders. The goal of this data protection policy is to assist users in pro...

  17. Protecting your files on the AFS file system

    CERN Multimedia

    2011-01-01

    The Andrew File System is a world-wide distributed file system linking hundreds of universities and organizations, including CERN. Files can be accessed from anywhere, via dedicated AFS client programs or via web interfaces that export the file contents on the web. Due to the ease of access to AFS it is of utmost importance to properly protect access to sensitive data in AFS. As the use of AFS access control mechanisms is not obvious to all users, passwords, private SSH keys or certificates have been exposed in the past. In one specific instance, this also led to bad publicity due to a journalist accessing supposedly "private" AFS folders (SonntagsZeitung 2009/11/08). This problem does not only affect the individual user but also has a bad impact on CERN's reputation when it comes to IT security. Therefore, all departments and LHC experiments agreed in April 2010 to apply more stringent folder protections to all AFS user folders. The goal of this data protection policy is to assist users in...

  18. The use of a VAX cluster for the DIII-D data acquisition system

    International Nuclear Information System (INIS)

    McHarg, B.B. Jr.

    1991-11-01

    The DIII-D tokamak is a large fusion energy research experiment funded by the Department of Energy. The experiment currently collects nearly 40 Mbytes of data from each shot of the experiment. In the past, most of this data was acquired through the MODCOMP Classic data acquisition computers and then transferred to a DEC VAX computer system for permanent archiving and storage. A much smaller amount of data was acquired from a few MicroVAX based data acquisition systems. In the last two years, MicroVAX based systems have become the standard means for adding new diagnostic data and account for half the total data. There are now 17 VAX systems of various types at the DIII-D facility. As more diagnostics and data are added, it takes increasing of time to merge the data into the central shot file. The system management of so many systems has become increasingly time consuming as well. To improve the efficiency of the overall data acquisition system, a mixed interconnect VAX cluster has been formed consisting of 16 VAX computers. In the cluster, the software protocol for passing data around the cluster is much more efficient than using DECnet. The cluster has also greatly simplified the procedure of backing up disks. Another big improvement is the use of a VAX console system which ties all the console ports of the computers into one central computer system which then manages the entire cluster

  19. Storage of sparse files using parallel log-structured file system

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary; Torres, Aaron

    2017-11-07

    A sparse file is stored without holes by storing a data portion of the sparse file using a parallel log-structured file system; and generating an index entry for the data portion, the index entry comprising a logical offset, physical offset and length of the data portion. The holes can be restored to the sparse file upon a reading of the sparse file. The data portion can be stored at a logical end of the sparse file. Additional storage efficiency can optionally be achieved by (i) detecting a write pattern for a plurality of the data portions and generating a single patterned index entry for the plurality of the patterned data portions; and/or (ii) storing the patterned index entries for a plurality of the sparse files in a single directory, wherein each entry in the single directory comprises an identifier of a corresponding sparse file.

  20. Computer Forensics Method in Analysis of Files Timestamps in Microsoft Windows Operating System and NTFS File System

    Directory of Open Access Journals (Sweden)

    Vesta Sergeevna Matveeva

    2013-02-01

    Full Text Available All existing file browsers displays 3 timestamps for every file in file system NTFS. Nowadays there are a lot of utilities that can manipulate temporal attributes to conceal the traces of file using. However every file in NTFS has 8 timestamps that are stored in file record and used in detecting the fact of attributes substitution. The authors suggest a method of revealing original timestamps after replacement and automated variant of it in case of a set of files.

  1. Small file aggregation in a parallel computing system

    Science.gov (United States)

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Grider, Gary; Zhang, Jingwang

    2014-09-02

    Techniques are provided for small file aggregation in a parallel computing system. An exemplary method for storing a plurality of files generated by a plurality of processes in a parallel computing system comprises aggregating the plurality of files into a single aggregated file; and generating metadata for the single aggregated file. The metadata comprises an offset and a length of each of the plurality of files in the single aggregated file. The metadata can be used to unpack one or more of the files from the single aggregated file.

  2. MR-AFS: a global hierarchical file-system

    International Nuclear Information System (INIS)

    Reuter, H.

    2000-01-01

    The next generation of fusion experiments will use object-oriented technology creating the need for world wide sharing of an underlying hierarchical file-system. The Andrew file system (AFS) is a well known and widely spread global distributed file-system. Multiple-resident-AFS (MR-AFS) combines the features of AFS with hierarchical storage management systems. Files in MR-AFS therefore may be migrated on secondary storage, such as roboted tape libraries. MR-AFS is in use at IPP for the current experiments and data originating from super-computer applications. Experiences and scalability issues are discussed

  3. Improving File System Performance by Striping

    Science.gov (United States)

    Lam, Terance L.; Kutler, Paul (Technical Monitor)

    1998-01-01

    This document discusses the performance and advantages of striped file systems on the SGI AD workstations. Performance of several striped file system configurations are compared and guidelines for optimal striping are recommended.

  4. Storing files in a parallel computing system using list-based index to identify replica files

    Science.gov (United States)

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Zhang, Zhenhua; Grider, Gary

    2015-07-21

    Improved techniques are provided for storing files in a parallel computing system using a list-based index to identify file replicas. A file and at least one replica of the file are stored in one or more storage nodes of the parallel computing system. An index for the file comprises at least one list comprising a pointer to a storage location of the file and a storage location of the at least one replica of the file. The file comprises one or more of a complete file and one or more sub-files. The index may also comprise a checksum value for one or more of the file and the replica(s) of the file. The checksum value can be evaluated to validate the file and/or the file replica(s). A query can be processed using the list.

  5. DMFS: A Data Migration File System for NetBSD

    Science.gov (United States)

    Studenmund, William

    2000-01-01

    I have recently developed DMFS, a Data Migration File System, for NetBSD. This file system provides kernel support for the data migration system being developed by my research group at NASA/Ames. The file system utilizes an underlying file store to provide the file backing, and coordinates user and system access to the files. It stores its internal metadata in a flat file, which resides on a separate file system. This paper will first describe our data migration system to provide a context for DMFS, then it will describe DMFS. It also will describe the changes to NetBSD needed to make DMFS work. Then it will give an overview of the file archival and restoration procedures, and describe how some typical user actions are modified by DMFS. Lastly, the paper will present simple performance measurements which indicate that there is little performance loss due to the use of the DMFS layer.

  6. Flexibility and Performance of Parallel File Systems

    Science.gov (United States)

    Kotz, David; Nieuwejaar, Nils

    1996-01-01

    As we gain experience with parallel file systems, it becomes increasingly clear that a single solution does not suit all applications. For example, it appears to be impossible to find a single appropriate interface, caching policy, file structure, or disk-management strategy. Furthermore, the proliferation of file-system interfaces and abstractions make applications difficult to port. We propose that the traditional functionality of parallel file systems be separated into two components: a fixed core that is standard on all platforms, encapsulating only primitive abstractions and interfaces, and a set of high-level libraries to provide a variety of abstractions and application-programmer interfaces (API's). We present our current and next-generation file systems as examples of this structure. Their features, such as a three-dimensional file structure, strided read and write interfaces, and I/O-node programs, are specifically designed with the flexibility and performance necessary to support a wide range of applications.

  7. Design and Implementation of a Metadata-rich File System

    Energy Technology Data Exchange (ETDEWEB)

    Ames, S; Gokhale, M B; Maltzahn, C

    2010-01-19

    Despite continual improvements in the performance and reliability of large scale file systems, the management of user-defined file system metadata has changed little in the past decade. The mismatch between the size and complexity of large scale data stores and their ability to organize and query their metadata has led to a de facto standard in which raw data is stored in traditional file systems, while related, application-specific metadata is stored in relational databases. This separation of data and semantic metadata requires considerable effort to maintain consistency and can result in complex, slow, and inflexible system operation. To address these problems, we have developed the Quasar File System (QFS), a metadata-rich file system in which files, user-defined attributes, and file relationships are all first class objects. In contrast to hierarchical file systems and relational databases, QFS defines a graph data model composed of files and their relationships. QFS incorporates Quasar, an XPATH-extended query language for searching the file system. Results from our QFS prototype show the effectiveness of this approach. Compared to the de facto standard, the QFS prototype shows superior ingest performance and comparable query performance on user metadata-intensive operations and superior performance on normal file metadata operations.

  8. Dynamic Non-Hierarchical File Systems for Exascale Storage

    Energy Technology Data Exchange (ETDEWEB)

    Long, Darrell E. [Univ. of California, Santa Cruz, CA (United States); Miller, Ethan L [Univ. of California, Santa Cruz, CA (United States)

    2015-02-24

    This constitutes the final report for “Dynamic Non-Hierarchical File Systems for Exascale Storage”. The ultimate goal of this project was to improve data management in scientific computing and high-end computing (HEC) applications, and to achieve this goal we proposed: to develop the first, HEC-targeted, file system featuring rich metadata and provenance collection, extreme scalability, and future storage hardware integration as core design goals, and to evaluate and develop a flexible non-hierarchical file system interface suitable for providing more powerful and intuitive data management interfaces to HEC and scientific computing users. Data management is swiftly becoming a serious problem in the scientific community – while copious amounts of data are good for obtaining results, finding the right data is often daunting and sometimes impossible. Scientists participating in a Department of Energy workshop noted that most of their time was spent “...finding, processing, organizing, and moving data and it’s going to get much worse”. Scientists should not be forced to become data mining experts in order to retrieve the data they want, nor should they be expected to remember the naming convention they used several years ago for a set of experiments they now wish to revisit. Ideally, locating the data you need would be as easy as browsing the web. Unfortunately, existing data management approaches are usually based on hierarchical naming, a 40 year-old technology designed to manage thousands of files, not exabytes of data. Today’s systems do not take advantage of the rich array of metadata that current high-end computing (HEC) file systems can gather, including content-based metadata and provenance1 information. As a result, current metadata search approaches are typically ad hoc and often work by providing a parallel management system to the “main” file system, as is done in Linux (the locate utility), personal computers, and enterprise search

  9. Collective operations in a file system based execution model

    Science.gov (United States)

    Shinde, Pravin; Van Hensbergen, Eric

    2013-02-19

    A mechanism is provided for group communications using a MULTI-PIPE synthetic file system. A master application creates a multi-pipe synthetic file in the MULTI-PIPE synthetic file system, the master application indicating a multi-pipe operation to be performed. The master application then writes a header-control block of the multi-pipe synthetic file specifying at least one of a multi-pipe synthetic file system name, a message type, a message size, a specific destination, or a specification of the multi-pipe operation. Any other application participating in the group communications then opens the same multi-pipe synthetic file. A MULTI-PIPE file system module then implements the multi-pipe operation as identified by the master application. The master application and the other applications then either read or write operation messages to the multi-pipe synthetic file and the MULTI-PIPE synthetic file system module performs appropriate actions.

  10. Performance of the Galley Parallel File System

    Science.gov (United States)

    Nieuwejaar, Nils; Kotz, David

    1996-01-01

    As the input/output (I/O) needs of parallel scientific applications increase, file systems for multiprocessors are being designed to provide applications with parallel access to multiple disks. Many parallel file systems present applications with a conventional Unix-like interface that allows the application to access multiple disks transparently. This interface conceals the parallism within the file system, which increases the ease of programmability, but makes it difficult or impossible for sophisticated programmers and libraries to use knowledge about their I/O needs to exploit that parallelism. Furthermore, most current parallel file systems are optimized for a different workload than they are being asked to support. We introduce Galley, a new parallel file system that is intended to efficiently support realistic parallel workloads. Initial experiments, reported in this paper, indicate that Galley is capable of providing high-performance 1/O to applications the applications that rely on them. In Section 3 we describe that access data in patterns that have been observed to be common.

  11. RAMA: A file system for massively parallel computers

    Science.gov (United States)

    Miller, Ethan L.; Katz, Randy H.

    1993-01-01

    This paper describes a file system design for massively parallel computers which makes very efficient use of a few disks per processor. This overcomes the traditional I/O bottleneck of massively parallel machines by storing the data on disks within the high-speed interconnection network. In addition, the file system, called RAMA, requires little inter-node synchronization, removing another common bottleneck in parallel processor file systems. Support for a large tertiary storage system can easily be integrated in lo the file system; in fact, RAMA runs most efficiently when tertiary storage is used.

  12. Specifying a Realistic File System

    Directory of Open Access Journals (Sweden)

    Sidney Amani

    2015-11-01

    Full Text Available We present the most interesting elements of the correctness specification of BilbyFs, a performant Linux flash file system. The BilbyFs specification supports asynchronous writes, a feature that has been overlooked by several file system verification projects, and has been used to verify the correctness of BilbyFs's fsync( C implementation. It makes use of nondeterminism to be concise and is shallowly-embedded in higher-order logic.

  13. Clockwise: A Mixed-Media File System

    NARCIS (Netherlands)

    Bosch, H.G.P.; Jansen, P.G.; Mullender, Sape J.

    This (short) paper presents the Clockwise, a mixed-media file system. The primary goal of the Clockwise is to provide a storage architecture that supports the storage and retrieval of best-effort and real-time file system data. Clockwise provides an abstraction called a dynamic partition that groups

  14. Optimizing Input/Output Using Adaptive File System Policies

    Science.gov (United States)

    Madhyastha, Tara M.; Elford, Christopher L.; Reed, Daniel A.

    1996-01-01

    Parallel input/output characterization studies and experiments with flexible resource management algorithms indicate that adaptivity is crucial to file system performance. In this paper we propose an automatic technique for selecting and refining file system policies based on application access patterns and execution environment. An automatic classification framework allows the file system to select appropriate caching and pre-fetching policies, while performance sensors provide feedback used to tune policy parameters for specific system environments. To illustrate the potential performance improvements possible using adaptive file system policies, we present results from experiments involving classification-based and performance-based steering.

  15. XCPU2 process management system

    Energy Technology Data Exchange (ETDEWEB)

    Ionkov, Latchesar [Los Alamos National Laboratory; Van Hensbergen, Eric [IBM AUSTIN RESEARCH LAB

    2009-01-01

    Xcpu2 is a new process management system that allows the users to specify custom file system for a running job. Most cluster management systems enforce single software distribution running on all nodes. Xcpu2 allows programs running on the cluster to work in environment identical to the user's desktop, using the same versions of the libraries and tools the user installed locally, and accessing the configuration file in the same places they are located on the desktop. Xcpu2 builds on our earlier work with the Xcpu system. Like Xcpu, Xcpu2's process management interface is represented as a set of files exported by a 9P file server. It supports heterogeneous clusters and multiple head nodes. Unlike Xcpu, it uses pull instead of push model. In this paper we describe the Xcpu2 clustering model, its operation and how the per-job filesystem configuration can be used to solve some of the common problems when running a cluster.

  16. Development of an integrated filing system for endoscopic images.

    Science.gov (United States)

    Fujino, M A; Ikeda, M; Yamamoto, Y; Kinose, T; Tachikawa, H; Morozumi, A; Sano, S; Kojima, Y; Nakamura, T; Kawai, T

    1991-01-01

    A new integrated filing system for endoscopic images has been developed, comprising a main image filing system and subsystems located at different stations. A hybrid filing system made up of both digital and analog filing devices was introduced to construct this system that combines the merits of the two filing methods. Each subsystem provided with a video processor, is equipped with a digital filing device, and routine images were recorded in the analog image filing device of the main system. The use of a multi-input adapter enabled simultaneous input of analog images from up to 8 video processors. Recorded magneto-optical disks make it possible to recall the digital images at any station in the hospital; the disks are copied without image degradation and also utilised for image processing. This system promises reliable storage and integrated, efficient management of endoscopic information. It also costs less to install than the so-called PACS (picture archiving and communication system), which connects all the stations of the hospital using optical fiber cables.

  17. 78 FR 21930 - Aquenergy Systems, Inc.; Notice of Intent To File License Application, Filing of Pre-Application...

    Science.gov (United States)

    2013-04-12

    ... Systems, Inc.; Notice of Intent To File License Application, Filing of Pre-Application Document, and Approving Use of the Traditional Licensing Process a. Type of Filing: Notice of Intent to File License...: November 11, 2012. d. Submitted by: Aquenergy Systems, Inc., a fully owned subsidiaries of Enel Green Power...

  18. Non-POSIX File System for LHCb Online Event Handling

    CERN Document Server

    Garnier, J C; Cherukuwada, S S

    2011-01-01

    LHCb aims to use its O(20000) CPU cores in the high level trigger (HLT) and its 120 TB Online storage system for data reprocessing during LHC shutdown periods. These periods can last a few days for technical maintenance or only a few hours during beam interfill gaps. These jobs run on files which are staged in from tape storage to the local storage buffer. The result are again one or more files. Efficient file writing and reading is essential for the performance of the system. Rather than using a traditional shared file-system such as NFS or CIFS we have implemented a custom, light-weight, non-Posix network file-system for the handling of these files. Streaming this file-system for the data-access allows to obtain high performance, while at the same time keep the resource consumption low and add nice features not found in NFS such as high-availability, transparent fail-over of the read and write service. The writing part of this streaming service is in successful use for the Online, real-time writing of the d...

  19. The design and development of GRASS file reservation system

    International Nuclear Information System (INIS)

    Huang Qiulan; Zhu Suijiang; Cheng Yaodong; Chen Gang

    2010-01-01

    GFRS (GRASS File Reservation System) is designed to improve the file access performance of GRASS (Grid-enabled Advanced Storage System) which is a Hierarchical Storage Management (HSM) system developed at Computing Center, Institute of High Energy Physics. GRASS can provide massive storage management and data migration, but the data migration policy is simply based factors such as pool water level, the intervals for migration and so on, so it is short of precise control over files. As for that, we design GFRS to implement user-based file reservation which is to reserve and keep the required files on disks for High Energy physicists. CFRS can improve file access speed for users by avoiding migrating frequently accessed files to tapes. In this paper we first give a brief introduction of GRASS system and then detailed architecture and implementation of GFRS. Experiments results from GFRS have shown good performance and a simple analysis is made based on it. (authors)

  20. Athos: Efficient Authentication of Outsourced File Systems

    DEFF Research Database (Denmark)

    Triandopoulos, Nikolaos; Goodrich, Michael T.; Papamanthou, Charalampos

    2008-01-01

    We study the problem of authenticated storage, where we wish to construct protocols that allow to outsource any complex file system to an untrusted server and yet ensure the file-system's integrity. We introduce Athos, a new, platform-independent and user-transparent architecture for authenticated...

  1. Research of Performance Linux Kernel File Systems

    Directory of Open Access Journals (Sweden)

    Andrey Vladimirovich Ostroukh

    2015-10-01

    Full Text Available The article describes the most common Linux Kernel File Systems. The research was carried out on a personal computer, the characteristics of which are written in the article. The study was performed on a typical workstation running GNU/Linux with below characteristics. On a personal computer for measuring the file performance, has been installed the necessary software. Based on the results, conclusions and proposed recommendations for use of file systems. Identified and recommended by the best ways to store data.

  2. Final Report for File System Support for Burst Buffers on HPC Systems

    Energy Technology Data Exchange (ETDEWEB)

    Yu, W. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Mohror, K. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2017-11-27

    Distributed burst buffers are a promising storage architecture for handling I/O workloads for exascale computing. As they are being deployed on more supercomputers, a file system that efficiently manages these burst buffers for fast I/O operations carries great consequence. Over the past year, FSU team has undertaken several efforts to design, prototype and evaluate distributed file systems for burst buffers on HPC systems. These include MetaKV: a Key-Value Store for Metadata Management of Distributed Burst Buffers, a user-level file system with multiple backends, and a specialized file system for large datasets of deep neural networks. Our progress for these respective efforts are elaborated further in this report.

  3. Very wide register : an asymmetric register file organization for low power embedded processors.

    NARCIS (Netherlands)

    Raghavan, P.; Lambrechts, A.; Jayapala, M.; Catthoor, F.; Verkest, D.T.M.L.; Corporaal, H.

    2007-01-01

    In current embedded systems processors, multi-ported register files are one of the most power hungry parts of the processor, even when they are clustered. This paper presents a novel register file architecture, which has single ported cells and asymmetric interfaces to the memory and to the

  4. Multi-level, automatic file management system using magnetic disk, mass storage system and magnetic tape

    International Nuclear Information System (INIS)

    Fujii, Minoru; Asai, Kiyoshi

    1979-12-01

    A simple, effective file management system using magnetic disk, mass storage system (MSS) and magnetic tape is described. Following are the concepts and techniques introduced in this file management system. (1) File distribution and continuity character of file references are closely approximated by memory retention function. A density function using the memory retention function is thus defined. (2) A method of computing the cost/benefit lines for magnetic disk, MSS and magnetic tape is presented. (3) A decision process of an optimal organization of file facilities incorporating file demands distribution to respective file devices, is presented. (4) A method of simple, practical, effective, automatic file management, incorporating multi-level file management, space management and file migration control, is proposed. (author)

  5. Distributed PACS using distributed file system with hierarchical meta data servers.

    Science.gov (United States)

    Hiroyasu, Tomoyuki; Minamitani, Yoshiyuki; Miki, Mitsunori; Yokouchi, Hisatake; Yoshimi, Masato

    2012-01-01

    In this research, we propose a new distributed PACS (Picture Archiving and Communication Systems) which is available to integrate several PACSs that exist in each medical institution. The conventional PACS controls DICOM file into one data-base. On the other hand, in the proposed system, DICOM file is separated into meta data and image data and those are stored individually. Using this mechanism, since file is not always accessed the entire data, some operations such as finding files, changing titles, and so on can be performed in high-speed. At the same time, as distributed file system is utilized, accessing image files can also achieve high-speed access and high fault tolerant. The introduced system has a more significant point. That is the simplicity to integrate several PACSs. In the proposed system, only the meta data servers are integrated and integrated system can be constructed. This system also has the scalability of file access with along to the number of file numbers and file sizes. On the other hand, because meta-data server is integrated, the meta data server is the weakness of this system. To solve this defect, hieratical meta data servers are introduced. Because of this mechanism, not only fault--tolerant ability is increased but scalability of file access is also increased. To discuss the proposed system, the prototype system using Gfarm was implemented. For evaluating the implemented system, file search operating time of Gfarm and NFS were compared.

  6. Tuning HDF5 subfiling performance on parallel file systems

    Energy Technology Data Exchange (ETDEWEB)

    Byna, Suren [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Chaarawi, Mohamad [Intel Corp. (United States); Koziol, Quincey [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Mainzer, John [The HDF Group (United States); Willmore, Frank [The HDF Group (United States)

    2017-05-12

    Subfiling is a technique used on parallel file systems to reduce locking and contention issues when multiple compute nodes interact with the same storage target node. Subfiling provides a compromise between the single shared file approach that instigates the lock contention problems on parallel file systems and having one file per process, which results in generating a massive and unmanageable number of files. In this paper, we evaluate and tune the performance of recently implemented subfiling feature in HDF5. In specific, we explain the implementation strategy of subfiling feature in HDF5, provide examples of using the feature, and evaluate and tune parallel I/O performance of this feature with parallel file systems of the Cray XC40 system at NERSC (Cori) that include a burst buffer storage and a Lustre disk-based storage. We also evaluate I/O performance on the Cray XC30 system, Edison, at NERSC. Our results show performance benefits of 1.2X to 6X performance advantage with subfiling compared to writing a single shared HDF5 file. We present our exploration of configurations, such as the number of subfiles and the number of Lustre storage targets to storing files, as optimization parameters to obtain superior I/O performance. Based on this exploration, we discuss recommendations for achieving good I/O performance as well as limitations with using the subfiling feature.

  7. Performance enhancement of a web-based picture archiving and communication system using commercial off-the-shelf server clusters.

    Science.gov (United States)

    Liu, Yan-Lin; Shih, Cheng-Ting; Chang, Yuan-Jen; Chang, Shu-Jun; Wu, Jay

    2014-01-01

    The rapid development of picture archiving and communication systems (PACSs) thoroughly changes the way of medical informatics communication and management. However, as the scale of a hospital's operations increases, the large amount of digital images transferred in the network inevitably decreases system efficiency. In this study, a server cluster consisting of two server nodes was constructed. Network load balancing (NLB), distributed file system (DFS), and structured query language (SQL) duplication services were installed. A total of 1 to 16 workstations were used to transfer computed radiography (CR), computed tomography (CT), and magnetic resonance (MR) images simultaneously to simulate the clinical situation. The average transmission rate (ATR) was analyzed between the cluster and noncluster servers. In the download scenario, the ATRs of CR, CT, and MR images increased by 44.3%, 56.6%, and 100.9%, respectively, when using the server cluster, whereas the ATRs increased by 23.0%, 39.2%, and 24.9% in the upload scenario. In the mix scenario, the transmission performance increased by 45.2% when using eight computer units. The fault tolerance mechanisms of the server cluster maintained the system availability and image integrity. The server cluster can improve the transmission efficiency while maintaining high reliability and continuous availability in a healthcare environment.

  8. Performance Enhancement of a Web-Based Picture Archiving and Communication System Using Commercial Off-the-Shelf Server Clusters

    Directory of Open Access Journals (Sweden)

    Yan-Lin Liu

    2014-01-01

    Full Text Available The rapid development of picture archiving and communication systems (PACSs thoroughly changes the way of medical informatics communication and management. However, as the scale of a hospital’s operations increases, the large amount of digital images transferred in the network inevitably decreases system efficiency. In this study, a server cluster consisting of two server nodes was constructed. Network load balancing (NLB, distributed file system (DFS, and structured query language (SQL duplication services were installed. A total of 1 to 16 workstations were used to transfer computed radiography (CR, computed tomography (CT, and magnetic resonance (MR images simultaneously to simulate the clinical situation. The average transmission rate (ATR was analyzed between the cluster and noncluster servers. In the download scenario, the ATRs of CR, CT, and MR images increased by 44.3%, 56.6%, and 100.9%, respectively, when using the server cluster, whereas the ATRs increased by 23.0%, 39.2%, and 24.9% in the upload scenario. In the mix scenario, the transmission performance increased by 45.2% when using eight computer units. The fault tolerance mechanisms of the server cluster maintained the system availability and image integrity. The server cluster can improve the transmission efficiency while maintaining high reliability and continuous availability in a healthcare environment.

  9. Secure File Allocation and Caching in Large-scale Distributed Systems

    DEFF Research Database (Denmark)

    Di Mauro, Alessio; Mei, Alessandro; Jajodia, Sushil

    2012-01-01

    In this paper, we present a file allocation and caching scheme that guarantees high assurance, availability, and load balancing in a large-scale distributed file system that can support dynamic updates of authorization policies. The scheme uses fragmentation and replication to store files with hi......-balancing, and reducing delay of read operations. The system offers a trade-off-between performance and security that is dynamically tunable according to the current level of threat. We validate our mechanisms with extensive simulations in an Internet-like network.......In this paper, we present a file allocation and caching scheme that guarantees high assurance, availability, and load balancing in a large-scale distributed file system that can support dynamic updates of authorization policies. The scheme uses fragmentation and replication to store files with high...... security requirements in a system composed of a majority of low-security servers. We develop mechanisms to fragment files, to allocate them into multiple servers, and to cache them as close as possible to their readers while preserving the security requirement of the files, providing load...

  10. Dynamically allocated virtual clustering management system

    Science.gov (United States)

    Marcus, Kelvin; Cannata, Jess

    2013-05-01

    The U.S Army Research Laboratory (ARL) has built a "Wireless Emulation Lab" to support research in wireless mobile networks. In our current experimentation environment, our researchers need the capability to run clusters of heterogeneous nodes to model emulated wireless tactical networks where each node could contain a different operating system, application set, and physical hardware. To complicate matters, most experiments require the researcher to have root privileges. Our previous solution of using a single shared cluster of statically deployed virtual machines did not sufficiently separate each user's experiment due to undesirable network crosstalk, thus only one experiment could be run at a time. In addition, the cluster did not make efficient use of our servers and physical networks. To address these concerns, we created the Dynamically Allocated Virtual Clustering management system (DAVC). This system leverages existing open-source software to create private clusters of nodes that are either virtual or physical machines. These clusters can be utilized for software development, experimentation, and integration with existing hardware and software. The system uses the Grid Engine job scheduler to efficiently allocate virtual machines to idle systems and networks. The system deploys stateless nodes via network booting. The system uses 802.1Q Virtual LANs (VLANs) to prevent experimentation crosstalk and to allow for complex, private networks eliminating the need to map each virtual machine to a specific switch port. The system monitors the health of the clusters and the underlying physical servers and it maintains cluster usage statistics for historical trends. Users can start private clusters of heterogeneous nodes with root privileges for the duration of the experiment. Users also control when to shutdown their clusters.

  11. 76 FR 61956 - Electronic Tariff Filing System (ETFS)

    Science.gov (United States)

    2011-10-06

    ...] Electronic Tariff Filing System (ETFS) AGENCY: Federal Communications Commission. ACTION: Final rule... with the Commission's Electronic Tariff Filing System (ETFS), Report and Order (Order). This notice is...: Pamela Arluk, Pricing Policy Division, Wireline Competition Bureau, at (202) 418-1520, or email: pamela...

  12. Reliable file sharing in distributed operating system using web RTC

    Science.gov (United States)

    Dukiya, Rajesh

    2017-12-01

    Since, the evolution of distributed operating system, distributed file system is come out to be important part in operating system. P2P is a reliable way in Distributed Operating System for file sharing. It was introduced in 1999, later it became a high research interest topic. Peer to Peer network is a type of network, where peers share network workload and other load related tasks. A P2P network can be a period of time connection, where a bunch of computers connected by a USB (Universal Serial Bus) port to transfer or enable disk sharing i.e. file sharing. Currently P2P requires special network that should be designed in P2P way. Nowadays, there is a big influence of browsers in our life. In this project we are going to study of file sharing mechanism in distributed operating system in web browsers, where we will try to find performance bottlenecks which our research will going to be an improvement in file sharing by performance and scalability in distributed file systems. Additionally, we will discuss the scope of Web Torrent file sharing and free-riding in peer to peer networks.

  13. Study and development of a document file system with selective access

    International Nuclear Information System (INIS)

    Mathieu, Jean-Claude

    1974-01-01

    The objective of this research thesis was to design and to develop a set of software aimed at an efficient management of a document file system by using methods of selective access to information. Thus, the three main aspects of file processing (creation, modification, reorganisation) have been addressed. The author first presents the main problems related to the development of a comprehensive automatic documentation system, and their conventional solutions. Some future aspects, notably dealing with the development of peripheral computer technology, are also evoked. He presents the characteristics of INIS bibliographic records provided by the IAEA which have been used to create the files. In the second part, he briefly describes the file system general organisation. This system is based on the use of two main files: an inverse file which contains for each descriptor a list of of numbers of files indexed by this descriptor, and a dictionary of descriptor or input file which gives access to the inverse file. The organisation of these both files is then describes in a detailed way. Other related or associated files are created, and the overall architecture and mechanisms integrated into the file data input software are described, as well as various processing applied to these different files. Performance and possible development are finally discussed

  14. Experiments towards model-based testing using Plan 9: Labelled transition file systems, stacking file systems, on-the-fly coverage measuring

    NARCIS (Netherlands)

    Belinfante, Axel; Guardiola, G.; Soriano, E.; Ballesteros, F.J.

    2006-01-01

    We report on experiments that we did on Plan 9/Inferno to gain more experience with the file-system-as-tool-interface approach. We reimplemented functionality that we earlier worked on in Unix, trying to use Plan 9 file system interfaces. The application domain for those experiments was model-based

  15. EFFICIENCY ANALYSIS OF HASHING METHODS FOR FILE SYSTEMS IN USER MODE

    Directory of Open Access Journals (Sweden)

    E. Y. Ivanov

    2013-05-01

    Full Text Available The article deals with characteristics and performance of interaction protocols between virtual file system and file system, their influence on processing power of microkernel operating systems. User mode implementation of ext2 file system for MINIX 3 OS is used to show that in microkernel operating systems file object identification time might increase up to 26 times in comparison with monolithic systems. Therefore, we present efficiency analysis of various hashing methods for file systems, running in user mode. Studies have shown that using hashing methods recommended in this paper it is possible to achieve competitive performance of the considered component of I/O stacks in microkernel and monolithic operating systems.

  16. NASA work unit system file maintenance manual

    Science.gov (United States)

    1972-01-01

    The NASA Work Unit System is a management information system for research tasks (i.e., work units) performed under NASA grants and contracts. It supplies profiles on research efforts and statistics on fund distribution. The file maintenance operator can add, delete and change records at a remote terminal or can submit punched cards to the computer room for batch update. The system is designed for file maintenance by a person with little or no knowledge of data processing techniques.

  17. Cluster Table - KAIKOcDNA | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available Description of data contents List of number of identical cDNA sequences that make up cluster. Data file File name: kaiko_cdna..._cluster.zip File URL: ftp://ftp.biosciencedbc.jp/archive/kaiko-cdna/LATEST/kaiko_cdna_clu...ster.zip File size: 453 KB Simple search URL http://togodb.biosciencedbc.jp/togodb/view/kaiko_cdna

  18. File-System Workload on a Scientific Multiprocessor

    Science.gov (United States)

    Kotz, David; Nieuwejaar, Nils

    1995-01-01

    Many scientific applications have intense computational and I/O requirements. Although multiprocessors have permitted astounding increases in computational performance, the formidable I/O needs of these applications cannot be met by current multiprocessors a their I/O subsystems. To prevent I/O subsystems from forever bottlenecking multiprocessors and limiting the range of feasible applications, new I/O subsystems must be designed. The successful design of computer systems (both hardware and software) depends on a thorough understanding of their intended use. A system designer optimizes the policies and mechanisms for the cases expected to most common in the user's workload. In the case of multiprocessor file systems, however, designers have been forced to build file systems based only on speculation about how they would be used, extrapolating from file-system characterizations of general-purpose workloads on uniprocessor and distributed systems or scientific workloads on vector supercomputers (see sidebar on related work). To help these system designers, in June 1993 we began the Charisma Project, so named because the project sought to characterize 1/0 in scientific multiprocessor applications from a variety of production parallel computing platforms and sites. The Charisma project is unique in recording individual read and write requests-in live, multiprogramming, parallel workloads (rather than from selected or nonparallel applications). In this article, we present the first results from the project: a characterization of the file-system workload an iPSC/860 multiprocessor running production, parallel scientific applications at NASA's Ames Research Center.

  19. Launching large computing applications on a disk-less cluster

    International Nuclear Information System (INIS)

    Schwemmer, Rainer; Caicedo Carvajal, Juan Manuel; Neufeld, Niko

    2011-01-01

    The LHCb Event Filter Farm system is based on a cluster of the order of 1.500 disk-less Linux nodes. Each node runs one instance of the filtering application per core. The amount of cores in our current production environment is 8 per machine for the old cluster and 12 per machine on extension of the cluster. Each instance has to load about 1.000 shared libraries, weighting 200 MB from several directory locations from a central repository. The repository is currently hosted on a SAN and exported via NFS. The libraries are all available in the local file system cache on every node. Loading a library still causes a huge number of requests to the server though, because the loader will try to probe every available path. Measurements show there are between 100.000-200.000 calls per application instance start up. Multiplied by the numbers of cores in the farm, this translates into a veritable DDoS attack on the servers, which lasts several minutes. Since the application is being restarted frequently, a better solution had to be found.scp Rolling out the software to the nodes is out of the question, because they have no disks and the software in it's entirety is too large to put into a ram disk. To solve this problem we developed a FUSE based file systems which acts as a permanent, controllable cache that keeps the essential files that are necessary in stock.

  20. How to detect trap cluster systems?

    International Nuclear Information System (INIS)

    Mandowski, Arkadiusz

    2008-01-01

    Spatially correlated traps and recombination centres (trap-recombination centre pairs and larger clusters) are responsible for many anomalous phenomena that are difficult to explain in the framework of both classical models, i.e. model of localized transitions (LT) and the simple trap model (STM), even with a number of discrete energy levels. However, these 'anomalous' effects may provide a good platform for identifying trap cluster systems. This paper considers selected cluster-type effects, mainly relating to an anomalous dependence of TL on absorbed dose in the system of isolated clusters (ICs). Some consequences for interacting cluster (IAC) systems, involving both localized and delocalized transitions occurring simultaneously, are also discussed

  1. Cluster - ClEST | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available switchLanguage; BLAST Search Image Search Home About Archive Update History Data ...e File name: clest_cluster.zip File URL: ftp://ftp.biosciencedbc.jp/archive/clest/LATEST/clest_cluster.zip F...ownload License Update History of This Database Site Policy | Contact Us Cluster - ClEST | LSDB Archive ...

  2. Development of EDFSRS: evaluated data files storage and retrieval system

    International Nuclear Information System (INIS)

    Hasegawa, Akira

    1985-07-01

    EDFSRS: Evaluated Data Files Storage and Retrieval System has been developed, which is a complete service system for the evaluated nuclear data files compiled in the major three formats: ENDF/B, UKNDL and KEDAK. This system intends to give efficient loading and maintenance of evaluated nuclear data files to the data base administrators and efficient retrievals to their users not only with the easiness but with the best confidence. It can give users all of the information available in these major three formats. The system consists of more than fifteen independent programs and some 150 Mega byte data files and index files (data-base) of the loaded data. In addition it is designed to be operated in the on-line TSS (Time Sharing System) mode, so that users can get any information from their desk top terminals. This report is prepared as a reference manual of the EDFSRS. (author)

  3. Oak Ridge Institutional Cluster Autotune Test Drive Report

    Energy Technology Data Exchange (ETDEWEB)

    Jibonananda, Sanyal [ORNL; New, Joshua Ryan [ORNL

    2014-02-01

    The Oak Ridge Institutional Cluster (OIC) provides general purpose computational resources for the ORNL staff to run computation heavy jobs that are larger than desktop applications but do not quite require the scale and power of the Oak Ridge Leadership Computing Facility (OLCF). This report details the efforts made and conclusions derived in performing a short test drive of the cluster resources on Phase 5 of the OIC. EnergyPlus was used in the analysis as a candidate user program and the overall software environment was evaluated against anticipated challenges experienced with resources such as the shared memory-Nautilus (JICS) and Titan (OLCF). The OIC performed within reason and was found to be acceptable in the context of running EnergyPlus simulations. The number of cores per node and the availability of scratch space per node allow non-traditional desktop focused applications to leverage parallel ensemble execution. Although only individual runs of EnergyPlus were executed, the software environment on the OIC appeared suitable to run ensemble simulations with some modifications to the Autotune workflow. From a standpoint of general usability, the system supports common Linux libraries, compilers, standard job scheduling software (Torque/Moab), and the OpenMPI library (the only MPI library) for MPI communications. The file system is a Panasas file system which literature indicates to be an efficient file system.

  4. Remote file inquiry (RFI) system

    Science.gov (United States)

    1975-01-01

    System interrogates and maintains user-definable data files from remote terminals, using English-like, free-form query language easily learned by persons not proficient in computer programming. System operates in asynchronous mode, allowing any number of inquiries within limitation of available core to be active concurrently.

  5. Digital Libraries: The Next Generation in File System Technology.

    Science.gov (United States)

    Bowman, Mic; Camargo, Bill

    1998-01-01

    Examines file sharing within corporations that use wide-area, distributed file systems. Applications and user interactions strongly suggest that the addition of services typically associated with digital libraries (content-based file location, strongly typed objects, representation of complex relationships between documents, and extrinsic…

  6. Overview and Status of the Ceph File System

    CERN Multimedia

    CERN. Geneva

    2017-01-01

    The Ceph file system (CephFS) is the POSIX-compatible distributed file system running on top of Ceph's powerful and stable object store. This presentation will give a general introduction of CephFS and detail the recent work the Ceph team has done to improve its stability and usability. In particular, we will cover directory fragmentation, multiple active metadata servers, and directory subtree pinning to metadata servers, features slated for stability in the imminent Luminous release. This talk will also give an overview of how we are measuring performance of multiple active metadata servers using large on-demand cloud deployments. The results will highlight how CephFS distributes metadata load across metadata servers to achieve scaling. About the speaker Patrick Donnelly is a software engineer at Red Hat, Inc. currently working on the Ceph distributed file system. In 2016 he completed his Ph.D. in computer science at the University of Notre Dame with a dissertation on the topic of file transfer management...

  7. The CERN analysis facility-a PROOF cluster for day-one physics analysis

    International Nuclear Information System (INIS)

    Grosse-Oetringhaus, J F

    2008-01-01

    ALICE (A Large Ion Collider Experiment) at the LHC plans to use a PROOF cluster at CERN (CAF - CERN Analysis Facility) for analysis. The system is especially aimed at the prototyping phase of analyses that need a high number of development iterations and thus require a short response time. Typical examples are the tuning of cuts during the development of an analysis as well as calibration and alignment. Furthermore, the use of an interactive system with very fast response will allow ALICE to extract physics observables out of first data quickly. An additional use case is fast event simulation and reconstruction. A test setup consisting of 40 machines is used for evaluation since May 2006. The PROOF system enables the parallel processing and xrootd the access to files distributed on the test cluster. An automatic staging system for files either catalogued in the ALICE file catalog or stored in the CASTOR mass storage system has been developed. The current setup and ongoing development towards disk quotas and CPU fairshare are described. Furthermore, the integration of PROOF into ALICE's software framework (AliRoot) is discussed

  8. 76 FR 70651 - Fee for Filing a Patent Application Other Than by the Electronic Filing System

    Science.gov (United States)

    2011-11-15

    ... government; or (3) preempt tribal law. Therefore, a tribal summary impact statement is not required under... 0651-AC64 Fee for Filing a Patent Application Other Than by the Electronic Filing System AGENCY: United..., that is not filed by electronic means as prescribed by the Director of the United States Patent and...

  9. Automatic management software for large-scale cluster system

    International Nuclear Information System (INIS)

    Weng Yunjian; Chinese Academy of Sciences, Beijing; Sun Gongxing

    2007-01-01

    At present, the large-scale cluster system faces to the difficult management. For example the manager has large work load. It needs to cost much time on the management and the maintenance of large-scale cluster system. The nodes in large-scale cluster system are very easy to be chaotic. Thousands of nodes are put in big rooms so that some managers are very easy to make the confusion with machines. How do effectively carry on accurate management under the large-scale cluster system? The article introduces ELFms in the large-scale cluster system. Furthermore, it is proposed to realize the large-scale cluster system automatic management. (authors)

  10. Evolution of clustered storage

    CERN Multimedia

    CERN. Geneva; Van de Vyvre, Pierre

    2007-01-01

    The session actually featured two presentations: * Evolution of clustered storage by Lance Hukill, Quantum Corporation * ALICE DAQ - Usage of a Cluster-File System: Quantum StorNext by Pierre Vande Vyvre, CERN-PH the second one prepared at short notice by Pierre (thanks!) to present how the Quantum technologies are being used in the ALICE experiment. The abstract to Mr Hukill's follows. Clustered Storage is a technology that is driven by business and mission applications. The evolution of Clustered Storage solutions starts first at the alignment between End-users needs and Industry trends: * Push-and-Pull between managing for today versus planning for tomorrow * Breaking down the real business problems to the core applications * Commoditization of clients, servers, and target devices * Interchangeability, Interoperability, Remote Access, Centralized control * Oh, and yes, there is a budget and the "real world" to deal with This presentation will talk through these needs and trends, and then ask the question, ...

  11. The Application of Clustering Techniques to Citation Data. Research Reports Series B No. 6.

    Science.gov (United States)

    Arms, William Y.; Arms, Caroline

    This report describes research carried out as part of the Design of Information Systems in the Social Sciences (DISISS) project. Cluster analysis techniques were applied to a machine readable file of bibliographic data in the form of cited journal titles in order to identify groupings which could be used to structure bibliographic files. Practical…

  12. File System Virtual Appliances

    Science.gov (United States)

    2010-05-01

    4 KB of data is read or written, data is copied back and forth using trampoline buffers — pages that are shared during proxy initialization — because...in 2008. CIO Magazine. 104 · File system virtual appliances [64] Megiddo, N. and Modha, D. S. 2003. ARC: A Self-Tuning, Low Over- head Replacement

  13. What Is PG-TRAK90? An Introduction to PGCC's Lifestyle Cluster System for Student Recruitment Targeting and Enrollment Analysis. Market Analysis MA94-1.

    Science.gov (United States)

    Boughan, Karl

    PG-TRAK90 is a cluster-based geographic marketing system designed by Maryland's Prince George's Community College (PGCC) to maximize educational marketing objectives. To create it, United States Census Bureau files containing over 200 demographic, housing, and lifecycle variables for 172 tracts in Prince George County (PGC) were reformatted into…

  14. AliEnFS - a Linux File System for the AliEn Grid Services

    OpenAIRE

    Peters, Andreas J.; Saiz, P.; Buncic, P.

    2003-01-01

    Among the services offered by the AliEn (ALICE Environment http://alien.cern.ch) Grid framework there is a virtual file catalogue to allow transparent access to distributed data-sets using various file transfer protocols. $alienfs$ (AliEn File System) integrates the AliEn file catalogue as a new file system type into the Linux kernel using LUFS, a hybrid user space file system framework (Open Source http://lufs.sourceforge.net). LUFS uses a special kernel interface level called VFS (Virtual F...

  15. The Future of the Andrew File System

    CERN Multimedia

    CERN. Geneva; Altman, Jeffrey

    2011-01-01

    The talk will discuss the ten operational capabilities that have made AFS unique in the distributed file system space and how these capabilities are being expanded upon to meet the needs of the 21st century. Derrick Brashear and Jeffrey Altman will present a technical road map of new features and technical innovations that are under development by the OpenAFS community and Your File System, Inc. funded by a U.S. Department of Energy Small Business Innovative Research grant. The talk will end with a comparison of AFS to its modern days competitors.

  16. Configuration Management File Manager Developed for Numerical Propulsion System Simulation

    Science.gov (United States)

    Follen, Gregory J.

    1997-01-01

    One of the objectives of the High Performance Computing and Communication Project's (HPCCP) Numerical Propulsion System Simulation (NPSS) is to provide a common and consistent way to manage applications, data, and engine simulations. The NPSS Configuration Management (CM) File Manager integrated with the Common Desktop Environment (CDE) window management system provides a common look and feel for the configuration management of data, applications, and engine simulations for U.S. engine companies. In addition, CM File Manager provides tools to manage a simulation. Features include managing input files, output files, textual notes, and any other material normally associated with simulation. The CM File Manager includes a generic configuration management Application Program Interface (API) that can be adapted for the configuration management repositories of any U.S. engine company.

  17. Storing files in a parallel computing system based on user-specified parser function

    Science.gov (United States)

    Faibish, Sorin; Bent, John M; Tzelnic, Percy; Grider, Gary; Manzanares, Adam; Torres, Aaron

    2014-10-21

    Techniques are provided for storing files in a parallel computing system based on a user-specified parser function. A plurality of files generated by a distributed application in a parallel computing system are stored by obtaining a parser from the distributed application for processing the plurality of files prior to storage; and storing one or more of the plurality of files in one or more storage nodes of the parallel computing system based on the processing by the parser. The plurality of files comprise one or more of a plurality of complete files and a plurality of sub-files. The parser can optionally store only those files that satisfy one or more semantic requirements of the parser. The parser can also extract metadata from one or more of the files and the extracted metadata can be stored with one or more of the plurality of files and used for searching for files.

  18. LHCB: Non-POSIX File System for the LHCB Online Event Handling

    CERN Multimedia

    Garnier, J-C; Cherukuwada, S S

    2010-01-01

    LHCb aims to use its O(20000) CPU cores in the High Level Trigger (HLT) and its 120 TB Online storage system for data reprocessing during LHC shutdown periods. These periods can last between a few days and several weeks during the winter shutdown or even only a few hours during beam interfill gaps. These jobs run on files which are staged in from tape storage to the local storage buffer. The result are again one or more files. Efficient file writing and reading is essential for the performance of the system. Rather than using a traditional shared filesystem such as NFS or CIFS we have implemented a custom, light-weight, non-Posix file-system for the handling of these files. Streaming this filesystem for the data-access allows to obtain high performance, while at the same time keep the resource consumption low and add nice features not found in NFS such as high-availability, transparent failover of the read and write service. The writing part of this file-system is in successful use for the Online, real-time w...

  19. NCPC Central Files Information System (CFIS)

    Data.gov (United States)

    National Capital Planning Commission — This dataset contains records from NCPC's Central Files Information System (CFIS), which is a comprehensive database of projects submitted to NCPC for design review...

  20. Properties of the open cluster system

    International Nuclear Information System (INIS)

    Janes, K.A.; Tilley, C.; Lynga, G.

    1988-01-01

    A system of weights corresponding to the precision of open cluster data is described. Using these weights, some properties of open clusters can be studied more accurately than was possible earlier. It is clear that there are three types of objects: unbound clusters, bound clusters in the thin disk, and older bound clusters. Galactic gradients of metallicity, longevity, and linear diameter are studied. Distributions at right angles to the galactic plane are discussed in the light of the different cluster types. The clumping of clusters in complexes is studied. An estimate of the selection effects influencing the present material of open cluster data is made in order to evaluate the role played by open clusters in the history of the galactic disk. 58 references

  1. Cluster computing software for GATE simulations

    International Nuclear Information System (INIS)

    Beenhouwer, Jan de; Staelens, Steven; Kruecker, Dirk; Ferrer, Ludovic; D'Asseler, Yves; Lemahieu, Ignace; Rannou, Fernando R.

    2007-01-01

    Geometry and tracking (GEANT4) is a Monte Carlo package designed for high energy physics experiments. It is used as the basis layer for Monte Carlo simulations of nuclear medicine acquisition systems in GEANT4 Application for Tomographic Emission (GATE). GATE allows the user to realistically model experiments using accurate physics models and time synchronization for detector movement through a script language contained in a macro file. The downside of this high accuracy is long computation time. This paper describes a platform independent computing approach for running GATE simulations on a cluster of computers in order to reduce the overall simulation time. Our software automatically creates fully resolved, nonparametrized macros accompanied with an on-the-fly generated cluster specific submit file used to launch the simulations. The scalability of GATE simulations on a cluster is investigated for two imaging modalities, positron emission tomography (PET) and single photon emission computed tomography (SPECT). Due to a higher sensitivity, PET simulations are characterized by relatively high data output rates that create rather large output files. SPECT simulations, on the other hand, have lower data output rates but require a long collimator setup time. Both of these characteristics hamper scalability as a function of the number of CPUs. The scalability of PET simulations is improved here by the development of a fast output merger. The scalability of SPECT simulations is improved by greatly reducing the collimator setup time. Accordingly, these two new developments result in higher scalability for both PET and SPECT simulations and reduce the computation time to more practical values

  2. Development of data file system for cardiovascular nuclear medicine

    International Nuclear Information System (INIS)

    Hayashida, Kohei; Nishimura, Tsunehiko; Uehara, Toshiisa; Nisawa, Yoshifumi.

    1985-01-01

    A computer-assisted filing system for storing and processing data from cardiac pool scintigraphy and myocardial scintigraphy has been developed. Individual patient data are stored with his (her) identification number (ID) into floppy discs successively in order of receiving scintigraphy. Data for 900 patients can be stored per floppy disc. Scintigraphic findings can be outputted in a uniform file format, and can be used as a reporting format. Output or retrieval of filed individual patient data is possible according to each examination, disease code or ID. This system seems to be used for prospective study in patients with cardiovascular diseases. (Namekawa, K.)

  3. Construction of the radiation oncology teaching files system for charged particle radiotherapy.

    Science.gov (United States)

    Masami, Mukai; Yutaka, Ando; Yasuo, Okuda; Naoto, Takahashi; Yoshihisa, Yoda; Hiroshi, Tsuji; Tadashi, Kamada

    2013-01-01

    Our hospital started the charged particle therapy since 1996. New institutions for charged particle therapy are planned in the world. Our hospital are accepting many visitors from those newly planned medical institutions and having many opportunities to provide with the training to them. Based upon our experiences, we have developed the radiation oncology teaching files system for charged particle therapy. We adopted the PowerPoint of Microsoft as a basic framework of our teaching files system. By using our export function of the viewer any physician can create teaching files easily and effectively. Now our teaching file system has 33 cases for clinical and physics contents. We expect that we can improve the safety and accuracy of charged particle therapy by using our teaching files system substantially.

  4. Screw-in forces during instrumentation by various file systems.

    Science.gov (United States)

    Ha, Jung-Hong; Kwak, Sang Won; Kim, Sung-Kyo; Kim, Hyeon-Cheol

    2016-11-01

    The purpose of this study was to compare the maximum screw-in forces generated during the movement of various Nickel-Titanium (NiTi) file systems. Forty simulated canals in resin blocks were randomly divided into 4 groups for the following instruments: Mtwo size 25/0.07 (MTW, VDW GmbH), Reciproc R25 (RPR, VDW GmbH), ProTaper Universal F2 (PTU, Dentsply Maillefer), and ProTaper Next X2 (PTN, Dentsply Maillefer, n = 10). All the artificial canals were prepared to obtain a standardized lumen by using ProTaper Universal F1. Screw-in forces were measured using a custom-made experimental device (AEndoS- k , DMJ system) during instrumentation with each NiTi file system using the designated movement. The rotation speed was set at 350 rpm with an automatic 4 mm pecking motion at a speed of 1 mm/sec. The pecking depth was increased by 1 mm for each pecking motion until the file reach the working length. Forces were recorded during file movement, and the maximum force was extracted from the data. Maximum screw-in forces were analyzed by one-way ANOVA and Tukey's post hoc comparison at a significance level of 95%. Reciproc and ProTaper Universal files generated the highest maximum screw-in forces among all the instruments while M-two and ProTaper Next showed the lowest ( p files with smaller cross-sectional area for higher flexibility is recommended.

  5. Distributed Data Management and Distributed File Systems

    CERN Document Server

    Girone, Maria

    2015-01-01

    The LHC program has been successful in part due to the globally distributed computing resources used for collecting, serving, processing, and analyzing the large LHC datasets. The introduction of distributed computing early in the LHC program spawned the development of new technologies and techniques to synchronize information and data between physically separated computing centers. Two of the most challenges services are the distributed file systems and the distributed data management systems. In this paper I will discuss how we have evolved from local site services to more globally independent services in the areas of distributed file systems and data management and how these capabilities may continue to evolve into the future. I will address the design choices, the motivations, and the future evolution of the computing systems used for High Energy Physics.

  6. [PVFS 2000: An operational parallel file system for Beowulf

    Science.gov (United States)

    Ligon, Walt

    2004-01-01

    The approach has been to develop Parallel Virtual File System version 2 (PVFS2) , retaining the basic philosophy of the original file system but completely rewriting the code. It shows the architecture of the server and client components. BMI - BMI is the network abstraction layer. It is designed with a common driver and modules for each protocol supported. The interface is non-blocking, and provides mechanisms for optimizations including pinning user buffers. Currently TCP/IP and GM(Myrinet) modules have been implemented. Trove -Trove is the storage abstraction layer. It provides for storing both data spaces and name/value pairs. Trove can also be implemented using different underlying storage mechanisms including native files, raw disk partitions, SQL and other databases. The current implementation uses native files for data spaces and Berkeley db for name/value pairs.

  7. Quark cluster model in the three-nucleon system

    International Nuclear Information System (INIS)

    Osman, A.

    1986-11-01

    The quark cluster model is used to investigate the structure of the three-nucleon systems. The nucleon-nucleon interaction is proposed considering the colour-nucleon clusters and incorporating the quark degrees of freedom. The quark-quark potential in the quark compound bag model agrees with the central force potentials. The confinement potential reduces the short-range repulsion. The colour van der Waals force is determined. Then, the probability of quark clusters in the three-nucleon bound state systems are numerically calculated using realistic nuclear wave functions. The results of the present calculations show that quarks cluster themselves in three-quark systems building the quark cluster model for the trinucleon system. (author)

  8. clusterMaker: a multi-algorithm clustering plugin for Cytoscape

    Directory of Open Access Journals (Sweden)

    Morris John H

    2011-11-01

    Full Text Available Abstract Background In the post-genomic era, the rapid increase in high-throughput data calls for computational tools capable of integrating data of diverse types and facilitating recognition of biologically meaningful patterns within them. For example, protein-protein interaction data sets have been clustered to identify stable complexes, but scientists lack easily accessible tools to facilitate combined analyses of multiple data sets from different types of experiments. Here we present clusterMaker, a Cytoscape plugin that implements several clustering algorithms and provides network, dendrogram, and heat map views of the results. The Cytoscape network is linked to all of the other views, so that a selection in one is immediately reflected in the others. clusterMaker is the first Cytoscape plugin to implement such a wide variety of clustering algorithms and visualizations, including the only implementations of hierarchical clustering, dendrogram plus heat map visualization (tree view, k-means, k-medoid, SCPS, AutoSOME, and native (Java MCL. Results Results are presented in the form of three scenarios of use: analysis of protein expression data using a recently published mouse interactome and a mouse microarray data set of nearly one hundred diverse cell/tissue types; the identification of protein complexes in the yeast Saccharomyces cerevisiae; and the cluster analysis of the vicinal oxygen chelate (VOC enzyme superfamily. For scenario one, we explore functionally enriched mouse interactomes specific to particular cellular phenotypes and apply fuzzy clustering. For scenario two, we explore the prefoldin complex in detail using both physical and genetic interaction clusters. For scenario three, we explore the possible annotation of a protein as a methylmalonyl-CoA epimerase within the VOC superfamily. Cytoscape session files for all three scenarios are provided in the Additional Files section. Conclusions The Cytoscape plugin cluster

  9. A technique for integrating remote minicomputers into a general computer's file system

    CERN Document Server

    Russell, R D

    1976-01-01

    This paper describes a simple technique for interfacing remote minicomputers used for real-time data acquisition into the file system of a central computer. Developed as part of the ORION system at CERN, this 'File Manager' subsystem enables a program in the minicomputer to access and manipulate files of any type as if they resided on a storage device attached to the minicomputer. Yet, completely transparent to the program, the files are accessed from disks on the central system via high-speed data links, with response times comparable to local storage devices. (6 refs).

  10. Lessons Learned in Deploying the World s Largest Scale Lustre File System

    Energy Technology Data Exchange (ETDEWEB)

    Dillow, David A [ORNL; Fuller, Douglas [ORNL; Wang, Feiyi [ORNL; Oral, H Sarp [ORNL; Zhang, Zhe [ORNL; Hill, Jason J [ORNL; Shipman, Galen M [ORNL

    2010-01-01

    The Spider system at the Oak Ridge National Laboratory's Leadership Computing Facility (OLCF) is the world's largest scale Lustre parallel file system. Envisioned as a shared parallel file system capable of delivering both the bandwidth and capacity requirements of the OLCF's diverse computational environment, the project had a number of ambitious goals. To support the workloads of the OLCF's diverse computational platforms, the aggregate performance and storage capacity of Spider exceed that of our previously deployed systems by a factor of 6x - 240 GB/sec, and 17x - 10 Petabytes, respectively. Furthermore, Spider supports over 26,000 clients concurrently accessing the file system, which exceeds our previously deployed systems by nearly 4x. In addition to these scalability challenges, moving to a center-wide shared file system required dramatically improved resiliency and fault-tolerance mechanisms. This paper details our efforts in designing, deploying, and operating Spider. Through a phased approach of research and development, prototyping, deployment, and transition to operations, this work has resulted in a number of insights into large-scale parallel file system architectures, from both the design and the operational perspectives. We present in this paper our solutions to issues such as network congestion, performance baselining and evaluation, file system journaling overheads, and high availability in a system with tens of thousands of components. We also discuss areas of continued challenges, such as stressed metadata performance and the need for file system quality of service alongside with our efforts to address them. Finally, operational aspects of managing a system of this scale are discussed along with real-world data and observations.

  11. The use of a VAX cluster for the DIII-D data acquisition system

    International Nuclear Information System (INIS)

    McHarg, B.B. Jr.

    1992-01-01

    This paper reports on the DIII-D tokamak, a large fusion energy research experiment funded by the Department of Energy. The experiment currently collects nearly 40 Mbytes of data from each shot of the experiment. In the past, most of this data was acquired through the MODCOMP Classic data acquisition computers and then transferred to a DEC VAX computer system for permanent archiving and storage. A much smaller amount of data was acquired from a few Micro VAX based data acquisition systems. In the last two years, MicroVAX based systems have become the standard means for adding new diagnostic data and account for half the total data. There are now 17 VAX systems of various types at the DIII-D facility. As more diagnostics and data are added, it takes increasing amounts of time to merge the data into the central shot file. The system management of so many systems has become increasingly time consuming as well. To improve the efficiency of the overall data acquisition system, a mixed interconnect VAX cluster has been formed consisting of 16 VAX computers

  12. Parallel file system with metadata distributed across partitioned key-value store c

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary; Torres, Aaron

    2017-09-19

    Improved techniques are provided for storing metadata associated with a plurality of sub-files associated with a single shared file in a parallel file system. The shared file is generated by a plurality of applications executing on a plurality of compute nodes. A compute node implements a Parallel Log Structured File System (PLFS) library to store at least one portion of the shared file generated by an application executing on the compute node and metadata for the at least one portion of the shared file on one or more object storage servers. The compute node is also configured to implement a partitioned data store for storing a partition of the metadata for the shared file, wherein the partitioned data store communicates with partitioned data stores on other compute nodes using a message passing interface. The partitioned data store can be implemented, for example, using Multidimensional Data Hashing Indexing Middleware (MDHIM).

  13. Stochastic Petri net analysis of a replicated file system

    Science.gov (United States)

    Bechta Dugan, Joanne; Ciardo, Gianfranco

    1989-01-01

    A stochastic Petri-net model of a replicated file system is presented for a distributed environment where replicated files reside on different hosts and a voting algorithm is used to maintain consistency. Witnesses, which simply record the status of the file but contain no data, can be used in addition to or in place of files to reduce overhead. A model sufficiently detailed to include file status (current or out-of-date), as well as failure and repair of hosts where copies or witnesses reside, is presented. The number of copies and witnesses is a parameter of the model. Two different majority protocols are examined, one where a majority of all copies and witnesses is necessary to form a quorum, and the other where only a majority of the copies and witnesses on operational hosts is needed. The latter, known as adaptive voting, is shown to increase file availability in most cases.

  14. A History of the Andrew File System

    CERN Multimedia

    CERN. Geneva; Altman, Jeffrey

    2011-01-01

    Derrick Brashear and Jeffrey Altman will present a technical history of the evolution of Andrew File System starting with the early days of the Andrew Project at Carnegie Mellon through the commercialization by Transarc Corporation and IBM and a decade of OpenAFS. The talk will be technical with a focus on the various decisions and implementation trade-offs that were made over the course of AFS versions 1 through 4, the development of the Distributed Computing Environment Distributed File System (DCE DFS), and the course of the OpenAFS development community. The speakers will also discuss the various AFS branches developed at the University of Michigan, Massachusetts Institute of Technology and Carnegie Mellon University.

  15. Generalized File Management System or Proto-DBMS?

    Science.gov (United States)

    Braniff, Tom

    1979-01-01

    The use of a data base management system (DBMS) as opposed to traditional data processing is discussed. The generalized file concept is viewed as an entry level step to the DBMS. The transition process from one system to the other is detailed. (SF)

  16. Supersymmetry for nuclear cluster systems

    International Nuclear Information System (INIS)

    Levai, G.; Cseh, J.; Van Isacker, P.

    2001-01-01

    A supersymmetry scheme is proposed for nuclear cluster systems. The bosonic sector of the superalgebra describes the relative motion of the clusters, while its fermionic sector is associated with their internal structure. An example of core+α configurations is discussed in which the core is a p-shell nucleus and the underlying superalgebra is U(4/12). The α-cluster states of the nuclei 20 Ne and 19 F are analysed and correlations between their spectra, electric quadrupole transitions, and one-nucleon transfer reactions are interpreted in terms of U(4/12) supersymmetry. (author)

  17. Evaluation of canal transportation after preparation with Reciproc single-file systems with or without glide path files.

    Science.gov (United States)

    Aydin, Ugur; Karataslioglu, Emrah

    2017-01-01

    Canal transportation is a common sequel caused by rotary instruments. The purpose of the present study is to evaluate the degree of transportation after the use of Reciproc single-file instruments with or without glide path files. Thirty resin blocks with L-shaped canals were divided into three groups ( n = 10). Group 1 - canals were prepared with Reciproc-25 file. Group 2 - glide path file-G1 was used before Reciproc. Group 3 - glide path files-G1 and G2 were used before Reciproc. Pre- and post-instrumentation images were superimposed under microscope, and resin removed from the inner and outer surfaces of the root canal was calculated throughout 10 points. Statistical analysis was performed with Kruskal-Wallis test and post hoc Dunn test. For coronal and middle one-thirds, there was no significant difference among groups ( P > 0.05). For apical section, transportation of Group 1 was significantly higher than other groups ( P files before Reciproc single-file system reduced the degree of apical canal transportation.

  18. 78 FR 63159 - Amendment to Certification of Nebraska's Central Filing System

    Science.gov (United States)

    2013-10-23

    ..., Raspberries, Rye, Seed crops, Sheep & Lambs, Silage, Sorghum Grain, Soybeans, Squash, Strawberries, Sugar....usda.gov ). Farm products covered by a State's central filing system are also identified through the GIPSA Web site. The Nebraska central filing system covers specified farm products. We originally...

  19. Apical extrusion of debris during the preparation of oval root canals: a comparative study between a full-sequence SAF system and a rotary file system supplemented by XP-endo finisher file.

    Science.gov (United States)

    Kfir, Anda; Moza-Levi, Rotem; Herteanu, Moran; Weissman, Amir; Wigler, Ronald

    2018-03-01

    The purpose of this study was to assess the amount of apically extruded debris during the preparation of oval canals with either a rotary file system supplemented by the XP-endo Finisher file or a full-sequence self-adjusting file (SAF) system. Sixty mandibular incisors were randomly assigned to two groups: group A: stage 1-glide path preparation with Pre-SAF instruments. Stage 2-cleaning and shaping with SAF. Group B: stage 1-glide path preparation with ProGlider file. Stage 2-cleaning and shaping with ProTaper Next system. Stage 3-Final cleaning with XP-endo Finisher file. The debris extruded during each of the stages was collected, and the debris weights were compared between the groups and between the stages within the groups using t tests with a significance level set at P file followed by XP-endo Finisher file extruded significantly more debris than a full-sequence SAF system. Each stage, in either procedure, had its own contribution to the extrusion of debris. Final preparation with XP-endo Finisher file contributes to the total amount of extruded debris, but the clinical relevance of the relative difference in the amount of apically extruded debris remains unclear.

  20. Recommending the heterogeneous cluster type multi-processor system computing

    International Nuclear Information System (INIS)

    Iijima, Nobukazu

    2010-01-01

    Real-time reactor simulator had been developed by reusing the equipment of the Musashi reactor and its performance improvement became indispensable for research tools to increase sampling rate with introduction of arithmetic units using multi-Digital Signal Processor(DSP) system (cluster). In order to realize the heterogeneous cluster type multi-processor system computing, combination of two kinds of Control Processor (CP) s, Cluster Control Processor (CCP) and System Control Processor (SCP), were proposed with Large System Control Processor (LSCP) for hierarchical cluster if needed. Faster computing performance of this system was well evaluated by simulation results for simultaneous execution of plural jobs and also pipeline processing between clusters, which showed the system led to effective use of existing system and enhancement of the cost performance. (T. Tanaka)

  1. Beyond a Terabyte File System

    Science.gov (United States)

    Powers, Alan K.

    1994-01-01

    The Numerical Aerodynamics Simulation Facility's (NAS) CRAY C916/1024 accesses a "virtual" on-line file system, which is expanding beyond a terabyte of information. This paper will present some options to fine tuning Data Migration Facility (DMF) to stretch the online disk capacity and explore the transitions to newer devices (STK 4490, ER90, RAID).

  2. Branch file system for nonconventional literature from the nuclear field

    International Nuclear Information System (INIS)

    Dvorakova, K.

    1982-01-01

    The branch filing system collects research and study reports, translations, trip reports, literature searches and information on scientific and technical events in Czechoslovakia. The method is described of filing, processing and use of the materials. (M.D.)

  3. 5 CFR 293.504 - Composition of, and access to, the Employee Medical File System.

    Science.gov (United States)

    2010-01-01

    ... Employee Medical File System. 293.504 Section 293.504 Administrative Personnel OFFICE OF PERSONNEL MANAGEMENT CIVIL SERVICE REGULATIONS PERSONNEL RECORDS Employee Medical File System Records § 293.504 Composition of, and access to, the Employee Medical File System. (a) All employee occupational medical records...

  4. Q-systems as cluster algebras

    International Nuclear Information System (INIS)

    Kedem, Rinat

    2008-01-01

    Q-systems first appeared in the analysis of the Bethe equations for the XXX model and generalized Heisenberg spin chains (Kirillov and Reshetikhin 1987 Zap. Nauchn. Sem. Leningr. Otd. Mat. Inst. Steklov. 160 211-21, 301). Such systems are known to exist for any simple Lie algebra and many other Kac-Moody algebras. We formulate the Q-system associated with any simple, simply-laced Lie algebras g in the language of cluster algebras (Fomin and Zelevinsky 2002 J. Am. Math. Soc. 15 497-529), and discuss the relation of the polynomiality property of the solutions of the Q-system in the initial variables, which follows from the representation-theoretical interpretation, to the Laurent phenomenon in cluster algebras (Fomin and Zelevinsky 2002 Adv. Appl. Math. 28 119-44)

  5. Evolutionary Game Theory-Based Evaluation of P2P File-Sharing Systems in Heterogeneous Environments

    Directory of Open Access Journals (Sweden)

    Yusuke Matsuda

    2010-01-01

    Full Text Available Peer-to-Peer (P2P file sharing is one of key technologies for achieving attractive P2P multimedia social networking. In P2P file-sharing systems, file availability is improved by cooperative users who cache and share files. Note that file caching carries costs such as storage consumption and processing load. In addition, users have different degrees of cooperativity in file caching and they are in different surrounding environments arising from the topological structure of P2P networks. With evolutionary game theory, this paper evaluates the performance of P2P file sharing systems in such heterogeneous environments. Using micro-macro dynamics, we analyze the impact of the heterogeneity of user selfishness on the file availability and system stability. Further, through simulation experiments with agent-based dynamics, we reveal how other aspects, for example, synchronization among nodes and topological structure, affect the system performance. Both analytical and simulation results show that the environmental heterogeneity contributes to the file availability and system stability.

  6. Electronic Document Management Using Inverted Files System

    Science.gov (United States)

    Suhartono, Derwin; Setiawan, Erwin; Irwanto, Djon

    2014-03-01

    The amount of documents increases so fast. Those documents exist not only in a paper based but also in an electronic based. It can be seen from the data sample taken by the SpringerLink publisher in 2010, which showed an increase in the number of digital document collections from 2003 to mid of 2010. Then, how to manage them well becomes an important need. This paper describes a new method in managing documents called as inverted files system. Related with the electronic based document, the inverted files system will closely used in term of its usage to document so that it can be searched over the Internet using the Search Engine. It can improve document search mechanism and document save mechanism.

  7. Electronic Document Management Using Inverted Files System

    Directory of Open Access Journals (Sweden)

    Suhartono Derwin

    2014-03-01

    Full Text Available The amount of documents increases so fast. Those documents exist not only in a paper based but also in an electronic based. It can be seen from the data sample taken by the SpringerLink publisher in 2010, which showed an increase in the number of digital document collections from 2003 to mid of 2010. Then, how to manage them well becomes an important need. This paper describes a new method in managing documents called as inverted files system. Related with the electronic based document, the inverted files system will closely used in term of its usage to document so that it can be searched over the Internet using the Search Engine. It can improve document search mechanism and document save mechanism.

  8. Architecture for Automated Tagging and Clustering of Song Files According to Mood

    OpenAIRE

    Singh, Puneet; Kapoor, Ashutosh; Kaushik, Vishal; Maringanti, Hima Bindu

    2012-01-01

    Music is one of the basic human needs for recreation and entertainment. As song files are digitalized now a days, and digital libraries are expanding continuously, which makes it difficult to recall a song. Thus need of a new classification system other than genre is very obvious and mood based classification system serves the purpose very well. In this paper we will present a well-defined architecture to classify songs into different mood-based categories, using audio content analysis, affec...

  9. 29 CFR 4902.11 - Specific exemptions: Office of Inspector General Investigative File System.

    Science.gov (United States)

    2010-07-01

    ... Investigative File System. 4902.11 Section 4902.11 Labor Regulations Relating to Labor (Continued) PENSION... General Investigative File System. (a) Criminal Law Enforcement. (1) Exemption. Under the authority... Inspector General Investigative File System—PBGC” from the provisions of 5 U.S.C. 552a (c)(3), (c)(4), (d)(1...

  10. Maintaining a Distributed File System by Collection and Analysis of Metrics

    Science.gov (United States)

    Bromberg, Daniel

    1997-01-01

    AFS(originally, Andrew File System) is a widely-deployed distributed file system product used by companies, universities, and laboratories world-wide. However, it is not trivial to operate: runing an AFS cell is a formidable task. It requires a team of dedicated and experienced system administratores who must manage a user base numbring in the thousands, rather than the smaller range of 10 to 500 faced by the typical system administrator.

  11. NASIS data base management system: IBM 360 TSS implementation. Volume 6: NASIS message file

    Science.gov (United States)

    1973-01-01

    The message file for the NASA Aerospace Safety Information System (NASIS) is discussed. The message file contains all the message and term explanations for the system. The data contained in the file can be broken down into three separate sections: (1) global terms, (2) local terms, and (3) system messages. The various terms are defined and their use within the system is explained.

  12. The self-adjusting file (SAF) system: An evidence-based update.

    Science.gov (United States)

    Metzger, Zvi

    2014-09-01

    Current rotary file systems are effective tools. Nevertheless, they have two main shortcomings: They are unable to effectively clean and shape oval canals and depend too much on the irrigant to do the cleaning, which is an unrealistic illusionThey may jeopardize the long-term survival of the tooth via unnecessary, excessive removal of sound dentin and creation of micro-cracks in the remaining root dentin. The new Self-adjusting File (SAF) technology uses a hollow, compressible NiTi file, with no central metal core, through which a continuous flow of irrigant is provided throughout the procedure. The SAF technology allows for effective cleaning of all root canals including oval canals, thus allowing for the effective disinfection and obturation of all canal morphologies. This technology uses a new concept of cleaning and shaping in which a uniform layer of dentin is removed from around the entire perimeter of the root canal, thus avoiding unnecessary excessive removal of sound dentin. Furthermore, the mode of action used by this file system does not apply the machining of all root canals to a circular bore, as do all other rotary file systems, and does not cause micro-cracks in the remaining root dentin. The new SAF technology allows for a new concept in cleaning and shaping root canals: Minimally Invasive 3D Endodontics.

  13. The self-adjusting file (SAF) system: An evidence-based update

    Science.gov (United States)

    Metzger, Zvi

    2014-01-01

    Current rotary file systems are effective tools. Nevertheless, they have two main shortcomings: They are unable to effectively clean and shape oval canals and depend too much on the irrigant to do the cleaning, which is an unrealistic illusionThey may jeopardize the long-term survival of the tooth via unnecessary, excessive removal of sound dentin and creation of micro-cracks in the remaining root dentin. The new Self-adjusting File (SAF) technology uses a hollow, compressible NiTi file, with no central metal core, through which a continuous flow of irrigant is provided throughout the procedure. The SAF technology allows for effective cleaning of all root canals including oval canals, thus allowing for the effective disinfection and obturation of all canal morphologies. This technology uses a new concept of cleaning and shaping in which a uniform layer of dentin is removed from around the entire perimeter of the root canal, thus avoiding unnecessary excessive removal of sound dentin. Furthermore, the mode of action used by this file system does not apply the machining of all root canals to a circular bore, as do all other rotary file systems, and does not cause micro-cracks in the remaining root dentin. The new SAF technology allows for a new concept in cleaning and shaping root canals: Minimally Invasive 3D Endodontics. PMID:25298639

  14. NASA ARCH- A FILE ARCHIVAL SYSTEM FOR THE DEC VAX

    Science.gov (United States)

    Scott, P. J.

    1994-01-01

    The function of the NASA ARCH system is to provide a permanent storage area for files that are infrequently accessed. The NASA ARCH routines were designed to provide a simple mechanism by which users can easily store and retrieve files. The user treats NASA ARCH as the interface to a black box where files are stored. There are only five NASA ARCH user commands, even though NASA ARCH employs standard VMS directives and the VAX BACKUP utility. Special care is taken to provide the security needed to insure file integrity over a period of years. The archived files may exist in any of three storage areas: a temporary buffer, the main buffer, and a magnetic tape library. When the main buffer fills up, it is transferred to permanent magnetic tape storage and deleted from disk. Files may be restored from any of the three storage areas. A single file, multiple files, or entire directories can be stored and retrieved. archived entities hold the same name, extension, version number, and VMS file protection scheme as they had in the user's account prior to archival. NASA ARCH is capable of handling up to 7 directory levels. Wildcards are supported. User commands include TEMPCOPY, DISKCOPY, DELETE, RESTORE, and DIRECTORY. The DIRECTORY command searches a directory of savesets covering all three archival areas, listing matches according to area, date, filename, or other criteria supplied by the user. The system manager commands include 1) ARCHIVE- to transfer the main buffer to duplicate magnetic tapes, 2) REPORTto determine when the main buffer is full enough to archive, 3) INCREMENT- to back up the partially filled main buffer, and 4) FULLBACKUP- to back up the entire main buffer. On-line help files are provided for all NASA ARCH commands. NASA ARCH is written in DEC VAX DCL for interactive execution and has been implemented on a DEC VAX computer operating under VMS 4.X. This program was developed in 1985.

  15. Clustering execution in a processing system to increase power savings

    Energy Technology Data Exchange (ETDEWEB)

    Bose, Pradip; Buyuktosunoglu, Alper; Jacobson, Hans M.; Vega, Augusto J.

    2018-04-03

    Embodiments relate to clustering execution in a processing system. An aspect includes accessing a control flow graph that defines a data dependency and an execution sequence of a plurality of tasks of an application that executes on a plurality of system components. The execution sequence of the tasks in the control flow graph is modified as a clustered control flow graph that clusters active and idle phases of a system component while maintaining the data dependency. The clustered control flow graph is sent to an operating system, where the operating system utilizes the clustered control flow graph for scheduling the tasks.

  16. Clustering execution in a processing system to increase power savings

    Science.gov (United States)

    Bose, Pradip; Buyuktosunoglu, Alper; Jacobson, Hans M.; Vega, Augusto J.

    2018-03-20

    Embodiments relate to clustering execution in a processing system. An aspect includes accessing a control flow graph that defines a data dependency and an execution sequence of a plurality of tasks of an application that executes on a plurality of system components. The execution sequence of the tasks in the control flow graph is modified as a clustered control flow graph that clusters active and idle phases of a system component while maintaining the data dependency. The clustered control flow graph is sent to an operating system, where the operating system utilizes the clustered control flow graph for scheduling the tasks.

  17. PHOBINS: an index file of photon production cross section data and its utility code system

    International Nuclear Information System (INIS)

    Hasegawa, Akira; Koyama, Kinji; Ido, Masaru; Hotta, Masakazu; Miyasaka, Shun-ichi

    1978-08-01

    The code System PHOBINS developed for reference of photon production cross sections is described in detail. The system is intended to grasp the present status of photon production data and present the information of available data. It consists of four utility routines, CREA, UP-DT, REF and BACK, and data files. These utility routines are used for making an index file of the photon production cross sections, updating the index file, searching the index file and producing a back-up file of the index file. In the index file of the photon production cross sections, a data base system is employed for efficient data management in economical storage, ease of updating and efficient reference. The present report is a reference manual of PHOBINS. (author)

  18. Dynamically Allocated Virtual Clustering Management System Users Guide

    Science.gov (United States)

    2016-11-01

    ARL-SR-0366 ● NOV 2016 US Army Research Laboratory Dynamically Allocated Virtual Clustering Management System User’s Guide by... Clustering Management System User’s Guide by Kelvin M Marcus Computational and Information Sciences Directorate, ARL...

  19. THE HST/ACS COMA CLUSTER SURVEY. IV. INTERGALACTIC GLOBULAR CLUSTERS AND THE MASSIVE GLOBULAR CLUSTER SYSTEM AT THE CORE OF THE COMA GALAXY CLUSTER

    International Nuclear Information System (INIS)

    Peng, Eric W.; Ferguson, Henry C.; Goudfrooij, Paul; Hammer, Derek; Lucey, John R.; Marzke, Ronald O.; Puzia, Thomas H.; Carter, David; Balcells, Marc; Bridges, Terry; Chiboucas, Kristin; Del Burgo, Carlos; Graham, Alister W.; Guzman, Rafael; Hudson, Michael J.; Matkovic, Ana

    2011-01-01

    Intracluster stellar populations are a natural result of tidal interactions in galaxy clusters. Measuring these populations is difficult, but important for understanding the assembly of the most massive galaxies. The Coma cluster of galaxies is one of the nearest truly massive galaxy clusters and is host to a correspondingly large system of globular clusters (GCs). We use imaging from the HST/ACS Coma Cluster Survey to present the first definitive detection of a large population of intracluster GCs (IGCs) that fills the Coma cluster core and is not associated with individual galaxies. The GC surface density profile around the central massive elliptical galaxy, NGC 4874, is dominated at large radii by a population of IGCs that extend to the limit of our data (R +4000 -5000 (systematic) IGCs out to this radius, and that they make up ∼70% of the central GC system, making this the largest GC system in the nearby universe. Even including the GC systems of other cluster galaxies, the IGCs still make up ∼30%-45% of the GCs in the cluster core. Observational limits from previous studies of the intracluster light (ICL) suggest that the IGC population has a high specific frequency. If the IGC population has a specific frequency similar to high-S N dwarf galaxies, then the ICL has a mean surface brightness of μ V ∼ 27 mag arcsec -2 and a total stellar mass of roughly 10 12 M sun within the cluster core. The ICL makes up approximately half of the stellar luminosity and one-third of the stellar mass of the central (NGC 4874+ICL) system. The color distribution of the IGC population is bimodal, with blue, metal-poor GCs outnumbering red, metal-rich GCs by a ratio of 4:1. The inner GCs associated with NGC 4874 also have a bimodal distribution in color, but with a redder metal-poor population. The fraction of red IGCs (20%), and the red color of those GCs, implies that IGCs can originate from the halos of relatively massive, L* galaxies, and not solely from the disruption of

  20. Shaping ability of 4 different single-file systems in simulated S-shaped canals.

    Science.gov (United States)

    Saleh, Abdulrahman Mohammed; Vakili Gilani, Pouyan; Tavanafar, Saeid; Schäfer, Edgar

    2015-04-01

    The aim of this study was to compare the shaping ability of 4 different single-file systems in simulated S-shaped canals. Sixty-four S-shaped canals in resin blocks were prepared to an apical size of 25 using Reciproc (VDW, Munich, Germany), WaveOne (Dentsply Maillefer, Ballaigues, Switzerland), OneShape (Micro Méga, Besançon, France), and F360 (Komet Brasseler, Lemgo, Germany) (n = 16 canals/group) systems. Composite images were made from the superimposition of pre- and postinstrumentation images. The amount of resin removed by each system was measured by using a digital template and image analysis software. Canal aberrations and the preparation time were also recorded. The data were statistically analyzed by using analysis of variance, Tukey, and chi-square tests. Canals prepared with the F360 and OneShape systems were better centered compared with the Reciproc and WaveOne systems. Reciproc and WaveOne files removed significantly greater amounts of resin from the inner side of both curvatures (P files was significantly faster compared with WaveOne and F360 files (P file instruments were safe to use and were able to prepare the canals efficiently. However, single-file systems that are less tapered seem to be more favorable when preparing S-shaped canals. Copyright © 2015 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  1. The European Southern Observatory-MIDAS table file system

    Science.gov (United States)

    Peron, M.; Grosbol, P.

    1992-01-01

    The new and substantially upgraded version of the Table File System in MIDAS is presented as a scientific database system. MIDAS applications for performing database operations on tables are discussed, for instance, the exchange of the data to and from the TFS, the selection of objects, the uncertainty joins across tables, and the graphical representation of data. This upgraded version of the TFS is a full implementation of the binary table extension of the FITS format; in addition, it also supports arrays of strings. Different storage strategies for optimal access of very large data sets are implemented and are addressed in detail. As a simple relational database, the TFS may be used for the management of personal data files. This opens the way to intelligent pipeline processing of large amounts of data. One of the key features of the Table File System is to provide also an extensive set of tools for the analysis of the final results of a reduction process. Column operations using standard and special mathematical functions as well as statistical distributions can be carried out; commands for linear regression and model fitting using nonlinear least square methods and user-defined functions are available. Finally, statistical tests of hypothesis and multivariate methods can also operate on tables.

  2. Methods and apparatus for multi-resolution replication of files in a parallel computing system using semantic information

    Science.gov (United States)

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Grider, Gary; Torres, Aaron

    2015-10-20

    Techniques are provided for storing files in a parallel computing system using different resolutions. A method is provided for storing at least one file generated by a distributed application in a parallel computing system. The file comprises one or more of a complete file and a sub-file. The method comprises the steps of obtaining semantic information related to the file; generating a plurality of replicas of the file with different resolutions based on the semantic information; and storing the file and the plurality of replicas of the file in one or more storage nodes of the parallel computing system. The different resolutions comprise, for example, a variable number of bits and/or a different sub-set of data elements from the file. A plurality of the sub-files can be merged to reproduce the file.

  3. Development of fibre channel disk clusters. Final report for period September 2, 1998 - March 17, 1999

    International Nuclear Information System (INIS)

    Dunn, W.L.; Justice, J.R.; Stockert, T.D.; Barker, A.R.; Yacout, A.M.

    1999-01-01

    This report documents the accomplishments of a Phase I project whose purpose was to demonstrate feasibility of developing inexpensive and fast data storage using multi-host Fibre Channel disk clusters. In Phase I, a working file system called ZFS was developed and tested. The ZFS approach was designed to be suited for high energy physics applications, but is general and flexible enough to be useful for other high-volume applications. The ZFS approach, which borrows from the networking concept of cut-through routing, uses Linux boxes and disk clusters in a Fibre Channel--Arbitrated Loop architecture. In ZFS, file locking and other meta-data level operations are carried out over the primary data network, after which all data are sent directly over a Fibre Channel between the workstation and the disk cluster. No intermediate server is required. Substantially higher throughputs than in traditional networked disk architectures have been demonstrated. The ZFS architecture is described and tests of the first implementation of ZFS at Fermilab are discussed. The current system is implemented for Linux and is being optimized for Fermilab's needs, but extensions to other operating systems and other data-intensive applications are clearly foreseen

  4. Subspace identification of distributed clusters of homogeneous systems

    NARCIS (Netherlands)

    Yu, C.; Verhaegen, M.H.G.

    2017-01-01

    This note studies the identification of a network comprised of interconnected clusters of LTI systems. Each cluster consists of homogeneous dynamical systems, and its interconnections with the rest of the network are unmeasurable. A subspace identification method is proposed for identifying a single

  5. Temperature increases on the external root surface during endodontic treatment using single file systems.

    Science.gov (United States)

    Özkocak, I; Taşkan, M M; Gökt Rk, H; Aytac, F; Karaarslan, E Şirin

    2015-01-01

    The aim of this study is to evaluate increases in temperature on the external root surface during endodontic treatment with different rotary systems. Fifty human mandibular incisors with a single root canal were selected. All root canals were instrumented using a size 20 Hedstrom file, and the canals were irrigated with 5% sodium hypochlorite solution. The samples were randomly divided into the following three groups of 15 teeth: Group 1: The OneShape Endodontic File no.: 25; Group 2: The Reciproc Endodontic File no.: 25; Group 3: The WaveOne Endodontic File no.: 25. During the preparation, the temperature changes were measured in the middle third of the roots using a noncontact infrared thermometer. The temperature data were transferred from the thermometer to the computer and were observed graphically. Statistical analysis was performed using the Kruskal-Wallis analysis of variance at a significance level of 0.05. The increases in temperature caused by the OneShape file system were lower than those of the other files (P file showed the highest temperature increases. However, there were no significant differences between the Reciproc and WaveOne files. The single file rotary systems used in this study may be recommended for clinical use.

  6. NASIS data base management system - IBM 360/370 OS MVT implementation. 6: NASIS message file

    Science.gov (United States)

    1973-01-01

    The message file for the NASA Aerospace Safety Information System (NASIS) is discussed. The message file contains all the message and term explanations for the system. The data contained in the file can be broken down into three separate sections: (1) global terms, (2) local terms, and (3) system messages. The various terms are defined and their use within the system is explained.

  7. Speaker Segmentation and Clustering Using Gender Information

    Science.gov (United States)

    2006-02-01

    used in the first stages of segmentation forder information in the clustering of the opposite-gender speaker diarization of news broadcasts. files, the...AFRL-HE-WP-TP-2006-0026 AIR FORCE RESEARCH LABORATORY Speaker Segmentation and Clustering Using Gender Information Brian M. Ore General Dynamics...COVERED (From - To) February 2006 ProceedinLgs 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER Speaker Segmentation and Clustering Using Gender Information 5b

  8. Rack Aware Data Placement for Network Consumption in Erasure-Coded Clustered Storage Systems

    Directory of Open Access Journals (Sweden)

    Bilin Shao

    2018-06-01

    Full Text Available The amount of encoded data replication in an erasure-coded clustered storage system has a great impact on the bandwidth consumption and network latency, mostly during data reconstruction. Aimed at the reasons that lead to the excess data transmission between racks, a rack aware data block placement method is proposed. In order to ensure rack-level fault tolerance and reduce the frequency and amount of the cross-rack data transmission during data reconstruction, the method deploys partial data block concentration to store the data blocks of a file in fewer racks. Theoretical analysis and simulation results show that our proposed strategy greatly reduces the frequency and data volume of the cross-rack transmission during data reconstruction. At the same time, it has better performance than the typical random distribution method in terms of network usage and data reconstruction efficiency.

  9. Assessment of apically extruded debris produced by the self-adjusting file system.

    Science.gov (United States)

    De-Deus, Gustavo André; Nogueira Leal Silva, Emmanuel João; Moreira, Edson Jorge; de Almeida Neves, Aline; Belladonna, Felipe Gonçalves; Tameirão, Michele

    2014-04-01

    This study was designed to quantitatively evaluate the amount of apically extruded debris by the Self-Adjusting-File system (SAF; ReDent-Nova, Ra'anana, Israel). Hand and rotary instruments were used as references for comparison. Sixty mesial roots of mandibular molars were randomly assigned to 3 groups (n = 20). The root canals were instrumented with hand files using a crown-down technique. The ProTaper (Dentsply Maillefer, Ballaigues, Switzerland) and SAF systems were used according to the manufacturers' instructions. Sodium hypochlorite was used as an irrigant, and the apically extruded debris was collected in preweighted glass vials and dried afterward. The mean weight of debris was assessed with a microbalance and statistically analyzed using 1-way analysis of variance and the post hoc Tukey multiple comparison test. Hand file instrumentation produced significantly more debris compared with the ProTaper and SAF systems (P system produced significantly more debris compared with the SAF system (P systems caused apical debris extrusion. SAF instrumentation was associated with less debris extrusion compared with the use of hand and rotary files. Copyright © 2014 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  10. Study and development of a document file system with selective access; Etude et realisation d'un systeme de fichiers documentaires a acces selectif

    Energy Technology Data Exchange (ETDEWEB)

    Mathieu, Jean-Claude

    1974-06-21

    The objective of this research thesis was to design and to develop a set of software aimed at an efficient management of a document file system by using methods of selective access to information. Thus, the three main aspects of file processing (creation, modification, reorganisation) have been addressed. The author first presents the main problems related to the development of a comprehensive automatic documentation system, and their conventional solutions. Some future aspects, notably dealing with the development of peripheral computer technology, are also evoked. He presents the characteristics of INIS bibliographic records provided by the IAEA which have been used to create the files. In the second part, he briefly describes the file system general organisation. This system is based on the use of two main files: an inverse file which contains for each descriptor a list of of numbers of files indexed by this descriptor, and a dictionary of descriptor or input file which gives access to the inverse file. The organisation of these both files is then describes in a detailed way. Other related or associated files are created, and the overall architecture and mechanisms integrated into the file data input software are described, as well as various processing applied to these different files. Performance and possible development are finally discussed.

  11. Nickel-Titanium Single-file System in Endodontics.

    Science.gov (United States)

    Dagna, Alberto

    2015-10-01

    This work describes clinical cases treated with a innovative single-use and single-file nickel-titanium (NiTi) system used in continuous rotation. Nickel-titanium files are commonly used for root canal treatment but they tend to break because of bending stresses and torsional stresses. Today new instruments used only for one treatment have been introduced. They help the clinician to make the root canal shaping easier and safer because they do not require sterilization and after use have to be discarded. A new sterile instrument is used for each treatment in order to reduce the possibility of fracture inside the canal. The new One Shape NiTi single-file instrument belongs to this group. One Shape is used for complete shaping of root canal after an adequate preflaring. Its protocol is simple and some clinical cases are presented. It is helpful for easy cases and reliable for difficult canals. After 2 years of clinical practice, One Shape seems to be helpful for the treatment of most of the root canals, with low risk of separation. After each treatment, the instrument is discarded and not sterilized in autoclave or re-used. This single-use file simplifies the endodontic therapy, because only one instrument is required for canal shaping of many cases. The respect of clinical protocol guarantees predictable good results.

  12. SLIB77, Source Library Data Compression and File Maintenance System

    International Nuclear Information System (INIS)

    Lunsford, A.

    1989-01-01

    Description of program or function: SLIB77 is a source librarian program designed to maintain FORTRAN source code in a compressed form on magnetic disk. The program was prepared to meet program maintenance requirements for ongoing program development and continual improvement of very large programs involving many programmers from a number of different organizations. SLIB77 automatically maintains in one file the source of the current program as well as all previous modifications. Although written originally for FORTRAN programs, SLIB77 is suitable for use with data files, text files, operating systems, and other programming languages, such as Ada, C and COBOL. It can handle libraries with records of up to 160-characters. Records are grouped into DECKS and assigned deck names by the user. SLIB77 assigns a number to each record in each DECK. Records can be deleted or restored singly or as a group within each deck. Modification records are grouped and assigned modification identification names by the user. The program assigns numbers to each new record within the deck. The program has two modes of execution, BATCH and EDIT. The BATCH mode is controlled by an input file and is used to make changes permanent and create new library files. The EDIT mode is controlled by interactive terminal input and a built-in line editor is used for modification of single decks. Transferring of a library from one computer system to another is accomplished using a Portable Library File created by SLIB77 in a BATCH run

  13. Properties of the disk system of globular clusters

    International Nuclear Information System (INIS)

    Armandroff, T.E.

    1989-01-01

    A large refined data sample is used to study the properties and origin of the disk system of globular clusters. A scale height for the disk cluster system of 800-1500 pc is found which is consistent with scale-height determinations for samples of field stars identified with the Galactic thick disk. A rotational velocity of 193 + or - 29 km/s and a line-of-sight velocity dispersion of 59 + or - 14 km/s have been found for the metal-rich clusters. 70 references

  14. Prefetching in file systems for MIMD multiprocessors

    Science.gov (United States)

    Kotz, David F.; Ellis, Carla Schlatter

    1990-01-01

    The question of whether prefetching blocks on the file into the block cache can effectively reduce overall execution time of a parallel computation, even under favorable assumptions, is considered. Experiments have been conducted with an interleaved file system testbed on the Butterfly Plus multiprocessor. Results of these experiments suggest that (1) the hit ratio, the accepted measure in traditional caching studies, may not be an adequate measure of performance when the workload consists of parallel computations and parallel file access patterns, (2) caching with prefetching can significantly improve the hit ratio and the average time to perform an I/O (input/output) operation, and (3) an improvement in overall execution time has been observed in most cases. In spite of these gains, prefetching sometimes results in increased execution times (a negative result, given the optimistic nature of the study). The authors explore why it is not trivial to translate savings on individual I/O requests into consistently better overall performance and identify the key problems that need to be addressed in order to improve the potential of prefetching techniques in the environment.

  15. DJFS: Providing Highly Reliable and High‐Performance File System with Small‐Sized

    Directory of Open Access Journals (Sweden)

    Junghoon Kim

    2017-11-01

    Full Text Available File systems and applications try to implement their own update protocols to guarantee data consistency, which is one of the most crucial aspects of computing systems. However, we found that the storage devices are substantially under‐utilized when preserving data consistency because they generate massive storage write traffic with many disk cache flush operations and force‐unit‐access (FUA commands. In this paper, we present DJFS (Delta‐Journaling File System that provides both a high level of performance and data consistency for different applications. We made three technical contributions to achieve our goal. First, to remove all storage accesses with disk cache flush operations and FUA commands, DJFS uses small‐sized NVRAM for a file system journal. Second, to reduce the access latency and space requirements of NVRAM, DJFS attempts to journal compress the differences in the modified blocks. Finally, to relieve explicit checkpointing overhead, DJFS aggressively reflects the checkpoint transactions to file system area in the unit of the specified region. Our evaluation on TPC‐C SQLite benchmark shows that, using our novel optimization schemes, DJFS outperforms Ext4 by up to 64.2 times with only 128 MB of NVRAM.

  16. Digital Stratigraphy: Contextual Analysis of File System Traces in Forensic Science.

    Science.gov (United States)

    Casey, Eoghan

    2017-12-28

    This work introduces novel methods for conducting forensic analysis of file allocation traces, collectively called digital stratigraphy. These in-depth forensic analysis methods can provide insight into the origin, composition, distribution, and time frame of strata within storage media. Using case examples and empirical studies, this paper illuminates the successes, challenges, and limitations of digital stratigraphy. This study also shows how understanding file allocation methods can provide insight into concealment activities and how real-world computer usage can complicate digital stratigraphy. Furthermore, this work explains how forensic analysts have misinterpreted traces of normal file system behavior as indications of concealment activities. This work raises awareness of the value of taking the overall context into account when analyzing file system traces. This work calls for further research in this area and for forensic tools to provide necessary information for such contextual analysis, such as highlighting mass deletion, mass copying, and potential backdating. © 2017 American Academy of Forensic Sciences.

  17. Cross-system log file analysis for hypothesis testing

    NARCIS (Netherlands)

    Glahn, Christian

    2008-01-01

    Glahn, C. (2008). Cross-system log file analysis for hypothesis testing. Presented at Empowering Learners for Lifelong Competence Development: pedagogical, organisational and technological issues. 4th TENCompetence Open Workshop. April, 10, 2008, Madrid, Spain.

  18. Cooperative storage of shared files in a parallel computing system with dynamic block size

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary

    2015-11-10

    Improved techniques are provided for parallel writing of data to a shared object in a parallel computing system. A method is provided for storing data generated by a plurality of parallel processes to a shared object in a parallel computing system. The method is performed by at least one of the processes and comprises: dynamically determining a block size for storing the data; exchanging a determined amount of the data with at least one additional process to achieve a block of the data having the dynamically determined block size; and writing the block of the data having the dynamically determined block size to a file system. The determined block size comprises, e.g., a total amount of the data to be stored divided by the number of parallel processes. The file system comprises, for example, a log structured virtual parallel file system, such as a Parallel Log-Structured File System (PLFS).

  19. The system of indicators for regional cluster formation assessment

    Directory of Open Access Journals (Sweden)

    A. A. Mantsaeva

    2016-01-01

    Full Text Available The article shows the result of working-out the cluster formation assessment system, and each indicator of this system reflect the specific clusters property - cooperation and efficiency Completeness and depth of the system of indicators provided by systematic approach and a representing of quantitative and qualitative aspects of cluster formation process. A feature of the technique is the use of indicators that require a special accounting and enable tracking of a certain stage of cluster development. Testing the system of indicators produced by the example on the tourism industry, which is due, firstly, the high development rate of the tourist services sphere in comparison with the branches of material production, and, secondly, the increased interest in the establishment of regional tourism and recreation clusters with the country's leadership. Quantitative indicators of the formation and development of tourism and recreation clusters – geographic proximity of companies cluster members, the effectiveness of the sector for the regional economy, innovation activity, exports of goods and services, intended for the regions of the South and the North Caucasian Federal District. Universality technique ensures its empirical base - official data from Rosstat, the Federal Agency for Tourism, as well as the results of mass opinion polls carried out in all regions of the country as part of the annual “"Monitoring the quality of public and municipal services” (on the Republic of Kalmykia material. In general, we believe that the application of the developed system of indicators will contribute to intensify and improve the quality of cluster policy, implemented by the regional executive bodies and local authorities.

  20. 29 CFR 1602.43 - Commission's remedy for school systems' or districts' failure to file report.

    Science.gov (United States)

    2010-07-01

    ...' failure to file report. Any school system or district failing or refusing to file report EEO-5 when... 29 Labor 4 2010-07-01 2010-07-01 false Commission's remedy for school systems' or districts' failure to file report. 1602.43 Section 1602.43 Labor Regulations Relating to Labor (Continued) EQUAL...

  1. An information retrieval system for research file data

    Science.gov (United States)

    Joan E. Lengel; John W. Koning

    1978-01-01

    Research file data have been successfully retrieved at the Forest Products Laboratory through a high-speed cross-referencing system involving the computer program FAMULUS as modified by the Madison Academic Computing Center at the University of Wisconsin. The method of data input, transfer to computer storage, system utilization, and effectiveness are discussed....

  2. Methods and apparatus for capture and storage of semantic information with sub-files in a parallel computing system

    Science.gov (United States)

    Faibish, Sorin; Bent, John M; Tzelnic, Percy; Grider, Gary; Torres, Aaron

    2015-02-03

    Techniques are provided for storing files in a parallel computing system using sub-files with semantically meaningful boundaries. A method is provided for storing at least one file generated by a distributed application in a parallel computing system. The file comprises one or more of a complete file and a plurality of sub-files. The method comprises the steps of obtaining a user specification of semantic information related to the file; providing the semantic information as a data structure description to a data formatting library write function; and storing the semantic information related to the file with one or more of the sub-files in one or more storage nodes of the parallel computing system. The semantic information provides a description of data in the file. The sub-files can be replicated based on semantically meaningful boundaries.

  3. An in vitro study: Evaluation of intracanal calcium hydroxide removal with different file systems

    Directory of Open Access Journals (Sweden)

    Atul Jain

    2017-01-01

    Full Text Available Background: Calcium hydroxide (Ca(OH2 is the most commonly used intracanal material; it needs to be removed in entirety before obturation. Several techniques have been used for the same including use of various hand and rotary files. Aim: This study was carried out to compare the efficacy of Hand K files and single and multiple rotary file system in removal of Ca(OH2. Methodology: Distobuccal root of 45 maxillary molars were selected on the basis of specified inclusion and exclusion criteria. They were divided into three groups - Group 1 (H and K file, Group 2 (HERO Shaper, and Group 3 (One Shape. Biomechanical preparation (BMP was carried out as per the manufacturer's instructions; 2.5% sodium hypochlorite was used as the irrigant and 17% ethylenediaminetetraacetic acid as the penultimate irrigant. Ca(OH2powder was mixed with normal saline to obtain a paste; canals were filled with this paste using a Lentulo spiral and were sealed. After 7 days, Ca(OH2was removed, using the same file system as that used for BMP. Samples were sectioned longitudinally and evaluated under a stereomicroscope. Statistical Analysis: Statistical analysis of the obtained data was carried out using one-way analysis of variance test. Results: HERO Shaper displayed better removal of Ca(OH2than One Shape and Hand K file. Moreover, removal was better in the middle third of canal than apical third. Conclusion: Multiple rotary file system (HERO Shaper is more effective in removal of Ca(OH2than the single file system (One Shape

  4. FOSSIL SYSTEMS IN THE 400d CLUSTER CATALOG

    International Nuclear Information System (INIS)

    Voevodkin, Alexey; Borozdin, Konstantin; Heitmann, Katrin; Habib, Salman; Vikhlinin, Alexey; Mescheryakov, Alexander; Burenin, Rodion; Hornstrup, Allan

    2010-01-01

    We report the discovery of seven new fossil systems in the 400d cluster survey. Our search targets nearby, z ≤ 0.2, and X-ray bright, L X ≥ 10 43 erg s -1 , clusters of galaxies. Where available, we measure the optical luminosities from Sloan Digital Sky Survey images, thereby obtaining uniform sets of both X-ray and optical data. Our selection criteria identify 12 fossil systems, out of which five are known from previous studies. While in general agreement with earlier results, our larger sample size allows us to put tighter constraints on the number density of fossil clusters. It has been previously reported that fossil groups are more X-ray bright than other X-ray groups of galaxies for the same optical luminosity. We find, however, that the X-ray brightness of massive fossil systems is consistent with that of the general population of galaxy clusters and follows the same L X -L opt scaling relation.

  5. The signatures of the parental cluster on field planetary systems

    Science.gov (United States)

    Cai, Maxwell Xu; Portegies Zwart, Simon; van Elteren, Arjen

    2018-03-01

    Due to the high stellar densities in young clusters, planetary systems formed in these environments are likely to have experienced perturbations from encounters with other stars. We carry out direct N-body simulations of multiplanet systems in star clusters to study the combined effects of stellar encounters and internal planetary dynamics. These planetary systems eventually become part of the Galactic field population as the parental cluster dissolves, which is where most presently known exoplanets are observed. We show that perturbations induced by stellar encounters lead to distinct signatures in the field planetary systems, most prominently, the excited orbital inclinations and eccentricities. Planetary systems that form within the cluster's half-mass radius are more prone to such perturbations. The orbital elements are most strongly excited in the outermost orbit, but the effect propagates to the entire planetary system through secular evolution. Planet ejections may occur long after a stellar encounter. The surviving planets in these reduced systems tend to have, on average, higher inclinations and larger eccentricities compared to systems that were perturbed less strongly. As soon as the parental star cluster dissolves, external perturbations stop affecting the escaped planetary systems, and further evolution proceeds on a relaxation time-scale. The outer regions of these ejected planetary systems tend to relax so slowly that their state carries the memory of their last strong encounter in the star cluster. Regardless of the stellar density, we observe a robust anticorrelation between multiplicity and mean inclination/eccentricity. We speculate that the `Kepler dichotomy' observed in field planetary systems is a natural consequence of their early evolution in the parental cluster.

  6. The Fifth Workshop on HPC Best Practices: File Systems and Archives

    Energy Technology Data Exchange (ETDEWEB)

    Hick, Jason; Hules, John; Uselton, Andrew

    2011-11-30

    The workshop on High Performance Computing (HPC) Best Practices on File Systems and Archives was the fifth in a series sponsored jointly by the Department Of Energy (DOE) Office of Science and DOE National Nuclear Security Administration. The workshop gathered technical and management experts for operations of HPC file systems and archives from around the world. Attendees identified and discussed best practices in use at their facilities, and documented findings for the DOE and HPC community in this report.

  7. Developing a File System Structure to Solve Healthy Big Data Storage and Archiving Problems Using a Distributed File System

    Directory of Open Access Journals (Sweden)

    Atilla Ergüzen

    2018-06-01

    Full Text Available Recently, the use of internet has become widespread, increasing the use of mobile phones, tablets, computers, Internet of Things (IoT devices and other digital sources. In the health sector with the help of new generation digital medical equipment, this digital world also has tended to grow in an unpredictable way in that it has nearly 10% of the global wide data itself and continues to keep grow beyond what the other sectors have. This progress has greatly enlarged the amount of produced data which cannot be resolved with conventional methods. In this work, an efficient model for the storage of medical images using a distributed file system structure has been developed. With this work, a robust, available, scalable, and serverless solution structure has been produced, especially for storing large amounts of data in the medical field. Furthermore, the security level of the system is extreme by use of static Internet protocol (IP, user credentials, and synchronously encrypted file contents. One of the most important key features of the system is high performance and easy scalability. In this way, the system can work with fewer hardware elements and be more robust than others that use name node architecture. According to the test results, it is seen that the performance of the designed system is better than 97% from a Not Only Structured Query Language (NoSQL system, 80% from a relational database management system (RDBMS, and 74% from an operating system (OS.

  8. Unbinding Transition of Probes in Single-File Systems

    Science.gov (United States)

    Bénichou, Olivier; Démery, Vincent; Poncet, Alexis

    2018-02-01

    Single-file transport, arising in quasi-one-dimensional geometries where particles cannot pass each other, is characterized by the anomalous dynamics of a probe, notably its response to an external force. In these systems, the motion of several probes submitted to different external forces, although relevant to mixtures of charged and neutral or active and passive objects, remains unexplored. Here, we determine how several probes respond to external forces. We rely on a hydrodynamic description of the symmetric exclusion process to obtain exact analytical results at long times. We show that the probes can either move as a whole, or separate into two groups moving away from each other. In between the two regimes, they separate with a different dynamical exponent, as t1 /4. This unbinding transition also occurs in several continuous single-file systems and is expected to be observable.

  9. Scalable I/O Systems via Node-Local Storage: Approaching 1 TB/sec File I/O

    Energy Technology Data Exchange (ETDEWEB)

    Bronevetsky, G; Moody, A

    2009-08-18

    In the race to PetaFLOP-speed supercomputing systems, the increase in computational capability has been accompanied by corresponding increases in CPU count, total RAM, and storage capacity. However, a proportional increase in storage bandwidth has lagged behind. In order to improve system reliability and to reduce maintenance effort for modern large-scale systems, system designers have opted to remove node-local storage from the compute nodes. Today's multi-TeraFLOP supercomputers are typically attached to parallel file systems that provide only tens of GBs/s of I/O bandwidth. As a result, such machines have access to much less than 1GB/s of I/O bandwidth per TeraFLOP of compute power, which is below the generally accepted limit required for a well-balanced system. In a many ways, the current I/O bottleneck limits the capabilities of modern supercomputers, specifically in terms of limiting their working sets and restricting fault tolerance techniques, which become critical on systems consisting of tens of thousands of components. This paper resolves the dilemma between high performance and high reliability by presenting an alternative system design which makes use of node-local storage to improve aggregate system I/O bandwidth. In this work, we focus on the checkpointing use-case and present an experimental evaluation of the Scalable Checkpoint/Restart (SCR) library, a new adaptive checkpointing library that uses node-local storage to significantly improve the checkpointing performance of large-scale supercomputers. Experiments show that SCR achieves unprecedented write speeds, reaching a measured 700GB/s of aggregate bandwidth on 8,752 processors and an estimated 1TB/s for a similarly structured machine of 12,500 processors. This corresponds to a speedup of over 70x compared to the bandwidth provided by the 10GB/s parallel file system the cluster uses. Further, SCR can adapt to an environment in which there is wide variation in performance or capacity among

  10. QCS: a system for querying, clustering and summarizing documents.

    Energy Technology Data Exchange (ETDEWEB)

    Dunlavy, Daniel M.; Schlesinger, Judith D. (Center for Computing Sciences, Bowie, MD); O' Leary, Dianne P. (University of Maryland, College Park, MD); Conroy, John M. (Center for Computing Sciences, Bowie, MD)

    2006-10-01

    Information retrieval systems consist of many complicated components. Research and development of such systems is often hampered by the difficulty in evaluating how each particular component would behave across multiple systems. We present a novel hybrid information retrieval system--the Query, Cluster, Summarize (QCS) system--which is portable, modular, and permits experimentation with different instantiations of each of the constituent text analysis components. Most importantly, the combination of the three types of components in the QCS design improves retrievals by providing users more focused information organized by topic. We demonstrate the improved performance by a series of experiments using standard test sets from the Document Understanding Conferences (DUC) along with the best known automatic metric for summarization system evaluation, ROUGE. Although the DUC data and evaluations were originally designed to test multidocument summarization, we developed a framework to extend it to the task of evaluation for each of the three components: query, clustering, and summarization. Under this framework, we then demonstrate that the QCS system (end-to-end) achieves performance as good as or better than the best summarization engines. Given a query, QCS retrieves relevant documents, separates the retrieved documents into topic clusters, and creates a single summary for each cluster. In the current implementation, Latent Semantic Indexing is used for retrieval, generalized spherical k-means is used for the document clustering, and a method coupling sentence 'trimming', and a hidden Markov model, followed by a pivoted QR decomposition, is used to create a single extract summary for each cluster. The user interface is designed to provide access to detailed information in a compact and useful format. Our system demonstrates the feasibility of assembling an effective IR system from existing software libraries, the usefulness of the modularity of the design

  11. QCS : a system for querying, clustering, and summarizing documents.

    Energy Technology Data Exchange (ETDEWEB)

    Dunlavy, Daniel M.

    2006-08-01

    Information retrieval systems consist of many complicated components. Research and development of such systems is often hampered by the difficulty in evaluating how each particular component would behave across multiple systems. We present a novel hybrid information retrieval system--the Query, Cluster, Summarize (QCS) system--which is portable, modular, and permits experimentation with different instantiations of each of the constituent text analysis components. Most importantly, the combination of the three types of components in the QCS design improves retrievals by providing users more focused information organized by topic. We demonstrate the improved performance by a series of experiments using standard test sets from the Document Understanding Conferences (DUC) along with the best known automatic metric for summarization system evaluation, ROUGE. Although the DUC data and evaluations were originally designed to test multidocument summarization, we developed a framework to extend it to the task of evaluation for each of the three components: query, clustering, and summarization. Under this framework, we then demonstrate that the QCS system (end-to-end) achieves performance as good as or better than the best summarization engines. Given a query, QCS retrieves relevant documents, separates the retrieved documents into topic clusters, and creates a single summary for each cluster. In the current implementation, Latent Semantic Indexing is used for retrieval, generalized spherical k-means is used for the document clustering, and a method coupling sentence ''trimming'', and a hidden Markov model, followed by a pivoted QR decomposition, is used to create a single extract summary for each cluster. The user interface is designed to provide access to detailed information in a compact and useful format. Our system demonstrates the feasibility of assembling an effective IR system from existing software libraries, the usefulness of the modularity of

  12. Ex Vivo Comparison of Mtwo and RaCe Rotary File Systems in Root Canal Deviation: One File Only versus the Conventional Method.

    Science.gov (United States)

    Aminsobhani, Mohsen; Razmi, Hasan; Nozari, Solmaz

    2015-07-01

    Cleaning and shaping of the root canal system is an important step in endodontic therapy. New instruments incorporate new preparation techniques that can improve the efficacy of cleaning and shaping. The aim of this study was to compare the efficacy of Mtwo and RaCe rotary file systems in straightening the canal curvature using only one file or the conventional method. Sixty mesial roots of extracted human mandibular molars were prepared by RaCe and Mtwo nickel-titanium (NiTi) rotary files using the conventional and only one rotary file methods. The working length was 18 mm and the curvatures of the root canals were between 15-45°. By superimposing x-ray images before and after the instrumentation, deviation of the canals was assessed using Adobe Photoshop CS3 software. Preparation time was recorded. Data were analyzed using three-way ANOVA and Tukey's post hoc test. There were no significant differences between RaCe and Mtwo or between the two root canal preparation methods in root canal deviation in buccolingual and mesiodistal radiographs (P>0.05). Changes of root canal curvature in >35° subgroups were significantly more than in other subgroups with smaller canal curvatures. Preparation time was shorter in one file only technique. According to the results, the two rotary systems and the two root canal preparation methods had equal efficacy in straightening the canals; but the preparation time was shorter in one file only group.

  13. Motivators, Barriers and Concerns in Adoption of Electronic Filing System: Survey Evidence from Malaysian Professional Accountants

    OpenAIRE

    Ming-Ling Lai; Kwai-Fatt Choong

    2010-01-01

    Problem statement: Worldwide, electronic filing (e-filing) system and its' adoption has attracted much attention, however, scholarly study on accounting professionals' acceptance of e-filing system is scant. Approach: This study aimed (i) to examine factors that motivated professional accountants to use e-filing (ii) to solicit their usage experience and (iii) to assess the barriers to adoption and other compliance considerations. The questionnaire survey was administered on 700 professionals...

  14. Optimum off-line trace synchronization of computer clusters

    International Nuclear Information System (INIS)

    Jabbarifar, Masoume; Dagenais, Michel; Roy, Robert; Sendi, Alireza Shameli

    2012-01-01

    A tracing and monitoring framework produces detailed execution trace files for a system. Each trace file contains events with associated timestamps based on the local clock of their respective system, which are not perfectly synchronized. To monitor all behavior in multi-core distributed systems, a global time reference is required, thus the need for traces synchronization techniques. The synchronization is time consuming when there is a cluster of many computers. In this paper we propose an optimized technique to reduce the total synchronization time. Compared with related techniques that have been used on kernel level traces, this method improves the performance while maintaining a high accuracy. It uses the packet rate and the hop count as two major criteria to focus the computation on more accurate network links during synchronization. These criteria, tested in real-word experiments, were identified as most important features of a network. Furthermore, we present numerical and analytical evaluation results, and compare these with previous methods demonstrating the accuracy and the performance of the method.

  15. The properties of the disk system of globular clusters

    Science.gov (United States)

    Armandroff, Taft E.

    1989-01-01

    A large refined data sample is used to study the properties and origin of the disk system of globular clusters. A scale height for the disk cluster system of 800-1500 pc is found which is consistent with scale-height determinations for samples of field stars identified with the Galactic thick disk. A rotational velocity of 193 + or - 29 km/s and a line-of-sight velocity dispersion of 59 + or - 14 km/s have been found for the metal-rich clusters.

  16. ATLAS, an integrated structural analysis and design system. Volume 4: Random access file catalog

    Science.gov (United States)

    Gray, F. P., Jr. (Editor)

    1979-01-01

    A complete catalog is presented for the random access files used by the ATLAS integrated structural analysis and design system. ATLAS consists of several technical computation modules which output data matrices to corresponding random access file. A description of the matrices written on these files is contained herein.

  17. Cluster Computing for Embedded/Real-Time Systems

    Science.gov (United States)

    Katz, D.; Kepner, J.

    1999-01-01

    Embedded and real-time systems, like other computing systems, seek to maximize computing power for a given price, and thus can significantly benefit from the advancing capabilities of cluster computing.

  18. The image database management system of teaching file using personal computer

    International Nuclear Information System (INIS)

    Shin, M. J.; Kim, G. W.; Chun, T. J.; Ahn, W. H.; Baik, S. K.; Choi, H. Y.; Kim, B. G.

    1995-01-01

    For the systemic management and easy using of teaching file in radiology department, the authors tried to do the setup of a database management system of teaching file using personal computer. We used a personal computer (IBM PC compatible, 486DX2) including a image capture card(Window vision, Dooin Elect, Seoul, Korea) and video camera recorder (8mm, CCD-TR105, Sony, Tokyo, Japan) for the acquisition and storage of images. We developed the database program by using Foxpro for Window 2.6(Microsoft, Seattle, USA) executed in the Window 3.1 (Microsoft, Seattle, USA). Each datum consisted of hospital number, name, sex, age, examination date, keyword, radiologic examination modalities, final diagnosis, radiologic findings, references and representative images. The images were acquired and stored as bitmap format (8 bitmap, 540 X 390 ∼ 545 X 414, 256 gray scale) and displayed on the 17 inch-flat monitor(1024 X 768, Samtron, Seoul, Korea). Without special devices, the images acquisition and storage could be done on the reading viewbox, simply. The image quality on the computer's monitor was less than the one of original film on the viewbox, but generally the characteristics of each lesions could be differentiated. Easy retrieval of data was possible for the purpose of teaching file system. Without high cost appliances, we could consummate the image database system of teaching file using personal computer with relatively inexpensive method

  19. Quantitative evaluation of apically extruded debris with different single-file systems: Reciproc, F360 and OneShape versus Mtwo.

    Science.gov (United States)

    Bürklein, S; Benten, S; Schäfer, E

    2014-05-01

    To assess in a laboratory setting the amount of apically extruded debris associated with different single-file nickel-titanium instrumentation systems compared to one multiple-file rotary system. Eighty human mandibular central incisors were randomly assigned to four groups (n = 20 teeth per group). The root canals were instrumented according to the manufacturers' instructions using the reciprocating single-file system Reciproc, the single-file rotary systems F360 and OneShape and the multiple-file rotary Mtwo instruments. The apically extruded debris was collected and dried in pre-weighed glass vials. The amount of debris was assessed with a micro balance and statistically analysed using anova and post hoc Student-Newman-Keuls test. The time required to prepare the canals with the different instruments was also recorded. Reciproc produced significantly more debris compared to all other systems (P file rotary systems and the multiple-file rotary system (P > 0.05). Instrumentation with the three single-file systems was significantly faster than with Mtwo (P systems caused apical debris extrusion. Rotary instrumentation was associated with less debris extrusion compared to reciprocal instrumentation. © 2013 International Endodontic Journal. Published by John Wiley & Sons Ltd.

  20. Supra-galactic colour patterns in globular cluster systems

    Science.gov (United States)

    Forte, Juan C.

    2017-07-01

    An analysis of globular cluster systems associated with galaxies included in the Virgo and Fornax Hubble Space Telescope-Advanced Camera Surveys reveals distinct (g - z) colour modulation patterns. These features appear on composite samples of globular clusters and, most evidently, in galaxies with absolute magnitudes Mg in the range from -20.2 to -19.2. These colour modulations are also detectable on some samples of globular clusters in the central galaxies NGC 1399 and NGC 4486 (and confirmed on data sets obtained with different instruments and photometric systems), as well as in other bright galaxies in these clusters. After discarding field contamination, photometric errors and statistical effects, we conclude that these supra-galactic colour patterns are real and reflect some previously unknown characteristic. These features suggest that the globular cluster formation process was not entirely stochastic but included a fraction of clusters that formed in a rather synchronized fashion over large spatial scales, and in a tentative time lapse of about 1.5 Gy at redshifts z between 2 and 4. We speculate that the putative mechanism leading to that synchronism may be associated with large scale feedback effects connected with violent star-forming events and/or with supermassive black holes.

  1. [Development and evaluation of the medical imaging distribution system with dynamic web application and clustering technology].

    Science.gov (United States)

    Yokohama, Noriya; Tsuchimoto, Tadashi; Oishi, Masamichi; Itou, Katsuya

    2007-01-20

    It has been noted that the downtime of medical informatics systems is often long. Many systems encounter downtimes of hours or even days, which can have a critical effect on daily operations. Such systems remain especially weak in the areas of database and medical imaging data. The scheme design shows the three-layer architecture of the system: application, database, and storage layers. The application layer uses the DICOM protocol (Digital Imaging and Communication in Medicine) and HTTP (Hyper Text Transport Protocol) with AJAX (Asynchronous JavaScript+XML). The database is designed to decentralize in parallel using cluster technology. Consequently, restoration of the database can be done not only with ease but also with improved retrieval speed. In the storage layer, a network RAID (Redundant Array of Independent Disks) system, it is possible to construct exabyte-scale parallel file systems that exploit storage spread. Development and evaluation of the test-bed has been successful in medical information data backup and recovery in a network environment. This paper presents a schematic design of the new medical informatics system that can be accommodated from a recovery and the dynamic Web application for medical imaging distribution using AJAX.

  2. Traveling cluster approximation for uncorrelated amorphous systems

    International Nuclear Information System (INIS)

    Kaplan, T.; Sen, A.K.; Gray, L.J.; Mills, R.

    1985-01-01

    In this paper, the authors apply the TCA concepts to spatially disordered, uncorrelated systems (e.g., fluids or amorphous metals without short-range order). This is the first approximation scheme for amorphous systems that takes cluster effects into account while preserving the Herglotz property for any amount of disorder. They have performed some computer calculations for the pair TCA, for the model case of delta-function potentials on a one-dimensional random chain. These results are compared with exact calculations (which, in principle, taken into account all cluster effects) and with the CPA, which is the single-site TCA. The density of states for the pair TCA clearly shows some improvement over the CPA, and yet, apparently, the pair approximation distorts some of the features of the exact results. They conclude that the effects of large clusters are much more important in an uncorrelated liquid metal than in a substitutional alloy. As a result, the pair TCA, which does quite a nice job for alloys, is not adequate for the liquid. Larger clusters must be treated exactly, and therefore an n-TCA with n > 2 must be used

  3. Automatic generation of configuration files for a distributed control system

    CERN Document Server

    Cupérus, J

    1995-01-01

    The CERN PS accelerator complex is composed of 9 interlinked accelerators for production and acceleration of various kinds of particles. The hardware is controlled through CAMAC, VME, G64, and GPIB modules, which in turn are controlled by more than 100 microprocessors in VME crates. To produce startup files for all these microprocessors, with the correct drivers, programs and parameters in each of them, is quite a challenge. The problem is solved by generating the startup files automatically from the description of the control system in a relational database. The generation process detects inconsistencies and incomplete information. Included in the startup files are data which are formally comments, but can be interpreted for run-time checking of interface modules and program activity.

  4. Comparison of canal transportation and centering ability of twisted files, Pathfile-ProTaper system, and stainless steel hand K-files by using computed tomography.

    Science.gov (United States)

    Gergi, Richard; Rjeily, Joe Abou; Sader, Joseph; Naaman, Alfred

    2010-05-01

    The purpose of this study was to compare canal transportation and centering ability of 2 rotary nickel-titanium (NiTi) systems (Twisted Files [TF] and Pathfile-ProTaper [PP]) with conventional stainless steel K-files. Ninety root canals with severe curvature and short radius were selected. Canals were divided randomly into 3 groups of 30 each. After preparation with TF, PP, and stainless steel files, the amount of transportation that occurred was assessed by using computed tomography. Three sections from apical, mid-root, and coronal levels of the canal were recorded. Amount of transportation and centering ability were assessed. The 3 groups were statistically compared with analysis of variance and Tukey honestly significant difference test. Less transportation and better centering ability occurred with TF rotary instruments (P < .0001). K-files showed the highest transportation followed by PP system. PP system showed significant transportation when compared with TF (P < .0001). The TF system was found to be the best for all variables measured in this study. Copyright (c) 2010 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  5. Cross-system log file analysis for hypothesis testing

    NARCIS (Netherlands)

    Glahn, Christian; Specht, Marcus; Schoonenboom, Judith; Sligte, Henk; Moghnieh, Ayman; Hernández-Leo, Davinia; Stefanov, Krassen; Lemmers, Ruud; Koper, Rob

    2008-01-01

    Glahn, C., Specht, M., Schoonenboom, J., Sligte, H., Moghnieh, A., Hernández-Leo, D. Stefanov, K., Lemmers, R., & Koper, R. (2008). Cross-system log file analysis for hypothesis testing. In H. Sligte & R. Koper (Eds.), Proceedings of the 4th TENCompetence Open Workshop. Empowering Learners for

  6. Evaluation of the incidence of microcracks caused by Mtwo and ProTaper Next rotary file systems versus the self-adjusting file: A scanning electron microscopic study.

    Science.gov (United States)

    Saha, Suparna Ganguly; Vijaywargiya, Neelam; Saxena, Divya; Saha, Mainak Kanti; Bharadwaj, Anuj; Dubey, Sandeep

    2017-01-01

    To evaluate the incidence of microcrack formation canal preparation with two rotary nickel-titanium systems Mtwo and ProTaper Next along with the self-adjusting file system. One hundred and twenty mandibular premolar teeth were selected. Standardized access cavities were prepared and the canals were manually prepared up to size 20 after coronal preflaring. The teeth were divided into three experimental groups and one control group ( n = 30). Group 1: The canals were prepared using Mtwo rotary files. Group 2: The canals were prepared with ProTaper Next files. Group 3: The canals were prepared with self-adjusting files. Group 4: The canals were unprepared and used as a control. The roots were sectioned horizontally 3, 6, and 9 mm from the apex and examined under a scanning electron microscope to check for the presence of microcracks. The Pearson's Chi-square test was applied. The highest incidence of microcracks were associated with the ProTaper Next group, 80% ( P = 0.00), followed by the Mtwo group, 70% ( P = 0.000), and the least number of microcracks was noted in the self-adjusting file group, 10% ( P = 0.068). No significant difference was found between the ProTaper Next and Mtwo groups ( P = 0.368) while a significant difference was observed between the ProTaper Next and self-adjusting file groups ( P = 0.000) as well as the Mtwo and self-adjusting file groups ( P = 0.000). All nickel-titanium rotary instrument systems were associated with microcracks. However, the self-adjusting file system had significantly fewer microcracks when compared with the Mtwo and ProTaper Next.

  7. Disability in the UN cluster system

    Directory of Open Access Journals (Sweden)

    Adele Perry

    2010-07-01

    Full Text Available The cluster system offers space for raising awareness among humanitarian actors and for putting disability on the agenda, but it impairs local and cross-cutting dynamics at field level.

  8. Integrated photometry of globular star clusters in the Vilnius system

    International Nuclear Information System (INIS)

    Zdanavichyus, K.V.

    1983-01-01

    Integrated colour indices in the Vilnius photometric system and newly determined colour excesses Esub(B-V) for 39 globular clusters are presented. It is shown that the coincidence of integrated spectral types are not a sufficient criterion for the identity of intrinsic colour indices of globular clusters. Relation of integrated colour indices with the slope of the giant branch S and with the horizontal branch morphological type D is investigated. Integrated colour indices of clusters with a blue horizontal branch show no correlation with either D or S. The increase of colour indices of the clusters of types D >= 4 correlates with the distribution of stars along the horizontal branch. Integrated photometry of globular star clusters in the Vilnius multicoloured photometric system permits to determine their colour excesses from some Q diagrams and normal colour index. Integral normal colour indexes and Q parameters for I globular star clusters of the Mironov group display small changes as compared to clusters of group 2. Colour indexes among star clusters having only red horizontal branches (D=7) change most considerably

  9. DSN Beowulf Cluster-Based VLBI Correlator

    Science.gov (United States)

    Rogstad, Stephen P.; Jongeling, Andre P.; Finley, Susan G.; White, Leslie A.; Lanyi, Gabor E.; Clark, John E.; Goodhart, Charles E.

    2009-01-01

    The NASA Deep Space Network (DSN) requires a broadband VLBI (very long baseline interferometry) correlator to process data routinely taken as part of the VLBI source Catalogue Maintenance and Enhancement task (CAT M&E) and the Time and Earth Motion Precision Observations task (TEMPO). The data provided by these measurements are a crucial ingredient in the formation of precision deep-space navigation models. In addition, a VLBI correlator is needed to provide support for other VLBI related activities for both internal and external customers. The JPL VLBI Correlator (JVC) was designed, developed, and delivered to the DSN as a successor to the legacy Block II Correlator. The JVC is a full-capability VLBI correlator that uses software processes running on multiple computers to cross-correlate two-antenna broadband noise data. Components of this new system (see Figure 1) consist of Linux PCs integrated into a Beowulf Cluster, an existing Mark5 data storage system, a RAID array, an existing software correlator package (SoftC) originally developed for Delta DOR Navigation processing, and various custom- developed software processes and scripts. Parallel processing on the JVC is achieved by assigning slave nodes of the Beowulf cluster to process separate scans in parallel until all scans have been processed. Due to the single stream sequential playback of the Mark5 data, some ramp-up time is required before all nodes can have access to required scan data. Core functions of each processing step are accomplished using optimized C programs. The coordination and execution of these programs across the cluster is accomplished using Pearl scripts, PostgreSQL commands, and a handful of miscellaneous system utilities. Mark5 data modules are loaded on Mark5 Data systems playback units, one per station. Data processing is started when the operator scans the Mark5 systems and runs a script that reads various configuration files and then creates an experiment-dependent status database

  10. Embedded system file transfer USB

    International Nuclear Information System (INIS)

    Jaoua, Mehdi

    2008-01-01

    The development of the communication series A emphasized new aspects of data exchange. The transfer of data, subject of my project of end of studies, consists in transferring from the files of a support of mass towards another via port USB. In first phase, I had like stain the realization of an embarked system allowing the communication between a key USB and final of communication such as a Pc. For this fact, I had to include/understand the operation of protocol USB and thus I could programmed a Peak to manage this communication. The second phase, will consist in extending this project towards a transmission de< donnees between two keys USB without intervention of a powerful machine equipped with an operating system pour rant to manage this transaction. (Author)

  11. B =5 Skyrmion as a two-cluster system

    Science.gov (United States)

    Gudnason, Sven Bjarke; Halcrow, Chris

    2018-06-01

    The classical B =5 Skyrmion can be approximated by a two-cluster system in which a B =1 Skyrmion is attached to a core B =4 Skyrmion. We quantize this system, allowing the B =1 to freely orbit the core. The configuration space is 11 dimensional but simplifies significantly after factoring out the overall spin and isospin degrees of freedom. We exactly solve the free quantum problem and then include an interaction potential between the Skyrmions numerically. The resulting energy spectrum is compared to the corresponding nuclei—the helium-5/lithium-5 isodoublet. We find approximate parity doubling not seen in the experimental data. In addition, we fail to obtain the correct ground-state spin. The framework laid out for this two-cluster system can readily be modified for other clusters and in particular for other B =4 n +1 nuclei, of which B =5 is the simplest example.

  12. Silvabase: A flexible data file management system

    Science.gov (United States)

    Lambing, Steven J.; Reynolds, Sandra J.

    1991-01-01

    The need for a more flexible and efficient data file management system for mission planning in the Mission Operations Laboratory (EO) at MSFC has spawned the development of Silvabase. Silvabase is a new data file structure based on a B+ tree data structure. This data organization allows for efficient forward and backward sequential reads, random searches, and appends to existing data. It also provides random insertions and deletions with reasonable efficiency, utilization of storage space well but not at the expense of speed, and performance of these functions on a large volume of data. Mission planners required that some data be keyed and manipulated in ways not found in a commercial product. Mission planning software is currently being converted to use Silvabase in the Spacelab and Space Station Mission Planning Systems. Silvabase runs on a Digital Equipment Corporation's popular VAX/VMS computers in VAX Fortran. Silvabase has unique features involving time histories and intervals such as in operations research. Because of its flexibility and unique capabilities, Silvabase could be used in almost any government or commercial application that requires efficient reads, searches, and appends in medium to large amounts of almost any kind of data.

  13. A convertor and user interface to import CAD files into worldtoolkit virtual reality systems

    Science.gov (United States)

    Wang, Peter Hor-Ching

    1996-01-01

    Virtual Reality (VR) is a rapidly developing human-to-computer interface technology. VR can be considered as a three-dimensional computer-generated Virtual World (VW) which can sense particular aspects of a user's behavior, allow the user to manipulate the objects interactively, and render the VW at real-time accordingly. The user is totally immersed in the virtual world and feel the sense of transforming into that VW. NASA/MSFC Computer Application Virtual Environments (CAVE) has been developing the space-related VR applications since 1990. The VR systems in CAVE lab are based on VPL RB2 system which consists of a VPL RB2 control tower, an LX eyephone, an Isotrak polhemus sensor, two Fastrak polhemus sensors, a folk of Bird sensor, and two VPL DG2 DataGloves. A dynamics animator called Body Electric from VPL is used as the control system to interface with all the input/output devices and to provide the network communications as well as VR programming environment. The RB2 Swivel 3D is used as the modelling program to construct the VW's. A severe limitation of the VPL VR system is the use of RB2 Swivel 3D, which restricts the files to a maximum of 1020 objects and doesn't have the advanced graphics texture mapping. The other limitation is that the VPL VR system is a turn-key system which does not provide the flexibility for user to add new sensors and C language interface. Recently, NASA/MSFC CAVE lab provides VR systems built on Sense8 WorldToolKit (WTK) which is a C library for creating VR development environments. WTK provides device drivers for most of the sensors and eyephones available on the VR market. WTK accepts several CAD file formats, such as Sense8 Neutral File Format, AutoCAD DXF and 3D Studio file format, Wave Front OBJ file format, VideoScape GEO file format, Intergraph EMS stereolithographics and CATIA Stereolithographics STL file formats. WTK functions are object-oriented in their naming convention, are grouped into classes, and provide easy C

  14. A Formal Approach to Exploiting Multi-Stage Attacks based on File-System Vulnerabilities of Web Applications (Extended Version)

    OpenAIRE

    De Meo, Federico; Viganò, Luca

    2017-01-01

    Web applications require access to the file-system for many different tasks. When analyzing the security of a web application, secu- rity analysts should thus consider the impact that file-system operations have on the security of the whole application. Moreover, the analysis should take into consideration how file-system vulnerabilities might in- teract with other vulnerabilities leading an attacker to breach into the web application. In this paper, we first propose a classification of file-...

  15. Design and application of remote file management system

    International Nuclear Information System (INIS)

    Zhu Haijun; Liu Dekang; Shen liren

    2006-01-01

    File transfer protocol can help users transfer files between computers on internet. FTP can not fulfill the needs of users in special occasions, so it needs programmer define file transfer protocol himself based on users. The method or realization and application for user-defined file transfer protocol is introduced. (authors)

  16. Virtual file system on NoSQL for processing high volumes of HL7 messages.

    Science.gov (United States)

    Kimura, Eizen; Ishihara, Ken

    2015-01-01

    The Standardized Structured Medical Information Exchange (SS-MIX) is intended to be the standard repository for HL7 messages that depend on a local file system. However, its scalability is limited. We implemented a virtual file system using NoSQL to incorporate modern computing technology into SS-MIX and allow the system to integrate local patient IDs from different healthcare systems into a universal system. We discuss its implementation using the database MongoDB and describe its performance in a case study.

  17. Enhanced K-means clustering with encryption on cloud

    Science.gov (United States)

    Singh, Iqjot; Dwivedi, Prerna; Gupta, Taru; Shynu, P. G.

    2017-11-01

    This paper tries to solve the problem of storing and managing big files over cloud by implementing hashing on Hadoop in big-data and ensure security while uploading and downloading files. Cloud computing is a term that emphasis on sharing data and facilitates to share infrastructure and resources.[10] Hadoop is an open source software that gives us access to store and manage big files according to our needs on cloud. K-means clustering algorithm is an algorithm used to calculate distance between the centroid of the cluster and the data points. Hashing is a algorithm in which we are storing and retrieving data with hash keys. The hashing algorithm is called as hash function which is used to portray the original data and later to fetch the data stored at the specific key. [17] Encryption is a process to transform electronic data into non readable form known as cipher text. Decryption is the opposite process of encryption, it transforms the cipher text into plain text that the end user can read and understand well. For encryption and decryption we are using Symmetric key cryptographic algorithm. In symmetric key cryptography are using DES algorithm for a secure storage of the files. [3

  18. The incidence of root microcracks caused by 3 different single-file systems versus the ProTaper system.

    Science.gov (United States)

    Liu, Rui; Hou, Ben Xiang; Wesselink, Paul R; Wu, Min-Kai; Shemesh, Hagay

    2013-08-01

    The aim of this study was to compare the incidence of root cracks observed at the apical root surface and/or in the canal wall after canal instrumentation with 3 single-file systems and the ProTaper system (Dentsply Maillefer, Ballaigues, Switzerland). One hundred mandibular incisors were selected. Twenty control teeth were coronally flared with Gates-Glidden drills (Dentsply Maillefer). No further preparation was made. The other 80 teeth were mounted in resin blocks with simulated periodontal ligaments, and the apex was exposed. They were divided into 4 experimental groups (n = 20); the root canals were first coronally flared with Gates-Glidden drills and then instrumented to the full working length with the ProTaper, OneShape (Micro-Mega, Besancon, France), Reciproc (VDW, Munich, Germany), or the Self-Adjusting File (ReDent-Nova, Ra'anana, Israel). The apical root surface and horizontal sections 2, 4, and 6 mm from the apex were observed under a microscope. The presence of cracks was noted. The chi-square test was performed to compare the appearance of cracked roots between the experimental groups. No cracks were found in the control teeth and teeth instrumented with the Self-Adjusting File. Cracks were found in 10 of 20 (50%), 7 of 20 (35%), and 1 of 20 (5%) teeth after canal instrumentation with the ProTaper, OneShape, and Reciproc files, respectively. The difference between the experimental groups was statistically significant (P File and Reciproc files caused less cracks than the ProTaper and OneShape files. Copyright © 2013 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  19. Database Objects vs Files: Evaluation of alternative strategies for managing large remote sensing data

    Science.gov (United States)

    Baru, Chaitan; Nandigam, Viswanath; Krishnan, Sriram

    2010-05-01

    Increasingly, the geoscience user community expects modern IT capabilities to be available in service of their research and education activities, including the ability to easily access and process large remote sensing datasets via online portals such as GEON (www.geongrid.org) and OpenTopography (opentopography.org). However, serving such datasets via online data portals presents a number of challenges. In this talk, we will evaluate the pros and cons of alternative storage strategies for management and processing of such datasets using binary large object implementations (BLOBs) in database systems versus implementation in Hadoop files using the Hadoop Distributed File System (HDFS). The storage and I/O requirements for providing online access to large datasets dictate the need for declustering data across multiple disks, for capacity as well as bandwidth and response time performance. This requires partitioning larger files into a set of smaller files, and is accompanied by the concomitant requirement for managing large numbers of file. Storing these sub-files as blobs in a shared-nothing database implemented across a cluster provides the advantage that all the distributed storage management is done by the DBMS. Furthermore, subsetting and processing routines can be implemented as user-defined functions (UDFs) on these blobs and would run in parallel across the set of nodes in the cluster. On the other hand, there are both storage overheads and constraints, and software licensing dependencies created by such an implementation. Another approach is to store the files in an external filesystem with pointers to them from within database tables. The filesystem may be a regular UNIX filesystem, a parallel filesystem, or HDFS. In the HDFS case, HDFS would provide the file management capability, while the subsetting and processing routines would be implemented as Hadoop programs using the MapReduce model. Hadoop and its related software libraries are freely available

  20. Modeling the formation of globular cluster systems in the Virgo cluster

    International Nuclear Information System (INIS)

    Li, Hui; Gnedin, Oleg Y.

    2014-01-01

    The mass distribution and chemical composition of globular cluster (GC) systems preserve fossil record of the early stages of galaxy formation. The observed distribution of GC colors within massive early-type galaxies in the ACS Virgo Cluster Survey (ACSVCS) reveals a multi-modal shape, which likely corresponds to a multi-modal metallicity distribution. We present a simple model for the formation and disruption of GCs that aims to match the ACSVCS data. This model tests the hypothesis that GCs are formed during major mergers of gas-rich galaxies and inherit the metallicity of their hosts. To trace merger events, we use halo merger trees extracted from a large cosmological N-body simulation. We select 20 halos in the mass range of 2 × 10 12 to 7 × 10 13 M ☉ and match them to 19 Virgo galaxies with K-band luminosity between 3 × 10 10 and 3 × 10 11 L ☉ . To set the [Fe/H] abundances, we use an empirical galaxy mass-metallicity relation. We find that a minimal merger ratio of 1:3 best matches the observed cluster metallicity distribution. A characteristic bimodal shape appears because metal-rich GCs are produced by late mergers between massive halos, while metal-poor GCs are produced by collective merger activities of less massive hosts at early times. The model outcome is robust to alternative prescriptions for cluster formation rate throughout cosmic time, but a gradual evolution of the mass-metallicity relation with redshift appears to be necessary to match the observed cluster metallicities. We also affirm the age-metallicity relation, predicted by an earlier model, in which metal-rich clusters are systematically several billion younger than their metal-poor counterparts.

  1. Considering User's Access Pattern in Multimedia File Systems

    Science.gov (United States)

    Cho, KyoungWoon; Ryu, YeonSeung; Won, Youjip; Koh, Kern

    2002-12-01

    Legacy buffer cache management schemes for multimedia server are grounded at the assumption that the application sequentially accesses the multimedia file. However, user access pattern may not be sequential in some circumstances, for example, in distance learning application, where the user may exploit the VCR-like function(rewind and play) of the system and accesses the particular segments of video repeatedly in the middle of sequential playback. Such a looping reference can cause a significant performance degradation of interval-based caching algorithms. And thus an appropriate buffer cache management scheme is required in order to deliver desirable performance even under the workload that exhibits looping reference behavior. We propose Adaptive Buffer cache Management(ABM) scheme which intelligently adapts to the file access characteristics. For each opened file, ABM applies either the LRU replacement or the interval-based caching depending on the Looping Reference Indicator, which indicates that how strong temporally localized access pattern is. According to our experiment, ABM exhibits better buffer cache miss ratio than interval-based caching or LRU, especially when the workload exhibits not only sequential but also looping reference property.

  2. Privacy Act System of Records: EPA Personnel Emergency Contact Files, EPA-44

    Science.gov (United States)

    Learn about the EPA Personnel Emergency Contact Files System, including including who is covered in the system, the purpose of data collection, routine uses for the system's records, and other security procedure.

  3. Fossil systems in the 400d cluster catalog

    DEFF Research Database (Denmark)

    Voevodkin, Alexey; Borozdin, Konstantin; Heitmann, Katrin

    2010-01-01

    We report the discovery of seven new fossil systems in the 400d cluster survey. Our search targets nearby, z ≤ 0.2, and X-ray bright, LX ≥ 10^43 erg s-1, clusters of galaxies. Where available, we measure the optical luminosities from Sloan Digital Sky Survey images, thereby obtaining uniform sets...

  4. Apically extruded debris with reciprocating single-file and full-sequence rotary instrumentation systems.

    Science.gov (United States)

    Bürklein, Sebastian; Schäfer, Edgar

    2012-06-01

    The purpose of this in vitro study was to assess the amount of apically extruded debris using rotary and reciprocating nickel-titanium instrumentation systems. Eighty human mandibular central incisors were randomly assigned to 4 groups (n = 20 teeth per group). The root canals were instrumented according to the manufacturers' instructions using the 2 reciprocating single-file systems Reciproc (VDW, Munich, Germany) and WaveOne (Dentsply Maillefer, Ballaigues, Switzerland) and the 2 full-sequence rotary Mtwo (VDW, Munich, Germany) and ProTaper (Dentsply Maillefer, Ballaigues, Switzerland) instruments. Bidistilled water was used as irrigant. The apically extruded debris was collected in preweighted glass vials using the Myers and Montgomery method. After drying, the mean weight of debris was assessed with a microbalance and statistically analyzed using analysis of variance and the post hoc Student-Newman-Keuls test. The time required to prepare the canals with the different instruments was also recorded. The reciprocating files produced significantly more debris compared with both rotary systems (P rotary instruments (P > .05), the reciprocating single-file system Reciproc produced significantly more debris compared with all other instruments (P Instrumentation was significantly faster using Reciproc than with all other instrument (P rotary instrumentation was associated with less debris extrusion compared with the use of reciprocating single-file systems. Copyright © 2012 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  5. Attitude Estimation in Fractionated Spacecraft Cluster Systems

    Science.gov (United States)

    Hadaegh, Fred Y.; Blackmore, James C.

    2011-01-01

    An attitude estimation was examined in fractioned free-flying spacecraft. Instead of a single, monolithic spacecraft, a fractionated free-flying spacecraft uses multiple spacecraft modules. These modules are connected only through wireless communication links and, potentially, wireless power links. The key advantage of this concept is the ability to respond to uncertainty. For example, if a single spacecraft module in the cluster fails, a new one can be launched at a lower cost and risk than would be incurred with onorbit servicing or replacement of the monolithic spacecraft. In order to create such a system, however, it is essential to know what the navigation capabilities of the fractionated system are as a function of the capabilities of the individual modules, and to have an algorithm that can perform estimation of the attitudes and relative positions of the modules with fractionated sensing capabilities. Looking specifically at fractionated attitude estimation with startrackers and optical relative attitude sensors, a set of mathematical tools has been developed that specify the set of sensors necessary to ensure that the attitude of the entire cluster ( cluster attitude ) can be observed. Also developed was a navigation filter that can estimate the cluster attitude if these conditions are satisfied. Each module in the cluster may have either a startracker, a relative attitude sensor, or both. An extended Kalman filter can be used to estimate the attitude of all modules. A range of estimation performances can be achieved depending on the sensors used and the topology of the sensing network.

  6. Log-Less Metadata Management on Metadata Server for Parallel File Systems

    Directory of Open Access Journals (Sweden)

    Jianwei Liao

    2014-01-01

    Full Text Available This paper presents a novel metadata management mechanism on the metadata server (MDS for parallel and distributed file systems. In this technique, the client file system backs up the sent metadata requests, which have been handled by the metadata server, so that the MDS does not need to log metadata changes to nonvolatile storage for achieving highly available metadata service, as well as better performance improvement in metadata processing. As the client file system backs up certain sent metadata requests in its memory, the overhead for handling these backup requests is much smaller than that brought by the metadata server, while it adopts logging or journaling to yield highly available metadata service. The experimental results show that this newly proposed mechanism can significantly improve the speed of metadata processing and render a better I/O data throughput, in contrast to conventional metadata management schemes, that is, logging or journaling on MDS. Besides, a complete metadata recovery can be achieved by replaying the backup logs cached by all involved clients, when the metadata server has crashed or gone into nonoperational state exceptionally.

  7. Securing the AliEn File Catalogue - Enforcing authorization with accountable file operations

    International Nuclear Information System (INIS)

    Schreiner, Steffen; Banerjee, Subho Sankar; Betev, Latchezar; Carminati, Federico; Vladimirovna Datskova, Olga; Furano, Fabrizio; Grigoras, Alina; Grigoras, Costin; Mendez Lorenzo, Patricia; Peters, Andreas Joachim; Saiz, Pablo; Bagnasco, Stefano; Zhu Jianlin

    2011-01-01

    The AliEn Grid Services, as operated by the ALICE Collaboration in its global physics analysis grid framework, is based on a central File Catalogue together with a distributed set of storage systems and the possibility to register links to external data resources. This paper describes several identified vulnerabilities in the AliEn File Catalogue access protocol regarding fraud and unauthorized file alteration and presents a more secure and revised design: a new mechanism, called LFN Booking Table, is introduced in order to keep track of access authorization in the transient state of files entering or leaving the File Catalogue. Due to a simplification of the original Access Envelope mechanism for xrootd-protocol-based storage systems, fundamental computational improvements of the mechanism were achieved as well as an up to 50% reduction of the credential's size. By extending the access protocol with signed status messages from the underlying storage system, the File Catalogue receives trusted information about a file's size and checksum and the protocol is no longer dependent on client trust. Altogether, the revised design complies with atomic and consistent transactions and allows for accountable, authentic, and traceable file operations. This paper describes these changes as part and beyond the development of AliEn version 2.19.

  8. Text File Comparator

    Science.gov (United States)

    Kotler, R. S.

    1983-01-01

    File Comparator program IFCOMP, is text file comparator for IBM OS/VScompatable systems. IFCOMP accepts as input two text files and produces listing of differences in pseudo-update form. IFCOMP is very useful in monitoring changes made to software at the source code level.

  9. Using electronic document management systems to manage highway project files.

    Science.gov (United States)

    2011-12-12

    "WisDOTs Bureau of Technical Services is interested in learning about the practices of other state departments of : transportation in developing and implementing an electronic document management system to manage highway : project files"

  10. Performance of the engineering analysis and data system 2 common file system

    Science.gov (United States)

    Debrunner, Linda S.

    1993-01-01

    The Engineering Analysis and Data System (EADS) was used from April 1986 to July 1993 to support large scale scientific and engineering computation (e.g. computational fluid dynamics) at Marshall Space Flight Center. The need for an updated system resulted in a RFP in June 1991, after which a contract was awarded to Cray Grumman. EADS II was installed in February 1993, and by July 1993 most users were migrated. EADS II is a network of heterogeneous computer systems supporting scientific and engineering applications. The Common File System (CFS) is a key component of this system. The CFS provides a seamless, integrated environment to the users of EADS II including both disk and tape storage. UniTree software is used to implement this hierarchical storage management system. The performance of the CFS suffered during the early months of the production system. Several of the performance problems were traced to software bugs which have been corrected. Other problems were associated with hardware. However, the use of NFS in UniTree UCFM software limits the performance of the system. The performance issues related to the CFS have led to a need to develop a greater understanding of the CFS organization. This paper will first describe the EADS II with emphasis on the CFS. Then, a discussion of mass storage systems will be presented, and methods of measuring the performance of the Common File System will be outlined. Finally, areas for further study will be identified and conclusions will be drawn.

  11. MOCCA Code for Star Cluster Simulation: Comparison with Optical Observations using COCOA

    OpenAIRE

    Askar, Abbas; Giersz, Mirek; Pych, Wojciech; Olech, Arkadiusz; Hypki, Arkadiusz

    2014-01-01

    We introduce and present preliminary results from COCOA (Cluster simulatiOn Comparison with ObservAtions) code for a star cluster after 12 Gyrs of evolution simulated using the MOCCA code. The COCOA code is being developed to quickly compare results of numerical simulations of star clusters with observational data. We use COCOA to obtain parameters of the projected cluster model. For comparison, a FITS file of the projected cluster was provided to observers so that they could use their observ...

  12. Uncoupling File System Components for Bridging Legacy and Modern Storage Architectures

    Science.gov (United States)

    Golpayegani, N.; Halem, M.; Tilmes, C.; Prathapan, S.; Earp, D. N.; Ashkar, J. S.

    2016-12-01

    Long running Earth Science projects can span decades of architectural changes in both processing and storage environments. As storage architecture designs change over decades such projects need to adjust their tools, systems, and expertise to properly integrate such new technologies with their legacy systems. Traditional file systems lack the necessary support to accommodate such hybrid storage infrastructure resulting in more complex tool development to encompass all possible storage architectures used for the project. The MODIS Adaptive Processing System (MODAPS) and the Level 1 and Atmospheres Archive and Distribution System (LAADS) is an example of a project spanning several decades which has evolved into a hybrid storage architecture. MODAPS/LAADS has developed the Lightweight Virtual File System (LVFS) which ensures a seamless integration of all the different storage architectures, including standard block based POSIX compliant storage disks, to object based architectures such as the S3 compliant HGST Active Archive System, and the Seagate Kinetic disks utilizing the Kinetic Protocol. With LVFS, all analysis and processing tools used for the project continue to function unmodified regardless of the underlying storage architecture enabling MODAPS/LAADS to easily integrate any new storage architecture without the costly need to modify existing tools to utilize such new systems. Most file systems are designed as a single application responsible for using metadata to organizing the data into a tree, determine the location for data storage, and a method of data retrieval. We will show how LVFS' unique approach of treating these components in a loosely coupled fashion enables it to merge different storage architectures into a single uniform storage system which bridges the underlying hybrid architecture.

  13. The role of the Pauli principle in three-cluster systems composed of identical clusters

    International Nuclear Information System (INIS)

    Lashko, Yu.A.; Filippov, G.F.

    2009-01-01

    Within the microscopic model based on the algebraic version of the resonating group method the role of the Pauli principle in the formation of continuum wave function of nuclear systems composed of three identical s-clusters has been investigated. Emphasis is placed upon the study of the exchange effects contained in the genuine three-cluster norm kernel. Three-fermion, three-boson, three-dineutron (3d ' ) and 3α systems are considered in detail. Simple analytical method of constructing the norm kernel for 3α system is suggested. The Pauli-allowed basis functions for the 3α and 3d ' systems are given in an explicit form and asymptotic behavior of these functions is established. Complete classification of the eigenfunctions and the eigenvalues of the 12 C norm kernel by the 8 Be=α+α eigenvalues has been given for the first time. Spectrum of the 12 C norm kernel is compared to that of the 5 H system.

  14. Cluster-based localization and tracking in ubiquitous computing systems

    CERN Document Server

    Martínez-de Dios, José Ramiro; Torres-González, Arturo; Ollero, Anibal

    2017-01-01

    Localization and tracking are key functionalities in ubiquitous computing systems and techniques. In recent years a very high variety of approaches, sensors and techniques for indoor and GPS-denied environments have been developed. This book briefly summarizes the current state of the art in localization and tracking in ubiquitous computing systems focusing on cluster-based schemes. Additionally, existing techniques for measurement integration, node inclusion/exclusion and cluster head selection are also described in this book.

  15. Managing Variant Calling Files the Big Data Way: Using HDFS and Apache Parquet

    NARCIS (Netherlands)

    Boufea, Aikaterini; Finkers, H.J.; Kaauwen, van M.P.W.; Kramer, M.R.; Athanasiadis, I.N.

    2017-01-01

    Big Data has been seen as a remedy for the efficient management of the ever-increasing genomic data. In this paper, we investigate the use of Apache Spark to store and process Variant Calling Files (VCF) on a Hadoop cluster. We demonstrate Tomatula, a software tool for converting VCF files to Apache

  16. Flying across Galaxy Clusters with Google Earth: additional imagery from SDSS co-added data

    Energy Technology Data Exchange (ETDEWEB)

    Hao, Jiangang; Annis, James; /Fermilab

    2010-10-01

    Galaxy clusters are spectacular. We provide a Google Earth compatible imagery for the deep co-added images from the Sloan Digital Sky Survey and make it a tool for examing galaxy clusters. Google Earth (in sky mode) provides a highly interactive environment for visualizing the sky. By encoding the galaxy cluster information into a kml/kmz file, one can use Google Earth as a tool for examining galaxy clusters and fly across them freely. However, the resolution of the images provided by Google Earth is not very high. This is partially because the major imagery google earth used is from Sloan Digital Sky Survey (SDSS) (SDSS collaboration 2000) and the resolutions have been reduced to speed up the web transferring. To have higher resolution images, you need to add your own images in a way that Google Earth can understand. The SDSS co-added data are the co-addition of {approx}100 scans of images from SDSS stripe 82 (Annis et al. 2010). It provides the deepest images based on SDSS and reach as deep as about redshift 1.0. Based on the co-added images, we created color images in a way as described by Lupton et al. (2004) and convert the color images to Google Earth compatible images using wcs2kml (Brewer et al. 2007). The images are stored at a public server at Fermi National Accelerator Laboratory and can be accessed by the public. To view those images in Google Earth, you need to download a kmz file, which contains the links to the color images, and then open the kmz file with your Google Earth. To meet different needs for resolutions, we provide three kmz files corresponding to low, medium and high resolution images. We recommend the high resolution one as long as you have a broadband Internet connection, though you should choose to download any of them, depending on your own needs and Internet speed. After you open the downloaded kmz file with Google Earth (in sky mode), it takes about 5 minutes (depending on your Internet connection and the resolution of images you

  17. Flying across Galaxy Clusters with Google Earth: additional imagery from SDSS co-added data

    International Nuclear Information System (INIS)

    Hao, Jiangang; Annis, James

    2010-01-01

    Galaxy clusters are spectacular. We provide a Google Earth compatible imagery for the deep co-added images from the Sloan Digital Sky Survey and make it a tool for examing galaxy clusters. Google Earth (in sky mode) provides a highly interactive environment for visualizing the sky. By encoding the galaxy cluster information into a kml/kmz file, one can use Google Earth as a tool for examining galaxy clusters and fly across them freely. However, the resolution of the images provided by Google Earth is not very high. This is partially because the major imagery google earth used is from Sloan Digital Sky Survey (SDSS) (SDSS collaboration 2000) and the resolutions have been reduced to speed up the web transferring. To have higher resolution images, you need to add your own images in a way that Google Earth can understand. The SDSS co-added data are the co-addition of ∼100 scans of images from SDSS stripe 82 (Annis et al. 2010). It provides the deepest images based on SDSS and reach as deep as about redshift 1.0. Based on the co-added images, we created color images in a way as described by Lupton et al. (2004) and convert the color images to Google Earth compatible images using wcs2kml (Brewer et al. 2007). The images are stored at a public server at Fermi National Accelerator Laboratory and can be accessed by the public. To view those images in Google Earth, you need to download a kmz file, which contains the links to the color images, and then open the kmz file with your Google Earth. To meet different needs for resolutions, we provide three kmz files corresponding to low, medium and high resolution images. We recommend the high resolution one as long as you have a broadband Internet connection, though you should choose to download any of them, depending on your own needs and Internet speed. After you open the downloaded kmz file with Google Earth (in sky mode), it takes about 5 minutes (depending on your Internet connection and the resolution of images you want

  18. Non-volatile main memory management methods based on a file system.

    Science.gov (United States)

    Oikawa, Shuichi

    2014-01-01

    There are upcoming non-volatile (NV) memory technologies that provide byte addressability and high performance. PCM, MRAM, and STT-RAM are such examples. Such NV memory can be used as storage because of its data persistency without power supply while it can be used as main memory because of its high performance that matches up with DRAM. There are a number of researches that investigated its uses for main memory and storage. They were, however, conducted independently. This paper presents the methods that enables the integration of the main memory and file system management for NV memory. Such integration makes NV memory simultaneously utilized as both main memory and storage. The presented methods use a file system as their basis for the NV memory management. We implemented the proposed methods in the Linux kernel, and performed the evaluation on the QEMU system emulator. The evaluation results show that 1) the proposed methods can perform comparably to the existing DRAM memory allocator and significantly better than the page swapping, 2) their performance is affected by the internal data structures of a file system, and 3) the data structures appropriate for traditional hard disk drives do not always work effectively for byte addressable NV memory. We also performed the evaluation of the effects caused by the longer access latency of NV memory by cycle-accurate full-system simulation. The results show that the effect on page allocation cost is limited if the increase of latency is moderate.

  19. Odysseus/DFS: Integration of DBMS and Distributed File System for Transaction Processing of Big Data

    OpenAIRE

    Kim, Jun-Sung; Whang, Kyu-Young; Kwon, Hyuk-Yoon; Song, Il-Yeol

    2014-01-01

    The relational DBMS (RDBMS) has been widely used since it supports various high-level functionalities such as SQL, schemas, indexes, and transactions that do not exist in the O/S file system. But, a recent advent of big data technology facilitates development of new systems that sacrifice the DBMS functionality in order to efficiently manage large-scale data. Those so-called NoSQL systems use a distributed file system, which support scalability and reliability. They support scalability of the...

  20. bcl::Cluster : A method for clustering biological molecules coupled with visualization in the Pymol Molecular Graphics System.

    Science.gov (United States)

    Alexander, Nathan; Woetzel, Nils; Meiler, Jens

    2011-02-01

    Clustering algorithms are used as data analysis tools in a wide variety of applications in Biology. Clustering has become especially important in protein structure prediction and virtual high throughput screening methods. In protein structure prediction, clustering is used to structure the conformational space of thousands of protein models. In virtual high throughput screening, databases with millions of drug-like molecules are organized by structural similarity, e.g. common scaffolds. The tree-like dendrogram structure obtained from hierarchical clustering can provide a qualitative overview of the results, which is important for focusing detailed analysis. However, in practice it is difficult to relate specific components of the dendrogram directly back to the objects of which it is comprised and to display all desired information within the two dimensions of the dendrogram. The current work presents a hierarchical agglomerative clustering method termed bcl::Cluster. bcl::Cluster utilizes the Pymol Molecular Graphics System to graphically depict dendrograms in three dimensions. This allows simultaneous display of relevant biological molecules as well as additional information about the clusters and the members comprising them.

  1. Schedulability-Driven Partitioning and Mapping for Multi-Cluster Real-Time Systems

    DEFF Research Database (Denmark)

    Pop, Paul; Eles, Petru; Peng, Zebo

    2004-01-01

    We present an approach to partitioning and mapping for multi-cluster embedded systems consisting of time-triggered and event-triggered clusters, interconnected via gateways. We have proposed a schedulability analysis for such systems, including a worst-case queuing delay analysis for the gateways...

  2. DICOM supported sofware configuration by XML files

    International Nuclear Information System (INIS)

    LucenaG, Bioing Fabian M; Valdez D, Andres E; Gomez, Maria E; Nasisi, Oscar H

    2007-01-01

    A method for the configuration of informatics systems that provide support to DICOM standards using XML files is proposed. The difference with other proposals is base on that this system does not code the information of a DICOM objects file, but codes the standard itself in an XML file. The development itself is the format for the XML files mentioned, in order that they can support what DICOM normalizes for multiple languages. In this way, the same configuration file (or files) can be use in different systems. Jointly the XML configuration file generated, we wrote also a set of CSS and XSL files. So the same file can be visualized in a standard browser, as a query system of DICOM standard, emerging use, that did not was a main objective but brings a great utility and versatility. We exposed also some uses examples of the configuration file mainly in relation with the load of DICOM information objects. Finally, at the conclusions we show the utility that the system has already provided when the edition of DICOM standard changes from 2006 to 2007

  3. Long term file migration. Part I: file reference patterns

    International Nuclear Information System (INIS)

    Smith, A.J.

    1978-08-01

    In most large computer installations, files are moved between on-line disk and mass storage (tape, integrated mass storage device) either automatically by the system or specifically at the direction of the user. This is the first of two papers which study the selection of algorithms for the automatic migration of files between mass storage and disk. The use of the text editor data sets at the Stanford Linear Accelerator Center (SLAC) computer installation is examined through the analysis of thirteen months of file reference data. Most files are used very few times. Of those that are used sufficiently frequently that their reference patterns may be examined, about a third show declining rates of reference during their lifetime; of the remainder, very few (about 5%) show correlated interreference intervals, and interreference intervals (in days) appear to be more skewed than would occur with the Bernoulli process. Thus, about two-thirds of all sufficiently active files appear to be referenced as a renewal process with a skewed interreference distribution. A large number of other file reference statistics (file lifetimes, interference distributions, moments, means, number of uses/file, file sizes, file rates of reference, etc.) are computed and presented. The results are applied in the following paper to the development and comparative evaluation of file migration algorithms. 17 figures, 13 tables

  4. Design Optimization of Multi-Cluster Embedded Systems for Real-Time Applications

    DEFF Research Database (Denmark)

    Pop, Paul; Eles, Petru; Peng, Zebo

    2004-01-01

    We present an approach to design optimization of multi-cluster embedded systems consisting of time-triggered and event-triggered clusters, interconnected via gateways. In this paper, we address design problems which are characteristic to multi-clusters: partitioning of the system functionality...... into time-triggered and event-triggered domains, process mapping, and the optimization of parameters corresponding to the communication protocol. We present several heuristics for solving these problems. Our heuristics are able to find schedulable implementations under limited resources, achieving...... an efficient utilization of the system. The developed algorithms are evaluated using extensive experiments and a real-life example....

  5. Design Optimization of Multi-Cluster Embedded Systems for Real-Time Applications

    DEFF Research Database (Denmark)

    Pop, Paul; Eles, Petru; Peng, Zebo

    2006-01-01

    We present an approach to design optimization of multi-cluster embedded systems consisting of time-triggered and event-triggered clusters, interconnected via gateways. In this paper, we address design problems which are characteristic to multi-clusters: partitioning of the system functionality...... into time-triggered and event-triggered domains, process mapping, and the optimization of parameters corresponding to the communication protocol. We present several heuristics for solving these problems. Our heuristics are able to find schedulable implementations under limited resources, achieving...... an efficient utilization of the system. The developed algorithms are evaluated using extensive experiments and a real-life example....

  6. Autonomic Cluster Management System (ACMS): A Demonstration of Autonomic Principles at Work

    Science.gov (United States)

    Baldassari, James D.; Kopec, Christopher L.; Leshay, Eric S.; Truszkowski, Walt; Finkel, David

    2005-01-01

    Cluster computing, whereby a large number of simple processors or nodes are combined together to apparently function as a single powerful computer, has emerged as a research area in its own right. The approach offers a relatively inexpensive means of achieving significant computational capabilities for high-performance computing applications, while simultaneously affording the ability to. increase that capability simply by adding more (inexpensive) processors. However, the task of manually managing and con.guring a cluster quickly becomes impossible as the cluster grows in size. Autonomic computing is a relatively new approach to managing complex systems that can potentially solve many of the problems inherent in cluster management. We describe the development of a prototype Automatic Cluster Management System (ACMS) that exploits autonomic properties in automating cluster management.

  7. TRANSPORT AND LOGISTICS CLUSTER IN AN ECONOMIC SYSTEM OF A REGION

    Directory of Open Access Journals (Sweden)

    I.G. Menshenina

    2008-09-01

    Full Text Available The main types of clusters are described in the article. The function of a transport and logistics model is also described using the theory of graphs. The relationship of clusters is shown in the economic system of a region, and the main role of transport and logistics cluster is emphasized as a good condition for the effective functioning of other clusters in the region.

  8. A design for a new catalog manager and associated file management for the Land Analysis System (LAS)

    Science.gov (United States)

    Greenhagen, Cheryl

    1986-01-01

    Due to the larger number of different types of files used in an image processing system, a mechanism for file management beyond the bounds of typical operating systems is necessary. The Transportable Applications Executive (TAE) Catalog Manager was written to meet this need. Land Analysis System (LAS) users at the EROS Data Center (EDC) encountered some problems in using the TAE catalog manager, including catalog corruption, networking difficulties, and lack of a reliable tape storage and retrieval capability. These problems, coupled with the complexity of the TAE catalog manager, led to the decision to design a new file management system for LAS, tailored to the needs of the EDC user community. This design effort, which addressed catalog management, label services, associated data management, and enhancements to LAS applications, is described. The new file management design will provide many benefits including improved system integration, increased flexibility, enhanced reliability, enhanced portability, improved performance, and improved maintainability.

  9. Eight months of clinical experience with the Self-Adjusting File system.

    Science.gov (United States)

    Solomonov, Michael

    2011-06-01

    The Self-Adjusting File (SAF) system (ReDent-Nova, Ra'anana, Israel) has been recently introduced for the simultaneous instrumentation and irrigation of root canals. The SAF is claimed to adapt itself three dimensionally to the root canal, including its cross-section. It is operated with a continuous flow of sodium hypochlorite that is delivered into the root canal through the hollow file and claimed to be activated by sonic agitation of the irrigant. Our aim was to present for the first time clinical cases prepared with the SAF system and to describe a clinical classification of canals, according to their difficulty, with recommendations for endodontic treatment sequences for each category. This report is based on the experience of a single endodontist, who used the system to treat more than 50 consecutive primary endodontic cases over the prior 8 months. A clinical classification was developed which enabled the operator to select a treatment protocol for easy and optimal glide path preparation to be effectively used with the SAF file in the various root canals encountered in the clinical environment. Clinical classification of canal difficulty makes root canal treatment sequences with the SAF simple and predictable. Many types of cases can be treated with the SAF system although a novice user is advised to advance slowly along the learning curve from simpler to more complicated canals. Copyright © 2011 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  10. FEDGROUP - A program system for producing group constants from evaluated nuclear data of files disseminated by IAEA

    International Nuclear Information System (INIS)

    Vertes, P.

    1976-06-01

    A program system for calculating group constants from several evaluated nuclear data files has been developed. These files are distributed by the Nuclear Data Section of IAEA. Our program system - FEDGROUP - has certain advantage over the well-known similar codes such as: 1. it requires only a medium sized computer />or approximately equal to 20000 words memory/, 2. it is easily adaptable to any type of computer, 3. it is flexible to the input evaluated nuclear data file and to the output group constant file. Nowadays, FEDGROUP calculates practically all types of group constants needed for reactor physics calculations by using the most frequent representations of evaluated data. (author)

  11. Comparative evaluation of effect of rotary and reciprocating single-file systems on pericervical dentin: A cone-beam computed tomography study.

    Science.gov (United States)

    Zinge, Priyanka Ramdas; Patil, Jayaprakash

    2017-01-01

    The aim of this study is to evaluate and compare the effect of one shape, Neolix rotary single-file systems and WaveOne, Reciproc reciprocating single-file systems on pericervical dentin (PCD) using cone-beam computed tomography (CBCT). A total of 40 freshly extracted mandibular premolars were collected and divided into two groups, namely, Group A - Rotary: A 1 - Neolix and A 2 - OneShape and Group B - Reciprocating: B 1 - WaveOne and B 2 - Reciproc. Preoperative scans of each were taken followed by conventional access cavity preparation and working length determination with 10-k file. Instrumentation of the canal was done according to the respective file system, and postinstrumentation CBCT scans of teeth were obtained. 90 μm thick slices were obtained 4 mm apical and coronal to the cementoenamel junction. The PCD thickness was calculated as the shortest distance from the canal outline to the closest adjacent root surface, which was measured in four surfaces, i.e., facial, lingual, mesial, and distal for all the groups in the two obtained scans. There was no significant difference found between rotary single-file systems and reciprocating single-file systems in their effect on PCD, but in Group B 2 , there was most significant loss of tooth structure in the mesial, lingual, and distal surface ( P file system removes more PCD as compared to other experimental groups, whereas Neolix single file system had the least effect on PCD.

  12. Schedulability-Driven Frame Packing for Multi-Cluster Distributed Embedded Systems

    DEFF Research Database (Denmark)

    Pop, Paul; Eles, Petru; Peng, Zebo

    2003-01-01

    We present an approach to frame packing for multi-cluster distributed embedded systems consisting of time-triggered and event-triggered clusters, interconnected via gateways. In our approach, the application messages are packed into frames such that the application is schedulable. Thus, we have...... also proposed a schedulability analysis for applications consisting of mixed event-triggered and time-triggered processes and messages, and a worst case queuing delay analysis for the gateways, responsible for routing inter-cluster traffic. Optimization heuristics for frame packing aiming at producing...... a schedulable system have been proposed. Extensive experiments and a real-life example show the efficiency of our frame-packing approach....

  13. Design of a Control System for Quality Maintenance on Cutting Edges of Files Production

    Directory of Open Access Journals (Sweden)

    E. Seabra

    2000-01-01

    Full Text Available The file cutting edges are the most important parameter that influence the performance of the filing operation. The practice shows that the most efficient way of generating these cutting edges is by penetration, by blow, of a cutting tool, which creates a plastic deformation on the file body. The penetration depth is probably the most important factor of the final quality of a file. In the existing machines of files manufacturing, this depth is manually adjusted by the operator, using specific mechanism. This means that files are manufactured on an empirical basis, relying on subjective factors, that do not allow to keep constant quality level of the production. In a research work, being developed in the University of Minho, it is intended to eliminate the subjectivity factors by the means of the evolution of the present “all-mechanical” system to a “mechatronic” one. In this paper, which is related with that research work, it presented a study of a round files production machine, regarding the identification, as well as the categorisation, of the operating parameters that affect the cutting edges production. They are, as well, defined and quantified those factors that influence the final quality of a round file.

  14. Heterologous expression of pikromycin biosynthetic gene cluster using Streptomyces artificial chromosome system.

    Science.gov (United States)

    Pyeon, Hye-Rim; Nah, Hee-Ju; Kang, Seung-Hoon; Choi, Si-Sun; Kim, Eung-Soo

    2017-05-31

    Heterologous expression of biosynthetic gene clusters of natural microbial products has become an essential strategy for titer improvement and pathway engineering of various potentially-valuable natural products. A Streptomyces artificial chromosomal conjugation vector, pSBAC, was previously successfully applied for precise cloning and tandem integration of a large polyketide tautomycetin (TMC) biosynthetic gene cluster (Nah et al. in Microb Cell Fact 14(1):1, 2015), implying that this strategy could be employed to develop a custom overexpression scheme of natural product pathway clusters present in actinomycetes. To validate the pSBAC system as a generally-applicable heterologous overexpression system for a large-sized polyketide biosynthetic gene cluster in Streptomyces, another model polyketide compound, the pikromycin biosynthetic gene cluster, was preciously cloned and heterologously expressed using the pSBAC system. A unique HindIII restriction site was precisely inserted at one of the border regions of the pikromycin biosynthetic gene cluster within the chromosome of Streptomyces venezuelae, followed by site-specific recombination of pSBAC into the flanking region of the pikromycin gene cluster. Unlike the previous cloning process, one HindIII site integration step was skipped through pSBAC modification. pPik001, a pSBAC containing the pikromycin biosynthetic gene cluster, was directly introduced into two heterologous hosts, Streptomyces lividans and Streptomyces coelicolor, resulting in the production of 10-deoxymethynolide, a major pikromycin derivative. When two entire pikromycin biosynthetic gene clusters were tandemly introduced into the S. lividans chromosome, overproduction of 10-deoxymethynolide and the presence of pikromycin, which was previously not detected, were both confirmed. Moreover, comparative qRT-PCR results confirmed that the transcription of pikromycin biosynthetic genes was significantly upregulated in S. lividans containing tandem

  15. 10 CFR 110.89 - Filing and service.

    Science.gov (United States)

    2010-01-01

    ...: Rulemakings and Adjudications Staff or via the E-Filing system, following the procedure set forth in 10 CFR 2.302. Filing by mail is complete upon deposit in the mail. Filing via the E-Filing system is completed... residence with some occupant of suitable age and discretion; (2) Following the requirements for E-Filing in...

  16. Exploring the Dynamics of Exoplanetary Systems in a Young Stellar Cluster

    Science.gov (United States)

    Thornton, Jonathan Daniel; Glaser, Joseph Paul; Wall, Joshua Edward

    2018-01-01

    I describe a dynamical simulation of planetary systems in a young star cluster. One rather arbitrary aspect of cluster simulations is the choice of initial conditions. These are typically chosen from some standard model, such as Plummer or King, or from a “fractal” distribution to try to model young clumpy systems. Here I adopt the approach of realizing an initial cluster model directly from a detailed magnetohydrodynamical model of cluster formation from a 1000-solar-mass interstellar gas cloud, with magnetic fields and radiative and wind feedback from massive stars included self-consistently. The N-body simulation of the stars and planets starts once star formation is largely over and feedback has cleared much of the gas from the region where the newborn stars reside. It continues until the cluster dissolves in the galactic field. Of particular interest is what would happen to the free-floating planets created in the gas cloud simulation. Are they captured by a star or are they ejected from the cluster? This method of building a dynamical cluster simulation directly from the results of a cluster formation model allows us to better understand the evolution of young star clusters and enriches our understanding of extrasolar planet development in them. These simulations were performed within the AMUSE simulation framework, and combine N-body, multiples and background potential code.

  17. Effect of Patency File on Transportation and Curve Straightening in Canal Preparation with ProTaper System.

    Science.gov (United States)

    Hasheminia, Seyed Mohsen; Farhadi, Nastaran; Shokraneh, Ali

    2013-01-01

    The aim of this ex vivo study was to evaluate the effect of using a patency file on apical transportation and curve straightening during canal instrumentation with the ProTaper rotary system. Seventy permanent mandibular first molars with mesiobuccal canals, measuring 18-23 mm in length and with a 25-40° curvature (according to the Schneider method), were selected. The working lengths were determined and the teeth were mounted and divided into two experimental groups: (A) prepared by the ProTaper system without using a patency file (n = 35) and (B) prepared by the ProTaper system using a patency file (n = 35). Radiographs taken before and after the preparation were imported into Photoshop software and the apical transportation, and curve straightening were measured. Data were analyzed using independent t-test. Partial correlation analysis was performed to evaluate the relationship between the initial curvature, transportation, and curve straightening (α = 0.05). Using a patency file during canal preparation significantly decreased both apical transportation and curve straightening (P system.

  18. 75 FR 27986 - Electronic Filing System-Web (EFS-Web) Contingency Option

    Science.gov (United States)

    2010-05-19

    ...] Electronic Filing System--Web (EFS-Web) Contingency Option AGENCY: United States Patent and Trademark Office... contingency option when the primary portal to EFS-Web has an unscheduled outage. Previously, the entire EFS-Web system is not available to the users during such an outage. The contingency option in EFS-Web will...

  19. Traveling-cluster approximation for uncorrelated amorphous systems

    International Nuclear Information System (INIS)

    Sen, A.K.; Mills, R.; Kaplan, T.; Gray, L.J.

    1984-01-01

    We have developed a formalism for including cluster effects in the one-electron Green's function for a positionally disordered (liquid or amorphous) system without any correlation among the scattering sites. This method is an extension of the technique known as the traveling-cluster approximation (TCA) originally obtained and applied to a substitutional alloy by Mills and Ratanavararaksa. We have also proved the appropriate fixed-point theorem, which guarantees, for a bounded local potential, that the self-consistent equations always converge upon iteration to a unique, Herglotz solution. To our knowledge, this is the only analytic theory for considering cluster effects. Furthermore, we have performed some computer calculations in the pair TCA, for the model case of delta-function potentials on a one-dimensional random chain. These results have been compared with ''exact calculations'' (which, in principle, take into account all cluster effects) and with the coherent-potential approximation (CPA), which is the single-site TCA. The density of states for the pair TCA clearly shows some improvement over the CPA and yet, apparently, the pair approximation distorts some of the features of the exact results

  20. File structure and organization in the automation system for operative account of equipment and materials in JINR

    International Nuclear Information System (INIS)

    Gulyaeva, N.D.; Markova, N.F.; Nikitina, V.I.; Tentyukova, G.N.

    1975-01-01

    The structure and organization of files in the information bank for the first variant of a JINR material and technical supply subsystem are described. Automated system of equipment operative stock-taking on the base of the SDS-6200 computer is developed. Information is stored on magnetic discs. The arrangement of each file depends on its purpose and structure of data. Access to the files can be arbitrary or consecutive. The files are divided into groups: primary document files, long-term reference, information on items that may change as a result of administrative decision [ru

  1. Advanced cluster methods for correlated-electron systems

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, Andre

    2015-04-27

    In this thesis, quantum cluster methods are used to calculate electronic properties of correlated-electron systems. A special focus lies in the determination of the ground state properties of a 3/4 filled triangular lattice within the one-band Hubbard model. At this filling, the electronic density of states exhibits a so-called van Hove singularity and the Fermi surface becomes perfectly nested, causing an instability towards a variety of spin-density-wave (SDW) and superconducting states. While chiral d+id-wave superconductivity has been proposed as the ground state in the weak coupling limit, the situation towards strong interactions is unclear. Additionally, quantum cluster methods are used here to investigate the interplay of Coulomb interactions and symmetry-breaking mechanisms within the nematic phase of iron-pnictide superconductors. The transition from a tetragonal to an orthorhombic phase is accompanied by a significant change in electronic properties, while long-range magnetic order is not established yet. The driving force of this transition may not only be phonons but also magnetic or orbital fluctuations. The signatures of these scenarios are studied with quantum cluster methods to identify the most important effects. Here, cluster perturbation theory (CPT) and its variational extention, the variational cluster approach (VCA) are used to treat the respective systems on a level beyond mean-field theory. Short-range correlations are incorporated numerically exactly by exact diagonalization (ED). In the VCA, long-range interactions are included by variational optimization of a fictitious symmetry-breaking field based on a self-energy functional approach. Due to limitations of ED, cluster sizes are limited to a small number of degrees of freedom. For the 3/4 filled triangular lattice, the VCA is performed for different cluster symmetries. A strong symmetry dependence and finite-size effects make a comparison of the results from different clusters difficult

  2. Techniques for Representation of Regional Clusters in Geographical In-formation Systems

    Directory of Open Access Journals (Sweden)

    Adriana REVEIU

    2011-01-01

    Full Text Available This paper provides an overview of visualization techniques adapted for regional clusters presentation in Geographic Information Systems. Clusters are groups of companies and insti-tutions co-located in a specific geographic region and linked by interdependencies in providing a related group of products and services. The regional clusters can be visualized by projecting the data into two-dimensional space or using parallel coordinates. Cluster membership is usually represented by different colours or by dividing clusters into several panels of a grille display. Taking into consideration regional clusters requirements and the multilevel administrative division of the Romania’s territory, I used two cartograms: NUTS2- regions and NUTS3- counties, to illustrate the tools for regional clusters representation.

  3. Browsing for Information on the Web and in the File System

    National Research Council Canada - National Science Library

    Seifert, Ethan; Stumpf, Simone; Herlocker, Jonathan; Wynn, Eleanor

    2007-01-01

    Browsing is one of the methods used for finding and refinding information on the web or in the file local system and there are opportunities to avoid this, particularly if that information is revisited frequently...

  4. Management system of ELHEP cluster machine for FEL photonics design

    Science.gov (United States)

    Zysik, Jacek; Poźniak, Krzysztof; Romaniuk, Ryszard

    2006-10-01

    A multipurpose, distributed MatLab calculations oriented, cluster machine was assembled in PERG/ELHEP laboratory at ISE/WUT. It is predicted mainly for advanced photonics and FPGA/DSP based systems design for Free Electron Laser. It will be used also for student projects for superconducting accelerator and FEL. Here we present one specific side of cluster design. For an intense, distributed daily work with the cluster, it is important to have a good interface and practical access to all machine resources. A complex management system was implemented in PERG laboratory. It helps all registered users to work using all necessary applications, communicate with other logged in people, check all the news and gather all necessary information about what is going on in the system, how it is utilized, etc. The system is also very practical for administrator purposes, it helps to keep controlling who is using the resources and for how long. It provides different privileges for different applications and many more. The system is introduced as a freeware, using open source code and can be modified by system operators or super-users who are interested in nonstandard system configuration.

  5. NetCDF based data archiving system applied to ITER Fast Plant System Control prototype

    International Nuclear Information System (INIS)

    Castro, R.; Vega, J.; Ruiz, M.; De Arcas, G.; Barrera, E.; López, J.M.; Sanz, D.; Gonçalves, B.; Santos, B.; Utzel, N.; Makijarvi, P.

    2012-01-01

    Highlights: ► Implementation of a data archiving solution for a Fast Plant System Controller (FPSC) for ITER CODAC. ► Data archiving solution based on scientific NetCDF-4 file format and Lustre storage clustering. ► EPICS control based solution. ► Tests results and detailed analysis of using NetCDF-4 and clustering technologies on fast acquisition data archiving. - Abstract: EURATOM/CIEMAT and Technical University of Madrid (UPM) have been involved in the development of a FPSC (Fast Plant System Control) prototype for ITER, based on PXIe (PCI eXtensions for Instrumentation). One of the main focuses of this project has been data acquisition and all the related issues, including scientific data archiving. Additionally, a new data archiving solution has been developed to demonstrate the obtainable performances and possible bottlenecks of scientific data archiving in Fast Plant System Control. The presented system implements a fault tolerant architecture over a GEthernet network where FPSC data are reliably archived on remote, while remaining accessible to be redistributed, within the duration of a pulse. The storing service is supported by a clustering solution to guaranty scalability, so that FPSC management and configuration may be simplified, and a unique view of all archived data provided. All the involved components have been integrated under EPICS (Experimental Physics and Industrial Control System), implementing in each case the necessary extensions, state machines and configuration process variables. The prototyped solution is based on the NetCDF-4 (Network Common Data Format) file format in order to incorporate important features, such as scientific data models support, huge size files management, platform independent codification, or single-writer/multiple-readers concurrency. In this contribution, a complete description of the above mentioned solution is presented, together with the most relevant results of the tests performed, while focusing in the

  6. 10 CFR 2.302 - Filing of documents.

    Science.gov (United States)

    2010-01-01

    ... this part shall be electronically transmitted through the E-Filing system, unless the Commission or... all methods of filing have been completed. (e) For filings by electronic transmission, the filer must... digital ID certificates, the NRC permits participants in the proceeding to access the E-Filing system to...

  7. MOCCA code for star cluster simulation: comparison with optical observations using COCOA

    Science.gov (United States)

    Askar, Abbas; Giersz, Mirek; Pych, Wojciech; Olech, Arkadiusz; Hypki, Arkadiusz

    2016-02-01

    We introduce and present preliminary results from COCOA (Cluster simulatiOn Comparison with ObservAtions) code for a star cluster after 12 Gyr of evolution simulated using the MOCCA code. The COCOA code is being developed to quickly compare results of numerical simulations of star clusters with observational data. We use COCOA to obtain parameters of the projected cluster model. For comparison, a FITS file of the projected cluster was provided to observers so that they could use their observational methods and techniques to obtain cluster parameters. The results show that the similarity of cluster parameters obtained through numerical simulations and observations depends significantly on the quality of observational data and photometric accuracy.

  8. File Managing and Program Execution in Web Operating Systems

    OpenAIRE

    Bravetti, Mario

    2010-01-01

    Web Operating Systems can be seen as an extension of traditional Operating Systems where the addresses used to manage files and execute programs (via the basic load/execution mechanism) are extended from local filesystem path-names to URLs. A first consequence is that, similarly as in traditional web technologies, executing a program at a given URL, can be done in two modalities: either the execution is performed client-side at the invoking machine (and relative URL addressing in the executed...

  9. Physics development of web-based tools for use in hardware clusters doing lattice physics

    International Nuclear Information System (INIS)

    Dreher, P.; Akers, W.; Chen, J.; Chen, Y.; Watson, C.

    2002-01-01

    Jefferson Lab and MIT are developing a set of web-based tools within the Lattice Hadron Physics Collaboration to allow lattice QCD theorists to treat the computational facilities located at the two sites as a single meta-facility. The prototype Lattice Portal provides researchers the ability to submit jobs to the cluster, browse data caches, and transfer files between cache and off-line storage. The user can view the configuration of the PBS servers and to monitor both the status of all batch queues as well as the jobs in each queue. Work is starting on expanding the present system to include job submissions at the meta-facility level (shared queue), as well as multi-site file transfers and enhanced policy-based data management capabilities

  10. Physics development of web-based tools for use in hardware clusters doing lattice physics

    International Nuclear Information System (INIS)

    Dreher, P.; Akers, Walt; Jian-ping Chen; Chen, Y.; William, A. Watson III

    2001-01-01

    Jefferson Lab and MIT are developing a set of web-based tools within the Lattice Hadron Physics Collaboration to allow lattice QCD theorists to treat the computational facilities located at the two sites as a single meta-facility. The prototype Lattice Portal provides researchers the ability to submit jobs to the cluster, browse data caches, and transfer files between cache and off-line storage. The user can view the configuration of the PBS servers and to monitor both the status of all batch queues as well as the jobs in each queue. Work is starting on expanding the present system to include job submissions at the meta-facility level (shared queue), as well as multi-site file transfers and enhanced policy-based data management capabilities

  11. Physics development of web-based tools for use in hardware clusters doing lattice physics

    Energy Technology Data Exchange (ETDEWEB)

    Dreher, P.; Akers, W.; Chen, J.; Chen, Y.; Watson, C

    2002-03-01

    Jefferson Lab and MIT are developing a set of web-based tools within the Lattice Hadron Physics Collaboration to allow lattice QCD theorists to treat the computational facilities located at the two sites as a single meta-facility. The prototype Lattice Portal provides researchers the ability to submit jobs to the cluster, browse data caches, and transfer files between cache and off-line storage. The user can view the configuration of the PBS servers and to monitor both the status of all batch queues as well as the jobs in each queue. Work is starting on expanding the present system to include job submissions at the meta-facility level (shared queue), as well as multi-site file transfers and enhanced policy-based data management capabilities.

  12. Endodontic treatment of mandibular molar with root dilaceration using Reciproc single-file system.

    Science.gov (United States)

    Meireles, Daniely Amorin; Bastos, Mariana Mena Barreto; Marques, André Augusto Franco; Garcia, Lucas da Fonseca Roberti; Sponchiado, Emílio Carlos

    2013-08-01

    Biomechanical preparation of root canals with accentuated curvature is challenging. New rotatory systems, such as Reciproc, require a shorter period of time to prepare curved canals, and became a viable alternative for endodontic treatment of teeth with root dilaceration. Thus, this study aimed to report a clinical case of endodontic therapy of root with accentuated dilaceration using Reciproc single-file system. Mandibular right second molar was diagnosed as asymptomatic irreversible pulpitis. Pulp chamber access was performed, and glide path was created with #10 K-file (Dentsply Maillefer) and PathFile #13, #16 and #19 (Dentsply Maillefer) up to the temporary working length. The working length measured corresponded to 20 mm in the mesio-buccal and mesio-lingual canals, and 22 mm in the distal canal. The R25 file (VDW GmbH) was used in all the canals for instrumentation and final preparation, followed by filling with Reciproc gutta-percha cones (VDW GmbH) and AH Plus sealer (Dentsply Maillefer), using thermal compaction technique. The case has been receiving follow-up for 6 mon and no painful symptomatology or periapical lesions have been found. Despite the difficulties, the treatment could be performed in a shorter period of time than the conventional methods.

  13. Endodontic treatment of mandibular molar with root dilaceration using Reciproc single-file system

    Directory of Open Access Journals (Sweden)

    Daniely Amorin Meireles

    2013-08-01

    Full Text Available Biomechanical preparation of root canals with accentuated curvature is challenging. New rotatory systems, such as Reciproc, require a shorter period of time to prepare curved canals, and became a viable alternative for endodontic treatment of teeth with root dilaceration. Thus, this study aimed to report a clinical case of endodontic therapy of root with accentuated dilaceration using Reciproc single-file system. Mandibular right second molar was diagnosed as asymptomatic irreversible pulpitis. Pulp chamber access was performed, and glide path was created with #10 K-file (Dentsply Maillefer and PathFile #13, #16 and #19 (Dentsply Maillefer up to the temporary working length. The working length measured corresponded to 20 mm in the mesio-buccal and mesio-lingual canals, and 22 mm in the distal canal. The R25 file (VDW GmbH was used in all the canals for instrumentation and final preparation, followed by filling with Reciproc gutta-percha cones (VDW GmbH and AH Plus sealer (Dentsply Maillefer, using thermal compaction technique. The case has been receiving follow-up for 6 mon and no painful symptomatology or periapical lesions have been found. Despite the difficulties, the treatment could be performed in a shorter period of time than the conventional methods.

  14. A Preliminary Study Application Clustering System in Acoustic Emission Monitoring

    Directory of Open Access Journals (Sweden)

    Saiful Bahari Nur Amira Afiza

    2017-01-01

    Full Text Available Acoustic Emission (AE is a non-destructive testing known as assessment on damage detection in structural engineering. It also can be used to discriminate the different types of damage occurring in a composite materials. The main problem associated with the data analysis is the discrimination between the different AE sources and analysis of the AE signal in order to identify the most critical damage mechanism. Clustering analysis is a technique in which the set of object are assigned to a group called cluster. The objective of the cluster analysis is to separate a set of data into several classes that reflect the internal structure of data. In this paper was used k-means algorithm for partitioned clustering method, numerous effort have been made to improve the performance of application k-means clustering algorithm. This paper presents a current review on application clustering system in Acoustic Emission.

  15. Fault detection of flywheel system based on clustering and principal component analysis

    Directory of Open Access Journals (Sweden)

    Wang Rixin

    2015-12-01

    Full Text Available Considering the nonlinear, multifunctional properties of double-flywheel with closed-loop control, a two-step method including clustering and principal component analysis is proposed to detect the two faults in the multifunctional flywheels. At the first step of the proposed algorithm, clustering is taken as feature recognition to check the instructions of “integrated power and attitude control” system, such as attitude control, energy storage or energy discharge. These commands will ask the flywheel system to work in different operation modes. Therefore, the relationship of parameters in different operations can define the cluster structure of training data. Ordering points to identify the clustering structure (OPTICS can automatically identify these clusters by the reachability-plot. K-means algorithm can divide the training data into the corresponding operations according to the reachability-plot. Finally, the last step of proposed model is used to define the relationship of parameters in each operation through the principal component analysis (PCA method. Compared with the PCA model, the proposed approach is capable of identifying the new clusters and learning the new behavior of incoming data. The simulation results show that it can effectively detect the faults in the multifunctional flywheels system.

  16. Cluster-based bulk metallic glass formation in Fe-Si-B-Nb alloy systems

    Energy Technology Data Exchange (ETDEWEB)

    Zhu, C L; Wang, Q; Li, F W; Li, Y H; Wang, Y M; Dong, C [State Key Laboratory of Materials Modification, Dalian University of Technology (DUT), Dalian 116024 (China); Zhang, W; Inoue, A, E-mail: dong@dlut.edu.c [Institute for Materials Research (IMR), Tohoku University, Katahira 2-1-1, Aoba-Ku, Sendai 980-8577 (Japan)

    2009-01-01

    Bulk metallic glass formations have been explored in Fe-B-Si-Nb alloy system using the so-called atomic cluster line approach in combination with minor alloying guideline. The atomic cluster line refers to a straight line linking binary cluster to the third element in a ternary system. The basic ternary compositions in Fe-B-Si system are determined by the inetersection points of two cluster lines, namely Fe-B cluster to Si and Fe-Si cluster to B, and then further alloyed with 3-5 at. % Nb for enhancing glass forming abilities. BMG rods with a diameter of 3 mm are formed under the case of minor Nb alloying the basic intersecting compositions of Fe{sub 8}B{sub 3}-Si with Fe{sub 12}Si-B and Fe{sub 8}B{sub 2}-Si with Fe{sub 9}Si-B. The BMGs also exhibit high Vickers hardness (H{sub v}) of 1130-1164 and high Young's modulous (E) of 170-180 GPa

  17. A multimembership catalogue for 1876 open clusters using UCAC4 data

    Science.gov (United States)

    Sampedro, L.; Dias, W. S.; Alfaro, E. J.; Monteiro, H.; Molino, A.

    2017-10-01

    The main objective of this work is to determine the cluster members of 1876 open clusters, using positions and proper motions of the astrometric fourth United States Naval Observatory (USNO) CCD Astrograph Catalog (UCAC4). For this purpose, we apply three different methods, all based on a Bayesian approach, but with different formulations: a purely parametric method, another completely non-parametric algorithm and a third, recently developed by Sampedro & Alfaro, using both formulations at different steps of the whole process. The first and second statistical moments of the members' phase-space subspace, obtained after applying the three methods, are compared for every cluster. Although, on average, the three methods yield similar results, there are also specific differences between them, as well as for some particular clusters. The comparison with other published catalogues shows good agreement. We have also estimated, for the first time, the mean proper motion for a sample of 18 clusters. The results are organized in a single catalogue formed by two main files, one with the most relevant information for each cluster, partially including that in UCAC4, and the other showing the individual membership probabilities for each star in the cluster area. The final catalogue, with an interface design that enables an easy interaction with the user, is available in electronic format at the Stellar Systems Group (SSG-IAA) web site (http://ssg.iaa.es/en/content/sampedro-cluster-catalog).

  18. VizieR Online Data Catalog: HST/ACS Coma Cluster Survey. VI. (den Brok+, 2011)

    Science.gov (United States)

    den Brok, M.; Peletier, R. F.; Valentijn, E. A.; Balcells, M.; Carter, D.; Erwin, P.; Ferguson, H. C.; Goudfrooij, P.; Graham, A. W.; Hammer, D.; Lucey, J. R.; Trentham, N.; Guzman, R.; Hoyos, C.; Verdoes Kleijn, G.; Jogee, S.; Karick, A. M.; Marinova, I.; Mouhcine, M.; Weinzirl, T.

    2018-01-01

    We have used the data from the HST/ACS Coma Cluster Survey, a deep two-passband imaging survey of the Coma cluster. A full description of the observations and data reduction can be found in Paper I (Carter et al., 2008ApJS..176..424C). We have derived colour gradients for a sample of confirmed or very likely Coma cluster members. (2 data files).

  19. Canal shaping of different single-file systems in curved root canals

    Directory of Open Access Journals (Sweden)

    Maurizio D'Amario

    2017-12-01

    Full Text Available Background/Purpose: This study compared maintenance of canal anatomy, occurrence of apical transportation, and working time observed after instrumentation with One Shape New Generation rotary system (Micro-Mega, with those observed after instrumentation with Reciproc (VDW and WaveOne (Dentsply-Maillefer reciprocating systems. Materials and methods: The mesial canals of 45 mandibular molars (curvature angles between 35° and 45° were selected. Specimens were randomly divided into three groups, and canal preparations were performed using One Shape, Reciproc, or WaveOne systems (size #25. A digital double radiographic technique was used to determine apical transportation and change in angle of curvature. Also, working time and instrument failures were recorded. Data were statistically analyzed. Results: During preparation, no file fractured. No statistically significant differences were found among groups. No system showed a significantly faster preparation time than others (P>0.05. All instruments maintained the original canal curvature well and were safe to use. Conclusion: Both continuous rotary instrument and reciprocating systems did not have any influence on the presence of apical transportation or caused an alteration in angle of canal curvature. Keywords: canal curvature, canal straightening, endodontics, reciprocating motion, single file instrumentation

  20. A case study in application I/O on Linux clusters

    International Nuclear Information System (INIS)

    Ross, R.; Nurmi, D.; Cheng, A.; Zingale, M.

    2001-01-01

    A critical but often ignored component of system performance is the I/O system. Today's applications expect a great deal from underlying storage systems and software, and both high performance distributed storage and high level interfaces have been developed to fill these needs. In this paper they discuss the I/O performance of a parallel scientific application on a Linux cluster, the FLASH astrophysics code. This application relies on three I/O software components to provide high performance parallel I/O on Linux clusters: the Parallel Virtual File System (PVFS), the ROMIO MPI-IO implementation, and the Hierarchical Data Format (HDF5) library. First they discuss the roles played by each of these components in providing an I/O solution. Next they discuss the FLASH I/O benchmark and point out its relevance. Following this they examine the performance of the benchmark, and through instrumentation of both the application and underlying system software code they discover the location of major software bottlenecks. They work around the most inhibiting of these bottlenecks, showing substantial performance improvement. Finally they point out similarities between the inefficiencies found here and those found in message passing systems, indicating that research in the message passing field could be leveraged to solve similar problems in high-level I/O interfaces

  1. XRootD popularity on hadoop clusters

    Science.gov (United States)

    Meoni, Marco; Boccali, Tommaso; Magini, Nicolò; Menichetti, Luca; Giordano, Domenico; CMS Collaboration

    2017-10-01

    Performance data and metadata of the computing operations at the CMS experiment are collected through a distributed monitoring infrastructure, currently relying on a traditional Oracle database system. This paper shows how to harness Big Data architectures in order to improve the throughput and the efficiency of such monitoring. A large set of operational data - user activities, job submissions, resources, file transfers, site efficiencies, software releases, network traffic, machine logs - is being injected into a readily available Hadoop cluster, via several data streamers. The collected metadata is further organized running fast arbitrary queries; this offers the ability to test several Map&Reduce-based frameworks and measure the system speed-up when compared to the original database infrastructure. By leveraging a quality Hadoop data store and enabling an analytics framework on top, it is possible to design a mining platform to predict dataset popularity and discover patterns and correlations.

  2. RRB's SVES Input File - Post Entitlement State Verification and Exchange System (PSSVES)

    Data.gov (United States)

    Social Security Administration — Several PSSVES request files are transmitted to SSA each year for processing in the State Verification and Exchange System (SVES). This is a first step in obtaining...

  3. Mode 3 knowledge production: Systems and systems theory, clusters and networks

    OpenAIRE

    Carayannis, Elias G.; Campbell, David F. J.; Rehman, Scheherazade S.

    2016-01-01

    With the comprehensive term of "Mode 3," we want to draw a conceptual link between systems and systems theory and want to demonstrate further how this can be applied to knowledge in the next steps. Systems can be understood as being composed of "elements", which are tied together by a "self-rationale". For innovation, often innovation clusters and innovation networks are being regarded as important. By leveraging systems theory for innovation concepts, one can implement references between the...

  4. Optimized distributed systems achieve significant performance improvement on sorted merging of massive VCF files.

    Science.gov (United States)

    Sun, Xiaobo; Gao, Jingjing; Jin, Peng; Eng, Celeste; Burchard, Esteban G; Beaty, Terri H; Ruczinski, Ingo; Mathias, Rasika A; Barnes, Kathleen; Wang, Fusheng; Qin, Zhaohui S

    2018-06-01

    Sorted merging of genomic data is a common data operation necessary in many sequencing-based studies. It involves sorting and merging genomic data from different subjects by their genomic locations. In particular, merging a large number of variant call format (VCF) files is frequently required in large-scale whole-genome sequencing or whole-exome sequencing projects. Traditional single-machine based methods become increasingly inefficient when processing large numbers of files due to the excessive computation time and Input/Output bottleneck. Distributed systems and more recent cloud-based systems offer an attractive solution. However, carefully designed and optimized workflow patterns and execution plans (schemas) are required to take full advantage of the increased computing power while overcoming bottlenecks to achieve high performance. In this study, we custom-design optimized schemas for three Apache big data platforms, Hadoop (MapReduce), HBase, and Spark, to perform sorted merging of a large number of VCF files. These schemas all adopt the divide-and-conquer strategy to split the merging job into sequential phases/stages consisting of subtasks that are conquered in an ordered, parallel, and bottleneck-free way. In two illustrating examples, we test the performance of our schemas on merging multiple VCF files into either a single TPED or a single VCF file, which are benchmarked with the traditional single/parallel multiway-merge methods, message passing interface (MPI)-based high-performance computing (HPC) implementation, and the popular VCFTools. Our experiments suggest all three schemas either deliver a significant improvement in efficiency or render much better strong and weak scalabilities over traditional methods. Our findings provide generalized scalable schemas for performing sorted merging on genetics and genomics data using these Apache distributed systems.

  5. Real Time Advanced Clustering System

    Directory of Open Access Journals (Sweden)

    Giuseppe Spampinato

    2017-05-01

    Full Text Available This paper describes a system to gather information from a stationary camera to identify moving objects. The proposed solution makes only use of motion vectors between adjacent frames, obtained from any algorithm. Starting from them, the system is able to retrieve clusters of moving objects in a scene acquired by an image sensor device. Since all the system is only based on optical flow, it is really simple and fast, to be easily integrated directly in low cost cameras. The experimental results show fast and robust performance of our method. The ANSI-C code has been tested on the ARM Cortex A15 CPU @2.32GHz, obtaining an impressive fps, about 3000 fps, excluding optical flow computation and I/O. Moreover, the system has been tested for different applications, cross traffic alert and video surveillance, in different conditions, indoor and outdoor, and with different lenses.

  6. VizieR Online Data Catalog: Tidal radii of 7 globular clusters (Lehmann+ 1997)

    Science.gov (United States)

    Lehmann, I.; Scholz, R.-D.

    1998-02-01

    We present new tidal radii for seven Galactic globular clusters using the method of automated star counts on Schmidt plates of the Tautenburg, Palomar and UK telescopes. The plates were fully scanned with the APM system in Cambridge (UK). Special account was given to a reliable background subtraction and the correction of crowding effects in the central cluster region. For the latter we used a new kind of crowding correction based on a statistical approach to the distribution of stellar images and the luminosity function of the cluster stars in the uncrowded area. The star counts were correlated with surface brightness profiles of different authors to obtain complete projected density profiles of the globular clusters. Fitting an empirical density law (King 1962AJ.....67..471K) we derived the following structural parameters: tidal radius rt, core radius rc and concentration parameter c. In the cases of NGC 5466, M 5, M 12, M 13 and M 15 we found an indication for a tidal tail around these objects (cf. Grillmair et al., 1995AJ....109.2553G). (1 data file).

  7. Clustering analysis of water distribution systems: identifying critical components and community impacts.

    Science.gov (United States)

    Diao, K; Farmani, R; Fu, G; Astaraie-Imani, M; Ward, S; Butler, D

    2014-01-01

    Large water distribution systems (WDSs) are networks with both topological and behavioural complexity. Thereby, it is usually difficult to identify the key features of the properties of the system, and subsequently all the critical components within the system for a given purpose of design or control. One way is, however, to more explicitly visualize the network structure and interactions between components by dividing a WDS into a number of clusters (subsystems). Accordingly, this paper introduces a clustering strategy that decomposes WDSs into clusters with stronger internal connections than external connections. The detected cluster layout is very similar to the community structure of the served urban area. As WDSs may expand along with urban development in a community-by-community manner, the correspondingly formed distribution clusters may reveal some crucial configurations of WDSs. For verification, the method is applied to identify all the critical links during firefighting for the vulnerability analysis of a real-world WDS. Moreover, both the most critical pipes and clusters are addressed, given the consequences of pipe failure. Compared with the enumeration method, the method used in this study identifies the same group of the most critical components, and provides similar criticality prioritizations of them in a more computationally efficient time.

  8. Radiological reporting system developed with FileMakerPro. Cooperation with HIS, RIS, and PACS

    International Nuclear Information System (INIS)

    Kawakami, Satoshi

    2004-01-01

    This article briefly describes our original radiological reporting system. This system was developed with the widely used database software FileMakerPro (ver 5.5). The reporting system can obtain information about patients and examinations from a radiology information system (RIS) by the Open DataBase Connectivity (ODBC) technique. By clicking the button on the reporting system, the corresponding Digital Imaging and Communications in Medicine (DICOM) images can be displayed on a picture archiving and communication system (PACS) workstation monitor. Reference images in JPEG format can be easily moved from PACS to the reporting system. Reports produced by the reporting system are distributed to the hospital information system (HIS) in Portable Document Format (PDF), through another web server. By utilizing the capacity of FileMakerPro, the human-machine interface of the system has been able to be improved easily. In addition, cooperation with HIS, RIS, and PACS could be constructed. Therefore, this original system would contribute to increasing the efficiency of radiological diagnosis. (author)

  9. Toward Millions of File System IOPS on Low-Cost, Commodity Hardware.

    Science.gov (United States)

    Zheng, Da; Burns, Randal; Szalay, Alexander S

    2013-01-01

    We describe a storage system that removes I/O bottlenecks to achieve more than one million IOPS based on a user-space file abstraction for arrays of commodity SSDs. The file abstraction refactors I/O scheduling and placement for extreme parallelism and non-uniform memory and I/O. The system includes a set-associative, parallel page cache in the user space. We redesign page caching to eliminate CPU overhead and lock-contention in non-uniform memory architecture machines. We evaluate our design on a 32 core NUMA machine with four, eight-core processors. Experiments show that our design delivers 1.23 million 512-byte read IOPS. The page cache realizes the scalable IOPS of Linux asynchronous I/O (AIO) and increases user-perceived I/O performance linearly with cache hit rates. The parallel, set-associative cache matches the cache hit rates of the global Linux page cache under real workloads.

  10. A Science Portal and Archive for Extragalactic Globular Cluster Systems Data

    Science.gov (United States)

    Young, Michael; Rhode, Katherine L.; Gopu, Arvind

    2015-01-01

    For several years we have been carrying out a wide-field imaging survey of the globular cluster populations of a sample of giant spiral, S0, and elliptical galaxies with distances of ~10-30 Mpc. We use mosaic CCD cameras on the WIYN 3.5-m and Kitt Peak 4-m telescopes to acquire deep BVR imaging of each galaxy and then analyze the data to derive global properties of the globular cluster system. In addition to measuring the total numbers, specific frequencies, spatial distributions, and color distributions for the globular cluster populations, we have produced deep, high-quality images and lists of tens to thousands of globular cluster candidates for the ~40 galaxies included in the survey.With the survey nearing completion, we have been exploring how to efficiently disseminate not only the overall results, but also all of the relevant data products, to the astronomical community. Here we present our solution: a scientific portal and archive for extragalactic globular cluster systems data. With a modern and intuitive web interface built on the same framework as the WIYN One Degree Imager Portal, Pipeline, and Archive (ODI-PPA), our system will provide public access to the survey results and the final stacked mosaic images of the target galaxies. In addition, the astrometric and photometric data for thousands of identified globular cluster candidates, as well as for all point sources detected in each field, will be indexed and searchable. Where available, spectroscopic follow-up data will be paired with the candidates. Advanced imaging tools will enable users to overlay the cluster candidates and other sources on the mosaic images within the web interface, while metadata charting tools will allow users to rapidly and seamlessly plot the survey results for each galaxy and the data for hundreds of thousands of individual sources. Finally, we will appeal to other researchers with similar data products and work toward making our portal a central repository for data

  11. Visual system of recovering and combination of information for ENDF (Evaluated Nuclear Data File) format libraries

    International Nuclear Information System (INIS)

    Ferreira, Claudia A.S. Velloso; Corcuera, Raquel A. Paviotti

    1997-01-01

    This report presents a data information retrieval and merger system for ENDF (Evaluated Nuclear Data File) format libraries, which can be run on personal computers under the Windows TM environment. The input is the name of an ENDF/B library, which can be chosen in a proper window. The system has a display function which allows the user to visualize the reaction data of a specific nuclide and to produce a printed copy of these data. The system allows the user to retrieve and/or combine evaluated data to create a single file of data in ENDF format, from a number of different files, each of which is in the ENDF format. The user can also create a mini-library from an ENDF/B library. This interactive and easy-to-handle system is a useful tool for Nuclear Data Centers and it is also of interest to nuclear and reactor physics researchers. (author)

  12. A Self-Organizing Spatial Clustering Approach to Support Large-Scale Network RTK Systems.

    Science.gov (United States)

    Shen, Lili; Guo, Jiming; Wang, Lei

    2018-06-06

    The network real-time kinematic (RTK) technique can provide centimeter-level real time positioning solutions and play a key role in geo-spatial infrastructure. With ever-increasing popularity, network RTK systems will face issues in the support of large numbers of concurrent users. In the past, high-precision positioning services were oriented towards professionals and only supported a few concurrent users. Currently, precise positioning provides a spatial foundation for artificial intelligence (AI), and countless smart devices (autonomous cars, unmanned aerial-vehicles (UAVs), robotic equipment, etc.) require precise positioning services. Therefore, the development of approaches to support large-scale network RTK systems is urgent. In this study, we proposed a self-organizing spatial clustering (SOSC) approach which automatically clusters online users to reduce the computational load on the network RTK system server side. The experimental results indicate that both the SOSC algorithm and the grid algorithm can reduce the computational load efficiently, while the SOSC algorithm gives a more elastic and adaptive clustering solution with different datasets. The SOSC algorithm determines the cluster number and the mean distance to cluster center (MDTCC) according to the data set, while the grid approaches are all predefined. The side-effects of clustering algorithms on the user side are analyzed with real global navigation satellite system (GNSS) data sets. The experimental results indicate that 10 km can be safely used as the cluster radius threshold for the SOSC algorithm without significantly reducing the positioning precision and reliability on the user side.

  13. A SURVEY ON DOCUMENT CLUSTERING APPROACH FOR COMPUTER FORENSIC ANALYSIS

    OpenAIRE

    Monika Raghuvanshi*, Rahul Patel

    2016-01-01

    In a forensic analysis, large numbers of files are examined. Much of the information comprises of in unstructured format, so it’s quite difficult task for computer forensic to perform such analysis. That’s why to do the forensic analysis of document within a limited period of time require a special approach such as document clustering. This paper review different document clustering algorithms methodologies for example K-mean, K-medoid, single link, complete link, average link in accorandance...

  14. Migrant Student Record Transfer System (MSRTS) [machine-readable data file].

    Science.gov (United States)

    Arkansas State Dept. of Education, Little Rock. General Education Div.

    The Migrant Student Record Transfer System (MSRTS) machine-readable data file (MRDF) is a collection of education and health data on more than 750,000 migrant children in grades K-12 in the United States (except Hawaii), the District of Columbia, and the outlying territories of Puerto Rico and the Mariana and Marshall Islands. The active file…

  15. Clinical Evaluation of Quality of Obturation and Instrumentation Time using Two Modified Rotary File Systems with Manual Instrumentation in Primary Teeth.

    Science.gov (United States)

    Govindaraju, Lavanya; Jeevanandan, Ganesh; Subramanian, Emg

    2017-09-01

    Pulp therapy in primary teeth has been performed using various instrumentation techniques. However, the conventional instrumentation technique used for root canal preparation in primary teeth is hand instrumentation. Various Nickel-Titanium (Ni-Ti) instruments are available to perform efficient root canal preparation in primary teeth. These Ni-Ti instruments has been designed to aid in better root canal preparation in permanent teeth but are rarely used in primary teeth. It is necessary to assess the feasibility of using these adult rotary files with a modified sequence in primary teeth. To compare the quality of obturation and instrumentation time during root canal preparation using hand files and modified rotary file systems in primary molars. Forty-five primary mandibular molars were randomly assigned to three experimental groups (n=15). Group I was instrumented using k-hand files, Group II with S2 ProTaper universal file and Group III with 0.25 tip 4% taper K3 rotary file. Standardized digital radiographs were taken before and after root canal instrumentation. Root canal preparation time was also recorded. Statistical analysis of the obtained data was done using SPSS Software version 17.0. An intergroup comparison of the instrumentation time and the quality of obturation was done using ANOVA and Chi-square test with the level of significance set at 0.05. No significant differences were noted with regard to the quality of obturation (p=0.791). However, a statistically significant difference was noted in the instrumentation time between the three groups (pProTaper rotary system had significantly lesser instrumentation time when compared to that of K3 rotary system and hand file system. The hand files, S2 ProTaper Universal and K3 0.25 tip 4% taper files systems performed similarly with respect to the quality of obturation. There was a significant difference in instrumentation time with manual instrumentation compared to the modified rotary file systems in primary

  16. A Binary System in the Hyades Cluster Hosting a Neptune-Sized Planet

    Science.gov (United States)

    Feinstein, Adina; Ciardi, David; Crossfield, Ian; Schlieder, Joshua; Petigura, Erik; David, Trevor J.; Bristow, Makennah; Patel, Rahul; Arnold, Lauren; Benneke, Björn; Christiansen, Jessie; Dressing, Courtney; Fulton, Benjamin; Howard, Andrew; Isaacson, Howard; Sinukoff, Evan; Thackeray, Beverly

    2018-01-01

    We report the discovery of a Neptune-size planet (Rp = 3.0Rearth) in the Hyades Cluster. The host star is in a binary system, comprising a K5V star and M7/8V star with a projected separation of 40 AU. The planet orbits the primary star with an orbital period of 17.3 days and a transit duration of 3 hours. The host star is bright (V = 11.2, J = 9.1) and so may be a good target for precise radial velocity measurements. The planet is the first Neptune-sized planet to be found orbiting in a binary system within an open cluster. The Hyades is the nearest star cluster to the Sun, has an age of 625-750 Myr, and forms one of the fundamental rungs in the distance ladder; understanding the planet population in such a well-studied cluster can help us understand and set contraints on the formation and evolution of planetary systems.

  17. Phase correlation and clustering of a nearest neighbour coupled oscillators system

    CERN Document Server

    Ei-Nashar, H F

    2002-01-01

    We investigated the phases in a system of nearest neighbour coupled oscillators before complete synchronization in frequency occurs. We found that when oscillators under the influence of coupling form a cluster of the same time-average frequency, their phases start to correlate. An order parameter, which measures this correlation, starts to grow at this stage until it reaches maximum. This means that a time-average phase locked state is reached between the oscillators inside the cluster of the same time- average frequency. At this strength the cluster attracts individual oscillators or a cluster to join in. We also observe that clustering in averaged frequencies orders the phases of the oscillators. This behavior is found at all the transition points studied.

  18. Phase correlation and clustering of a nearest neighbour coupled oscillators system

    International Nuclear Information System (INIS)

    EI-Nashar, Hassan F.

    2002-09-01

    We investigated the phases in a system of nearest neighbour coupled oscillators before complete synchronization in frequency occurs. We found that when oscillators under the influence of coupling form a cluster of the same time-average frequency, their phases start to correlate. An order parameter, which measures this correlation, starts to grow at this stage until it reaches maximum. This means that a time-average phase locked state is reached between the oscillators inside the cluster of the same time- average frequency. At this strength the cluster attracts individual oscillators or a cluster to join in. We also observe that clustering in averaged frequencies orders the phases of the oscillators. This behavior is found at all the transition points studied. (author)

  19. Cluster Synchronization of Diffusively Coupled Nonlinear Systems: A Contraction-Based Approach

    Science.gov (United States)

    Aminzare, Zahra; Dey, Biswadip; Davison, Elizabeth N.; Leonard, Naomi Ehrich

    2018-04-01

    Finding the conditions that foster synchronization in networked nonlinear systems is critical to understanding a wide range of biological and mechanical systems. However, the conditions proved in the literature for synchronization in nonlinear systems with linear coupling, such as has been used to model neuronal networks, are in general not strict enough to accurately determine the system behavior. We leverage contraction theory to derive new sufficient conditions for cluster synchronization in terms of the network structure, for a network where the intrinsic nonlinear dynamics of each node may differ. Our result requires that network connections satisfy a cluster-input-equivalence condition, and we explore the influence of this requirement on network dynamics. For application to networks of nodes with FitzHugh-Nagumo dynamics, we show that our new sufficient condition is tighter than those found in previous analyses that used smooth or nonsmooth Lyapunov functions. Improving the analytical conditions for when cluster synchronization will occur based on network configuration is a significant step toward facilitating understanding and control of complex networked systems.

  20. Nuclear clustering - a cluster core model study

    International Nuclear Information System (INIS)

    Paul Selvi, G.; Nandhini, N.; Balasubramaniam, M.

    2015-01-01

    Nuclear clustering, similar to other clustering phenomenon in nature is a much warranted study, since it would help us in understanding the nature of binding of the nucleons inside the nucleus, closed shell behaviour when the system is highly deformed, dynamics and structure at extremes. Several models account for the clustering phenomenon of nuclei. We present in this work, a cluster core model study of nuclear clustering in light mass nuclei

  1. Electronic licensing filing system development and implementation experience

    International Nuclear Information System (INIS)

    Walderhaug, J.

    1993-01-01

    The Electronic Licensing Filing System (ELFS) is a microcomputer-based integrated document search and retrieval system for the Nuclear Regulatory Affairs Division of Southern California Edison (SCE). ELFS allows the user access to the current licensing basis of a subject by providing an easily searchable electronic information data base consisting of regulatory correspondence, design-bases documentation, licensing documents [updated final safety and analysis report (UFSAR) and technical specifications], and regulatory guidance or directives [10CFR, generic letters, bulletins, notices, circulars, regulatory guides, policy statements, and selected US Nuclear Regulatory Commission (NRC) regulations]. It is used in the preparation of correspondence and submittals to the NRC, 50.59 safety evaluations, design-bases reconstitution, and commitment tracking and management

  2. Pseudo-potential method for taking into account the Pauli principle in cluster systems

    International Nuclear Information System (INIS)

    Krasnopol'skii, V.M.; Kukulin, V.I.

    1975-01-01

    In order to take account of the Pauli principle in cluster systems (such as 3α, α + α + n) a convenient method of renormalization of the cluster-cluster deep attractive potentials with forbidden states is suggested. The renormalization consists of adding projectors upon the occupied states with an infinite coupling constant to the initial deep potential which means that we pass to pseudo-potentials. The pseudo-potential approach in projecting upon the noneigenstates is shown to be equivalent to the orthogonality condition model of Saito et al. The orthogonality of the many-particle wave function to the forbidden states of each two-cluster sub-system is clearly demonstrated

  3. 75 FR 48629 - Electronic Tariff Filing System (ETFS)

    Science.gov (United States)

    2010-08-11

    ... composition requirements of our rules. This would ensure that all tariffs have a basic uniformity that will... transmittal: (1) A summary of the filing's basic rates, terms and conditions; (2) A statement concerning... filing, the basis of ratemaking employed, and economic information to support the changed or new matter...

  4. Cluster analysis in systems of magnetic spheres and cubes

    Energy Technology Data Exchange (ETDEWEB)

    Pyanzina, E.S., E-mail: elena.pyanzina@urfu.ru [Ural Federal University, Lenin Av. 51, Ekaterinburg (Russian Federation); Gudkova, A.V. [Ural Federal University, Lenin Av. 51, Ekaterinburg (Russian Federation); Donaldson, J.G. [University of Vienna, Sensengasse 8, Vienna (Austria); Kantorovich, S.S. [Ural Federal University, Lenin Av. 51, Ekaterinburg (Russian Federation); University of Vienna, Sensengasse 8, Vienna (Austria)

    2017-06-01

    In the present work we use molecular dynamics simulations and graph-theory based cluster analysis to compare self-assembly in systems of magnetic spheres, and cubes where the dipole moment is oriented along the side of the cube in the [001] crystallographic direction. We show that under the same conditions cubes aggregate far less than their spherical counterparts. This difference can be explained in terms of the volume of phase space in which the formation of the bond is thermodynamically advantageous. It follows that this volume is much larger for a dipolar sphere than for a dipolar cube. - Highlights: • A comparison of the degree of self-assembly in systems of magnetic spheres and cubes. • Spheres are more likely to form larger clusters than cubes. • Differences in microstructure will manifest in the magnetic response of each system.

  5. Privacy Act System of Records: Freedom of Information Act Request and Appeal File, EPA-9

    Science.gov (United States)

    Learn more about the Freedom of Information Act Request and Appeal File System, including who is covered in the system, the purpose of data collection, routine uses for the system's records, and other security procedures.

  6. Analyzing Log Files using Data-Mining

    Directory of Open Access Journals (Sweden)

    Marius Mihut

    2008-01-01

    Full Text Available Information systems (i.e. servers, applications and communication devices create a large amount of monitoring data that are saved as log files. For analyzing them, a data-mining approach is helpful. This article presents the steps which are necessary for creating an ‘analyzing instrument’, based on an open source software called Waikato Environment for Knowledge Analysis (Weka [1]. For exemplification, a system log file created by a Windows-based operating system, is used as input file.

  7. CryptoCache: A Secure Sharable File Cache for Roaming Users

    DEFF Research Database (Denmark)

    Jensen, Christian D.

    2000-01-01

    . Conventional distributed file systems cache everything locally or not at all; there is no possibility to cache files on nearby nodes.In this paper we present the design of a secure cache system called CryptoCache that allows roaming users to cache files on untrusted file hosting servers. The system allows...... flexible sharing of cached files among unauthenticated users, i.e. unlike most distributed file systems CryptoCache does not require a global authentication framework.Files are encrypted when they are transferred over the network and while stored on untrusted servers. The system uses public key......Small mobile computers are now sufficiently powerful to run many applications, but storage capacity remains limited so working files cannot be cached or stored locally. Even if files can be stored locally, the mobile device is not powerful enough to act as server in collaborations with other users...

  8. Electrochemical impedance spectroscopy investigation on the clinical lifetime of ProTaper rotary file system.

    Science.gov (United States)

    Penta, Virgil; Pirvu, Cristian; Demetrescu, Ioana

    2014-01-01

    The main objective of the current paper is to show that electrochemical impedance spectroscopy (EIS) could be a method for evaluating and predicting of ProTaper rotary file system clinical lifespan. This particular aspect of everyday use of the endodontic files is of great importance in each dental practice and has profound clinical implications. The method used for quantification resides in the electrochemical impedance spectroscopy theory and has in its main focus the characteristics of the surface titanium oxide layer. This electrochemical technique has been adapted successfully to identify the quality of the Ni-Ti files oxide layer. The modification of this protective layer induces changes in corrosion behavior of the alloy modifying the impedance value of the file. In order to assess the method, 14 ProTaper sets utilized on different patients in a dental clinic have been submitted for testing using EIS. The information obtained in regard to the surface oxide layer has offered an indication of use and proves that the said layer evolves with each clinical application. The novelty of this research is related to an electrochemical technique successfully adapted for Ni-Ti file investigation and correlation with surface and clinical aspects.

  9. A Self-Organizing Spatial Clustering Approach to Support Large-Scale Network RTK Systems

    Directory of Open Access Journals (Sweden)

    Lili Shen

    2018-06-01

    Full Text Available The network real-time kinematic (RTK technique can provide centimeter-level real time positioning solutions and play a key role in geo-spatial infrastructure. With ever-increasing popularity, network RTK systems will face issues in the support of large numbers of concurrent users. In the past, high-precision positioning services were oriented towards professionals and only supported a few concurrent users. Currently, precise positioning provides a spatial foundation for artificial intelligence (AI, and countless smart devices (autonomous cars, unmanned aerial-vehicles (UAVs, robotic equipment, etc. require precise positioning services. Therefore, the development of approaches to support large-scale network RTK systems is urgent. In this study, we proposed a self-organizing spatial clustering (SOSC approach which automatically clusters online users to reduce the computational load on the network RTK system server side. The experimental results indicate that both the SOSC algorithm and the grid algorithm can reduce the computational load efficiently, while the SOSC algorithm gives a more elastic and adaptive clustering solution with different datasets. The SOSC algorithm determines the cluster number and the mean distance to cluster center (MDTCC according to the data set, while the grid approaches are all predefined. The side-effects of clustering algorithms on the user side are analyzed with real global navigation satellite system (GNSS data sets. The experimental results indicate that 10 km can be safely used as the cluster radius threshold for the SOSC algorithm without significantly reducing the positioning precision and reliability on the user side.

  10. Discovery of four gravitational lensing systems by clusters in the SDSS DR6

    International Nuclear Information System (INIS)

    Wen Zhonglue; Han Jinlin; Xu Xiangyang; Jiang Yunying; Guo Zhiqing; Wang Pengfei; Liu Fengshan

    2009-01-01

    We report the discovery of 4 strong gravitational lensing systems by visual inspections of the Sloan Digital Sky Survey images of galaxy clusters in Data Release 6 (SDSS DR6). Two of the four systems show Einstein rings while the others show tangential giant arcs. These arcs or rings have large angular separations (> 8) from the bright central galaxies and show bluer color compared with the red cluster galaxies. In addition, we found 5 probable and 4 possible lenses by galaxy clusters. (letters)

  11. Comparison of the amount of apical debris extrusion associated with different retreatment systems and supplementary file application during retreatment process.

    Science.gov (United States)

    Çiçek, Ersan; Koçak, Mustafa Murat; Koçak, Sibel; Sağlam, Baran Can

    2016-01-01

    The type of instrument affects the amount of debris extruded. The aim of this study was to compare the effect of retreatment systems and supplementary file application on the amount of apical debris extrusion. Forty-eight extracted mandibular premolars with a single canal and similar length were selected. The root canals were prepared with the ProTaper Universal system with a torque-controlled engine. The root canals were dried and were obturated using Gutta-percha and sealer. The specimens were randomly divided into four equal groups according to the retreatment procedures (Group 1, Mtwo retreatment files; Group 2, Mtwo retreatment files + Mtwo rotary file #30 supplementary file; Group 3, ProTaper Universal retreatment (PTUR) files; and Group 4, PTUR files + ProTaper F3 supplementary file). The extruded debris during instrumentation were collected into preweighed Eppendorf tubes. The amount of apically extruded debris was calculated by subtracting the initial weight of the tube from the final weight. Three consecutive weights were obtained for each tube. No statistically significant difference was found in the amount of apically extruded debris between Groups 1 and 3 (P = 0.590). A significant difference was observed between Groups 1 and 2 (P file significantly increased the amount of apically extruded debris.

  12. Evaluation of apical extrusion of debris and irrigant using two new reciprocating and one continuous rotation single file systems.

    Science.gov (United States)

    Nayak, Gurudutt; Singh, Inderpreet; Shetty, Shashit; Dahiya, Surya

    2014-05-01

    Apical extrusion of debris and irrigants during cleaning and shaping of the root canal is one of the main causes of periapical inflammation and postoperative flare-ups. The purpose of this study was to quantitatively measure the amount of debris and irrigants extruded apically in single rooted canals using two reciprocating and one rotary single file nickel-titanium instrumentation systems. Sixty human mandibular premolars, randomly assigned to three groups (n = 20) were instrumented using two reciprocating (Reciproc and Wave One) and one rotary (One Shape) single-file nickel-titanium systems. Bidistilled water was used as irrigant with traditional needle irrigation delivery system. Eppendorf tubes were used as test apparatus for collection of debris and irrigant. The volume of extruded irrigant was collected and quantified via 0.1-mL increment measure supplied on the disposable plastic insulin syringe. The liquid inside the tubes was dried and the mean weight of debris was assessed using an electronic microbalance. The data were statistically analysed using Kruskal-Wallis nonparametric test and Mann Whitney U test with Bonferroni adjustment. P-values less than 0.05 were considered significant. The Reciproc file system produced significantly more debris compared with OneShape file system (P0.05). Extrusion of irrigant was statistically insignificant irrespective of the instrument or instrumentation technique used (P >0.05). Although all systems caused apical extrusion of debris and irrigant, continuous rotary instrumentation was associated with less extrusion as compared with the use of reciprocating file systems.

  13. 76 FR 9780 - Notification of Deletion of System of Records; EPA Parking Control Office File (EPA-10) and EPA...

    Science.gov (United States)

    2011-02-22

    ... System of Records; EPA Parking Control Office File (EPA-10) and EPA Transit and Guaranteed Ride Home... Environmental Protection Agency (EPA) is deleting the systems of records for EPA Parking Control Office File... through the EPA Internet under the ``Federal Register'' listings at http://www.epa.gov/fedrgstr/ . Dated...

  14. DIGITAL ONCOLOGY PATIENT RECORD - HETEROGENEOUS FILE BASED APPROACH

    Directory of Open Access Journals (Sweden)

    Nikolay Sapundzhiev

    2010-12-01

    Full Text Available Introduction: Oncology patients need extensive follow-up and meticulous documentation. The aim of this study was to introduce a simple, platform independent file based system for documentation of diagnostic and therapeutic procedures in oncology patients and test its function.Material and methods: A file-name based system of the type M1M2M3.F2 was introduced, where M1 is a unique identifier for the patient, M2 is the date of the clinical intervention/event, M3 is an identifier for the author of the medical record and F2 is the specific software generated file-name extension.Results: This system is in use at 5 institutions, where a total of 11 persons on 14 different workstations inputted 16591 entries (files for 2370. The merge process was tested on 2 operating systems - when copied together all files sort up as expected by patient, and for each patient in a chronological order, providing a digital cumulative patient record, which contains heterogeneous file formats.Conclusion: The file based approach for storing heterogeneous digital patient related information is an reliable system, which can handle open-source, proprietary, general and custom file formats and seems to be easily scalable. Further development of software for automatic checks of the integrity and searching and indexing of the files is expected to produce a more user-friendly environment

  15. An asynchronous writing method for restart files in the gysela code in prevision of exascale systems*

    Directory of Open Access Journals (Sweden)

    Thomine O.

    2013-12-01

    Full Text Available The present work deals with an optimization procedure developed in the full-f global GYrokinetic SEmi-LAgrangian code (GYSELA. Optimizing the writing of the restart files is necessary to reduce the computing impact of crashes. These files require a very large memory space, and particularly so for very large mesh sizes. The limited bandwidth of the data pipe between the comput- ing nodes and the storage system induces a non-scalable part in the GYSELA code, which increases with the mesh size. Indeed the transfer time of RAM to data depends linearly on the files size. The necessity of non synchronized writing-in-file procedure is therefore crucial. A new GYSELA module has been developed. This asynchronous procedure allows the frequent writ- ing of the restart files, whilst preventing a severe slowing down due to the limited writing bandwidth. This method has been improved to generate a checksum control of the restart files, and automatically rerun the code in case of a crash for any cause.

  16. Estimating Motor Carrier Management Information System Crash File Underreporting from Carrier Records.

    Science.gov (United States)

    2017-08-01

    This FMCSA-sponsored research investigated the claim that motor carriers have a substantial number of crashes in their own records that are not contained in the Motor Carrier Management Information System (MCMIS) crash file. Based on the results of t...

  17. Schedulability Analysis and Optimization for the Synthesis of Multi-Cluster Distributed Embedded Systems

    DEFF Research Database (Denmark)

    Pop, Paul; Eles, Petru; Peng, Zebo

    2003-01-01

    We present an approach to schedulability analysis for the synthesis of multi-cluster distributed embedded systems consisting of time-triggered and event-triggered clusters, interconnected via gateways. We have also proposed a buffer size and worst case queuing delay analysis for the gateways......, responsible for routing inter-cluster traffic. Optimization heuristics for the priority assignment and synthesis of bus access parameters aimed at producing a schedulable system with minimal buffer needs have been proposed. Extensive experiments and a real-life example show the efficiency of our approaches....

  18. Schedulability Analysis and Optimization for the Synthesis of Multi-Cluster Distributed Embedded Systems

    DEFF Research Database (Denmark)

    Pop, Paul; Eles, Petru; Peng, Zebo

    2003-01-01

    An approach to schedulability analysis for the synthesis of multi-cluster distributed embedded systems consisting of time-triggered and event-triggered clusters, interconnected via gateways, is presented. A buffer size and worst case queuing delay analysis for the gateways, responsible for routing...... inter-cluster traffic, is also proposed. Optimisation heuristics for the priority assignment and synthesis of bus access parameters aimed at producing a schedulable system with minimal buffer needs have been proposed. Extensive experiments and a real-life example show the efficiency of the approaches....

  19. DEVELOPMENT OF TRANSPORT SUBSYSTEM STREAMING DATA REPLICATION CLUSTER IN CORBA-SYSTEM WITH ZEROMQ TECHNOLOGY

    Directory of Open Access Journals (Sweden)

    F. A. Kozlov

    2013-03-01

    Full Text Available The article deals with the peculiarities of distributed cluster system creation with streaming data replication. Ways of replication cluster implementation in CORBA-systems with ZeroMq technology are presented. Major advantages of ZeroMQ technology over similar technologies are considered in this type distributed systems creation.

  20. 75 FR 17700 - Energy Efficient Building Systems Regional Innovation Cluster Initiative-Joint Federal Funding...

    Science.gov (United States)

    2010-04-07

    ... economically dynamic regional innovation cluster focused on energy efficient buildings technologies and systems...-risk, high-reward research that overcomes technology challenges through approaches that span basic... DEPARTMENT OF ENERGY Energy Efficient Building Systems Regional Innovation Cluster Initiative...

  1. Evaluated neutronic file for indium

    International Nuclear Information System (INIS)

    Smith, A.B.; Chiba, S.; Smith, D.L.; Meadows, J.W.; Guenther, P.T.; Lawson, R.D.; Howerton, R.J.

    1990-01-01

    A comprehensive evaluated neutronic data file for elemental indium is documented. This file, extending from 10 -5 eV to 20 MeV, is presented in the ENDF/B-VI format, and contains all neutron-induced processes necessary for the vast majority of neutronic applications. In addition, an evaluation of the 115 In(n,n') 116m In dosimetry reaction is presented as a separate file. Attention is given in quantitative values, with corresponding uncertainty information. These files have been submitted for consideration as a part of the ENDF/B-VI national evaluated-file system. 144 refs., 10 figs., 4 tabs

  2. Effect of Instrumentation Length and Instrumentation Systems: Hand Versus Rotary Files on Apical Crack Formation - An In vitro Study.

    Science.gov (United States)

    Devale, Madhuri R; Mahesh, M C; Bhandary, Shreetha

    2017-01-01

    Stresses generated during root canal instrumentation have been reported to cause apical cracks. The smaller, less pronounced defects like cracks can later propagate into vertical root fracture, when the tooth is subjected to repeated stresses from endodontic or restorative procedures. This study evaluated occurrence of apical cracks with stainless steel hand files, rotary NiTi RaCe and K3 files at two different instrumentation lengths. In the present in vitro study, 60 mandibular premolars were mounted in resin blocks with simulated periodontal ligament. Apical 3 mm of the root surfaces were exposed and stained using India ink. Preoperative images of root apices were obtained at 100x using stereomicroscope. The teeth were divided into six groups of 10 each. First two groups were instrumented with stainless steel files, next two groups with rotary NiTi RaCe files and the last two groups with rotary NiTi K3 files. The instrumentation was carried out till the apical foramen (Working Length-WL) and 1 mm short of the apical foramen (WL-1) with each file system. After root canal instrumentation, postoperative images of root apices were obtained. Preoperative and postoperative images were compared and the occurrence of cracks was recorded. Descriptive statistical analysis and Chi-square tests were used to analyze the results. Apical root cracks were seen in 30%, 35% and 20% of teeth instrumented with K-files, RaCe files and K3 files respectively. There was no statistical significance among three instrumentation systems in the formation of apical cracks (p=0.563). Apical cracks were seen in 40% and 20% of teeth instrumented with K-files; 60% and 10% of teeth with RaCe files and 40% and 0% of teeth with K3 files at WL and WL-1 respectively. For groups instrumented with hand files there was no statistical significance in number of cracks at WL and WL-1 (p=0.628). But for teeth instrumented with RaCe files and K3 files significantly more number of cracks were seen at WL than

  3. FPGA cluster for high-performance AO real-time control system

    Science.gov (United States)

    Geng, Deli; Goodsell, Stephen J.; Basden, Alastair G.; Dipper, Nigel A.; Myers, Richard M.; Saunter, Chris D.

    2006-06-01

    Whilst the high throughput and low latency requirements for the next generation AO real-time control systems have posed a significant challenge to von Neumann architecture processor systems, the Field Programmable Gate Array (FPGA) has emerged as a long term solution with high performance on throughput and excellent predictability on latency. Moreover, FPGA devices have highly capable programmable interfacing, which lead to more highly integrated system. Nevertheless, a single FPGA is still not enough: multiple FPGA devices need to be clustered to perform the required subaperture processing and the reconstruction computation. In an AO real-time control system, the memory bandwidth is often the bottleneck of the system, simply because a vast amount of supporting data, e.g. pixel calibration maps and the reconstruction matrix, need to be accessed within a short period. The cluster, as a general computing architecture, has excellent scalability in processing throughput, memory bandwidth, memory capacity, and communication bandwidth. Problems, such as task distribution, node communication, system verification, are discussed.

  4. Clustering of tethered satellite system simulation data by an adaptive neuro-fuzzy algorithm

    Science.gov (United States)

    Mitra, Sunanda; Pemmaraju, Surya

    1992-01-01

    Recent developments in neuro-fuzzy systems indicate that the concepts of adaptive pattern recognition, when used to identify appropriate control actions corresponding to clusters of patterns representing system states in dynamic nonlinear control systems, may result in innovative designs. A modular, unsupervised neural network architecture, in which fuzzy learning rules have been embedded is used for on-line identification of similar states. The architecture and control rules involved in Adaptive Fuzzy Leader Clustering (AFLC) allow this system to be incorporated in control systems for identification of system states corresponding to specific control actions. We have used this algorithm to cluster the simulation data of Tethered Satellite System (TSS) to estimate the range of delta voltages necessary to maintain the desired length rate of the tether. The AFLC algorithm is capable of on-line estimation of the appropriate control voltages from the corresponding length error and length rate error without a priori knowledge of their membership functions and familarity with the behavior of the Tethered Satellite System.

  5. Evaluation of apical extrusion of debris and irrigant using two new reciprocating and one continuous rotation single file systems.

    Directory of Open Access Journals (Sweden)

    Gurudutt Nayak

    2014-06-01

    Full Text Available Apical extrusion of debris and irrigants during cleaning and shaping of the root canal is one of the main causes of periapical inflammation and postoperative flare-ups. The purpose of this study was to quantitatively measure the amount of debris and irrigants extruded apically in single rooted canals using two reciprocating and one rotary single file nickel-titanium instrumentation systems.Sixty human mandibular premolars, randomly assigned to three groups (n = 20 were instrumented using two reciprocating (Reciproc and Wave One and one rotary (One Shape single-file nickel-titanium systems. Bidistilled water was used as irrigant with traditional needle irrigation delivery system. Eppendorf tubes were used as test apparatus for collection of debris and irrigant. The volume of extruded irrigant was collected and quantified via 0.1-mL increment measure supplied on the disposable plastic insulin syringe. The liquid inside the tubes was dried and the mean weight of debris was assessed using an electronic microbalance. The data were statistically analysed using Kruskal-Wallis nonparametric test and Mann Whitney U test with Bonferroni adjustment. P-values less than 0.05 were considered significant.The Reciproc file system produced significantly more debris compared with OneShape file system (P0.05. Extrusion of irrigant was statistically insignificant irrespective of the instrument or instrumentation technique used (P >0.05.Although all systems caused apical extrusion of debris and irrigant, continuous rotary instrumentation was associated with less extrusion as compared with the use of reciprocating file systems.

  6. Evaluation of Single File Systems Reciproc, Oneshape, and WaveOne using Cone Beam Computed Tomography -An In Vitro Study.

    Science.gov (United States)

    Dhingra, Annil; Ruhal, Nidhi; Miglani, Anjali

    2015-04-01

    Successful endodontic therapy depends on many factor, one of the most important step in any root canal treatment is root canal preparation. In addition, respecting the original shape of the canal is of the same importance; otherwise, canal aberrations such as transportation will be created. The purpose of this study is to compare and evaluate Reciprocating WaveOne ,Reciproc and Rotary Oneshape Single File Instrumentation System On Cervical Dentin Thickness, Cross Sectional Area and Canal Transportation on First Mandibular Molar Using Cone Beam Computed Tomography. Sixty Mandibular First Molars extracted due to periodontal reason was collected from the Department of Oral and Maxillofacial. Teeth were prepared using one rotary and two reciprocating single file system. Teeth were divided into 3 groups 20 teeth in each group. Pre instrumentation and Post instrumentation scans was done and evaluated for three parameters Canal Transportation, Cervical Dentinal Thickness, Cross-sectional Area. Results were analysed statistically using ANOVA, Post-Hoc Tukey analysis. The change in cross-sectional area after filing showed significant difference at 0mm, 1mm, 2mm and 7mm (pfile system over a distance of 7 mm (starting from 0mm and then evaluation at 1mm, 2mm, 3mm, 5mm and 7mm), the results showed a significant difference among the file systems at various lengths (p= 0.014, 0.046, 0.004, 0.028, 0.005 & 0.029 respectively). Mean value of cervical dentinal removal is maximum at all the levels for oneshape and minimum for waveone showing the better quality of waveone and reciproc over oneshape file system. Significant difference was found at 9mm, 11mm and 12mm between all the three file systems (p<0.001,< 0.001, <0.001). It was concluded that reciprocating motion is better than rotary motion in all the three parameters Canal Transportation, Cross-sectional Area, Cervical Dentinal Thickness.

  7. Influence of system temperature on the micro-structures and dynamics of dust clusters in dusty plasmas

    Energy Technology Data Exchange (ETDEWEB)

    Song, Y. L.; Huang, F., E-mail: huangfeng@cau.edu.cn [College of Science, China Agricultural University, Beijing 100083 (China); He, Y. F.; Wu, L. [College of Information and Electrical Engineering, China Agricultural University, Beijing 100083 (China); Liu, Y. H. [School of Physics and Optoelectronic Engineering, Ludong University, Yantai 264025 (China); Chen, Z. Y. [Department of Physics, Beijing University of Chemical Technology, Beijing 100029 (China); Yu, M. Y. [Institute for Fusion Theory and Simulation, Zhejiang University, Hangzhou 310027 (China); Institute for Theoretical Physics I, Ruhr University, D-44801 Bochum (Germany)

    2015-06-15

    Influence of the system temperature on the micro-structures and dynamics of dust clusters in dusty plasmas is investigated through laboratory experiment and molecular dynamics simulation. The micro-structures, defect numbers, and pair correlation function of the dust clusters are studied for different system temperatures. The dust grains' trajectories, the mean square displacement, and the corresponding self-diffusion coefficient of the clusters are calculated for different temperatures for illustrating the phase properties of the dust clusters. The simulation results confirm that with the increase in system temperature, the micro-structures and dynamics of dust clusters are gradually changed, which qualitatively agree with experimental results.

  8. FURTHER DEFINITION OF THE MASS-METALLICITY RELATION IN GLOBULAR CLUSTER SYSTEMS AROUND BRIGHTEST CLUSTER GALAXIES

    International Nuclear Information System (INIS)

    Cockcroft, Robert; Harris, William E.; Wehner, Elizabeth M. H.; Whitmore, Bradley C.; Rothberg, Barry

    2009-01-01

    We combine the globular cluster (GC) data for 15 brightest cluster galaxies and use this material to trace the mass-metallicity relations (MMRs) in their globular cluster systems (GCSs). This work extends previous studies which correlate the properties of the MMR with those of the host galaxy. Our combined data sets show a mean trend for the metal-poor subpopulation that corresponds to a scaling of heavy-element abundance with cluster mass Z ∼ M 0.30±0.05 . No trend is seen for the metal-rich subpopulation which has a scaling relation that is consistent with zero. We also find that the scaling exponent is independent of the GCS specific frequency and host galaxy luminosity, except perhaps for dwarf galaxies. We present new photometry in (g',i') obtained with Gemini/GMOS for the GC populations around the southern giant ellipticals NGC 5193 and IC 4329. Both galaxies have rich cluster populations which show up as normal, bimodal sequences in the color-magnitude diagram. We test the observed MMRs and argue that they are statistically real, and not an artifact caused by the method we used. We also argue against asymmetric contamination causing the observed MMR as our mean results are no different from other contamination-free studies. Finally, we compare our method to the standard bimodal fitting method (KMM or RMIX) and find our results are consistent. Interpretation of these results is consistent with recent models for GC formation in which the MMR is determined by GC self-enrichment during their brief formation period.

  9. Pathway enrichment and co-expression cluster analysis - FANTOM5 | LSDB Archive [Life Science Database Archive metadata

    Lifescience Database Archive (English)

    Full Text Available switchLanguage; BLAST Search Image Search Home About Archive Update History Data List Contact us FANTOM...lusters File URL: ftp://ftp.biosciencedbc.jp/archive/fantom5/datafiles/phase1.3/extra/Co-expression_clusters...ite Policy | Contact Us Pathway enrichment and co-expression cluster analysis - FANTOM5 | LSDB Archive ...

  10. Comparison of quality of obturation and instrumentation time using hand files and two rotary file systems in primary molars: A single-blinded randomized controlled trial.

    Science.gov (United States)

    Govindaraju, Lavanya; Jeevanandan, Ganesh; Subramanian, E M G

    2017-01-01

    In permanent dentition, different rotary systems are used for canal cleaning and shaping. Rotary instrumentation in pediatric dentistry is an emerging concept. A very few studies have compared the efficiency of rotary instrumentation for canal preparation in primary teeth. Hence, this study was performed to compare the obturation quality and instrumentation time of two rotary files systems - Protaper, Mtwo with hand files in primary molars. Forty-five primary mandibular molars were randomly allotted to one of the three groups. Instrumentation was done using K-files in Group 1; Protaper in Group 2; and Mtwo in Group 3. Instrumentation time was recorded. The canal filling quality was assessed as underfill, optimal fill, and overfill. Statistical analysis was done using Chi-square, ANOVA, and post hoc Tukey test. No significant difference was observed in the quality of obturation among three groups. Intergroup comparison of the instrumentation time showed a statistically significant difference between the three groups. The use of rotary instrumentation in primary teeth results in marked reduction in the instrumentation time and improves the quality of obturation.

  11. EFEKTIVITAS MANAJEMEN E-SYSTEM DALAM PELAYANAN PERPAJAKAN TERHADAP KEPATUHAN WAJIB PAJAK DALAM PELAPORAN SPT (E-FILING

    Directory of Open Access Journals (Sweden)

    Pujo Gunarso

    2017-04-01

    Full Text Available This purposes of this study is to analyze the influence of management effectiveness of e-taxation system to tax compliance in terms of reporting such as (e-filing. This research was conducted at the Tax Office Batu region in the tax year 2015. The research method uses a descriptive quantitative method. The results of this study explains that the management of e-filing for the realization of the Key Performance Index e-filing using point based on the rules of the Directorate General of Taxation number: PER-03 / PJ / 2015 has been effective. �

  12. New data visualization of the LHC Era Monitoring (Lemon) system

    International Nuclear Information System (INIS)

    Ivan, Fedorko; Veronique, Lefebure; Daniel, Lenkes; Omar, Pera Mira

    2012-01-01

    In the last few years, new requirements have been received for visualization of monitoring data: advanced graphics, flexibility in configuration and decoupling of the presentation layer from the monitoring repository. Lemonweb is the data visualization component of the LHC Era Monitoring (Lemon) system. Lemonweb consists of two subcomponents: a data collector and a web visualization interface. The data collector is a daemon, implemented in Python, responsible for data gathering from the central monitoring repository and storing into time series data structures. Data is stored on disk in Round Robin Database (RRD) files: one file per monitored entity, with set of entity related metrics. Entities may be grouped into a hierarchical structure, called “clusters” and supporting mathematical operations over entities and clusters (e.g. cluster A + cluster B /clusters C – entity XY). Using the configuration information, a cluster definition is evaluated in the collector engine and, at runtime, a sequence of data selects is built, to optimize access to the central monitoring repository. In this article, an overview of the design and architecture as well as highlights of some implemented features will be presented.

  13. Ground-Based Global Navigation Satellite System (GNSS) GLONASS Broadcast Ephemeris Data (hourly files) from NASA CDDIS

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset consists of ground-based Global Navigation Satellite System (GNSS) GLObal NAvigation Satellite System (GLONASS) Broadcast Ephemeris Data (hourly files)...

  14. Clustering and Recurring Anomaly Identification: Recurring Anomaly Detection System (ReADS)

    Science.gov (United States)

    McIntosh, Dawn

    2006-01-01

    This viewgraph presentation reviews the Recurring Anomaly Detection System (ReADS). The Recurring Anomaly Detection System is a tool to analyze text reports, such as aviation reports and maintenance records: (1) Text clustering algorithms group large quantities of reports and documents; Reduces human error and fatigue (2) Identifies interconnected reports; Automates the discovery of possible recurring anomalies; (3) Provides a visualization of the clusters and recurring anomalies We have illustrated our techniques on data from Shuttle and ISS discrepancy reports, as well as ASRS data. ReADS has been integrated with a secure online search

  15. Security in the CernVM File System and the Frontier Distributed Database Caching System

    International Nuclear Information System (INIS)

    Dykstra, D; Blomer, J

    2014-01-01

    Both the CernVM File System (CVMFS) and the Frontier Distributed Database Caching System (Frontier) distribute centrally updated data worldwide for LHC experiments using http proxy caches. Neither system provides privacy or access control on reading the data, but both control access to updates of the data and can guarantee the authenticity and integrity of the data transferred to clients over the internet. CVMFS has since its early days required digital signatures and secure hashes on all distributed data, and recently Frontier has added X.509-based authenticity and integrity checking. In this paper we detail and compare the security models of CVMFS and Frontier.

  16. Security in the CernVM File System and the Frontier Distributed Database Caching System

    Science.gov (United States)

    Dykstra, D.; Blomer, J.

    2014-06-01

    Both the CernVM File System (CVMFS) and the Frontier Distributed Database Caching System (Frontier) distribute centrally updated data worldwide for LHC experiments using http proxy caches. Neither system provides privacy or access control on reading the data, but both control access to updates of the data and can guarantee the authenticity and integrity of the data transferred to clients over the internet. CVMFS has since its early days required digital signatures and secure hashes on all distributed data, and recently Frontier has added X.509-based authenticity and integrity checking. In this paper we detail and compare the security models of CVMFS and Frontier.

  17. 47 CFR 61.14 - Method of filing publications.

    Science.gov (United States)

    2010-10-01

    ... 47 Telecommunication 3 2010-10-01 2010-10-01 false Method of filing publications. 61.14 Section 61...) TARIFFS Rules for Electronic Filing § 61.14 Method of filing publications. (a) Publications filed... date of a publication received by the Electronic Tariff Filing System will be determined by the date...

  18. Multi-agent grid system Agent-GRID with dynamic load balancing of cluster nodes

    Science.gov (United States)

    Satymbekov, M. N.; Pak, I. T.; Naizabayeva, L.; Nurzhanov, Ch. A.

    2017-12-01

    In this study the work presents the system designed for automated load balancing of the contributor by analysing the load of compute nodes and the subsequent migration of virtual machines from loaded nodes to less loaded ones. This system increases the performance of cluster nodes and helps in the timely processing of data. A grid system balances the work of cluster nodes the relevance of the system is the award of multi-agent balancing for the solution of such problems.

  19. Active Galactic Nuclei Feedback and Clusters Biman B. Nath

    Indian Academy of Sciences (India)

    entropy into the ICM gas. The recently determined universal pressure pro- file of the ICM gas has been used and after comparing with the entropy profile of the gas from gravitational effects of the dark matter halo, the additional entropy injected by non-gravitational sources, as a function of the total cluster mass is determined.

  20. Publicity in secured transactions law: Towards a European public notice filing system for non-possessory security rights in movable assets?

    NARCIS (Netherlands)

    Hamwijk, D.J.Y.

    2014-01-01

    There is a growing call to introduce a European public notice filing system for security rights in movable goods comparable to the notice filing system of Art. 9 UCC. A proposal to this effect has been adopted in Book IX DCFR, which represents a comprehensive framework of rules for proprietary

  1. On-Board File Management and Its Application in Flight Operations

    Science.gov (United States)

    Kuo, N.

    1998-01-01

    In this paper, the author presents the minimum functions required for an on-board file management system. We explore file manipulation processes and demonstrate how the file transfer along with the file management system will be utilized to support flight operations and data delivery.

  2. Diametrical clustering for identifying anti-correlated gene clusters.

    Science.gov (United States)

    Dhillon, Inderjit S; Marcotte, Edward M; Roshan, Usman

    2003-09-01

    Clustering genes based upon their expression patterns allows us to predict gene function. Most existing clustering algorithms cluster genes together when their expression patterns show high positive correlation. However, it has been observed that genes whose expression patterns are strongly anti-correlated can also be functionally similar. Biologically, this is not unintuitive-genes responding to the same stimuli, regardless of the nature of the response, are more likely to operate in the same pathways. We present a new diametrical clustering algorithm that explicitly identifies anti-correlated clusters of genes. Our algorithm proceeds by iteratively (i). re-partitioning the genes and (ii). computing the dominant singular vector of each gene cluster; each singular vector serving as the prototype of a 'diametric' cluster. We empirically show the effectiveness of the algorithm in identifying diametrical or anti-correlated clusters. Testing the algorithm on yeast cell cycle data, fibroblast gene expression data, and DNA microarray data from yeast mutants reveals that opposed cellular pathways can be discovered with this method. We present systems whose mRNA expression patterns, and likely their functions, oppose the yeast ribosome and proteosome, along with evidence for the inverse transcriptional regulation of a number of cellular systems.

  3. Clinical antibacterial effectiveness of the self-adjusting file system.

    Science.gov (United States)

    Neves, M A S; Rôças, I N; Siqueira, J F

    2014-04-01

    To evaluate in vivo the antibacterial effectiveness of the self-adjusting file (SAF) using molecular methods. Root canals from single-rooted teeth with apical periodontitis were instrumented using the SAF system under continuous irrigation with 2.5% NaOCl. DNA extracts from samples taken before and after instrumentation were subjected to quantitative analysis of total bacteria counts and levels of streptococci by quantitative real-time polymerase chain reaction (qPCR). The reverse-capture checkerboard assay was also used to identify 28 bacterial taxa before (S1) and after (S2) SAF instrumentation. SAF was also compared with a conventional hand nickel-titanium instrumentation technique for total bacterial reduction. Data from qPCR were analysed statistically within groups using the Wilcoxon matched pairs test and between groups using the Mann-Whitney U-test and the Fisher's exact test, with significance level set at P file significantly reduced the total bacterial counts from a mean number of 1.96 × 10(7) cells to 1.34 × 10(4) cells (P system was significantly superior to the 95.1% reduction obtained by hand instrumentation (P system succeeded in significantly reducing the streptococcal levels, but four cases still harboured these bacteria in S2. Checkerboard analysis revealed that not only streptococci but also some anaerobic and even as-yet-uncultivated bacteria may resist the effects of chemomechanical procedures. The SAF instrumentation system was highly effective in reducing bacterial populations from infected root canals and performed significantly better than hand instrumentation. However, because half of the samples still had detectable bacteria after preparation with SAF, supplementary disinfection is still required to maximize bacterial elimination. © 2013 International Endodontic Journal. Published by John Wiley & Sons Ltd.

  4. Effect of Instrumentation Length and Instrumentation Systems: Hand Versus Rotary Files on Apical Crack Formation – An In vitro Study

    Science.gov (United States)

    Mahesh, MC; Bhandary, Shreetha

    2017-01-01

    Introduction Stresses generated during root canal instrumentation have been reported to cause apical cracks. The smaller, less pronounced defects like cracks can later propagate into vertical root fracture, when the tooth is subjected to repeated stresses from endodontic or restorative procedures. Aim This study evaluated occurrence of apical cracks with stainless steel hand files, rotary NiTi RaCe and K3 files at two different instrumentation lengths. Materials and Methods In the present in vitro study, 60 mandibular premolars were mounted in resin blocks with simulated periodontal ligament. Apical 3 mm of the root surfaces were exposed and stained using India ink. Preoperative images of root apices were obtained at 100x using stereomicroscope. The teeth were divided into six groups of 10 each. First two groups were instrumented with stainless steel files, next two groups with rotary NiTi RaCe files and the last two groups with rotary NiTi K3 files. The instrumentation was carried out till the apical foramen (Working Length-WL) and 1 mm short of the apical foramen (WL-1) with each file system. After root canal instrumentation, postoperative images of root apices were obtained. Preoperative and postoperative images were compared and the occurrence of cracks was recorded. Descriptive statistical analysis and Chi-square tests were used to analyze the results. Results Apical root cracks were seen in 30%, 35% and 20% of teeth instrumented with K-files, RaCe files and K3 files respectively. There was no statistical significance among three instrumentation systems in the formation of apical cracks (p=0.563). Apical cracks were seen in 40% and 20% of teeth instrumented with K-files; 60% and 10% of teeth with RaCe files and 40% and 0% of teeth with K3 files at WL and WL-1 respectively. For groups instrumented with hand files there was no statistical significance in number of cracks at WL and WL-1 (p=0.628). But for teeth instrumented with RaCe files and K3 files

  5. THE ACS FORNAX CLUSTER SURVEY. X. COLOR GRADIENTS OF GLOBULAR CLUSTER SYSTEMS IN EARLY-TYPE GALAXIES

    International Nuclear Information System (INIS)

    Liu Chengze; Peng, Eric W.; Jordan, Andres; Ferrarese, Laura; Blakeslee, John P.; Cote, Patrick; Mei, Simona

    2011-01-01

    We use the largest homogeneous sample of globular clusters (GCs), drawn from the ACS Virgo Cluster Survey (ACSVCS) and ACS Fornax Cluster Survey (ACSFCS), to investigate the color gradients of GC systems in 76 early-type galaxies. We find that most GC systems possess an obvious negative gradient in (g-z) color with radius (bluer outward), which is consistent with previous work. For GC systems displaying color bimodality, both metal-rich and metal-poor GC subpopulations present shallower but significant color gradients on average, and the mean color gradients of these two subpopulations are of roughly equal strength. The field of view of ACS mainly restricts us to measuring the inner gradients of the studied GC systems. These gradients, however, can introduce an aperture bias when measuring the mean colors of GC subpopulations from relatively narrow central pointings. Inferred corrections to previous work imply a reduced significance for the relation between the mean color of metal-poor GCs and their host galaxy luminosity. The GC color gradients also show a dependence with host galaxy mass where the gradients are weakest at the ends of the mass spectrum-in massive galaxies and dwarf galaxies-and strongest in galaxies of intermediate mass, around a stellar mass of M * ∼10 10 M sun . We also measure color gradients for field stars in the host galaxies. We find that GC color gradients are systematically steeper than field star color gradients, but the shape of the gradient-mass relation is the same for both. If gradients are caused by rapid dissipational collapse and weakened by merging, these color gradients support a picture where the inner GC systems of most intermediate-mass and massive galaxies formed early and rapidly with the most massive galaxies having experienced greater merging. The lack of strong gradients in the GC systems of dwarfs, which probably have not experienced many recent major mergers, suggests that low-mass halos were inefficient at retaining

  6. Securing recommender systems against shilling attacks using social-based clustering

    KAUST Repository

    Zhang, Xiangliang

    2013-07-01

    Recommender systems (RS) have been found supportive and practical in e-commerce and been established as useful aiding services. Despite their great adoption in the user communities, RS are still vulnerable to unscrupulous producers who try to promote their products by shilling the systems. With the advent of social networks new sources of information have been made available which can potentially render RS more resistant to attacks. In this paper we explore the information provided in the form of social links with clustering for diminishing the impact of attacks. We propose two algorithms, CluTr and WCluTr, to combine clustering with "trust" among users. We demonstrate that CluTr and WCluTr enhance the robustness of RS by experimentally evaluating them on data from a public consumer recommender system Epinions.com. © 2013 Springer Science+Business Media New York & Science Press, China.

  7. Comparison Of Keyword Based Clustering Of Web Documents By Using Openstack 4j And By Traditional Method

    Directory of Open Access Journals (Sweden)

    Shiza Anand

    2015-08-01

    Full Text Available As the number of hypertext documents are increasing continuously day by day on world wide web. Therefore clustering methods will be required to bind documents into the clusters repositories according to the similarity lying between the documents. Various clustering methods exist such as Hierarchical Based K-means Fuzzy Logic Based Centroid Based etc. These keyword based clustering methods takes much more amount of time for creating containers and putting documents in their respective containers. These traditional methods use File Handling techniques of different programming languages for creating repositories and transferring web documents into these containers. In contrast openstack4j SDK is a new technique for creating containers and shifting web documents into these containers according to the similarity in much more less amount of time as compared to the traditional methods. Another benefit of this technique is that this SDK understands and reads all types of files such as jpg html pdf doc etc. This paper compares the time required for clustering of documents by using openstack4j and by traditional methods and suggests various search engines to adopt this technique for clustering so that they give result to the user querries in less amount of time.

  8. User and Document Group Approach of Clustering in Tagging Systems

    DEFF Research Database (Denmark)

    Pan, Rong; Xu, Guandong; Dolog, Peter

    2010-01-01

    In this paper, we propose a spectral clustering approach for users and documents group modeling in order to capture the common preference and relatedness of users and documents, and to reduce the time complexity of similarity calculations. In experiments, we investigate the selection of the optim...... amount of clusters. We also show a reduction of the time consuming in calculating the similarity for the recommender systems by selecting a centroid first, and then compare the inside item on behalf of each group....

  9. Cluster analysis of autoantibodies in 852 patients with systemic lupus erythematosus from a single center.

    Science.gov (United States)

    Artim-Esen, Bahar; Çene, Erhan; Şahinkaya, Yasemin; Ertan, Semra; Pehlivan, Özlem; Kamali, Sevil; Gül, Ahmet; Öcal, Lale; Aral, Orhan; Inanç, Murat

    2014-07-01

    Associations between autoantibodies and clinical features have been described in systemic lupus erythematosus (SLE). Herein, we aimed to define autoantibody clusters and their clinical correlations in a large cohort of patients with SLE. We analyzed 852 patients with SLE who attended our clinic. Seven autoantibodies were selected for cluster analysis: anti-DNA, anti-Sm, anti-RNP, anticardiolipin (aCL) immunoglobulin (Ig)G or IgM, lupus anticoagulant (LAC), anti-Ro, and anti-La. Two-step clustering and Kaplan-Meier survival analyses were used. Five clusters were identified. A cluster consisted of patients with only anti-dsDNA antibodies, a cluster of anti-Sm and anti-RNP, a cluster of aCL IgG/M and LAC, and a cluster of anti-Ro and anti-La antibodies. Analysis revealed 1 more cluster that consisted of patients who did not belong to any of the clusters formed by antibodies chosen for cluster analysis. Sm/RNP cluster had significantly higher incidence of pulmonary hypertension and Raynaud phenomenon. DsDNA cluster had the highest incidence of renal involvement. In the aCL/LAC cluster, there were significantly more patients with neuropsychiatric involvement, antiphospholipid syndrome, autoimmune hemolytic anemia, and thrombocytopenia. According to the Systemic Lupus International Collaborating Clinics damage index, the highest frequency of damage was in the aCL/LAC cluster. Comparison of 10 and 20 years survival showed reduced survival in the aCL/LAC cluster. This study supports the existence of autoantibody clusters with distinct clinical features in SLE and shows that forming clinical subsets according to autoantibody clusters may be useful in predicting the outcome of the disease. Autoantibody clusters in SLE may exhibit differences according to the clinical setting or population.

  10. 47 CFR 1.1704 - Station files.

    Science.gov (United States)

    2010-10-01

    ... System (COALS) § 1.1704 Station files. Applications, notifications, correspondence, electronic filings... Television Relay Service (CARS) are maintained by the Commission in COALS and the Public Reference Room...

  11. Performance comparison analysis library communication cluster system using merge sort

    Science.gov (United States)

    Wulandari, D. A. R.; Ramadhan, M. E.

    2018-04-01

    Begins by using a single processor, to increase the speed of computing time, the use of multi-processor was introduced. The second paradigm is known as parallel computing, example cluster. The cluster must have the communication potocol for processing, one of it is message passing Interface (MPI). MPI have many library, both of them OPENMPI and MPICH2. Performance of the cluster machine depend on suitable between performance characters of library communication and characters of the problem so this study aims to analyze the comparative performances libraries in handling parallel computing process. The case study in this research are MPICH2 and OpenMPI. This case research execute sorting’s problem to know the performance of cluster system. The sorting problem use mergesort method. The research method is by implementing OpenMPI and MPICH2 on a Linux-based cluster by using five computer virtual then analyze the performance of the system by different scenario tests and three parameters for to know the performance of MPICH2 and OpenMPI. These performances are execution time, speedup and efficiency. The results of this study showed that the addition of each data size makes OpenMPI and MPICH2 have an average speed-up and efficiency tend to increase but at a large data size decreases. increased data size doesn’t necessarily increased speed up and efficiency but only execution time example in 100000 data size. OpenMPI has a execution time greater than MPICH2 example in 1000 data size average execution time with MPICH2 is 0,009721 and OpenMPI is 0,003895 OpenMPI can customize communication needs.

  12. Scientific Cluster Deployment and Recovery - Using puppet to simplify cluster management

    Science.gov (United States)

    Hendrix, Val; Benjamin, Doug; Yao, Yushu

    2012-12-01

    Deployment, maintenance and recovery of a scientific cluster, which has complex, specialized services, can be a time consuming task requiring the assistance of Linux system administrators, network engineers as well as domain experts. Universities and small institutions that have a part-time FTE with limited time for and knowledge of the administration of such clusters can be strained by such maintenance tasks. This current work is the result of an effort to maintain a data analysis cluster (DAC) with minimal effort by a local system administrator. The realized benefit is the scientist, who is the local system administrator, is able to focus on the data analysis instead of the intricacies of managing a cluster. Our work provides a cluster deployment and recovery process (CDRP) based on the puppet configuration engine allowing a part-time FTE to easily deploy and recover entire clusters with minimal effort. Puppet is a configuration management system (CMS) used widely in computing centers for the automatic management of resources. Domain experts use Puppet's declarative language to define reusable modules for service configuration and deployment. Our CDRP has three actors: domain experts, a cluster designer and a cluster manager. The domain experts first write the puppet modules for the cluster services. A cluster designer would then define a cluster. This includes the creation of cluster roles, mapping the services to those roles and determining the relationships between the services. Finally, a cluster manager would acquire the resources (machines, networking), enter the cluster input parameters (hostnames, IP addresses) and automatically generate deployment scripts used by puppet to configure it to act as a designated role. In the event of a machine failure, the originally generated deployment scripts along with puppet can be used to easily reconfigure a new machine. The cluster definition produced in our CDRP is an integral part of automating cluster deployment

  13. Ground-Based Global Navigation Satellite System GLONASS (GLObal NAvigation Satellite System) Combined Broadcast Ephemeris Data (daily files) from NASA CDDIS

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset consists of ground-based Global Navigation Satellite System (GNSS) GLONASS Combined Broadcast Ephemeris Data (daily files of all distinct navigation...

  14. 75 FR 68340 - Combined Notice of Filings # 1

    Science.gov (United States)

    2010-11-05

    ... Project, LP; Southern Company--Florida LLC. Description: Report of Non-Material Change in Estimated Coal... Energy Ohio, Inc. submits tariff filing per 35: Compliance Filing to be effective 7/29/2010. Filed Date..., Inc. Description: Midwest Independent Transmission System Operator, Inc. submits tariff filing per 35...

  15. A Clustering Methodology of Web Log Data for Learning Management Systems

    Science.gov (United States)

    Valsamidis, Stavros; Kontogiannis, Sotirios; Kazanidis, Ioannis; Theodosiou, Theodosios; Karakos, Alexandros

    2012-01-01

    Learning Management Systems (LMS) collect large amounts of data. Data mining techniques can be applied to analyse their web data log files. The instructors may use this data for assessing and measuring their courses. In this respect, we have proposed a methodology for analysing LMS courses and students' activity. This methodology uses a Markov…

  16. Hydrogen spillover on DV (555-777) graphene – vanadium cluster system: First principles study

    Energy Technology Data Exchange (ETDEWEB)

    Kumar, E. Mathan, E-mail: ranjit.t@res.srmuniv.ac.in, E-mail: mathanranjitha@gmail.com; Thapa, Ranjit, E-mail: ranjit.t@res.srmuniv.ac.in, E-mail: mathanranjitha@gmail.com [SRM Research Institute, SRM University, Kattankulathur, Tamil Nadu - 603203 (India); P, Sabarikirishwaran [Department of Physics and Nanotechnology, SRM University, Kattankulathur, Tamil Nadu - 603203 (India)

    2015-06-24

    Using dispersion corrected density functional theory (DFT+D), the interaction of Vanadium adatom and cluster with divacancy (555-777) defective graphene sheet has been studied elaborately. We explore the prospect of hydrogen storage on V{sub 4} cluster adsorbed divacancy graphene system. It has been observed that V{sub 4} cluster (acting as a catalyst) can dissociate the H{sub 2} molecule into H atoms with very low barrier energy. We introduce the spillover of the atomic hydrogen throughout the surface via external mediator gallane (GaH{sub 3}) to form a hydrogenated system.

  17. Evaluation of apically extruded debris during root canal preparation in primary molar teeth using three different rotary systems and hand files.

    Science.gov (United States)

    Topçuoğlu, Gamze; Topçuoğlu, Hüseyin Sinan; Akpek, Firdevs

    2016-09-01

    To assess the amount of debris extruded apically during root canal preparation using various nickel titanium instrumentation systems and hand files in primary molar teeth. Sixty extracted primary first mandibular molar human teeth were randomly assigned to four groups (n = 15 teeth for each group). The canals were then instrumented with the following instrument systems: Revo-S, Mtwo, ProTaper Next, and hand files. Apically extruded debris during instrumentation was collected into pre-weighed Eppendorf tubes. The Eppendorf tubes were then stored in an incubator at 70°C for 5 days. The weight of the dry extruded debris was established by subtracting the pre-instrumentation and post-instrumentation weight of the Eppendorf tubes for each group. The data were analyzed using one-way analysis of variance (anova) and Tukey's post hoc tests. ProTaper Next files were associated with less apically extruded debris than the Mtwo, Revo-S, and hand files (P Hand files extruded more debris than Mtwo and Revo-S instruments (P 0.05). All instruments were associated with apical extrusion of debris. ProTaper Next files caused less debris extrusion compared to the other systems used. © 2015 BSPD, IAPD and John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  18. Comparative Analysis of Canal Centering Ability of Different Single File Systems Using Cone Beam Computed Tomography- An In-Vitro Study.

    Science.gov (United States)

    Agarwal, Rolly S; Agarwal, Jatin; Jain, Pradeep; Chandra, Anil

    2015-05-01

    The ability of an endodontic instrument to remain centered in the root canal system is one of the most important characteristic influencing the clinical performance of a particular file system. Thus, it is important to assess the canal centering ability of newly introduced single file systems before they can be considered a viable replacement of full-sequence rotary file systems. The aim of the study was to compare the canal transportation, centering ability, and time taken for preparation of curved root canals after instrumentation with single file systems One Shape and Wave One, using cone-beam computed tomography (CBCT). Sixty mesiobuccal canals of mandibular molars with an angle of curvature ranging from 20(o) to 35(o) were divided into three groups of 20 samples each: ProTaper PT (group I) - full-sequence rotary control group, OneShape OS (group II)- single file continuous rotation, WaveOne WO - single file reciprocal motion (group III). Pre instrumentation and post instrumentation three-dimensional CBCT images were obtained from root cross-sections at 3mm, 6mm and 9mm from the apex. Scanned images were then accessed to determine canal transportation and centering ability. The data collected were evaluated using one-way analysis of variance (ANOVA) with Tukey's honestly significant difference test. It was observed that there were no differences in the magnitude of transportation between the rotary instruments (p >0.05) at both 3mm as well as 6mm from the apex. At 9 mm from the apex, Group I PT showed significantly higher mean canal transportation and lower centering ability (0.19±0.08 and 0.39±0.16), as compared to Group II OS (0.12±0.07 and 0.54±0.24) and Group III WO (0.13±0.06 and 0.55±0.18) while the differences between OS and WO were not statistically significant. It was concluded that there was minor difference between the tested groups. Single file systems demonstrated average canal transportation and centering ability comparable to full sequence

  19. Comparative Analysis of Canal Centering Ability of Different Single File Systems Using Cone Beam Computed Tomography- An In-Vitro Study

    Science.gov (United States)

    Agarwal, Jatin; Jain, Pradeep; Chandra, Anil

    2015-01-01

    Background The ability of an endodontic instrument to remain centered in the root canal system is one of the most important characteristic influencing the clinical performance of a particular file system. Thus, it is important to assess the canal centering ability of newly introduced single file systems before they can be considered a viable replacement of full-sequence rotary file systems. Aim The aim of the study was to compare the canal transportation, centering ability, and time taken for preparation of curved root canals after instrumentation with single file systems One Shape and Wave One, using cone-beam computed tomography (CBCT). Materials and Methods Sixty mesiobuccal canals of mandibular molars with an angle of curvature ranging from 20o to 35o were divided into three groups of 20 samples each: ProTaper PT (group I) – full-sequence rotary control group, OneShape OS (group II)- single file continuous rotation, WaveOne WO – single file reciprocal motion (group III). Pre instrumentation and post instrumentation three-dimensional CBCT images were obtained from root cross-sections at 3mm, 6mm and 9mm from the apex. Scanned images were then accessed to determine canal transportation and centering ability. The data collected were evaluated using one-way analysis of variance (ANOVA) with Tukey’s honestly significant difference test. Results It was observed that there were no differences in the magnitude of transportation between the rotary instruments (p >0.05) at both 3mm as well as 6mm from the apex. At 9 mm from the apex, Group I PT showed significantly higher mean canal transportation and lower centering ability (0.19±0.08 and 0.39±0.16), as compared to Group II OS (0.12±0.07 and 0.54±0.24) and Group III WO (0.13±0.06 and 0.55±0.18) while the differences between OS and WO were not statistically significant Conclusion It was concluded that there was minor difference between the tested groups. Single file systems demonstrated average canal

  20. Average correlation clustering algorithm (ACCA) for grouping of co-regulated genes with similar pattern of variation in their expression values.

    Science.gov (United States)

    Bhattacharya, Anindya; De, Rajat K

    2010-08-01

    Distance based clustering algorithms can group genes that show similar expression values under multiple experimental conditions. They are unable to identify a group of genes that have similar pattern of variation in their expression values. Previously we developed an algorithm called divisive correlation clustering algorithm (DCCA) to tackle this situation, which is based on the concept of correlation clustering. But this algorithm may also fail for certain cases. In order to overcome these situations, we propose a new clustering algorithm, called average correlation clustering algorithm (ACCA), which is able to produce better clustering solution than that produced by some others. ACCA is able to find groups of genes having more common transcription factors and similar pattern of variation in their expression values. Moreover, ACCA is more efficient than DCCA with respect to the time of execution. Like DCCA, we use the concept of correlation clustering concept introduced by Bansal et al. ACCA uses the correlation matrix in such a way that all genes in a cluster have the highest average correlation values with the genes in that cluster. We have applied ACCA and some well-known conventional methods including DCCA to two artificial and nine gene expression datasets, and compared the performance of the algorithms. The clustering results of ACCA are found to be more significantly relevant to the biological annotations than those of the other methods. Analysis of the results show the superiority of ACCA over some others in determining a group of genes having more common transcription factors and with similar pattern of variation in their expression profiles. Availability of the software: The software has been developed using C and Visual Basic languages, and can be executed on the Microsoft Windows platforms. The software may be downloaded as a zip file from http://www.isical.ac.in/~rajat. Then it needs to be installed. Two word files (included in the zip file) need to

  1. Cluster fusion algorithm: application to Lennard-Jones clusters

    DEFF Research Database (Denmark)

    Solov'yov, Ilia; Solov'yov, Andrey V.; Greiner, Walter

    2006-01-01

    paths up to the cluster size of 150 atoms. We demonstrate that in this way all known global minima structures of the Lennard-Jones clusters can be found. Our method provides an efficient tool for the calculation and analysis of atomic cluster structure. With its use we justify the magic number sequence......We present a new general theoretical framework for modelling the cluster structure and apply it to description of the Lennard-Jones clusters. Starting from the initial tetrahedral cluster configuration, adding new atoms to the system and absorbing its energy at each step, we find cluster growing...... for the clusters of noble gas atoms and compare it with experimental observations. We report the striking correspondence of the peaks in the dependence of the second derivative of the binding energy per atom on cluster size calculated for the chain of the Lennard-Jones clusters based on the icosahedral symmetry...

  2. Cluster fusion algorithm: application to Lennard-Jones clusters

    DEFF Research Database (Denmark)

    Solov'yov, Ilia; Solov'yov, Andrey V.; Greiner, Walter

    2008-01-01

    paths up to the cluster size of 150 atoms. We demonstrate that in this way all known global minima structures of the Lennard-Jones clusters can be found. Our method provides an efficient tool for the calculation and analysis of atomic cluster structure. With its use we justify the magic number sequence......We present a new general theoretical framework for modelling the cluster structure and apply it to description of the Lennard-Jones clusters. Starting from the initial tetrahedral cluster configuration, adding new atoms to the system and absorbing its energy at each step, we find cluster growing...... for the clusters of noble gas atoms and compare it with experimental observations. We report the striking correspondence of the peaks in the dependence of the second derivative of the binding energy per atom on cluster size calculated for the chain of the Lennard-Jones clusters based on the icosahedral symmetry...

  3. VizieR Online Data Catalog: Clusters of galaxies in SDSS-III (Wen+, 2012)

    Science.gov (United States)

    Wen, Z. L.; Han, J. L.; Liu, F. S.

    2012-06-01

    Wen et al. (2009, Cat. J/ApJS/183/197) identified 39668 galaxy clusters from the SDSS DR6 by the discrimination of member galaxies of clusters using photometric redshifts of galaxies. Wen & Han (2011ApJ...734...68W) improved the method and successfully identified the high-redshift clusters from the deep fields of the Canada-France-Hawaii Telescope (CFHT) Wide survey, the CHFT Deep survey, the Cosmic Evolution Survey, and the Spitzer Wide-area InfraRed Extragalactic survey. Here, we follow and improve the algorithm to identify clusters from SDSS-III (SDSS Data Release 8; Aihara et al. 2011ApJS..193...29A, see Cat. II/306). (1 data file).

  4. Effect of coronal flaring on apical extrusion of debris during root canal instrumentation using single-file systems.

    Science.gov (United States)

    Topçuoğlu, H S; Üstün, Y; Akpek, F; Aktı, A; Topçuoğlu, G

    2016-09-01

    To evaluate the effect of coronal flaring on the amount of debris extruded apically during root canal preparation using the Reciproc, WaveOne (WO) and OneShape (OS) single-file systems. Ninety extracted single-rooted mandibular incisor teeth were randomly assigned to six groups (n = 15 for each group) for canal instrumentation. Endodontic access cavities were prepared in each tooth. In three of the six groups, coronal flaring was not performed; coronal flaring was performed with Gates-Glidden drills on all teeth in the remaining three groups. The canals were then instrumented with one or other of the following single-file instrument systems: Reciproc, WO and OS. Debris extruded apically during instrumentation was collected into pre-weighed Eppendorf tubes. The tubes were then stored in an incubator at 70 °C for 5 days. The weight of the dry extruded debris was established by subtracting the pre-instrumentation and post-instrumentation weight of the Eppendorf tubes for each group. Data were analysed using one-way analysis of variance (anova) and Tukey's post hoc tests (P = 0.05). Reciproc and WO files without coronal flaring produced significantly more debris compared with the other groups (P  0.05). All single-file systems caused apical extrusion of debris. Performing coronal flaring prior to canal preparation reduced the amount of apically extruded debris when using Reciproc or WO systems. © 2015 International Endodontic Journal. Published by John Wiley & Sons Ltd.

  5. Xbox one file system data storage: A forensic analysis

    OpenAIRE

    Gravel, Caitlin Elizabeth

    2015-01-01

    The purpose of this research was to answer the question, how does the file system of the Xbox One store data on its hard disk? This question is the main focus of the exploratory research and results sought. The research is focused on digital forensic investigators and experts. An out of the box Xbox One gaming console was used in the research. Three test cases were created as viable scenarios an investigator could come across in a search and seizure of evidence. The three test cases were then...

  6. 77 FR 60418 - Combined Notice of Filings

    Science.gov (United States)

    2012-10-03

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings Take notice that the Commission has received the following Natural Gas Pipeline Rate and Refund Report filings...-1064-000. Applicants: Venice Gathering System, L.L.C. Description: NAESB 2.0 Compliance Filing to be...

  7. 77 FR 9226 - Physical Systems Integration, LLC; Supplemental Notice That Initial Market-Based Rate Filing...

    Science.gov (United States)

    2012-02-16

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission [Docket No. ER12-1013-000] Physical Systems Integration, LLC; Supplemental Notice That Initial Market-Based Rate Filing Includes Request for... Physical Systems Integration, LLC's application for market-based rate authority, with an accompanying rate...

  8. 75 FR 3939 - Merit Systems Protection Board (MSPB) Provides Notice of Opportunity To File Amicus Briefs

    Science.gov (United States)

    2010-01-25

    ... MERIT SYSTEMS PROTECTION BOARD Merit Systems Protection Board (MSPB) Provides Notice of Opportunity To File Amicus Briefs AGENCY: Merit Systems Protection Board. ACTION: Notice. SUMMARY: Pursuant to 5 U.S.C. 7521 and 5 CFR 1201.131, the Merit Systems Protection Board (MSPB) is providing notice of...

  9. VizieR Online Data Catalog: NGC 6802 dwarf cluster members and non-members (Tang+, 2017)

    Science.gov (United States)

    Tang, B.; Geisler, D.; Friel, E.; Villanova, S.; Smiljanic, R.; Casey, A. R.; Randich, S.; Magrini, L.; San, Roman I.; Munoz, C.; Cohen, R. E.; Mauro, F.; Bragaglia, A.; Donati, P.; Tautvaisiene, G.; Drazdauskas, A.; Zenoviene, R.; Snaith, O.; Sousa, S.; Adibekyan, V.; Costado, M. T.; Blanco-Cuaresma, S.; Jimenez-Esteban, F.; Carraro, G.; Zwitter, T.; Francois, P.; Jofre, P.; Sordo, R.; Gilmore, G.; Flaccomio, E.; Koposov, S.; Korn, A. J.; Lanzafame, A. C.; Pancino, E.; Bayo, A.; Damiani, F.; Franciosini, E.; Hourihane, A.; Lardo, C.; Lewis, J.; Monaco, L.; Morbidelli, L.; Prisinzano, L.; Sacco, G.; Worley, C. C.; Zaggia, S.

    2016-11-01

    The dwarf stars in NGC 6802 observed by GIRAFFE spectrograph are separated into four tables: 1. cluster members in the lower main sequence; 2. cluster members in the upper main sequence; 3. non-member dwarfs in the lower main sequence; 4. non-member dwarfs in the upper main sequence. The star coordinates, V band magnitude, V-I color, and radial velocity are given. (4 data files).

  10. Development of a utility system for nuclear reaction data file: WinNRDF

    International Nuclear Information System (INIS)

    Aoyama, Shigeyoshi; Ohbayasi, Yosihide; Masui, Hiroshi; Chiba, Masaki; Kato, Kiyoshi; Ohnishi, Akira

    2000-01-01

    A utility system, WinNRDF, is developed for charged particle nuclear reaction data of NRDF (Nuclear Reaction Data File) on the Windows interface. By using this system, we can easily search the experimental data of a charged particle nuclear reaction in NRDF than old retrieval systems on the mainframe and also see graphically the experimental data on GUI (Graphical User Interface). We adopted a mechanism of making a new index of keywords to put to practical use of the time dependent properties of the NRDF database. (author)

  11. VizieR Online Data Catalog: LX-T relation in galaxy clusters (Hilton+, 2012)

    Science.gov (United States)

    Hilton, M.; Romer, A. K.; Kay, S. T.; Mehrtens, N.; Lloyd-Davies, E. J.; Thomas, P. A.; Short, C. J.; Mayers, J. A.; Rooney, P. J.; Stott, J. P.; Collins, C. A.; Harrison, C. D.; Hoyle, B.; Liddle, A. R.; Mann, R. G.; Miller, C. J.; Sahlen, M.; Viana, P. T. P.; Davidson, M.; Hosmer, M.; Nichol, R. C.; Sabirli, K.; Stanford, S. A.; West, M. J.

    2013-05-01

    XCS-DR1 (http://www.xcs-home.org/datareleases) is constructed from 5776 XMM observations, publicly available before 2010 July. In this work, we use only the subsample of 211 XCS-DR1 clusters with spectroscopic redshifts. (1 data file).

  12. Sparse Reconstruction of the Merging A520 Cluster System

    Energy Technology Data Exchange (ETDEWEB)

    Peel, Austin [Département d’Astrophysique, IRFU, CEA, Université Paris-Saclay, F-91191 Gif-sur-Yvette (France); Lanusse, François [McWilliams Center for Cosmology, Department of Physics, Carnegie Mellon University, Pittsburgh, PA 15213 (United States); Starck, Jean-Luc, E-mail: austin.peel@cea.fr [Université Paris Diderot, AIM, Sorbonne Paris Cité, CEA, CNRS, F-91191 Gif-sur-Yvette (France)

    2017-09-20

    Merging galaxy clusters present a unique opportunity to study the properties of dark matter in an astrophysical context. These are rare and extreme cosmic events in which the bulk of the baryonic matter becomes displaced from the dark matter halos of the colliding subclusters. Since all mass bends light, weak gravitational lensing is a primary tool to study the total mass distribution in such systems. Combined with X-ray and optical analyses, mass maps of cluster mergers reconstructed from weak-lensing observations have been used to constrain the self-interaction cross-section of dark matter. The dynamically complex Abell 520 (A520) cluster is an exceptional case, even among merging systems: multi-wavelength observations have revealed a surprising high mass-to-light concentration of dark mass, the interpretation of which is difficult under the standard assumption of effectively collisionless dark matter. We revisit A520 using a new sparsity-based mass-mapping algorithm to independently assess the presence of the puzzling dark core. We obtain high-resolution mass reconstructions from two separate galaxy shape catalogs derived from Hubble Space Telescope observations of the system. Our mass maps agree well overall with the results of previous studies, but we find important differences. In particular, although we are able to identify the dark core at a certain level in both data sets, it is at much lower significance than has been reported before using the same data. As we cannot confirm the detection in our analysis, we do not consider A520 as posing a significant challenge to the collisionless dark matter scenario.

  13. The Feasibility of Using Cluster Analysis to Examine Log Data from Educational Video Games. CRESST Report 790

    Science.gov (United States)

    Kerr, Deirdre; Chung, Gregory K. W. K.; Iseli, Markus R.

    2011-01-01

    Analyzing log data from educational video games has proven to be a challenging endeavor. In this paper, we examine the feasibility of using cluster analysis to extract information from the log files that is interpretable in both the context of the game and the context of the subject area. If cluster analysis can be used to identify patterns of…

  14. 78 FR 14530 - Combined Notice of Filings #2

    Science.gov (United States)

    2013-03-06

    ...: California Independent System Operator Corporation submits tariff filing per 35.13(a)(2)(iii): 2013-02-27 Pay... per 35.17(b): 2013-02-28--OASIS Att J Errata to be effective 4/15/2013. Filed Date: 2/27/13. Accession... Due: 5 p.m. ET 3/20/13. The filings are accessible in the Commission's eLibrary system by clicking on...

  15. Design and Implementation of File Access and Control System Based on Dynamic Web

    Institute of Scientific and Technical Information of China (English)

    GAO Fuxiang; YAO Lan; BAO Shengfei; YU Ge

    2006-01-01

    A dynamic Web application, which can help the departments of enterprise to collaborate with each other conveniently, is proposed. Several popular design solutions are introduced at first. Then, dynamic Web system is chosen for developing the file access and control system. Finally, the paper gives the detailed process of the design and implementation of the system, which includes some key problems such as solutions of document management and system security. Additionally, the limitations of the system as well as the suggestions of further improvement are also explained.

  16. Improving clustering with metabolic pathway data.

    Science.gov (United States)

    Milone, Diego H; Stegmayer, Georgina; López, Mariana; Kamenetzky, Laura; Carrari, Fernando

    2014-04-10

    It is a common practice in bioinformatics to validate each group returned by a clustering algorithm through manual analysis, according to a-priori biological knowledge. This procedure helps finding functionally related patterns to propose hypotheses for their behavior and the biological processes involved. Therefore, this knowledge is used only as a second step, after data are just clustered according to their expression patterns. Thus, it could be very useful to be able to improve the clustering of biological data by incorporating prior knowledge into the cluster formation itself, in order to enhance the biological value of the clusters. A novel training algorithm for clustering is presented, which evaluates the biological internal connections of the data points while the clusters are being formed. Within this training algorithm, the calculation of distances among data points and neurons centroids includes a new term based on information from well-known metabolic pathways. The standard self-organizing map (SOM) training versus the biologically-inspired SOM (bSOM) training were tested with two real data sets of transcripts and metabolites from Solanum lycopersicum and Arabidopsis thaliana species. Classical data mining validation measures were used to evaluate the clustering solutions obtained by both algorithms. Moreover, a new measure that takes into account the biological connectivity of the clusters was applied. The results of bSOM show important improvements in the convergence and performance for the proposed clustering method in comparison to standard SOM training, in particular, from the application point of view. Analyses of the clusters obtained with bSOM indicate that including biological information during training can certainly increase the biological value of the clusters found with the proposed method. It is worth to highlight that this fact has effectively improved the results, which can simplify their further analysis.The algorithm is available as a

  17. Ground-Based Global Navigation Satellite System Combined Broadcast Ephemeris Data (daily files) from NASA CDDIS

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset consists of ground-based Global Navigation Satellite System (GNSS) Combined Broadcast Ephemeris Data (daily files of all distinct navigation messages...

  18. Histological evaluation of the cleaning effectiveness of two reciprocating single-file systems in severely curved root canals: Reciproc versus WaveOne.

    Science.gov (United States)

    Carvalho, Maira de Souza; Junior, Emílio Carlos Sponchiado; Bitencourt Garrido, Angela Delfina; Roberti Garcia, Lucas da Fonseca; Franco Marques, André Augusto

    2015-01-01

    The aim of this study was to evaluate the cleaning effectiveness achieved with two reciprocating single-file systems in severely curved root canals: Reciproc and WaveOne. Twenty-five mesial roots of mandibular molars were randomly separated into two groups, according to the instrumentation system used. The negative control group consisted of five specimens that were not instrumented. The mesial canals (buccal and lingual) in Reciproc Group were instrumented with file R25 and the WaveOne group with the Primary file. The samples were submitted to histological processing and analyzed under a digital microscope. The WaveOne group presented a larger amount of debris than the Reciproc Group, however, without statistically significant difference (P > 0.05). A larger amount of debris in the control group was observed, with statistically significant difference to Reciproc and WaveOne groups (P file instrumentation systems presented similar effectiveness for root canal cleaning.

  19. Vertical Load Induced with Twisted File Adaptive System during Canal Shaping.

    Science.gov (United States)

    Jamleh, Ahmed; Alfouzan, Khalid

    2016-12-01

    To evaluate the vertical load induced with the Twisted File Adaptive (TFA; SybronEndo, Orange, CA) system during canal shaping of extracted teeth by comparing it with the Twisted File (TF, SybronEndo), ProTaper Next (PTN; Dentsply Maillefer, Ballaigues, Switzerland), and ProTaper Universal (PTU, Dentsply Maillefer) systems. Fifty-two root canals were shaped using the TFA, TF, PTN, or PTU systems (n = 13 for each system). They were shaped gently according to the manufacturers' instructions. During canal shaping, vertical loads were recorded and shown in 2 directions, apically and coronally directed loads. The vertical peak loads of 3 instrumentation stages were used for comparison. The effects of rotary systems on the mean positive and negative peak loads were analyzed statistically using the Kruskal-Wallis and Mann-Whitney tests at a confidence level of 95%. The overall pattern of the instantaneous loads appeared to increase with the use of successive instruments within the system. During canal shaping in all groups, the apically and coronally directed peak loads ranged from 0.84-7.55 N and 2.16-2.79 N, respectively. There were significant differences in both peak loads among the tested systems at each instrumentation stage. TFA had the lowest apically directed peak loads. In terms of coronally directed peak loads, the TFA and TF had a significantly lower amount of loads developed with their instruments than PTN and PTU. The choice of instrument system had an influence on the loads developed during canal shaping. TFA instruments were associated favorably with the lowest values of peak loads followed by TF, PTN, and PTU. Copyright © 2016 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  20. Intrinsic Variability in Multiple Systems and Clusters: Open Questions

    Science.gov (United States)

    Lampens, P.

    2006-04-01

    It is most interesting and rewarding to probe the stellar structure of stars which belong to a system originating from the same parent cloud as this provides additional and more accurate constraints for the models. New results on pulsating components in multiple systems and clusters are beginning to emerge regularly. Based on concrete studies, I will present still unsolved problems and discuss some of the issues which may affect our understanding of the pulsation physics in such systems but also in general.

  1. Shaping ability of the conventional nickel-titanium and reciprocating nickel-titanium file systems: a comparative study using micro-computed tomography.

    Science.gov (United States)

    Hwang, Young-Hye; Bae, Kwang-Shik; Baek, Seung-Ho; Kum, Kee-Yeon; Lee, WooCheol; Shon, Won-Jun; Chang, Seok Woo

    2014-08-01

    This study used micro-computed tomographic imaging to compare the shaping ability of Mtwo (VDW, Munich, Germany), a conventional nickel-titanium file system, and Reciproc (VDW), a reciprocating file system morphologically similar to Mtwo. Root canal shaping was performed on the mesiobuccal and distobuccal canals of extracted maxillary molars. In the RR group (n = 15), Reciproc was used in a reciprocating motion (150° counterclockwise/30° clockwise, 300 rpm); in the MR group, Mtwo was used in a reciprocating motion (150° clockwise/30° counterclockwise, 300 rpm); and in the MC group, Mtwo was used in a continuous rotating motion (300 rpm). Micro-computed tomographic images taken before and after canal shaping were used to analyze canal volume change and the degree of transportation at the cervical, middle, and apical levels. The time required for canal shaping was recorded. Afterward, each file was analyzed using scanning electron microscopy. No statistically significant differences were found among the 3 groups in the time for canal shaping or canal volume change (P > .05). Transportation values of the RR and MR groups were not significantly different at any level. However, the transportation value of the MC group was significantly higher than both the RR and MR groups at the cervical and apical levels (P file deformation was observed for 1 file in group RR (1/15), 3 files in group MR (3/15), and 5 files in group MC (5/15). In terms of shaping ability, Mtwo used in a reciprocating motion was not significantly different from the Reciproc system. Copyright © 2014 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  2. Decay data file based on the ENSDF file

    Energy Technology Data Exchange (ETDEWEB)

    Katakura, J. [Japan Atomic Energy Research Inst., Tokai, Ibaraki (Japan). Tokai Research Establishment

    1997-03-01

    A decay data file with the JENDL (Japanese Evaluated Nuclear Data Library) format based on the ENSDF (Evaluated Nuclear Structure Data File) file was produced as a tentative one of special purpose files of JENDL. The problem using the ENSDF file as primary source data of the JENDL decay data file is presented. (author)

  3. A secure file manager for UNIX

    Energy Technology Data Exchange (ETDEWEB)

    DeVries, R.G.

    1990-12-31

    The development of a secure file management system for a UNIX-based computer facility with supercomputers and workstations is described. Specifically, UNIX in its usual form does not address: (1) Operation which would satisfy rigorous security requirements. (2) Online space management in an environment where total data demands would be many times the actual online capacity. (3) Making the file management system part of a computer network in which users of any computer in the local network could retrieve data generated on any other computer in the network. The characteristics of UNIX can be exploited to develop a portable, secure file manager which would operate on computer systems ranging from workstations to supercomputers. Implementation considerations making unusual use of UNIX features, rather than requiring extensive internal system changes, are described, and implementation using the Cray Research Inc. UNICOS operating system is outlined.

  4. Meant to make a difference, the clinical experience of minimally invasive endodontics with the self-adjusting file system in India.

    Science.gov (United States)

    Pawar, Ajinkya M; Pawar, Mansing G; Kokate, Sharad R

    2014-01-01

    The vital steps in any endodontic treatment are thorough mechanical shaping and chemical cleaning followed by obtaining a fluid tight impervious seal by an inert obturating material. For the past two decades, introduction and use of rotary nickel-titanium (Ni-Ti) files have changed our concepts of endodontic treatment from conventional to contemporary. They have reported good success rates, but still have many drawbacks. The Self-Adjusting File (SAF) introduces a new era in endodontics by performing the vital steps of shaping and cleaning simultaneously. The SAF is a hollow file in design that adapts itself three-dimensionally to the root canal and is a single file system, made up of Ni-Ti lattice. The case series presented in the paper report the clinical experience, while treating primary endodontic cases with the SAF system in India.

  5. Meant to make a difference, the clinical experience of minimally invasive endodontics with the self-adjusting file system in India

    Directory of Open Access Journals (Sweden)

    Ajinkya M Pawar

    2014-01-01

    Full Text Available The vital steps in any endodontic treatment are thorough mechanical shaping and chemical cleaning followed by obtaining a fluid tight impervious seal by an inert obturating material. For the past two decades, introduction and use of rotary nickel-titanium (Ni-Ti files have changed our concepts of endodontic treatment from conventional to contemporary. They have reported good success rates, but still have many drawbacks. The Self-Adjusting File (SAF introduces a new era in endodontics by performing the vital steps of shaping and cleaning simultaneously. The SAF is a hollow file in design that adapts itself three-dimensionally to the root canal and is a single file system, made up of Ni-Ti lattice. The case series presented in the paper report the clinical experience, while treating primary endodontic cases with the SAF system in India.

  6. Cyclic fatigue resistance tests of Nickel-Titanium rotary files using simulated canal and weight loading conditions

    Directory of Open Access Journals (Sweden)

    Ok-In Cho

    2013-02-01

    Full Text Available Objectives This study compared the cyclic fatigue resistance of nickel-titanium (NiTi files obtained in a conventional test using a simulated canal with a newly developed method that allows the application of constant fatigue load conditions. Materials and Methods ProFile and K3 files of #25/.06, #30/.06, and #40/.04 were selected. Two types of testing devices were built to test their fatigue performance. The first (conventional device prescribed curvature inside a simulated canal (C-test, the second new device exerted a constant load (L-test whilst allowing any resulting curvature. Ten new instruments of each size and brand were tested with each device. The files were rotated until fracture and the number of cycles to failure (NCF was determined. The NCF were subjected to one-way ANOVA and Duncan's post-hoc test for each method. Spearman's rank correlation coefficient was computed to examine any association between methods. Results Spearman's rank correlation coefficient (ρ = -0.905 showed a significant negative correlation between methods. Groups with significant difference after the L-test divided into 4 clusters, whilst the C-test gave just 2 clusters. From the L-test, considering the negative correlation of NCF, K3 gave a significantly lower fatigue resistance than ProFile as in the C-test. K3 #30/.06 showed a lower fatigue resistance than K3 #25/.06, which was not found by the C-test. Variation in fatigue test methodology resulted in different cyclic fatigue resistance rankings for various NiTi files. Conclusions The new methodology standardized the load during fatigue testing, allowing determination fatigue behavior under constant load conditions.

  7. Аdaptive clustering algorithm for recommender systems

    OpenAIRE

    Stekh, Yu.; Artsibasov, V.

    2012-01-01

    In this article adaptive clustering algorithm for recommender systems is developed. Розроблено адаптивний алгоритм кластеризації для рекомендаційних систем.

  8. Clustering of galaxies near damped Lyman-alpha systems with (z) = 2.6

    Science.gov (United States)

    Wolfe, A. M

    1993-01-01

    The galaxy two-point correlation function, xi, at (z) = 2.6 is determined by comparing the number of Ly-alpha-emitting galaxies in narrowband CCD fields selected for the presence of damped L-alpha absorption to their number in randomly selected control fields. Comparisons between the presented determination of (xi), a density-weighted volume average of xi, and model predictions for (xi) at large redshifts show that models in which the clustering pattern is fixed in proper coordinates are highly unlikely, while better agreement is obtained if the clustering pattern is fixed in comoving coordinates. Therefore, clustering of Ly-alpha-emitting galaxies around damped Ly-alpha systems at large redshifts is strong. It is concluded that the faint blue galaxies are drawn from a parent population different from normal galaxies, the presumed offspring of damped Ly-alpha systems.

  9. Vibrations Generated by Several Nickel-titanium Endodontic File Systems during Canal Shaping in an Ex Vivo Model.

    Science.gov (United States)

    Choi, Dong-Min; Kim, Jin-Woo; Park, Se-Hee; Cho, Kyung-Mo; Kwak, Sang Won; Kim, Hyeon-Cheol

    2017-07-01

    This study aimed to compare the vibration generated by several nickel-titanium (NiTi) file systems and transmitted to teeth under 2 different motions (continuous rotation motion and reciprocating motion). Sixty J-shaped resin blocks (Endo Training Bloc-J; Dentsply Maillefer, Ballaigues, Switzerland) were trimmed to a root-shaped form and divided into 2 groups according to the types of electric motors: WaveOne motor (WOM, Dentsply Maillefer) and X-Smart Plus motor (XSM, Dentsply Maillefer). Each group was further subdivided into 3 subgroups (n = 10 each) according to the designated file systems: ProTaper Next (PTN, Dentsply Maillefer), ProTaper Universal (PTU, Dentsply Maillefer), and WaveOne (WOP, Dentsply Maillefer) systems. Vibration was measured during the pecking motion using an accelerometer attached to a predetermined consistent position. The average vibration values were subjected to 2-way analysis of variance as well as the t test and Duncan test for post hoc comparison at the 95% confidence interval. Both motor types and instrument types produced significantly different ranges of average vibrations. Regardless of the instrument types, the WOM group generated greater vibration than the XSM group (P file system may generate greater vibration than the continuous rotation NiTi file systems. The motor type also has a significant effect to amplify the vibrations. Copyright © 2017 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  10. Special issue on cluster algebras in mathematical physics

    Science.gov (United States)

    Di Francesco, Philippe; Gekhtman, Michael; Kuniba, Atsuo; Yamazaki, Masahito

    2014-02-01

    This is a call for contributions to a special issue of Journal of Physics A: Mathematical and Theoretical dedicated to cluster algebras in mathematical physics. Over the ten years since their introduction by Fomin and Zelevinsky, the theory of cluster algebras has witnessed a spectacular growth, first and foremost due to the many links that have been discovered with a wide range of subjects in mathematics and, increasingly, theoretical and mathematical physics. The main motivation of this special issue is to gather together reviews, recent developments and open problems, mainly from a mathematical physics viewpoint, into a single comprehensive issue. We expect that such a special issue will become a valuable reference for the broad scientific community working in mathematical and theoretical physics. The issue will consist of invited review articles and contributed papers containing new results on the interplays of cluster algebras with mathematical physics. Editorial policy The Guest Editors for this issue are Philippe Di Francesco, Michael Gekhtman, Atsuo Kuniba and Masahito Yamazaki. The areas and topics for this issue include, but are not limited to: discrete integrable systems arising from cluster mutations cluster structure on Poisson varieties cluster algebras and soliton interactions cluster positivity conjecture Y-systems in the thermodynamic Bethe ansatz and Zamolodchikov's periodicity conjecture T-system of transfer matrices of integrable lattice models dilogarithm identities in conformal field theory wall crossing in 4d N = 2 supersymmetric gauge theories 4d N = 1 quiver gauge theories described by networks scattering amplitudes of 4d N = 4 theories 3d N = 2 gauge theories described by flat connections on 3-manifolds integrability of dimer/Ising models on graphs. All contributions will be refereed and processed according to the usual procedure of the journal. Guidelines for preparation of contributions The deadline for contributed papers is 31 March

  11. Use of WebDAV to Support a Virtual File System in a Coalition Environment

    National Research Council Canada - National Science Library

    Bradney, Jeremiah A

    2006-01-01

    .... By enabling the use of WebDAV in MYSEA, this thesis provides a means for fulfilling the above requirement for secure remote access by creating a virtual web-based file system accessible from the MYSEA MLS network...

  12. EPA FRS Facilities State Single File CSV Download

    Science.gov (United States)

    This page provides state comma separated value (CSV) files containing key information of all facilities and sites within the Facility Registry System (FRS). Each state zip file contains a single CSV file of key facility-level information.

  13. Scientific Cluster Deployment and Recovery – Using puppet to simplify cluster management

    International Nuclear Information System (INIS)

    Hendrix, Val; Yao Yushu; Benjamin, Doug

    2012-01-01

    Deployment, maintenance and recovery of a scientific cluster, which has complex, specialized services, can be a time consuming task requiring the assistance of Linux system administrators, network engineers as well as domain experts. Universities and small institutions that have a part-time FTE with limited time for and knowledge of the administration of such clusters can be strained by such maintenance tasks. This current work is the result of an effort to maintain a data analysis cluster (DAC) with minimal effort by a local system administrator. The realized benefit is the scientist, who is the local system administrator, is able to focus on the data analysis instead of the intricacies of managing a cluster. Our work provides a cluster deployment and recovery process (CDRP) based on the puppet configuration engine allowing a part-time FTE to easily deploy and recover entire clusters with minimal effort. Puppet is a configuration management system (CMS) used widely in computing centers for the automatic management of resources. Domain experts use Puppet's declarative language to define reusable modules for service configuration and deployment. Our CDRP has three actors: domain experts, a cluster designer and a cluster manager. The domain experts first write the puppet modules for the cluster services. A cluster designer would then define a cluster. This includes the creation of cluster roles, mapping the services to those roles and determining the relationships between the services. Finally, a cluster manager would acquire the resources (machines, networking), enter the cluster input parameters (hostnames, IP addresses) and automatically generate deployment scripts used by puppet to configure it to act as a designated role. In the event of a machine failure, the originally generated deployment scripts along with puppet can be used to easily reconfigure a new machine. The cluster definition produced in our CDRP is an integral part of automating cluster deployment

  14. Cleaning capacity of hybrid instrumentation technique using reamer with alternating cutting edges system files: Histological analysis

    Directory of Open Access Journals (Sweden)

    Emilio Carlos Sponchiado Junior

    2014-01-01

    Full Text Available Aim: The aim of the following study is to evaluate the cleaning capacity of a hybrid instrumentation technique using Reamer with Alternating Cutting Edges (RaCe system files in the apical third of mesial roots of mandibular molars. Materials and Methods: Twenty teeth were selected and separated into two groups (n = 20 according to instrumentation technique as follows: BioRaCe - chemomechanical preparation with K-type files #10 and #15; and files BioRaCe BR0, BR1, BR2, BR3, and BR4; HybTec - hybrid instrumentation technique with K-type files #10 and #15 in the working length, #20 at 2 mm, #25 at 3 mm, cervical preparation with Largo burs #1 and #2; apical preparation with K-type files #15, #20, and #25 and RaCe files #25.04 and #30.04. The root canals were irrigated with 1 ml of 2.5% sodium hypochlorite at each change of instrument. The specimens were histologically processed and photographed under light optical microscope. The images were inserted onto an integration grid to count the amount of debris present in the root canal. Results: BioRaCe presented the highest percentage of debris in the apical third, however, with no statistically significant difference for HybTec (P > 0.05. Conclusions: The hybrid technique presented similar cleaning capacity as the technique recommended by the manufacturer.

  15. Customized recommendations for production management clusters of North American automatic milking systems.

    Science.gov (United States)

    Tremblay, Marlène; Hess, Justin P; Christenson, Brock M; McIntyre, Kolby K; Smink, Ben; van der Kamp, Arjen J; de Jong, Lisanne G; Döpfer, Dörte

    2016-07-01

    Automatic milking systems (AMS) are implemented in a variety of situations and environments. Consequently, there is a need to characterize individual farming practices and regional challenges to streamline management advice and objectives for producers. Benchmarking is often used in the dairy industry to compare farms by computing percentile ranks of the production values of groups of farms. Grouping for conventional benchmarking is commonly limited to the use of a few factors such as farms' geographic region or breed of cattle. We hypothesized that herds' production data and management information could be clustered in a meaningful way using cluster analysis and that this clustering approach would yield better peer groups of farms than benchmarking methods based on criteria such as country, region, breed, or breed and region. By applying mixed latent-class model-based cluster analysis to 529 North American AMS dairy farms with respect to 18 significant risk factors, 6 clusters were identified. Each cluster (i.e., peer group) represented unique management styles, challenges, and production patterns. When compared with peer groups based on criteria similar to the conventional benchmarking standards, the 6 clusters better predicted milk produced (kilograms) per robot per day. Each cluster represented a unique management and production pattern that requires specialized advice. For example, cluster 1 farms were those that recently installed AMS robots, whereas cluster 3 farms (the most northern farms) fed high amounts of concentrates through the robot to compensate for low-energy feed in the bunk. In addition to general recommendations for farms within a cluster, individual farms can generate their own specific goals by comparing themselves to farms within their cluster. This is very comparable to benchmarking but adds the specific characteristics of the peer group, resulting in better farm management advice. The improvement that cluster analysis allows for is

  16. Influence of operator's experience on root canal shaping ability with a rotary nickel-titanium single-file reciprocating motion system.

    Science.gov (United States)

    Muñoz, Estefanía; Forner, Leopoldo; Llena, Carmen

    2014-04-01

    The aim of this study was to evaluate the influence of the operator's experience on the shaping of double-curvature simulated root canals with a nickel-titanium single-file reciprocating motion system. Sixty double-curvature root canals simulated in methacrylate blocks were prepared by 10 students without any experience in endodontics and by 10 professionals who had studied endodontics at the postgraduate level. The Reciproc-VDW system's R25 file was used in the root canal preparation. The blocks were photographed before and after the instrumentation, and the time of instrumentation was also evaluated. Changes in root canal dimensions were analyzed in 6 positions. Significant differences (P file reciprocating motion system Reciproc is not seen to be influenced by the operator's experience regarding the increase of the canal area. Previous training and the need to acquire experience are important in the use of this system, in spite of its apparent simplicity. Copyright © 2014 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  17. Evaluated nuclear-data file for niobium

    International Nuclear Information System (INIS)

    Smith, A.B.; Smith, D.L.; Howerton, R.J.

    1985-03-01

    A comprehensive evaluated nuclear-data file for elemental niobium is provided in the ENDF/B format. This file, extending over the energy range 10 -11 -20 MeV, is suitable for comprehensive neutronic calculations, particulary those dealing with fusion-energy systems. It also provides dosimetry information. Attention is given to the internal consistancy of the file, energy balance, and the quantitative specification of uncertainties. Comparisons are made with experimental data and previous evaluated files. The results of integral tests are described and remaining outstanding problem areas are cited. 107 refs

  18. In vitro comparison rate of dental root canal transportation using two single file systems on the simulated resin blocks

    Directory of Open Access Journals (Sweden)

    Mohammad Javad Etesami

    2016-07-01

    Full Text Available Background and Aims: Cleaning and shaping is one of the most important stages in endodontic treatment. Single-file systems save time and reduce the risk of transmission of pathogens. This in vitro study was aimed to compare the rate of canal transportation after the preparation of the stimulated resin root canal with two single-file systems, namely Waveone and Reciproc. Materials and Methods: Thirty stimulated resin root canal blocks with size 8/0. 02 K file were randomly divided into two study groups. The preparation in Group A and Group B was performed using Reciproc and Waveone files, respectively. Pre and post- preparation photographs were taken and the images were superimposed to evaluate the inner and outer wall’s curvature tendency at three points (apical, middle and coronal using AutoCad pragram. Data were analyzed using T-test. Results: Based on the results, the degree of transportation in the inner and outer walls of the canal was less at the level of 3 millimeters (P0.05. Conclusion: Waveone showed better performance in the middle third of canal and this system maybe recommended.

  19. 75 FR 7464 - Energy Efficient Building Systems Regional Innovation Cluster Initiative-Joint Federal Funding...

    Science.gov (United States)

    2010-02-19

    ... a regional innovation cluster focused on innovation in energy efficient building technologies and... technology challenges through approaches that span basic research to engineering development to... DEPARTMENT OF ENERGY Energy Efficient Building Systems Regional Innovation Cluster Initiative...

  20. 48 CFR 750.7110-5 - Contract files.

    Science.gov (United States)

    2010-10-01

    ... 48 Federal Acquisition Regulations System 5 2010-10-01 2010-10-01 false Contract files. 750.7110-5 Section 750.7110-5 Federal Acquisition Regulations System AGENCY FOR INTERNATIONAL DEVELOPMENT CONTRACT... Interests of the United States 750.7110-5 Contract files. The fully executed action memorandum indicating...

  1. Titanium-II: an evaluated nuclear data file

    International Nuclear Information System (INIS)

    Philis, C.; Howerton, R.; Smith, A.B.

    1977-06-01

    A comprehensive evaluated nuclear data file for elemental titanium is outlined including definition of the data base, the evaluation procedures and judgments, and the final evaluated results. The file describes all significant neutron-induced reactions with elemental titanium and the associated photon-production processes to incident neutron energies of 20.0 MeV. In addition, isotopic-reaction files, consistent with the elemental file, are separately defined for those processes which are important to applied considerations of material-damage and neutron-dosimetry. The file is formulated in the ENDF format. This report formally documents the evaluation and, together with the numerical file, is submitted for consideration as a part of the ENDF/B-V evaluated file system. 20 figures, 9 tables

  2. Novel Functions of MicroRNA-17-92 Cluster in the Endocrine System.

    Science.gov (United States)

    Wan, Shan; Chen, Xiang; He, Yuedong; Yu, Xijie

    2018-01-01

    MiR-17-92 cluster is coded by MIR17HG in chromosome 13, which is highly conserved in vertebrates. Published literatures have proved that miR-17-92 cluster critically regulates tumorigenesis and metastasis. Recent researches showed that the miR-17-92 cluster also plays novel functions in the endocrine system. To summarize recent findings on the physiological and pathological roles of miR-17-92 cluster in bone, lipid and glucose metabolisms. MiR-17-92 cluster plays significant regulatory roles in bone development and metabolism through regulating the differentiation and function of osteoblasts and osteoclasts. In addition, miR-17- 92 cluster is nearly involved in every aspect of lipid metabolism. Last but not the least, the miR-17-92 cluster is closely bound up with pancreatic beta cell function, development of type 1 diabetes and insulin resistance. However, whether miR-17-92 cluster is involved in the communication among bone, fat and glucose metabolisms remains unknown. Growing evidence indicates that miR-17-92 cluster plays significant roles in bone, lipid and glucose metabolisms through a variety of signaling pathways. Fully understanding its modulating mechanisms may necessarily facilitate to comprehend the clinical and molecule features of some metabolic disorders such as osteoporosis, arthrosclerosis and diabetes mellitus. It may provide new drug targets to prevent and cure these disorders. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.

  3. Cluster electric spectroscopy of colloid chemical oxyhydrate systems

    CERN Document Server

    Sucharev, Yu I

    2015-01-01

    This monograph deals with the shape of Liesegang operator and its respective phase diagrams of spontaneous surges and analyzed properties of cluster attractors. It describes the influence of pulsation noise or self-organization current of gel systems in a magnetic field on singularities of optic parameters of yttrium oxyhydrate, as well as on kinetic curves of changes in optic density of oxyhydrate systems, sorptive properties of d- and f-elements, and the structural organization of their colloids. This monograph is meant for postgraduate students, magisters, researchers, and those interested

  4. Comparative evaluation of debris extruded apically by using, Protaper retreatment file, K3 file and H-file with solvent in endodontic retreatment

    Directory of Open Access Journals (Sweden)

    Chetna Arora

    2012-01-01

    Full Text Available Aim: The aim of this study was to evaluate the apical extrusion of debris comparing 2 engine driven systems and hand instrumentation technique during root canal retreatment. Materials and Methods: Forty five human permanent mandibular premolars were prepared using the step-back technique, obturated with gutta-percha/zinc oxide eugenol sealer and cold lateral condensation technique. The teeth were divided into three groups: Group A: Protaper retreatment file, Group B: K3, file Group C: H-file with tetrachloroethylene. All the canals were irrigated with 20ml distilled water during instrumentation. Debris extruded along with the irrigating solution during retreatment procedure was carefully collected in preweighed Eppendorf tubes. The tubes were stored in an incubator for 5 days, placed in a desiccator and then re-weighed. Weight of dry debris was calculated by subtracting the weight of the tube before instrumentation and from the weight of the tube after instrumentation. Data was analyzed using Two Way ANOVA and Post Hoc test. Results : There was statistically significant difference in the apical extrusion of debris between hand instrumentation and protaper retreatment file and K3 file. The amount of extruded debris caused by protaper retreatment file and K3 file instrumentation technique was not statistically significant. All the three instrumentation techniques produced apically extruded debris and irrigant. Conclusion: The best way to minimize the extrusion of debris is by adapting crown down technique therefore the use of rotary technique (Protaper retreatment file, K3 file is recommended.

  5. Evaluation of Single File Systems Reciproc, Oneshape, and WaveOne using Cone Beam Computed Tomography –An In Vitro Study

    Science.gov (United States)

    Dhingra, Annil; Miglani, Anjali

    2015-01-01

    Background Successful endodontic therapy depends on many factor, one of the most important step in any root canal treatment is root canal preparation. In addition, respecting the original shape of the canal is of the same importance; otherwise, canal aberrations such as transportation will be created. Aim The purpose of this study is to compare and evaluate Reciprocating WaveOne ,Reciproc and Rotary Oneshape Single File Instrumentation System On Cervical Dentin Thickness, Cross Sectional Area and Canal Transportation on First Mandibular Molar Using Cone Beam Computed Tomography. Materials and Methods Sixty Mandibular First Molars extracted due to periodontal reason was collected from the Department of Oral and Maxillofacial. Teeth were prepared using one rotary and two reciprocating single file system. Teeth were divided into 3 groups 20 teeth in each group. Pre instrumentation and Post instrumentation scans was done and evaluated for three parameters Canal Transportation, Cervical Dentinal Thickness, Cross-sectional Area. Results were analysed statistically using ANOVA, Post-Hoc Tukey analysis. Results The change in cross-sectional area after filing showed significant difference at 0mm, 1mm, 2mm and 7mm (pfile system over a distance of 7 mm (starting from 0mm and then evaluation at 1mm, 2mm, 3mm, 5mm and 7mm), the results showed a significant difference among the file systems at various lengths (p= 0.014, 0.046, 0.004, 0.028, 0.005 & 0.029 respectively). Mean value of cervical dentinal removal is maximum at all the levels for oneshape and minimum for waveone showing the better quality of waveone and reciproc over oneshape file system. Significant difference was found at 9mm, 11mm and 12mm between all the three file systems (p<0.001,< 0.001, <0.001). Conclusion It was concluded that reciprocating motion is better than rotary motion in all the three parameters Canal Transportation, Cross-sectional Area, Cervical Dentinal Thickness. PMID:26023639

  6. Manual versus computerized filing system in nuclear medicine - a regression. A simple but effectiv system for patient calling in in-vivo diagnosis

    International Nuclear Information System (INIS)

    Spitz, J.

    1985-01-01

    A patient-filing and calling system is described which uses colour-coded data carriers and a special board. The system proved to be useful for stock display and handling of labelled activity ready for injection in patients. (orig.) [de

  7. The HectoMAP Cluster Survey. I. redMaPPer Clusters

    Science.gov (United States)

    Sohn, Jubee; Geller, Margaret J.; Rines, Kenneth J.; Hwang, Ho Seong; Utsumi, Yousuke; Diaferio, Antonaldo

    2018-04-01

    We use the dense HectoMAP redshift survey to explore the properties of 104 redMaPPer cluster candidates. The redMaPPer systems in HectoMAP cover the full range of richness and redshift (0.08 systems included in the Subaru/Hyper Suprime-Cam public data release are bona fide clusters. The median number of spectroscopic members per cluster is ∼20. We include redshifts of 3547 member candidates listed in the redMaPPer catalog whether they are cluster members or not. We evaluate the redMaPPer membership probability spectroscopically. The purity (number of real systems) in redMaPPer exceeds 90% even at the lowest richness. Three massive galaxy clusters (M ∼ 2 × 1013 M ⊙) associated with X-ray emission in the HectoMAP region are not included in the public redMaPPer catalog with λ rich > 20, because they lie outside the cuts for this catalog.

  8. Folksonomies and clustering in the collaborative system CiteULike

    Science.gov (United States)

    Capocci, Andrea; Caldarelli, Guido

    2008-06-01

    We analyze CiteULike, an online collaborative tagging system where users bookmark and annotate scientific papers. Such a system can be naturally represented as a tri-partite graph whose nodes represent papers, users and tags connected by individual tag assignments. The semantics of tags is studied here, in order to uncover the hidden relationships between tags. We find that the clustering coefficient can be used to analyze the semantical patterns among tags.

  9. Folksonomies and clustering in the collaborative system CiteULike

    International Nuclear Information System (INIS)

    Capocci, Andrea; Caldarelli, Guido

    2008-01-01

    We analyze CiteULike, an online collaborative tagging system where users bookmark and annotate scientific papers. Such a system can be naturally represented as a tri-partite graph whose nodes represent papers, users and tags connected by individual tag assignments. The semantics of tags is studied here, in order to uncover the hidden relationships between tags. We find that the clustering coefficient can be used to analyze the semantical patterns among tags

  10. Folksonomies and clustering in the collaborative system CiteULike

    Energy Technology Data Exchange (ETDEWEB)

    Capocci, Andrea [Dip. di Informatica e Sistemistica Universita ' Sapienza' , via Ariosto, 25 00185 Rome (Italy); Caldarelli, Guido [SMC Centre, CNR-INFM, Dip. di Fisica, Universita ' Sapienza' , P.le A. Moro 5, 00185-Rome (Italy)

    2008-06-06

    We analyze CiteULike, an online collaborative tagging system where users bookmark and annotate scientific papers. Such a system can be naturally represented as a tri-partite graph whose nodes represent papers, users and tags connected by individual tag assignments. The semantics of tags is studied here, in order to uncover the hidden relationships between tags. We find that the clustering coefficient can be used to analyze the semantical patterns among tags.

  11. Globular Clusters - Guides to Galaxies

    CERN Document Server

    Richtler, Tom; Joint ESO-FONDAP Workshop on Globular Clusters

    2009-01-01

    The principal question of whether and how globular clusters can contribute to a better understanding of galaxy formation and evolution is perhaps the main driving force behind the overall endeavour of studying globular cluster systems. Naturally, this splits up into many individual problems. The objective of the Joint ESO-FONDAP Workshop on Globular Clusters - Guides to Galaxies was to bring together researchers, both observational and theoretical, to present and discuss the most recent results. Topics covered in these proceedings are: internal dynamics of globular clusters and interaction with host galaxies (tidal tails, evolution of cluster masses), accretion of globular clusters, detailed descriptions of nearby cluster systems, ultracompact dwarfs, formations of massive clusters in mergers and elsewhere, the ACS Virgo survey, galaxy formation and globular clusters, dynamics and kinematics of globular cluster systems and dark matter-related problems. With its wide coverage of the topic, this book constitute...

  12. Cluster consensus in discrete-time networks of multiagents with inter-cluster nonidentical inputs.

    Science.gov (United States)

    Han, Yujuan; Lu, Wenlian; Chen, Tianping

    2013-04-01

    In this paper, cluster consensus of multiagent systems is studied via inter-cluster nonidentical inputs. Here, we consider general graph topologies, which might be time-varying. The cluster consensus is defined by two aspects: intracluster synchronization, the state at which differences between each pair of agents in the same cluster converge to zero, and inter-cluster separation, the state at which agents in different clusters are separated. For intra-cluster synchronization, the concepts and theories of consensus, including the spanning trees, scramblingness, infinite stochastic matrix product, and Hajnal inequality, are extended. As a result, it is proved that if the graph has cluster spanning trees and all vertices self-linked, then the static linear system can realize intra-cluster synchronization. For the time-varying coupling cases, it is proved that if there exists T > 0 such that the union graph across any T-length time interval has cluster spanning trees and all graphs has all vertices self-linked, then the time-varying linear system can also realize intra-cluster synchronization. Under the assumption of common inter-cluster influence, a sort of inter-cluster nonidentical inputs are utilized to realize inter-cluster separation, such that each agent in the same cluster receives the same inputs and agents in different clusters have different inputs. In addition, the boundedness of the infinite sum of the inputs can guarantee the boundedness of the trajectory. As an application, we employ a modified non-Bayesian social learning model to illustrate the effectiveness of our results.

  13. Influence of a glide path on apical extrusion of debris during canal preparation using single-file systems in curved canals.

    Science.gov (United States)

    Topçuoğlu, H S; Düzgün, S; Akpek, F; Topçuoğlu, G; Aktı, A

    2016-06-01

    To evaluate the effect of a glide path on the amount of apically extruded debris during canal preparation using single-file systems in curved canals. Ninety extracted mandibular molar teeth were randomly assigned to six groups (n = 15 for each group) for canal instrumentation. Endodontic access cavities were prepared in each tooth. In three of the six groups, a glide path was not created whereas a glide path was created using PathFile instruments on the mesial canals of all teeth in the remaining three groups. The mesial canals of the teeth were then instrumented with the following single-file instrument systems: WaveOne, Reciproc and OneShape. Debris extruded apically during instrumentation was collected into pre-weighed Eppendorf tubes. The tubes were then stored in an incubator at 70 °C for 5 days. The weight of the dry extruded debris was established by subtracting the pre-instrumentation and post-instrumentation weight of the Eppendorf tubes for each group. The data obtained were analysed using one-way analysis of variance (anova) and Tukey's post hoc tests. The OneShape file was associated with less debris extrusion than the Reciproc and WaveOne files when canal instrumentation was performed without a glide path (P files (P > 0.05). There was no significant difference amongst the OneShape, Reciproc and WaveOne files when a glide path was created before canal preparation in curved root canals (P > 0.05). All systems extruded significantly less debris in groups with a glide path than in groups without a glide path (P < 0.05). All instruments were associated with apical extrusion of debris. Creating a glide path prior to canal instrumentation reduced the amount of apically extruded debris in curved canals. © 2015 International Endodontic Journal. Published by John Wiley & Sons Ltd.

  14. Technological advances in endodontics: treatment of a mandibular molar with internal root resorption using a reciprocating single-file system.

    Science.gov (United States)

    de Souza, Samir Noronha; Marques, André Augusto Franco; Sponchiado-Júnior, EmÍlio Carlos; Roberti Garcia, Lucas da Fonseca; da Frota, Matheus Franco; de Carvalho, Fredson Márcio Acris

    2017-01-01

    The field of endodontics has become increasingly successful due to technological advances that allow clinicians to solve clinical cases that would have been problematic a few years ago. Despite such advances, endodontic treatment of teeth with internal root resorption remains challenging. This article presents a clinical case in which a reciprocating single-file system was used for endodontic treatment of a mandibular molar with internal root resorption. Radiographic examination revealed the presence of internal root resorption in the distobuccal root canal of the mandibular right first molar. A reciprocating single-file system was used for root canal instrumentation and final preparation, and filling was obtained through a thermal compaction technique. No painful symptoms or periapical lesions were observed in 12 months of follow-up. The results indicate that a reciprocating single-file system is an adequate alternative for root canal instrumentation, particularly in teeth with internal root resorption.

  15. Efficiency of the Self Adjusting File, WaveOne, Reciproc, ProTaper and hand files in root canal debridement.

    Science.gov (United States)

    Topcu, K Meltem; Karatas, Ertugrul; Ozsu, Damla; Ersoy, Ibrahim

    2014-07-01

    The aim of this study was to compare the canal debridement capabilities of three single file systems, ProTaper, and K-files in oval-shaped canals. Seventy-five extracted human mandibular central incisors with oval-shaped root canals were selected. A radiopaque contrast medium (Metapex; Meta Biomed Co. Ltd., Chungcheongbuk-do, Korea) was introduced into the canal systems and the self-adjusting file (SAF), WaveOne, Reciproc, ProTaper, and K-files were used for the instrumentation of the canals. The percentage of removed contrast medium was calculated using pre- and post-operative radiographs. An overall comparison between the groups revealed that the hand file (HF) and SAF groups presented the lowest percentage of removed contrast medium, whereas the WaveOne group showed the highest percentage (P ProTaper group removed more contrast medium than the SAF and HF groups (P < 0.05). None of the instruments was able to remove the contrast medium completely. WaveOne performed significantly better than other groups.

  16. 77 FR 71412 - Combined Notice of Filings #2

    Science.gov (United States)

    2012-11-30

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings 2 Take notice that the Commission has received the following Natural Gas Pipeline Rate and Refund Report filings: Filings Instituting Proceedings Docket Numbers: RP13-301-000. Applicants: Venice Gathering System, LLC...

  17. European BWR R and D cluster for innovative passive safety systems

    International Nuclear Information System (INIS)

    Hicken, E.F.; Lensa, W. von

    1996-01-01

    The main technological innovation trends for future nuclear power plants tend towards a broader use of passive safety systems for the prevention, mitigation and managing of severe accident scenarios. Several approaches have been undertaken in a number of European countries to study and demonstrate the feasibility and charateristics of innovative passive safety systems. The European BWR R and D Cluster combines those experimental and analytical efforts that are mainly directed to the introduction of passive safety systems into boiling water reactor technology. The Cluster is grouped around thermohydraulic test facilities in Europe for the qualification of innovative BWR safety systems, also taking into account especially the operating experience of the nuclear power plant Dodewaard and other BWRs, which already incorporated some passive safety features. The background, the objectives, the structure of the project and the work programme are presented in this paper as well as an outline of the significance of the expected results. (orig.) [de

  18. Parallel checksumming of data chunks of a shared data object using a log-structured file system

    Science.gov (United States)

    Bent, John M.; Faibish, Sorin; Grider, Gary

    2016-09-06

    Checksum values are generated and used to verify the data integrity. A client executing in a parallel computing system stores a data chunk to a shared data object on a storage node in the parallel computing system. The client determines a checksum value for the data chunk; and provides the checksum value with the data chunk to the storage node that stores the shared object. The data chunk can be stored on the storage node with the corresponding checksum value as part of the shared object. The storage node may be part of a Parallel Log-Structured File System (PLFS), and the client may comprise, for example, a Log-Structured File System client on a compute node or burst buffer. The checksum value can be evaluated when the data chunk is read from the storage node to verify the integrity of the data that is read.

  19. ClusterControl: a web interface for distributing and monitoring bioinformatics applications on a Linux cluster.

    Science.gov (United States)

    Stocker, Gernot; Rieder, Dietmar; Trajanoski, Zlatko

    2004-03-22

    ClusterControl is a web interface to simplify distributing and monitoring bioinformatics applications on Linux cluster systems. We have developed a modular concept that enables integration of command line oriented program into the application framework of ClusterControl. The systems facilitate integration of different applications accessed through one interface and executed on a distributed cluster system. The package is based on freely available technologies like Apache as web server, PHP as server-side scripting language and OpenPBS as queuing system and is available free of charge for academic and non-profit institutions. http://genome.tugraz.at/Software/ClusterControl

  20. A COMPUTER CLUSTER SYSTEM FOR PSEUDO-PARALLEL EXECUTION OF GEANT4 SERIAL APPLICATION

    Directory of Open Access Journals (Sweden)

    Memmo Federici

    2013-12-01

    Full Text Available Simulation of the interactions between particles and matter in studies for developing X-rays detectors generally requires very long calculation times (up to several days or weeks. These times are often a serious limitation for the success of the simulations and for the accuracy of the simulated models. One of the tools used by the scientific community to perform these simulations is Geant4 (Geometry And Tracking [2, 3]. On the best of experience in the design of the AVES cluster computing system, Federici et al. [1], the IAPS (Istituto di Astrofisica e Planetologia Spaziali INAF laboratories were able to develop a cluster computer system dedicated to Geant 4. The Cluster is easy to use and easily expandable, and thanks to the design criteria adopted it achieves an excellent compromise between performance and cost. The management software developed for the Cluster splits the single instance of simulation on the cores available, allowing the use of software written for serial computation to reach a computing speed similar to that obtainable from a native parallel software. The simulations carried out on the Cluster showed an increase in execution time by a factor of 20 to 60 compared to the times obtained with the use of a single PC of medium quality.

  1. VizieR Online Data Catalog: HST/ACS Coma cluster survey. II. (Hammer+, 2010)

    NARCIS (Netherlands)

    Hammer, D.; Verdoes Kleijn, G.; Hoyos, C.; den Brok, M.; Balcells, M.; Ferguson, H. C.; Goudfrooij, P.; Carter, D.; Guzman, R.; Peletier, R. F.; Smith, R. J.; Graham, A. W.; Trentham, N.; Peng, E.; Puzia, T. H.; Lucey, J. R.; Jogee, S.; Aguerri, A. L.; Batcheldor, D.; Bridges, T. J.; Chiboucas, K.; Davies, J. I.; Del Burgo, C.; Erwin, P.; Hornschemeier, A.; Hudson, M. J.; Huxor, A.; Jenkins, L.; Karick, A.; Khosroshahi, H.; Kourkchi, E.; Komiyama, Y.; Lotz, J.; Marzke, R. O.; Marinova, I.; Matkovic, A.; Merritt, D.; Miller, B. W.; Miller, N. A.; Mobasher, B.; Mouhcine, M.; Okamura, S.; Percival, S.; Phillipps, S.; Poggianti, B. M.; Price, J.; Sharples, R. M.; Tully, R. B.; Valentijn, E.

    2010-01-01

    This data release contains catalogs for the ACS Images in F475W and F814W bands of 25 fields in the Coma cluster of galaxies. Each field is about 202x202arcsec. Please see the release notes for further details. (25 data files).

  2. File access prediction using neural networks.

    Science.gov (United States)

    Patra, Prashanta Kumar; Sahu, Muktikanta; Mohapatra, Subasish; Samantray, Ronak Kumar

    2010-06-01

    One of the most vexing issues in design of a high-speed computer is the wide gap of access times between the memory and the disk. To solve this problem, static file access predictors have been used. In this paper, we propose dynamic file access predictors using neural networks to significantly improve upon the accuracy, success-per-reference, and effective-success-rate-per-reference by using neural-network-based file access predictor with proper tuning. In particular, we verified that the incorrect prediction has been reduced from 53.11% to 43.63% for the proposed neural network prediction method with a standard configuration than the recent popularity (RP) method. With manual tuning for each trace, we are able to improve upon the misprediction rate and effective-success-rate-per-reference using a standard configuration. Simulations on distributed file system (DFS) traces reveal that exact fit radial basis function (RBF) gives better prediction in high end system whereas multilayer perceptron (MLP) trained with Levenberg-Marquardt (LM) backpropagation outperforms in system having good computational capability. Probabilistic and competitive predictors are the most suitable for work stations having limited resources to deal with and the former predictor is more efficient than the latter for servers having maximum system calls. Finally, we conclude that MLP with LM backpropagation algorithm has better success rate of file prediction than those of simple perceptron, last successor, stable successor, and best k out of m predictors.

  3. Archive Inventory Management System (AIMS) — A Fast, Metrics Gathering Framework for Validating and Gaining Insight from Large File-Based Data Archives

    Science.gov (United States)

    Verma, R. V.

    2018-04-01

    The Archive Inventory Management System (AIMS) is a software package for understanding the distribution, characteristics, integrity, and nuances of files and directories in large file-based data archives on a continuous basis.

  4. Correlations in clusters and related systems. New perspectives on the many-body problem

    International Nuclear Information System (INIS)

    Connerade, J.P.

    1996-01-01

    The contents of the present volume are the proceedings of an Adriatico Research Conference, held at the International Centre for Theoretical Physics in Trieste from 26 to 29 July 1994. The theme of the conference covered many aspects of cooperative effects, beginning with giant resonances in many-electron systems, and particularly in new objects such as metallic clusters, in which collective electron dynamics are a novel feature. The relationship of these resonances with comparable features in nuclear and solid state physics was extensively discussed. Related effects, such as instabilities of valence both in clusters and in solids were explored. Clusters allow one to track the evolution of certain properties from the free atom to the solid state limits as a function of size. The giant resonances concerned not only intra-atomic excitations, but also correlated motions of all delocalized electrons within the cluster. Other systems with unusual properties, such as negative ions, in which correlations play an important role, were also considered. Finally, dynamical effects and the possible interactions between electron-electron correlations and high laser fields were envisaged

  5. Proposed Fuzzy-NN Algorithm with LoRaCommunication Protocol for Clustered Irrigation Systems

    Directory of Open Access Journals (Sweden)

    Sotirios Kontogiannis

    2017-11-01

    Full Text Available Modern irrigation systems utilize sensors and actuators, interconnected together as a single entity. In such entities, A.I. algorithms are implemented, which are responsible for the irrigation process. In this paper, the authors present an irrigation Open Watering System (OWS architecture that spatially clusters the irrigation process into autonomous irrigation sections. Authors’ OWS implementation includes a Neuro-Fuzzy decision algorithm called FITRA, which originates from the Greek word for seed. In this paper, the FITRA algorithm is described in detail, as are experimentation results that indicate significant water conservations from the use of the FITRA algorithm. Furthermore, the authors propose a new communication protocol over LoRa radio as an alternative low-energy and long-range OWS clusters communication mechanism. The experimental scenarios confirm that the FITRA algorithm provides more efficient irrigation on clustered areas than existing non-clustered, time scheduled or threshold adaptive algorithms. This is due to the FITRA algorithm’s frequent monitoring of environmental conditions, fuzzy and neural network adaptation as well as adherence to past irrigation preferences.

  6. Directed clustering coefficient as a measure of systemic risk in complex banking networks

    Science.gov (United States)

    Tabak, Benjamin M.; Takami, Marcelo; Rocha, Jadson M. C.; Cajueiro, Daniel O.; Souza, Sergio R. S.

    2014-01-01

    Recent literature has focused on the study of systemic risk in complex networks. It is clear now, after the crisis of 2008, that the aggregate behavior of the interaction among agents is not straightforward and it is very difficult to predict. Contributing to this debate, this paper shows that the directed clustering coefficient may be used as a measure of systemic risk in complex networks. Furthermore, using data from the Brazilian interbank network, we show that the directed clustering coefficient is negatively correlated with domestic interest rates.

  7. 76 FR 2368 - Balance Power Systems, LLC; Supplemental Notice That Initial Market-Based Rate Filing Includes...

    Science.gov (United States)

    2011-01-13

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission [Docket No. ER11-2331-000] Balance Power Systems, LLC; Supplemental Notice That Initial Market-Based Rate Filing Includes Request for Blanket... proceeding of Balance Power Systems, LLC's application for market-based rate authority, with an accompanying...

  8. Reduced-dimension power allocation over clustered channels in cognitive radios system under co-channel interference

    KAUST Repository

    Ben Ghorbel, Mahdi

    2014-05-12

    The objective of this paper is to propose a reduceddimension resource allocation scheme in the context of cognitive radio system in presence of co-channel interference between users. We assume a multicarrier transmission for both the primary and secondary systems. Instead of optimizing the powers over all sub-carriers, the sub-carriers are grouped into clusters of sub-carriers, where the power of each sub-carrier is directly related to the power of the correspondent cluster. The power optimization is done only over the set of clusters instead of all sub-carriers which can significantly reduce the complexity of the resource allocation problem. The performance loss of the reduced dimension solution with respect to the optimal solution, where the optimization is carried over all active sub-carriers, allows trading-off complexity versus performance. Numerical evaluation indeed revealed that a limited performance loss occurs by optimizing over a reduced set of clusters instead of the full optimization in the context of cognitive radio systems.

  9. GEODOC: the GRID document file, record structure and data element description

    Energy Technology Data Exchange (ETDEWEB)

    Trippe, T.; White, V.; Henderson, F.; Phillips, S.

    1975-11-06

    The purpose of this report is to describe the information structure of the GEODOC file. GEODOC is a computer based file which contains the descriptive cataloging and indexing information for all documents processed by the National Geothermal Information Resource Group. This file (along with other GRID files) is managed by DBMS, the Berkeley Data Base Management System. Input for the system is prepared using the IRATE Text Editing System with its extended (12 bit) character set, or punched cards.

  10. Incidence of microcracks in maxillary first premolars after instrumentation with three different mechanized file systems: a comparative ex vivo study.

    Science.gov (United States)

    Kfir, A; Elkes, D; Pawar, A; Weissman, A; Tsesis, I

    2017-01-01

    The objective of this study is to determine the potential for microcracks in the radicular dentin of first maxillary premolars using three different mechanized endodontic instrumentation systems. Eighty extracted maxillary first premolars with two root canals and no externally visible microcracks were selected. Root canal instrumentation was performed with either the ProTaper file system, the WaveOne primary file, or the self-adjusting file (SAF). Teeth with intact roots served as controls. The roots were cut into segments and examined with an intensive, small-diameter light source that was applied diagonally to the entire periphery of the root slice under ×20 magnification; the presence of microcracks and fractures was recorded. Pearson's chi-square method was used for statistical analysis, and significance was set at p systems, respectively, while no microcracks were present in the roots treated with the SAF (p = 0.008 and p = 0.035, respectively). Intact teeth presented with cracks in 5 % of the roots. The intensive, small-diameter light source revealed microcracks that could not be detected when using the microscope's light alone. Within the limitations of this study, it could be concluded that mechanized root canal instrumentation with the ProTaper and WaveOne systems in maxillary first premolars causes microcracks in the radicular dentin, while the use of the SAF file causes no such microcracks. Rotary and reciprocating files with large tapers may cause microcracks in the radicular dentin of maxillary first premolars. Less aggressive methods should be considered for these teeth.

  11. THE M33 GLOBULAR CLUSTER SYSTEM WITH PAndAS DATA: THE LAST OUTER HALO CLUSTER?

    International Nuclear Information System (INIS)

    Cockcroft, Robert; Harris, William E.; Ferguson, Annette M. N.

    2011-01-01

    We use CFHT/MegaCam data to search for outer halo star clusters in M33 as part of the Pan-Andromeda Archaeological Survey. This work extends previous studies out to a projected radius of 50 kpc and covers over 40 deg 2 . We find only one new unambiguous star cluster in addition to the five previously known in the M33 outer halo (10 kpc ≤ r ≤ 50 kpc). Although we identify 2440 cluster candidates of various degrees of confidence from our objective image search procedure, almost all of these are likely background contaminants, mostly faint unresolved galaxies. We measure the luminosity, color, and structural parameters of the new cluster in addition to the five previously known outer halo clusters. At a projected radius of 22 kpc, the new cluster is slightly smaller, fainter, and redder than all but one of the other outer halo clusters, and has g' ∼ 19.9, (g' - i') ∼ 0.6, concentration parameter c ∼ 1.0, a core radius r c ∼ 3.5 pc, and a half-light radius r h ∼ 5.5 pc. For M33 to have so few outer halo clusters compared to M31 suggests either tidal stripping of M33's outer halo clusters by M31, or a very different, much calmer accretion history of M33.

  12. Ground-Based Global Navigation Satellite System (GNSS) GPS Broadcast Ephemeris Data (daily files) from NASA CDDIS

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset consists of ground-based Global Navigation Satellite System (GNSS) GPS Broadcast Ephemeris Data (daily files) from the NASA Crustal Dynamics Data...

  13. Reliability analysis of a replication with limited number of journaling files

    International Nuclear Information System (INIS)

    Kimura, Mitsutaka; Imaizumi, Mitsuhiro; Nakagawa, Toshio

    2013-01-01

    Recently, replication mechanisms using journaling files have been widely used for the server systems. We have already discussed the model of asynchronous replication system using journaling files [8]. This paper formulates a stochastic model of a server system with replication considering the number of transmitting journaling files. The server updates the storage database and transmits the journaling file when a client requests the data update. The server transmits the database content to a backup site either at a constant time or after a constant number of transmitting journaling files. We derive the expected number of the replication and of transmitting journaling files. Further, we calculate the expected cost and discuss optimal replication interval to minimize it. Finally, numerical examples are given

  14. Apical Transportation, Centering Ability, and Cleaning Effectiveness of Reciprocating Single-file System Associated with Different Glide Path Techniques.

    Science.gov (United States)

    de Carvalho, Guilherme Moreira; Sponchiado Junior, Emílio Carlos; Garrido, Angela Delfina Bittencourt; Lia, Raphael Carlos Comelli; Garcia, Lucas da Fonseca Roberti; Marques, André Augusto Franco

    2015-12-01

    The aim of this study was to evaluate the apical transportation, the centering ability, and the cleaning effectiveness of a reciprocating single-file system associated to different glide path techniques. The mesial root canals of 52 mandibular molars were randomly distributed into 4 groups (n = 13) according to the different glide path techniques used before biomechanical preparation with Reciproc System (RS): KF/RS (sizes 10 and 15 K-files), NGP/RS (no glide path, only reciprocating system), PF/RS (sizes 13, 16, and 19 PathFile instruments), and NP (no preparation). Cone-beam computed tomography analysis was performed before and after instrumentation for apical third images acquisition. Apical transportation and its direction were evaluated by using the formula D = (X1 - X2) - (Y1 - Y2), and the centering ability was analyzed by the formula CC = (X1 - X2/Y1 - Y2 or Y1 - Y2/X1 - X2). The samples were submitted to histologic processing and analyzed under a digital microscope for debris quantification. The values were statistically analyzed (Kruskal-Wallis, the Dunn multiple comparisons test, P .05). Groups had a tendency toward transportation in the mesial direction. No technique had perfect centering ability (=1.0), with no significant difference among them. KF/RS had larger amount of debris, with statistically significant difference in comparison with NGP/RS (P > .05). The different glide path techniques promoted minimal apical transportation, and the reciprocating single-file system tested remained relatively centralized within the root canal. Also, the different techniques interfered in the cleaning effectiveness of the reciprocating system. Copyright © 2015 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  15. Development of a user-centered radiology teaching file system

    Science.gov (United States)

    dos Santos, Marcelo; Fujino, Asa

    2011-03-01

    Learning radiology requires systematic and comprehensive study of a large knowledge base of medical images. In this work is presented the development of a digital radiology teaching file system. The proposed system has been created in order to offer a set of customized services regarding to users' contexts and their informational needs. This has been done by means of an electronic infrastructure that provides easy and integrated access to all relevant patient data at the time of image interpretation, so that radiologists and researchers can examine all available data to reach well-informed conclusions, while protecting patient data privacy and security. The system is presented such as an environment which implements a distributed clinical database, including medical images, authoring tools, repository for multimedia documents, and also a peer-reviewed model which assures dataset quality. The current implementation has shown that creating clinical data repositories on networked computer environments points to be a good solution in terms of providing means to review information management practices in electronic environments and to create customized and contextbased tools for users connected to the system throughout electronic interfaces.

  16. Evaluation of the Self-Adjusting File system (SAF) for the instrumentation of primary molar root canals: a micro-computed tomographic study.

    Science.gov (United States)

    Kaya, E; Elbay, M; Yiğit, D

    2017-06-01

    The Self-Adjusting File (SAF) system has been recommended for use in permanent teeth since it offers more conservative and effective root-canal preparation when compared to traditional rotary systems. However, no study had evaluated the usage of SAF in primary teeth. The aim of this study was to evaluate and compare the use of SAF, K file (manual instrumentation) and Profile (traditional rotary instrumentation) systems for primary-tooth root-canal preparation in terms of instrumentation time and amounts of dentin removed using micro-computed tomography (μCT) technology. Study Design: The study was conducted with 60 human primary mandibular second molar teeth divided into 3 groups according to instrumentation technique: Group I: SAF (n=20); Group II: K file (n=20); Group III; Profile (n=20). Teeth were embedded in acrylic blocks and scanned with a μCT scanner prior to instrumentation. All distal root canals were prepared up to size 30 for K file,.04/30 for Profile and 2 mm thickness, size 25 for SAF; instrumentation time was recorded for each tooth, and a second μCT scan was performed after instrumentation was complete. Amounts of dentin removed were measured using the three-dimensional images by calculating the difference in root-canal volume before and after preparation. Data was statistically analysed using the Kolmogorov-Smirnov and Kruskal-Wallis tests. Manual instrumentation (K file) resulted in significantly more dentin removal when compared to rotary instrumentation (Profile and SAF), while the SAF system generated significantly less dentin removal than both manual instrumentation (K file) and traditional rotary instrumentation (Profile) (psystems. Within the experimental conditions of the present study, the SAF seems as a useful system for root-canal instrumentation in primary molars because it removed less dentin than other systems, which is especially important for the relatively thin-walled canals of primary teeth, and because it involves less

  17. A Measurement Study of the Structured Overlay Network in P2P File-Sharing Systems

    Directory of Open Access Journals (Sweden)

    Mo Zhou

    2007-01-01

    Full Text Available The architecture of P2P file-sharing applications has been developing to meet the needs of large scale demands. The structured overlay network, also known as DHT, has been used in these applications to improve the scalability, and robustness of the system, and to make it free from single-point failure. We believe that the measurement study of the overlay network used in the real file-sharing P2P systems can provide guidance for the designing of such systems, and improve the performance of the system. In this paper, we perform the measurement in two different aspects. First, a modified client is designed to provide view to the overlay network from a single-user vision. Second, the instances of crawler programs deployed in many nodes managed to crawl the user information of the overlay network as much as possible. We also find a vulnerability in the overlay network, combined with the character of the DNS service, a more serious DDoS attack can be launched.

  18. Application of the dynamically allocated virtual clustering management system to emulated tactical network experimentation

    Science.gov (United States)

    Marcus, Kelvin

    2014-06-01

    The U.S Army Research Laboratory (ARL) has built a "Network Science Research Lab" to support research that aims to improve their ability to analyze, predict, design, and govern complex systems that interweave the social/cognitive, information, and communication network genres. Researchers at ARL and the Network Science Collaborative Technology Alliance (NS-CTA), a collaborative research alliance funded by ARL, conducted experimentation to determine if automated network monitoring tools and task-aware agents deployed within an emulated tactical wireless network could potentially increase the retrieval of relevant data from heterogeneous distributed information nodes. ARL and NS-CTA required the capability to perform this experimentation over clusters of heterogeneous nodes with emulated wireless tactical networks where each node could contain different operating systems, application sets, and physical hardware attributes. Researchers utilized the Dynamically Allocated Virtual Clustering Management System (DAVC) to address each of the infrastructure support requirements necessary in conducting their experimentation. The DAVC is an experimentation infrastructure that provides the means to dynamically create, deploy, and manage virtual clusters of heterogeneous nodes within a cloud computing environment based upon resource utilization such as CPU load, available RAM and hard disk space. The DAVC uses 802.1Q Virtual LANs (VLANs) to prevent experimentation crosstalk and to allow for complex private networks. Clusters created by the DAVC system can be utilized for software development, experimentation, and integration with existing hardware and software. The goal of this paper is to explore how ARL and the NS-CTA leveraged the DAVC to create, deploy and manage multiple experimentation clusters to support their experimentation goals.

  19. Ground-Based Global Navigation Satellite System Mixed Broadcast Ephemeris Data (sub-hourly files) from NASA CDDIS

    Data.gov (United States)

    National Aeronautics and Space Administration — This dataset consists of ground-based Global Navigation Satellite System (GNSS) Mixed Broadcast Ephemeris Data (sub-hourly files) from the NASA Crustal Dynamics Data...

  20. Nodal aberration theory for wild-filed asymmetric optical systems

    Science.gov (United States)

    Chen, Yang; Cheng, Xuemin; Hao, Qun

    2016-10-01

    Nodal Aberration Theory (NAT) was used to calculate the zero field position in Full Field Display (FFD) for the given aberration term. Aiming at wide-filed non-rotational symmetric decentered optical systems, we have presented the nodal geography behavior of the family of third-order and fifth-order aberrations. Meanwhile, we have calculated the wavefront aberration expressions when one optical element in the system is tilted, which was not at the entrance pupil. By using a three-piece-cellphone lens example in optical design software CodeV, the nodal geography is testified under several situations; and the wavefront aberrations are calculated when the optical element is tilted. The properties of the nodal aberrations are analyzed by using Fringe Zernike coefficients, which are directly related with the wavefront aberration terms and usually obtained by real ray trace and wavefront surface fitting.

  1. 10 CFR 13.26 - Filing and service of papers.

    Science.gov (United States)

    2010-01-01

    ... found in the E-Filing Guidance and on the NRC Web site at http://www.nrc.gov/site-help/e-submittals.html... electronically to the E-Filing system. In addition, optical storage media (OSM) containing the entire filing must... document (e.g., motion to quash subpoena). (6) Filing is complete when the filer performs the last act that...

  2. A scalable and practical one-pass clustering algorithm for recommender system

    Science.gov (United States)

    Khalid, Asra; Ghazanfar, Mustansar Ali; Azam, Awais; Alahmari, Saad Ali

    2015-12-01

    KMeans clustering-based recommendation algorithms have been proposed claiming to increase the scalability of recommender systems. One potential drawback of these algorithms is that they perform training offline and hence cannot accommodate the incremental updates with the arrival of new data, making them unsuitable for the dynamic environments. From this line of research, a new clustering algorithm called One-Pass is proposed, which is a simple, fast, and accurate. We show empirically that the proposed algorithm outperforms K-Means in terms of recommendation and training time while maintaining a good level of accuracy.

  3. 78 FR 67927 - Foreign Trade Regulations (FTR): Mandatory Automated Export System Filing for All Shipments...

    Science.gov (United States)

    2013-11-13

    ... Internet at [email protected] ). FOR FURTHER INFORMATION CONTACT: Nick Orsini, Chief, Foreign Trade Division... 0607-AA50 Foreign Trade Regulations (FTR): Mandatory Automated Export System Filing for All Shipments... approval of new information collection requirements. [[Page 67928

  4. NJOY99, Data Processing System of Evaluated Nuclear Data Files ENDF Format

    International Nuclear Information System (INIS)

    2000-01-01

    1 - Description of program or function: The NJOY nuclear data processing system is a modular computer code used for converting evaluated nuclear data in the ENDF format into libraries useful for applications calculations. Because the Evaluated Nuclear Data File (ENDF) format is used all around the world (e.g., ENDF/B-VI in the US, JEF-2.2 in Europe, JENDL-3.2 in Japan, BROND-2.2 in Russia), NJOY gives its users access to a wide variety of the most up-to-date nuclear data. NJOY provides comprehensive capabilities for processing evaluated data, and it can serve applications ranging from continuous-energy Monte Carlo (MCNP), through deterministic transport codes (DANT, ANISN, DORT), to reactor lattice codes (WIMS, EPRI). NJOY handles a wide variety of nuclear effects, including resonances, Doppler broadening, heating (KERMA), radiation damage, thermal scattering (even cold moderators), gas production, neutrons and charged particles, photo-atomic interactions, self shielding, probability tables, photon production, and high-energy interactions (to 150 MeV). Output can include printed listings, special library files for applications, and Postscript graphics (plus color). More information on NJOY is available from the developer's home page at http://t2.lanl.gov/tour/tourbus.html. Follow the Tourbus section of the Tour area to find notes from the ICTP lectures held at Trieste in March 2000 on the ENDF format and on the NJOY code. NJOY contains the following modules: NJOY directs the flow of data through the other modules and contains a library of common functions and subroutines used by the other modules. RECONR reconstructs pointwise (energy-dependent) cross sections from ENDF resonance parameters and interpolation schemes. BROADR Doppler broadens and thins pointwise cross sections. UNRESR computes effective self-shielded pointwise cross sections in the unresolved energy range. HEATR generates pointwise heat production cross sections (KERMA coefficients) and radiation

  5. The version control service for ATLAS data acquisition configuration files

    CERN Document Server

    Soloviev, Igor; The ATLAS collaboration

    2012-01-01

    To configure data taking session the ATLAS systems and detectors store more than 160 MBytes of data acquisition related configuration information in OKS XML files [1]. The total number of the files exceeds 1300 and they are updated by many system experts. In the past from time to time after such updates we had experienced problems caused by XML syntax errors or inconsistent state of files from a point of view of the overall ATLAS configuration. It was not always possible to know who made a modification causing problems or how to go back to a previous version of the modified file. Few years ago a special service addressing these issues has been implemented and deployed on ATLAS Point-1. It excludes direct write access to XML files stored in a central database repository. Instead, for an update the files are copied into a user repository, validated after modifications and committed using a version control system. The system's callback updates the central repository. Also, it keeps track of all modifications pro...

  6. 75 FR 54869 - Combined Notice of Filings 3

    Science.gov (United States)

    2010-09-09

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings 3 September 02, 2010. Take notice that the Commission has received the following Natural Gas Pipeline Rate and...: Venice Gathering System, L.L.C. Description: Venice Gathering System, L.L.C. submits tariff filing per...

  7. WinHPC System Configuration | High-Performance Computing | NREL

    Science.gov (United States)

    ), login node (WinHPC02) and worker/compute nodes. The head node acts as the file, DNS, and license server . The login node is where the users connect to access the cluster. Node 03 has dual Intel Xeon E5530 2008 R2 HPC Edition. The login node, WinHPC02, is where users login to access the system. This is where

  8. Properties of an ionised-cluster beam from a vaporised-cluster ion source

    International Nuclear Information System (INIS)

    Takagi, T.; Yamada, I.; Sasaki, A.

    1978-01-01

    A new type of ion source vaporised-metal cluster ion source, has been developed for deposition and epitaxy. A cluster consisting of 10 2 to 10 3 atoms coupled loosely together is formed by adiabatic expansion ejecting the vapour of materials into a high-vacuum region through the nozzle of a heated crucible. The clusters are ionised by electron bombardment and accelerated with neutral clusters toward a substrate. In this paper, mechanisms of cluster formation experimental results of the cluster size (atoms/cluster) and its distribution, and characteristics of the cluster ion beams are reported. The size is calculated from the kinetic equation E = (1/2)mNVsub(ej) 2 , where E is the cluster beam energy, Vsub(ej) is the ejection velocity, m is the mass of atom and N is the cluster size. The energy and the velocity of the cluster are measured by an electrostatic 127 0 energy analyser and a rotating disc system, respectively. The cluster size obtained for Ag is about 5 x 10 2 to 2 x 10 3 atoms. The retarding potential method is used to confirm the results for Ag. The same dependence on cluster size for metals such as Ag, Cu and Pb has been obtained in previous experiments. In the cluster state the cluster ion beam is easily produced by electron bombardment. About 50% of ionised clusters are obtained under typical operation conditions, because of the large ionisation cross sections of the clusters. To obtain a uniform spatial distribution, the ionising electrode system is also discussed. The new techniques are termed ionised-cluster beam deposition (ICBD) and epitaxy (ICBE). (author)

  9. 77 FR 13114 - Combined Notice of Filings #1

    Science.gov (United States)

    2012-03-05

    .../12 Docket Numbers: ER12-1013-001 Applicants: Physical Systems Integration, LLC Description: Physical Systems Integration, LLC--Amendment to MBR Application to be effective 3/1/2012. Filed Date: 2/24/12... Market-Based Rate Tariff of Hampton Lumber Mills-Washington, Inc. Filed Date: 2/27/12 Accession Number...

  10. DATA Act File C Award Financial - Social Security

    Data.gov (United States)

    Social Security Administration — The DATA Act Information Model Schema Reporting Submission Specification File C. File C includes the agency award information from the financial accounting system at...

  11. Image Steganography of Multiple File Types with Encryption and Compression Algorithms

    Directory of Open Access Journals (Sweden)

    Ernest Andreigh C. Centina

    2017-05-01

    Full Text Available The goals of this study were to develop a system intended for securing files through the technique of image steganography integrated with cryptography by utilizing ZLIB Algorithm for compressing and decompressing secret files, DES Algorithm for encryption and decryption, and Least Significant Bit Algorithm for file embedding and extraction to avoid compromise on highly confidential files from exploits of unauthorized persons. Ensuing to this, the system is in acc ordance with ISO 9126 international quality standards. Every quality criteria of the system was evaluated by 10 Information Technology professionals, and the arithmetic Mean and Standard Deviation of the survey were computed. The result exhibits that m ost of them strongly agreed that the system is excellently effective based on Functionality, Reliability, Usability, Efficiency, Maintainability and Portability conformance to ISO 9126 standards. The system was found to be a useful tool for both governmen t agencies and private institutions for it could keep not only the message secret but also the existence of that particular message or file et maintaining the privacy of highly confidential and sensitive files from unauthorized access.

  12. The sluggs survey: HST/ACS mosaic imaging of the NGC 3115 globular cluster system

    Energy Technology Data Exchange (ETDEWEB)

    Jennings, Zachary G.; Romanowsky, Aaron J.; Brodie, Jean P.; Arnold, Jacob A. [University of California Observatories, Santa Cruz, CA 95064 (United States); Strader, Jay [Department of Physics and Astronomy, Michigan State University, East Lansing, Michigan, MI 48824 (United States); Lin, Dacheng; Irwin, Jimmy A.; Wong, Ka-Wah [Department of Physics and Astronomy, University of Alabama, Box 870324, Tuscaloosa, AL 35487 (United States); Sivakoff, Gregory R., E-mail: zgjennin@ucsc.edu [Department of Physics, University of Alberta, Edmonton, Alberta T6G 2E1 (Canada)

    2014-08-01

    We present Hubble Space Telescope/Advanced Camera for Surveys (HST/ACS) g and z photometry and half-light radii R {sub h} measurements of 360 globular cluster (GC) candidates around the nearby S0 galaxy NGC 3115. We also include Subaru/Suprime-Cam g, r, and i photometry of 421 additional candidates. The well-established color bimodality of the GC system is obvious in the HST/ACS photometry. We find evidence for a 'blue tilt' in the blue GC subpopulation, wherein the GCs in the blue subpopulation get redder as luminosity increases, indicative of a mass-metallicity relationship. We find a color gradient in both the red and blue subpopulations, with each group of clusters becoming bluer at larger distances from NGC 3115. The gradient is of similar strength in both subpopulations, but is monotonic and more significant for the blue clusters. On average, the blue clusters have ∼10% larger R {sub h} than the red clusters. This average difference is less than is typically observed for early-type galaxies but does match that measured in the literature for the Sombrero Galaxy (M104), suggesting that morphology and inclination may affect the measured size difference between the red and blue clusters. However, the scatter on the R {sub h} measurements is large. We also identify 31 clusters more extended than typical GCs, which we term ultra-compact dwarf (UCD) candidates. Many of these objects are actually considerably fainter than typical UCDs. While it is likely that a significant number will be background contaminants, six of these UCD candidates are spectroscopically confirmed as NGC 3115 members. To explore the prevalence of low-mass X-ray binaries in the GC system, we match our ACS and Suprime-Cam detections to corresponding Chandra X-ray sources. We identify 45 X-ray-GC matches: 16 among the blue subpopulation and 29 among the red subpopulation. These X-ray/GC coincidence fractions are larger than is typical for most GC systems, probably due to the increased

  13. 47 CFR 1.10007 - What applications can I file electronically?

    Science.gov (United States)

    2010-10-01

    ... International Bureau Filing System § 1.10007 What applications can I file electronically? (a) For a complete... the following formats: Word, Adobe Acrobat, Excel and Text. (c) For paper filing rules and procedures...

  14. Molecular depth profiling of multi-layer systems with cluster ion sources

    Energy Technology Data Exchange (ETDEWEB)

    Cheng, Juan [Department of Chemistry, Penn State University, University Park, PA 16802 (United States); Winograd, Nicholas [Department of Chemistry, Penn State University, University Park, PA 16802 (United States)]. E-mail: nxw@psu.edu

    2006-07-30

    Cluster bombardment of molecular films has created new opportunities for SIMS research. To more quantitatively examine the interaction of cluster beams with organic materials, we have developed a reproducible platform consisting of a well-defined sugar film (trehalose) doped with peptides. Molecular depth profiles have been acquired with these systems using C{sub 60} {sup +} bombardment. In this study, we utilize this platform to determine the feasibility of examining buried interfaces for multi-layer systems. Using C{sub 60} {sup +} at 20 keV, several systems have been tested including Al/trehalose/Si, Al/trehalose/Al/Si, Ag/trehalose/Si and ice/trehalose/Si. The results show that there can be interactions between the layers during the bombardment process that prevent a simple interpretation of the depth profile. We find so far that the best results are obtained when the mass of the overlayer atoms is less than or nearly equal to the mass of the atoms in buried molecules. In general, these observations suggest that C{sub 60} {sup +} bombardment can be successfully applied to interface characterization of multi-layer systems if the systems are carefully chosen.

  15. Cluster management.

    Science.gov (United States)

    Katz, R

    1992-11-01

    Cluster management is a management model that fosters decentralization of management, develops leadership potential of staff, and creates ownership of unit-based goals. Unlike shared governance models, there is no formal structure created by committees and it is less threatening for managers. There are two parts to the cluster management model. One is the formation of cluster groups, consisting of all staff and facilitated by a cluster leader. The cluster groups function for communication and problem-solving. The second part of the cluster management model is the creation of task forces. These task forces are designed to work on short-term goals, usually in response to solving one of the unit's goals. Sometimes the task forces are used for quality improvement or system problems. Clusters are groups of not more than five or six staff members, facilitated by a cluster leader. A cluster is made up of individuals who work the same shift. For example, people with job titles who work days would be in a cluster. There would be registered nurses, licensed practical nurses, nursing assistants, and unit clerks in the cluster. The cluster leader is chosen by the manager based on certain criteria and is trained for this specialized role. The concept of cluster management, criteria for choosing leaders, training for leaders, using cluster groups to solve quality improvement issues, and the learning process necessary for manager support are described.

  16. 47 CFR 1.1113 - Filing locations.

    Science.gov (United States)

    2010-10-01

    ... by the ULS electronic filing system on FCC Form 159) and submit such number with the payment in order... received more than ten (10) days after electronic filing of an application on a Bureau/Office electronic... Television Relay Service (CARS) that are submitted electronically via the Cable Operations and Licensing...

  17. Effect of Policy Analysis on Indonesia’s Maritime Cluster Development Using System Dynamics Modeling

    Science.gov (United States)

    Nursyamsi, A.; Moeis, A. O.; Komarudin

    2018-03-01

    As an archipelago with two third of its territory consist of water, Indonesia should address more attention to its maritime industry development. One of the catalyst to fasten the maritime industry growth is by developing a maritime cluster. The purpose of this research is to gain understanding of the effect if Indonesia implement maritime cluster policy to the growth of maritime economic and its role to enhance the maritime cluster performance, hence enhancing Indonesia’s maritime industry as well. The result of the constructed system dynamic model simulation shows that with the effect of maritime cluster, the growth of employment rate and maritime economic is much bigger that the business as usual case exponentially. The result implies that the government should act fast to form a legitimate cluster maritime organizer institution so that there will be a synergize, sustainable, and positive maritime cluster environment that will benefit the performance of Indonesia’s maritime industry.

  18. Super computer made with Linux cluster

    International Nuclear Information System (INIS)

    Lee, Jeong Hun; Oh, Yeong Eun; Kim, Jeong Seok

    2002-01-01

    This book consists of twelve chapters, which introduce super computer made with Linux cluster. The contents of this book are Linux cluster, the principle of cluster, design of Linux cluster, general things for Linux, building up terminal server and client, Bear wolf cluster by Debian GNU/Linux, cluster system with red hat, Monitoring system, application programming-MPI, on set-up and install application programming-PVM, with PVM programming and XPVM application programming-open PBS with composition and install and set-up and GRID with GRID system, GSI, GRAM, MDS, its install and using of tool kit

  19. Derived virtual devices: a secure distributed file system mechanism

    Science.gov (United States)

    VanMeter, Rodney; Hotz, Steve; Finn, Gregory

    1996-01-01

    This paper presents the design of derived virtual devices (DVDs). DVDs are the mechanism used by the Netstation Project to provide secure shared access to network-attached peripherals distributed in an untrusted network environment. DVDs improve Input/Output efficiency by allowing user processes to perform I/O operations directly from devices without intermediate transfer through the controlling operating system kernel. The security enforced at the device through the DVD mechanism includes resource boundary checking, user authentication, and restricted operations, e.g., read-only access. To illustrate the application of DVDs, we present the interactions between a network-attached disk and a file system designed to exploit the DVD abstraction. We further discuss third-party transfer as a mechanism intended to provide for efficient data transfer in a typical NAP environment. We show how DVDs facilitate third-party transfer, and provide the security required in a more open network environment.

  20. Experience on QA in the CernVM File System

    CERN Multimedia

    CERN. Geneva; MEUSEL, Rene

    2015-01-01

    The CernVM-File System (CVMFS) delivers experiment software installations to thousands of globally distributed nodes in the WLCG and beyond. In recent years it became a mission-critical component for offline data processing of the LHC experiments and many other collaborations. From a software engineering perspective, CVMFS is a medium-sized C++ system-level project. Following the growth of the project, we introduced a number of measures to improve the code quality, testability, and maintainability. In particular, we found very useful code reviews through github pull requests and automated unit- and integration testing. We are also transitioning to a test-driven development for new features and bug fixes. These processes are supported by a number of tools, such as Google Test, Jenkins, Docker, and others. We would like to share our experience on problems we encountered and on which processes and tools worked well for us.

  1. Clustering of near clusters versus cluster compactness

    International Nuclear Information System (INIS)

    Yu Gao; Yipeng Jing

    1989-01-01

    The clustering properties of near Zwicky clusters are studied by using the two-point angular correlation function. The angular correlation functions for compact and medium compact clusters, for open clusters, and for all near Zwicky clusters are estimated. The results show much stronger clustering for compact and medium compact clusters than for open clusters, and that open clusters have nearly the same clustering strength as galaxies. A detailed study of the compactness-dependence of correlation function strength is worth investigating. (author)

  2. System of HPC content archiving

    Science.gov (United States)

    Bogdanov, A.; Ivashchenko, A.

    2017-12-01

    This work is aimed to develop a system, that will effectively solve the problem of storing and analyzing files containing text data, by using modern software development tools, techniques and approaches. The main challenge of storing a large number of text documents defined at the problem formulation stage, have to be resolved with such functionality as full text search and document clustering depends on their contents. Main system features could be described with notions of distributed multilevel architecture, flexibility and interchangeability of components, achieved through the standard functionality incapsulation in independent executable modules.

  3. Reduced-dimension power allocation over clustered channels in cognitive radios system under co-channel interference

    KAUST Repository

    Ben Ghorbel, Mahdi; Guenach, Mamoun; Alouini, Mohamed-Slim

    2014-01-01

    and secondary systems. Instead of optimizing the powers over all sub-carriers, the sub-carriers are grouped into clusters of sub-carriers, where the power of each sub-carrier is directly related to the power of the correspondent cluster. The power optimization

  4. A Coupled User Clustering Algorithm Based on Mixed Data for Web-Based Learning Systems

    Directory of Open Access Journals (Sweden)

    Ke Niu

    2015-01-01

    Full Text Available In traditional Web-based learning systems, due to insufficient learning behaviors analysis and personalized study guides, a few user clustering algorithms are introduced. While analyzing the behaviors with these algorithms, researchers generally focus on continuous data but easily neglect discrete data, each of which is generated from online learning actions. Moreover, there are implicit coupled interactions among the data but are frequently ignored in the introduced algorithms. Therefore, a mass of significant information which can positively affect clustering accuracy is neglected. To solve the above issues, we proposed a coupled user clustering algorithm for Wed-based learning systems by taking into account both discrete and continuous data, as well as intracoupled and intercoupled interactions of the data. The experiment result in this paper demonstrates the outperformance of the proposed algorithm.

  5. An evaluated neutronic data file for elemental zirconium

    International Nuclear Information System (INIS)

    Smith, A.B.; Chiba, S.

    1994-09-01

    A comprehensive evaluated neutronic data file for elemental zirconium is derived and presented in the ENDF/B-VI formats. The derivation is based upon measured microscopic nuclear data, augmented by model calculations as necessary. The primary objective is a quality contemporary file suitable for fission-reactor development extending from conventional thermal to fast and innovative systems. This new file is a significant improvement over previously available evaluated zirconium files, in part, as a consequence of extensive new experimental measurements reported elsewhere

  6. Conjunctive Conceptual Clustering: A Methodology and Experimentation.

    Science.gov (United States)

    1987-09-01

    observing a typical restaurant table on vhich there are such objects as food on a plate, a salad, utensils, salt and pepper, napkins , a ase with flowers, a...colored graph has nodes and inks that match only if they have corre-ponding link-olor and node-color labelg 4w 80 [SEtexture sa lif ba p S M i e d If...LINK LINK LINK LINK LINK 9 0 1 OPENdRECT RECTLOD 1 2 CL 10 0 0 LINK LINK INK LINK LINK ,~ . 0.5. Input file for attribute-based clustering The

  7. Cluster-cluster clustering

    International Nuclear Information System (INIS)

    Barnes, J.; Dekel, A.; Efstathiou, G.; Frenk, C.S.; Yale Univ., New Haven, CT; California Univ., Santa Barbara; Cambridge Univ., England; Sussex Univ., Brighton, England)

    1985-01-01

    The cluster correlation function xi sub c(r) is compared with the particle correlation function, xi(r) in cosmological N-body simulations with a wide range of initial conditions. The experiments include scale-free initial conditions, pancake models with a coherence length in the initial density field, and hybrid models. Three N-body techniques and two cluster-finding algorithms are used. In scale-free models with white noise initial conditions, xi sub c and xi are essentially identical. In scale-free models with more power on large scales, it is found that the amplitude of xi sub c increases with cluster richness; in this case the clusters give a biased estimate of the particle correlations. In the pancake and hybrid models (with n = 0 or 1), xi sub c is steeper than xi, but the cluster correlation length exceeds that of the points by less than a factor of 2, independent of cluster richness. Thus the high amplitude of xi sub c found in studies of rich clusters of galaxies is inconsistent with white noise and pancake models and may indicate a primordial fluctuation spectrum with substantial power on large scales. 30 references

  8. Shaping ability of single-file reciprocating and heat-treated multifile rotary systems: a micro-CT study.

    Science.gov (United States)

    Marceliano-Alves, M F V; Sousa-Neto, M D; Fidel, S R; Steier, L; Robinson, J P; Pécora, J D; Versiani, M A

    2015-12-01

    To investigate changes in three-dimensional geometry, in various cross-sectional morphological parameters and in the centring ability of root canals prepared with different preparation systems using microcomputed tomographic imaging technology. Sixty-four mesial canals of mandibular molars were matched based on similar morphological dimensions using micro-CT evaluation and assigned to four experimental groups (n = 16), according to the canal preparation technique: Reciproc, WaveOne, Twisted File and HyFlex CM systems. Changes in several 2D (area, perimeter, form factor, roundness, minor and major diameter) and 3D [volume, surface area, structure model index (SMI)] morphological parameters, as well as canal transportation, were compared with preoperative values using Kruskal-Wallis and anovapost hoc Tukey's tests with the significance level set at 5%. Preparation significantly increased all tested parameters in the experimental groups. No significant differences were observed between groups regarding changes in volume, surface area, SMI, form factor and roundness of the root canal after preparation (P > 0.05). In the apical third, the Reciproc group had significantly greater changes in canal area, perimeter, major and minor diameters than the other groups (P File and HyFlex CM systems were associated with significantly less transportation than the reciprocating instruments, Reciproc and WaveOne (P File and HyFlex CM systems were able to maintain the original canal anatomy with less canal transportation than Reciproc and WaveOne; however, these differences are unlikely to be of clinical significance. © 2014 International Endodontic Journal. Published by John Wiley & Sons Ltd.

  9. Resistance–temperature relation and atom cluster estimation of In–Bi system melts

    International Nuclear Information System (INIS)

    Geng, Haoran; Wang Zhiming; Zhou Yongzhi; Li Cancan

    2012-01-01

    Highlights: ► A testing device was adopted to measure the electrical resistivity of In–Bi system melts. ► A basically linear relation exists between the resistivity and temperature of In x Bi 100−x melts in measured temperature range. ► Based on Novakovic's assumption, the content of InBi atomic cluster in In x Bi 100−x melt is estimated with ρ ≈ ρ InBi x InBi + ρ m (1 − x InBi ) equation. - Abstract: A testing device for the resistivity of high-temperature melt was adopted to measure the l resistivity of In–Bi system melts at different temperatures. It can be concluded from the analysis and calculation of the experimental results that the resistivity of In x Bi 100−x (x = 0–100) melt is in linear relationship with temperature within the experiment temperature range. The resistivity of the melt decreases with the increasing content of In. The fair consistency of resistivity of In–Bi system melt is found in the heating and cooling processes. On the basis of Novakovic's assumption, we approximately estimated the content of InBi atom clusters in In x Bi 100−x melts with the resistivity data by equation ρ ≈ ρ InBi x InBi + ρ m (1 − x InBi ). In the whole components interval, the content corresponds well with the mole fraction of InBi clusters calculated by Novakovic in the thermodynamic approach. The mole fraction of InBi type atom clusters in the melts reaches the maximum at the point of stoichiometric composition In 50 Bi 50 .

  10. 76 FR 55375 - Combined Notice of Filings

    Science.gov (United States)

    2011-09-07

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings Take notice that the Commission has received the following Natural Gas Pipeline Rate and Refund Report filings..., September 12, 2011. Docket Numbers: RP11-2454-000. Applicants: Venice Gathering System, LLC. Description...

  11. 76 FR 39085 - Combined Notice of Filings

    Science.gov (United States)

    2011-07-05

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings Take notice that the Commission has received the following Natural Gas Pipeline Rate and Refund Report filings..., July 05, 2011. Docket Numbers: RP11-2205-000. Applicants: Venice Gathering System, LLC. Description...

  12. Retrieval with Clustering in a Case-Based Reasoning System for Radiotherapy Treatment Planning

    Science.gov (United States)

    Khussainova, Gulmira; Petrovic, Sanja; Jagannathan, Rupa

    2015-05-01

    Radiotherapy treatment planning aims to deliver a sufficient radiation dose to cancerous tumour cells while sparing healthy organs in the tumour surrounding area. This is a trial and error process highly dependent on the medical staff's experience and knowledge. Case-Based Reasoning (CBR) is an artificial intelligence tool that uses past experiences to solve new problems. A CBR system has been developed to facilitate radiotherapy treatment planning for brain cancer. Given a new patient case the existing CBR system retrieves a similar case from an archive of successfully treated patient cases with the suggested treatment plan. The next step requires adaptation of the retrieved treatment plan to meet the specific demands of the new case. The CBR system was tested by medical physicists for the new patient cases. It was discovered that some of the retrieved cases were not suitable and could not be adapted for the new cases. This motivated us to revise the retrieval mechanism of the existing CBR system by adding a clustering stage that clusters cases based on their tumour positions. A number of well-known clustering methods were investigated and employed in the retrieval mechanism. Results using real world brain cancer patient cases have shown that the success rate of the new CBR retrieval is higher than that of the original system.

  13. Retrieval with Clustering in a Case-Based Reasoning System for Radiotherapy Treatment Planning

    International Nuclear Information System (INIS)

    Khussainova, Gulmira; Petrovic, Sanja; Jagannathan, Rupa

    2015-01-01

    Radiotherapy treatment planning aims to deliver a sufficient radiation dose to cancerous tumour cells while sparing healthy organs in the tumour surrounding area. This is a trial and error process highly dependent on the medical staff's experience and knowledge. Case-Based Reasoning (CBR) is an artificial intelligence tool that uses past experiences to solve new problems. A CBR system has been developed to facilitate radiotherapy treatment planning for brain cancer. Given a new patient case the existing CBR system retrieves a similar case from an archive of successfully treated patient cases with the suggested treatment plan. The next step requires adaptation of the retrieved treatment plan to meet the specific demands of the new case. The CBR system was tested by medical physicists for the new patient cases. It was discovered that some of the retrieved cases were not suitable and could not be adapted for the new cases. This motivated us to revise the retrieval mechanism of the existing CBR system by adding a clustering stage that clusters cases based on their tumour positions. A number of well-known clustering methods were investigated and employed in the retrieval mechanism. Results using real world brain cancer patient cases have shown that the success rate of the new CBR retrieval is higher than that of the original system. (paper)

  14. SU-F-T-172: A Method for Log File QA On An IBA Proteus System for Patient Specific Spot Scanning Quality Assurance

    Energy Technology Data Exchange (ETDEWEB)

    Tang, S; Ho, M; Chen, C; Mah, D [ProCure NJ, Somerset, NJ (United States); Rice, I; Doan, D; Mac Rae, B [IBA, Somerset, NJ (United States)

    2016-06-15

    Purpose: The use of log files to perform patient specific quality assurance for both protons and IMRT has been established. Here, we extend that approach to a proprietary log file format and compare our results to measurements in phantom. Our goal was to generate a system that would permit gross errors to be found within 3 fractions until direct measurements. This approach could eventually replace direct measurements. Methods: Spot scanning protons pass through multi-wire ionization chambers which provide information about the charge, location, and size of each delivered spot. We have generated a program that calculates the dose in phantom from these log files and compares the measurements with the plan. The program has 3 different spot shape models: single Gaussian, double Gaussian and the ASTROID model. The program was benchmarked across different treatment sites for 23 patients and 74 fields. Results: The dose calculated from the log files were compared to those generate by the treatment planning system (Raystation). While the dual Gaussian model often gave better agreement, overall, the ASTROID model gave the most consistent results. Using a 5%–3 mm gamma with a 90% passing criteria and excluding doses below 20% of prescription all patient samples passed. However, the degree of agreement of the log file approach was slightly worse than that of the chamber array measurement approach. Operationally, this implies that if the beam passes the log file model, it should pass direct measurement. Conclusion: We have established and benchmarked a model for log file QA in an IBA proteus plus system. The choice of optimal spot model for a given class of patients may be affected by factors such as site, field size, and range shifter and will be investigated further.

  15. Disk cloning program 'Dolly+' for system management of PC Linux cluster

    International Nuclear Information System (INIS)

    Atsushi Manabe

    2001-01-01

    The Dolly+ is a Linux application program to clone files and disk partition image from a PC to many others. By using several techniques such as logical ring connection, multi threading and pipelining, it could achieve high performance and scalability. For example, in typical condition, installations to a hundred PCs takes almost equivalent time for two PCs. Together with the Intel PXE and the RedHat kickstart, automatic and very fast system installation and upgrading could be performed

  16. 75 FR 6728 - The Merit Systems Protection Board (MSPB) is Providing Notice of the Opportunity to File Amicus...

    Science.gov (United States)

    2010-02-10

    ... MERIT SYSTEMS PROTECTION BOARD The Merit Systems Protection Board (MSPB) is Providing Notice of the Opportunity to File Amicus Briefs in the Matters of Conyers v. Department of Defense, Docket No...: Merit Systems Protection Board. ACTION: Notice. SUMMARY: On January 25, 2010, the MSPB published in the...

  17. VizieR Online Data Catalog: 24um excesses in clusters & membership of NGC2244 (Meng+, 2017)

    Science.gov (United States)

    Meng, H. Y. A.; Rieke, G. H.; Su, K. Y. L.; Gaspar, A.

    2017-09-01

    We re-measured the Spitzer/MIPS 24um photometry for all the clusters except for a few noted in Appendix A. We obtained the 24um data from the Spitzer Heritage Archive between 2004 Jan 28 and 2008 Oct 23. (3 data files).

  18. Hierarchical Clustering of Large Databases and Classification of Antibiotics at High Noise Levels

    Directory of Open Access Journals (Sweden)

    Alexander V. Yarkov

    2008-12-01

    Full Text Available A new algorithm for divisive hierarchical clustering of chemical compounds based on 2D structural fragments is suggested. The algorithm is deterministic, and given a random ordering of the input, will always give the same clustering and can process a database up to 2 million records on a standard PC. The algorithm was used for classification of 1,183 antibiotics mixed with 999,994 random chemical structures. Similarity threshold, at which best separation of active and non active compounds took place, was estimated as 0.6. 85.7% of the antibiotics were successfully classified at this threshold with 0.4% of inaccurate compounds. A .sdf file was created with the probe molecules for clustering of external databases.

  19. A Human Activity Recognition System Based on Dynamic Clustering of Skeleton Data

    Directory of Open Access Journals (Sweden)

    Alessandro Manzi

    2017-05-01

    Full Text Available Human activity recognition is an important area in computer vision, with its wide range of applications including ambient assisted living. In this paper, an activity recognition system based on skeleton data extracted from a depth camera is presented. The system makes use of machine learning techniques to classify the actions that are described with a set of a few basic postures. The training phase creates several models related to the number of clustered postures by means of a multiclass Support Vector Machine (SVM, trained with Sequential Minimal Optimization (SMO. The classification phase adopts the X-means algorithm to find the optimal number of clusters dynamically. The contribution of the paper is twofold. The first aim is to perform activity recognition employing features based on a small number of informative postures, extracted independently from each activity instance; secondly, it aims to assess the minimum number of frames needed for an adequate classification. The system is evaluated on two publicly available datasets, the Cornell Activity Dataset (CAD-60 and the Telecommunication Systems Team (TST Fall detection dataset. The number of clusters needed to model each instance ranges from two to four elements. The proposed approach reaches excellent performances using only about 4 s of input data (~100 frames and outperforms the state of the art when it uses approximately 500 frames on the CAD-60 dataset. The results are promising for the test in real context.

  20. A Human Activity Recognition System Based on Dynamic Clustering of Skeleton Data.

    Science.gov (United States)

    Manzi, Alessandro; Dario, Paolo; Cavallo, Filippo

    2017-05-11

    Human activity recognition is an important area in computer vision, with its wide range of applications including ambient assisted living. In this paper, an activity recognition system based on skeleton data extracted from a depth camera is presented. The system makes use of machine learning techniques to classify the actions that are described with a set of a few basic postures. The training phase creates several models related to the number of clustered postures by means of a multiclass Support Vector Machine (SVM), trained with Sequential Minimal Optimization (SMO). The classification phase adopts the X-means algorithm to find the optimal number of clusters dynamically. The contribution of the paper is twofold. The first aim is to perform activity recognition employing features based on a small number of informative postures, extracted independently from each activity instance; secondly, it aims to assess the minimum number of frames needed for an adequate classification. The system is evaluated on two publicly available datasets, the Cornell Activity Dataset (CAD-60) and the Telecommunication Systems Team (TST) Fall detection dataset. The number of clusters needed to model each instance ranges from two to four elements. The proposed approach reaches excellent performances using only about 4 s of input data (~100 frames) and outperforms the state of the art when it uses approximately 500 frames on the CAD-60 dataset. The results are promising for the test in real context.

  1. Files synchronization from a large number of insertions and deletions

    Science.gov (United States)

    Ellappan, Vijayan; Kumari, Savera

    2017-11-01

    Synchronization between different versions of files is becoming a major issue that most of the applications are facing. To make the applications more efficient a economical algorithm is developed from the previously used algorithm of “File Loading Algorithm”. I am extending this algorithm in three ways: First, dealing with non-binary files, Second backup is generated for uploaded files and lastly each files are synchronized with insertions and deletions. User can reconstruct file from the former file with minimizing the error and also provides interactive communication by eliminating the frequency without any disturbance. The drawback of previous system is overcome by using synchronization, in which multiple copies of each file/record is created and stored in backup database and is efficiently restored in case of any unwanted deletion or loss of data. That is, to introduce a protocol that user B may use to reconstruct file X from file Y with suitably low probability of error. Synchronization algorithms find numerous areas of use, including data storage, file sharing, source code control systems, and cloud applications. For example, cloud storage services such as Drop box synchronize between local copies and cloud backups each time users make changes to local versions. Similarly, synchronization tools are necessary in mobile devices. Specialized synchronization algorithms are used for video and sound editing. Synchronization tools are also capable of performing data duplication.

  2. Canal shaping with WaveOne Primary reciprocating files and ProTaper system: a comparative study.

    Science.gov (United States)

    Berutti, Elio; Chiandussi, Giorgio; Paolino, Davide Salvatore; Scotti, Nicola; Cantatore, Giuseppe; Castellucci, Arnaldo; Pasqualini, Damiano

    2012-04-01

    This study compared the canal curvature and axis modification after instrumentation with WaveOne Primary reciprocating files (Dentsply Maillefer, Ballaigues, Switzerland) and nickel-titanium (NiTi) rotary ProTaper (Dentsply Maillefer). Thirty ISO 15, 0.02 taper, Endo Training Blocks (Dentsply Maillefer) were used. In all specimens, the glide path was achieved with PathFile 1, 2, and 3 (Dentsply Maillefer) at the working length (WL). Specimens were then assigned to 1 of 2 groups for shaping: specimens in group 1 were shaped with ProTaper S1-S2-F1-F2 at the WL and specimens in group 2 were shaped with WaveOne Primary reciprocating files at the WL. Pre- and postinstrumentation digital images were superimposed and processed with Matlab r2010b (The MathWorks Inc, Natick, MA) software to analyze the curvature-radius ratio (CRr) and the relative axis error (rAe), representing canal curvature modification. Data were analyzed with one-way balanced analyses of variance at 2 levels (P < .05). The instrument factor was extremely significant for both the CRr parameter (F(1) = 9.59, P = .004) and the rAe parameter (F(1) = 13.55, P = .001). Canal modifications are reduced when the new WaveOne NiTi single-file system is used. Copyright © 2012 American Association of Endodontists. Published by Elsevier Inc. All rights reserved.

  3. 77 FR 71408 - Combined Notice of Filings

    Science.gov (United States)

    2012-11-30

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings Take notice that the Commission has received the following Natural Gas Pipeline Rate and Refund Report filings....m. ET 12/3/12. Docket Numbers: RP13-301-000. Applicants: Venice Gathering System, LLC. Description...

  4. 78 FR 50408 - Combined Notice of Filings

    Science.gov (United States)

    2013-08-19

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings Take notice that the Commission has received the following Natural Gas Pipeline Rate and Refund Report filings... Numbers: RP13-1239-000. Applicants: Venice Gathering System, L.L.C. Description: Order No. 776 Compliance...

  5. Dynamical transitions in large systems of mean field-coupled Landau-Stuart oscillators: Extensive chaos and cluster states.

    Science.gov (United States)

    Ku, Wai Lim; Girvan, Michelle; Ott, Edward

    2015-12-01

    In this paper, we study dynamical systems in which a large number N of identical Landau-Stuart oscillators are globally coupled via a mean-field. Previously, it has been observed that this type of system can exhibit a variety of different dynamical behaviors. These behaviors include time periodic cluster states in which each oscillator is in one of a small number of groups for which all oscillators in each group have the same state which is different from group to group, as well as a behavior in which all oscillators have different states and the macroscopic dynamics of the mean field is chaotic. We argue that this second type of behavior is "extensive" in the sense that the chaotic attractor in the full phase space of the system has a fractal dimension that scales linearly with N and that the number of positive Lyapunov exponents of the attractor also scales linearly with N. An important focus of this paper is the transition between cluster states and extensive chaos as the system is subjected to slow adiabatic parameter change. We observe discontinuous transitions between the cluster states (which correspond to low dimensional dynamics) and the extensively chaotic states. Furthermore, examining the cluster state, as the system approaches the discontinuous transition to extensive chaos, we find that the oscillator population distribution between the clusters continually evolves so that the cluster state is always marginally stable. This behavior is used to reveal the mechanism of the discontinuous transition. We also apply the Kaplan-Yorke formula to study the fractal structure of the extensively chaotic attractors.

  6. FRAMES User Defined Body Burden Concentration File Module Documentation

    International Nuclear Information System (INIS)

    Pelton, Mitchell A.; Rutz, Frederick C.; Eslinger, Melany A.; Gelston, Gariann M.

    2001-01-01

    The Framework for Risk Analysis in Multimedia Environmental Systems (FRAMES) Body Burden Concentration File (BBF) contains time-varying, instantaneous, constituent concentrations for body burden by contaminant. This report contains the requirements for this file and will be used by software engineers and testers to ensure that the file inputs properly.

  7. 77 FR 12367 - Agency Information Collection and Reporting Activities; Electronic Filing of Bank Secrecy Act...

    Science.gov (United States)

    2012-02-29

    ... capability of electronically filing BSA reports through its system called BSA E-Filing. Effective August 2011... Accounts (FBAR) report. BSA E-Filing is a secure, web-based electronic filing system. It is a flexible... filing institutions or individuals, thereby providing a significant improvement in data quality. BSA E...

  8. Apical extrusion of bacteria when using reciprocating single-file and rotary multifile instrumentation systems.

    Science.gov (United States)

    Tinoco, J M; De-Deus, G; Tinoco, E M B; Saavedra, F; Fidel, R A S; Sassone, L M

    2014-06-01

    To evaluate ex vivo, apical bacterial extrusion associated with two reciprocating single-file systems (WaveOne and Reciproc) compared with a conventional multifile rotary system (BioRace). Forty-five human single-rooted mandibular incisors were used. Endodontic access cavities were prepared, and root canals were contaminated with an Enterococcus faecalis suspension. Following incubation at 37 °C for thirty days, the contaminated teeth were divided into three groups of 15 specimens each (G1 - Reciproc, G2 - WaveOne and G3 - BioRace). Positive and negative controls consisted of 5 infected teeth and 3 uninfected incisors that were instrumented with one of the tested NiTi systems, respectively. Bacteria extruded from the apical foramen during instrumentation were collected into vials containing 0.9% NaCl. The microbiological samples were taken from the vials and incubated in brain heart agar medium for 24 h. The resulting bacterial titre, in colony-forming units (CFU) per mL, was determined, and these data were analysed by Wilcoxon matched-pairs signed rank test and Kruskal-Wallis H-test. The level of significance was set at α = 0.05. No significant difference was found in the number of CFU between the two reciprocating systems (P = 0.41). The conventional multifile rotary system group was associated with significantly higher CFU than both of the two reciprocating groups (P = 0.01). All instrumentation systems extruded bacteria beyond the foramen. However, both reciprocating single-file systems extruded fewer bacteria apically than the conventional multifile rotary system. © 2013 International Endodontic Journal. Published by John Wiley & Sons Ltd.

  9. 75 FR 25885 - The Merit Systems Protection Board (MSPB) is Providing Notice of the Opportunity to File Amicus...

    Science.gov (United States)

    2010-05-10

    ... MERIT SYSTEMS PROTECTION BOARD The Merit Systems Protection Board (MSPB) is Providing Notice of the Opportunity to File Amicus Briefs in the Matter of Larry L. Evans v. Department of Veterans Affairs, MSPB Docket Number AT-3330-09-0953- I-1. AGENCY: Merit Systems Protection Board. ACTION: Notice...

  10. 75 FR 20007 - The Merit Systems Protection Board (MSPB) Is Providing Notice of the Opportunity To File Amicus...

    Science.gov (United States)

    2010-04-16

    ... MERIT SYSTEMS PROTECTION BOARD The Merit Systems Protection Board (MSPB) Is Providing Notice of the Opportunity To File Amicus Briefs in the Matter of Hyginus U. Aguzie v. Office of Personnel Management, MSPB Docket Number DC-0731- 09-0261-R-1 AGENCY: Merit Systems Protection Board. ACTION: Notice...

  11. 75 FR 65467 - Combined Notice of Filings No. 1

    Science.gov (United States)

    2010-10-25

    ... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission Combined Notice of Filings No. 1 October 19, 2010. Take notice that the Commission has received the following Natural Gas Pipeline Rate and...: Venice Gathering System, L.L.C. Description: Venice Gathering System, L.L.C. submits tariff filing per...

  12. Extracting Aggregation Free Energies of Mixed Clusters from Simulations of Small Systems: Application to Ionic Surfactant Micelles.

    Science.gov (United States)

    Zhang, X; Patel, L A; Beckwith, O; Schneider, R; Weeden, C J; Kindt, J T

    2017-11-14

    Micelle cluster distributions from molecular dynamics simulations of a solvent-free coarse-grained model of sodium octyl sulfate (SOS) were analyzed using an improved method to extract equilibrium association constants from small-system simulations containing one or two micelle clusters at equilibrium with free surfactants and counterions. The statistical-thermodynamic and mathematical foundations of this partition-enabled analysis of cluster histograms (PEACH) approach are presented. A dramatic reduction in computational time for analysis was achieved through a strategy similar to the selector variable method to circumvent the need for exhaustive enumeration of the possible partitions of surfactants and counterions into clusters. Using statistics from a set of small-system (up to 60 SOS molecules) simulations as input, equilibrium association constants for micelle clusters were obtained as a function of both number of surfactants and number of associated counterions through a global fitting procedure. The resulting free energies were able to accurately predict micelle size and charge distributions in a large (560 molecule) system. The evolution of micelle size and charge with SOS concentration as predicted by the PEACH-derived free energies and by a phenomenological four-parameter model fit, along with the sensitivity of these predictions to variations in cluster definitions, are analyzed and discussed.

  13. 14 CFR 406.113 - Filing documents with the Docket Management System (DMS) and sending documents to the...

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 4 2010-01-01 2010-01-01 false Filing documents with the Docket Management System (DMS) and sending documents to the administrative law judge and Assistant Chief Counsel for... Management System (DMS) and sending documents to the administrative law judge and Assistant Chief Counsel for...

  14. A Centralized Control and Dynamic Dispatch Architecture for File Integrity Analysis

    Directory of Open Access Journals (Sweden)

    Ronald DeMara

    2006-02-01

    Full Text Available The ability to monitor computer file systems for unauthorized changes is a powerful administrative tool. Ideally this task could be performed remotely under the direction of the administrator to allow on-demand checking, and use of tailorable reporting and exception policies targeted to adjustable groups of network elements. This paper introduces M-FICA, a Mobile File Integrity and Consistency Analyzer as a prototype to achieve this capability using mobile agents. The M-FICA file tampering detection approach uses MD5 message digests to identify file changes. Two agent types, Initiator and Examiner, are used to perform file integrity tasks. An Initiator travels to client systems, computes a file digest, then stores those digests in a database file located on write-once media. An Examiner agent computes a new digest to compare with the original digests in the database file. Changes in digest values indicate that the file contents have been modified. The design and evaluation results for a prototype developed in the Concordia agent framework are described.

  15. Shaping Ability of Reciproc, WaveOne GOLD, and HyFlex EDM Single-file Systems in Simulated S-shaped Canals.

    Science.gov (United States)

    Özyürek, Taha; Yılmaz, Koray; Uslu, Gülşah

    2017-05-01

    The aim of the present study was to compare the shaping ability of Reciproc (RPC; VDW, Munich, Germany), HyFlex EDM (HEDM; Coltene/Whaledent AG, Altstätten, Switzerland), and WaveOne GOLD (WOG; Dentsply Maillefer, Ballaigues, Switzerland) nickel-titanium (NiTi) files made of different NiTi alloys in S-shaped simulated canals. Sixty S-shaped canals in resin blocks were prepared to an apical size of 0.25 mm using RPC R25, WOG Primary, and HEDM OneFile (n = 20 canal/per group) systems. Composite images were made from the superimposition of pre- and postinstrumentation images. The amount of resin removed by each system was measured using a digital template and image analysis software in 22 different points. Canal aberrations were also recorded. Data were statistically analyzed using the Kruskal-Wallis and post hoc Dunn tests at the 5% level. NiTi file fracture was not observed during shaping of the simulated canals although a danger zone formation in 1 sample and a ledge in 1 sample were observed in the RPC group. There was no statistically significant difference between the WOG and HEDM groups' apical, medial, and coronal regions (P > .05). However, it was determined that the RPC group removed a statistically significantly higher amount of resin from all the canal regions when compared with the WOG and HEDM groups (P < .05). Within the limitation of the present study, it was determined that all of the tested NiTi files caused various levels of resin removal. However, WOG and HEDM NiTi files were found to cause a lower level of resin removal than RPC NiTi files. Copyright © 2017 American Association of Endodontists. All rights reserved.

  16. Evolution of highly compact binary stellar systems in globular clusters

    International Nuclear Information System (INIS)

    Krolik, J.H.; Meiksin, A.; Joss, P.C.

    1984-01-01

    We have calculated the secular evolution of a highly compact binary stellar system, composed of a collapsed object and a low-mass secondary star, in the core of a globular cluster. The binary evolves under the combined influences of (i) gravitational radiation losses from the system, (ii) the evolution of the secondary star, (iii) the resultant gradual mass transfer, if any, from the secondary to the collapsed object, and (iv) occasional encounters with passing field stars. We calculate all these effects in detail, utilizing some simplifying approximations appropriate to low-mass secondaries. The times of encounters with field stars, and the initial parameter specifying those encounters, were chosen by use of a Monte Carlo technique; the subsequent gravitational interactions were calculated utilzing a three-body integrator, and the changes in the binary orbital parmeters were thereby determined. We carried out a total of 20 such evolutionary calculations for each of two cluster core densities (1 and 3 x 10 3 stars pc -3 ). Each calculation was continued until the binary was disrupted or until 2 x 10 10 yr had elapsed

  17. VizieR Online Data Catalog: Stellar mass of brightest cluster galaxies (Bellstedt+, 2016)

    Science.gov (United States)

    Bellstedt, S.; Lidman, C.; Muzzin, A.; Franx, M.; Guatelli, S.; Hill, A. R.; Hoekstra, H.; Kurinsky, N.; Labbe, I.; Marchesini, D.; Marsan, Z. C.; Safavi-Naeini, M.; Sifon, C.; Stefanon, M.; van de Sande, J.; van Dokkum, P.; Weigel, C.

    2017-11-01

    We utilize a sample of 98 newly imaged galaxy clusters from the RELICS (REd Lens Infrared Cluster Survey) survey within this study. The data were collected during six observing runs on three instruments over a period spanning from 2013 October to 2015 March. The instruments utilized were the SofI2 camera on the New Technology Telescope at the European Southern Observatory (ESO) La Silla Observatory in Chile, WHIRC3 on the WIYN telescope at the Kitt Peak National Observatory and LIRIS4 on the William Herschel Telescope (WHT) in La Palma, Spain. (2 data files).

  18. Controlling P2P File-Sharing Networks Traffic

    OpenAIRE

    García Pineda, Miguel; HAMMOUMI, MOHAMMED; Canovas Solbes, Alejandro; Lloret, Jaime

    2011-01-01

    Since the appearance of Peer-To-Peer (P2P) file-sharing networks some time ago, many Internet users have chosen this technology to share and search programs, videos, music, documents, etc. The total number of P2P file-sharing users has been increasing and decreasing in the last decade depending on the creation or end of some well known P2P file-sharing systems. P2P file-sharing networks traffic is currently overloading some data networks and it is a major headache for netw...

  19. Cluster Bulleticity

    OpenAIRE

    Massey, Richard; Kitching, Thomas; Nagai, Daisuke

    2010-01-01

    The unique properties of dark matter are revealed during collisions between clusters of galaxies, such as the bullet cluster (1E 0657−56) and baby bullet (MACS J0025−12). These systems provide evidence for an additional, invisible mass in the separation between the distributions of their total mass, measured via gravitational lensing, and their ordinary ‘baryonic’ matter, measured via its X-ray emission. Unfortunately, the information available from these systems is limited by their rarity. C...

  20. A new Self-Adaptive disPatching System for local clusters

    Science.gov (United States)

    Kan, Bowen; Shi, Jingyan; Lei, Xiaofeng

    2015-12-01

    The scheduler is one of the most important components of a high performance cluster. This paper introduces a self-adaptive dispatching system (SAPS) based on Torque[1] and Maui[2]. It promotes cluster resource utilization and improves the overall speed of tasks. It provides some extra functions for administrators and users. First of all, in order to allow the scheduling of GPUs, a GPU scheduling module based on Torque and Maui has been developed. Second, SAPS analyses the relationship between the number of queueing jobs and the idle job slots, and then tunes the priority of users’ jobs dynamically. This means more jobs run and fewer job slots are idle. Third, integrating with the monitoring function, SAPS excludes nodes in error states as detected by the monitor, and returns them to the cluster after the nodes have recovered. In addition, SAPS provides a series of function modules including a batch monitoring management module, a comprehensive scheduling accounting module and a real-time alarm module. The aim of SAPS is to enhance the reliability and stability of Torque and Maui. Currently, SAPS has been running stably on a local cluster at IHEP (Institute of High Energy Physics, Chinese Academy of Sciences), with more than 12,000 cpu cores and 50,000 jobs running each day. Monitoring has shown that resource utilization has been improved by more than 26%, and the management work for both administrator and users has been reduced greatly.

  1. Archiving and retrieval of experimental data using SAN based centralized storage system for SST-1

    Energy Technology Data Exchange (ETDEWEB)

    Bhandarkar, Manisha, E-mail: manisha@ipr.res.in; Masand, Harish; Kumar, Aveg; Patel, Kirit; Dhongde, Jasraj; Gulati, Hitesh; Mahajan, Kirti; Chudasama, Hitesh; Pradhan, Subrata

    2016-11-15

    Highlights: • SAN (Storage Area Network) based centralized data storage system of SST-1 has envisaged to address the need of centrally availability of SST-1 storage system to archive/retrieve experimental data for the authenticated users for 24 × 7. • The SAN based data storage system has been designed/configured with 3-tiered architecture and GFS cluster file system with multipath support. • The adopted SAN based data storage for SST-1 is a modular, robust, and allows future expandability. • Important considerations has been taken like, Handling of varied Data writing speed from different subsystems to central storage, Simultaneous read access of the bulk experimental and as well as essential diagnostic data, The life expectancy of data, How often data will be retrieved and how fast it will be needed, How much historical data should be maintained at storage. - Abstract: SAN (Storage Area Network, a high-speed, block level storage device) based centralized data storage system of SST-1 (Steady State superconducting Tokamak) has envisaged to address the need of availability of SST-1 operation & experimental data centrally for archival as well as retrieval [2]. Considering the initial data volume requirement, ∼10 TB (Terabytes) capacity of SAN based data storage system has configured/installed with optical fiber backbone with compatibility considerations of existing Ethernet network of SST-1. The SAN based data storage system has been designed/configured with 3-tiered architecture and GFS (Global File System) cluster file system with multipath support. Tier-1 is of ∼3 TB (frequent access and low data storage capacity) comprises of Fiber channel (FC) based hard disks for optimum throughput. Tier-2 is of ∼6 TB (less frequent access and high data storage capacity) comprises of SATA based hard disks. Tier-3 will be planned later to store offline historical data. In the SAN configuration two tightly coupled storage servers (with cluster configuration) are

  2. Archiving and retrieval of experimental data using SAN based centralized storage system for SST-1

    International Nuclear Information System (INIS)

    Bhandarkar, Manisha; Masand, Harish; Kumar, Aveg; Patel, Kirit; Dhongde, Jasraj; Gulati, Hitesh; Mahajan, Kirti; Chudasama, Hitesh; Pradhan, Subrata

    2016-01-01

    Highlights: • SAN (Storage Area Network) based centralized data storage system of SST-1 has envisaged to address the need of centrally availability of SST-1 storage system to archive/retrieve experimental data for the authenticated users for 24 × 7. • The SAN based data storage system has been designed/configured with 3-tiered architecture and GFS cluster file system with multipath support. • The adopted SAN based data storage for SST-1 is a modular, robust, and allows future expandability. • Important considerations has been taken like, Handling of varied Data writing speed from different subsystems to central storage, Simultaneous read access of the bulk experimental and as well as essential diagnostic data, The life expectancy of data, How often data will be retrieved and how fast it will be needed, How much historical data should be maintained at storage. - Abstract: SAN (Storage Area Network, a high-speed, block level storage device) based centralized data storage system of SST-1 (Steady State superconducting Tokamak) has envisaged to address the need of availability of SST-1 operation & experimental data centrally for archival as well as retrieval [2]. Considering the initial data volume requirement, ∼10 TB (Terabytes) capacity of SAN based data storage system has configured/installed with optical fiber backbone with compatibility considerations of existing Ethernet network of SST-1. The SAN based data storage system has been designed/configured with 3-tiered architecture and GFS (Global File System) cluster file system with multipath support. Tier-1 is of ∼3 TB (frequent access and low data storage capacity) comprises of Fiber channel (FC) based hard disks for optimum throughput. Tier-2 is of ∼6 TB (less frequent access and high data storage capacity) comprises of SATA based hard disks. Tier-3 will be planned later to store offline historical data. In the SAN configuration two tightly coupled storage servers (with cluster configuration) are

  3. Patient Treatment File (PTF)

    Data.gov (United States)

    Department of Veterans Affairs — This database is part of the National Medical Information System (NMIS). The Patient Treatment File (PTF) contains a record for each inpatient care episode provided...

  4. Search for Formation Criteria for Globular Cluster Systems

    Science.gov (United States)

    Nuritdinov, S. N.; Mirtadjieva, K. T.; Tadjibaev, I. U.

    2005-01-01

    Star cluster formation is a major mode of star formation in the extreme conditions of interacting galaxies and violent starbursts. By studying ages and metallicities of young metal-enhanced star clusters in mergers / merger remnants we can learn about the violent star formation history of these galaxies and eventually about galaxy formation and evolution. We will present a new set of evolutionary synthesis models of our GALEV code specially developed to account for the gaseous emission of presently forming star clusters and an advanced tool to compare large model grids with multi-color broad-band observations becoming presently available in large amounts. Such observations are an ecomonic way to determine the parameters of young star clusters as will be shown in the presentation. First results of newly-born clusters in mergers and starburst galaxies are presented and compared to the well-studied old globulars and interpreted in the framework of galaxy formation / evolution.

  5. Evaluated nuclear data file of Th-232

    International Nuclear Information System (INIS)

    Meadows, J.; Poenitz, W.; Smith, A.; Smith, D.; Whalen, J.; Howerton, R.

    1977-09-01

    An evaluated nuclear data file for thorium is described. The file extends over the energy range 0.049 (i.e., the inelastic-scattering threshold) to 20.0 MeV and is formulated within the framework of the ENDF system. The input data base, the evaluation procedures and judgments, and ancillary experiments carried out in conjunction with the evaluation are outlined. The file includes: neutron total cross sections, neutron scattering processes, neutron radiative capture cross sections, fission cross sections, (n;2n) and (n;3n) processes, fission properties (e.g., nu-bar and delayed neutron emission) and photon production processes. Regions of uncertainty are pointed out particularly where new measured results would be of value. The file is extended to thermal energies using previously reported resonance evaluations thereby providing a complete file for neutronic calculations. Integral data tests indicated that the file was suitable for neutronic calculations in the MeV range

  6. AIDEN: A Density Conscious Artificial Immune System for Automatic Discovery of Arbitrary Shape Clusters in Spatial Patterns

    Directory of Open Access Journals (Sweden)

    Vishwambhar Pathak

    2012-11-01

    Full Text Available Recent efforts in modeling of dynamics of the natural immune cells leading to artificial immune systems (AIS have ignited contemporary research interest in finding out its analogies to real world problems. The AIS models have been vastly exploited to develop dependable robust
    solutions to clustering. Most of the traditional clustering methods bear limitations in their capability to detect clusters of arbitrary shapes in a fully unsupervised manner. In this paper the recognition and communication dynamics of T Cell Receptors, the recognizing elements in innate immune
    system, has been modeled with a kernel density estimation method. The model has been shown to successfully discover non spherical clusters in spatial patterns. Modeling the cohesion of the antibodies and pathogens with ‘local influence’ measure inducts comprehensive extension of the
    antibody representation ball (ARB, which in turn corresponds to controlled expansion of clusters and prevents overfitting.

  7. London SPAN version 4 parameter file format

    International Nuclear Information System (INIS)

    2004-06-01

    Powernext SA is a Multilateral Trading Facility in charge of managing the French power exchange through an optional and anonymous organised trading system. Powernext SA collaborates with the clearing organization LCH.Clearnet SA to secure and facilitate the transactions. The French Standard Portfolio Analysis of Risk (SPAN) is a system used by LCH.Clearnet to calculate the initial margins from and for its clearing members. SPAN is a computerized system which calculates the impact of several possible variations of rates and volatility on by-product portfolios. The initial margin call is equal to the maximum probable loss calculated by the system. This document contains details of the format of the London SPAN version 4 parameter file. This file contains all the parameters and risk arrays required to calculate SPAN margins. London SPAN Version 4 is an upgrade from Version 3, which is also known as LME SPAN. This document contains the full revised file specification, highlighting the changes from Version 3 to Version 4

  8. The high performance cluster computing system for BES offline data analysis

    International Nuclear Information System (INIS)

    Sun Yongzhao; Xu Dong; Zhang Shaoqiang; Yang Ting

    2004-01-01

    A high performance cluster computing system (EPCfarm) is introduced, which used for BES offline data analysis. The setup and the characteristics of the hardware and software of EPCfarm are described. The PBS, a queue management package, and the performance of EPCfarm is presented also. (authors)

  9. JENDL special purpose file

    International Nuclear Information System (INIS)

    Nakagawa, Tsuneo

    1995-01-01

    In JENDL-3,2, the data on all the reactions having significant cross section over the neutron energy from 0.01 meV to 20 MeV are given for 340 nuclides. The object range of application extends widely, such as the neutron engineering, shield and others of fast reactors, thermal neutron reactors and nuclear fusion reactors. This is a general purpose data file. On the contrary to this, the file in which only the data required for a specific application field are collected is called special purpose file. The file for dosimetry is a typical special purpose file. The Nuclear Data Center, Japan Atomic Energy Research Institute, is making ten kinds of JENDL special purpose files. The files, of which the working groups of Sigma Committee are in charge, are listed. As to the format of the files, ENDF format is used similarly to JENDL-3,2. Dosimetry file, activation cross section file, (α, n) reaction data file, fusion file, actinoid file, high energy data file, photonuclear data file, PKA/KERMA file, gas production cross section file and decay data file are described on their contents, the course of development and their verification. Dosimetry file and gas production cross section file have been completed already. As for the others, the expected time of completion is shown. When these files are completed, they are opened to the public. (K.I.)

  10. Explicitly-correlated ring-coupled-cluster-doubles theory: Including exchange for computations on closed-shell systems

    Energy Technology Data Exchange (ETDEWEB)

    Hehn, Anna-Sophia; Holzer, Christof; Klopper, Wim, E-mail: klopper@kit.edu

    2016-11-10

    Highlights: • Ring-coupled-cluster-doubles approach now implemented with exchange terms. • Ring-coupled-cluster-doubles approach now implemented with F12 functions. • Szabo–Ostlund scheme (SO2) implemented for use in SAPT. • Fast convergence to the limit of a complete basis. • Implementation in the TURBOMOLE program system. - Abstract: Random-phase-approximation (RPA) methods have proven to be powerful tools in electronic-structure theory, being non-empirical, computationally efficient and broadly applicable to a variety of molecular systems including small-gap systems, transition-metal compounds and dispersion-dominated complexes. Applications are however hindered due to the slow basis-set convergence of the electron-correlation energy with the one-electron basis. As a remedy, we present approximate explicitly-correlated RPA approaches based on the ring-coupled-cluster-doubles formulation including exchange contributions. Test calculations demonstrate that the basis-set convergence of correlation energies is drastically accelerated through the explicitly-correlated approach, reaching 99% of the basis-set limit with triple-zeta basis sets. When implemented in close analogy to early work by Szabo and Ostlund [36], the new explicitly-correlated ring-coupled-cluster-doubles approach including exchange has the perspective to become a valuable tool in the framework of symmetry-adapted perturbation theory (SAPT) for the computation of dispersion energies of molecular complexes of weakly interacting closed-shell systems.

  11. Grid collector: An event catalog with automated file management

    International Nuclear Information System (INIS)

    Wu, Kesheng; Zhang, Wei-Ming; Sim, Alexander; Gu, Junmin; Shoshani, Arie

    2003-01-01

    High Energy Nuclear Physics (HENP) experiments such as STAR at BNL and ATLAS at CERN produce large amounts of data that are stored as files on mass storage systems in computer centers. In these files, the basic unit of data is an event. Analysis is typically performed on a selected set of events. The files containing these events have to be located, copied from mass storage systems to disks before analysis, and removed when no longer needed. These file management tasks are tedious and time consuming. Typically, all events contained in the files are read into memory before a selection is made. Since the time to read the events dominate the overall execution time, reading the unwanted event needlessly increases the analysis time. The Grid Collector is a set of software modules that works together to address these two issues. It automates the file management tasks and provides ''direct'' access to the selected events for analyses. It is currently integrated with the STAR analysis framework. The users can select events based on tags, such as, ''production date between March 10 and 20, and the number of charged tracks > 100.'' The Grid Collector locates the files containing relevant events, transfers the files across the Grid if necessary, and delivers the events to the analysis code through the familiar iterators. There has been some research efforts to address the file management issues, the Grid Collector is unique in that it addresses the event access issue together with the file management issues. This makes it more useful to a large variety of users

  12. Grid collector an event catalog with automated file management

    CERN Document Server

    Ke Sheng Wu; Sim, A; Jun Min Gu; Shoshani, A

    2004-01-01

    High Energy Nuclear Physics (HENP) experiments such as STAR at BNL and ATLAS at CERN produce large amounts of data that are stored as files on mass storage systems in computer centers. In these files, the basic unit of data is an event. Analysis is typically performed on a selected set of events. The files containing these events have to be located, copied from mass storage systems to disks before analysis, and removed when no longer needed. These file management tasks are tedious and time consuming. Typically, all events contained in the files are read into memory before a selection is made. Since the time to read the events dominate the overall execution time, reading the unwanted event needlessly increases the analysis time. The Grid Collector is a set of software modules that works together to address these two issues. It automates the file management tasks and provides "direct" access to the selected events for analyses. It is currently integrated with the STAR analysis framework. The users can select ev...

  13. EVIDENCE FOR AN ACCRETION ORIGIN FOR THE OUTER HALO GLOBULAR CLUSTER SYSTEM OF M31

    International Nuclear Information System (INIS)

    Mackey, A. D.; Huxor, A. P.; Ferguson, A. M. N.; Irwin, M. J.; Chapman, S. C.; Tanvir, N. R.; McConnachie, A. W.; Ibata, R. A.; Lewis, G. F.

    2010-01-01

    We use a sample of newly discovered globular clusters from the Pan-Andromeda Archaeological Survey (PAndAS) in combination with previously cataloged objects to map the spatial distribution of globular clusters in the M31 halo. At projected radii beyond ∼30 kpc, where large coherent stellar streams are readily distinguished in the field, there is a striking correlation between these features and the positions of the globular clusters. Adopting a simple Monte Carlo approach, we test the significance of this association by computing the probability that it could be due to the chance alignment of globular clusters smoothly distributed in the M31 halo. We find that the likelihood of this possibility is low, below 1%, and conclude that the observed spatial coincidence between globular clusters and multiple tidal debris streams in the outer halo of M31 reflects a genuine physical association. Our results imply that the majority of the remote globular cluster system of M31 has been assembled as a consequence of the accretion of cluster-bearing satellite galaxies. This constitutes the most direct evidence to date that the outer halo globular cluster populations in some galaxies are largely accreted.

  14. Large-Scale Multi-Dimensional Document Clustering on GPU Clusters

    Energy Technology Data Exchange (ETDEWEB)

    Cui, Xiaohui [ORNL; Mueller, Frank [North Carolina State University; Zhang, Yongpeng [ORNL; Potok, Thomas E [ORNL

    2010-01-01

    Document clustering plays an important role in data mining systems. Recently, a flocking-based document clustering algorithm has been proposed to solve the problem through simulation resembling the flocking behavior of birds in nature. This method is superior to other clustering algorithms, including k-means, in the sense that the outcome is not sensitive to the initial state. One limitation of this approach is that the algorithmic complexity is inherently quadratic in the number of documents. As a result, execution time becomes a bottleneck with large number of documents. In this paper, we assess the benefits of exploiting the computational power of Beowulf-like clusters equipped with contemporary Graphics Processing Units (GPUs) as a means to significantly reduce the runtime of flocking-based document clustering. Our framework scales up to over one million documents processed simultaneously in a sixteennode GPU cluster. Results are also compared to a four-node cluster with higher-end GPUs. On these clusters, we observe 30X-50X speedups, which demonstrates the potential of GPU clusters to efficiently solve massive data mining problems. Such speedups combined with the scalability potential and accelerator-based parallelization are unique in the domain of document-based data mining, to the best of our knowledge.

  15. Effect of mesoscopic fluctuations on equation of state in cluster-forming systems

    Directory of Open Access Journals (Sweden)

    A. Ciach

    2012-06-01

    Full Text Available Equation of state for systems with particles self-assembling into aggregates is derived within a mesoscopic theory combining density functional and field-theoretic approaches. We focus on the effect of mesoscopic fluctuations in the disordered phase. The pressure - volume fraction isotherms are calculated explicitly for two forms of the short-range attraction long-range repulsion potential. Mesoscopic fluctuations lead to an increased pressure in each case, except for very small volume fractions. When large clusters are formed, the mechanical instability of the system is present at much higher temperature than found in mean-field approximation. In this case phase separation competes with the formation of periodic phases (colloidal crystals. In the case of small clusters, no mechanical instability associated with separation into dilute and dense phases appears.

  16. Classification Of Cluster Area Forsatellite Image

    Directory of Open Access Journals (Sweden)

    Thwe Zin Phyo

    2015-06-01

    Full Text Available Abstract This paper describes area classification for Landsat7 satellite image. The main purpose of this system is to classify the area of each cluster contained in a satellite image. To classify this image firstly need to clusterthe satellite image into different land cover types. Clustering is an unsupervised learning method that aimsto classify an image into homogeneous regions. This system is implemented based on color features with K-means clustering unsupervised algorithm. This method does not need to train image before clustering.The clusters of satellite image are grouped into a set of three clusters for Landsat7 satellite image. For this work the combined band 432 from Landsat7 satellite is used as an input. Satellite imageMandalay area in 2001 is chosen to test the segmentation method. After clustering a specific range for three clustered images must be defined in order to obtain greenland water and urbanbalance.This system is implemented by using MATLAB programming language.

  17. Global Navigation Satellite System (GNSS) Final Clock Product (5 minute resolution, daily files, generated weekly) from NASA CDDIS

    Data.gov (United States)

    National Aeronautics and Space Administration — This derived product set consists of Global Navigation Satellite System Final Satellite and Receiver Clock Product (5-minute granularity, daily files, generated...

  18. Linked cluster expansions for open quantum systems on a lattice

    Science.gov (United States)

    Biella, Alberto; Jin, Jiasen; Viyuela, Oscar; Ciuti, Cristiano; Fazio, Rosario; Rossini, Davide

    2018-01-01

    We propose a generalization of the linked-cluster expansions to study driven-dissipative quantum lattice models, directly accessing the thermodynamic limit of the system. Our method leads to the evaluation of the desired extensive property onto small connected clusters of a given size and topology. We first test this approach on the isotropic spin-1/2 Hamiltonian in two dimensions, where each spin is coupled to an independent environment that induces incoherent spin flips. Then we apply it to the study of an anisotropic model displaying a dissipative phase transition from a magnetically ordered to a disordered phase. By means of a Padé analysis on the series expansions for the average magnetization, we provide a viable route to locate the phase transition and to extrapolate the critical exponent for the magnetic susceptibility.

  19. MULTIAGENT IMITATION MODEL OF A REGIONAL CONSTRUCTION CLUSTER AS A HETERARCHICAL SYSTEM

    Directory of Open Access Journals (Sweden)

    Anufriev Dmitriy Petrovich

    2018-01-01

    Full Text Available Subject: a regional construction cluster, which is viewed as a complex system territorially localized within the region, consisting of interconnected and complementary enterprises of construction and related industries that are united with local institutions, authorities and cooperating enterprises by heterarchic relations. Research objectives: development of multi-agent simulation model that allows us to examine the business-processes in the regional construction cluster as a complex heterarchical system. Materials and methods: we formulate the mathematical problem for description of processes in a heterarchic system as in a special multi-agent queueing network. Conclusions: the article substantiates application of the decentralized approach which is based on the use of agent methodology. Several types of agents that model elementary organizational structures have been developed. We describe the functional core of the multi-agent simulation model characterizing the heterarchic organizational model. Using the Fishman-Kivia criterion, the adequacy of the logical functioning of the developed model was established.

  20. Spatial clustering of pixels of a multispectral image

    Science.gov (United States)

    Conger, James Lynn

    2014-08-19

    A method and system for clustering the pixels of a multispectral image is provided. A clustering system computes a maximum spectral similarity score for each pixel that indicates the similarity between that pixel and the most similar neighboring. To determine the maximum similarity score for a pixel, the clustering system generates a similarity score between that pixel and each of its neighboring pixels and then selects the similarity score that represents the highest similarity as the maximum similarity score. The clustering system may apply a filtering criterion based on the maximum similarity score so that pixels with similarity scores below a minimum threshold are not clustered. The clustering system changes the current pixel values of the pixels in a cluster based on an averaging of the original pixel values of the pixels in the cluster.