WorldWideScience

Sample records for component based heuristic

  1. Prediction-based dynamic load-sharing heuristics

    Science.gov (United States)

    Goswami, Kumar K.; Devarakonda, Murthy; Iyer, Ravishankar K.

    1993-01-01

    The authors present dynamic load-sharing heuristics that use predicted resource requirements of processes to manage workloads in a distributed system. A previously developed statistical pattern-recognition method is employed for resource prediction. While nonprediction-based heuristics depend on a rapidly changing system status, the new heuristics depend on slowly changing program resource usage patterns. Furthermore, prediction-based heuristics can be more effective since they use future requirements rather than just the current system state. Four prediction-based heuristics, two centralized and two distributed, are presented. Using trace driven simulations, they are compared against random scheduling and two effective nonprediction based heuristics. Results show that the prediction-based centralized heuristics achieve up to 30 percent better response times than the nonprediction centralized heuristic, and that the prediction-based distributed heuristics achieve up to 50 percent improvements relative to their nonprediction counterpart.

  2. Advances in heuristically based generalized perturbation theory

    International Nuclear Information System (INIS)

    Gandini, A.

    1994-01-01

    A distinctive feature of heuristically based generalized perturbation theory methodology consists in the systematic use of importance conservation concepts. As well known, this use leads to fundamental reciprocity relationship. Instead, the alternative variational and differential one approaches make a consistent use of the properties and adjoint functions. The equivalence between the importance and the adjoint functions have been demonstrated in important cases. There are some instances, however, in which the commonly known operator governing the adjoint function are not adequate. In this paper ways proposed to generalize this rules, as adopted with the heuristic generalized perturbation theory methodology, are illustrated. When applied to the neutron/nuclide field characterizing the core evolution in a power reactor system, in which also an intensive control variable (ρ) is defined, these rules leas to an orthogonality relationship connected to this same control variable. A set of ρ-mode eigenfunctions may be correspondingly defined and an extended concept of reactivity (generalizing that commonly associated with the multiplication factor) proposed as more directly indicative of the controllability of a critical reactor system. (author). 25 refs

  3. Case-Based Reasoning as a Heuristic Selector in a Hyper-Heuristic for Course Timetabling Problems

    OpenAIRE

    Petrovic, Sanja; Qu, Rong

    2002-01-01

    This paper studies Knowledge Discovery (KD) using Tabu Search and Hill Climbing within Case-Based Reasoning (CBR) as a hyper-heuristic method for course timetabling problems. The aim of the hyper-heuristic is to choose the best heuristic(s) for given timetabling problems according to the knowledge stored in the case base. KD in CBR is a 2-stage iterative process on both case representation and the case base. Experimental results are analysed and related research issues for future work are dis...

  4. Knowledge discovery in hyper-heuristic using case-based reasoning on course timetabling

    OpenAIRE

    Burke, Edmund; MacCarthy, Bart L.; Petrovic, Sanja; Qu, Rong

    2002-01-01

    This paper presents a new hyper-heuristic method using Case-Based Reasoning (CBR) for solving course timetabling problems. The term Hyper-heuristics has recently been employed to refer to 'heuristics that choose heuristics' rather than heuristics that operate directly on given problems. One of the overriding motivations of hyper-heuristic methods is the attempt to develop techniques that can operate with greater generality than is currently possible. The basic idea behind this is that we main...

  5. The afforestation problem: a heuristic method based on simulated annealing

    DEFF Research Database (Denmark)

    Vidal, Rene Victor Valqui

    1992-01-01

    This paper presents the afforestation problem, that is the location and design of new forest compartments to be planted in a given area. This optimization problem is solved by a two-step heuristic method based on simulated annealing. Tests and experiences with this method are also presented....

  6. Seismic active control by a heuristic-based algorithm

    International Nuclear Information System (INIS)

    Tang, Yu.

    1996-01-01

    A heuristic-based algorithm for seismic active control is generalized to permit consideration of the effects of control-structure interaction and actuator dynamics. Control force is computed at onetime step ahead before being applied to the structure. Therefore, the proposed control algorithm is free from the problem of time delay. A numerical example is presented to show the effectiveness of the proposed control algorithm. Also, two indices are introduced in the paper to assess the effectiveness and efficiency of control laws

  7. Automated Detection of Heuristics and Biases among Pathologists in a Computer-Based System

    Science.gov (United States)

    Crowley, Rebecca S.; Legowski, Elizabeth; Medvedeva, Olga; Reitmeyer, Kayse; Tseytlin, Eugene; Castine, Melissa; Jukic, Drazen; Mello-Thoms, Claudia

    2013-01-01

    The purpose of this study is threefold: (1) to develop an automated, computer-based method to detect heuristics and biases as pathologists examine virtual slide cases, (2) to measure the frequency and distribution of heuristics and errors across three levels of training, and (3) to examine relationships of heuristics to biases, and biases to…

  8. Application of the heuristically based GPT theory to termohydraulic problems

    International Nuclear Information System (INIS)

    Alvim, A.C.M.

    1988-01-01

    Application of heuristically based generalized perturbation theory (GPT) to the thermohydraulic (generally nonlinear) field is here illustrated. After a short description of the general methodology, the (linear) equations governing the importance function relevant to a generic multichannel problem are derived, within the physical model adopted in the COBRA IV-I Code. These equations are put in a form which should benefit of the calculational scheme of the original COBRA Code in the sense that only minor changes of it (mostly implying physical constants and source terms redefinitions) should be necessary for their solutions. (author) [pt

  9. Solving Inventory Routing Problems Using Location Based Heuristics

    Directory of Open Access Journals (Sweden)

    Paweł Hanczar

    2014-01-01

    Full Text Available Inventory routing problems (IRPs occur where vendor managed inventory replenishment strategies are implemented in supply chains. These problems are characterized by the presence of both transportation and inventory considerations, either as parameters or constraints. The research presented in this paper aims at extending IRP formulation developed on the basis of location based heuristics proposed by Bramel and Simchi-Levi and continued by Hanczar. In the first phase of proposed algorithms, mixed integer programming is used to determine the partitioning of customers as well as dates and quantities of deliveries. Then, using 2-opt algorithm for solving the traveling sales-person problem the optimal routes for each partition are determined. In the main part of research the classical formulation is extended by additional constraints (visit spacing, vehicle filling rate, driver (vehicle consistency, and heterogeneous fleet of vehicles as well as the additional criteria are discussed. Then the impact of using each of proposed extensions for solution possibilities is evaluated. The results of computational tests are presented and discussed. Obtained results allow to conclude that the location based heuristics should be considered when solving real life instances of IRP. (original abstract

  10. Design and usability of heuristic-based deliberation tools for women facing amniocentesis.

    Science.gov (United States)

    Durand, Marie-Anne; Wegwarth, Odette; Boivin, Jacky; Elwyn, Glyn

    2012-03-01

    Evidence suggests that in decision contexts characterized by uncertainty and time constraints (e.g. health-care decisions), fast and frugal decision-making strategies (heuristics) may perform better than complex rules of reasoning. To examine whether it is possible to design deliberation components in decision support interventions using simple models (fast and frugal heuristics). The 'Take The Best' heuristic (i.e. selection of a 'most important reason') and 'The Tallying' integration algorithm (i.e. unitary weighing of pros and cons) were used to develop two deliberation components embedded in a Web-based decision support intervention for women facing amniocentesis testing. Ten researchers (recruited from 15), nine health-care providers (recruited from 28) and ten pregnant women (recruited from 14) who had recently been offered amniocentesis testing appraised evolving versions of 'your most important reason' (Take The Best) and 'weighing it up' (Tallying). Most researchers found the tools useful in facilitating decision making although emphasized the need for simple instructions and clear layouts. Health-care providers however expressed concerns regarding the usability and clarity of the tools. By contrast, 7 out of 10 pregnant women found the tools useful in weighing up the pros and cons of each option, helpful in structuring and clarifying their thoughts and visualizing their decision efforts. Several pregnant women felt that 'weighing it up' and 'your most important reason' were not appropriate when facing such a difficult and emotional decision. Theoretical approaches based on fast and frugal heuristics can be used to develop deliberation tools that provide helpful support to patients facing real-world decisions about amniocentesis. © 2011 Blackwell Publishing Ltd.

  11. A Modularity Degree Based Heuristic Community Detection Algorithm

    Directory of Open Access Journals (Sweden)

    Dongming Chen

    2014-01-01

    Full Text Available A community in a complex network can be seen as a subgroup of nodes that are densely connected. Discovery of community structures is a basic problem of research and can be used in various areas, such as biology, computer science, and sociology. Existing community detection methods usually try to expand or collapse the nodes partitions in order to optimize a given quality function. These optimization function based methods share the same drawback of inefficiency. Here we propose a heuristic algorithm (MDBH algorithm based on network structure which employs modularity degree as a measure function. Experiments on both synthetic benchmarks and real-world networks show that our algorithm gives competitive accuracy with previous modularity optimization methods, even though it has less computational complexity. Furthermore, due to the use of modularity degree, our algorithm naturally improves the resolution limit in community detection.

  12. Memory-Based Simple Heuristics as Attribute Substitution: Competitive Tests of Binary Choice Inference Models

    Science.gov (United States)

    Honda, Hidehito; Matsuka, Toshihiko; Ueda, Kazuhiro

    2017-01-01

    Some researchers on binary choice inference have argued that people make inferences based on simple heuristics, such as recognition, fluency, or familiarity. Others have argued that people make inferences based on available knowledge. To examine the boundary between heuristic and knowledge usage, we examine binary choice inference processes in…

  13. Combination of graph heuristics in producing initial solution of curriculum based course timetabling problem

    Science.gov (United States)

    Wahid, Juliana; Hussin, Naimah Mohd

    2016-08-01

    The construction of population of initial solution is a crucial task in population-based metaheuristic approach for solving curriculum-based university course timetabling problem because it can affect the convergence speed and also the quality of the final solution. This paper presents an exploration on combination of graph heuristics in construction approach in curriculum based course timetabling problem to produce a population of initial solutions. The graph heuristics were set as single and combination of two heuristics. In addition, several ways of assigning courses into room and timeslot are implemented. All settings of heuristics are then tested on the same curriculum based course timetabling problem instances and are compared with each other in terms of number of population produced. The result shows that combination of saturation degree followed by largest degree heuristic produce the highest number of population of initial solutions. The results from this study can be used in the improvement phase of algorithm that uses population of initial solutions.

  14. Using Heuristic Task Analysis to Create Web-Based Instructional Design Theory

    Science.gov (United States)

    Fiester, Herbert R.

    2010-01-01

    The first purpose of this study was to identify procedural and heuristic knowledge used when creating web-based instruction. The second purpose of this study was to develop suggestions for improving the Heuristic Task Analysis process, a technique for eliciting, analyzing, and representing expertise in cognitively complex tasks. Three expert…

  15. Memory-Based Decision-Making with Heuristics: Evidence for a Controlled Activation of Memory Representations

    Science.gov (United States)

    Khader, Patrick H.; Pachur, Thorsten; Meier, Stefanie; Bien, Siegfried; Jost, Kerstin; Rosler, Frank

    2011-01-01

    Many of our daily decisions are memory based, that is, the attribute information about the decision alternatives has to be recalled. Behavioral studies suggest that for such decisions we often use simple strategies (heuristics) that rely on controlled and limited information search. It is assumed that these heuristics simplify decision-making by…

  16. Automated detection of heuristics and biases among pathologists in a computer-based system.

    Science.gov (United States)

    Crowley, Rebecca S; Legowski, Elizabeth; Medvedeva, Olga; Reitmeyer, Kayse; Tseytlin, Eugene; Castine, Melissa; Jukic, Drazen; Mello-Thoms, Claudia

    2013-08-01

    The purpose of this study is threefold: (1) to develop an automated, computer-based method to detect heuristics and biases as pathologists examine virtual slide cases, (2) to measure the frequency and distribution of heuristics and errors across three levels of training, and (3) to examine relationships of heuristics to biases, and biases to diagnostic errors. The authors conducted the study using a computer-based system to view and diagnose virtual slide cases. The software recorded participant responses throughout the diagnostic process, and automatically classified participant actions based on definitions of eight common heuristics and/or biases. The authors measured frequency of heuristic use and bias across three levels of training. Biases studied were detected at varying frequencies, with availability and search satisficing observed most frequently. There were few significant differences by level of training. For representativeness and anchoring, the heuristic was used appropriately as often or more often than it was used in biased judgment. Approximately half of the diagnostic errors were associated with one or more biases. We conclude that heuristic use and biases were observed among physicians at all levels of training using the virtual slide system, although their frequencies varied. The system can be employed to detect heuristic use and to test methods for decreasing diagnostic errors resulting from cognitive biases.

  17. Evolutionary algorithm based heuristic scheme for nonlinear heat transfer equations.

    Science.gov (United States)

    Ullah, Azmat; Malik, Suheel Abdullah; Alimgeer, Khurram Saleem

    2018-01-01

    In this paper, a hybrid heuristic scheme based on two different basis functions i.e. Log Sigmoid and Bernstein Polynomial with unknown parameters is used for solving the nonlinear heat transfer equations efficiently. The proposed technique transforms the given nonlinear ordinary differential equation into an equivalent global error minimization problem. Trial solution for the given nonlinear differential equation is formulated using a fitness function with unknown parameters. The proposed hybrid scheme of Genetic Algorithm (GA) with Interior Point Algorithm (IPA) is opted to solve the minimization problem and to achieve the optimal values of unknown parameters. The effectiveness of the proposed scheme is validated by solving nonlinear heat transfer equations. The results obtained by the proposed scheme are compared and found in sharp agreement with both the exact solution and solution obtained by Haar Wavelet-Quasilinearization technique which witnesses the effectiveness and viability of the suggested scheme. Moreover, the statistical analysis is also conducted for investigating the stability and reliability of the presented scheme.

  18. Evolutionary algorithm based heuristic scheme for nonlinear heat transfer equations.

    Directory of Open Access Journals (Sweden)

    Azmat Ullah

    Full Text Available In this paper, a hybrid heuristic scheme based on two different basis functions i.e. Log Sigmoid and Bernstein Polynomial with unknown parameters is used for solving the nonlinear heat transfer equations efficiently. The proposed technique transforms the given nonlinear ordinary differential equation into an equivalent global error minimization problem. Trial solution for the given nonlinear differential equation is formulated using a fitness function with unknown parameters. The proposed hybrid scheme of Genetic Algorithm (GA with Interior Point Algorithm (IPA is opted to solve the minimization problem and to achieve the optimal values of unknown parameters. The effectiveness of the proposed scheme is validated by solving nonlinear heat transfer equations. The results obtained by the proposed scheme are compared and found in sharp agreement with both the exact solution and solution obtained by Haar Wavelet-Quasilinearization technique which witnesses the effectiveness and viability of the suggested scheme. Moreover, the statistical analysis is also conducted for investigating the stability and reliability of the presented scheme.

  19. Probabilistic dual heuristic programming-based adaptive critic

    Science.gov (United States)

    Herzallah, Randa

    2010-02-01

    Adaptive critic (AC) methods have common roots as generalisations of dynamic programming for neural reinforcement learning approaches. Since they approximate the dynamic programming solutions, they are potentially suitable for learning in noisy, non-linear and non-stationary environments. In this study, a novel probabilistic dual heuristic programming (DHP)-based AC controller is proposed. Distinct to current approaches, the proposed probabilistic (DHP) AC method takes uncertainties of forward model and inverse controller into consideration. Therefore, it is suitable for deterministic and stochastic control problems characterised by functional uncertainty. Theoretical development of the proposed method is validated by analytically evaluating the correct value of the cost function which satisfies the Bellman equation in a linear quadratic control problem. The target value of the probabilistic critic network is then calculated and shown to be equal to the analytically derived correct value. Full derivation of the Riccati solution for this non-standard stochastic linear quadratic control problem is also provided. Moreover, the performance of the proposed probabilistic controller is demonstrated on linear and non-linear control examples.

  20. Petri nets SM-cover-based on heuristic coloring algorithm

    Science.gov (United States)

    Tkacz, Jacek; Doligalski, Michał

    2015-09-01

    In the paper, coloring heuristic algorithm of interpreted Petri nets is presented. Coloring is used to determine the State Machines (SM) subnets. The present algorithm reduces the Petri net in order to reduce the computational complexity and finds one of its possible State Machines cover. The proposed algorithm uses elements of interpretation of Petri nets. The obtained result may not be the best, but it is sufficient for use in rapid prototyping of logic controllers. Found SM-cover will be also used in the development of algorithms for decomposition, and modular synthesis and implementation of parallel logic controllers. Correctness developed heuristic algorithm was verified using Gentzen formal reasoning system.

  1. Sequence-based heuristics for faster annotation of non-coding RNA families.

    Science.gov (United States)

    Weinberg, Zasha; Ruzzo, Walter L

    2006-01-01

    Non-coding RNAs (ncRNAs) are functional RNA molecules that do not code for proteins. Covariance Models (CMs) are a useful statistical tool to find new members of an ncRNA gene family in a large genome database, using both sequence and, importantly, RNA secondary structure information. Unfortunately, CM searches are extremely slow. Previously, we created rigorous filters, which provably sacrifice none of a CM's accuracy, while making searches significantly faster for virtually all ncRNA families. However, these rigorous filters make searches slower than heuristics could be. In this paper we introduce profile HMM-based heuristic filters. We show that their accuracy is usually superior to heuristics based on BLAST. Moreover, we compared our heuristics with those used in tRNAscan-SE, whose heuristics incorporate a significant amount of work specific to tRNAs, where our heuristics are generic to any ncRNA. Performance was roughly comparable, so we expect that our heuristics provide a high-quality solution that--unlike family-specific solutions--can scale to hundreds of ncRNA families. The source code is available under GNU Public License at the supplementary web site.

  2. A priority-based heuristic algorithm (PBHA for optimizing integrated process planning and scheduling problem

    Directory of Open Access Journals (Sweden)

    Muhammad Farhan Ausaf

    2015-12-01

    Full Text Available Process planning and scheduling are two important components of a manufacturing setup. It is important to integrate them to achieve better global optimality and improved system performance. To find optimal solutions for integrated process planning and scheduling (IPPS problem, numerous algorithm-based approaches exist. Most of these approaches try to use existing meta-heuristic algorithms for solving the IPPS problem. Although these approaches have been shown to be effective in optimizing the IPPS problem, there is still room for improvement in terms of quality of solution and algorithm efficiency, especially for more complicated problems. Dispatching rules have been successfully utilized for solving complicated scheduling problems, but haven’t been considered extensively for the IPPS problem. This approach incorporates dispatching rules with the concept of prioritizing jobs, in an algorithm called priority-based heuristic algorithm (PBHA. PBHA tries to establish job and machine priority for selecting operations. Priority assignment and a set of dispatching rules are simultaneously used to generate both the process plans and schedules for all jobs and machines. The algorithm was tested for a series of benchmark problems. The proposed algorithm was able to achieve superior results for most complex problems presented in recent literature while utilizing lesser computational resources.

  3. Heuristic algorithm for single resource constrained project scheduling problem based on the dynamic programming

    Directory of Open Access Journals (Sweden)

    Stanimirović Ivan

    2009-01-01

    Full Text Available We introduce a heuristic method for the single resource constrained project scheduling problem, based on the dynamic programming solution of the knapsack problem. This method schedules projects with one type of resources, in the non-preemptive case: once started an activity is not interrupted and runs to completion. We compare the implementation of this method with well-known heuristic scheduling method, called Minimum Slack First (known also as Gray-Kidd algorithm, as well as with Microsoft Project.

  4. A MILP-based heuristic for a commercial train timetabling problem

    OpenAIRE

    Gestrelius, Sara; Aronsson, Martin; Peterson, Anders

    2017-01-01

    Using mathematical methods to support the yearly timetable planning process has many advantages. Unfortunately, the train timetabling problem for large geographical areas and many trains is intractable for optimization models alone. In this paper, we therefore present a MILP-based heuristic that has been designed to generate good-enough timetables for large geographical areas and many trains. In the incremental fix and release heuristic (IFRH), trains are added to the timetable in batches. Fo...

  5. Choosing a heuristic and root node for edge ordering in BDD-based network reliability analysis

    International Nuclear Information System (INIS)

    Mo, Yuchang; Xing, Liudong; Zhong, Farong; Pan, Zhusheng; Chen, Zhongyu

    2014-01-01

    In the Binary Decision Diagram (BDD)-based network reliability analysis, heuristics have been widely used to obtain a reasonably good ordering of edge variables. Orderings generated using different heuristics can lead to dramatically different sizes of BDDs, and thus dramatically different running times and memory usages for the analysis of the same network. Unfortunately, due to the nature of the ordering problem (i.e., being an NP-complete problem) no formal guidelines or rules are available for choosing a good heuristic or for choosing a high-performance root node to perform edge searching using a particular heuristic. In this work, we make novel contributions by proposing heuristic and root node selection methods based on the concept of boundary sets for the BDD-based network reliability analysis. Empirical studies show that the proposed selection methods can help to generate high-performance edge ordering for most of studied cases, enabling the efficient BDD-based reliability analysis of large-scale networks. The proposed methods are demonstrated on different types of networks, including square lattice networks, torus lattice networks and de Bruijn networks

  6. Proportional reasoning as a heuristic-based process: time constraint and dual task considerations.

    Science.gov (United States)

    Gillard, Ellen; Van Dooren, Wim; Schaeken, Walter; Verschaffel, Lieven

    2009-01-01

    The present study interprets the overuse of proportional solution methods from a dual process framework. Dual process theories claim that analytic operations involve time-consuming executive processing, whereas heuristic operations are fast and automatic. In two experiments to test whether proportional reasoning is heuristic-based, the participants solved "proportional" problems, for which proportional solution methods provide correct answers, and "nonproportional" problems known to elicit incorrect answers based on the assumption of proportionality. In Experiment 1, the available solution time was restricted. In Experiment 2, the executive resources were burdened with a secondary task. Both manipulations induced an increase in proportional answers and a decrease in correct answers to nonproportional problems. These results support the hypothesis that the choice for proportional methods is heuristic-based.

  7. PROBLEM SOLVING IN SCHOOL MATHEMATICS BASED ON HEURISTIC STRATEGIES

    Directory of Open Access Journals (Sweden)

    NOVOTNÁ, Jarmila

    2014-03-01

    Full Text Available The paper describes one of the ways of developing pupils’ creative approach to problem solving. The described experiment is a part of a longitudinal research focusing on improvement of culture of problem solving by pupils. It deals with solving of problems using the following heuristic strategies: Analogy, Guess – check – revise, Systematic experimentation, Problem reformulation, Solution drawing, Way back and Use of graphs of functions. Most attention is paid to the question whether short-term work, in this case only over the period of three months, can result in improvement of pupils’ abilities to solve problems whose solving algorithms are easily accessible. It also answers the question which strategies pupils will prefer and with what results. The experiment shows that even short-term work can bear positive results as far as pupils’ approach to problem solving is concerned.

  8. An ordering heuristic to develop the binary decision diagram based on structural importance

    International Nuclear Information System (INIS)

    Bartlett, L.M.; Andrews, J.D.

    2001-01-01

    Fault tree analysis is often used to assess risks within industrial systems. The technique is commonly used although there are associated limitations in terms of accuracy and efficiency when dealing with large fault tree structures. The most recent approach to aid the analysis of the fault tree diagram is the Binary Decision Diagram (BDD) methodology. To utilise the technique the fault tree structure needs to be converted into the BDD format. Converting the fault tree requires the basic events of the tree to be placed in an ordering. The ordering of the basic events is critical to the resulting size of the BDD, and ultimately affects the performance and benefits of this technique. A number of heuristic approaches have been developed to produce an optimal ordering permutation for a specific tree. These heuristic approaches do not always yield a minimal BDD structure for all trees. This paper looks at a heuristic that is based on the structural importance measure of each basic event. Comparing the resulting size of the BDD with the smallest generated from a set of six alternative ordering heuristics, this new structural heuristic produced a BDD of smaller or equal dimension on 77% of trials

  9. An LP-based heuristic for the fixed charge transportation problem

    DEFF Research Database (Denmark)

    Klose, Andreas

    2007-01-01

    The fixed charge transportation problem consists in finding a minimum cost network flow from a set of suppliers to a set of customers. Beside costs proportional to quantities transported, transportation costs also include a fixed charge. The paper describes a linear programming based heuristic...... approach for computing lower and upper bounds on the minimal cost. To this end, the LP relaxation is iteratively strengthened by means of adding cuts; in each iteration the current LP solution is then used to guide a local search heuristic. In addition to standard polyhedral cuts as lifted cover...

  10. Trends Concerning Four Misconceptions in Students' Intuitively-Based Probabilistic Reasoning Sourced in the Heuristic of Representativeness

    Science.gov (United States)

    Kustos, Paul Nicholas

    2010-01-01

    Student difficulty in the study of probability arises in intuitively-based misconceptions derived from heuristics. One such heuristic, the one of note for this research study, is that of representativeness, in which an individual informally assesses the probability of an event based on the degree to which the event is similar to the sample from…

  11. A column generation-based heuristic for rostering with work patterns

    DEFF Research Database (Denmark)

    Lusby, Richard Martin; Dohn, Anders Høeg; Range, Troels Martin

    2012-01-01

    This paper addresses the Ground Crew Rostering Problem with Work Patterns, an important manpower planning problem arising in the ground operations of airline companies. We present a cutting stock-based integer programming formulation of the problem and describe a powerful heuristic decomposition...

  12. Socially rational agents in spatial land use planning: a heuristic proposal based negotiation mechanism

    NARCIS (Netherlands)

    Ghavami, S.M.; Taleai, M.; Arentze, T.A.

    2016-01-01

    This paper introduces a novel heuristic based negotiation model for urban land use planning by using multi-agent systems. The model features two kinds of agents: facilitator and advocate. Facilitator agent runs the negotiation according to a certain protocol that defines the procedure. Two roles are

  13. Indoor Trajectory Tracking Scheme Based on Delaunay Triangulation and Heuristic Information in Wireless Sensor Networks.

    Science.gov (United States)

    Qin, Junping; Sun, Shiwen; Deng, Qingxu; Liu, Limin; Tian, Yonghong

    2017-06-02

    Object tracking and detection is one of the most significant research areas for wireless sensor networks. Existing indoor trajectory tracking schemes in wireless sensor networks are based on continuous localization and moving object data mining. Indoor trajectory tracking based on the received signal strength indicator ( RSSI ) has received increased attention because it has low cost and requires no special infrastructure. However, RSSI tracking introduces uncertainty because of the inaccuracies of measurement instruments and the irregularities (unstable, multipath, diffraction) of wireless signal transmissions in indoor environments. Heuristic information includes some key factors for trajectory tracking procedures. This paper proposes a novel trajectory tracking scheme based on Delaunay triangulation and heuristic information (TTDH). In this scheme, the entire field is divided into a series of triangular regions. The common side of adjacent triangular regions is regarded as a regional boundary. Our scheme detects heuristic information related to a moving object's trajectory, including boundaries and triangular regions. Then, the trajectory is formed by means of a dynamic time-warping position-fingerprint-matching algorithm with heuristic information constraints. Field experiments show that the average error distance of our scheme is less than 1.5 m, and that error does not accumulate among the regions.

  14. The Effectiveness of Local Culture-Based Mathematical Heuristic-KR Learning towards Enhancing Student's Creative Thinking Skill

    Science.gov (United States)

    Tandiseru, Selvi Rajuaty

    2015-01-01

    The problem in this research is the lack of creative thinking skills of students. One of the learning models that is expected to enhance student's creative thinking skill is the local culture-based mathematical heuristic-KR learning model (LC-BMHLM). Heuristic-KR is a learning model which was introduced by Krulik and Rudnick (1995) that is the…

  15. A derived heuristics based multi-objective optimization procedure for micro-grid scheduling

    Science.gov (United States)

    Li, Xin; Deb, Kalyanmoy; Fang, Yanjun

    2017-06-01

    With the availability of different types of power generators to be used in an electric micro-grid system, their operation scheduling as the load demand changes with time becomes an important task. Besides satisfying load balance constraints and the generator's rated power, several other practicalities, such as limited availability of grid power and restricted ramping of power output from generators, must all be considered during the operation scheduling process, which makes it difficult to decide whether the optimization results are accurate and satisfactory. In solving such complex practical problems, heuristics-based customized optimization algorithms are suggested. However, due to nonlinear and complex interactions of variables, it is difficult to come up with heuristics in such problems off-hand. In this article, a two-step strategy is proposed in which the first task deciphers important heuristics about the problem and the second task utilizes the derived heuristics to solve the original problem in a computationally fast manner. Specifically, the specific operation scheduling is considered from a two-objective (cost and emission) point of view. The first task develops basic and advanced level knowledge bases offline from a series of prior demand-wise optimization runs and then the second task utilizes them to modify optimized solutions in an application scenario. Results on island and grid connected modes and several pragmatic formulations of the micro-grid operation scheduling problem clearly indicate the merit of the proposed two-step procedure.

  16. Project Scheduling Heuristics-Based Standard PSO for Task-Resource Assignment in Heterogeneous Grid

    Directory of Open Access Journals (Sweden)

    Ruey-Maw Chen

    2011-01-01

    Full Text Available The task scheduling problem has been widely studied for assigning resources to tasks in heterogeneous grid environment. Effective task scheduling is an important issue for the performance of grid computing. Meanwhile, the task scheduling problem is an NP-complete problem. Hence, this investigation introduces a named “standard“ particle swarm optimization (PSO metaheuristic approach to efficiently solve the task scheduling problems in grid. Meanwhile, two promising heuristics based on multimode project scheduling are proposed to help in solving interesting scheduling problems. They are the best performance resource heuristic and the latest finish time heuristic. These two heuristics applied to the PSO scheme are for speeding up the search of the particle and improving the capability of finding a sound schedule. Moreover, both global communication topology and local ring communication topology are also investigated for efficient study of proposed scheme. Simulation results demonstrate that the proposed approach in this investigation can successfully solve the task-resource assignment problems in grid computing and similar scheduling problems.

  17. SPARSE: quadratic time simultaneous alignment and folding of RNAs without sequence-based heuristics.

    Science.gov (United States)

    Will, Sebastian; Otto, Christina; Miladi, Milad; Möhl, Mathias; Backofen, Rolf

    2015-08-01

    RNA-Seq experiments have revealed a multitude of novel ncRNAs. The gold standard for their analysis based on simultaneous alignment and folding suffers from extreme time complexity of [Formula: see text]. Subsequently, numerous faster 'Sankoff-style' approaches have been suggested. Commonly, the performance of such methods relies on sequence-based heuristics that restrict the search space to optimal or near-optimal sequence alignments; however, the accuracy of sequence-based methods breaks down for RNAs with sequence identities below 60%. Alignment approaches like LocARNA that do not require sequence-based heuristics, have been limited to high complexity ([Formula: see text] quartic time). Breaking this barrier, we introduce the novel Sankoff-style algorithm 'sparsified prediction and alignment of RNAs based on their structure ensembles (SPARSE)', which runs in quadratic time without sequence-based heuristics. To achieve this low complexity, on par with sequence alignment algorithms, SPARSE features strong sparsification based on structural properties of the RNA ensembles. Following PMcomp, SPARSE gains further speed-up from lightweight energy computation. Although all existing lightweight Sankoff-style methods restrict Sankoff's original model by disallowing loop deletions and insertions, SPARSE transfers the Sankoff algorithm to the lightweight energy model completely for the first time. Compared with LocARNA, SPARSE achieves similar alignment and better folding quality in significantly less time (speedup: 3.7). At similar run-time, it aligns low sequence identity instances substantially more accurate than RAF, which uses sequence-based heuristics. © The Author 2015. Published by Oxford University Press.

  18. SPARSE: quadratic time simultaneous alignment and folding of RNAs without sequence-based heuristics

    Science.gov (United States)

    Will, Sebastian; Otto, Christina; Miladi, Milad; Möhl, Mathias; Backofen, Rolf

    2015-01-01

    Motivation: RNA-Seq experiments have revealed a multitude of novel ncRNAs. The gold standard for their analysis based on simultaneous alignment and folding suffers from extreme time complexity of O(n6). Subsequently, numerous faster ‘Sankoff-style’ approaches have been suggested. Commonly, the performance of such methods relies on sequence-based heuristics that restrict the search space to optimal or near-optimal sequence alignments; however, the accuracy of sequence-based methods breaks down for RNAs with sequence identities below 60%. Alignment approaches like LocARNA that do not require sequence-based heuristics, have been limited to high complexity (≥ quartic time). Results: Breaking this barrier, we introduce the novel Sankoff-style algorithm ‘sparsified prediction and alignment of RNAs based on their structure ensembles (SPARSE)’, which runs in quadratic time without sequence-based heuristics. To achieve this low complexity, on par with sequence alignment algorithms, SPARSE features strong sparsification based on structural properties of the RNA ensembles. Following PMcomp, SPARSE gains further speed-up from lightweight energy computation. Although all existing lightweight Sankoff-style methods restrict Sankoff’s original model by disallowing loop deletions and insertions, SPARSE transfers the Sankoff algorithm to the lightweight energy model completely for the first time. Compared with LocARNA, SPARSE achieves similar alignment and better folding quality in significantly less time (speedup: 3.7). At similar run-time, it aligns low sequence identity instances substantially more accurate than RAF, which uses sequence-based heuristics. Availability and implementation: SPARSE is freely available at http://www.bioinf.uni-freiburg.de/Software/SPARSE. Contact: backofen@informatik.uni-freiburg.de Supplementary information: Supplementary data are available at Bioinformatics online. PMID:25838465

  19. PENDEKATAN ACTIVITY-BASED COSTING DAN METODE PENCARIAN HEURISTIC UNTUK MENYELESAIKAN PROBLEM PEMILIHAN PERALATAN PADA FLEXIBLE MANUFACTURING SYSTEMS (FMS

    Directory of Open Access Journals (Sweden)

    Gregorius Satia Budhi

    2002-01-01

    Full Text Available The application of Activity Based Costing (ABC approach to select the set-machine that is used in the production of Flexible Manufacture System (FMS based on technical and economical criteria can be useful for producers to design FMS by considering the minimum production cost. In the other hand, Heuristic Search is known to have a short searching time. Algorithm Heuristic that using ABC approach as the weight in finding the solution to shorten the equipment selection time during the design / redesign process of the FMS in less than exponential time was designed in this research. The increasing speed is useful because with the faster time in design / redesign process, therefore the flexibility level of part variety that can be processed will become better. Theoretical and empirical analysis in Algorithm Heuristic shows that time searching to get appropriate set of equipment is not too long, so that we can assume that the designed Algorithm Heuristic can be implemented in the real world. By comparing the empirical result of Algorithm Heuristic to the Algorithm Exhaustive, we can also assume that Algorithm Heuristic that using ABC method as the weight for finding solution can optimise the equipment selection problem of FMS based on economical criteria too. Abstract in Bahasa Indonesia : Penggunaan pendekatan Activity Based Costing (ABC untuk memilih set mesin yang digunakan dalam produksi pada Flexible Manufacture Systems (FMS berdasar atas kriteria teknis dan ekonomis, dapat membantu pelaku produksi untuk mendisain FMS dengan pertimbangan minimalisasi biaya produksi. Sementara itu, Heuristic Search dikenal memiliki waktu pencarian yang singkat. Pada riset ini didisain sebuah Algoritma Heuristic yang menggunakan pendekatan ABC sebagai bobot dalam pencarian solusi, untuk mempersingkat waktu pemilihan peralatan saat desain/redisain FMS dalam waktu kurang dari waktu Eksponensial. Peningkatan kecepatan ini bermanfaat, karena dengan cepatnya waktu

  20. The Heuristic Model Based on LPR in the Context of Material Conversion

    Directory of Open Access Journals (Sweden)

    Wilk-Kołodziejczyk D.

    2017-09-01

    Full Text Available High complexity of the physical and chemical processes occurring in liquid metal is the reason why it is so difficult, impossible even sometimes, to make analytical models of these phenomena. In this situation, the use of heuristic models based on the experimental data and experience of technicians is fully justified since, in an approximate manner at least, they allow predicting the mechanical properties of the metal manufactured under given process conditions. The study presents a methodology applicable in the design of a heuristic model based on the formalism of the logic of plausible reasoning (LPR. The problem under consideration consists in finding a technological variant of the process that will give the desired product parameters while minimizing the cost of production. The conducted tests have shown the effectiveness of the proposed approach.

  1. A variable neighborhood descent based heuristic to solve the capacitated location-routing problem

    Directory of Open Access Journals (Sweden)

    M. S. Jabal-Ameli

    2011-01-01

    Full Text Available Location-routing problem (LRP is established as a new research area in the context of location analysis. The primary concern of LRP is on locating facilities and routing of vehicles among established facilities and existing demand points. In this work, we address the capacitated LRP which arises in many practical applications within logistics and supply chain management. The objective is to minimize the overall system costs which include the fixed costs of opening depots and using vehicles at each depot site, and the variable costs associated with delivery activities. A novel heuristic is proposed which is based on variable neighborhood descent (VND algorithm to solve the resulted problem. The computational study indicates that the proposed VND based heuristic is highly competitive with the existing solution algorithms in terms of solution quality.

  2. Project Scheduling Heuristics-Based Standard PSO for Task-Resource Assignment in Heterogeneous Grid

    OpenAIRE

    Chen, Ruey-Maw; Wang, Chuin-Mu

    2011-01-01

    The task scheduling problem has been widely studied for assigning resources to tasks in heterogeneous grid environment. Effective task scheduling is an important issue for the performance of grid computing. Meanwhile, the task scheduling problem is an NP-complete problem. Hence, this investigation introduces a named “standard“ particle swarm optimization (PSO) metaheuristic approach to efficiently solve the task scheduling problems in grid. Meanwhile, two promising heuristics based on multimo...

  3. Robotics Vision-based Heuristic Reasoning for Underwater Target Tracking and Navigation

    OpenAIRE

    Kia, Chua; Arshad, Mohd Rizal

    2006-01-01

    This paper presents a robotics vision-based heuristic reasoning system for underwater target tracking and navigation. This system is introduced to improve the level of automation of underwater Remote Operated Vehicles (ROVs) operations. A prototype which combines computer vision with an underwater robotics system is successfully designed and developed to perform target tracking and intelligent navigation. This study focuses on developing image processing algorithms and fuzzy inference system ...

  4. State of the art in HGPT (Heuristically Based Generalized Perturbation) methodology

    International Nuclear Information System (INIS)

    Gandini, A.

    1993-01-01

    A distinctive feature of heuristically based generalized perturbation theory (HGPT) methodology consists in the systematic use of importance conservation concepts. As well known, this use leads to fundamental reciprocity relationships from which perturbation, or sensitivity, expressions can be derived. The state of the art of the HGPT methodology is here illustrated. The application to a number of specific nonlinear fields of interest is commented. (author)

  5. A marketing science perspective on recognition-based heuristics (and the fast-and-frugal paradigm

    Directory of Open Access Journals (Sweden)

    John Hauser

    2011-07-01

    Full Text Available Marketing science seeks to prescribe better marketing strategies (advertising, product development, pricing, etc.. To do so we rely on models of consumer decisions grounded in empirical observations. Field experience suggests that recognition-based heuristics help consumers to choose which brands to consider and purchase in frequently-purchased categories, but other heuristics are more relevant in durable-goods categories. Screening with recognition is a rational screening rule when advertising is a signal of product quality, when observing other consumers makes it easy to learn decision rules, and when firms react to engineering-design constraints by offering brands such that a high-level on one product feature implies a low level on another product feature. Experience with applications and field experiments suggests four fruitful research topics: deciding how to decide (endogeneity, learning decision rules by self-reflection, risk reduction, and the difference between utility functions and decision rules. These challenges also pose methodological cautions.

  6. Design Heuristics for Authentic Simulation-Based Learning Games

    Science.gov (United States)

    Ney, Muriel; Gonçalves, Celso; Balacheff, Nicolas

    2014-01-01

    Simulation games are games for learning based on a reference in the real world. We propose a model for authenticity in this context as a result of a compromise among learning, playing and realism. In the health game used to apply this model, students interact with characters in the game through phone messages, mail messages, SMS and video.…

  7. Agent-based transportation planning compared with scheduling heuristics

    NARCIS (Netherlands)

    Mes, Martijn R.K.; van der Heijden, Matthijs C.; van Harten, Aart

    2004-01-01

    Here we consider the problem of dynamically assigning vehicles to transportation orders that have di¤erent time windows and should be handled in real time. We introduce a new agent-based system for the planning and scheduling of these transportation networks. Intelligent vehicle agents schedule

  8. Reconciliation of Decision-Making Heuristics Based on Decision Trees Topologies and Incomplete Fuzzy Probabilities Sets.

    Science.gov (United States)

    Doubravsky, Karel; Dohnal, Mirko

    2015-01-01

    Complex decision making tasks of different natures, e.g. economics, safety engineering, ecology and biology, are based on vague, sparse, partially inconsistent and subjective knowledge. Moreover, decision making economists / engineers are usually not willing to invest too much time into study of complex formal theories. They require such decisions which can be (re)checked by human like common sense reasoning. One important problem related to realistic decision making tasks are incomplete data sets required by the chosen decision making algorithm. This paper presents a relatively simple algorithm how some missing III (input information items) can be generated using mainly decision tree topologies and integrated into incomplete data sets. The algorithm is based on an easy to understand heuristics, e.g. a longer decision tree sub-path is less probable. This heuristic can solve decision problems under total ignorance, i.e. the decision tree topology is the only information available. But in a practice, isolated information items e.g. some vaguely known probabilities (e.g. fuzzy probabilities) are usually available. It means that a realistic problem is analysed under partial ignorance. The proposed algorithm reconciles topology related heuristics and additional fuzzy sets using fuzzy linear programming. The case study, represented by a tree with six lotteries and one fuzzy probability, is presented in details.

  9. Reconciliation of Decision-Making Heuristics Based on Decision Trees Topologies and Incomplete Fuzzy Probabilities Sets.

    Directory of Open Access Journals (Sweden)

    Karel Doubravsky

    Full Text Available Complex decision making tasks of different natures, e.g. economics, safety engineering, ecology and biology, are based on vague, sparse, partially inconsistent and subjective knowledge. Moreover, decision making economists / engineers are usually not willing to invest too much time into study of complex formal theories. They require such decisions which can be (rechecked by human like common sense reasoning. One important problem related to realistic decision making tasks are incomplete data sets required by the chosen decision making algorithm. This paper presents a relatively simple algorithm how some missing III (input information items can be generated using mainly decision tree topologies and integrated into incomplete data sets. The algorithm is based on an easy to understand heuristics, e.g. a longer decision tree sub-path is less probable. This heuristic can solve decision problems under total ignorance, i.e. the decision tree topology is the only information available. But in a practice, isolated information items e.g. some vaguely known probabilities (e.g. fuzzy probabilities are usually available. It means that a realistic problem is analysed under partial ignorance. The proposed algorithm reconciles topology related heuristics and additional fuzzy sets using fuzzy linear programming. The case study, represented by a tree with six lotteries and one fuzzy probability, is presented in details.

  10. Section 3. General issues in management : Heuristics or experience-based techniques for making accounting judgments and learning

    OpenAIRE

    Schiller, Stefan

    2013-01-01

    The purpose of this paper is to further the development of initial accounting for internally generated intangible assets, relevant to both academics and practitioners, examining what happens when accountants are given principles-based discretion. This paper draws on existing insights into heuristics or experience-based techniques for making accounting judgments. Knowledge about judgment under uncertainty, and the general framework offered by the heuristics and biases program in particular, fo...

  11. Slime moulds use heuristics based on within-patch experience to decide when to leave.

    Science.gov (United States)

    Latty, Tanya; Beekman, Madeleine

    2015-04-15

    Animals foraging in patchy, non-renewing or slowly renewing environments must make decisions about how long to remain within a patch. Organisms can use heuristics ('rules of thumb') based on available information to decide when to leave the patch. Here, we investigated proximate patch-departure heuristics in two species of giant, brainless amoeba: the slime moulds Didymium bahiense and Physarum polycephalum. We explicitly tested the importance of information obtained through experience by eliminating chemosensory cues of patch quality. In P. polycephalum, patch departure was influenced by the consumption of high, and to a much lesser extent low, quality food items such that engulfing a food item increased patch-residency time. Physarum polycephalum also tended to forage for longer in darkened, 'safe' patches. In D. bahiense, engulfment of any food item increased patch residency irrespective of that food item's quality. Exposure to light had no effect on the patch-residency time of D. bahiense. Given that these organisms lack a brain, our results illustrate how the use of simple heuristics can give the impression that individuals make sophisticated foraging decisions. © 2015. Published by The Company of Biologists Ltd.

  12. Augmented neural networks and problem structure-based heuristics for the bin-packing problem

    Science.gov (United States)

    Kasap, Nihat; Agarwal, Anurag

    2012-08-01

    In this article, we report on a research project where we applied augmented-neural-networks (AugNNs) approach for solving the classical bin-packing problem (BPP). AugNN is a metaheuristic that combines a priority rule heuristic with the iterative search approach of neural networks to generate good solutions fast. This is the first time this approach has been applied to the BPP. We also propose a decomposition approach for solving harder BPP, in which subproblems are solved using a combination of AugNN approach and heuristics that exploit the problem structure. We discuss the characteristics of problems on which such problem structure-based heuristics could be applied. We empirically show the effectiveness of the AugNN and the decomposition approach on many benchmark problems in the literature. For the 1210 benchmark problems tested, 917 problems were solved to optimality and the average gap between the obtained solution and the upper bound for all the problems was reduced to under 0.66% and computation time averaged below 33 s per problem. We also discuss the computational complexity of our approach.

  13. Heuristic hybrid game approach for fleet condition-based maintenance planning

    International Nuclear Information System (INIS)

    Feng, Qiang; Bi, Xiong; Zhao, Xiujie; Chen, Yiran; Sun, Bo

    2017-01-01

    The condition-based maintenance (CBM) method is commonly used to select appropriate maintenance opportunities according to equipment status over a period of time. The CBM of aircraft fleets is a fleet maintenance planning problem. In this problem, mission requirements, resource constraints, and aircraft statuses are considered to find an optimal strategy set. Given that the maintenance strategies for each aircraft are finite, fleet CBM can be treated as a combinatorial optimization problem. In this study, the process of making a decision on the CBM of military fleets is analyzed. The fleet CBM problem is treated as a two-stage dynamic decision-making problem. Aircraft are divided into dispatch and standby sets; thus, the problem scale is significantly reduced. A heuristic hybrid game (HHG) approach comprising a competition game and a cooperative game is proposed on the basis of heuristic rule. In the dispatch set, a competition game approach is proposed to search for a local optimal strategy matrix. A cooperative game method for the two sets is also proposed to ensure global optimization. Finally, a case study regarding a fleet comprising 20 aircraft is conducted, with the results proving that the approach efficiently generates outcomes that meet the mission risk-oriented schedule requirement. - Highlights: • A new heuristic hybrid game method for fleet condition-based maintenance is proposed. • The problem is simplified by hierarchical solving based on dispatch and standby set. • The local optimal solution is got by competition game algorithm for dispatch set. • The global optimal solution is got by cooperative game algorithm between two sets.

  14. Yet Another Method for Image Segmentation based on Histograms and Heuristics

    Directory of Open Access Journals (Sweden)

    Horia-Nicolai L. Teodorescu

    2012-07-01

    Full Text Available We introduce a method for image segmentation that requires little computations, yet providing comparable results to other methods. While the proposed method resembles to the known ones based on histograms, it is still different in the use of the gray level distribution. When to the basic procedure we add several heuristic rules, the method produces results that, in some cases, may outperform the results produced by the known methods. The paper reports preliminary results. More details on the method, improvements, and results will be presented in a future paper.

  15. A heuristic two-dimensional presentation of microsatellite-based data applied to dogs and wolves

    Directory of Open Access Journals (Sweden)

    Foerster Martin

    2007-07-01

    Full Text Available Abstract Methods based on genetic distance matrices usually lose information during the process of tree-building by converting a multi-dimensional matrix into a phylogenetic tree. We applied a heuristic method of two-dimensional presentation to achieve a better resolution of the relationship between breeds and individuals investigated. Four hundred and nine individuals from nine German dog breed populations and one free-living wolf population were analysed with a marker set of 23 microsatellites. The result of the two-dimensional presentation was partly comparable with and complemented a model-based analysis that uses genotype patterns. The assignment test and the neighbour-joining tree based on allele sharing estimate allocated 99% and 97% of the individuals according to their breed, respectively. The application of the two-dimensional presentation to distances on the basis of the proportion of shared alleles resulted in comparable and further complementary insight into inferred population structure by multilocus genotype data. We expect that the inference of population structure in domesticated species with complex breeding histories can be strongly supported by the two-dimensional presentation based on the described heuristic method.

  16. Automated Segmentation of Coronary Arteries Based on Statistical Region Growing and Heuristic Decision Method

    Directory of Open Access Journals (Sweden)

    Yun Tian

    2016-01-01

    Full Text Available The segmentation of coronary arteries is a vital process that helps cardiovascular radiologists detect and quantify stenosis. In this paper, we propose a fully automated coronary artery segmentation from cardiac data volume. The method is built on a statistics region growing together with a heuristic decision. First, the heart region is extracted using a multi-atlas-based approach. Second, the vessel structures are enhanced via a 3D multiscale line filter. Next, seed points are detected automatically through a threshold preprocessing and a subsequent morphological operation. Based on the set of detected seed points, a statistics-based region growing is applied. Finally, results are obtained by setting conservative parameters. A heuristic decision method is then used to obtain the desired result automatically because parameters in region growing vary in different patients, and the segmentation requires full automation. The experiments are carried out on a dataset that includes eight-patient multivendor cardiac computed tomography angiography (CTA volume data. The DICE similarity index, mean distance, and Hausdorff distance metrics are employed to compare the proposed algorithm with two state-of-the-art methods. Experimental results indicate that the proposed algorithm is capable of performing complete, robust, and accurate extraction of coronary arteries.

  17. Generic component failure data base

    International Nuclear Information System (INIS)

    Eide, S.A.; Calley, M.B.

    1992-01-01

    This report discusses comprehensive component generic failure data base which has been developed for light water reactor probabilistic risk assessments. The Nuclear Computerized Library for Assessing Reactor Reliability (NUCLARR) was used to generate component failure rates. Using this approach, most of the failure rates are based on actual plant data rather then existing estimates

  18. Comparative study of heuristics algorithms in solving flexible job shop scheduling problem with condition based maintenance

    Directory of Open Access Journals (Sweden)

    Yahong Zheng

    2014-05-01

    Full Text Available Purpose: This paper focuses on a classic optimization problem in operations research, the flexible job shop scheduling problem (FJSP, to discuss the method to deal with uncertainty in a manufacturing system.Design/methodology/approach: In this paper, condition based maintenance (CBM, a kind of preventive maintenance, is suggested to reduce unavailability of machines. Different to the simultaneous scheduling algorithm (SSA used in the previous article (Neale & Cameron,1979, an inserting algorithm (IA is applied, in which firstly a pre-schedule is obtained through heuristic algorithm and then maintenance tasks are inserted into the pre-schedule scheme.Findings: It is encouraging that a new better solution for an instance in benchmark of FJSP is obtained in this research. Moreover, factually SSA used in literature for solving normal FJSPPM (FJSP with PM is not suitable for the dynamic FJSPPM. Through application in the benchmark of normal FJSPPM, it is found that although IA obtains inferior results compared to SSA used in literature, it performs much better in executing speed.Originality/value: Different to traditional scheduling of FJSP, uncertainty of machines is taken into account, which increases the complexity of the problem. An inserting algorithm (IA is proposed to solve the dynamic scheduling problem. It is stated that the quality of the final result depends much on the quality of the pre-schedule obtained during the procedure of solving a normal FJSP. In order to find the best solution of FJSP, a comparative study of three heuristics is carried out, the integrated GA, ACO and ABC. In the comparative study, we find that GA performs best in the three heuristic algorithms. Meanwhile, a new better solution for an instance in benchmark of FJSP is obtained in this research.

  19. Can the inherence heuristic explain vitalistic reasoning?

    Science.gov (United States)

    Bastian, Brock

    2014-10-01

    Inherence is an important component of psychological essentialism. By drawing on vitalism as a way in which to explain this link, however, the authors appear to conflate causal explanations based on fixed features with those based on general causal forces. The disjuncture between these two types of explanatory principles highlights potential new avenues for the inherence heuristic.

  20. Thermodynamic heuristics with case-based reasoning: combined insights for RNA pseudoknot secondary structure.

    Science.gov (United States)

    Al-Khatib, Ra'ed M; Rashid, Nur'Aini Abdul; Abdullah, Rosni

    2011-08-01

    The secondary structure of RNA pseudoknots has been extensively inferred and scrutinized by computational approaches. Experimental methods for determining RNA structure are time consuming and tedious; therefore, predictive computational approaches are required. Predicting the most accurate and energy-stable pseudoknot RNA secondary structure has been proven to be an NP-hard problem. In this paper, a new RNA folding approach, termed MSeeker, is presented; it includes KnotSeeker (a heuristic method) and Mfold (a thermodynamic algorithm). The global optimization of this thermodynamic heuristic approach was further enhanced by using a case-based reasoning technique as a local optimization method. MSeeker is a proposed algorithm for predicting RNA pseudoknot structure from individual sequences, especially long ones. This research demonstrates that MSeeker improves the sensitivity and specificity of existing RNA pseudoknot structure predictions. The performance and structural results from this proposed method were evaluated against seven other state-of-the-art pseudoknot prediction methods. The MSeeker method had better sensitivity than the DotKnot, FlexStem, HotKnots, pknotsRG, ILM, NUPACK and pknotsRE methods, with 79% of the predicted pseudoknot base-pairs being correct.

  1. Heuristic evaluation of paper-based Web pages: a simplified inspection usability methodology.

    Science.gov (United States)

    Allen, Mureen; Currie, Leanne M; Bakken, Suzanne; Patel, Vimla L; Cimino, James J

    2006-08-01

    Online medical information, when presented to clinicians, must be well-organized and intuitive to use, so that the clinicians can conduct their daily work efficiently and without error. It is essential to actively seek to produce good user interfaces that are acceptable to the user. This paper describes the methodology used to develop a simplified heuristic evaluation (HE) suitable for the evaluation of screen shots of Web pages, the development of an HE instrument used to conduct the evaluation, and the results of the evaluation of the aforementioned screen shots. In addition, this paper presents examples of the process of categorizing problems identified by the HE and the technological solutions identified to resolve these problems. Four usability experts reviewed 18 paper-based screen shots and made a total of 108 comments. Each expert completed the task in about an hour. We were able to implement solutions to approximately 70% of the violations. Our study found that a heuristic evaluation using paper-based screen shots of a user interface was expeditious, inexpensive, and straightforward to implement.

  2. Heuristic guidelines for making effective augmented-reality based training instruction

    International Nuclear Information System (INIS)

    Yim, Ho Bin

    2009-02-01

    As industrial plants and factories age, their maintenance requirements increase. Because maintenance mistakes directly increase the operating costs of a power plant, maintenance quality is of significant concern to plant management. By law, all personnel working with nuclear technology must be re-trained every three years; however, as the statistical data show, the number of shutdown accidents at nuclear power plants (NPPs) due to maintenance failure is still high and needs to be reduced. Industries have started to adopt various technologies to increase the speed and accuracy of maintenance. Among those technologies, augmented reality (AR) is the latest multimedia presentation technology to be applied to plant maintenance, and it offers superior intuitiveness and user interactivity over other conventional multimedia. The objectives of this empirical study are to measure the optimum amounts of information to be delivered at a time and to identify what types of information enhance the learning ability of novices and to suggest heuristic guidelines by which to make effective AR training instructions. In addition, AR training instructions will be composed based on the suggested heuristic guidelines and on the Cognitive Theory of Multimedia Learning, which provides guidelines for multimedia instructions, and the efficiency of those guidelines will be compared with that of other AR instructions. Cognitive Load Theory addresses how working memory is structured. Intrinsic cognitive load refers to the cognitive load that is inherent in information to be learned. It depends on the basic amount of processing required for understanding a presentation. Intrinsic cognitive load is determined by the complexity of the learning material. Extraneous cognitive load refers to the cognitive load created through the presentation, format and delivery of the instruction. Germane cognitive load is the remaining part of the cognitive load, and it helps the learner transfer information from

  3. Heuristic guidelines for making effective augmented-reality based training instruction

    Energy Technology Data Exchange (ETDEWEB)

    Yim, Ho Bin

    2009-02-15

    As industrial plants and factories age, their maintenance requirements increase. Because maintenance mistakes directly increase the operating costs of a power plant, maintenance quality is of significant concern to plant management. By law, all personnel working with nuclear technology must be re-trained every three years; however, as the statistical data show, the number of shutdown accidents at nuclear power plants (NPPs) due to maintenance failure is still high and needs to be reduced. Industries have started to adopt various technologies to increase the speed and accuracy of maintenance. Among those technologies, augmented reality (AR) is the latest multimedia presentation technology to be applied to plant maintenance, and it offers superior intuitiveness and user interactivity over other conventional multimedia. The objectives of this empirical study are to measure the optimum amounts of information to be delivered at a time and to identify what types of information enhance the learning ability of novices and to suggest heuristic guidelines by which to make effective AR training instructions. In addition, AR training instructions will be composed based on the suggested heuristic guidelines and on the Cognitive Theory of Multimedia Learning, which provides guidelines for multimedia instructions, and the efficiency of those guidelines will be compared with that of other AR instructions. Cognitive Load Theory addresses how working memory is structured. Intrinsic cognitive load refers to the cognitive load that is inherent in information to be learned. It depends on the basic amount of processing required for understanding a presentation. Intrinsic cognitive load is determined by the complexity of the learning material. Extraneous cognitive load refers to the cognitive load created through the presentation, format and delivery of the instruction. Germane cognitive load is the remaining part of the cognitive load, and it helps the learner transfer information from

  4. Derived heuristics-based consistent optimization of material flow in a gold processing plant

    Science.gov (United States)

    Myburgh, Christie; Deb, Kalyanmoy

    2018-01-01

    Material flow in a chemical processing plant often follows complicated control laws and involves plant capacity constraints. Importantly, the process involves discrete scenarios which when modelled in a programming format involves if-then-else statements. Therefore, a formulation of an optimization problem of such processes becomes complicated with nonlinear and non-differentiable objective and constraint functions. In handling such problems using classical point-based approaches, users often have to resort to modifications and indirect ways of representing the problem to suit the restrictions associated with classical methods. In a particular gold processing plant optimization problem, these facts are demonstrated by showing results from MATLAB®'s well-known fmincon routine. Thereafter, a customized evolutionary optimization procedure which is capable of handling all complexities offered by the problem is developed. Although the evolutionary approach produced results with comparatively less variance over multiple runs, the performance has been enhanced by introducing derived heuristics associated with the problem. In this article, the development and usage of derived heuristics in a practical problem are presented and their importance in a quick convergence of the overall algorithm is demonstrated.

  5. Optimal Control of Complex Systems Based on Improved Dual Heuristic Dynamic Programming Algorithm

    Directory of Open Access Journals (Sweden)

    Hui Li

    2017-01-01

    Full Text Available When applied to solving the data modeling and optimal control problems of complex systems, the dual heuristic dynamic programming (DHP technique, which is based on the BP neural network algorithm (BP-DHP, has difficulty in prediction accuracy, slow convergence speed, poor stability, and so forth. In this paper, a dual DHP technique based on Extreme Learning Machine (ELM algorithm (ELM-DHP was proposed. Through constructing three kinds of network structures, the paper gives the detailed realization process of the DHP technique in the ELM. The controller designed upon the ELM-DHP algorithm controlled a molecular distillation system with complex features, such as multivariability, strong coupling, and nonlinearity. Finally, the effectiveness of the algorithm is verified by the simulation that compares DHP and HDP algorithms based on ELM and BP neural network. The algorithm can also be applied to solve the data modeling and optimal control problems of similar complex systems.

  6. A Novel Approach for Bi-Level Segmentation of Tuberculosis Bacilli Based on Meta-Heuristic Algorithms

    Directory of Open Access Journals (Sweden)

    AYAS, S.

    2018-02-01

    Full Text Available Image thresholding is the most crucial step in microscopic image analysis to distinguish bacilli objects causing of tuberculosis disease. Therefore, several bi-level thresholding algorithms are widely used to increase the bacilli segmentation accuracy. However, bi-level microscopic image thresholding problem has not been solved using optimization algorithms. This paper introduces a novel approach for the segmentation problem using heuristic algorithms and presents visual and quantitative comparisons of heuristic and state-of-art thresholding algorithms. In this study, well-known heuristic algorithms such as Firefly Algorithm, Particle Swarm Optimization, Cuckoo Search, Flower Pollination are used to solve bi-level microscopic image thresholding problem, and the results are compared with the state-of-art thresholding algorithms such as K-Means, Fuzzy C-Means, Fast Marching. Kapur's entropy is chosen as the entropy measure to be maximized. Experiments are performed to make comparisons in terms of evaluation metrics and execution time. The quantitative results are calculated based on ground truth segmentation. According to the visual results, heuristic algorithms have better performance and the quantitative results are in accord with the visual results. Furthermore, experimental time comparisons show the superiority and effectiveness of the heuristic algorithms over traditional thresholding algorithms.

  7. Heuristic decision making in medicine

    Science.gov (United States)

    Marewski, Julian N.; Gigerenzer, Gerd

    2012-01-01

    Can less information be more helpful when it comes to making medical decisions? Contrary to the common intuition that more information is always better, the use of heuristics can help both physicians and patients to make sound decisions. Heuristics are simple decision strategies that ignore part of the available information, basing decisions on only a few relevant predictors. We discuss: (i) how doctors and patients use heuristics; and (ii) when heuristics outperform information-greedy methods, such as regressions in medical diagnosis. Furthermore, we outline those features of heuristics that make them useful in health care settings. These features include their surprising accuracy, transparency, and wide accessibility, as well as the low costs and little time required to employ them. We close by explaining one of the statistical reasons why heuristics are accurate, and by pointing to psychiatry as one area for future research on heuristics in health care. PMID:22577307

  8. Heuristic decision making in medicine.

    Science.gov (United States)

    Marewski, Julian N; Gigerenzer, Gerd

    2012-03-01

    Can less information be more helpful when it comes to making medical decisions? Contrary to the common intuition that more information is always better, the use of heuristics can help both physicians and patients to make sound decisions. Heuristics are simple decision strategies that ignore part of the available information, basing decisions on only a few relevant predictors. We discuss: (i) how doctors and patients use heuristics; and (ii) when heuristics outperform information-greedy methods, such as regressions in medical diagnosis. Furthermore, we outline those features of heuristics that make them useful in health care settings. These features include their surprising accuracy, transparency, and wide accessibility, as well as the low costs and little time required to employ them. We close by explaining one of the statistical reasons why heuristics are accurate, and by pointing to psychiatry as one area for future research on heuristics in health care.

  9. Evidence-based Heuristics for Evaluating Demands on eHealth Literacy and Usability in a Mobile Consumer Health Application.

    Science.gov (United States)

    Monkman, Helen; Griffith, Janessa; Kushniruk, Andre W

    2015-01-01

    Heuristic evaluations have proven to be valuable for identifying usability issues in systems. Commonly used sets of heuritics exist; however, they may not always be the most suitable, given the specific goal of the analysis. One such example is seeking to evaluate the demands on eHealth literacy and usability of consumer health information systems. In this study, eight essential heuristics and three optional heuristics subsumed from the evidence on eHealth/health literacy and usability were tested for their utility in assessing a mobile blood pressure tracking application (app). This evaluation revealed a variety of ways the design of the app could both benefit and impede users with limited eHealth literacy. This study demonstrated the utility of a low-cost, single evaluation approach for identifying both eHealth literacy and usability issues based on existing evidence in the literature.

  10. The Memory State Heuristic: A Formal Model Based on Repeated Recognition Judgments

    Science.gov (United States)

    Castela, Marta; Erdfelder, Edgar

    2017-01-01

    The recognition heuristic (RH) theory predicts that, in comparative judgment tasks, if one object is recognized and the other is not, the recognized one is chosen. The memory-state heuristic (MSH) extends the RH by assuming that choices are not affected by recognition judgments per se, but by the memory states underlying these judgments (i.e.,…

  11. A heuristics-based solution to the continuous berth allocation and crane assignment problem

    Directory of Open Access Journals (Sweden)

    Mohammad Hamdy Elwany

    2013-12-01

    Full Text Available Effective utilization plans for various resources at a container terminal are essential to reducing the turnaround time of cargo vessels. Among the scarcest resources are the berth and its associated cranes. Thus, two important optimization problems arise, which are the berth allocation and quay crane assignment problems. The berth allocation problem deals with the generation of a berth plan, which determines where and when a ship has to berth alongside the quay. The quay crane assignment problem addresses the problem of determining how many and which quay crane(s will serve each vessel. In this paper, an integrated heuristics-based solution methodology is proposed that tackles both problems simultaneously. The preliminary experimental results show that the proposed approach yields high quality solutions to such an NP-hard problem in a reasonable computational time suggesting its suitability for practical use.

  12. The role of the control variable in the heuristically based generalized perturbation theory (HGPT)

    International Nuclear Information System (INIS)

    Gandini, A.

    1995-01-01

    The heuristically based generalized perturbation theory (HGPT) applied to the neutron field of a reactor system is discussed in relation to the criticality reset procedure. This procedure is implicit within the GPT methodology, corresponding to the so called filtering of the importance function relevant to the neutron field from the fundamental mode contamination. It is common practice to use the so called γ-mode filter. In order to account for any possible reset option, a general definition is introduced of an intensive control variable (ρ) entering into the governing equations, and correspondingly a fundamental ρ-mode filtering of the importance function is defined, relevant to the real criticality reset mechanism (control) adopted. A simple example illustrates the need in many circumstances of interest of taking into proper account the correct filtering so to avoid significant inaccuracies in the sensitivity calculation results

  13. Case-based clinical reasoning in feline medicine: 3: Use of heuristics and illness scripts.

    Science.gov (United States)

    Whitehead, Martin L; Canfield, Paul J; Johnson, Robert; O'Brien, Carolyn R; Malik, Richard

    2016-05-01

    This is Article 3 of a three-part series on clinical reasoning that encourages practitioners to explore and understand how they think and make case-based decisions. It is hoped that, in the process, they will learn to trust their intuition but, at the same time, put in place safeguards to diminish the impact of bias and misguided logic on their diagnostic decision-making. Article 1, published in the January 2016 issue of JFMS, discussed the relative merits and shortcomings of System 1 thinking (immediate and unconscious) and System 2 thinking (effortful and analytical). In Article 2, published in the March 2016 issue, ways of managing cognitive error, particularly the negative impact of bias, in making a diagnosis were examined. This final article explores the use of heuristics (mental short cuts) and illness scripts in diagnostic reasoning. © The Author(s) 2016.

  14. Adaptive Event-Triggered Control Based on Heuristic Dynamic Programming for Nonlinear Discrete-Time Systems.

    Science.gov (United States)

    Dong, Lu; Zhong, Xiangnan; Sun, Changyin; He, Haibo

    2017-07-01

    This paper presents the design of a novel adaptive event-triggered control method based on the heuristic dynamic programming (HDP) technique for nonlinear discrete-time systems with unknown system dynamics. In the proposed method, the control law is only updated when the event-triggered condition is violated. Compared with the periodic updates in the traditional adaptive dynamic programming (ADP) control, the proposed method can reduce the computation and transmission cost. An actor-critic framework is used to learn the optimal event-triggered control law and the value function. Furthermore, a model network is designed to estimate the system state vector. The main contribution of this paper is to design a new trigger threshold for discrete-time systems. A detailed Lyapunov stability analysis shows that our proposed event-triggered controller can asymptotically stabilize the discrete-time systems. Finally, we test our method on two different discrete-time systems, and the simulation results are included.

  15. Special relativity a heuristic approach

    CERN Document Server

    Hassani, Sadri

    2017-01-01

    Special Relativity: A Heuristic Approach provides a qualitative exposition of relativity theory on the basis of the constancy of the speed of light. Using Einstein's signal velocity as the defining idea for the notion of simultaneity and the fact that the speed of light is independent of the motion of its source, chapters delve into a qualitative exposition of the relativity of time and length, discuss the time dilation formula using the standard light clock, explore the Minkowski four-dimensional space-time distance based on how the time dilation formula is derived, and define the components of the two-dimensional space-time velocity, amongst other topics. Provides a heuristic derivation of the Minkowski distance formula Uses relativistic photography to see Lorentz transformation and vector algebra manipulation in action Includes worked examples to elucidate and complement the topic being discussed Written in a very accessible style

  16. Evidence-based and heuristic approaches for customization of care in cardiometabolic syndrome after spinal cord injury.

    Science.gov (United States)

    Nash, Mark S; Cowan, Rachel E; Kressler, Jochen

    2012-09-01

    Component and coalesced health risks of the cardiometabolic syndrome (CMS) are commonly reported in persons with spinal cord injuries (SCIs). These CMS hazards are also co-morbid with physical deconditioning and elevated pro-atherogenic inflammatory cytokines, both of which are common after SCI and worsen the prognosis for all-cause cardiovascular disease. This article describes a systematic procedure for individualized CMS risk assessment after SCI, and emphasizes evidence-based and intuition-centered countermeasures to disease. A unified approach will propose therapeutic lifestyle intervention as a routine plan for aggressive primary prevention in this risk-susceptible population. Customization of dietary and exercise plans then follow, identifying shortfalls in diet and activity patterns, and ways in which these healthy lifestyles can be more substantially embraced by both stakeholders with SCI and their health care providers. In cases where lifestyle intervention utilizing diet and exercise is unsuccessful in countering risks, available pharmacotherapies and a preferred therapeutic agent are proposed according to authoritative standards. The over-arching purpose of the monograph is to create an operational framework in which existing evidence-based approaches or heuristic modeling becomes best practice. In this way persons with SCI can lead more active and healthy lives.

  17. A simple heuristic for Internet-based evidence search in primary care: a randomized controlled trial

    Directory of Open Access Journals (Sweden)

    Eberbach A

    2016-08-01

    Full Text Available Andreas Eberbach,1 Annette Becker,1 Justine Rochon,2 Holger Finkemeler,1Achim Wagner,3 Norbert Donner-Banzhoff1 1Department of Family and Community Medicine, Philipp University of Marburg, Marburg, Germany; 2Institute of Medical Biometry and Informatics, University of Heidelberg, Heidelberg, Germany; 3Department of Sport Medicine, Justus-Liebig-University of Giessen, Giessen, Germany Background: General practitioners (GPs are confronted with a wide variety of clinical questions, many of which remain unanswered. Methods: In order to assist GPs in finding quick, evidence-based answers, we developed a learning program (LP with a short interactive workshop based on a simple ­three-step-heuristic to improve their search and appraisal competence (SAC. We evaluated the LP ­effectiveness with a randomized controlled trial (RCT. Participants (intervention group [IG] n=20; ­control group [CG] n=31 rated acceptance and satisfaction and also answered 39 ­knowledge ­questions to assess their SAC. We controlled for previous knowledge in content areas covered by the test. Results: Main outcome – SAC: within both groups, the pre–post test shows significant (P=0.00 improvements in correctness (IG 15% vs CG 11% and confidence (32% vs 26% to find evidence-based answers. However, the SAC difference was not significant in the RCT. Other measures: Most workshop participants rated “learning atmosphere” (90%, “skills acquired” (90%, and “relevancy to my practice” (86% as good or very good. The ­LP-recommendations were implemented by 67% of the IG, whereas 15% of the CG already conformed to LP recommendations spontaneously (odds ratio 9.6, P=0.00. After literature search, the IG showed a (not significantly higher satisfaction regarding “time spent” (IG 80% vs CG 65%, “quality of information” (65% vs 54%, and “amount of information” (53% vs 47%.Conclusion: Long-standing established GPs have a good SAC. Despite high acceptance, strong

  18. Multi-Criteria Optimization of the Deployment of a Grid for Rural Electrification Based on a Heuristic Method

    Science.gov (United States)

    Ortiz-Matos, L.; Aguila-Tellez, A.; Hincapié-Reyes, R. C.; González-Sanchez, J. W.

    2017-07-01

    In order to design electrification systems, recent mathematical models solve the problem of location, type of electrification components, and the design of possible distribution microgrids. However, due to the amount of points to be electrified increases, the solution to these models require high computational times, thereby becoming unviable practice models. This study posed a new heuristic method for the electrification of rural areas in order to solve the problem. This heuristic algorithm presents the deployment of rural electrification microgrids in the world, by finding routes for optimal placement lines and transformers in transmission and distribution microgrids. The challenge is to obtain a display with equity in losses, considering the capacity constraints of the devices and topology of the land at minimal economic cost. An optimal scenario ensures the electrification of all neighbourhoods to a minimum investment cost in terms of the distance between electric conductors and the amount of transformation devices.

  19. Heuristics for Multidimensional Packing Problems

    DEFF Research Database (Denmark)

    Egeblad, Jens

    for a minimum height container required for the items. The main contributions of the thesis are three new heuristics for strip-packing and knapsack packing problems where items are both rectangular and irregular. In the two first papers we describe a heuristic for the multidimensional strip-packing problem...... that is based on a relaxed placement principle. The heuristic starts with a random overlapping placement of items and large container dimensions. From the overlapping placement overlap is reduced iteratively until a non-overlapping placement is found and a new problem is solved with a smaller container size...... of this heuristic are among the best published in the literature both for two- and three-dimensional strip-packing problems for irregular shapes. In the third paper, we introduce a heuristic for two- and three-dimensional rectangular knapsack packing problems. The two-dimensional heuristic uses the sequence pair...

  20. Robotics Vision-based Heuristic Reasoning for Underwater Target Tracking and Navigation

    Directory of Open Access Journals (Sweden)

    Chua Kia

    2005-09-01

    Full Text Available This paper presents a robotics vision-based heuristic reasoning system for underwater target tracking and navigation. This system is introduced to improve the level of automation of underwater Remote Operated Vehicles (ROVs operations. A prototype which combines computer vision with an underwater robotics system is successfully designed and developed to perform target tracking and intelligent navigation. This study focuses on developing image processing algorithms and fuzzy inference system for the analysis of the terrain. The vision system developed is capable of interpreting underwater scene by extracting subjective uncertainties of the object of interest. Subjective uncertainties are further processed as multiple inputs of a fuzzy inference system that is capable of making crisp decisions concerning where to navigate. The important part of the image analysis is morphological filtering. The applications focus on binary images with the extension of gray-level concepts. An open-loop fuzzy control system is developed for classifying the traverse of terrain. The great achievement is the system's capability to recognize and perform target tracking of the object of interest (pipeline in perspective view based on perceived condition. The effectiveness of this approach is demonstrated by computer and prototype simulations. This work is originated from the desire to develop robotics vision system with the ability to mimic the human expert's judgement and reasoning when maneuvering ROV in the traverse of the underwater terrain.

  1. State-Transition-Aware Spilling Heuristic for MLC STT-RAM-Based Registers

    Directory of Open Access Journals (Sweden)

    Yuanhui Ni

    2017-01-01

    Full Text Available Multilevel Cell Spin-Transfer Torque Random Access Memory (MLC STT-RAM is a promising nonvolatile memory technology to build registers for its natural immunity to electromagnetic radiation in rad-hard space environment. Unlike traditional SRAM-based registers, MLC STT-RAM exhibits unbalanced write state transitions due to the fact that the magnetization directions of hard and soft domains cannot be flipped independently. This feature leads to nonuniform costs of write states in terms of latency and energy. However, current SRAM-targeting register allocations do not have a clear understanding of the impact of the different write state-transition costs. As a result, those approaches heuristically select variables to be spilled without considering the spilling priority imposed by MLC STT-RAM. Aiming to address this limitation, this paper proposes a state-transition-aware spilling cost minimization (SSCM policy, to save power when MLC STT-RAM is employed in register design. Specifically, the spilling cost model is first constructed according to the linear combination of different state-transition frequencies. Directed by the proposed cost model, the compiler picks up spilling candidates to achieve lower power and higher performance. Experimental results show that the proposed SSCM technique can save energy by 19.4% and improve the lifetime by 23.2% of MLC STT-RAM-based register design.

  2. Robotics Vision-based Heuristic Reasoning for Underwater Target Tracking and Navigation

    Directory of Open Access Journals (Sweden)

    Chua Kia

    2008-11-01

    Full Text Available This paper presents a robotics vision-based heuristic reasoning system for underwater target tracking and navigation. This system is introduced to improve the level of automation of underwater Remote Operated Vehicles (ROVs operations. A prototype which combines computer vision with an underwater robotics system is successfully designed and developed to perform target tracking and intelligent navigation. This study focuses on developing image processing algorithms and fuzzy inference system for the analysis of the terrain. The vision system developed is capable of interpreting underwater scene by extracting subjective uncertainties of the object of interest. Subjective uncertainties are further processed as multiple inputs of a fuzzy inference system that is capable of making crisp decisions concerning where to navigate. The important part of the image analysis is morphological filtering. The applications focus on binary images with the extension of gray-level concepts. An open-loop fuzzy control system is developed for classifying the traverse of terrain. The great achievement is the system's capability to recognize and perform target tracking of the object of interest (pipeline in perspective view based on perceived condition. The effectiveness of this approach is demonstrated by computer and prototype simulations. This work is originated from the desire to develop robotics vision system with the ability to mimic the human expert's judgement and reasoning when maneuvering ROV in the traverse of the underwater terrain.

  3. Action dependent heuristic dynamic programming based residential energy scheduling with home energy inter-exchange

    International Nuclear Information System (INIS)

    Xu, Yancai; Liu, Derong; Wei, Qinglai

    2015-01-01

    Highlights: • The algorithm is developed in the two-household energy management environment. • We develop the absent energy penalty cost for the first time. • The algorithm has ability to keep adapting in real-time operations. • Its application can lower total costs and achieve better load balancing. - Abstract: Residential energy scheduling is a hot topic nowadays in the background of energy saving and environmental protection worldwide. To achieve this objective, a new residential energy scheduling algorithm is developed for energy management, based on action dependent heuristic dynamic programming. The algorithm works under the circumstance of residential real-time pricing and two adjacent housing units with energy inter-exchange, which can reduce the overall cost and enhance renewable energy efficiency after long-term operation. It is designed to obtain the optimal control policy to manage the directions and amounts of electricity energy flux. The algorithm’s architecture is mainly constructed based on neural networks, denoting the learned characteristics in the linkage of layers. To get close to real situations, many constraints such as maximum charging/discharging power of batteries are taken into account. The absent energy penalty cost is developed for the first time as a part of the performance index function. When the environment changes, the residential energy scheduling algorithm gains new features and keeps adapting in real-time operations. Simulation results show that the developed algorithm is beneficial to energy conversation

  4. A heuristic ranking approach on capacity benefit margin determination using Pareto-based evolutionary programming technique.

    Science.gov (United States)

    Othman, Muhammad Murtadha; Abd Rahman, Nurulazmi; Musirin, Ismail; Fotuhi-Firuzabad, Mahmud; Rajabi-Ghahnavieh, Abbas

    2015-01-01

    This paper introduces a novel multiobjective approach for capacity benefit margin (CBM) assessment taking into account tie-line reliability of interconnected systems. CBM is the imperative information utilized as a reference by the load-serving entities (LSE) to estimate a certain margin of transfer capability so that a reliable access to generation through interconnected system could be attained. A new Pareto-based evolutionary programming (EP) technique is used to perform a simultaneous determination of CBM for all areas of the interconnected system. The selection of CBM at the Pareto optimal front is proposed to be performed by referring to a heuristic ranking index that takes into account system loss of load expectation (LOLE) in various conditions. Eventually, the power transfer based available transfer capability (ATC) is determined by considering the firm and nonfirm transfers of CBM. A comprehensive set of numerical studies are conducted on the modified IEEE-RTS79 and the performance of the proposed method is numerically investigated in detail. The main advantage of the proposed technique is in terms of flexibility offered to an independent system operator in selecting an appropriate solution of CBM simultaneously for all areas.

  5. A Heuristic Ranking Approach on Capacity Benefit Margin Determination Using Pareto-Based Evolutionary Programming Technique

    Directory of Open Access Journals (Sweden)

    Muhammad Murtadha Othman

    2015-01-01

    Full Text Available This paper introduces a novel multiobjective approach for capacity benefit margin (CBM assessment taking into account tie-line reliability of interconnected systems. CBM is the imperative information utilized as a reference by the load-serving entities (LSE to estimate a certain margin of transfer capability so that a reliable access to generation through interconnected system could be attained. A new Pareto-based evolutionary programming (EP technique is used to perform a simultaneous determination of CBM for all areas of the interconnected system. The selection of CBM at the Pareto optimal front is proposed to be performed by referring to a heuristic ranking index that takes into account system loss of load expectation (LOLE in various conditions. Eventually, the power transfer based available transfer capability (ATC is determined by considering the firm and nonfirm transfers of CBM. A comprehensive set of numerical studies are conducted on the modified IEEE-RTS79 and the performance of the proposed method is numerically investigated in detail. The main advantage of the proposed technique is in terms of flexibility offered to an independent system operator in selecting an appropriate solution of CBM simultaneously for all areas.

  6. First and higher order, heuristically based generalized perturbation theory (HGPT) with optional control reset variable

    International Nuclear Information System (INIS)

    Gandini, A.

    1996-01-01

    The heuristically based generalized perturbation theory (HGPT), to first and higher order, applied to the neutron field of a reactor system, is discussed in relation to the criticality reset procedure. This procedure is implicit within the GPT methodology, corresponding to the so called filtering of the importance function relevant to the neutron field from the fundamental mode contamination. It is common practice to use the so called ''lambda''-mode filter. In order to account for any possible reset option, a general definition is introduced of an intensive control variable (ρ) entering into the governing equations, and correspondingly a fundamental ρ-mode filtering of the importance function is defined, relevant to the real criticality reset (control) mechanism adopted. A simple example illustrates the need to take into account the correct filtering, so as to avoid significant inaccuracies in the sensitivity calculation results. The extension of this filtering technique to other functions entering into the GPT perturbative formulations at first and higher order is also discussed. (author)

  7. Ignorance- versus evidence-based decision making: a decision time analysis of the recognition heuristic.

    Science.gov (United States)

    Hilbig, Benjamin E; Pohl, Rüdiger F

    2009-09-01

    According to part of the adaptive toolbox notion of decision making known as the recognition heuristic (RH), the decision process in comparative judgments-and its duration-is determined by whether recognition discriminates between objects. By contrast, some recently proposed alternative models predict that choices largely depend on the amount of evidence speaking for each of the objects and that decision times thus depend on the evidential difference between objects, or the degree of conflict between options. This article presents 3 experiments that tested predictions derived from the RH against those from alternative models. All experiments used naturally recognized objects without teaching participants any information and thus provided optimal conditions for application of the RH. However, results supported the alternative, evidence-based models and often conflicted with the RH. Recognition was not the key determinant of decision times, whereas differences between objects with respect to (both positive and negative) evidence predicted effects well. In sum, alternative models that allow for the integration of different pieces of information may well provide a better account of comparative judgments. (c) 2009 APA, all rights reserved.

  8. Component-based development process and component lifecycle

    NARCIS (Netherlands)

    Crnkovic, I.; Chaudron, M.R.V.; Larsson, S.

    2006-01-01

    The process of component- and component-based system development differs in many significant ways from the "classical" development process of software systems. The main difference is in the separation of the development process of components from the development process of systems. This fact has a

  9. A CDT-Based Heuristic Zone Design Approach for Economic Census Investigators

    Directory of Open Access Journals (Sweden)

    Changixu Cheng

    2015-01-01

    Full Text Available This paper addresses a special zone design problem for economic census investigators that is motivated by a real-world application. This paper presented a heuristic multikernel growth approach via Constrained Delaunay Triangulation (CDT. This approach not only solved the barriers problem but also dealt with the polygon data in zoning procedure. In addition, it uses a new heuristic method to speed up the zoning process greatly on the premise of the required quality of zoning. At last, two special instances for economic census were performed, highlighting the performance of this approach.

  10. A similarity score-based two-phase heuristic approach to solve the dynamic cellular facility layout for manufacturing systems

    Science.gov (United States)

    Kumar, Ravi; Singh, Surya Prakash

    2017-11-01

    The dynamic cellular facility layout problem (DCFLP) is a well-known NP-hard problem. It has been estimated that the efficient design of DCFLP reduces the manufacturing cost of products by maintaining the minimum material flow among all machines in all cells, as the material flow contributes around 10-30% of the total product cost. However, being NP hard, solving the DCFLP optimally is very difficult in reasonable time. Therefore, this article proposes a novel similarity score-based two-phase heuristic approach to solve the DCFLP optimally considering multiple products in multiple times to be manufactured in the manufacturing layout. In the first phase of the proposed heuristic, a machine-cell cluster is created based on similarity scores between machines. This is provided as an input to the second phase to minimize inter/intracell material handling costs and rearrangement costs over the entire planning period. The solution methodology of the proposed approach is demonstrated. To show the efficiency of the two-phase heuristic approach, 21 instances are generated and solved using the optimization software package LINGO. The results show that the proposed approach can optimally solve the DCFLP in reasonable time.

  11. Heuristic guidelines and experimental evaluation of effective augmented-reality based instructions for maintenance in nuclear power plants

    International Nuclear Information System (INIS)

    Yim, Ho Bin; Seong, Poong Hyun

    2010-01-01

    Research highlights: → Augmented reality (AR) instructions were built for NPPs maintenance personnel. → 4-5 pieces of information at a time were optimum for AR instructions in this study. → A large variance in mode no. 5 implies these were also found to be critical amount. → Heuristic guidelines were suggested to make AR instructions more effective. - Abstract: As industrial plants and factories age, their maintenance requirements increase. Because maintenance mistakes directly increase the operating costs of a power plant, maintenance quality is significant concern to plant management. By law, all personnel working with nuclear technology must be re-trained every three years in Korea; however, as the statistical data show, the number of shutdown accidents at nuclear power plants (NPPs) due to maintenance failure is still high and needs to be reduced. Industries have started to adopt various technologies to increase the speed and accuracy of maintenance. Among those technologies, augmented reality (AR) is the latest multimedia presentation technology to be applied to plant maintenance, and it offers superior intuitiveness and user interactivity over other conventional multimedia. This empirical study aims to measure the optimum amounts of information to be delivered at a time and to identify what types of information enhance the learning ability of novices and to suggest heuristic guidelines by which to make effective AR training instructions. In the first experiment, the optimum amount of information in an AR learning environment for novices was found to be 4-5 pieces of information in a chunk by comparing results between a pre-test and an after-test. This result implies that intentionally made chunks help novices learn more effectively. In the second experiment, the AR training instruction based on the suggested heuristic guidelines was slightly more effective than other AR training instructions. Maintenance in nuclear power plants can be more reliable

  12. Improving the Bin Packing Heuristic through Grammatical Evolution Based on Swarm Intelligence

    Directory of Open Access Journals (Sweden)

    Marco Aurelio Sotelo-Figueroa

    2014-01-01

    Full Text Available In recent years Grammatical Evolution (GE has been used as a representation of Genetic Programming (GP which has been applied to many optimization problems such as symbolic regression, classification, Boolean functions, constructed problems, and algorithmic problems. GE can use a diversity of searching strategies including Swarm Intelligence (SI. Particle Swarm Optimisation (PSO is an algorithm of SI that has two main problems: premature convergence and poor diversity. Particle Evolutionary Swarm Optimization (PESO is a recent and novel algorithm which is also part of SI. PESO uses two perturbations to avoid PSO’s problems. In this paper we propose using PESO and PSO in the frame of GE as strategies to generate heuristics that solve the Bin Packing Problem (BPP; it is possible however to apply this methodology to other kinds of problems using another Grammar designed for that problem. A comparison between PESO, PSO, and BPP’s heuristics is performed through the nonparametric Friedman test. The main contribution of this paper is proposing a Grammar to generate online and offline heuristics depending on the test instance trying to improve the heuristics generated by other grammars and humans; it also proposes a way to implement different algorithms as search strategies in GE like PESO to obtain better results than those obtained by PSO.

  13. Example-Based Learning in Heuristic Domains: A Cognitive Load Theory Account

    Science.gov (United States)

    Renkl, Alexander; Hilbert, Tatjana; Schworm, Silke

    2009-01-01

    One classical instructional effect of cognitive load theory (CLT) is the worked-example effect. Although the vast majority of studies have focused on well-structured and algorithmic sub-domains of mathematics or physics, more recent studies have also analyzed learning with examples from complex domains in which only heuristic solution strategies…

  14. A Slicing Tree Representation and QCP-Model-Based Heuristic Algorithm for the Unequal-Area Block Facility Layout Problem

    Directory of Open Access Journals (Sweden)

    Mei-Shiang Chang

    2013-01-01

    Full Text Available The facility layout problem is a typical combinational optimization problem. In this research, a slicing tree representation and a quadratically constrained program model are combined with harmony search to develop a heuristic method for solving the unequal-area block layout problem. Because of characteristics of slicing tree structure, we propose a regional structure of harmony memory to memorize facility layout solutions and two kinds of harmony improvisation to enhance global search ability of the proposed heuristic method. The proposed harmony search based heuristic is tested on 10 well-known unequal-area facility layout problems from the literature. The results are compared with the previously best-known solutions obtained by genetic algorithm, tabu search, and ant system as well as exact methods. For problems O7, O9, vC10Ra, M11*, and Nug12, new best solutions are found. For other problems, the proposed approach can find solutions that are very similar to previous best-known solutions.

  15. Formalization in Component Based Development

    DEFF Research Database (Denmark)

    Holmegaard, Jens Peter; Knudsen, John; Makowski, Piotr

    2006-01-01

    We present a unifying conceptual framework for components, component interfaces, contracts and composition of components by focusing on the collection of properties or qualities that they must share. A specific property, such as signature, functionality behaviour or timing is an aspect. Each aspect...... may be specified in a formal language convenient for its purpose and, in principle, unrelated to languages for other aspects. Each aspect forms its own semantic domain, although a semantic domain may be parameterized by values derived from other aspects. The proposed conceptual framework is introduced...

  16. Familiarity and recollection in heuristic decision making.

    Science.gov (United States)

    Schwikert, Shane R; Curran, Tim

    2014-12-01

    Heuristics involve the ability to utilize memory to make quick judgments by exploiting fundamental cognitive abilities. In the current study we investigated the memory processes that contribute to the recognition heuristic and the fluency heuristic, which are both presumed to capitalize on the byproducts of memory to make quick decisions. In Experiment 1, we used a city-size comparison task while recording event-related potentials (ERPs) to investigate the potential contributions of familiarity and recollection to the 2 heuristics. ERPs were markedly different for recognition heuristic-based decisions and fluency heuristic-based decisions, suggesting a role for familiarity in the recognition heuristic and recollection in the fluency heuristic. In Experiment 2, we coupled the same city-size comparison task with measures of subjective preexperimental memory for each stimulus in the task. Although previous literature suggests the fluency heuristic relies on recognition speed alone, our results suggest differential contributions of recognition speed and recollected knowledge to these decisions, whereas the recognition heuristic relies on familiarity. Based on these results, we created a new theoretical framework that explains decisions attributed to both heuristics based on the underlying memory associated with the choice options. PsycINFO Database Record (c) 2014 APA, all rights reserved.

  17. Heuristic reasoning

    CERN Document Server

    2015-01-01

    How can we advance knowledge? Which methods do we need in order to make new discoveries? How can we rationally evaluate, reconstruct and offer discoveries as a means of improving the ‘method’ of discovery itself? And how can we use findings about scientific discovery to boost funding policies, thus fostering a deeper impact of scientific discovery itself? The respective chapters in this book provide readers with answers to these questions. They focus on a set of issues that are essential to the development of types of reasoning for advancing knowledge, such as models for both revolutionary findings and paradigm shifts; ways of rationally addressing scientific disagreement, e.g. when a revolutionary discovery sparks considerable disagreement inside the scientific community; frameworks for both discovery and inference methods; and heuristics for economics and the social sciences.

  18. Component protection based automatic control

    International Nuclear Information System (INIS)

    Otaduy, P.J.

    1992-01-01

    Control and safety systems as well as operation procedures are designed on the basis of critical process parameters limits. The expectation is that short and long term mechanical damage and process failures will be avoided by operating the plant within the specified constraints envelopes. In this paper, one of the Advanced Liquid Metal Reactor (ALMR) design duty cycles events is discussed to corroborate that the time has come to explicitly make component protection part of the control system. Component stress assessment and aging data should be an integral part of the control system. Then transient trajectory planning and operating limits could be aimed at minimizing component specific and overall plant component damage cost functions. The impact of transients on critical components could then be managed according to plant lifetime design goals. The need for developing methodologies for online transient trajectory planning and assessment of operating limits in order to facilitate the explicit incorporation of damage assessment capabilities to the plant control and protection systems is discussed. 12 refs

  19. Heuristic introduction to gravitational waves

    International Nuclear Information System (INIS)

    Sandberg, V.D.

    1982-01-01

    The purpose of this article is to provide a rough and somewhat heuristic theoretical background and introduction to gravitational radiation, its generation, and its detection based on Einstein's general theory of relativity

  20. Paranoid thinking as a heuristic.

    Science.gov (United States)

    Preti, Antonio; Cella, Matteo

    2010-08-01

    Paranoid thinking can be viewed as a human heuristic used by individuals to deal with uncertainty during stressful situations. Under stress, individuals are likely to emphasize the threatening value of neutral stimuli and increase the reliance on paranoia-based heuristic to interpreter events and guide their decisions. Paranoid thinking can also be activated by stress arising from the possibility of losing a good opportunity; this may result in an abnormal allocation of attentional resources to social agents. A better understanding of the interplay between cognitive heuristics and emotional processes may help to detect situations in which paranoid thinking is likely to exacerbate and improve intervention for individuals with delusional disorders.

  1. Heuristics for Hierarchical Partitioning with Application to Model Checking

    DEFF Research Database (Denmark)

    Möller, Michael Oliver; Alur, Rajeev

    2001-01-01

    Given a collection of connected components, it is often desired to cluster together parts of strong correspondence, yielding a hierarchical structure. We address the automation of this process and apply heuristics to battle the combinatorial and computational complexity. We define a cost function...... that captures the quality of a structure relative to the connections and favors shallow structures with a low degree of branching. Finding a structure with minimal cost is NP-complete. We present a greedy polynomial-time algorithm that approximates good solutions incrementally by local evaluation of a heuristic...... function. We argue for a heuristic function based on four criteria: the number of enclosed connections, the number of components, the number of touched connections and the depth of the structure. We report on an application in the context of formal verification, where our algorithm serves as a preprocessor...

  2. QuickVina: accelerating AutoDock Vina using gradient-based heuristics for global optimization.

    Science.gov (United States)

    Handoko, Stephanus Daniel; Ouyang, Xuchang; Su, Chinh Tran To; Kwoh, Chee Keong; Ong, Yew Soon

    2012-01-01

    Predicting binding between macromolecule and small molecule is a crucial phase in the field of rational drug design. AutoDock Vina, one of the most widely used docking software released in 2009, uses an empirical scoring function to evaluate the binding affinity between the molecules and employs the iterated local search global optimizer for global optimization, achieving a significantly improved speed and better accuracy of the binding mode prediction compared its predecessor, AutoDock 4. In this paper, we propose further improvement in the local search algorithm of Vina by heuristically preventing some intermediate points from undergoing local search. Our improved version of Vina-dubbed QVina-achieved a maximum acceleration of about 25 times with the average speed-up of 8.34 times compared to the original Vina when tested on a set of 231 protein-ligand complexes while maintaining the optimal scores mostly identical. Using our heuristics, larger number of different ligands can be quickly screened against a given receptor within the same time frame.

  3. Path-Wise Test Data Generation Based on Heuristic Look-Ahead Methods

    Directory of Open Access Journals (Sweden)

    Ying Xing

    2014-01-01

    Full Text Available Path-wise test data generation is generally considered an important problem in the automation of software testing. In essence, it is a constraint optimization problem, which is often solved by search methods such as backtracking algorithms. In this paper, the backtracking algorithm branch and bound and state space search in artificial intelligence are introduced to tackle the problem of path-wise test data generation. The former is utilized to explore the space of potential solutions and the latter is adopted to construct the search tree dynamically. Heuristics are employed in the look-ahead stage of the search. Dynamic variable ordering is presented with a heuristic rule to break ties, values of a variable are determined by the monotonicity analysis on branching conditions, and maintaining path consistency is achieved through analysis on the result of interval arithmetic. An optimization method is also proposed to reduce the search space. The results of empirical experiments show that the search is conducted in a basically backtrack-free manner, which ensures both test data generation with promising performance and its excellence over some currently existing static and dynamic methods in terms of coverage. The results also demonstrate that the proposed method is applicable in engineering.

  4. A component-based groupware development methodology

    NARCIS (Netherlands)

    Guareis de farias, Cléver; Ferreira Pires, Luis; van Sinderen, Marten J.

    2000-01-01

    Software development in general and groupware applications in particular can greatly benefit from the reusability and interoperability aspects associated with software components. Component-based software development enables the construction of software artefacts by assembling prefabricated,

  5. Heuristics in Conflict Resolution

    OpenAIRE

    Drescher, Christian; Gebser, Martin; Kaufmann, Benjamin; Schaub, Torsten

    2010-01-01

    Modern solvers for Boolean Satisfiability (SAT) and Answer Set Programming (ASP) are based on sophisticated Boolean constraint solving techniques. In both areas, conflict-driven learning and related techniques constitute key features whose application is enabled by conflict analysis. Although various conflict analysis schemes have been proposed, implemented, and studied both theoretically and practically in the SAT area, the heuristic aspects involved in conflict analysis have not yet receive...

  6. Heuristic Drift-based Model of the Power Scrape-off width in H-mode Tokamaks

    International Nuclear Information System (INIS)

    Goldston, Robert J.

    2011-01-01

    An heuristic model for the plasma scrape-off width in H-mode plasmas is introduced. Grad B and curv B drifts into the SOL are balanced against sonic parallel flows out of the SOL, to the divertor plates. The overall particle flow pattern posited is a modification for open field lines of Pfirsch-Shlueter flows to include sinks to the divertors. These assumptions result in an estimated SOL width of ∼ 2αρ p /R. They also result in a first-principles calculation of the particle confinement time of H-mode plasmas, qualitatively consistent with experimental observations. It is next assumed that anomalous perpendicular electron thermal diffusivity is the dominant source of heat flux across the separatrix, investing the SOL width, defined above, with heat from the main plasma. The separatrix temperature is calculated based on a two-point model balancing power input to the SOL with Spitzer-Haerm parallel thermal conduction losses to the divertor. This results in a heuristic closed-form prediction for the power scrape-off width that is in reasonable quantitative agreement both in absolute magnitude and in scaling with recent experimental data from deuterium plasmas. Further work should include full numerical calculations, including all magnetic and electric drifts, as well as more thorough comparison with experimental data.

  7. An Heuristic Drift-Based Model of the Power Scrape-Off Width in H-Mode Tokamaks

    International Nuclear Information System (INIS)

    Goldston, Robert J.

    2011-01-01

    An heuristic model for the plasma scrape-off width in H-mode plasmas is introduced. Grad B and curv B drifts into the SOL are balanced against sonic parallel flows out of the SOL, to the divertor plates. The overall mass flow pattern posited is a modification for open field lines of Pfirsch-Shlueter flows to include sinks to the divertors. These assumptions result in an estimated SOL width of 2αρ p /R. They also result in a first-principles calculation of the particle confinement time of H-mode plasmas, qualitatively consistent with experimental observations. It is next assumed that anomalous perpendicular electron thermal diffusivity is the dominant source of heat flux across the separatrix, investing the SOL width, defined above, with heat from the main plasma. The separatrix temperature is calculated based on a two-point model balancing power input to the SOL with Spitzer-Haerm parallel thermal conduction losses to the divertor. This results in an heuristic closed-form prediction for the power scrape-off width that is in remarkable quantitative agreement both in absolute magnitude and in scaling with recent experimental data. Further work should include full numerical calculations, including all magnetic and electric drifts, as well as more thorough comparison with experimental data.

  8. A subjective framework for seat comfort based on a heuristic multi criteria decision making technique and anthropometry.

    Science.gov (United States)

    Fazlollahtabar, Hamed

    2010-12-01

    Consumer expectations for automobile seat comfort continue to rise. With this said, it is evident that the current automobile seat comfort development process, which is only sporadically successful, needs to change. In this context, there has been growing recognition of the need for establishing theoretical and methodological automobile seat comfort. On the other hand, seat producer need to know the costumer's required comfort to produce based on their interests. The current research methodologies apply qualitative approaches due to anthropometric specifications. The most significant weakness of these approaches is the inexact extracted inferences. Despite the qualitative nature of the consumer's preferences there are some methods to transform the qualitative parameters into numerical value which could help seat producer to improve or enhance their products. Nonetheless this approach would help the automobile manufacturer to provide their seats from the best producer regarding to the consumers idea. In this paper, a heuristic multi criteria decision making technique is applied to make consumers preferences in the numeric value. This Technique is combination of Analytical Hierarchy Procedure (AHP), Entropy method, and Technique for Order Preference by Similarity to an Ideal Solution (TOPSIS). A case study is conducted to illustrate the applicability and the effectiveness of the proposed heuristic approach. Copyright © 2010 Elsevier Ltd. All rights reserved.

  9. A Meta-Heuristic Regression-Based Feature Selection for Predictive Analytics

    Directory of Open Access Journals (Sweden)

    Bharat Singh

    2014-11-01

    Full Text Available A high-dimensional feature selection having a very large number of features with an optimal feature subset is an NP-complete problem. Because conventional optimization techniques are unable to tackle large-scale feature selection problems, meta-heuristic algorithms are widely used. In this paper, we propose a particle swarm optimization technique while utilizing regression techniques for feature selection. We then use the selected features to classify the data. Classification accuracy is used as a criterion to evaluate classifier performance, and classification is accomplished through the use of k-nearest neighbour (KNN and Bayesian techniques. Various high dimensional data sets are used to evaluate the usefulness of the proposed approach. Results show that our approach gives better results when compared with other conventional feature selection algorithms.

  10. Neural bases for basic processes in heuristic problem solving: Take solving Sudoku puzzles as an example.

    Science.gov (United States)

    Qin, Yulin; Xiang, Jie; Wang, Rifeng; Zhou, Haiyan; Li, Kuncheng; Zhong, Ning

    2012-12-01

    Newell and Simon postulated that the basic steps in human problem-solving involve iteratively applying operators to transform the state of the problem to eventually achieve a goal. To check the neural basis of this framework, the present study focused on the basic processes in human heuristic problem-solving that the participants identified the current problem state and then recalled and applied the corresponding heuristic rules to change the problem state. A new paradigm, solving simplified Sudoku puzzles, was developed for an event-related functional magnetic resonance imaging (fMRI) study in problem solving. Regions of interest (ROIs), including the left prefrontal cortex, the bilateral posterior parietal cortex, the anterior cingulated cortex, the bilateral caudate nuclei, the bilateral fusiform, as well as the bilateral frontal eye fields, were found to be involved in the task. To obtain convergent evidence, in addition to traditional statistical analysis, we used the multivariate voxel classification method to check the accuracy of the predictions for the condition of the task from the blood oxygen level dependent (BOLD) response of the ROIs, using a new classifier developed in this study for fMRI data. To reveal the roles that the ROIs play in problem solving, we developed an ACT-R computational model of the information-processing processes in human problem solving, and tried to predict the BOLD response of the ROIs from the task. Advances in human problem-solving research after Newell and Simon are then briefly discussed. © 2012 The Institute of Psychology, Chinese Academy of Sciences and Blackwell Publishing Asia Pty Ltd.

  11. The memory state heuristic: A formal model based on repeated recognition judgments.

    Science.gov (United States)

    Castela, Marta; Erdfelder, Edgar

    2017-02-01

    The recognition heuristic (RH) theory predicts that, in comparative judgment tasks, if one object is recognized and the other is not, the recognized one is chosen. The memory-state heuristic (MSH) extends the RH by assuming that choices are not affected by recognition judgments per se, but by the memory states underlying these judgments (i.e., recognition certainty, uncertainty, or rejection certainty). Specifically, the larger the discrepancy between memory states, the larger the probability of choosing the object in the higher state. The typical RH paradigm does not allow estimation of the underlying memory states because it is unknown whether the objects were previously experienced or not. Therefore, we extended the paradigm by repeating the recognition task twice. In line with high threshold models of recognition, we assumed that inconsistent recognition judgments result from uncertainty whereas consistent judgments most likely result from memory certainty. In Experiment 1, we fitted 2 nested multinomial models to the data: an MSH model that formalizes the relation between memory states and binary choices explicitly and an approximate model that ignores the (unlikely) possibility of consistent guesses. Both models provided converging results. As predicted, reliance on recognition increased with the discrepancy in the underlying memory states. In Experiment 2, we replicated these results and found support for choice consistency predictions of the MSH. Additionally, recognition and choice latencies were in agreement with the MSH in both experiments. Finally, we validated critical parameters of our MSH model through a cross-validation method and a third experiment. (PsycINFO Database Record (c) 2017 APA, all rights reserved).

  12. A Genetic Algorithm-based Heuristic for Part-Feeding Mobile Robot Scheduling Problem

    DEFF Research Database (Denmark)

    Dang, Vinh Quang; Nielsen, Izabela Ewa; Bocewicz, Grzegorz

    2012-01-01

    This present study deals with the problem of sequencing feeding tasks of a single mobile robot with manipulation arm which is able to provide parts or components for feeders of machines in a manufacturing cell. The mobile robot has to be scheduled in order to keep machines within the cell producing...... products without any shortage of parts. A method based on the characteristics of feeders and inspired by the (s, Q) inventory system, is thus applied to define time windows for feeding tasks of the robot. The performance criterion is to minimize total traveling time of the robot in a given planning horizon...

  13. Heuristic knowledge-based planning for single-isocenter stereotactic radiosurgery to multiple brain metastases.

    Science.gov (United States)

    Ziemer, Benjamin P; Sanghvi, Parag; Hattangadi-Gluth, Jona; Moore, Kevin L

    2017-10-01

    plan quality improvements were quantified by calculating the difference between SRS quality metrics (QMs): ΔQM = QM clinical  - QM KBP . In addition to GM, investigated QMs were: volume of brain receiving ≥ 10 Gy (V 10 Gy ), volume of brain receiving ≥ 5 Gy (ΔV 5 Gy ), heterogeneity index (HI), dose to 0.1 cc of the brainstem (D 0.1 cc ), dose to 1% of the optic chiasm (D 1% ), and interlesion dose (D IL ). In addition to this quantitative analysis, overall plan quality was assessed via blinded plan comparison of the manual and KBP treatment plans by SRS-specializing physicians. A dose combination factor of n = 8 yielded an integrated dose profile RMS difference of 2.9% across the 41-patient cohort. Multimet dose predictions exhibited ΔGM = 0.07 ± 0.10 cm against the clinical sample, implying either further normal tissue sparing was possible or that dose predictions were slightly overestimating achievable dose gradients. The latter is the more likely explanation, as this bias vanished when dose predictions were converted to deliverable KBP plans ΔGM = 0.00 ± 0.08 cm. Remaining QMs were nearly identical or showed modest improvements in the KBP sample. Equivalent QMs included: ΔV 10 Gy  = 0.37 ± 3.78 cc, ΔHI = 0.02 ± 0.08 and ΔD IL  = -2.22 ± 171.4 cGy. The KBP plans showed a greater degree of normal tissue sparing as indicated by brain ΔV 5 Gy  = 4.11± 24.05 cc, brainstem ΔD 0.1 cc  = 42.8 ± 121.4 cGy, and chiasm ΔD 1%  = 50.8 ± 83.0 cGy. In blinded review by SRS-specializing physicians, KBP-generated plans were deemed equivalent or superior in 32/41(78.1%) of the cases. Heuristic KBP-driven automated planning in linac-based, single-isocenter treatments for multiple brain metastases maintained or exceeded overall plan quality. © 2017 American Association of Physicists in Medicine.

  14. Heuristic algorithms for joint optimization of unicast and anycast traffic in elastic optical network–based large–scale computing systems

    Directory of Open Access Journals (Sweden)

    Markowski Marcin

    2017-09-01

    Full Text Available In recent years elastic optical networks have been perceived as a prospective choice for future optical networks due to better adjustment and utilization of optical resources than is the case with traditional wavelength division multiplexing networks. In the paper we investigate the elastic architecture as the communication network for distributed data centers. We address the problems of optimization of routing and spectrum assignment for large-scale computing systems based on an elastic optical architecture; particularly, we concentrate on anycast user to data center traffic optimization. We assume that computational resources of data centers are limited. For this offline problems we formulate the integer linear programming model and propose a few heuristics, including a meta-heuristic algorithm based on a tabu search method. We report computational results, presenting the quality of approximate solutions and efficiency of the proposed heuristics, and we also analyze and compare some data center allocation scenarios.

  15. Graphene-based spintronic components

    OpenAIRE

    Zeng, Minggang; Shen, Lei; Su, Haibin; Zhou, Miao; Zhang, Chun; Feng, Yuanping

    2010-01-01

    A major challenge of spintronics is in generating, controlling and detecting spin-polarized current. Manipulation of spin-polarized current, in particular, is difficult. We demonstrate here, based on calculated transport properties of graphene nanoribbons, that nearly +-100% spin-polarized current can be generated in zigzag graphene nanoribbons (ZGNRs) and tuned by a source-drain voltage in the bipolar spin diode, in addition to magnetic configurations of the electrodes. This unusual transpor...

  16. The artifacts of component-based development

    International Nuclear Information System (INIS)

    Rizwan, M.; Qureshi, J.; Hayat, S.A.

    2007-01-01

    Component based development idea was floated in a conference name Mass Produced Software Components in 1968 (1). Since then engineering and scientific libraries are developed to reuse the previously developed functions. This concept is now widely used in SW development as component based development (CBD). Component-based software engineering (CBSE) is used to develop/ assemble software from existing components (2). Software developed using components is called component where (3). This paper presents different architectures of CBD such as Active X, common object request broker architecture (CORBA), remote method invocation (RMI) and simple object access protocol (SOAP). The overall objective of this paper is to support the practice of CBD by comparing its advantages and disadvantages. This paper also evaluates object oriented process model to adapt it for CBD. (author)

  17. Distance selection based on relevance feedback in the context of CBIR using the SFS meta-heuristic with one round

    Directory of Open Access Journals (Sweden)

    Mawloud Mosbah

    2017-03-01

    Full Text Available In this paper, we address the selection in the context of Content Based-Image Retrieval (CBIR. Instead of addressing features’ selection issue, we deal here with distance selection as a novel paradigm poorly addressed within CBIR field. Whereas distance concept is a very precise and sharp mathematical tool, we extend the study to weak distances: Similarity, quasi-distance, and divergence. Therefore, as many as eighteen (18 such measures as considered: distances: {Euclidian, …}, similarities{Ruzika, …}, quasi-distances: {Neyman-X2, …} and divergences: {Jeffrey, …}. We specifically propose a hybrid system based on the Sequential Forward Selector (SFS meta-heuristic with one round and relevance feedback. The experiments conducted on the Wang database (Corel-1K using color moments as a signature show that our system yields promising results in terms of effectiveness.

  18. Coupled Heuristic Prediction of Long Lead-Time Accumulated Total Inflow of a Reservoir during Typhoons Using Deterministic Recurrent and Fuzzy Inference-Based Neural Network

    Directory of Open Access Journals (Sweden)

    Chien-Lin Huang

    2015-11-01

    Full Text Available This study applies Real-Time Recurrent Learning Neural Network (RTRLNN and Adaptive Network-based Fuzzy Inference System (ANFIS with novel heuristic techniques to develop an advanced prediction model of accumulated total inflow of a reservoir in order to solve the difficulties of future long lead-time highly varied uncertainty during typhoon attacks while using a real-time forecast. For promoting the temporal-spatial forecasted precision, the following original specialized heuristic inputs were coupled: observed-predicted inflow increase/decrease (OPIID rate, total precipitation, and duration from current time to the time of maximum precipitation and direct runoff ending (DRE. This study also investigated the temporal-spatial forecasted error feature to assess the feasibility of the developed models, and analyzed the output sensitivity of both single and combined heuristic inputs to determine whether the heuristic model is susceptible to the impact of future forecasted uncertainty/errors. Validation results showed that the long lead-time–predicted accuracy and stability of the RTRLNN-based accumulated total inflow model are better than that of the ANFIS-based model because of the real-time recurrent deterministic routing mechanism of RTRLNN. Simulations show that the RTRLNN-based model with coupled heuristic inputs (RTRLNN-CHI, average error percentage (AEP/average forecast lead-time (AFLT: 6.3%/49 h can achieve better prediction than the model with non-heuristic inputs (AEP of RTRLNN-NHI and ANFIS-NHI: 15.2%/31.8% because of the full consideration of real-time hydrological initial/boundary conditions. Besides, the RTRLNN-CHI model can promote the forecasted lead-time above 49 h with less than 10% of AEP which can overcome the previous forecasted limits of 6-h AFLT with above 20%–40% of AEP.

  19. Toward a More Usable Home-Based Video Telemedicine System: A Heuristic Evaluation of the Clinician User Interfaces of Home-Based Video Telemedicine Systems.

    Science.gov (United States)

    Agnisarman, Sruthy; Narasimha, Shraddhaa; Chalil Madathil, Kapil; Welch, Brandon; Brinda, Fnu; Ashok, Aparna; McElligott, James

    2017-04-24

    Telemedicine is the use of technology to provide and support health care when distance separates the clinical service and the patient. Home-based telemedicine systems involve the use of such technology for medical support and care connecting the patient from the comfort of their homes with the clinician. In order for such a system to be used extensively, it is necessary to understand not only the issues faced by the patients in using them but also the clinician. The aim of this study was to conduct a heuristic evaluation of 4 telemedicine software platforms-Doxy.me, Polycom, Vidyo, and VSee-to assess possible problems and limitations that could affect the usability of the system from the clinician's perspective. It was found that 5 experts individually evaluated all four systems using Nielsen's list of heuristics, classifying the issues based on a severity rating scale. A total of 46 unique problems were identified by the experts. The heuristics most frequently violated were visibility of system status and Error prevention amounting to 24% (11/46 issues) each. Esthetic and minimalist design was second contributing to 13% (6/46 issues) of the total errors. Heuristic evaluation coupled with a severity rating scale was found to be an effective method for identifying problems with the systems. Prioritization of these problems based on the rating provides a good starting point for resolving the issues affecting these platforms. There is a need for better transparency and a more streamlined approach for how physicians use telemedicine systems. Visibility of the system status and speaking the users' language are keys for achieving this. ©Sruthy Agnisarman, Shraddhaa Narasimha, Kapil Chalil Madathil, Brandon Welch, FNU Brinda, Aparna Ashok, James McElligott. Originally published in JMIR Human Factors (http://humanfactors.jmir.org), 24.04.2017.

  20. A new Nawaz-Enscore-Ham-based heuristic for permutation flow-shop problems with bicriteria of makespan and machine idle time

    Science.gov (United States)

    Liu, Weibo; Jin, Yan; Price, Mark

    2016-10-01

    A new heuristic based on the Nawaz-Enscore-Ham algorithm is proposed in this article for solving a permutation flow-shop scheduling problem. A new priority rule is proposed by accounting for the average, mean absolute deviation, skewness and kurtosis, in order to fully describe the distribution style of processing times. A new tie-breaking rule is also introduced for achieving effective job insertion with the objective of minimizing both makespan and machine idle time. Statistical tests illustrate better solution quality of the proposed algorithm compared to existing benchmark heuristics.

  1. Event based neutron activation spectroscopy and analysis algorithm using MLE and meta-heuristics

    International Nuclear Information System (INIS)

    Wallace, B.

    2014-01-01

    Techniques used in neutron activation analysis are often dependent on the experimental setup. In the context of developing a portable and high efficiency detection array, good energy resolution and half-life discrimination are difficult to obtain with traditional methods given the logistic and financial constraints. An approach different from that of spectrum addition and standard spectroscopy analysis was needed. The use of multiple detectors prompts the need for a flexible storage of acquisition data to enable sophisticated post processing of information. Analogously to what is done in heavy ion physics, gamma detection counts are stored as two-dimensional events. This enables post-selection of energies and time frames without the need to modify the experimental setup. This method of storage also permits the use of more complex analysis tools. Given the nature of the problem at hand, a light and efficient analysis code had to be devised. A thorough understanding of the physical and statistical processes involved was used to create a statistical model. Maximum likelihood estimation was combined with meta-heuristics to produce a sophisticated curve-fitting algorithm. Simulated and experimental data were fed into the analysis code prompting positive results in terms of half-life discrimination, peak identification and noise reduction. The code was also adapted to other fields of research such as heavy ion identification of the quasi-target (QT) and quasi-particle (QP). The approach used seems to be able to translate well into other fields of research. (author)

  2. A heuristic-based approach for reliability importance assessment of energy producers

    International Nuclear Information System (INIS)

    Akhavein, A.; Fotuhi Firuzabad, M.

    2011-01-01

    Reliability of energy supply is one of the most important issues of service quality. On one hand, customers usually have different expectations for service reliability and price. On the other hand, providing different level of reliability at load points is a challenge for system operators. In order to take reasonable decisions and obviate reliability implementation difficulties, market players need to know impacts of their assets on system and load-point reliabilities. One tool to specify reliability impacts of assets is the criticality or reliability importance measure by which system components can be ranked based on their effect on reliability. Conventional methods for determination of reliability importance are essentially on the basis of risk sensitivity analysis and hence, impose prohibitive calculation burden in large power systems. An approach is proposed in this paper to determine reliability importance of energy producers from perspective of consumers or distribution companies in a composite generation and transmission system. In the presented method, while avoiding immense computational burden, the energy producers are ranked based on their rating, unavailability and impact on power flows in the lines connecting to the considered load points. Study results on the IEEE reliability test system show successful application of the proposed method. - Research highlights: → Required reliability level at load points is a concern in modern power systems. → It is important to assess reliability importance of energy producers or generators. → Generators can be ranked based on their impacts on power flow to a selected area. → Ranking of generators is an efficient tool to assess their reliability importance.

  3. Characterizing web heuristics

    NARCIS (Netherlands)

    de Jong, Menno D.T.; van der Geest, Thea

    2000-01-01

    This article is intended to make Web designers more aware of the qualities of heuristics by presenting a framework for analyzing the characteristics of heuristics. The framework is meant to support Web designers in choosing among alternative heuristics. We hope that better knowledge of the

  4. Laser based refurbishment of steel mill components

    CSIR Research Space (South Africa)

    Kazadi, P

    2006-03-01

    Full Text Available Laser refurbishment capabilities were demonstrated and promising results were obtained for repair of distance sleeves, foot rolls, descaler cassette, idler rolls. Based on the cost projections and the results of the in-situ testing, components which...

  5. Structured Performance Analysis for Component Based Systems

    OpenAIRE

    Salmi , N.; Moreaux , Patrice; Ioualalen , M.

    2012-01-01

    International audience; The Component Based System (CBS) paradigm is now largely used to design software systems. In addition, performance and behavioural analysis remains a required step for the design and the construction of efficient systems. This is especially the case of CBS, which involve interconnected components running concurrent processes. % This paper proposes a compositional method for modeling and structured performance analysis of CBS. Modeling is based on Stochastic Well-formed...

  6. Using the Science Writing Heuristic To Move toward an Inquiry-Based Laboratory Curriculum: An Example from Physical Equilibrium.

    Science.gov (United States)

    Rudd, James A., II; Greenbowe, Thomas J.; Hand, Brian M.; Legg, Margaret J.

    2001-01-01

    Investigates the effects of the Science Writing Heuristic (SWH) format on student's achievement, thinking abilities and motivation. Focuses on distribution equilibrium and assesses student understanding by studying metacognitive and practical factors. (Contains 17 references.) (Author/YDS)

  7. Heuristic decision making.

    Science.gov (United States)

    Gigerenzer, Gerd; Gaissmaier, Wolfgang

    2011-01-01

    As reflected in the amount of controversy, few areas in psychology have undergone such dramatic conceptual changes in the past decade as the emerging science of heuristics. Heuristics are efficient cognitive processes, conscious or unconscious, that ignore part of the information. Because using heuristics saves effort, the classical view has been that heuristic decisions imply greater errors than do "rational" decisions as defined by logic or statistical models. However, for many decisions, the assumptions of rational models are not met, and it is an empirical rather than an a priori issue how well cognitive heuristics function in an uncertain world. To answer both the descriptive question ("Which heuristics do people use in which situations?") and the prescriptive question ("When should people rely on a given heuristic rather than a complex strategy to make better judgments?"), formal models are indispensable. We review research that tests formal models of heuristic inference, including in business organizations, health care, and legal institutions. This research indicates that (a) individuals and organizations often rely on simple heuristics in an adaptive way, and (b) ignoring part of the information can lead to more accurate judgments than weighting and adding all information, for instance for low predictability and small samples. The big future challenge is to develop a systematic theory of the building blocks of heuristics as well as the core capacities and environmental structures these exploit.

  8. Heuristic Environments

    Directory of Open Access Journals (Sweden)

    Ines Giunta

    2011-10-01

    Full Text Available The purpose of this research is the identification of a paradigm which fixes the basic concepts and the type of logical relationships between them, whereby direct, govern and evaluate choises on new technologies. The contribution is based on the assumption that the complexity of knowledge is correlated with the complexity of the learning environment. From the existence of this correlation will descend a series of consequences that contribute to the definition of a theoretical construct in which the logical categories of learning become the guiding criteria on which to design learning environments and, consequently, also the indicators on by which to evaluate its effectiveness.

  9. Comparison of Heuristics for Inhibitory Rule Optimization

    KAUST Repository

    Alsolami, Fawaz; Chikalov, Igor; Moshkov, Mikhail

    2014-01-01

    Friedman test with Nemenyi post-hoc are used to compare the greedy algorithms statistically against each other for length and coverage. The experiments are carried out on real datasets from UCI Machine Learning Repository. For leading heuristics, the constructed rules are compared with optimal ones obtained based on dynamic programming approach. The results seem to be promising for the best heuristics: the average relative difference between length (coverage) of constructed and optimal rules is at most 2.27% (7%, respectively). Furthermore, the quality of classifiers based on sets of inhibitory rules constructed by the considered heuristics are compared against each other, and the results show that the three best heuristics from the point of view classification accuracy coincides with the three well-performed heuristics from the point of view of rule length minimization.

  10. Modeling reproductive decisions with simple heuristics

    Directory of Open Access Journals (Sweden)

    Peter Todd

    2013-10-01

    Full Text Available BACKGROUND Many of the reproductive decisions that humans make happen without much planning or forethought, arising instead through the use of simple choice rules or heuristics that involve relatively little information and processing. Nonetheless, these heuristic-guided decisions are typically beneficial, owing to humans' ecological rationality - the evolved fit between our constrained decision mechanisms and the adaptive problems we face. OBJECTIVE This paper reviews research on the ecological rationality of human decision making in the domain of reproduction, showing how fertility-related decisions are commonly made using various simple heuristics matched to the structure of the environment in which they are applied, rather than being made with information-hungry mechanisms based on optimization or rational economic choice. METHODS First, heuristics for sequential mate search are covered; these heuristics determine when to stop the process of mate search by deciding that a good-enough mate who is also mutually interested has been found, using a process of aspiration-level setting and assessing. These models are tested via computer simulation and comparison to demographic age-at-first-marriage data. Next, a heuristic process of feature-based mate comparison and choice is discussed, in which mate choices are determined by a simple process of feature-matching with relaxing standards over time. Parental investment heuristics used to divide resources among offspring are summarized. Finally, methods for testing the use of such mate choice heuristics in a specific population over time are then described.

  11. Prediction of retention in micellar electrokinetic chromatography based on molecular structural descriptors by using the heuristic method

    International Nuclear Information System (INIS)

    Liu Huanxiang; Yao Xiaojun; Liu Mancang; Hu Zhide; Fan Botao

    2006-01-01

    Based on calculated molecular descriptors from the solutes' structure alone, the micelle-water partition coefficients of 103 solutes in micellar electrokinetic chromatography (MEKC) were predicted using the heuristic method (HM). At the same time, in order to show the influence of different molecular descriptors on the micelle-water partition of solute and to well understand the retention mechanism in MEKC, HM was used to build several multivariable linear models using different numbers of molecular descriptors. The best 6-parameter model gave the following results: the square of correlation coefficient R 2 was 0.958 and the mean relative error was 3.98%, which proved that the predictive values were in good agreement with the experimental results. From the built model, it can be concluded that the hydrophobic, H-bond, polar interactions of solutes with the micellar and aqueous phases are the main factors that determine their partitioning behavior. In addition, this paper provided a simple, fast and effective method for predicting the retention of the solutes in MEKC from their structures and gave some insight into structural features related to the retention of the solutes

  12. Methodology based on genetic heuristics for in-vivo characterizing the patient-specific biomechanical behavior of the breast tissues.

    Science.gov (United States)

    Lago, M A; Rúperez, M J; Martínez-Martínez, F; Martínez-Sanchis, S; Bakic, P R; Monserrat, C

    2015-11-30

    This paper presents a novel methodology to in-vivo estimate the elastic constants of a constitutive model proposed to characterize the mechanical behavior of the breast tissues. An iterative search algorithm based on genetic heuristics was constructed to in-vivo estimate these parameters using only medical images, thus avoiding invasive measurements of the mechanical response of the breast tissues. For the first time, a combination of overlap and distance coefficients were used for the evaluation of the similarity between a deformed MRI of the breast and a simulation of that deformation. The methodology was validated using breast software phantoms for virtual clinical trials, compressed to mimic MRI-guided biopsies. The biomechanical model chosen to characterize the breast tissues was an anisotropic neo-Hookean hyperelastic model. Results from this analysis showed that the algorithm is able to find the elastic constants of the constitutive equations of the proposed model with a mean relative error of about 10%. Furthermore, the overlap between the reference deformation and the simulated deformation was of around 95% showing the good performance of the proposed methodology. This methodology can be easily extended to characterize the real biomechanical behavior of the breast tissues, which means a great novelty in the field of the simulation of the breast behavior for applications such as surgical planing, surgical guidance or cancer diagnosis. This reveals the impact and relevance of the presented work.

  13. New heuristic method for the optimization of product mix based on theory of constraints and knapsack problem

    Directory of Open Access Journals (Sweden)

    Vinicius Amorim Sobreiro

    2013-06-01

    Full Text Available The definition of the product mix provides the allocation of the productive resources in the manufacture process and the optimization of productive system. However, the definition of the product mix is a problem of the NP-complete, in other words, of difficult solution. Taking this into account, with the aid of the Theory of Constraints - TOC, some constructive heuristics have been presented to help to solve this problem. Thus, the objective in this paper is to propose a new heuristics to provide better solutions when compared with the main heuristics presented in the literature, TOC-h of Fredendall and Lea. To accomplish this comparison, simulations were accomplished with the objective of identifying the production mix with the best throughput, considering CPU time and the characteristics of the productive ambient. The results show that the heuristics proposal was more satisfactory when compared to TOC-h and it shows good solution when compared with the optimum solution. This fact evidence the importance of the heuristics proposal in the definition of product mix.

  14. A Variable-Selection Heuristic for K-Means Clustering.

    Science.gov (United States)

    Brusco, Michael J.; Cradit, J. Dennis

    2001-01-01

    Presents a variable selection heuristic for nonhierarchical (K-means) cluster analysis based on the adjusted Rand index for measuring cluster recovery. Subjected the heuristic to Monte Carlo testing across more than 2,200 datasets. Results indicate that the heuristic is extremely effective at eliminating masking variables. (SLD)

  15. HEURISTICS IN DECISION MAKING

    Directory of Open Access Journals (Sweden)

    I PUTU SUDANA

    2011-01-01

    Full Text Available Prinsip heuristics tidak dapat dikatakan sebagai sebuah pendekatanpengambilan keputusan yang non-rasional, karena penerapan atau penggunaanyang unconscious atau subtle mind tidak dapat dianggap sebagai tindakanyang irrational. Dengan alasan tersebut, terdapat cukup alasan untukmenyatakan bahwa pengklasifikasian pendekatan-pendekatan keputusansemestinya menggunakan terminologi analytical dan experiential, dan bukanmemakai istilah rational dan non-rational seperti yang umumnya diikuti.Penerapan pendekatan heuristics dapat ditemukan pada berbagai disiplin,termasuk bisnis dan akuntansi. Topik heuristics semestinya mendapatperhatian yang cukup luas dari para periset di bidang akuntansi. Bidangbehavioral research in accounting menawarkan banyak kemungkinan untukdikaji, karena prinsip heuristics bertautan erat dengan aspek manusia sebagaipelaku dalam pengambilan keputusan.

  16. Component-Based Cartoon Face Generation

    Directory of Open Access Journals (Sweden)

    Saman Sepehri Nejad

    2016-11-01

    Full Text Available In this paper, we present a cartoon face generation method that stands on a component-based facial feature extraction approach. Given a frontal face image as an input, our proposed system has the following stages. First, face features are extracted using an extended Active Shape Model. Outlines of the components are locally modified using edge detection, template matching and Hermit interpolation. This modification enhances the diversity of output and accuracy of the component matching required for cartoon generation. Second, to bring cartoon-specific features such as shadows, highlights and, especially, stylish drawing, an array of various face photographs and corresponding hand-drawn cartoon faces are collected. These cartoon templates are automatically decomposed into cartoon components using our proposed method for parameterizing cartoon samples, which is fast and simple. Then, using shape matching methods, the appropriate cartoon component is selected and deformed to fit the input face. Finally, a cartoon face is rendered in a vector format using the rendering rules of the selected template. Experimental results demonstrate effectiveness of our approach in generating life-like cartoon faces.

  17. Selection of components based on their importance

    International Nuclear Information System (INIS)

    Stvan, F.

    2004-12-01

    A proposal is presented for sorting components of the Dukovany nuclear power plant with respect to their importance. The classification scheme includes property priority, property criticality and property structure. Each area has its criteria with weight coefficients to calculate the importance of each component by the Risk Priority Number method. The aim of the process is to generate a list of components in order of operating and safety importance, which will help spend funds to ensure operation and safety in an optimal manner. This proposal is linked to a proposal for a simple database which should serve to enter information and perform assessments. The present stage focused on a safety assessment of components categorized in safety classes BT1, BT2 and BT3 pursuant to Decree No. 76. Assessment was performed based ona PSE study for Level 1. The database includes inputs for entering financial data, which are represented by a potential damage resulting from the given failure and by the loss of MWh in financial terms. In a next input, the failure incidence intensity and time of correction can be entered. Information regarding the property structure, represented by the degree of backup and reparability of the component, is the last input available

  18. Heuristic Synthesis of Reversible Logic – A Comparative Study

    Directory of Open Access Journals (Sweden)

    Chua Shin Cheng

    2014-01-01

    Full Text Available Reversible logic circuits have been historically motivated by theoretical research in low-power, and recently attracted interest as components of the quantum algorithm, optical computing and nanotechnology. However due to the intrinsic property of reversible logic, traditional irreversible logic design and synthesis methods cannot be carried out. Thus a new set of algorithms are developed correctly to synthesize reversible logic circuit. This paper presents a comprehensive literature review with comparative study on heuristic based reversible logic synthesis. It reviews a range of heuristic based reversible logic synthesis techniques reported by researchers (BDD-based, cycle-based, search-based, non-search-based, rule-based, transformation-based, and ESOP-based. All techniques are described in detail and summarized in a table based on their features, limitation, library used and their consideration metric. Benchmark comparison of gate count and quantum cost are analysed for each synthesis technique. Comparing the synthesis algorithm outputs over the years, it can be observed that different approach has been used for the synthesis of reversible circuit. However, the improvements are not significant. Quantum cost and gate count has improved over the years, but arguments and debates are still on certain issues such as the issue of garbage outputs that remain the same. This paper provides the information of all heuristic based synthesis of reversible logic method proposed over the years. All techniques are explained in detail and thus informative for new reversible logic researchers and bridging the knowledge gap in this area.

  19. Comparison of Heuristics for Inhibitory Rule Optimization

    KAUST Repository

    Alsolami, Fawaz

    2014-09-13

    Knowledge representation and extraction are very important tasks in data mining. In this work, we proposed a variety of rule-based greedy algorithms that able to obtain knowledge contained in a given dataset as a series of inhibitory rules containing an expression “attribute ≠ value” on the right-hand side. The main goal of this paper is to determine based on rule characteristics, rule length and coverage, whether the proposed rule heuristics are statistically significantly different or not; if so, we aim to identify the best performing rule heuristics for minimization of rule length and maximization of rule coverage. Friedman test with Nemenyi post-hoc are used to compare the greedy algorithms statistically against each other for length and coverage. The experiments are carried out on real datasets from UCI Machine Learning Repository. For leading heuristics, the constructed rules are compared with optimal ones obtained based on dynamic programming approach. The results seem to be promising for the best heuristics: the average relative difference between length (coverage) of constructed and optimal rules is at most 2.27% (7%, respectively). Furthermore, the quality of classifiers based on sets of inhibitory rules constructed by the considered heuristics are compared against each other, and the results show that the three best heuristics from the point of view classification accuracy coincides with the three well-performed heuristics from the point of view of rule length minimization.

  20. Scientific data base for safeguards components

    International Nuclear Information System (INIS)

    Hall, R.C.; Jones, R.D.

    1978-01-01

    The need to store and maintain vast amounts of data and the desire to avoid nonfunctional redundancy have provided an impetus for modern data base technology. Large-scale data base management systems (DBMS) have emerged during the past two decades evolving from earlier generalized file processing systems. This evolution has primarily involved certain business applications (e.g., production control, payroll, order entry) because of their high volume data processing characterization. Current data base technology, however, is becoming increasingly concerned with generality. Many diverse applications, including scientific ones, are benefiting from the generalized data base management software which has resulted. The concept of a data base management system is examined. The three common models which have been proposed for organizing data and relationships are identified: the network model, the hierarchical model, and the relational model. A specific implementation using a hierarchical data base management system is described. This is the data base for safeguards components which has been developed at Sandia Laboratories using the System 2000 developed by MRI Systems Corporation. Its organization, components, and functions are presented. The various interfaces it permits to user programs (e.g., safeguards automated facility evaluation software) and interactive terminal users are described

  1. A survey on bio inspired meta heuristic based clustering protocols for wireless sensor networks

    Science.gov (United States)

    Datta, A.; Nandakumar, S.

    2017-11-01

    Recent studies have shown that utilizing a mobile sink to harvest and carry data from a Wireless Sensor Network (WSN) can improve network operational efficiency as well as maintain uniform energy consumption by the sensor nodes in the network. Due to Sink mobility, the path between two sensor nodes continuously changes and this has a profound effect on the operational longevity of the network and a need arises for a protocol which utilizes minimal resources in maintaining routes between the mobile sink and the sensor nodes. Swarm Intelligence based techniques inspired by the foraging behavior of ants, termites and honey bees can be artificially simulated and utilized to solve real wireless network problems. The author presents a brief survey on various bio inspired swarm intelligence based protocols used in routing data in wireless sensor networks while outlining their general principle and operation.

  2. A Heuristic Model for the Active Galactic Nucleus Based on the Planck Vacuum Theory

    Directory of Open Access Journals (Sweden)

    Daywitt W. C.

    2009-07-01

    Full Text Available The standard explanation for an active galactic nucleus (AGN is a "central engine" consisting of a hot accretion disk surrounding a supermassive black hole. Energy is generated by the gravitational infall of material which is heated to high temperatures in this dissipative accretion disk. What follows is an alternative model for the AGN based on the Planck vacuum (PV theory, where both the energy of the AGN and its variable luminosity are explained in terms of a variable photon flux emanating from the PV.

  3. A Heuristic Approach to Proposal-Based Negotiation: with Applications in Fashion Supply Chain Management

    Directory of Open Access Journals (Sweden)

    Stefania Costantini

    2013-01-01

    multi-issue negotiation between two autonomous competitive software agents proposed by Cadoli. This model is based on the view of negotiation spaces (or “areas”, representing the admissible values of the goods involved in the process as convex regions. However, in order to speed up the negotiation process and guarantee convergence, there was the restriction of potential agreements to vertices included in the intersection of the two areas. We present and assess experimentally an extension to Cadoli's approach where, for both participating agents, interaction is no longer vertex based, or at least not necessarily so. This eliminates the asymmetry among parties and the limitation to polyhedral negotiation areas. The extension can be usefully integrated to Cadoli's framework, thus obtaining an enhanced algorithm that can be effective in many practical cases. We present and discuss a number of experiments, aimed at assessing how parameters influence the performance of the algorithm and how they relate to each other. We discuss the usefulness of the approach in relevant application fields, such as, for instance, supply chain management in the fashion industry, which is a field of growing importance in economy and e-commerce.

  4. Process optimization of a non-circular drawing sequence based on multi-surrogate assisted meta-heuristic algorithms

    Energy Technology Data Exchange (ETDEWEB)

    Pholdee, Nantiwat; Bureerat, Su Jin [Khon Kaen University, Khon Kaen (Thailand); Baek, Hyun Moo [DTaQ, Changwon (Korea, Republic of); Im, Yong Taek [KAIST, Daejeon (Korea, Republic of)

    2015-08-15

    Process optimization of a Non-circular drawing (NCD) sequence of a pearlitic steel wire was performed to improve the mechanical properties of a drawn wire based on surrogate assisted meta-heuristic algorithms. The objective function was introduced to minimize inhomogeneity of effective strain distribution at the cross-section of the drawn wire, which could deteriorate delamination characteristics of the drawn wires. The design variables introduced were die geometry and reduction of area of the NCD sequence. Several surrogate models and their combinations with the weighted sum technique were utilized. In the process optimization of the NCD sequence, the surrogate models were used to predict effective strain distributions at the cross-section of the drawn wire. Optimization using Differential evolution (DE) algorithm was performed, while the objective function was calculated from the predicted effective strains. The accuracy of all surrogate models was investigated, while optimum results were compared with the previous study available in the literature. It was found that hybrid surrogate models can improve prediction accuracy compared to a single surrogate model. The best result was obtained from the combination of Kriging (KG) and Support vector regression (SVR) models, while the second best was obtained from the combination of four surrogate models: Polynomial response surface (PRS), Radial basic function (RBF), KG, and SVR. The optimum results found in this study showed better effective strain homogeneity at the cross-section of the drawn wire with the same total reduction of area of the previous work available in the literature for fewer number of passes. The multi-surrogate models with the weighted sum technique were found to be powerful in improving the delamination characteristics of the drawn wire and reducing the production cost.

  5. Designing Visual Aids That Promote Risk Literacy: A Systematic Review of Health Research and Evidence-Based Design Heuristics.

    Science.gov (United States)

    Garcia-Retamero, Rocio; Cokely, Edward T

    2017-06-01

    Background Effective risk communication is essential for informed decision making. Unfortunately, many people struggle to understand typical risk communications because they lack essential decision-making skills. Objective The aim of this study was to review the literature on the effect of numeracy on risk literacy, decision making, and health outcomes, and to evaluate the benefits of visual aids in risk communication. Method We present a conceptual framework describing the influence of numeracy on risk literacy, decision making, and health outcomes, followed by a systematic review of the benefits of visual aids in risk communication for people with different levels of numeracy and graph literacy. The systematic review covers scientific research published between January 1995 and April 2016, drawn from the following databases: Web of Science, PubMed, PsycINFO, ERIC, Medline, and Google Scholar. Inclusion criteria were investigation of the effect of numeracy and/or graph literacy, and investigation of the effect of visual aids or comparison of their effect with that of numerical information. Thirty-six publications met the criteria, providing data on 27,885 diverse participants from 60 countries. Results Transparent visual aids robustly improved risk understanding in diverse individuals by encouraging thorough deliberation, enhancing cognitive self-assessment, and reducing conceptual biases in memory. Improvements in risk understanding consistently produced beneficial changes in attitudes, behavioral intentions, trust, and healthy behaviors. Visual aids were found to be particularly beneficial for vulnerable and less skilled individuals. Conclusion Well-designed visual aids tend to be highly effective tools for improving informed decision making among diverse decision makers. We identify five categories of practical, evidence-based guidelines for heuristic evaluation and design of effective visual aids.

  6. A Simulated Annealing-Based Heuristic Algorithm for Job Shop Scheduling to Minimize Lateness

    Directory of Open Access Journals (Sweden)

    Rui Zhang

    2013-04-01

    Full Text Available A decomposition-based optimization algorithm is proposed for solving large job shop scheduling problems with the objective of minimizing the maximum lateness. First, we use the constraint propagation theory to derive the orientation of a portion of disjunctive arcs. Then we use a simulated annealing algorithm to find a decomposition policy which satisfies the maximum number of oriented disjunctive arcs. Subsequently, each subproblem (corresponding to a subset of operations as determined by the decomposition policy is successively solved with a simulated annealing algorithm, which leads to a feasible solution to the original job shop scheduling problem. Computational experiments are carried out for adapted benchmark problems, and the results show the proposed algorithm is effective and efficient in terms of solution quality and time performance.

  7. Assessing the use of cognitive heuristic representativeness in clinical reasoning.

    Science.gov (United States)

    Payne, Velma L; Crowley, Rebecca S; Crowley, Rebecca

    2008-11-06

    We performed a pilot study to investigate use of the cognitive heuristic Representativeness in clinical reasoning. We tested a set of tasks and assessments to determine whether subjects used the heuristics in reasoning, to obtain initial frequencies of heuristic use and related cognitive errors, and to collect cognitive process data using think-aloud techniques. The study investigates two aspects of the Representativeness heuristic - judging by perceived frequency and representativeness as causal beliefs. Results show that subjects apply both aspects of the heuristic during reasoning, and make errors related to misapplication of these heuristics. Subjects in this study rarely used base rates, showed significant variability in their recall of base rates, demonstrated limited ability to use provided base rates, and favored causal data in diagnosis. We conclude that the tasks and assessments we have developed provide a suitable test-bed to study the cognitive processes underlying heuristic errors.

  8. Assessing Use of Cognitive Heuristic Representativeness in Clinical Reasoning

    Science.gov (United States)

    Payne, Velma L.; Crowley, Rebecca S.

    2008-01-01

    We performed a pilot study to investigate use of the cognitive heuristic Representativeness in clinical reasoning. We tested a set of tasks and assessments to determine whether subjects used the heuristics in reasoning, to obtain initial frequencies of heuristic use and related cognitive errors, and to collect cognitive process data using think-aloud techniques. The study investigates two aspects of the Representativeness heuristic - judging by perceived frequency and representativeness as causal beliefs. Results show that subjects apply both aspects of the heuristic during reasoning, and make errors related to misapplication of these heuristics. Subjects in this study rarely used base rates, showed significant variability in their recall of base rates, demonstrated limited ability to use provided base rates, and favored causal data in diagnosis. We conclude that the tasks and assessments we have developed provide a suitable test-bed to study the cognitive processes underlying heuristic errors. PMID:18999140

  9. A comparison of heuristic and model-based clustering methods for dietary pattern analysis.

    Science.gov (United States)

    Greve, Benjamin; Pigeot, Iris; Huybrechts, Inge; Pala, Valeria; Börnhorst, Claudia

    2016-02-01

    Cluster analysis is widely applied to identify dietary patterns. A new method based on Gaussian mixture models (GMM) seems to be more flexible compared with the commonly applied k-means and Ward's method. In the present paper, these clustering approaches are compared to find the most appropriate one for clustering dietary data. The clustering methods were applied to simulated data sets with different cluster structures to compare their performance knowing the true cluster membership of observations. Furthermore, the three methods were applied to FFQ data assessed in 1791 children participating in the IDEFICS (Identification and Prevention of Dietary- and Lifestyle-Induced Health Effects in Children and Infants) Study to explore their performance in practice. The GMM outperformed the other methods in the simulation study in 72 % up to 100 % of cases, depending on the simulated cluster structure. Comparing the computationally less complex k-means and Ward's methods, the performance of k-means was better in 64-100 % of cases. Applied to real data, all methods identified three similar dietary patterns which may be roughly characterized as a 'non-processed' cluster with a high consumption of fruits, vegetables and wholemeal bread, a 'balanced' cluster with only slight preferences of single foods and a 'junk food' cluster. The simulation study suggests that clustering via GMM should be preferred due to its higher flexibility regarding cluster volume, shape and orientation. The k-means seems to be a good alternative, being easier to use while giving similar results when applied to real data.

  10. [Technique-based game for daycare visitors with and without dementia : Effects, heuristics and correlates].

    Science.gov (United States)

    Ehret, S; Putze, F; Miller-Teynor, H; Kruse, A; Schultz, T

    2017-01-01

    Playing of old people with or without dementia have not yet been substantially investigated. This study deals with the acceptance and impact of a tablet-based memory game, which was played on a weekly or semiweekly basis by visitors in two daycare units. Within the framework of focus groups the technical system was adapted for elderly users. The video-assisted data at the level of the game and the dynamics were investigated with respect to interaction and communication. The analysis of psychological observation forms and game protocols, which were conducted over a period of 3 months, indicated different effects of the game on psychosocial and cognitive activation. The individual memory cards in particular served as an intensification of communication and a stimulation of episodic memory. Finally, with video analysis during the whole game setting three theoretical relationship patterns of the spheres playing and speech could be depicted. Coherence, separation and incoherence of playing and speech are different forms of interaction in which individual and collaborative competences of people with and without dementia can be visualized. Furthermore, the study provides evidence for the cultural theory of playing by Huizinga.

  11. Automatic Generation of Heuristics for Scheduling

    Science.gov (United States)

    Morris, Robert A.; Bresina, John L.; Rodgers, Stuart M.

    1997-01-01

    This paper presents a technique, called GenH, that automatically generates search heuristics for scheduling problems. The impetus for developing this technique is the growing consensus that heuristics encode advice that is, at best, useful in solving most, or typical, problem instances, and, at worst, useful in solving only a narrowly defined set of instances. In either case, heuristic problem solvers, to be broadly applicable, should have a means of automatically adjusting to the idiosyncrasies of each problem instance. GenH generates a search heuristic for a given problem instance by hill-climbing in the space of possible multi-attribute heuristics, where the evaluation of a candidate heuristic is based on the quality of the solution found under its guidance. We present empirical results obtained by applying GenH to the real world problem of telescope observation scheduling. These results demonstrate that GenH is a simple and effective way of improving the performance of an heuristic scheduler.

  12. A decomposition heuristics based on multi-bottleneck machines for large-scale job shop scheduling problems

    Directory of Open Access Journals (Sweden)

    Yingni Zhai

    2014-10-01

    Full Text Available Purpose: A decomposition heuristics based on multi-bottleneck machines for large-scale job shop scheduling problems (JSP is proposed.Design/methodology/approach: In the algorithm, a number of sub-problems are constructed by iteratively decomposing the large-scale JSP according to the process route of each job. And then the solution of the large-scale JSP can be obtained by iteratively solving the sub-problems. In order to improve the sub-problems' solving efficiency and the solution quality, a detection method for multi-bottleneck machines based on critical path is proposed. Therewith the unscheduled operations can be decomposed into bottleneck operations and non-bottleneck operations. According to the principle of “Bottleneck leads the performance of the whole manufacturing system” in TOC (Theory Of Constraints, the bottleneck operations are scheduled by genetic algorithm for high solution quality, and the non-bottleneck operations are scheduled by dispatching rules for the improvement of the solving efficiency.Findings: In the process of the sub-problems' construction, partial operations in the previous scheduled sub-problem are divided into the successive sub-problem for re-optimization. This strategy can improve the solution quality of the algorithm. In the process of solving the sub-problems, the strategy that evaluating the chromosome's fitness by predicting the global scheduling objective value can improve the solution quality.Research limitations/implications: In this research, there are some assumptions which reduce the complexity of the large-scale scheduling problem. They are as follows: The processing route of each job is predetermined, and the processing time of each operation is fixed. There is no machine breakdown, and no preemption of the operations is allowed. The assumptions should be considered if the algorithm is used in the actual job shop.Originality/value: The research provides an efficient scheduling method for the

  13. Pitfalls in Teaching Judgment Heuristics

    Science.gov (United States)

    Shepperd, James A.; Koch, Erika J.

    2005-01-01

    Demonstrations of judgment heuristics typically focus on how heuristics can lead to poor judgments. However, exclusive focus on the negative consequences of heuristics can prove problematic. We illustrate the problem with the representativeness heuristic and present a study (N = 45) that examined how examples influence understanding of the…

  14. Combined Heuristic Attack Strategy on Complex Networks

    Directory of Open Access Journals (Sweden)

    Marek Šimon

    2017-01-01

    Full Text Available Usually, the existence of a complex network is considered an advantage feature and efforts are made to increase its robustness against an attack. However, there exist also harmful and/or malicious networks, from social ones like spreading hoax, corruption, phishing, extremist ideology, and terrorist support up to computer networks spreading computer viruses or DDoS attack software or even biological networks of carriers or transport centers spreading disease among the population. New attack strategy can be therefore used against malicious networks, as well as in a worst-case scenario test for robustness of a useful network. A common measure of robustness of networks is their disintegration level after removal of a fraction of nodes. This robustness can be calculated as a ratio of the number of nodes of the greatest remaining network component against the number of nodes in the original network. Our paper presents a combination of heuristics optimized for an attack on a complex network to achieve its greatest disintegration. Nodes are deleted sequentially based on a heuristic criterion. Efficiency of classical attack approaches is compared to the proposed approach on Barabási-Albert, scale-free with tunable power-law exponent, and Erdős-Rényi models of complex networks and on real-world networks. Our attack strategy results in a faster disintegration, which is counterbalanced by its slightly increased computational demands.

  15. A systematic approach for component-based software development

    NARCIS (Netherlands)

    Guareis de farias, Cléver; van Sinderen, Marten J.; Ferreira Pires, Luis

    2000-01-01

    Component-based software development enables the construction of software artefacts by assembling prefabricated, configurable and independently evolving building blocks, called software components. This paper presents an approach for the development of component-based software artefacts. This

  16. A Combined Approach for Component-based Software Design

    NARCIS (Netherlands)

    Guareis de farias, Cléver; van Sinderen, Marten J.; Ferreira Pires, Luis; Quartel, Dick; Baldoni, R.

    2001-01-01

    Component-based software development enables the construction of software artefacts by assembling binary units of production, distribution and deployment, the so-called software components. Several approaches addressing component-based development have been proposed recently. Most of these

  17. An integral heuristic

    International Nuclear Information System (INIS)

    Kauffman, L.H.

    1990-01-01

    This paper gives a heuristic derivation of the skein relation for the Homfly polynomial in an integral formalism. The derivation is formally correct but highly simplified. In the light of Witten's proposal for invariants of links via functional integrals, it is useful to have a formal pattern to compare with the complexities of the full approach. The formalism is a heuristic. However, it is closely related to the actual structure of the Witten functional integral

  18. A heuristic forecasting model for stock decision

    OpenAIRE

    Zhang, D.; Jiang, Q.; Li, X.

    2005-01-01

    This paper describes a heuristic forecasting model based on neural networks for stock decision-making. Some heuristic strategies are presented for enhancing the learning capability of neural networks and obtaining better trading performance. The China Shanghai Composite Index is used as case study. The forecasting model can forecast the buying and selling signs according to the result of neural network prediction. Results are compared with a benchmark buy-and-hold strategy. ...

  19. Component failure data base of TRIGA reactors

    International Nuclear Information System (INIS)

    Djuricic, M.

    2004-10-01

    This compilation provides failure data such as first criticality, component type description (reactor component, population, cumulative calendar time, cumulative operating time, demands, failure mode, failures, failure rate, failure probability) and specific information on each type of component of TRIGA Mark-II reactors in Austria, Bangladesh, Germany, Finland, Indonesia, Italy, Indonesia, Slovenia and Romania. (nevyjel)

  20. Heuristic decision model for intelligent nuclear power systems design

    International Nuclear Information System (INIS)

    Nassersharif, B.; Portal, M.G.; Gaeta, M.J.

    1989-01-01

    The objective of this project was to investigate intelligent nuclear power systems design. A theoretical model of the design process has been developed. A fundamental process in this model is the heuristic decision making for design (i.e., selection of methods, components, materials, etc.). Rule-based expert systems do not provide the completeness that is necessary to generate good design. A new method, based on the fuzzy set theory, has been developed and is presented here. A feedwater system knowledge base (KB) was developed for a prototype software experiment to benchmark the theory

  1. A Priority Rule-Based Heuristic for Resource Investment Project Scheduling Problem with Discounted Cash Flows and Tardiness Penalties

    Directory of Open Access Journals (Sweden)

    Amir Abbas Najafi

    2009-01-01

    Full Text Available Resource investment problem with discounted cash flows (RIPDCFs is a class of project scheduling problem. In RIPDCF, the availability levels of the resources are considered decision variables, and the goal is to find a schedule such that the net present value of the project cash flows optimizes. In this paper, we consider a new RIPDCF in which tardiness of project is permitted with defined penalty. We mathematically formulated the problem and developed a heuristic method to solve it. The results of the performance analysis of the proposed method show an effective solution approach to the problem.

  2. The Study on Heuristic and Algorithmic Processing Creative Training on Product Design Education: Based on the Bags & Luggage design Course

    Directory of Open Access Journals (Sweden)

    Zhang Ziran

    2018-01-01

    Full Text Available In the study, context-creativity of Teresa M. Amabile was used as The foundation to apply it in the bags & luggage design course. Moreover, the sectional creative training education mode of prior heuristic task and postpositional algorithmic task was proposed. 26 junior students in product design were used as the trial objects. The Consensual Technique for Creativity (CAT was considered as the scoring standard of creative performance. In the end, the sectional theoretical framework of effective creative training in product design was finally proposed.

  3. Identifying Onboarding Heuristics for Free-to-Play Mobile Games

    DEFF Research Database (Denmark)

    Thomsen, Line Ebdrup; Weigert Petersen, Falko; Drachen, Anders

    2016-01-01

    a set of heuristics for the design of onboarding phases in mobile games is presented. The heuristics are identified by a lab-based mixed-methods experiment, utilizing lightweight psycho-physiological measures together with self-reported player responses, across three titles that cross the genres...... of puzzle games, base builders and arcade games, and utilize different onboarding phase design approaches. Results showcase how heuristics can be used to design engaging onboarding phases in mobile games....

  4. The E-health Literacy Demands of Australia's My Health Record: A Heuristic Evaluation of Usability.

    Science.gov (United States)

    Walsh, Louisa; Hemsley, Bronwyn; Allan, Meredith; Adams, Natalie; Balandin, Susan; Georgiou, Andrew; Higgins, Isabel; McCarthy, Shaun; Hill, Sophie

    2017-01-01

    My Health Record is Australia's electronic personal health record system, which was introduced in July 2012. As of August 2017, approximately 21 percent of Australia's total population was registered to use My Health Record. Internationally, usability issues have been shown to negatively influence the uptake and use of electronic health record systems, and this scenario may particularly affect people who have low e-health literacy. It is likely that usability issues are negatively affecting the uptake and use of My Health Record in Australia. To identify potential e-health literacy-related usability issues within My Health Record through a heuristic evaluation method. Between September 14 and October 12, 2016, three of the authors conducted a heuristic evaluation of the two consumer-facing components of My Health Record-the information website and the electronic health record itself. These two components were evaluated against two sets of heuristics-the Health Literacy Online checklist and the Monkman Heuristics. The Health Literacy Online checklist and Monkman Heuristics are evidence-based checklists of web design elements with a focus on design for audiences with low health literacy. During this heuristic evaluation, the investigators individually navigated through the consumer-facing components of My Health Record, recording instances where the My Health Record did not conform to the checklist criteria. After the individual evaluations were completed, the investigators conferred and aggregated their results. From this process, a list of usability violations was constructed. When evaluated against the Health Literacy Online Checklist, the information website demonstrated violations in 12 of 35 criteria, and the electronic health record demonstrated violations in 16 of 35 criteria. When evaluated against the Monkman Heuristics, the information website demonstrated violations in 7 of 11 criteria, and the electronic health record demonstrated violations in 9 of 11

  5. Lifestyles Based on Health Components in Iran

    Directory of Open Access Journals (Sweden)

    Babaei

    2016-07-01

    Full Text Available Context Lifestyle is a way employed by people, groups and nations and is formed in specific geographical, economic, political, cultural and religious texts. Health depends on lifestyle and is essential to preserve and promote health and improve lifestyle. Objectives The present study aimed to investigate lifestyle based on health-oriented components in Iran. Data Sources The research was conducted through E-banks including scientific information database (SID, Iran medical science databank (Iran Medex, Iran journal databank (Magiran and other databases such as Elsevier, PubMed and google scholar meta search engine regarding the subject from 2000 to 2014. Moreover, Official Iranian statistics and information were applied. The search terms used included lifestyle, health, health promoting behaviors, health-oriented lifestyle and lifestyle in Iran. Study Selection In the primary research, many papers were observed out of which 157 (120 in Farsi and 37 in English were selected. Data Extraction Following the careful study of these papers and excluding the unqualified papers, 19 papers with thorough information and higher relevance with the research purpose were selected. Results After examining articles based on the selected keywords and search strategies, 215 articles (134 in Farsi and 81 in English were obtained. Components of lifestyle and health are increasing in recent years; therefore, 8 (42% and 11 (58% articles were published during 2005 - 2010 and 2011 - 2014, respectively. Among them, there were 3 (16%, 8 (42%, 2 (10.5%, 2 (10.5% and 0 articles on the review of literature, descriptive-analytic, qualitative, analytic and descriptive articles, respectively. Conclusions Due to positive effect of healthy lifestyle on health promotion of individuals, it would be better for the government to provide comprehensive programs and policies in the society to enhance awareness of people about positive effects of health-oriented lifestyle on life and

  6. The Component-Based Application for GAMESS

    Energy Technology Data Exchange (ETDEWEB)

    Peng, Fang [Iowa State Univ., Ames, IA (United States)

    2007-01-01

    GAMESS, a quantum chetnistry program for electronic structure calculations, has been freely shared by high-performance application scientists for over twenty years. It provides a rich set of functionalities and can be run on a variety of parallel platforms through a distributed data interface. While a chemistry computation is sophisticated and hard to develop, the resource sharing among different chemistry packages will accelerate the development of new computations and encourage the cooperation of scientists from universities and laboratories. Common Component Architecture (CCA) offers an enviromnent that allows scientific packages to dynamically interact with each other through components, which enable dynamic coupling of GAMESS with other chetnistry packages, such as MPQC and NWChem. Conceptually, a cotnputation can be constructed with "plug-and-play" components from scientific packages and require more than componentizing functions/subroutines of interest, especially for large-scale scientific packages with a long development history. In this research, we present our efforts to construct cotnponents for GAMESS that conform to the CCA specification. The goal is to enable the fine-grained interoperability between three quantum chemistry programs, GAMESS, MPQC and NWChem, via components. We focus on one of the three packages, GAMESS; delineate the structure of GAMESS computations, followed by our approaches to its component development. Then we use GAMESS as the driver to interoperate integral components from the other tw"o packages, arid show the solutions for interoperability problems along with preliminary results. To justify the versatility of the design, the Tuning and Analysis Utility (TAU) components have been coupled with GAMESS and its components, so that the performance of GAMESS and its components may be analyzed for a wide range of systetn parameters.

  7. Heuristic drift-based model of the power scrape-off width in low-gas-puff H-mode tokamaks

    International Nuclear Information System (INIS)

    Goldston, R.J.

    2012-01-01

    A heuristic model for the plasma scrape-off width in low-gas-puff tokamak H-mode plasmas is introduced. Grad B and curv B drifts into the scrape-off layer (SOL) are balanced against near-sonic parallel flows out of the SOL, to the divertor plates. The overall particle flow pattern posited is a modification for open field lines of Pfirsch–Schlüter flows to include order-unity sinks to the divertors. These assumptions result in an estimated SOL width of ∼2aρ p /R. They also result in a first-principles calculation of the particle confinement time of H-mode plasmas, qualitatively consistent with experimental observations. It is next assumed that anomalous perpendicular electron thermal diffusivity is the dominant source of heat flux across the separatrix, investing the SOL width, derived above, with heat from the main plasma. The separatrix temperature is calculated based on a two-point model balancing power input to the SOL with Spitzer–Härm parallel thermal conduction losses to the divertor. This results in a heuristic closed-form prediction for the power scrape-off width that is in reasonable quantitative agreement both in absolute magnitude and in scaling with recent experimental data. Further work should include full numerical calculations, including all magnetic and electric drifts, as well as more thorough comparison with experimental data.

  8. A Novel Spatial-Temporal Voronoi Diagram-Based Heuristic Approach for Large-Scale Vehicle Routing Optimization with Time Constraints

    Directory of Open Access Journals (Sweden)

    Wei Tu

    2015-10-01

    Full Text Available Vehicle routing optimization (VRO designs the best routes to reduce travel cost, energy consumption, and carbon emission. Due to non-deterministic polynomial-time hard (NP-hard complexity, many VROs involved in real-world applications require too much computing effort. Shortening computing time for VRO is a great challenge for state-of-the-art spatial optimization algorithms. From a spatial-temporal perspective, this paper presents a spatial-temporal Voronoi diagram-based heuristic approach for large-scale vehicle routing problems with time windows (VRPTW. Considering time constraints, a spatial-temporal Voronoi distance is derived from the spatial-temporal Voronoi diagram to find near neighbors in the space-time searching context. A Voronoi distance decay strategy that integrates a time warp operation is proposed to accelerate local search procedures. A spatial-temporal feature-guided search is developed to improve unpromising micro route structures. Experiments on VRPTW benchmarks and real-world instances are conducted to verify performance. The results demonstrate that the proposed approach is competitive with state-of-the-art heuristics and achieves high-quality solutions for large-scale instances of VRPTWs in a short time. This novel approach will contribute to spatial decision support community by developing an effective vehicle routing optimization method for large transportation applications in both public and private sectors.

  9. Usability of a patient education and motivation tool using heuristic evaluation.

    Science.gov (United States)

    Joshi, Ashish; Arora, Mohit; Dai, Liwei; Price, Kathleen; Vizer, Lisa; Sears, Andrew

    2009-11-06

    Computer-mediated educational applications can provide a self-paced, interactive environment to deliver educational content to individuals about their health condition. These programs have been used to deliver health-related information about a variety of topics, including breast cancer screening, asthma management, and injury prevention. We have designed the Patient Education and Motivation Tool (PEMT), an interactive computer-based educational program based on behavioral, cognitive, and humanistic learning theories. The tool is designed to educate users and has three key components: screening, learning, and evaluation. The objective of this tutorial is to illustrate a heuristic evaluation using a computer-based patient education program (PEMT) as a case study. The aims were to improve the usability of PEMT through heuristic evaluation of the interface; to report the results of these usability evaluations; to make changes based on the findings of the usability experts; and to describe the benefits and limitations of applying usability evaluations to PEMT. PEMT was evaluated by three usability experts using Nielsen's usability heuristics while reviewing the interface to produce a list of heuristic violations with severity ratings. The violations were sorted by heuristic and ordered from most to least severe within each heuristic. A total of 127 violations were identified with a median severity of 3 (range 0 to 4 with 0 = no problem to 4 = catastrophic problem). Results showed 13 violations for visibility (median severity = 2), 38 violations for match between system and real world (median severity = 2), 6 violations for user control and freedom (median severity = 3), 34 violations for consistency and standards (median severity = 2), 11 violations for error severity (median severity = 3), 1 violation for recognition and control (median severity = 3), 7 violations for flexibility and efficiency (median severity = 2), 9 violations for aesthetic and minimalist design

  10. A HYBRID HEURISTIC ALGORITHM FOR THE CLUSTERED TRAVELING SALESMAN PROBLEM

    Directory of Open Access Journals (Sweden)

    Mário Mestria

    2016-04-01

    Full Text Available ABSTRACT This paper proposes a hybrid heuristic algorithm, based on the metaheuristics Greedy Randomized Adaptive Search Procedure, Iterated Local Search and Variable Neighborhood Descent, to solve the Clustered Traveling Salesman Problem (CTSP. Hybrid Heuristic algorithm uses several variable neighborhood structures combining the intensification (using local search operators and diversification (constructive heuristic and perturbation routine. In the CTSP, the vertices are partitioned into clusters and all vertices of each cluster have to be visited contiguously. The CTSP is -hard since it includes the well-known Traveling Salesman Problem (TSP as a special case. Our hybrid heuristic is compared with three heuristics from the literature and an exact method. Computational experiments are reported for different classes of instances. Experimental results show that the proposed hybrid heuristic obtains competitive results within reasonable computational time.

  11. The development of component-based information systems

    CERN Document Server

    Cesare, Sergio de; Macredie, Robert

    2015-01-01

    This work provides a comprehensive overview of research and practical issues relating to component-based development information systems (CBIS). Spanning the organizational, developmental, and technical aspects of the subject, the original research included here provides fresh insights into successful CBIS technology and application. Part I covers component-based development methodologies and system architectures. Part II analyzes different aspects of managing component-based development. Part III investigates component-based development versus commercial off-the-shelf products (COTS), includi

  12. Using tree diversity to compare phylogenetic heuristics.

    Science.gov (United States)

    Sul, Seung-Jin; Matthews, Suzanne; Williams, Tiffani L

    2009-04-29

    Evolutionary trees are family trees that represent the relationships between a group of organisms. Phylogenetic heuristics are used to search stochastically for the best-scoring trees in tree space. Given that better tree scores are believed to be better approximations of the true phylogeny, traditional evaluation techniques have used tree scores to determine the heuristics that find the best scores in the fastest time. We develop new techniques to evaluate phylogenetic heuristics based on both tree scores and topologies to compare Pauprat and Rec-I-DCM3, two popular Maximum Parsimony search algorithms. Our results show that although Pauprat and Rec-I-DCM3 find the trees with the same best scores, topologically these trees are quite different. Furthermore, the Rec-I-DCM3 trees cluster distinctly from the Pauprat trees. In addition to our heatmap visualizations of using parsimony scores and the Robinson-Foulds distance to compare best-scoring trees found by the two heuristics, we also develop entropy-based methods to show the diversity of the trees found. Overall, Pauprat identifies more diverse trees than Rec-I-DCM3. Overall, our work shows that there is value to comparing heuristics beyond the parsimony scores that they find. Pauprat is a slower heuristic than Rec-I-DCM3. However, our work shows that there is tremendous value in using Pauprat to reconstruct trees-especially since it finds identical scoring but topologically distinct trees. Hence, instead of discounting Pauprat, effort should go in improving its implementation. Ultimately, improved performance measures lead to better phylogenetic heuristics and will result in better approximations of the true evolutionary history of the organisms of interest.

  13. Optimal restoration strategies based on heuristic techniques for electrical distribution networks; Estrategias de restablecimiento optimas basadas en tecnicas heuristicas para redes de distribucion electrica

    Energy Technology Data Exchange (ETDEWEB)

    Cruz Castrejon, J. A; Islas Perez, E; Espinosa Reza, A; Garcia Mendoza, R [Instituto de Investigaciones Electricas, Cuernavaca, Morelos (Mexico)]. E-mails: adrian.cruz@iie.org.mx; eislas@iie.org.mx; aer@iie.org.mx; rgarcia@iie.org.mx

    2013-03-15

    In this paper we present a proposed solution to the problem of finding alternatives to reset faults in radial distribution networks power systems. This solution uses a deterministic method based on the definition of heuristics and whose main objectives are to improve execution time and solution quality. This search is based on the alternate repetition of two stages: a stage that attempts to reset the unconnected areas and other areas trying ballasting overloaded. [Spanish] En este articulo se presenta una propuesta de solucion al problema de busqueda de alternativas de restablecimiento para fallas en redes de distribucion radiales en sistemas electricos de potencia. Esta solucion utiliza un metodo deterministico basado en la definicion de heuristicas y cuyos objetivos principales son: mejorar el tiempo de ejecucion y calidad de la solucion. Esta busqueda se basa en la repeticion alternada de dos etapas: una etapa que intenta restablecer las areas desconectadas y otra que intenta deslastrar las areas sobrecargadas.

  14. A refinement driven component-based design

    DEFF Research Database (Denmark)

    Chen, Zhenbang; Liu, Zhiming; Ravn, Anders Peter

    2007-01-01

    the work on the Common Component Modelling Example (CoCoME). This gives evidence that the formal techniques developed in rCOS can be integrated into a model-driven development process and shows where it may be integrated in computer-aided software engineering (CASE) tools for adding formally supported...

  15. Proposing New Heuristic Approaches for Preventive Maintenance Scheduling

    Directory of Open Access Journals (Sweden)

    majid Esmailian

    2013-08-01

    Full Text Available The purpose of preventive maintenance management is to perform a series of tasks that prevent or minimize production breakdowns and improve reliability of production facilities. An important objective of preventive maintenance management is to minimize downtime of production facilities. In order to accomplish this objective, personnel should efficiently allocate resources and determine an effective maintenance schedule. Gopalakrishnan (1997 developed a mathematical model and four heuristic approaches to solve the preventive maintenance scheduling problem of assigning skilled personnel to work with tasks that require a set of corresponding skills. However, there are several limitations in the prior work in this area of research. The craft combination problem has not been solved because the craft combination is assumed as given. The craft combination problem concerns the computation of all combinations of assigning multi skilled workers to accomplishment of a particular task. In fact, determining craft combinations is difficult because of the exponential number of craft combinations that are possible. This research provides a heuristic approach for determining the craft combination and four new heuristic approach solution for the preventive maintenance scheduling problem with multi skilled workforce constraints. In order to examine the new heuristic approach and to compare the new heuristic approach with heuristic approach of Gopalakrishnan (1997, 81 standard problems have been generated based on the criterion suggested by from Gopalakrishnan (1997. The average solution quality (SQ of the new heuristic approaches is 1.86% and in old heuristic approaches is 8.32%. The solution time of new heuristic approaches are shorter than old heuristic approaches. The solution time of new heuristic approaches is 0.78 second and old heuristic approaches is 6.43 second, but the solution time of mathematical model provided by Gopalakrishnan (1997 is 152 second.

  16. Educational Technologies Based on Software Components

    Directory of Open Access Journals (Sweden)

    Marian DARDALA

    2006-01-01

    Full Text Available Informatics technologies allow to easily develop and adapt e-learning systems. In order to be used by any users, the system must be developed to permit the lessons construction in a dynamic way. The component technology is a solution to this problem and offers the possibility to define the basic objects that will be connected at the run time to develop the personalized e-lessons.

  17. Quantifying Heuristic Bias: Anchoring, Availability, and Representativeness.

    Science.gov (United States)

    Richie, Megan; Josephson, S Andrew

    2018-01-01

    Construct: Authors examined whether a new vignette-based instrument could isolate and quantify heuristic bias. Heuristics are cognitive shortcuts that may introduce bias and contribute to error. There is no standardized instrument available to quantify heuristic bias in clinical decision making, limiting future study of educational interventions designed to improve calibration of medical decisions. This study presents validity data to support a vignette-based instrument quantifying bias due to the anchoring, availability, and representativeness heuristics. Participants completed questionnaires requiring assignment of probabilities to potential outcomes of medical and nonmedical scenarios. The instrument randomly presented scenarios in one of two versions: Version A, encouraging heuristic bias, and Version B, worded neutrally. The primary outcome was the difference in probability judgments for Version A versus Version B scenario options. Of 167 participants recruited, 139 enrolled. Participants assigned significantly higher mean probability values to Version A scenario options (M = 9.56, SD = 3.75) than Version B (M = 8.98, SD = 3.76), t(1801) = 3.27, p = .001. This result remained significant analyzing medical scenarios alone (Version A, M = 9.41, SD = 3.92; Version B, M = 8.86, SD = 4.09), t(1204) = 2.36, p = .02. Analyzing medical scenarios by heuristic revealed a significant difference between Version A and B for availability (Version A, M = 6.52, SD = 3.32; Version B, M = 5.52, SD = 3.05), t(404) = 3.04, p = .003, and representativeness (Version A, M = 11.45, SD = 3.12; Version B, M = 10.67, SD = 3.71), t(396) = 2.28, p = .02, but not anchoring. Stratifying by training level, students maintained a significant difference between Version A and B medical scenarios (Version A, M = 9.83, SD = 3.75; Version B, M = 9.00, SD = 3.98), t(465) = 2.29, p = .02, but not residents or attendings. Stratifying by heuristic and training level, availability maintained

  18. Design Optimization of Mechanical Components Using an Enhanced Teaching-Learning Based Optimization Algorithm with Differential Operator

    Directory of Open Access Journals (Sweden)

    B. Thamaraikannan

    2014-01-01

    Full Text Available This paper studies in detail the background and implementation of a teaching-learning based optimization (TLBO algorithm with differential operator for optimization task of a few mechanical components, which are essential for most of the mechanical engineering applications. Like most of the other heuristic techniques, TLBO is also a population-based method and uses a population of solutions to proceed to the global solution. A differential operator is incorporated into the TLBO for effective search of better solutions. To validate the effectiveness of the proposed method, three typical optimization problems are considered in this research: firstly, to optimize the weight in a belt-pulley drive, secondly, to optimize the volume in a closed coil helical spring, and finally to optimize the weight in a hollow shaft. have been demonstrated. Simulation result on the optimization (mechanical components problems reveals the ability of the proposed methodology to find better optimal solutions compared to other optimization algorithms.

  19. Does the inherence heuristic take us to psychological essentialism?

    Science.gov (United States)

    Marmodoro, Anna; Murphy, Robin A; Baker, A G

    2014-10-01

    We argue that the claim that essence-based causal explanations emerge, hydra-like, from an inherence heuristic is incomplete. No plausible mechanism for the transition from concrete properties, or cues, to essences is provided. Moreover, the fundamental shotgun and storytelling mechanisms of the inherence heuristic are not clearly enough specified to distinguish them, developmentally, from associative or causal networks.

  20. Component-based framework for subsurface simulations

    International Nuclear Information System (INIS)

    Palmer, B J; Fang, Yilin; Hammond, Glenn; Gurumoorthi, Vidhya

    2007-01-01

    Simulations in the subsurface environment represent a broad range of phenomena covering an equally broad range of scales. Developing modelling capabilities that can integrate models representing different phenomena acting at different scales present formidable challenges both from the algorithmic and computer science perspective. This paper will describe the development of an integrated framework that will be used to combine different models into a single simulation. Initial work has focused on creating two frameworks, one for performing smooth particle hydrodynamics (SPH) simulations of fluid systems, the other for performing grid-based continuum simulations of reactive subsurface flow. The SPH framework is based on a parallel code developed for doing pore scale simulations, the continuum grid-based framework is based on the STOMP (Subsurface Transport Over Multiple Phases) code developed at PNNL Future work will focus on combining the frameworks together to perform multiscale, multiphysics simulations of reactive subsurface flow

  1. Motor heuristics and embodied choices: how to choose and act.

    Science.gov (United States)

    Raab, Markus

    2017-08-01

    Human performance requires choosing what to do and how to do it. The goal of this theoretical contribution is to advance understanding of how the motor and cognitive components of choices are intertwined. From a holistic perspective I extend simple heuristics that have been tested in cognitive tasks to motor tasks, coining the term motor heuristics. Similarly I extend the concept of embodied cognition, that has been tested in simple sensorimotor processes changing decisions, to complex sport behavior coining the term embodied choices. Thus both motor heuristics and embodied choices explain complex behavior such as studied in sport and exercise psychology. Copyright © 2017 Elsevier Ltd. All rights reserved.

  2. Component-based development of software language engineering tools

    NARCIS (Netherlands)

    Ssanyu, J.; Hemerik, C.

    2011-01-01

    In this paper we outline how Software Language Engineering (SLE) could benefit from Component-based Software Development (CBSD) techniques and present an architecture aimed at developing a coherent set of lightweight SLE components, fitting into a general-purpose component framework. In order to

  3. Base isolation strategies for structures and components

    International Nuclear Information System (INIS)

    Varma, Veto; Reddy, G.R.; Vaze, K.K.; Kushwaha, H.S.

    2003-08-01

    In the present report the effect of laminated rubber bearing (LRB) system on the dynamic response of the structure was studied. A LRB system was designed and tested in the laboratory for its dynamic characteristics. Finite element analysis was also performed and based on this analysis, isolator for PHWR nuclear power plant was designed. Analysis of the building was performed with and without isolator. Comparison of responses was made in terms of frequencies, accelerations and displacements and floor response spectra. (author)

  4. Fourth Graders' Heuristic Problem-Solving Behavior.

    Science.gov (United States)

    Lee, Kil S.

    1982-01-01

    Eight boys and eight girls from a rural elementary school participated in the investigation. Specific heuristics were adopted from Polya; and the students selected represented two substages of Piaget's concrete operational stage. Five hypotheses were generated, based on observed results and the study's theoretical rationale. (MP)

  5. Seismic component fragility data base for IPEEE

    International Nuclear Information System (INIS)

    Bandyopadhyay, K.; Hofmayer, C.

    1990-01-01

    Seismic probabilistic risk assessment or a seismic margin study will require a reliable data base of seismic fragility of various equipment classes. Brookhaven National Laboratory (BNL) has selected a group of equipment and generically evaluated the seismic fragility of each equipment class by use of existing test data. This paper briefly discusses the evaluation methodology and the fragility results. The fragility analysis results when used in the Individual Plant Examination for External Events (IPEEE) Program for nuclear power plants are expected to provide insights into seismic vulnerabilities of equipment for earthquakes beyond the design basis. 3 refs., 1 fig., 1 tab

  6. A Tutorial on Heuristic Methods

    DEFF Research Database (Denmark)

    Vidal, Rene Victor Valqui; Werra, D. de; Silver, E.

    1980-01-01

    In this paper we define a heuristic method as a procedure for solving a well-defined mathematical problem by an intuitive approach in which the structure of the problem can be interpreted and exploited intelligently to obtain a reasonable solution. Issues discussed include: (i) the measurement...... of the quality of a heuristic method, (ii) different types of heuristic procedures, (iii) the interactive role of human beings and (iv) factors that may influence the choice or testing of heuristic methods. A large number of references are included....

  7. Motor heuristics and embodied choices: how to choose and act

    OpenAIRE

    Raab, M

    2017-01-01

    © 2017 Elsevier LtdHuman performance requires choosing what to do and how to do it. The goal of this theoretical contribution is to advance understanding of how the motor and cognitive components of choices are intertwined. From a holistic perspective I extend simple heuristics that have been tested in cognitive tasks to motor tasks, coining the term motor heuristics. Similarly I extend the concept of embodied cognition, that has been tested in simple sensorimotor processes changing decisions...

  8. Component-Based Approach in Learning Management System Development

    Science.gov (United States)

    Zaitseva, Larisa; Bule, Jekaterina; Makarov, Sergey

    2013-01-01

    The paper describes component-based approach (CBA) for learning management system development. Learning object as components of e-learning courses and their metadata is considered. The architecture of learning management system based on CBA being developed in Riga Technical University, namely its architecture, elements and possibilities are…

  9. IAEA's experience in compiling a generic component reliability data base

    International Nuclear Information System (INIS)

    Tomic, B.; Lederman, L.

    1991-01-01

    Reliability data are essential in probabilistic safety assessment, with component reliability parameters being particularly important. Component failure data which is plant specific would be most appropriate but this is rather limited. However, similar components are used in different designs. Generic data, that is all data that is not plant specific to the plant being analyzed but which relates to components more generally, is important. The International Atomic Energy Agency has compiled the Generic Component Reliability Data Base from data available in the open literature. It is part of the IAEA computer code package for fault/event tree analysis. The Data Base contains 1010 different records including most of the components used in probabilistic safety analyses of nuclear power plants. The data base input was quality controlled and data sources noted. The data compilation procedure and problems associated with using generic data are explained. (UK)

  10. Generalized structured component analysis a component-based approach to structural equation modeling

    CERN Document Server

    Hwang, Heungsun

    2014-01-01

    Winner of the 2015 Sugiyama Meiko Award (Publication Award) of the Behaviormetric Society of Japan Developed by the authors, generalized structured component analysis is an alternative to two longstanding approaches to structural equation modeling: covariance structure analysis and partial least squares path modeling. Generalized structured component analysis allows researchers to evaluate the adequacy of a model as a whole, compare a model to alternative specifications, and conduct complex analyses in a straightforward manner. Generalized Structured Component Analysis: A Component-Based Approach to Structural Equation Modeling provides a detailed account of this novel statistical methodology and its various extensions. The authors present the theoretical underpinnings of generalized structured component analysis and demonstrate how it can be applied to various empirical examples. The book enables quantitative methodologists, applied researchers, and practitioners to grasp the basic concepts behind this new a...

  11. Heuristics, biases and traps in managerial decision making

    Directory of Open Access Journals (Sweden)

    Peter Gál

    2013-01-01

    Full Text Available The aim of the paper is to demonstrate the impact of heuristics, biases and psychological traps on the decision making. Heuristics are unconscious routines people use to cope with the complexity inherent in most decision situations. They serve as mental shortcuts that help people to simplify and structure the information encountered in the world. These heuristics could be quite useful in some situations, while in others they can lead to severe and systematic errors, based on significant deviations from the fundamental principles of statistics, probability and sound judgment. This paper focuses on illustrating the existence of the anchoring, availability, and representativeness heuristics, originally described by Tversky & Kahneman in the early 1970’s. The anchoring heuristic is a tendency to focus on the initial information, estimate or perception (even random or irrelevant number as a starting point. People tend to give disproportionate weight to the initial information they receive. The availability heuristic explains why highly imaginable or vivid information have a disproportionate effect on people’s decisions. The representativeness heuristic causes that people rely on highly specific scenarios, ignore base rates, draw conclusions based on small samples and neglect scope. Mentioned phenomena are illustrated and supported by evidence based on the statistical analysis of the results of a questionnaire.

  12. Heuristics Reasoning in Diagnostic Judgment.

    Science.gov (United States)

    O'Neill, Eileen S.

    1995-01-01

    Describes three heuristics--short-cut mental strategies that streamline information--relevant to diagnostic reasoning: accessibility, similarity, and anchoring and adjustment. Analyzes factors thought to influence heuristic reasoning and presents interventions to be tested for nursing practice and education. (JOW)

  13. Reexamining Our Bias against Heuristics

    Science.gov (United States)

    McLaughlin, Kevin; Eva, Kevin W.; Norman, Geoff R.

    2014-01-01

    Using heuristics offers several cognitive advantages, such as increased speed and reduced effort when making decisions, in addition to allowing us to make decision in situations where missing data do not allow for formal reasoning. But the traditional view of heuristics is that they trade accuracy for efficiency. Here the authors discuss sources…

  14. HEURISTIC APPROACHES FOR PORTFOLIO OPTIMIZATION

    OpenAIRE

    Manfred Gilli, Evis Kellezi

    2000-01-01

    The paper first compares the use of optimization heuristics to the classical optimization techniques for the selection of optimal portfolios. Second, the heuristic approach is applied to problems other than those in the standard mean-variance framework where the classical optimization fails.

  15. A methodology to design heuristics for model selection based on the characteristics of data: Application to investigate when the Negative Binomial Lindley (NB-L) is preferred over the Negative Binomial (NB).

    Science.gov (United States)

    Shirazi, Mohammadali; Dhavala, Soma Sekhar; Lord, Dominique; Geedipally, Srinivas Reddy

    2017-10-01

    Safety analysts usually use post-modeling methods, such as the Goodness-of-Fit statistics or the Likelihood Ratio Test, to decide between two or more competitive distributions or models. Such metrics require all competitive distributions to be fitted to the data before any comparisons can be accomplished. Given the continuous growth in introducing new statistical distributions, choosing the best one using such post-modeling methods is not a trivial task, in addition to all theoretical or numerical issues the analyst may face during the analysis. Furthermore, and most importantly, these measures or tests do not provide any intuitions into why a specific distribution (or model) is preferred over another (Goodness-of-Logic). This paper ponders into these issues by proposing a methodology to design heuristics for Model Selection based on the characteristics of data, in terms of descriptive summary statistics, before fitting the models. The proposed methodology employs two analytic tools: (1) Monte-Carlo Simulations and (2) Machine Learning Classifiers, to design easy heuristics to predict the label of the 'most-likely-true' distribution for analyzing data. The proposed methodology was applied to investigate when the recently introduced Negative Binomial Lindley (NB-L) distribution is preferred over the Negative Binomial (NB) distribution. Heuristics were designed to select the 'most-likely-true' distribution between these two distributions, given a set of prescribed summary statistics of data. The proposed heuristics were successfully compared against classical tests for several real or observed datasets. Not only they are easy to use and do not need any post-modeling inputs, but also, using these heuristics, the analyst can attain useful information about why the NB-L is preferred over the NB - or vice versa- when modeling data. Copyright © 2017 Elsevier Ltd. All rights reserved.

  16. Real Time Engineering Analysis Based on a Generative Component Implementation

    DEFF Research Database (Denmark)

    Kirkegaard, Poul Henning; Klitgaard, Jens

    2007-01-01

    The present paper outlines the idea of a conceptual design tool with real time engineering analysis which can be used in the early conceptual design phase. The tool is based on a parametric approach using Generative Components with embedded structural analysis. Each of these components uses the g...

  17. Algorithmic fault tree construction by component-based system modeling

    International Nuclear Information System (INIS)

    Majdara, Aref; Wakabayashi, Toshio

    2008-01-01

    Computer-aided fault tree generation can be easier, faster and less vulnerable to errors than the conventional manual fault tree construction. In this paper, a new approach for algorithmic fault tree generation is presented. The method mainly consists of a component-based system modeling procedure an a trace-back algorithm for fault tree synthesis. Components, as the building blocks of systems, are modeled using function tables and state transition tables. The proposed method can be used for a wide range of systems with various kinds of components, if an inclusive component database is developed. (author)

  18. Verifying Embedded Systems using Component-based Runtime Observers

    DEFF Research Database (Denmark)

    Guan, Wei; Marian, Nicolae; Angelov, Christo K.

    against formally specified properties. This paper presents a component-based design method for runtime observers, which are configured from instances of prefabricated reusable components---Predicate Evaluator (PE) and Temporal Evaluator (TE). The PE computes atomic propositions for the TE; the latter...... is a reconfigurable component processing a data structure, representing the state transition diagram of a non-deterministic state machine, i.e. a Buchi automaton derived from a system property specified in Linear Temporal Logic (LTL). Observer components have been implemented using design models and design patterns...

  19. Flexible optical network components based on densely integrated microring resonators

    NARCIS (Netherlands)

    Geuzebroek, D.H.

    2005-01-01

    This thesis addresses the design, realization and characterization of reconfigurable optical network components based on multiple microring resonators. Since thermally tunable microring resonators can be used as wavelength selective space switches, very compact devices with high complexity and

  20. Component-Based Software Engineering and Runtime Type Definition

    OpenAIRE

    A. R. Shakurov

    2011-01-01

    The component-based approach to software engineering, its current implementations and their limitations are discussed. A new extended architecture for such systems is presented. Its main architectural concepts and principles are considered.

  1. Imprecise system reliability and component importance based on survival signature

    International Nuclear Information System (INIS)

    Feng, Geng; Patelli, Edoardo; Beer, Michael; Coolen, Frank P.A.

    2016-01-01

    The concept of the survival signature has recently attracted increasing attention for performing reliability analysis on systems with multiple types of components. It opens a new pathway for a structured approach with high computational efficiency based on a complete probabilistic description of the system. In practical applications, however, some of the parameters of the system might not be defined completely due to limited data, which implies the need to take imprecisions of component specifications into account. This paper presents a methodology to include explicitly the imprecision, which leads to upper and lower bounds of the survival function of the system. In addition, the approach introduces novel and efficient component importance measures. By implementing relative importance index of each component without or with imprecision, the most critical component in the system can be identified depending on the service time of the system. Simulation method based on survival signature is introduced to deal with imprecision within components, which is precise and efficient. Numerical example is presented to show the applicability of the approach for systems. - Highlights: • Survival signature is a novel way for system reliability and component importance • High computational efficiency based on a complete description of system. • Include explicitly the imprecision, which leads to bounds of the survival function. • A novel relative importance index is proposed as importance measure. • Allows to identify critical components depending on the service time of the system.

  2. Parallel PDE-Based Simulations Using the Common Component Architecture

    International Nuclear Information System (INIS)

    McInnes, Lois C.; Allan, Benjamin A.; Armstrong, Robert; Benson, Steven J.; Bernholdt, David E.; Dahlgren, Tamara L.; Diachin, Lori; Krishnan, Manoj Kumar; Kohl, James A.; Larson, J. Walter; Lefantzi, Sophia; Nieplocha, Jarek; Norris, Boyana; Parker, Steven G.; Ray, Jaideep; Zhou, Shujia

    2006-01-01

    The complexity of parallel PDE-based simulations continues to increase as multimodel, multiphysics, and multi-institutional projects become widespread. A goal of component based software engineering in such large-scale simulations is to help manage this complexity by enabling better interoperability among various codes that have been independently developed by different groups. The Common Component Architecture (CCA) Forum is defining a component architecture specification to address the challenges of high-performance scientific computing. In addition, several execution frameworks, supporting infrastructure, and general purpose components are being developed. Furthermore, this group is collaborating with others in the high-performance computing community to design suites of domain-specific component interface specifications and underlying implementations. This chapter discusses recent work on leveraging these CCA efforts in parallel PDE-based simulations involving accelerator design, climate modeling, combustion, and accidental fires and explosions. We explain how component technology helps to address the different challenges posed by each of these applications, and we highlight how component interfaces built on existing parallel toolkits facilitate the reuse of software for parallel mesh manipulation, discretization, linear algebra, integration, optimization, and parallel data redistribution. We also present performance data to demonstrate the suitability of this approach, and we discuss strategies for applying component technologies to both new and existing applications

  3. Cloud Based Big Data Infrastructure: Architectural Components and Automated Provisioning

    OpenAIRE

    Demchenko, Yuri; Turkmen, Fatih; Blanchet, Christophe; Loomis, Charles; Laat, Caees de

    2016-01-01

    This paper describes the general architecture and functional components of the cloud based Big Data Infrastructure (BDI). The proposed BDI architecture is based on the analysis of the emerging Big Data and data intensive technologies and supported by the definition of the Big Data Architecture Framework (BDAF) that defines the following components of the Big Data technologies: Big Data definition, Data Management including data lifecycle and data structures, Big Data Infrastructure (generical...

  4. Comparative study of heuristic evaluation and usability testing methods.

    Science.gov (United States)

    Thyvalikakath, Thankam Paul; Monaco, Valerie; Thambuganipalle, Himabindu; Schleyer, Titus

    2009-01-01

    Usability methods, such as heuristic evaluation, cognitive walk-throughs and user testing, are increasingly used to evaluate and improve the design of clinical software applications. There is still some uncertainty, however, as to how those methods can be used to support the development process and evaluation in the most meaningful manner. In this study, we compared the results of a heuristic evaluation with those of formal user tests in order to determine which usability problems were detected by both methods. We conducted heuristic evaluation and usability testing on four major commercial dental computer-based patient records (CPRs), which together cover 80% of the market for chairside computer systems among general dentists. Both methods yielded strong evidence that the dental CPRs have significant usability problems. An average of 50% of empirically-determined usability problems were identified by the preceding heuristic evaluation. Some statements of heuristic violations were specific enough to precisely identify the actual usability problem that study participants encountered. Other violations were less specific, but still manifested themselves in usability problems and poor task outcomes. In this study, heuristic evaluation identified a significant portion of problems found during usability testing. While we make no assumptions about the generalizability of the results to other domains and software systems, heuristic evaluation may, under certain circumstances, be a useful tool to determine design problems early in the development cycle.

  5. Structural Sustainability - Heuristic Approach

    Science.gov (United States)

    Rostański, Krzysztof

    2017-10-01

    Nowadays, we are faced with a challenge of having to join building structures with elements of nature, which seems to be the paradigm of modern planning and design. The questions arise, however, with reference to the following categories: the leading idea, the relation between elements of nature and buildings, the features of a structure combining such elements and, finally, our perception of this structure. If we consider both the overwhelming globalization and our attempts to preserve local values, the only reasonable solution is to develop naturalistic greenery. It can add its uniqueness to any building and to any developed area. Our holistic model, presented in this paper, contains the above mentioned categories within the scope of naturalism. The model is divided into principles, actions related, and possible effects to be obtained. It provides a useful tool for determining the ways and priorities of our design. Although it is not possible to consider all possible actions and solutions in order to support sustainability in any particular design, we can choose, however, a proper mode for our design according to the local conditions by turning to the heuristic method, which helps to choose priorities and targets. Our approach is an attempt to follow the ways of nature as in the natural environment it is optimal solutions that appear and survive, idealism being the domain of mankind only. We try to describe various natural processes in a manner comprehensible to us, which is always a generalization. Such definitions, however, called artificial by naturalists, are presented as art or the current state of knowledge by artists and engineers. Reality, in fact, is always more complicated than its definitions. The heuristic method demonstrates the way how to optimize our design. It requires that all possible information about the local environment should be gathered, as the more is known, the fewer mistakes are made. Following the unquestionable principles, we can

  6. A heuristic for efficient data distribution management in distributed simulation

    Science.gov (United States)

    Gupta, Pankaj; Guha, Ratan K.

    2005-05-01

    In this paper, we propose an algorithm for reducing the complexity of region matching and efficient multicasting in data distribution management component of High Level Architecture (HLA) Run Time Infrastructure (RTI). The current data distribution management (DDM) techniques rely on computing the intersection between the subscription and update regions. When a subscription region and an update region of different federates overlap, RTI establishes communication between the publisher and the subscriber. It subsequently routes the updates from the publisher to the subscriber. The proposed algorithm computes the update/subscription regions matching for dynamic allocation of multicast group. It provides new multicast routines that exploit the connectivity of federation by communicating updates regarding interactions and routes information only to those federates that require them. The region-matching problem in DDM reduces to clique-covering problem using the connections graph abstraction where the federations represent the vertices and the update/subscribe relations represent the edges. We develop an abstract model based on connection graph for data distribution management. Using this abstract model, we propose a heuristic for solving the region-matching problem of DDM. We also provide complexity analysis of the proposed heuristics.

  7. Estimating the relative weights of visual and auditory tau versus heuristic-based cues for time-to-contact judgments in realistic, familiar scenes by older and younger adults.

    Science.gov (United States)

    Keshavarz, Behrang; Campos, Jennifer L; DeLucia, Patricia R; Oberfeld, Daniel

    2017-04-01

    Estimating time to contact (TTC) involves multiple sensory systems, including vision and audition. Previous findings suggested that the ratio of an object's instantaneous optical size/sound intensity to its instantaneous rate of change in optical size/sound intensity (τ) drives TTC judgments. Other evidence has shown that heuristic-based cues are used, including final optical size or final sound pressure level. Most previous studies have used decontextualized and unfamiliar stimuli (e.g., geometric shapes on a blank background). Here we evaluated TTC estimates by using a traffic scene with an approaching vehicle to evaluate the weights of visual and auditory TTC cues under more realistic conditions. Younger (18-39 years) and older (65+ years) participants made TTC estimates in three sensory conditions: visual-only, auditory-only, and audio-visual. Stimuli were presented within an immersive virtual-reality environment, and cue weights were calculated for both visual cues (e.g., visual τ, final optical size) and auditory cues (e.g., auditory τ, final sound pressure level). The results demonstrated the use of visual τ as well as heuristic cues in the visual-only condition. TTC estimates in the auditory-only condition, however, were primarily based on an auditory heuristic cue (final sound pressure level), rather than on auditory τ. In the audio-visual condition, the visual cues dominated overall, with the highest weight being assigned to visual τ by younger adults, and a more equal weighting of visual τ and heuristic cues in older adults. Overall, better characterizing the effects of combined sensory inputs, stimulus characteristics, and age on the cues used to estimate TTC will provide important insights into how these factors may affect everyday behavior.

  8. Exact and Heuristic Algorithms for Runway Scheduling

    Science.gov (United States)

    Malik, Waqar A.; Jung, Yoon C.

    2016-01-01

    This paper explores the Single Runway Scheduling (SRS) problem with arrivals, departures, and crossing aircraft on the airport surface. Constraints for wake vortex separations, departure area navigation separations and departure time window restrictions are explicitly considered. The main objective of this research is to develop exact and heuristic based algorithms that can be used in real-time decision support tools for Air Traffic Control Tower (ATCT) controllers. The paper provides a multi-objective dynamic programming (DP) based algorithm that finds the exact solution to the SRS problem, but may prove unusable for application in real-time environment due to large computation times for moderate sized problems. We next propose a second algorithm that uses heuristics to restrict the search space for the DP based algorithm. A third algorithm based on a combination of insertion and local search (ILS) heuristics is then presented. Simulation conducted for the east side of Dallas/Fort Worth International Airport allows comparison of the three proposed algorithms and indicates that the ILS algorithm performs favorably in its ability to find efficient solutions and its computation times.

  9. Isocyanide based multi component reactions in combinatorial chemistry.

    NARCIS (Netherlands)

    Dömling, A.

    1998-01-01

    Although usually regarded as a recent development, the combinatorial approach to the synthesis of libraries of new drug candidates was first described as early as 1961 using the isocyanide-based one-pot multicomponent Ugi reaction. Isocyanide-based multi component reactions (MCR's) markedly differ

  10. Independent component analysis based filtering for penumbral imaging

    International Nuclear Information System (INIS)

    Chen Yenwei; Han Xianhua; Nozaki, Shinya

    2004-01-01

    We propose a filtering based on independent component analysis (ICA) for Poisson noise reduction. In the proposed filtering, the image is first transformed to ICA domain and then the noise components are removed by a soft thresholding (shrinkage). The proposed filter, which is used as a preprocessing of the reconstruction, has been successfully applied to penumbral imaging. Both simulation results and experimental results show that the reconstructed image is dramatically improved in comparison to that without the noise-removing filters

  11. Heuristic reasoning and relative incompleteness

    OpenAIRE

    Treur, J.

    1993-01-01

    In this paper an approach is presented in which heuristic reasoning is interpreted as strategic reasoning. This type of reasoning enables one to derive which hypothesis to investigate, and which observable information to acquire next (to be able to verify the chosen hypothesis). A compositional architecture for reasoning systems that perform such heuristic reasoning is introduced, called SIX (for Strategic Interactive eXpert systems). This compositional architecture enables user interaction a...

  12. Economic Load Dispatch - A Comparative Study on Heuristic Optimization Techniques With an Improved Coordinated Aggregation-Based PSO

    DEFF Research Database (Denmark)

    Vlachogiannis, Ioannis (John); Lee, KY

    2009-01-01

    In this paper an improved coordinated aggregation-based particle swarm optimization (ICA-PSO) algorithm is introduced for solving the optimal economic load dispatch (ELD) problem in power systems. In the ICA-PSO algorithm each particle in the swarm retains a memory of its best position ever...... encountered, and is attracted only by other particles with better achievements than its own with the exception of the particle with the best achievement, which moves randomly. Moreover, the population size is increased adaptively, the number of search intervals for the particles is selected adaptively...

  13. A heuristic framework for understanding the role of participatory decision making in community-based non-profits.

    Science.gov (United States)

    Bess, Kimberly D; Perkins, Douglas D; Cooper, Daniel G; Jones, Diana L

    2011-06-01

    This paper explores the role of member participation in decision-making (PDM) from an organizational learning (OL) perspective. Community-based organizations (CBOs) serve as mediators between the individual and the local community, often providing the means for community member participation and benefiting organizationally from members' input. Community psychologists have recognized these benefits; however, the field has paid less attention to the role participation plays in increasing CBOs' capacity to meet community needs. We present a framework for exploring how CBO contextual factors influence the use of participatory decision-making structures and practices, and how these affect OL. We then use the framework to examine PDM in qualitative case study analysis of four CBOs: a youth development organization, a faith-based social action coalition, a low-income neighborhood organization, and a large human service agency. We found that organizational form, energy, and culture each had a differential impact on participation in decision making within CBOs. We highlight how OL is constrained in CBOs and document how civic aims and voluntary membership enhanced participation and learning.

  14. A Personalized Rolling Optimal Charging Schedule for Plug-In Hybrid Electric Vehicle Based on Statistical Energy Demand Analysis and Heuristic Algorithm

    Directory of Open Access Journals (Sweden)

    Fanrong Kong

    2017-09-01

    Full Text Available To alleviate the emission of greenhouse gas and the dependence on fossil fuel, Plug-in Hybrid Electrical Vehicles (PHEVs have gained an increasing popularity in current decades. Due to the fluctuating electricity prices in the power market, a charging schedule is very influential to driving cost. Although the next-day electricity prices can be obtained in a day-ahead power market, a driving plan is not easily made in advance. Although PHEV owners can input a next-day plan into a charging system, e.g., aggregators, day-ahead, it is a very trivial task to do everyday. Moreover, the driving plan may not be very accurate. To address this problem, in this paper, we analyze energy demands according to a PHEV owner’s historical driving records and build a personalized statistic driving model. Based on the model and the electricity spot prices, a rolling optimization strategy is proposed to help make a charging decision in the current time slot. On one hand, by employing a heuristic algorithm, the schedule is made according to the situations in the following time slots. On the other hand, however, after the current time slot, the schedule will be remade according to the next tens of time slots. Hence, the schedule is made by a dynamic rolling optimization, but it only decides the charging decision in the current time slot. In this way, the fluctuation of electricity prices and driving routine are both involved in the scheduling. Moreover, it is not necessary for PHEV owners to input a day-ahead driving plan. By the optimization simulation, the results demonstrate that the proposed method is feasible to help owners save charging costs and also meet requirements for driving.

  15. Web-based education for low-literate parents in Neonatal Intensive Care Unit: development of a website and heuristic evaluation and usability testing.

    Science.gov (United States)

    Choi, Jeungok; Bakken, Suzanne

    2010-08-01

    Low health literacy has been associated with poor health-related outcomes. The purposes are to report the development of a website for low-literate parents in the Neonatal Intensive Care Unit (NICU), and the findings of heuristic evaluation and a usability testing of this website. To address low literacy of NICU parents, multimedia educational Website using visual aids (e.g., pictographs, photographs), voice-recorded text message in addition to a simplified text was developed. The text was created at the 5th grade readability level. The heuristic evaluation was conducted by three usability experts using 10 heuristics. End-users' performance was measured by counting the time spent completing tasks and number of errors, as well as recording users' perception of ease of use and usefulness (PEUU) in a sample of 10 NICU parents. Three evaluators identified 82 violations across the 10 heuristics. All violations, however, received scores visuals on the Website were well accepted by low-literate users and agreement of visuals with text improved understanding of the educational materials over that with text alone. The findings suggest that using concrete and realistic pictures and pictographs with clear captions would maximize the benefit of visuals. One emerging theme was "simplicity" in design (e.g., limited use of colors, one font type and size), content (e.g., avoid lengthy text), and technical features (e.g., limited use of pop-ups). The heuristic evaluation by usability experts and the usability test with actual users provided complementary expertise, which can give a richer assessment of a design for low literacy Website. These results facilitated design modification and implementation of solutions by categorizing and prioritizing the usability problems.

  16. Nuclear component design ontology building based on ASME codes

    International Nuclear Information System (INIS)

    Bao Shiyi; Zhou Yu; He Shuyan

    2005-01-01

    The adoption of ontology analysis in the study of concept knowledge acquisition and representation for the nuclear component design process based on computer-supported cooperative work (CSCW) makes it possible to share and reuse numerous concept knowledge of multi-disciplinary domains. A practical ontology building method is accordingly proposed based on Protege knowledge model in combination with both top-down and bottom-up approaches together with Formal Concept Analysis (FCA). FCA exhibits its advantages in the way it helps establish and improve taxonomic hierarchy of concepts and resolve concept conflict occurred in modeling multi-disciplinary domains. With Protege-3.0 as the ontology building tool, a nuclear component design ontology based ASME codes is developed by utilizing the ontology building method. The ontology serves as the basis to realize concept knowledge sharing and reusing of nuclear component design. (authors)

  17. IAEA's experience in compiling a generic component reliability data base

    International Nuclear Information System (INIS)

    Tomic, B.; Lederman, L.

    1988-01-01

    Reliability data are an essential part of probabilistic safety assessment. The quality of data can determine the quality of the study as a whole. It is obvious that component failure data originated from the plant being analyzed would be most appropriate. However, in few cases complete reliance on plant experience is possible, mainly because of the rather limited operating experience. Nuclear plants, although of different design, often use fairly similar components, so some of the experience could be combined and transferred from one plant to another. In addition information about component failures is available also from experts with knowledge on component design, manufacturing and operation. That bring us to the importance of assessing generic data. (Generic is meant to be everything that is not plant specific regarding the plant being analyzed). The generic data available in the open literature, can be divided in three broad categories. The first one includes data base used in previous analysis. These can be plant specific or updated from generic with plant specific information (latter case deserve special attention). The second one is based on compilation of plants' operating experience usually based on some kind of event reporting system. The third category includes data sources based on expert opinions (single or aggregate) or combination of expert opinions and other nuclear and non-nuclear experience. This paper reflects insights gained in compiling data from generic data sources and highlights advantages and pitfalls of using generic component reliability data in PSAs

  18. A Chaotic Particle Swarm Optimization-Based Heuristic for Market-Oriented Task-Level Scheduling in Cloud Workflow Systems.

    Science.gov (United States)

    Li, Xuejun; Xu, Jia; Yang, Yun

    2015-01-01

    Cloud workflow system is a kind of platform service based on cloud computing. It facilitates the automation of workflow applications. Between cloud workflow system and its counterparts, market-oriented business model is one of the most prominent factors. The optimization of task-level scheduling in cloud workflow system is a hot topic. As the scheduling is a NP problem, Ant Colony Optimization (ACO) and Particle Swarm Optimization (PSO) have been proposed to optimize the cost. However, they have the characteristic of premature convergence in optimization process and therefore cannot effectively reduce the cost. To solve these problems, Chaotic Particle Swarm Optimization (CPSO) algorithm with chaotic sequence and adaptive inertia weight factor is applied to present the task-level scheduling. Chaotic sequence with high randomness improves the diversity of solutions, and its regularity assures a good global convergence. Adaptive inertia weight factor depends on the estimate value of cost. It makes the scheduling avoid premature convergence by properly balancing between global and local exploration. The experimental simulation shows that the cost obtained by our scheduling is always lower than the other two representative counterparts.

  19. A Chaotic Particle Swarm Optimization-Based Heuristic for Market-Oriented Task-Level Scheduling in Cloud Workflow Systems

    Directory of Open Access Journals (Sweden)

    Xuejun Li

    2015-01-01

    Full Text Available Cloud workflow system is a kind of platform service based on cloud computing. It facilitates the automation of workflow applications. Between cloud workflow system and its counterparts, market-oriented business model is one of the most prominent factors. The optimization of task-level scheduling in cloud workflow system is a hot topic. As the scheduling is a NP problem, Ant Colony Optimization (ACO and Particle Swarm Optimization (PSO have been proposed to optimize the cost. However, they have the characteristic of premature convergence in optimization process and therefore cannot effectively reduce the cost. To solve these problems, Chaotic Particle Swarm Optimization (CPSO algorithm with chaotic sequence and adaptive inertia weight factor is applied to present the task-level scheduling. Chaotic sequence with high randomness improves the diversity of solutions, and its regularity assures a good global convergence. Adaptive inertia weight factor depends on the estimate value of cost. It makes the scheduling avoid premature convergence by properly balancing between global and local exploration. The experimental simulation shows that the cost obtained by our scheduling is always lower than the other two representative counterparts.

  20. Heuristics and Cognitive Error in Medical Imaging.

    Science.gov (United States)

    Itri, Jason N; Patel, Sohil H

    2018-05-01

    The field of cognitive science has provided important insights into mental processes underlying the interpretation of imaging examinations. Despite these insights, diagnostic error remains a major obstacle in the goal to improve quality in radiology. In this article, we describe several types of cognitive bias that lead to diagnostic errors in imaging and discuss approaches to mitigate cognitive biases and diagnostic error. Radiologists rely on heuristic principles to reduce complex tasks of assessing probabilities and predicting values into simpler judgmental operations. These mental shortcuts allow rapid problem solving based on assumptions and past experiences. Heuristics used in the interpretation of imaging studies are generally helpful but can sometimes result in cognitive biases that lead to significant errors. An understanding of the causes of cognitive biases can lead to the development of educational content and systematic improvements that mitigate errors and improve the quality of care provided by radiologists.

  1. Towards a Component Based Model for Database Systems

    Directory of Open Access Journals (Sweden)

    Octavian Paul ROTARU

    2004-02-01

    Full Text Available Due to their effectiveness in the design and development of software applications and due to their recognized advantages in terms of reusability, Component-Based Software Engineering (CBSE concepts have been arousing a great deal of interest in recent years. This paper presents and extends a component-based approach to object-oriented database systems (OODB introduced by us in [1] and [2]. Components are proposed as a new abstraction level for database system, logical partitions of the schema. In this context, the scope is introduced as an escalated property for transactions. Components are studied from the integrity, consistency, and concurrency control perspective. The main benefits of our proposed component model for OODB are the reusability of the database design, including the access statistics required for a proper query optimization, and a smooth information exchange. The integration of crosscutting concerns into the component database model using aspect-oriented techniques is also discussed. One of the main goals is to define a method for the assessment of component composition capabilities. These capabilities are restricted by the component’s interface and measured in terms of adaptability, degree of compose-ability and acceptability level. The above-mentioned metrics are extended from database components to generic software components. This paper extends and consolidates into one common view the ideas previously presented by us in [1, 2, 3].[1] Octavian Paul Rotaru, Marian Dobre, Component Aspects in Object Oriented Databases, Proceedings of the International Conference on Software Engineering Research and Practice (SERP’04, Volume II, ISBN 1-932415-29-7, pages 719-725, Las Vegas, NV, USA, June 2004.[2] Octavian Paul Rotaru, Marian Dobre, Mircea Petrescu, Integrity and Consistency Aspects in Component-Oriented Databases, Proceedings of the International Symposium on Innovation in Information and Communication Technology (ISIICT

  2. Impact of heuristics in clustering large biological networks.

    Science.gov (United States)

    Shafin, Md Kishwar; Kabir, Kazi Lutful; Ridwan, Iffatur; Anannya, Tasmiah Tamzid; Karim, Rashid Saadman; Hoque, Mohammad Mozammel; Rahman, M Sohel

    2015-12-01

    Traditional clustering algorithms often exhibit poor performance for large networks. On the contrary, greedy algorithms are found to be relatively efficient while uncovering functional modules from large biological networks. The quality of the clusters produced by these greedy techniques largely depends on the underlying heuristics employed. Different heuristics based on different attributes and properties perform differently in terms of the quality of the clusters produced. This motivates us to design new heuristics for clustering large networks. In this paper, we have proposed two new heuristics and analyzed the performance thereof after incorporating those with three different combinations in a recently celebrated greedy clustering algorithm named SPICi. We have extensively analyzed the effectiveness of these new variants. The results are found to be promising. Copyright © 2015 Elsevier Ltd. All rights reserved.

  3. Component based modelling of piezoelectric ultrasonic actuators for machining applications

    International Nuclear Information System (INIS)

    Saleem, A; Ahmed, N; Salah, M; Silberschmidt, V V

    2013-01-01

    Ultrasonically Assisted Machining (UAM) is an emerging technology that has been utilized to improve the surface finishing in machining processes such as turning, milling, and drilling. In this context, piezoelectric ultrasonic transducers are being used to vibrate the cutting tip while machining at predetermined amplitude and frequency. However, modelling and simulation of these transducers is a tedious and difficult task. This is due to the inherent nonlinearities associated with smart materials. Therefore, this paper presents a component-based model of ultrasonic transducers that mimics the nonlinear behaviour of such a system. The system is decomposed into components, a mathematical model of each component is created, and the whole system model is accomplished by aggregating the basic components' model. System parameters are identified using Finite Element technique which then has been used to simulate the system in Matlab/SIMULINK. Various operation conditions are tested and performed to demonstrate the system performance

  4. Context sensitivity and ambiguity in component-based systems design

    Energy Technology Data Exchange (ETDEWEB)

    Bespalko, S.J.; Sindt, A.

    1997-10-01

    Designers of components-based, real-time systems need to guarantee to correctness of soft-ware and its output. Complexity of a system, and thus the propensity for error, is best characterized by the number of states a component can encounter. In many cases, large numbers of states arise where the processing is highly dependent on context. In these cases, states are often missed, leading to errors. The following are proposals for compactly specifying system states which allow the factoring of complex components into a control module and a semantic processing module. Further, the need for methods that allow for the explicit representation of ambiguity and uncertainty in the design of components is discussed. Presented herein are examples of real-world problems which are highly context-sensitive or are inherently ambiguous.

  5. Heuristic space diversity management in a meta-hyper-heuristic framework

    CSIR Research Space (South Africa)

    Grobler, J

    2014-07-01

    Full Text Available This paper introduces the concept of heuristic space diversity and investigates various strategies for the management of heuristic space diversity within the context of a meta-hyper-heuristic algorithm. Evaluation on a diverse set of floating...

  6. Heuristic space diversity control for improved meta-hyper-heuristic performance

    CSIR Research Space (South Africa)

    Grobler, J

    2015-04-01

    Full Text Available This paper expands on the concept of heuristic space diversity and investigates various strategies for the management of heuristic space diversity within the context of a meta-hyper-heuristic algorithm in search of greater performance benefits...

  7. New insights into diversification of hyper-heuristics.

    Science.gov (United States)

    Ren, Zhilei; Jiang, He; Xuan, Jifeng; Hu, Yan; Luo, Zhongxuan

    2014-10-01

    There has been a growing research trend of applying hyper-heuristics for problem solving, due to their ability of balancing the intensification and the diversification with low level heuristics. Traditionally, the diversification mechanism is mostly realized by perturbing the incumbent solutions to escape from local optima. In this paper, we report our attempt toward providing a new diversification mechanism, which is based on the concept of instance perturbation. In contrast to existing approaches, the proposed mechanism achieves the diversification by perturbing the instance under solving, rather than the solutions. To tackle the challenge of incorporating instance perturbation into hyper-heuristics, we also design a new hyper-heuristic framework HIP-HOP (recursive acronym of HIP-HOP is an instance perturbation-based hyper-heuristic optimization procedure), which employs a grammar guided high level strategy to manipulate the low level heuristics. With the expressive power of the grammar, the constraints, such as the feasibility of the output solution could be easily satisfied. Numerical results and statistical tests over both the Ising spin glass problem and the p -median problem instances show that HIP-HOP is able to achieve promising performances. Furthermore, runtime distribution analysis reveals that, although being relatively slow at the beginning, HIP-HOP is able to achieve competitive solutions once given sufficient time.

  8. Use of Heuristics to Facilitate Scientific Discovery Learning in a Simulation Learning Environment in a Physics Domain

    Science.gov (United States)

    Veermans, Koen; van Joolingen, Wouter; de Jong, Ton

    2006-01-01

    This article describes a study into the role of heuristic support in facilitating discovery learning through simulation-based learning. The study compares the use of two such learning environments in the physics domain of collisions. In one learning environment (implicit heuristics) heuristics are only used to provide the learner with guidance…

  9. Management of Globally Distributed Component-Based Software Development Projects

    NARCIS (Netherlands)

    J. Kotlarsky (Julia)

    2005-01-01

    textabstractGlobally Distributed Component-Based Development (GD CBD) is expected to become a promising area, as increasing numbers of companies are setting up software development in a globally distributed environment and at the same time are adopting CBD methodologies. Being an emerging area, the

  10. Monitoring extensions for component-based distributed software

    NARCIS (Netherlands)

    Diakov, N.K.; Papir, Z.; van Sinderen, Marten J.; Quartel, Dick

    2000-01-01

    This paper defines a generic class of monitoring extensions to component-based distributed enterprise software. Introducing a monitoring extension to a legacy application system can be very costly. In this paper, we identify the minimum support for application monitoring within the generic

  11. Integration of Simulink Models with Component-based Software Models

    DEFF Research Database (Denmark)

    Marian, Nicolae

    2008-01-01

    Model based development aims to facilitate the development of embedded control systems by emphasizing the separation of the design level from the implementation level. Model based design involves the use of multiple models that represent different views of a system, having different semantics...... of abstract system descriptions. Usually, in mechatronics systems, design proceeds by iterating model construction, model analysis, and model transformation. Constructing a MATLAB/Simulink model, a plant and controller behavior is simulated using graphical blocks to represent mathematical and logical...... constraints. COMDES (Component-based Design of Software for Distributed Embedded Systems) is such a component-based system framework developed by the software engineering group of Mads Clausen Institute for Product Innovation (MCI), University of Southern Denmark. Once specified, the software model has...

  12. Feature-based component model for design of embedded systems

    Science.gov (United States)

    Zha, Xuan Fang; Sriram, Ram D.

    2004-11-01

    An embedded system is a hybrid of hardware and software, which combines software's flexibility and hardware real-time performance. Embedded systems can be considered as assemblies of hardware and software components. An Open Embedded System Model (OESM) is currently being developed at NIST to provide a standard representation and exchange protocol for embedded systems and system-level design, simulation, and testing information. This paper proposes an approach to representing an embedded system feature-based model in OESM, i.e., Open Embedded System Feature Model (OESFM), addressing models of embedded system artifacts, embedded system components, embedded system features, and embedded system configuration/assembly. The approach provides an object-oriented UML (Unified Modeling Language) representation for the embedded system feature model and defines an extension to the NIST Core Product Model. The model provides a feature-based component framework allowing the designer to develop a virtual embedded system prototype through assembling virtual components. The framework not only provides a formal precise model of the embedded system prototype but also offers the possibility of designing variation of prototypes whose members are derived by changing certain virtual components with different features. A case study example is discussed to illustrate the embedded system model.

  13. The affect heuristic in occupational safety.

    Science.gov (United States)

    Savadori, Lucia; Caovilla, Jessica; Zaniboni, Sara; Fraccaroli, Franco

    2015-07-08

    The affect heuristic is a rule of thumb according to which, in the process of making a judgment or decision, people use affect as a cue. If a stimulus elicits positive affect then risks associated to that stimulus are viewed as low and benefits as high; conversely, if the stimulus elicits negative affect, then risks are perceived as high and benefits as low. The basic tenet of this study is that affect heuristic guides worker's judgment and decision making in a risk situation. The more the worker likes her/his organization the less she/he will perceive the risks as high. A sample of 115 employers and 65 employees working in small family agricultural businesses completed a questionnaire measuring perceived safety costs, psychological safety climate, affective commitment and safety compliance. A multi-sample structural analysis supported the thesis that safety compliance can be explained through an affect-based heuristic reasoning, but only for employers. Positive affective commitment towards their family business reduced employers' compliance with safety procedures by increasing the perceived cost of implementing them.

  14. When decision heuristics and science collide.

    Science.gov (United States)

    Yu, Erica C; Sprenger, Amber M; Thomas, Rick P; Dougherty, Michael R

    2014-04-01

    The ongoing discussion among scientists about null-hypothesis significance testing and Bayesian data analysis has led to speculation about the practices and consequences of "researcher degrees of freedom." This article advances this debate by asking the broader questions that we, as scientists, should be asking: How do scientists make decisions in the course of doing research, and what is the impact of these decisions on scientific conclusions? We asked practicing scientists to collect data in a simulated research environment, and our findings show that some scientists use data collection heuristics that deviate from prescribed methodology. Monte Carlo simulations show that data collection heuristics based on p values lead to biases in estimated effect sizes and Bayes factors and to increases in both false-positive and false-negative rates, depending on the specific heuristic. We also show that using Bayesian data collection methods does not eliminate these biases. Thus, our study highlights the little appreciated fact that the process of doing science is a behavioral endeavor that can bias statistical description and inference in a manner that transcends adherence to any particular statistical framework.

  15. Heuristic errors in clinical reasoning.

    Science.gov (United States)

    Rylander, Melanie; Guerrasio, Jeannette

    2016-08-01

    Errors in clinical reasoning contribute to patient morbidity and mortality. The purpose of this study was to determine the types of heuristic errors made by third-year medical students and first-year residents. This study surveyed approximately 150 clinical educators inquiring about the types of heuristic errors they observed in third-year medical students and first-year residents. Anchoring and premature closure were the two most common errors observed amongst third-year medical students and first-year residents. There was no difference in the types of errors observed in the two groups. Errors in clinical reasoning contribute to patient morbidity and mortality Clinical educators perceived that both third-year medical students and first-year residents committed similar heuristic errors, implying that additional medical knowledge and clinical experience do not affect the types of heuristic errors made. Further work is needed to help identify methods that can be used to reduce heuristic errors early in a clinician's education. © 2015 John Wiley & Sons Ltd.

  16. Neural basis of scientific innovation induced by heuristic prototype.

    Directory of Open Access Journals (Sweden)

    Junlong Luo

    Full Text Available A number of major inventions in history have been based on bionic imitation. Heuristics, by applying biological systems to the creation of artificial devices and machines, might be one of the most critical processes in scientific innovation. In particular, prototype heuristics propositions that innovation may engage automatic activation of a prototype such as a biological system to form novel associations between a prototype's function and problem-solving. We speculated that the cortical dissociation between the automatic activation and forming novel associations in innovation is critical point to heuristic creativity. In the present study, novel and old scientific innovations (NSI and OSI were selected as experimental materials in using learning-testing paradigm to explore the neural basis of scientific innovation induced by heuristic prototype. College students were required to resolve NSI problems (to which they did not know the answers and OSI problems (to which they knew the answers. From two fMRI experiments, our results showed that the subjects could resolve NSI when provided with heuristic prototypes. In Experiment 1, it was found that the lingual gyrus (LG; BA18 might be related to prototype heuristics in college students resolving NSI after learning a relative prototype. In Experiment 2, the LG (BA18 and precuneus (BA31 were significantly activated for NSI compared to OSI when college students learned all prototypes one day before the test. In addition, the mean beta-values of these brain regions of NSI were all correlated with the behavior accuracy of NSI. As our hypothesis indicated, the findings suggested that the LG might be involved in forming novel associations using heuristic information, while the precuneus might be involved in the automatic activation of heuristic prototype during scientific innovation.

  17. Neural basis of scientific innovation induced by heuristic prototype.

    Science.gov (United States)

    Luo, Junlong; Li, Wenfu; Qiu, Jiang; Wei, Dongtao; Liu, Yijun; Zhang, Qinlin

    2013-01-01

    A number of major inventions in history have been based on bionic imitation. Heuristics, by applying biological systems to the creation of artificial devices and machines, might be one of the most critical processes in scientific innovation. In particular, prototype heuristics propositions that innovation may engage automatic activation of a prototype such as a biological system to form novel associations between a prototype's function and problem-solving. We speculated that the cortical dissociation between the automatic activation and forming novel associations in innovation is critical point to heuristic creativity. In the present study, novel and old scientific innovations (NSI and OSI) were selected as experimental materials in using learning-testing paradigm to explore the neural basis of scientific innovation induced by heuristic prototype. College students were required to resolve NSI problems (to which they did not know the answers) and OSI problems (to which they knew the answers). From two fMRI experiments, our results showed that the subjects could resolve NSI when provided with heuristic prototypes. In Experiment 1, it was found that the lingual gyrus (LG; BA18) might be related to prototype heuristics in college students resolving NSI after learning a relative prototype. In Experiment 2, the LG (BA18) and precuneus (BA31) were significantly activated for NSI compared to OSI when college students learned all prototypes one day before the test. In addition, the mean beta-values of these brain regions of NSI were all correlated with the behavior accuracy of NSI. As our hypothesis indicated, the findings suggested that the LG might be involved in forming novel associations using heuristic information, while the precuneus might be involved in the automatic activation of heuristic prototype during scientific innovation.

  18. Component-based software for high-performance scientific computing

    Energy Technology Data Exchange (ETDEWEB)

    Alexeev, Yuri; Allan, Benjamin A; Armstrong, Robert C; Bernholdt, David E; Dahlgren, Tamara L; Gannon, Dennis; Janssen, Curtis L; Kenny, Joseph P; Krishnan, Manojkumar; Kohl, James A; Kumfert, Gary; McInnes, Lois Curfman; Nieplocha, Jarek; Parker, Steven G; Rasmussen, Craig; Windus, Theresa L

    2005-01-01

    Recent advances in both computational hardware and multidisciplinary science have given rise to an unprecedented level of complexity in scientific simulation software. This paper describes an ongoing grass roots effort aimed at addressing complexity in high-performance computing through the use of Component-Based Software Engineering (CBSE). Highlights of the benefits and accomplishments of the Common Component Architecture (CCA) Forum and SciDAC ISIC are given, followed by an illustrative example of how the CCA has been applied to drive scientific discovery in quantum chemistry. Thrusts for future research are also described briefly.

  19. Component-based software for high-performance scientific computing

    International Nuclear Information System (INIS)

    Alexeev, Yuri; Allan, Benjamin A; Armstrong, Robert C; Bernholdt, David E; Dahlgren, Tamara L; Gannon, Dennis; Janssen, Curtis L; Kenny, Joseph P; Krishnan, Manojkumar; Kohl, James A; Kumfert, Gary; McInnes, Lois Curfman; Nieplocha, Jarek; Parker, Steven G; Rasmussen, Craig; Windus, Theresa L

    2005-01-01

    Recent advances in both computational hardware and multidisciplinary science have given rise to an unprecedented level of complexity in scientific simulation software. This paper describes an ongoing grass roots effort aimed at addressing complexity in high-performance computing through the use of Component-Based Software Engineering (CBSE). Highlights of the benefits and accomplishments of the Common Component Architecture (CCA) Forum and SciDAC ISIC are given, followed by an illustrative example of how the CCA has been applied to drive scientific discovery in quantum chemistry. Thrusts for future research are also described briefly

  20. Integration of Simulink Models with Component-based Software Models

    Directory of Open Access Journals (Sweden)

    MARIAN, N.

    2008-06-01

    Full Text Available Model based development aims to facilitate the development of embedded control systems by emphasizing the separation of the design level from the implementation level. Model based design involves the use of multiple models that represent different views of a system, having different semantics of abstract system descriptions. Usually, in mechatronics systems, design proceeds by iterating model construction, model analysis, and model transformation. Constructing a MATLAB/Simulink model, a plant and controller behavior is simulated using graphical blocks to represent mathematical and logical constructs and process flow, then software code is generated. A Simulink model is a representation of the design or implementation of a physical system that satisfies a set of requirements. A software component-based system aims to organize system architecture and behavior as a means of computation, communication and constraints, using computational blocks and aggregates for both discrete and continuous behavior, different interconnection and execution disciplines for event-based and time-based controllers, and so on, to encompass the demands to more functionality, at even lower prices, and with opposite constraints. COMDES (Component-based Design of Software for Distributed Embedded Systems is such a component-based system framework developed by the software engineering group of Mads Clausen Institute for Product Innovation (MCI, University of Southern Denmark. Once specified, the software model has to be analyzed. One way of doing that is to integrate in wrapper files the model back into Simulink S-functions, and use its extensive simulation features, thus allowing an early exploration of the possible design choices over multiple disciplines. The paper describes a safe translation of a restricted set of MATLAB/Simulink blocks to COMDES software components, both for continuous and discrete behavior, and the transformation of the software system into the S

  1. Heuristic method of fabricating counter electrodes in dye-sensitized solar cells based on a PEDOT:PSS layer as a catalytic material

    International Nuclear Information System (INIS)

    Edalati, Sh; Houshangi far, A; Torabi, N; Baneshi, Z; Behjat, A

    2017-01-01

    Poly(3,4-ethylendioxythiophene):poly(styrene sulfonate) (PEDOT:PSS) was deposited on a fluoride-doped tin oxide glass substrate using a heuristic method to fabricate platinum-free counter electrodes for dye-sensitized solar cells (DSSCs). In this heuristic method a thin layer of PEDOT:PPS is obtained by spin coating the PEDOT:PSS on a Cu substrate and then removing the substrate with FeCl 3 . The characteristics of the deposited PEDOT:PSS were studied by energy dispersive x-ray analysis and scanning electron microscopy, which revealed the micro-electronic specifications of the cathode. The aforementioned DSSCs exhibited a solar conversion efficiency of 3.90%, which is far higher than that of DSSCs with pure PEDOT:PSS (1.89%). This enhancement is attributed not only to the micro-electronic specifications but also to the HNO 3 treatment through our heuristic method. The results of cyclic voltammetry, electrochemical impedance spectroscopy (EIS) and Tafel polarization plots show the modified cathode has a dual function, including excellent conductivity and electrocatalytic activity for iodine reduction. (paper)

  2. A probabilistic model for component-based shape synthesis

    KAUST Repository

    Kalogerakis, Evangelos

    2012-07-01

    We present an approach to synthesizing shapes from complex domains, by identifying new plausible combinations of components from existing shapes. Our primary contribution is a new generative model of component-based shape structure. The model represents probabilistic relationships between properties of shape components, and relates them to learned underlying causes of structural variability within the domain. These causes are treated as latent variables, leading to a compact representation that can be effectively learned without supervision from a set of compatibly segmented shapes. We evaluate the model on a number of shape datasets with complex structural variability and demonstrate its application to amplification of shape databases and to interactive shape synthesis. © 2012 ACM 0730-0301/2012/08-ART55.

  3. Empirical projection-based basis-component decomposition method

    Science.gov (United States)

    Brendel, Bernhard; Roessl, Ewald; Schlomka, Jens-Peter; Proksa, Roland

    2009-02-01

    Advances in the development of semiconductor based, photon-counting x-ray detectors stimulate research in the domain of energy-resolving pre-clinical and clinical computed tomography (CT). For counting detectors acquiring x-ray attenuation in at least three different energy windows, an extended basis component decomposition can be performed in which in addition to the conventional approach of Alvarez and Macovski a third basis component is introduced, e.g., a gadolinium based CT contrast material. After the decomposition of the measured projection data into the basis component projections, conventional filtered-backprojection reconstruction is performed to obtain the basis-component images. In recent work, this basis component decomposition was obtained by maximizing the likelihood-function of the measurements. This procedure is time consuming and often unstable for excessively noisy data or low intrinsic energy resolution of the detector. Therefore, alternative procedures are of interest. Here, we introduce a generalization of the idea of empirical dual-energy processing published by Stenner et al. to multi-energy, photon-counting CT raw data. Instead of working in the image-domain, we use prior spectral knowledge about the acquisition system (tube spectra, bin sensitivities) to parameterize the line-integrals of the basis component decomposition directly in the projection domain. We compare this empirical approach with the maximum-likelihood (ML) approach considering image noise and image bias (artifacts) and see that only moderate noise increase is to be expected for small bias in the empirical approach. Given the drastic reduction of pre-processing time, the empirical approach is considered a viable alternative to the ML approach.

  4. Penjadwalan Produksi Garment Menggunakan Algoritma Heuristic Pour

    Directory of Open Access Journals (Sweden)

    Rizal Rachman

    2018-04-01

    resources efficiently. The basic calculation of Scheduling using Heuristic Pour algorithm. The research stages consist of data collection, standard time calculation, total time calculation based on job, scheduling with company start method, scheduling with Pour Heuristic method. Based on the results of scheduling using Pour Heuristik obtained savings compared with the current company method, so it can be used as an alternative method in scheduling the process of production process in Garment company. Keywords: Production Scheduling, Algorithms, Heuristic Pour.

  5. Heuristic Search Theory and Applications

    CERN Document Server

    Edelkamp, Stefan

    2011-01-01

    Search has been vital to artificial intelligence from the very beginning as a core technique in problem solving. The authors present a thorough overview of heuristic search with a balance of discussion between theoretical analysis and efficient implementation and application to real-world problems. Current developments in search such as pattern databases and search with efficient use of external memory and parallel processing units on main boards and graphics cards are detailed. Heuristic search as a problem solving tool is demonstrated in applications for puzzle solving, game playing, constra

  6. The role of heuristics in automated theorem proving J.A Robinson's resolution principle

    OpenAIRE

    Coderschi, Roberto

    1996-01-01

    The aim of this paper is to show how J.A. Robinson's resolution principle was perceived and discussed in the AI community between the mid sixties and the first seventies. During this time the so called ``heuristic search paradigm" was still influential in the AI community, and both resolution principle and certain resolution based, apparently human-like, search strategies were matched with those problem solving heuristic procedures which were representative of the AI heuristic search paradigm.

  7. Perceived breast cancer risk: Heuristic reasoning and search for a dominance structure

    OpenAIRE

    Katapodi, M. C.; Facione, N. C.; Humphreys, J. C.; Dodd, MJ.

    2005-01-01

    Studies suggest that people construct their risk perceptions by using inferential rules called heuristics. The purpose of this study was to identify heuristics that influence perceived breast cancer risk. We examined 11 interviews from women of diverse ethnic/cultural backgrounds who were recruited from community settings. Narratives in which women elaborated about their own breast cancer risk were analyzed with Argument and Heuristic Reasoning Analysis methodology, which is based on applied ...

  8. Polarized BRDF for coatings based on three-component assumption

    Science.gov (United States)

    Liu, Hong; Zhu, Jingping; Wang, Kai; Xu, Rong

    2017-02-01

    A pBRDF(polarized bidirectional reflection distribution function) model for coatings is given based on three-component reflection assumption in order to improve the polarized scattering simulation capability for space objects. In this model, the specular reflection is given based on microfacet theory, the multiple reflection and volume scattering are given separately according to experimental results. The polarization of specular reflection is considered from Fresnel's law, and both multiple reflection and volume scattering are assumed depolarized. Simulation and measurement results of two satellite coating samples SR107 and S781 are given to validate that the pBRDF modeling accuracy can be significantly improved by the three-component model given in this paper.

  9. Secure wireless embedded systems via component-based design

    DEFF Research Database (Denmark)

    Hjorth, T.; Torbensen, R.

    2010-01-01

    This paper introduces the method secure-by-design as a way of constructing wireless embedded systems using component-based modeling frameworks. This facilitates design of secure applications through verified, reusable software. Following this method we propose a security framework with a secure c......, with full support for confidentiality, authentication, and integrity using keypairs. The approach has been demonstrated in a multi-platform home automation prototype that can remotely unlock a door using a PDA over the Internet....

  10. Research on Air Quality Evaluation based on Principal Component Analysis

    Science.gov (United States)

    Wang, Xing; Wang, Zilin; Guo, Min; Chen, Wei; Zhang, Huan

    2018-01-01

    Economic growth has led to environmental capacity decline and the deterioration of air quality. Air quality evaluation as a fundamental of environmental monitoring and air pollution control has become increasingly important. Based on the principal component analysis (PCA), this paper evaluates the air quality of a large city in Beijing-Tianjin-Hebei Area in recent 10 years and identifies influencing factors, in order to provide reference to air quality management and air pollution control.

  11. Implementing components of the routines-based model

    OpenAIRE

    McWilliam, Robin; Fernández Valero, Rosa

    2015-01-01

    The MBR is comprised of 17 components that can generally be grouped into practices related to (a) functional assessment and intervention planning (for example, Routines-Based Interview), (b) organization of services (including location and staffing), (c) service delivery to children and families (using a consultative approach with families and teachers, integrated therapy), (d) classroom organization (for example, classroom zones), and (e) supervision and training through ch...

  12. Heuristics for the Robust Coloring Problem

    Directory of Open Access Journals (Sweden)

    Miguel Ángel Gutiérrez Andrade

    2011-03-01

    Full Text Available Let $G$ and $\\bar{G}$ be complementary graphs. Given a penalty function defined on the edges of $G$, we will say that the rigidity of a $k$-coloring of $G$ is the sum of the penalties of the edges of G joining vertices of the same color. Based on the previous definition, the Robust Coloring Problem (RCP is stated as the search of the minimum rigidity $k$-coloring. In this work a comparison of heuristics based on simulated annealing, GRASP and scatter search is presented. These are the best results for the RCP that have been obtained.

  13. Empirical usability testing in a component-based environment : improving test efficiency with component-specific usability measures

    NARCIS (Netherlands)

    Brinkman, W.P.; Haakma, R.; Bouwhuis, D.G.; Bastide, R.; Palanque, P.; Roth, J.

    2005-01-01

    This paper addresses the issue of usability testing in a component-based software engineering environment, specifically measuring the usability of different versions of a component in a more powerful manner than other, more holistic, usability methods. Three component-specific usability measures are

  14. Multi-spectrometer calibration transfer based on independent component analysis.

    Science.gov (United States)

    Liu, Yan; Xu, Hao; Xia, Zhenzhen; Gong, Zhiyong

    2018-02-26

    Calibration transfer is indispensable for practical applications of near infrared (NIR) spectroscopy due to the need for precise and consistent measurements across different spectrometers. In this work, a method for multi-spectrometer calibration transfer is described based on independent component analysis (ICA). A spectral matrix is first obtained by aligning the spectra measured on different spectrometers. Then, by using independent component analysis, the aligned spectral matrix is decomposed into the mixing matrix and the independent components of different spectrometers. These differing measurements between spectrometers can then be standardized by correcting the coefficients within the independent components. Two NIR datasets of corn and edible oil samples measured with three and four spectrometers, respectively, were used to test the reliability of this method. The results of both datasets reveal that spectra measurements across different spectrometers can be transferred simultaneously and that the partial least squares (PLS) models built with the measurements on one spectrometer can predict that the spectra can be transferred correctly on another.

  15. Heuristic Biases in Mathematical Reasoning

    Science.gov (United States)

    Inglis, Matthew; Simpson, Adrian

    2005-01-01

    In this paper we briefly describe the dual process account of reasoning, and explain the role of heuristic biases in human thought. Concentrating on the so-called matching bias effect, we describe a piece of research that indicates a correlation between success at advanced level mathematics and an ability to override innate and misleading…

  16. Heuristic reasoning and relative incompleteness

    NARCIS (Netherlands)

    Treur, J.

    1993-01-01

    In this paper an approach is presented in which heuristic reasoning is interpreted as strategic reasoning. This type of reasoning enables one to derive which hypothesis to investigate, and which observable information to acquire next (to be able to verify the chosen hypothesis). A compositional

  17. Cognitive components underpinning the development of model-based learning.

    Science.gov (United States)

    Potter, Tracey C S; Bryce, Nessa V; Hartley, Catherine A

    2017-06-01

    Reinforcement learning theory distinguishes "model-free" learning, which fosters reflexive repetition of previously rewarded actions, from "model-based" learning, which recruits a mental model of the environment to flexibly select goal-directed actions. Whereas model-free learning is evident across development, recruitment of model-based learning appears to increase with age. However, the cognitive processes underlying the development of model-based learning remain poorly characterized. Here, we examined whether age-related differences in cognitive processes underlying the construction and flexible recruitment of mental models predict developmental increases in model-based choice. In a cohort of participants aged 9-25, we examined whether the abilities to infer sequential regularities in the environment ("statistical learning"), maintain information in an active state ("working memory") and integrate distant concepts to solve problems ("fluid reasoning") predicted age-related improvements in model-based choice. We found that age-related improvements in statistical learning performance did not mediate the relationship between age and model-based choice. Ceiling performance on our working memory assay prevented examination of its contribution to model-based learning. However, age-related improvements in fluid reasoning statistically mediated the developmental increase in the recruitment of a model-based strategy. These findings suggest that gradual development of fluid reasoning may be a critical component process underlying the emergence of model-based learning. Copyright © 2016 The Authors. Published by Elsevier Ltd.. All rights reserved.

  18. Aeromagnetic Compensation Algorithm Based on Principal Component Analysis

    Directory of Open Access Journals (Sweden)

    Peilin Wu

    2018-01-01

    Full Text Available Aeromagnetic exploration is an important exploration method in geophysics. The data is typically measured by optically pumped magnetometer mounted on an aircraft. But any aircraft produces significant levels of magnetic interference. Therefore, aeromagnetic compensation is important in aeromagnetic exploration. However, multicollinearity of the aeromagnetic compensation model degrades the performance of the compensation. To address this issue, a novel aeromagnetic compensation method based on principal component analysis is proposed. Using the algorithm, the correlation in the feature matrix is eliminated and the principal components are using to construct the hyperplane to compensate the platform-generated magnetic fields. The algorithm was tested using a helicopter, and the obtained improvement ratio is 9.86. The compensated quality is almost the same or slightly better than the ridge regression. The validity of the proposed method was experimentally demonstrated.

  19. Robust LOD scores for variance component-based linkage analysis.

    Science.gov (United States)

    Blangero, J; Williams, J T; Almasy, L

    2000-01-01

    The variance component method is now widely used for linkage analysis of quantitative traits. Although this approach offers many advantages, the importance of the underlying assumption of multivariate normality of the trait distribution within pedigrees has not been studied extensively. Simulation studies have shown that traits with leptokurtic distributions yield linkage test statistics that exhibit excessive Type I error when analyzed naively. We derive analytical formulae relating the deviation from the expected asymptotic distribution of the lod score to the kurtosis and total heritability of the quantitative trait. A simple correction constant yields a robust lod score for any deviation from normality and for any pedigree structure, and effectively eliminates the problem of inflated Type I error due to misspecification of the underlying probability model in variance component-based linkage analysis.

  20. A Component Based Approach to Scientific Workflow Management

    CERN Document Server

    Le Goff, Jean-Marie; Baker, Nigel; Brooks, Peter; McClatchey, Richard

    2001-01-01

    CRISTAL is a distributed scientific workflow system used in the manufacturing and production phases of HEP experiment construction at CERN. The CRISTAL project has studied the use of a description driven approach, using meta- modelling techniques, to manage the evolving needs of a large physics community. Interest from such diverse communities as bio-informatics and manufacturing has motivated the CRISTAL team to re-engineer the system to customize functionality according to end user requirements but maximize software reuse in the process. The next generation CRISTAL vision is to build a generic component architecture from which a complete software product line can be generated according to the particular needs of the target enterprise. This paper discusses the issues of adopting a component product line based approach and our experiences of software reuse.

  1. A component based approach to scientific workflow management

    International Nuclear Information System (INIS)

    Baker, N.; Brooks, P.; McClatchey, R.; Kovacs, Z.; LeGoff, J.-M.

    2001-01-01

    CRISTAL is a distributed scientific workflow system used in the manufacturing and production phases of HEP experiment construction at CERN. The CRISTAL project has studied the use of a description driven approach, using meta-modelling techniques, to manage the evolving needs of a large physics community. Interest from such diverse communities as bio-informatics and manufacturing has motivated the CRISTAL team to re-engineer the system to customize functionality according to end user requirements but maximize software reuse in the process. The next generation CRISTAL vision is to build a generic component architecture from which a complete software product line can be generated according to the particular needs of the target enterprise. This paper discusses the issues of adopting a component product line based approach and our experiences of software reuse

  2. Efficient heuristics for the Rural Postman Problem

    Directory of Open Access Journals (Sweden)

    GW Groves

    2005-06-01

    Full Text Available A local search framework for the (undirected Rural Postman Problem (RPP is presented in this paper. The framework allows local search approaches that have been applied successfully to the well–known Travelling Salesman Problem also to be applied to the RPP. New heuristics for the RPP, based on this framework, are introduced and these are capable of solving significantly larger instances of the RPP than have been reported in the literature. Test results are presented for a number of benchmark RPP instances in a bid to compare efficiency and solution quality against known methods.

  3. Advances in heuristic signal processing and applications

    CERN Document Server

    Chatterjee, Amitava; Siarry, Patrick

    2013-01-01

    There have been significant developments in the design and application of algorithms for both one-dimensional signal processing and multidimensional signal processing, namely image and video processing, with the recent focus changing from a step-by-step procedure of designing the algorithm first and following up with in-depth analysis and performance improvement to instead applying heuristic-based methods to solve signal-processing problems. In this book the contributing authors demonstrate both general-purpose algorithms and those aimed at solving specialized application problems, with a spec

  4. Laser-Based Maintenance and Repair Technologies for Reactor Components

    International Nuclear Information System (INIS)

    Masaki Yoda; Naruhiko Mukai; Makoto Ochiai; Masataka Tamura; Satoshi Okada; Katsuhiko Sato; Motohiko Kimura; Yuji Sano; Noboru Saito; Seishi Shima; Tetsuo Yamamoto

    2004-01-01

    Toshiba has developed various laser-based maintenance and repair technologies and applied them to existing nuclear power plants. Laser-based technology is considered to be the best tool for remote processing in nuclear power plants, and particularly so for the maintenance and repair of reactor core components. Accessibility could be drastically improved by a simple handling system owing to the absence of reactive force against laser irradiation and the flexible optical fiber. For the preventive maintenance, laser peening (LP) technology was developed and applied to reactor components in operating BWR plants. LP is a novel process to improve residual stress from tensile to compressive on material surface layer by irradiating focused high-power laser pulses in water. We have developed a fiber-delivered LP system as a preventive maintenance measure against stress corrosion cracking (SCC). Laser ultrasonic testing (LUT) has a great potential to be applied to the remote inspection of reactor components. Laser-induced surface acoustic wave (SAW) inspection system was developed using a compact probe with a multi-mode optical fiber and an interferometer. The developed system successfully detected a micro slit of 0.5 mm depth on weld metal and heat-affected zone (HAZ). An artificial SCC was also detected by the system. We are developing a new LP system combined with LUT to treat the inner surface of bottom-mounted instruments (BMI) of PWR plants. Underwater laser seal welding (LSW) technology was also developed to apply surface crack. LSW is expected to isolate the crack tip from corrosive water environment and to stop the propagation of the crack. Rapid heating and cooling of the process minimize the heat effect, which extends the applicability to neutron-irradiated material. This paper describes recent advances in the development and application of such laser-based technologies. (authors)

  5. Principal Component Analysis Based Measure of Structural Holes

    Science.gov (United States)

    Deng, Shiguo; Zhang, Wenqing; Yang, Huijie

    2013-02-01

    Based upon principal component analysis, a new measure called compressibility coefficient is proposed to evaluate structural holes in networks. This measure incorporates a new effect from identical patterns in networks. It is found that compressibility coefficient for Watts-Strogatz small-world networks increases monotonically with the rewiring probability and saturates to that for the corresponding shuffled networks. While compressibility coefficient for extended Barabasi-Albert scale-free networks decreases monotonically with the preferential effect and is significantly large compared with that for corresponding shuffled networks. This measure is helpful in diverse research fields to evaluate global efficiency of networks.

  6. Bayou Choctaw Well Integrity Grading Component Based on Geomechanical Simulation

    Energy Technology Data Exchange (ETDEWEB)

    Park, Byoung [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States). Geotechnology & Engineering Dept.

    2016-09-08

    This letter report provides a Bayou Choctaw (BC) Strategic Petroleum Reserve (SPR) well grading system based on the geomechanical simulation. The analyses described in this letter were used to evaluate the caverns’ geomechanical effect on wellbore integrity, which is an important component in the well integrity grading system recently developed by Roberts et al. [2015]. Using these analyses, the wellbores for caverns BC-17 and 20 are expected to be significantly impacted by cavern geomechanics, BC-18 and 19 are expected to be medium impacted; and the other caverns are expected to be less impacted.

  7. Reasoning by analogy as an aid to heuristic theorem proving.

    Science.gov (United States)

    Kling, R. E.

    1972-01-01

    When heuristic problem-solving programs are faced with large data bases that contain numbers of facts far in excess of those needed to solve any particular problem, their performance rapidly deteriorates. In this paper, the correspondence between a new unsolved problem and a previously solved analogous problem is computed and invoked to tailor large data bases to manageable sizes. This paper outlines the design of an algorithm for generating and exploiting analogies between theorems posed to a resolution-logic system. These algorithms are believed to be the first computationally feasible development of reasoning by analogy to be applied to heuristic theorem proving.

  8. Availability Allocation of Networked Systems Using Markov Model and Heuristics Algorithm

    Directory of Open Access Journals (Sweden)

    Ruiying Li

    2014-01-01

    Full Text Available It is a common practice to allocate the system availability goal to reliability and maintainability goals of components in the early design phase. However, the networked system availability is difficult to be allocated due to its complex topology and multiple down states. To solve these problems, a practical availability allocation method is proposed. Network reliability algebraic methods are used to derive the availability expression of the networked topology on the system level, and Markov model is introduced to determine that on the component level. A heuristic algorithm is proposed to obtain the reliability and maintainability allocation values of components. The principles applied in the AGREE reliability allocation method, proposed by the Advisory Group on Reliability of Electronic Equipment, and failure rate-based maintainability allocation method persist in our allocation method. A series system is used to verify the new algorithm, and the result shows that the allocation based on the heuristic algorithm is quite accurate compared to the traditional one. Moreover, our case study of a signaling system number 7 shows that the proposed allocation method is quite efficient for networked systems.

  9. [Development and application of component-based Chinese medicine theory].

    Science.gov (United States)

    Zhang, Jun-Hua; Fan, Guan-Wei; Zhang, Han; Fan, Xiao-Hui; Wang, Yi; Liu, Li-Mei; Li, Chuan; Gao, Yue; Gao, Xiu-Mei; Zhang, Bo-Li

    2017-11-01

    Traditional Chinese medicine (TCM) prescription is the main therapies for disease prevention and treatment in Chinese medicine. Following the guidance of the theory of TCM and developing drug by composing prescriptions of TCM materials and pieces, it is a traditional application mode of TCM, and still widely used in clinic. TCM prescription has theoretical advantages and rich clinical application experience in dealing with multi-factor complex diseases, but scientific research is relatively weak. The lack of scientific cognition of the effective substances and mechanism of Chinese medicine leads to insufficient understanding of the efficacy regularity, which affects the stability of effect and hinders the improvement of quality of Chinese medicinal products. Component-based Chinese medicine (CCM) is an innovation based on inheritance, which breaks through the tradition of experience-based prescription and realize the transformation of compatibility from herbal pieces to components. CCM is an important achievement during the research process of modernization of Chinese medicine. Under the support of three national "973" projects, in order to reveal the scientific connotation of the prescription compatibility theory and develop innovative Chinese drugs, we have launched theoretical innovation and technological innovation around the "two relatively clear", and opened up the research field of CCM. CCM is an innovation based on inheritance, breaking through the tradition of experience based prescription, and realizing the transformation from compatibility of herbal pieces to component compatibility, which is an important achievement of the modernization of traditional Chinese medicine. In the past more than 10 years, with the deepening of research and the expansion of application, the theory and methods of CCM and efficacy-oriented compatibility have been continuously improved. The value of CCM is not only in developing new drug, more important is to build a

  10. Nonlinear Process Fault Diagnosis Based on Serial Principal Component Analysis.

    Science.gov (United States)

    Deng, Xiaogang; Tian, Xuemin; Chen, Sheng; Harris, Chris J

    2018-03-01

    Many industrial processes contain both linear and nonlinear parts, and kernel principal component analysis (KPCA), widely used in nonlinear process monitoring, may not offer the most effective means for dealing with these nonlinear processes. This paper proposes a new hybrid linear-nonlinear statistical modeling approach for nonlinear process monitoring by closely integrating linear principal component analysis (PCA) and nonlinear KPCA using a serial model structure, which we refer to as serial PCA (SPCA). Specifically, PCA is first applied to extract PCs as linear features, and to decompose the data into the PC subspace and residual subspace (RS). Then, KPCA is performed in the RS to extract the nonlinear PCs as nonlinear features. Two monitoring statistics are constructed for fault detection, based on both the linear and nonlinear features extracted by the proposed SPCA. To effectively perform fault identification after a fault is detected, an SPCA similarity factor method is built for fault recognition, which fuses both the linear and nonlinear features. Unlike PCA and KPCA, the proposed method takes into account both linear and nonlinear PCs simultaneously, and therefore, it can better exploit the underlying process's structure to enhance fault diagnosis performance. Two case studies involving a simulated nonlinear process and the benchmark Tennessee Eastman process demonstrate that the proposed SPCA approach is more effective than the existing state-of-the-art approach based on KPCA alone, in terms of nonlinear process fault detection and identification.

  11. Integration of Simulink Models with Component-based Software Models

    DEFF Research Database (Denmark)

    Marian, Nicolae; Top, Søren

    2008-01-01

    , communication and constraints, using computational blocks and aggregates for both discrete and continuous behaviour, different interconnection and execution disciplines for event-based and time-based controllers, and so on, to encompass the demands to more functionality, at even lower prices, and with opposite...... to be analyzed. One way of doing that is to integrate in wrapper files the model back into Simulink S-functions, and use its extensive simulation features, thus allowing an early exploration of the possible design choices over multiple disciplines. The paper describes a safe translation of a restricted set...... of MATLAB/Simulink blocks to COMDES software components, both for continuous and discrete behaviour, and the transformation of the software system into the S-functions. The general aim of this work is the improvement of multi-disciplinary development of embedded systems with the focus on the relation...

  12. Sensitivity analysis of a greedy heuristic for knapsack problems

    NARCIS (Netherlands)

    Ghosh, D; Chakravarti, N; Sierksma, G

    2006-01-01

    In this paper, we carry out parametric analysis as well as a tolerance limit based sensitivity analysis of a greedy heuristic for two knapsack problems-the 0-1 knapsack problem and the subset sum problem. We carry out the parametric analysis based on all problem parameters. In the tolerance limit

  13. Managing Heuristics as a Method of Inquiry in Autobiographical Graphic Design Theses

    Science.gov (United States)

    Ings, Welby

    2011-01-01

    This article draws on case studies undertaken in postgraduate research at AUT University, Auckland. It seeks to address a number of issues related to heuristic inquiries employed by graphic design students who use autobiographical approaches when developing research-based theses. For this type of thesis, heuristics as a system of inquiry may…

  14. Heuristic Processes in Ratings of Leader Behavior: Assessing Item-Induced Availability Biases.

    Science.gov (United States)

    Binning, John F.; Fernandez, Guadalupe

    Since observers' memory-based ratings of organizational phenomena provide data in research and decision-making contexts, bias in observers' judgments must be examined. A study was conducted to explore the extent to which leader behavior ratings are more generally biased by the availability heuristic. The availability heuristic is operative when a…

  15. Neural Signatures of Rational and Heuristic Choice Strategies: A Single Trial ERP Analysis

    Directory of Open Access Journals (Sweden)

    Szymon Wichary

    2017-08-01

    Full Text Available In multi-attribute choice, people use heuristics to simplify decision problems. We studied the use of heuristic and rational strategies and their electrophysiological correlates. Since previous work linked the P3 ERP component to attention and decision making, we were interested whether the amplitude of this component is associated with decision strategy use. To this end, we recorded EEG when participants performed a two-alternative choice task, where they could acquire decision cues in a sequential manner and use them to make choices. We classified participants’ choices as consistent with a rational Weighted Additive rule (WADD or a simple heuristic Take The Best (TTB. Participants differed in their preference for WADD and TTB. Using a permutation-based single trial approach, we analyzed EEG responses to consecutive decision cues and their relation to the individual strategy preference. The preference for WADD over TTB was associated with overall higher signal amplitudes to decision cues in the P3 time window. Moreover, the preference for WADD was associated with similar P3 amplitudes to consecutive cues, whereas the preference for TTB was associated with substantial decreases in P3 amplitudes to consecutive cues. We also found that the preference for TTB was associated with enhanced N1 component to cues that discriminated decision alternatives, suggesting very early attention allocation to such cues by TTB users. Our results suggest that preference for either WADD or TTB has an early neural signature reflecting differences in attentional weighting of decision cues. In light of recent findings and hypotheses regarding P3, we interpret these results as indicating the involvement of catecholamine arousal systems in shaping predecisional information processing and strategy selection.

  16. Neural Signatures of Rational and Heuristic Choice Strategies: A Single Trial ERP Analysis.

    Science.gov (United States)

    Wichary, Szymon; Magnuski, Mikołaj; Oleksy, Tomasz; Brzezicka, Aneta

    2017-01-01

    In multi-attribute choice, people use heuristics to simplify decision problems. We studied the use of heuristic and rational strategies and their electrophysiological correlates. Since previous work linked the P3 ERP component to attention and decision making, we were interested whether the amplitude of this component is associated with decision strategy use. To this end, we recorded EEG when participants performed a two-alternative choice task, where they could acquire decision cues in a sequential manner and use them to make choices. We classified participants' choices as consistent with a rational Weighted Additive rule (WADD) or a simple heuristic Take The Best (TTB). Participants differed in their preference for WADD and TTB. Using a permutation-based single trial approach, we analyzed EEG responses to consecutive decision cues and their relation to the individual strategy preference. The preference for WADD over TTB was associated with overall higher signal amplitudes to decision cues in the P3 time window. Moreover, the preference for WADD was associated with similar P3 amplitudes to consecutive cues, whereas the preference for TTB was associated with substantial decreases in P3 amplitudes to consecutive cues. We also found that the preference for TTB was associated with enhanced N1 component to cues that discriminated decision alternatives, suggesting very early attention allocation to such cues by TTB users. Our results suggest that preference for either WADD or TTB has an early neural signature reflecting differences in attentional weighting of decision cues. In light of recent findings and hypotheses regarding P3, we interpret these results as indicating the involvement of catecholamine arousal systems in shaping predecisional information processing and strategy selection.

  17. Inhibitory mechanism of the matching heuristic in syllogistic reasoning.

    Science.gov (United States)

    Tse, Ping Ping; Moreno Ríos, Sergio; García-Madruga, Juan Antonio; Bajo Molina, María Teresa

    2014-11-01

    A number of heuristic-based hypotheses have been proposed to explain how people solve syllogisms with automatic processes. In particular, the matching heuristic employs the congruency of the quantifiers in a syllogism—by matching the quantifier of the conclusion with those of the two premises. When the heuristic leads to an invalid conclusion, successful solving of these conflict problems requires the inhibition of automatic heuristic processing. Accordingly, if the automatic processing were based on processing the set of quantifiers, no semantic contents would be inhibited. The mental model theory, however, suggests that people reason using mental models, which always involves semantic processing. Therefore, whatever inhibition occurs in the processing implies the inhibition of the semantic contents. We manipulated the validity of the syllogism and the congruency of the quantifier of its conclusion with those of the two premises according to the matching heuristic. A subsequent lexical decision task (LDT) with related words in the conclusion was used to test any inhibition of the semantic contents after each syllogistic evaluation trial. In the LDT, the facilitation effect of semantic priming diminished after correctly solved conflict syllogisms (match-invalid or mismatch-valid), but was intact after no-conflict syllogisms. The results suggest the involvement of an inhibitory mechanism of semantic contents in syllogistic reasoning when there is a conflict between the output of the syntactic heuristic and actual validity. Our results do not support a uniquely syntactic process of syllogistic reasoning but fit with the predictions based on mental model theory. Copyright © 2014 Elsevier B.V. All rights reserved.

  18. Generalized perturbation theory (GPT) methods. A heuristic approach

    International Nuclear Information System (INIS)

    Gandini, A.

    1987-01-01

    Wigner first proposed a perturbation theory as early as 1945 to study fundamental quantities such as the reactivity worths of different materials. The first formulation, CPT, for conventional perturbation theory is based on universal quantum mechanics concepts. Since that early conception, significant contributions have been made to CPT, in particular, Soodak, who rendered a heuristic interpretation of the adjoint function, (referred to as the GPT method for generalized perturbation theory). The author illustrates the GPT methodology in a variety of linear and nonlinear domains encountered in nuclear reactor analysis. The author begins with the familiar linear neutron field and then generalizes the methodology to other linear and nonlinear fields, using heuristic arguments. The author believes that the inherent simplicity and elegance of the heuristic derivation, although intended here for reactor physics problems might be usefully adopted in collateral fields and includes such examples

  19. Modelling raster-based monthly water balance components for Europe

    Energy Technology Data Exchange (ETDEWEB)

    Ulmen, C.

    2000-11-01

    The terrestrial runoff component is a comparatively small but sensitive and thus significant quantity in the global energy and water cycle at the interface between landmass and atmosphere. As opposed to soil moisture and evapotranspiration which critically determine water vapour fluxes and thus water and energy transport, it can be measured as an integrated quantity over a large area, i.e. the river basin. This peculiarity makes terrestrial runoff ideally suited for the calibration, verification and validation of general circulation models (GCMs). Gauging stations are not homogeneously distributed in space. Moreover, time series are not necessarily continuously measured nor do they in general have overlapping time periods. To overcome this problems with regard to regular grid spacing used in GCMs, different methods can be applied to transform irregular data to regular so called gridded runoff fields. The present work aims to directly compute the gridded components of the monthly water balance (including gridded runoff fields) for Europe by application of the well-established raster-based macro-scale water balance model WABIMON used at the Federal Institute of Hydrology, Germany. Model calibration and validation is performed by separated examination of 29 representative European catchments. Results indicate a general applicability of the model delivering reliable overall patterns and integrated quantities on a monthly basis. For time steps less then too weeks further research and structural improvements of the model are suggested. (orig.)

  20. Deriving a Set of Privacy Specific Heuristics for the Assessment of PHRs (Personal Health Records).

    Science.gov (United States)

    Furano, Riccardo F; Kushniruk, Andre; Barnett, Jeff

    2017-01-01

    With the emergence of personal health record (PHR) platforms becoming more widely available, this research focused on the development of privacy heuristics to assess PHRs regarding privacy. Existing sets of heuristics are typically not application specific and do not address patient-centric privacy as a main concern prior to undergoing PHR procurement. A set of privacy specific heuristics were developed based on a scoping review of the literature. An internet-based commercially available, vendor specific PHR application was evaluated using the derived set of privacy specific heuristics. The proposed set of privacy specific derived heuristics is explored in detail in relation to ISO 29100. The assessment of the internet-based commercially available, vendor specific PHR application indicated numerous violations. These violations were noted within the study. It is argued that the new derived privacy heuristics should be used in addition to Nielsen's well-established set of heuristics. Privacy specific heuristics could be used to assess PHR portal system-level privacy mechanisms in the procurement process of a PHR application and may prove to be a beneficial form of assessment to prevent the selection of a PHR platform with a poor privacy specific interface design.

  1. Hyper-heuristic applied to nuclear reactor core design

    International Nuclear Information System (INIS)

    Domingos, R P; Platt, G M

    2013-01-01

    The design of nuclear reactors gives rises to a series of optimization problems because of the need for high efficiency, availability and maintenance of security levels. Gradient-based techniques and linear programming have been applied, as well as genetic algorithms and particle swarm optimization. The nonlinearity, multimodality and lack of knowledge about the problem domain makes de choice of suitable meta-heuristic models particularly challenging. In this work we solve the optimization problem of a nuclear reactor core design through the application of an optimal sequence of meta-heuritics created automatically. This combinatorial optimization model is known as hyper-heuristic.

  2. Job shop scheduling with makespan objective: A heuristic approach

    Directory of Open Access Journals (Sweden)

    Mohsen Ziaee

    2014-04-01

    Full Text Available Job shop has been considered as one of the most challenging scheduling problems and there are literally tremendous efforts on reducing the complexity of solution procedure for solving job shop problem. This paper presents a heuristic method to minimize makespan for different jobs in a job shop scheduling. The proposed model is based on a constructive procedure to obtain good quality schedules, very quickly. The performance of the proposed model of this paper is examined on standard benchmarks from the literature in order to evaluate its performance. Computational results show that, despite its simplicity, the proposed heuristic is computationally efficient and practical approach for the problem.

  3. High Q, Miniaturized LCP-Based Passive Components

    KAUST Repository

    Shamim, Atif

    2014-10-16

    Various methods and systems are provided for high Q, miniaturized LCP-based passive components. In one embodiment, among others, a spiral inductor includes a center connection and a plurality of inductors formed on a liquid crystal polymer (LCP) layer, the plurality of inductors concentrically spiraling out from the center connection. In another embodiment, a vertically intertwined inductor includes first and second inductors including a first section disposed on a side of the LCP layer forming a fraction of a turn and a second section disposed on another side of the LCP layer. At least a portion of the first section of the first inductor is substantially aligned with at least a portion of the second section of the second inductor and at least a portion of the first section of the second inductor is substantially aligned with at least a portion of the second section of the first inductor.

  4. High Q, Miniaturized LCP-Based Passive Components

    KAUST Repository

    Shamim, Atif; Arabi, Eyad A.

    2014-01-01

    Various methods and systems are provided for high Q, miniaturized LCP-based passive components. In one embodiment, among others, a spiral inductor includes a center connection and a plurality of inductors formed on a liquid crystal polymer (LCP) layer, the plurality of inductors concentrically spiraling out from the center connection. In another embodiment, a vertically intertwined inductor includes first and second inductors including a first section disposed on a side of the LCP layer forming a fraction of a turn and a second section disposed on another side of the LCP layer. At least a portion of the first section of the first inductor is substantially aligned with at least a portion of the second section of the second inductor and at least a portion of the first section of the second inductor is substantially aligned with at least a portion of the second section of the first inductor.

  5. Iris recognition based on robust principal component analysis

    Science.gov (United States)

    Karn, Pradeep; He, Xiao Hai; Yang, Shuai; Wu, Xiao Hong

    2014-11-01

    Iris images acquired under different conditions often suffer from blur, occlusion due to eyelids and eyelashes, specular reflection, and other artifacts. Existing iris recognition systems do not perform well on these types of images. To overcome these problems, we propose an iris recognition method based on robust principal component analysis. The proposed method decomposes all training images into a low-rank matrix and a sparse error matrix, where the low-rank matrix is used for feature extraction. The sparsity concentration index approach is then applied to validate the recognition result. Experimental results using CASIA V4 and IIT Delhi V1iris image databases showed that the proposed method achieved competitive performances in both recognition accuracy and computational efficiency.

  6. An auto-focusing heuristic model to increase the reliability of a scientific mission

    International Nuclear Information System (INIS)

    Gualdesi, Lavinio

    2006-01-01

    Researchers invest a lot of time and effort on the design and development of components used in a scientific mission. To capitalize on this investment and on the operational experience of the researchers, it is useful to adopt a quantitative data base to monitor the history and usage of the components. This work describes a model to monitor the reliability level of components. The model is very flexible and allows users to compose systems using the same components in different configurations as required by each mission. This tool provides availability and reliability figures for the configuration requested, derived from historical data of the components' previous performance. The system is based on preliminary checklists to establish standard operating procedures (SOP) for all components life phases. When an infringement to the SOP occurs, a quantitative ranking is provided in order to quantify the risk associated with this deviation. The final agreement between field data and expected performance of the component makes the model converge onto a heuristic monitoring system. The model automatically focuses on points of failure at the detailed component element level, calculates risks, provides alerts when a demonstrated risk to safety is encountered, and advises when there is a mismatch between component performance and mission requirements. This model also helps the mission to focus resources on critical tasks where they are most needed

  7. Performance of nickel base superalloy components in gas turbines

    DEFF Research Database (Denmark)

    Dahl, Kristian Vinter

    2006-01-01

    The topic of this thesis is the microstructural behaviour of hot section components in the industrial gas turbine......The topic of this thesis is the microstructural behaviour of hot section components in the industrial gas turbine...

  8. A Geographical Heuristic Routing Protocol for VANETs

    Science.gov (United States)

    Urquiza-Aguiar, Luis; Tripp-Barba, Carolina; Aguilar Igartua, Mónica

    2016-01-01

    Vehicular ad hoc networks (VANETs) leverage the communication system of Intelligent Transportation Systems (ITS). Recently, Delay-Tolerant Network (DTN) routing protocols have increased their popularity among the research community for being used in non-safety VANET applications and services like traffic reporting. Vehicular DTN protocols use geographical and local information to make forwarding decisions. However, current proposals only consider the selection of the best candidate based on a local-search. In this paper, we propose a generic Geographical Heuristic Routing (GHR) protocol that can be applied to any DTN geographical routing protocol that makes forwarding decisions hop by hop. GHR includes in its operation adaptations simulated annealing and Tabu-search meta-heuristics, which have largely been used to improve local-search results in discrete optimization. We include a complete performance evaluation of GHR in a multi-hop VANET simulation scenario for a reporting service. Our study analyzes all of the meaningful configurations of GHR and offers a statistical analysis of our findings by means of MANOVA tests. Our results indicate that the use of a Tabu list contributes to improving the packet delivery ratio by around 5% to 10%. Moreover, if Tabu is used, then the simulated annealing routing strategy gets a better performance than the selection of the best node used with carry and forwarding (default operation). PMID:27669254

  9. Evaluating the hydrological consistency of satellite based water cycle components

    KAUST Repository

    Lopez Valencia, Oliver Miguel

    2016-06-15

    Advances in multi-satellite based observations of the earth system have provided the capacity to retrieve information across a wide-range of land surface hydrological components and provided an opportunity to characterize terrestrial processes from a completely new perspective. Given the spatial advantage that space-based observations offer, several regional-to-global scale products have been developed, offering insights into the multi-scale behaviour and variability of hydrological states and fluxes. However, one of the key challenges in the use of satellite-based products is characterizing the degree to which they provide realistic and representative estimates of the underlying retrieval: that is, how accurate are the hydrological components derived from satellite observations? The challenge is intrinsically linked to issues of scale, since the availability of high-quality in-situ data is limited, and even where it does exist, is generally not commensurate to the resolution of the satellite observation. Basin-scale studies have shown considerable variability in achieving water budget closure with any degree of accuracy using satellite estimates of the water cycle. In order to assess the suitability of this type of approach for evaluating hydrological observations, it makes sense to first test it over environments with restricted hydrological inputs, before applying it to more hydrological complex basins. Here we explore the concept of hydrological consistency, i.e. the physical considerations that the water budget impose on the hydrologic fluxes and states to be temporally and spatially linked, to evaluate the reproduction of a set of large-scale evaporation (E) products by using a combination of satellite rainfall (P) and Gravity Recovery and Climate Experiment (GRACE) observations of storage change, focusing on arid and semi-arid environments, where the hydrological flows can be more realistically described. Our results indicate no persistent hydrological

  10. Heuristic Approach for Balancing Shift Schedules

    International Nuclear Information System (INIS)

    Kim, Dae Ho; Yun, Young Su; Lee, Yong Hee

    2005-01-01

    In this paper, a heuristic approach for balancing shift schedules is proposed. For the shift schedules, various constraints which have usually been considered in realworld industry are used, and the objective is to minimize the differences of the workloads in each workgroup. The constraints and objective function are implemented in the proposed heuristic approach. Using a simple instance, the efficiency of the proposed heuristic approach is proved

  11. Heuristic Model Of The Composite Quality Index Of Environmental Assessment

    Science.gov (United States)

    Khabarov, A. N.; Knyaginin, A. A.; Bondarenko, D. V.; Shepet, I. P.; Korolkova, L. N.

    2017-01-01

    The goal of the paper is to present the heuristic model of the composite environmental quality index based on the integrated application of the elements of utility theory, multidimensional scaling, expert evaluation and decision-making. The composite index is synthesized in linear-quadratic form, it provides higher adequacy of the results of the assessment preferences of experts and decision-makers.

  12. 78 FR 6344 - Certain Wireless Communications Base Stations and Components Thereof Notice of Receipt of...

    Science.gov (United States)

    2013-01-30

    ... INTERNATIONAL TRADE COMMISSION Certain Wireless Communications Base Stations and Components.... International Trade Commission has received a complaint entitled Certain Wireless Communications Base Stations... communications base stations and components thereof. The complaint names as respondents Telefonaktiebolaget LM...

  13. A single cognitive heuristic process meets the complexity of domain-specific moral heuristics.

    Science.gov (United States)

    Dubljević, Veljko; Racine, Eric

    2014-10-01

    The inherence heuristic (a) offers modest insights into the complex nature of both the is-ought tension in moral reasoning and moral reasoning per se, and (b) does not reflect the complexity of domain-specific moral heuristics. Formal and general in nature, we contextualize the process described as "inherence heuristic" in a web of domain-specific heuristics (e.g., agent specific; action specific; consequences specific).

  14. Application of fuzzy-MOORA method: Ranking of components for reliability estimation of component-based software systems

    Directory of Open Access Journals (Sweden)

    Zeeshan Ali Siddiqui

    2016-01-01

    Full Text Available Component-based software system (CBSS development technique is an emerging discipline that promises to take software development into a new era. As hardware systems are presently being constructed from kits of parts, software systems may also be assembled from components. It is more reliable to reuse software than to create. It is the glue code and individual components reliability that contribute to the reliability of the overall system. Every component contributes to overall system reliability according to the number of times it is being used, some components are of critical usage, known as usage frequency of component. The usage frequency decides the weight of each component. According to their weights, each component contributes to the overall reliability of the system. Therefore, ranking of components may be obtained by analyzing their reliability impacts on overall application. In this paper, we propose the application of fuzzy multi-objective optimization on the basis of ratio analysis, Fuzzy-MOORA. The method helps us find the best suitable alternative, software component, from a set of available feasible alternatives named software components. It is an accurate and easy to understand tool for solving multi-criteria decision making problems that have imprecise and vague evaluation data. By the use of ratio analysis, the proposed method determines the most suitable alternative among all possible alternatives, and dimensionless measurement will realize the job of ranking of components for estimating CBSS reliability in a non-subjective way. Finally, three case studies are shown to illustrate the use of the proposed technique.

  15. Heuristics and stock buying decision: Evidence from Malaysian and Pakistani stock markets

    Directory of Open Access Journals (Sweden)

    Habib Hussain Khan

    2017-06-01

    Full Text Available Applying both qualitative and quantitative approaches, we examine whether or not investors fall prey to three heuristics; namely, anchoring and adjustment, representativeness, and availability, while investing in stocks. We also compare investors' vulnerability to these heuristics based on their economic association, their type and demographic factors such as income, education and experience. For the data collection, a self-constructed questionnaire was administered to investors in the Malaysian and Pakistani stock exchanges. Data has been analyzed through description, correlation and regression analysis. The results indicate that all three heuristics are likely to affect the investors' stock buying decisions. The effect of heuristics is similar across the sample countries, the type of investors, and the income groups. However, the investors with a higher level of education and more experience are less likely to be affected by the heuristics.

  16. Cnn Based Retinal Image Upscaling Using Zero Component Analysis

    Science.gov (United States)

    Nasonov, A.; Chesnakov, K.; Krylov, A.

    2017-05-01

    The aim of the paper is to obtain high quality of image upscaling for noisy images that are typical in medical image processing. A new training scenario for convolutional neural network based image upscaling method is proposed. Its main idea is a novel dataset preparation method for deep learning. The dataset contains pairs of noisy low-resolution images and corresponding noiseless highresolution images. To achieve better results at edges and textured areas, Zero Component Analysis is applied to these images. The upscaling results are compared with other state-of-the-art methods like DCCI, SI-3 and SRCNN on noisy medical ophthalmological images. Objective evaluation of the results confirms high quality of the proposed method. Visual analysis shows that fine details and structures like blood vessels are preserved, noise level is reduced and no artifacts or non-existing details are added. These properties are essential in retinal diagnosis establishment, so the proposed algorithm is recommended to be used in real medical applications.

  17. Service oriented architecture assessment based on software components

    Directory of Open Access Journals (Sweden)

    Mahnaz Amirpour

    2016-01-01

    Full Text Available Enterprise architecture, with detailed descriptions of the functions of information technology in the organization, tries to reduce the complexity of technology applications resulting in tools with greater efficiency in achieving the objectives of the organization. Enterprise architecture consists of a set of models describing this technology in different components performance as well as various aspects of the applications in any organization. Therefore, information technology development and maintenance management can perform well within organizations. This study aims to suggest a method to identify different types of services in service-oriented architecture analysis step that applies some previous approaches in an integrated form and, based on the principles of software engineering, to provide a simpler and more transparent approach through the expression of analysis details. Advantages and disadvantages of proposals should be evaluated before the implementation and costs allocation. Evaluation methods can better identify strengths and weaknesses of the current situation apart from selecting appropriate model out of several suggestions, and clarify this technology development solution for organizations in the future. We will be able to simulate data and processes flow within the organization by converting the output of the model to colored Petri nets and evaluate and test it by examining various inputs to enterprise architecture before implemented in terms of reliability and response time. A model of application has been studied for the proposed model and the results can describe and design architecture for data.

  18. Internet MEMS design tools based on component technology

    Science.gov (United States)

    Brueck, Rainer; Schumer, Christian

    1999-03-01

    The micro electromechanical systems (MEMS) industry in Europe is characterized by small and medium sized enterprises specialized on products to solve problems in specific domains like medicine, automotive sensor technology, etc. In this field of business the technology driven design approach known from micro electronics is not appropriate. Instead each design problem aims at its own, specific technology to be used for the solution. The variety of technologies at hand, like Si-surface, Si-bulk, LIGA, laser, precision engineering requires a huge set of different design tools to be available. No single SME can afford to hold licenses for all these tools. This calls for a new and flexible way of designing, implementing and distributing design software. The Internet provides a flexible manner of offering software access along with methodologies of flexible licensing e.g. on a pay-per-use basis. New communication technologies like ADSL, TV cable of satellites as carriers promise to offer a bandwidth sufficient even for interactive tools with graphical interfaces in the near future. INTERLIDO is an experimental tool suite for process specification and layout verification for lithography based MEMS technologies to be accessed via the Internet. The first version provides a Java implementation even including a graphical editor for process specification. Currently, a new version is brought into operation that is based on JavaBeans component technology. JavaBeans offers the possibility to realize independent interactive design assistants, like a design rule checking assistants, a process consistency checking assistants, a technology definition assistants, a graphical editor assistants, etc. that may reside distributed over the Internet, communicating via Internet protocols. Each potential user thus is able to configure his own dedicated version of a design tool set dedicated to the requirements of the current problem to be solved.

  19. Conspicuous Waste and Representativeness Heuristic

    Directory of Open Access Journals (Sweden)

    Tatiana M. Shishkina

    2017-12-01

    Full Text Available The article deals with the similarities between conspicuous waste and representativeness heuristic. The conspicuous waste is analyzed according to the classic Veblen’ interpretation as a strategy to increase social status through conspicuous consumption and conspicuous leisure. In “The Theory of the Leisure Class” Veblen introduced two different types of utility – conspicuous and functional. The article focuses on the possible benefits of the analysis of conspicuous utility not only in terms of institutional economic theory, but also in terms of behavioral economics. To this end, the representativeness heuristics is considered, on the one hand, as a way to optimize the decision-making process, which allows to examine it in comparison with procedural rationality by Simon. On the other hand, it is also analyzed as cognitive bias within the Kahneman and Twersky’ approach. The article provides the analysis of the patterns in the deviations from the rational behavior strategy that could be observed in case of conspicuous waste both in modern market economies in the form of conspicuous consumption and in archaic economies in the form of gift-exchange. The article also focuses on the marketing strategies for luxury consumption’ advertisement. It highlights the impact of the symbolic capital (in Bourdieu’ interpretation on the social and symbolic payments that actors get from the act of conspicuous waste. This allows to perform a analysis of conspicuous consumption both as a rational way to get the particular kind of payments, and, at the same time, as a form of institutionalized cognitive bias.

  20. Heuristic Evaluation on Mobile Interfaces: A New Checklist

    Science.gov (United States)

    Yáñez Gómez, Rosa; Cascado Caballero, Daniel; Sevillano, José-Luis

    2014-01-01

    The rapid evolution and adoption of mobile devices raise new usability challenges, given their limitations (in screen size, battery life, etc.) as well as the specific requirements of this new interaction. Traditional evaluation techniques need to be adapted in order for these requirements to be met. Heuristic evaluation (HE), an Inspection Method based on evaluation conducted by experts over a real system or prototype, is based on checklists which are desktop-centred and do not adequately detect mobile-specific usability issues. In this paper, we propose a compilation of heuristic evaluation checklists taken from the existing bibliography but readapted to new mobile interfaces. Selecting and rearranging these heuristic guidelines offer a tool which works well not just for evaluation but also as a best-practices checklist. The result is a comprehensive checklist which is experimentally evaluated as a design tool. This experimental evaluation involved two software engineers without any specific knowledge about usability, a group of ten users who compared the usability of a first prototype designed without our heuristics, and a second one after applying the proposed checklist. The results of this experiment show the usefulness of the proposed checklist for avoiding usability gaps even with nontrained developers. PMID:25295300

  1. Heuristic Evaluation on Mobile Interfaces: A New Checklist

    Directory of Open Access Journals (Sweden)

    Rosa Yáñez Gómez

    2014-01-01

    Full Text Available The rapid evolution and adoption of mobile devices raise new usability challenges, given their limitations (in screen size, battery life, etc. as well as the specific requirements of this new interaction. Traditional evaluation techniques need to be adapted in order for these requirements to be met. Heuristic evaluation (HE, an Inspection Method based on evaluation conducted by experts over a real system or prototype, is based on checklists which are desktop-centred and do not adequately detect mobile-specific usability issues. In this paper, we propose a compilation of heuristic evaluation checklists taken from the existing bibliography but readapted to new mobile interfaces. Selecting and rearranging these heuristic guidelines offer a tool which works well not just for evaluation but also as a best-practices checklist. The result is a comprehensive checklist which is experimentally evaluated as a design tool. This experimental evaluation involved two software engineers without any specific knowledge about usability, a group of ten users who compared the usability of a first prototype designed without our heuristics, and a second one after applying the proposed checklist. The results of this experiment show the usefulness of the proposed checklist for avoiding usability gaps even with nontrained developers.

  2. Multistage principal component analysis based method for abdominal ECG decomposition

    International Nuclear Information System (INIS)

    Petrolis, Robertas; Krisciukaitis, Algimantas; Gintautas, Vladas

    2015-01-01

    Reflection of fetal heart electrical activity is present in registered abdominal ECG signals. However this signal component has noticeably less energy than concurrent signals, especially maternal ECG. Therefore traditionally recommended independent component analysis, fails to separate these two ECG signals. Multistage principal component analysis (PCA) is proposed for step-by-step extraction of abdominal ECG signal components. Truncated representation and subsequent subtraction of cardio cycles of maternal ECG are the first steps. The energy of fetal ECG component then becomes comparable or even exceeds energy of other components in the remaining signal. Second stage PCA concentrates energy of the sought signal in one principal component assuring its maximal amplitude regardless to the orientation of the fetus in multilead recordings. Third stage PCA is performed on signal excerpts representing detected fetal heart beats in aim to perform their truncated representation reconstructing their shape for further analysis. The algorithm was tested with PhysioNet Challenge 2013 signals and signals recorded in the Department of Obstetrics and Gynecology, Lithuanian University of Health Sciences. Results of our method in PhysioNet Challenge 2013 on open data set were: average score: 341.503 bpm 2 and 32.81 ms. (paper)

  3. A flexible framework for sparse simultaneous component based data integration

    Directory of Open Access Journals (Sweden)

    Van Deun Katrijn

    2011-11-01

    Full Text Available Abstract 1 Background High throughput data are complex and methods that reveal structure underlying the data are most useful. Principal component analysis, frequently implemented as a singular value decomposition, is a popular technique in this respect. Nowadays often the challenge is to reveal structure in several sources of information (e.g., transcriptomics, proteomics that are available for the same biological entities under study. Simultaneous component methods are most promising in this respect. However, the interpretation of the principal and simultaneous components is often daunting because contributions of each of the biomolecules (transcripts, proteins have to be taken into account. 2 Results We propose a sparse simultaneous component method that makes many of the parameters redundant by shrinking them to zero. It includes principal component analysis, sparse principal component analysis, and ordinary simultaneous component analysis as special cases. Several penalties can be tuned that account in different ways for the block structure present in the integrated data. This yields known sparse approaches as the lasso, the ridge penalty, the elastic net, the group lasso, sparse group lasso, and elitist lasso. In addition, the algorithmic results can be easily transposed to the context of regression. Metabolomics data obtained with two measurement platforms for the same set of Escherichia coli samples are used to illustrate the proposed methodology and the properties of different penalties with respect to sparseness across and within data blocks. 3 Conclusion Sparse simultaneous component analysis is a useful method for data integration: First, simultaneous analyses of multiple blocks offer advantages over sequential and separate analyses and second, interpretation of the results is highly facilitated by their sparseness. The approach offered is flexible and allows to take the block structure in different ways into account. As such

  4. A flexible framework for sparse simultaneous component based data integration.

    Science.gov (United States)

    Van Deun, Katrijn; Wilderjans, Tom F; van den Berg, Robert A; Antoniadis, Anestis; Van Mechelen, Iven

    2011-11-15

    High throughput data are complex and methods that reveal structure underlying the data are most useful. Principal component analysis, frequently implemented as a singular value decomposition, is a popular technique in this respect. Nowadays often the challenge is to reveal structure in several sources of information (e.g., transcriptomics, proteomics) that are available for the same biological entities under study. Simultaneous component methods are most promising in this respect. However, the interpretation of the principal and simultaneous components is often daunting because contributions of each of the biomolecules (transcripts, proteins) have to be taken into account. We propose a sparse simultaneous component method that makes many of the parameters redundant by shrinking them to zero. It includes principal component analysis, sparse principal component analysis, and ordinary simultaneous component analysis as special cases. Several penalties can be tuned that account in different ways for the block structure present in the integrated data. This yields known sparse approaches as the lasso, the ridge penalty, the elastic net, the group lasso, sparse group lasso, and elitist lasso. In addition, the algorithmic results can be easily transposed to the context of regression. Metabolomics data obtained with two measurement platforms for the same set of Escherichia coli samples are used to illustrate the proposed methodology and the properties of different penalties with respect to sparseness across and within data blocks. Sparse simultaneous component analysis is a useful method for data integration: First, simultaneous analyses of multiple blocks offer advantages over sequential and separate analyses and second, interpretation of the results is highly facilitated by their sparseness. The approach offered is flexible and allows to take the block structure in different ways into account. As such, structures can be found that are exclusively tied to one data platform

  5. The Probability Heuristics Model of Syllogistic Reasoning.

    Science.gov (United States)

    Chater, Nick; Oaksford, Mike

    1999-01-01

    Proposes a probability heuristic model for syllogistic reasoning and confirms the rationality of this heuristic by an analysis of the probabilistic validity of syllogistic reasoning that treats logical inference as a limiting case of probabilistic inference. Meta-analysis and two experiments involving 40 adult participants and using generalized…

  6. Cooperative heuristic multi-agent planning

    NARCIS (Netherlands)

    De Weerdt, M.M.; Tonino, J.F.M.; Witteveen, C.

    2001-01-01

    In this paper we will use the framework to study cooperative heuristic multi-agent planning. During the construction of their plans, the agents use a heuristic function inspired by the FF planner (l3l). At any time in the process of planning the agents may exchange available resources, or they may

  7. "A Heuristic for Visual Thinking in History"

    Science.gov (United States)

    Staley, David J.

    2007-01-01

    This article details a heuristic history teachers can use in assigning and evaluating multimedia projects in history. To use this heuristic successfully, requires more than simply following the steps in the list or stages in a recipe: in many ways, it requires a reorientation in what it means to think like an historian. This article, as much as…

  8. Effective Heuristics for New Venture Formation

    NARCIS (Netherlands)

    Kraaijenbrink, Jeroen

    2010-01-01

    Entrepreneurs are often under time pressure and may only have a short window of opportunity to launch their new venture. This means they often have no time for rational analytical decisions and rather rely on heuristics. Past research on entrepreneurial heuristics has primarily focused on predictive

  9. Religion, Heuristics, and Intergenerational Risk Management

    OpenAIRE

    Rupert Read; Nassim Nicholas Taleb

    2014-01-01

    Religions come with risk-​managing interdicts and heuristics, and they carry such interdicts and heuristics across generations. We remark on such facets of religion in relation to a propensity among some decision scientists and others to regard practices that they cannot understand as being irrational, biased, and so on.

  10. A probabilistic model for component-based shape synthesis

    KAUST Repository

    Kalogerakis, Evangelos; Chaudhuri, Siddhartha; Koller, Daphne; Koltun, Vladlen

    2012-01-01

    represents probabilistic relationships between properties of shape components, and relates them to learned underlying causes of structural variability within the domain. These causes are treated as latent variables, leading to a compact representation

  11. Perceived breast cancer risk: heuristic reasoning and search for a dominance structure.

    Science.gov (United States)

    Katapodi, Maria C; Facione, Noreen C; Humphreys, Janice C; Dodd, Marylin J

    2005-01-01

    Studies suggest that people construct their risk perceptions by using inferential rules called heuristics. The purpose of this study was to identify heuristics that influence perceived breast cancer risk. We examined 11 interviews from women of diverse ethnic/cultural backgrounds who were recruited from community settings. Narratives in which women elaborated about their own breast cancer risk were analyzed with Argument and Heuristic Reasoning Analysis methodology, which is based on applied logic. The availability, simulation, representativeness, affect, and perceived control heuristics, and search for a dominance structure were commonly used for making risk assessments. Risk assessments were based on experiences with an abnormal breast symptom, experiences with affected family members and friends, beliefs about living a healthy lifestyle, and trust in health providers. Assessment of the potential threat of a breast symptom was facilitated by the search for a dominance structure. Experiences with family members and friends were incorporated into risk assessments through the availability, simulation, representativeness, and affect heuristics. Mistrust in health providers led to an inappropriate dependence on the perceived control heuristic. Identified heuristics appear to create predictable biases and suggest that perceived breast cancer risk is based on common cognitive patterns.

  12. Evaluating Heuristics for Planning Effective and Efficient Inspections

    Science.gov (United States)

    Shull, Forrest J.; Seaman, Carolyn B.; Diep, Madeline M.; Feldmann, Raimund L.; Godfrey, Sara H.; Regardie, Myrna

    2010-01-01

    A significant body of knowledge concerning software inspection practice indicates that the value of inspections varies widely both within and across organizations. Inspection effectiveness and efficiency can be measured in numerous ways, and may be affected by a variety of factors such as Inspection planning, the type of software, the developing organization, and many others. In the early 1990's, NASA formulated heuristics for inspection planning based on best practices and early NASA inspection data. Over the intervening years, the body of data from NASA inspections has grown. This paper describes a multi-faceted exploratory analysis performed on this · data to elicit lessons learned in general about conducting inspections and to recommend improvements to the existing heuristics. The contributions of our results include support for modifying some of the original inspection heuristics (e.g. Increasing the recommended page rate), evidence that Inspection planners must choose between efficiency and effectiveness, as a good tradeoff between them may not exist, and Identification of small subsets of inspections for which new inspection heuristics are needed. Most Importantly, this work illustrates the value of collecting rich data on software Inspections, and using it to gain insight into, and Improve, inspection practice.

  13. Heuristics for Relevancy Ranking of Earth Dataset Search Results

    Science.gov (United States)

    Lynnes, Christopher; Quinn, Patrick; Norton, James

    2016-01-01

    As the Variety of Earth science datasets increases, science researchers find it more challenging to discover and select the datasets that best fit their needs. The most common way of search providers to address this problem is to rank the datasets returned for a query by their likely relevance to the user. Large web page search engines typically use text matching supplemented with reverse link counts, semantic annotations and user intent modeling. However, this produces uneven results when applied to dataset metadata records simply externalized as a web page. Fortunately, data and search provides have decades of experience in serving data user communities, allowing them to form heuristics that leverage the structure in the metadata together with knowledge about the user community. Some of these heuristics include specific ways of matching the user input to the essential measurements in the dataset and determining overlaps of time range and spatial areas. Heuristics based on the novelty of the datasets can prioritize later, better versions of data over similar predecessors. And knowledge of how different user types and communities use data can be brought to bear in cases where characteristics of the user (discipline, expertise) or their intent (applications, research) can be divined. The Earth Observing System Data and Information System has begun implementing some of these heuristics in the relevancy algorithm of its Common Metadata Repository search engine.

  14. How cognitive heuristics can explain social interactions in spatial movement.

    Science.gov (United States)

    Seitz, Michael J; Bode, Nikolai W F; Köster, Gerta

    2016-08-01

    The movement of pedestrian crowds is a paradigmatic example of collective motion. The precise nature of individual-level behaviours underlying crowd movements has been subject to a lively debate. Here, we propose that pedestrians follow simple heuristics rooted in cognitive psychology, such as 'stop if another step would lead to a collision' or 'follow the person in front'. In other words, our paradigm explicitly models individual-level behaviour as a series of discrete decisions. We show that our cognitive heuristics produce realistic emergent crowd phenomena, such as lane formation and queuing behaviour. Based on our results, we suggest that pedestrians follow different cognitive heuristics that are selected depending on the context. This differs from the widely used approach of capturing changes in behaviour via model parameters and leads to testable hypotheses on changes in crowd behaviour for different motivation levels. For example, we expect that rushed individuals more often evade to the side and thus display distinct emergent queue formations in front of a bottleneck. Our heuristics can be ranked according to the cognitive effort that is required to follow them. Therefore, our model establishes a direct link between behavioural responses and cognitive effort and thus facilitates a novel perspective on collective behaviour. © 2016 The Author(s).

  15. Heuristic Scheduling Algorithm Oriented Dynamic Tasks for Imaging Satellites

    Directory of Open Access Journals (Sweden)

    Maocai Wang

    2014-01-01

    Full Text Available Imaging satellite scheduling is an NP-hard problem with many complex constraints. This paper researches the scheduling problem for dynamic tasks oriented to some emergency cases. After the dynamic properties of satellite scheduling were analyzed, the optimization model is proposed in this paper. Based on the model, two heuristic algorithms are proposed to solve the problem. The first heuristic algorithm arranges new tasks by inserting or deleting them, then inserting them repeatedly according to the priority from low to high, which is named IDI algorithm. The second one called ISDR adopts four steps: insert directly, insert by shifting, insert by deleting, and reinsert the tasks deleted. Moreover, two heuristic factors, congestion degree of a time window and the overlapping degree of a task, are employed to improve the algorithm’s performance. Finally, a case is given to test the algorithms. The results show that the IDI algorithm is better than ISDR from the running time point of view while ISDR algorithm with heuristic factors is more effective with regard to algorithm performance. Moreover, the results also show that our method has good performance for the larger size of the dynamic tasks in comparison with the other two methods.

  16. A Heuristic T-S Fuzzy Model for the Pumped-Storage Generator-Motor Using Variable-Length Tree-Seed Algorithm-Based Competitive Agglomeration

    Directory of Open Access Journals (Sweden)

    Jianzhong Zhou

    2018-04-01

    Full Text Available With the fast development of artificial intelligence techniques, data-driven modeling approaches are becoming hotspots in both academic research and engineering practice. This paper proposes a novel data-driven T-S fuzzy model to precisely describe the complicated dynamic behaviors of pumped storage generator motor (PSGM. In premise fuzzy partition of the proposed T-S fuzzy model, a novel variable-length tree-seed algorithm based competitive agglomeration (VTSA-CA algorithm is presented to determine the optimal number of clusters automatically and improve the fuzzy clustering performances. Besides, in order to promote modeling accuracy of PSGM, the input and output formats in the T-S fuzzy model are selected by an economical parameter controlled auto-regressive (CAR model derived from a high-order transfer function of PSGM considering the distributed components in the water diversion system of the power plant. The effectiveness and superiority of the T-S fuzzy model for PSGM under different working conditions are validated by performing comparative studies with both practical data and the conventional mechanistic model.

  17. Cognitive biases and heuristics in medical decision making: a critical review using a systematic search strategy.

    Science.gov (United States)

    Blumenthal-Barby, J S; Krieger, Heather

    2015-05-01

    The role of cognitive biases and heuristics in medical decision making is of growing interest. The purpose of this study was to determine whether studies on cognitive biases and heuristics in medical decision making are based on actual or hypothetical decisions and are conducted with populations that are representative of those who typically make the medical decision; to categorize the types of cognitive biases and heuristics found and whether they are found in patients or in medical personnel; and to critically review the studies based on standard methodological quality criteria. Data sources were original, peer-reviewed, empirical studies on cognitive biases and heuristics in medical decision making found in Ovid Medline, PsycINFO, and the CINAHL databases published in 1980-2013. Predefined exclusion criteria were used to identify 213 studies. During data extraction, information was collected on type of bias or heuristic studied, respondent population, decision type, study type (actual or hypothetical), study method, and study conclusion. Of the 213 studies analyzed, 164 (77%) were based on hypothetical vignettes, and 175 (82%) were conducted with representative populations. Nineteen types of cognitive biases and heuristics were found. Only 34% of studies (n = 73) investigated medical personnel, and 68% (n = 145) confirmed the presence of a bias or heuristic. Each methodological quality criterion was satisfied by more than 50% of the studies, except for sample size and validated instruments/questions. Limitations are that existing terms were used to inform search terms, and study inclusion criteria focused strictly on decision making. Most of the studies on biases and heuristics in medical decision making are based on hypothetical vignettes, raising concerns about applicability of these findings to actual decision making. Biases and heuristics have been underinvestigated in medical personnel compared with patients. © The Author(s) 2014.

  18. A combined Component-Based Approach for the Design of Distributed Software Systems

    NARCIS (Netherlands)

    Guareis de farias, Cléver; Ferreira Pires, Luis; van Sinderen, Marten J.; Quartel, Dick; Yang, H.; Gupta, S.

    2001-01-01

    Component-based software development enables the construction of software artefacts by assembling binary units of production, distribution and deployment, the so-called components. Several approaches to component-based development have been proposed recently. Most of these approaches are based on

  19. Heuristic Strategies in Systems Biology

    Directory of Open Access Journals (Sweden)

    Fridolin Gross

    2016-06-01

    Full Text Available Systems biology is sometimes presented as providing a superior approach to the problem of biological complexity. Its use of ‘unbiased’ methods and formal quantitative tools might lead to the impression that the human factor is effectively eliminated. However, a closer look reveals that this impression is misguided. Systems biologists cannot simply assemble molecular information and compute biological behavior. Instead, systems biology’s main contribution is to accelerate the discovery of mechanisms by applying models as heuristic tools. These models rely on a variety of idealizing and simplifying assumptions in order to be efficient for this purpose. The strategies of systems biologists are similar to those of experimentalists in that they attempt to reduce the complexity of the discovery process. Analyzing and comparing these strategies, or ‘heuristics’, reveals the importance of the human factor in computational approaches and helps to situate systems biology within the epistemic landscape of the life sciences.

  20. Heuristic approach to train rescheduling

    Directory of Open Access Journals (Sweden)

    Mladenović Snežana

    2007-01-01

    Full Text Available Starting from the defined network topology and the timetable assigned beforehand, the paper considers a train rescheduling in respond to disturbances that have occurred. Assuming that the train trips are jobs, which require the elements of infrastructure - resources, it was done by the mapping of the initial problem into a special case of job shop scheduling problem. In order to solve the given problem, a constraint programming approach has been used. A support to fast finding "enough good" schedules is offered by original separation, bound and search heuristic algorithms. In addition, to improve the time performance, instead of the actual objective function with a large domain, a surrogate objective function is used with a smaller domain, if there is such. .

  1. THE HEURISTIC FUNCTION OF SPORT

    Directory of Open Access Journals (Sweden)

    Adam Petrović

    2012-09-01

    Full Text Available Being a significant area of human activity, sport has multiple functions. One of the more important functions of sport, especially top sport, is the inventive heuristic function. Creative work, being a process of creating new values, represents a significant possibility for advancement of sport. This paper aims at pointing at the various dimensions of human creative work, at the creative work which can be seen in sport (in a narrow sense and at the scientific and practical areas which borderline sport. The method of theoretical analysis of different approaches to the phenomenon of creative work , both in general and in sport, was applied in this paper. This area can be systematized according to various criterion : the level of creative work, different fields where it appears, the subjects of creative work - creators etc. Case analysis shows that the field of creative work in sport is widening and deepening constantly. There are different levels of creativity not only in the system of training and competition, but in a wider social context of sport as well. As a process of human spirit and mind the creative work belongs not just to athletes and coaches, but also to all the people and social groups who's creative power manifests itself in sport. The classification of creative work in sport according to various criterion allows for heuristic function of sport to be explained comprehensively and to create an image how do the sparks of human spirit improve the micro cosmos of sport. A thorough classification of creative work in sport allows for a detailed analysis of all the elements of creative work and each of it’s area in sport. In this way the progress in sport , as a consequence of innovations in both competitions and athletes’ training and of everything that goes with those activities, can be guided into the needed direction more easily as well as studied and applied.

  2. Interface Heuristics and Style Guide Design: An Air Battle Management Case Study

    National Research Council Canada - National Science Library

    Nelson, W. T; Bolia, Robert S

    2005-01-01

    .... An analysis of the content of extensive operator interviews from all relevant platforms preceded the production of a compact style guide based on a few simple heuristics and populated with wire frame...

  3. Optimized LTE cell planning for multiple user density subareas using meta-heuristic algorithms

    KAUST Repository

    Ghazzai, Hakim; Yaacoub, Elias E.; Alouini, Mohamed-Slim

    2014-01-01

    Base station deployment in cellular networks is one of the most fundamental problems in network design. This paper proposes a novel method for the cell planning problem for the fourth generation 4G-LTE cellular networks using meta heuristic

  4. FC-TLBO: fully constrained meta-heuristic algorithm for abundance ...

    Indian Academy of Sciences (India)

    Omprakash Tembhurne

    hyperspectral unmixing; meta-heuristic approach; teaching-learning-based optimisation (TLBO). 1. ... area of research due to its real-time applications. Satellite .... describes the detailed methodology of proposed FC-TLBO. Section 4 contains ...

  5. Opportunity-based block replacement: the single component case

    NARCIS (Netherlands)

    R. Dekker (Rommert); E. Smeitink

    1991-01-01

    textabstractIn this paper we consider a block replacement model in which a component can be replaced preventively at maintenance opportunities only. Maintenance opportunities occur randomly and are modelled through a renewal process. In the first, theoretical part of the paper we derive an

  6. Component Architectures and Web-Based Learning Environments

    Science.gov (United States)

    Ferdig, Richard E.; Mishra, Punya; Zhao, Yong

    2004-01-01

    The Web has caught the attention of many educators as an efficient communication medium and content delivery system. But we feel there is another aspect of the Web that has not been given the attention it deserves. We call this aspect of the Web its "component architecture." Briefly it means that on the Web one can develop very complex…

  7. Direct heuristic dynamic programming for damping oscillations in a large power system.

    Science.gov (United States)

    Lu, Chao; Si, Jennie; Xie, Xiaorong

    2008-08-01

    This paper applies a neural-network-based approximate dynamic programming method, namely, the direct heuristic dynamic programming (direct HDP), to a large power system stability control problem. The direct HDP is a learning- and approximation-based approach to addressing nonlinear coordinated control under uncertainty. One of the major design parameters, the controller learning objective function, is formulated to directly account for network-wide low-frequency oscillation with the presence of nonlinearity, uncertainty, and coupling effect among system components. Results include a novel learning control structure based on the direct HDP with applications to two power system problems. The first case involves static var compensator supplementary damping control, which is used to provide a comprehensive evaluation of the learning control performance. The second case aims at addressing a difficult complex system challenge by providing a new solution to a large interconnected power network oscillation damping control problem that frequently occurs in the China Southern Power Grid.

  8. A learning heuristic for space mapping and searching self-organizing systems using adaptive mesh refinement

    Science.gov (United States)

    Phillips, Carolyn L.

    2014-09-01

    In a complex self-organizing system, small changes in the interactions between the system's components can result in different emergent macrostructures or macrobehavior. In chemical engineering and material science, such spontaneously self-assembling systems, using polymers, nanoscale or colloidal-scale particles, DNA, or other precursors, are an attractive way to create materials that are precisely engineered at a fine scale. Changes to the interactions can often be described by a set of parameters. Different contiguous regions in this parameter space correspond to different ordered states. Since these ordered states are emergent, often experiment, not analysis, is necessary to create a diagram of ordered states over the parameter space. By issuing queries to points in the parameter space (e.g., performing a computational or physical experiment), ordered states can be discovered and mapped. Queries can be costly in terms of resources or time, however. In general, one would like to learn the most information using the fewest queries. Here we introduce a learning heuristic for issuing queries to map and search a two-dimensional parameter space. Using a method inspired by adaptive mesh refinement, the heuristic iteratively issues batches of queries to be executed in parallel based on past information. By adjusting the search criteria, different types of searches (for example, a uniform search, exploring boundaries, sampling all regions equally) can be flexibly implemented. We show that this method will densely search the space, while preferentially targeting certain features. Using numerical examples, including a study simulating the self-assembly of complex crystals, we show how this heuristic can discover new regions and map boundaries more accurately than a uniformly distributed set of queries.

  9. Combined heuristic with fuzzy system to transmission system expansion planning

    Energy Technology Data Exchange (ETDEWEB)

    Silva Sousa, Aldir; Asada, Eduardo N. [University of Sao Paulo, Sao Carlos School of Engineering, Department of Electrical Engineering Av. Trabalhador Sao-carlense, 400, 13566-590 Sao Carlos, SP (Brazil)

    2011-01-15

    A heuristic algorithm that employs fuzzy logic is proposed to the power system transmission expansion planning problem. The algorithm is based on the divide to conquer strategy, which is controlled by the fuzzy system. The algorithm provides high quality solutions with the use of fuzzy decision making, which is based on nondeterministic criteria to guide the search. The fuzzy system provides a self-adjusting mechanism that eliminates the manual adjustment of parameters to each system being solved. (author)

  10. A Computational Investigation of Heuristic Algorithms for 2-Edge-Connectivity Augmentation

    DEFF Research Database (Denmark)

    Bang-Jensen, Jørgen; Chiarandini, Marco; Morling, Peter

    2010-01-01

    an equivalent set covering   formulation.  The results indicate that exact solutions by means of a   basic integer programming model can be obtained in reasonably short   time even on networks with 800 vertices and around 287,000   edges. Alternatively, an advanced heuristic algorithm based on   subgradient...... programming, simple construction heuristics and   metaheuristics. As part of the design of heuristics, we consider   different neighborhood structures for local search, among which is a very   large scale neighborhood. In all cases, we exploit approaches through   the graph formulation as well as through...

  11. A two-component NZRI metamaterial based rectangular cloak

    Directory of Open Access Journals (Sweden)

    Sikder Sunbeam Islam

    2015-10-01

    Full Text Available A new two-component, near zero refractive index (NZRI metamaterial is presented for electromagnetic rectangular cloaking operation in the microwave range. In the basic design a pi-shaped, metamaterial was developed and its characteristics were investigated for the two major axes (x and z-axis wave propagation through the material. For the z-axis wave propagation, it shows more than 2 GHz bandwidth and for the x-axis wave propagation; it exhibits more than 1 GHz bandwidth of NZRI property. The metamaterial was then utilized in designing a rectangular cloak where a metal cylinder was cloaked perfectly in the C-band area of microwave regime. The experimental result was provided for the metamaterial and the cloak and these results were compared with the simulated results. This is a novel and promising design for its two-component NZRI characteristics and rectangular cloaking operation in the electromagnetic paradigm.

  12. Component Pin Recognition Using Algorithms Based on Machine Learning

    Science.gov (United States)

    Xiao, Yang; Hu, Hong; Liu, Ze; Xu, Jiangchang

    2018-04-01

    The purpose of machine vision for a plug-in machine is to improve the machine’s stability and accuracy, and recognition of the component pin is an important part of the vision. This paper focuses on component pin recognition using three different techniques. The first technique involves traditional image processing using the core algorithm for binary large object (BLOB) analysis. The second technique uses the histogram of oriented gradients (HOG), to experimentally compare the effect of the support vector machine (SVM) and the adaptive boosting machine (AdaBoost) learning meta-algorithm classifiers. The third technique is the use of an in-depth learning method known as convolution neural network (CNN), which involves identifying the pin by comparing a sample to its training. The main purpose of the research presented in this paper is to increase the knowledge of learning methods used in the plug-in machine industry in order to achieve better results.

  13. Systematic Heuristic Evaluation of Computerized Consultation Order Templates: Clinicians' and Human Factors Engineers' Perspectives.

    Science.gov (United States)

    Savoy, April; Patel, Himalaya; Flanagan, Mindy E; Weiner, Michael; Russ, Alissa L

    2017-08-01

    We assessed the usability of consultation order templates and identified problems to prioritize in design efforts for improving referral communication. With a sample of 26 consultation order templates, three evaluators performed a usability heuristic evaluation. The evaluation used 14 domain-independent heuristics and the following three supplemental references: 1 new domain-specific heuristic, 6 usability goals, and coded clinicians' statements regarding ease of use for 10 sampled templates. Evaluators found 201 violations, a mean of 7.7 violations per template. Minor violations outnumbered major violations almost twofold, 115 (57%) to 62 (31%). Approximately 68% of violations were linked to 5 heuristics: aesthetic and minimalist design (17%), error prevention (16%), consistency and standards (14%), recognition rather than recall (11%), and meet referrers' information needs (10%). Severe violations were attributed mostly to meet referrers' information needs and recognition rather than recall. Recorded violations yielded potential negative consequences for efficiency, effectiveness, safety, learnability, and utility. Evaluators and clinicians demonstrated 80% agreement in usability assessment. Based on frequency and severity of usability heuristic violations, the consultation order templates reviewed may impede clinical efficiency and risk patient safety. Results support the following design considerations: communicate consultants' requirements, facilitate information seeking, and support communication. While the most frequent heuristic violations involved interaction design and presentation, the most severe violations lacked information desired by referring clinicians. Violations related to templates' inability to support referring clinicians' information needs had the greatest potential negative impact on efficiency and safety usability goals. Heuristics should be prioritized in future design efforts.

  14. Mathematical models and heuristic solutions for container positioning problems in port terminals

    DEFF Research Database (Denmark)

    Kallehauge, Louise Sibbesen

    2008-01-01

    presents an efficient solution algorithm for the CPP. Based on a number of new concepts, an event-based construction heuristic is developed and its ability to solve real-life problem instances is established. The backbone of the algorithm is a list of events, corresponding to a sequence of operations...... by constructing mathematical programming formulations of the problem and developing an efficient heuristic algorithm for its solution. The thesis consists of an introduction, two main chapters concerning new mathematical formulations and a new heuristic for the CPP, technical issues, computational results...... concerning the subject is reviewed. The research presented in this thesis is divided into two main parts: Construction and investigation of new mathematical programming formulations of the CPP and development and implementation of a new event-based heuristic for the problem. The first part presents three...

  15. Learning Algorithms for Audio and Video Processing: Independent Component Analysis and Support Vector Machine Based Approaches

    National Research Council Canada - National Science Library

    Qi, Yuan

    2000-01-01

    In this thesis, we propose two new machine learning schemes, a subband-based Independent Component Analysis scheme and a hybrid Independent Component Analysis/Support Vector Machine scheme, and apply...

  16. A service component-based accounting and charging architecture to support interim mechanisms across multiple domains

    NARCIS (Netherlands)

    Le, M. van; Beijnum, B.J.F. van; Huitema, G.B.

    2004-01-01

    Today, telematics services are often compositions of different chargeable service components offered by different service providers. To enhance component-based accounting and charging, the service composition information is used to match with the corresponding charging structure of a service

  17. Evaluating the usability of an interactive, bi-lingual, touchscreen-enabled breastfeeding educational programme: application of Nielson's heuristics.

    Science.gov (United States)

    Joshi, Ashish; Perin, Douglas M Puricelli; Amadi, Chioma; Trout, Kate

    2015-03-05

    The study purpose was to conduct heuristic evaluation of an interactive, bilingual touchscreen-enabled breastfeeding educational programme for Hispanic women living in rural settings in Nebraska. Three raters conducted the evaluation during May 2013 using principles of Nielson's heuristics. A total of 271 screens were evaluated and included: interface (n = 5), programme sections (n = 223) and educational content (n = 43). A total of 97 heuristic violations were identified and were mostly related to interface (8 violations/5 screens) and programme components (89 violations/266 screens). The most common heuristic violations reported were recognition rather than recall (62%, n = 60), consistency and standards (14%, n = 14) and match between the system and real world (9%, n = 9). Majority of the heuristic violations had minor usability issues (73%, n = 71). The only grade 4 heuristic violation reported was due to the visibility of system status in the assessment modules. The results demonstrated that the system was more consistent with Nielsen's usability heuristics. With Nielsen's usability heuristics, it is possible to identify problems in a timely manner, and help facilitate the identification and prioritisation of problems needing urgent attention at an earlier stage before the final deployment of the system.

  18. The Effect of Incentive Structure on Heuristic Decision Making: The Proportion Heuristic

    OpenAIRE

    Robert Oxoby

    2007-01-01

    When making judgments, individuals often utilize heuristics to interpret information. We report on a series of experiments designed to test the ways in which incentive mechanisms influence the use of a particular heuristic in decision-making. Specifically, we demonstrate how information regarding the number of available practice problems influences the behaviors of individuals preparing for an exam (the proportion heuristic). More importantly the extent to which this information influences be...

  19. An Heuristic Framework for Non-Conscious Reasoning

    Directory of Open Access Journals (Sweden)

    Felipe Lara-Rosano

    2017-11-01

    Full Text Available Human non-conscious reasoning is one of the most successful procedures evolved for the purposes of solving everyday problems in an efficient way. This is why the field of artificial intelligence should analyze, formalize and emulate the multiple ways of non-conscious reasoning with the purpose of applying them in human problem solving tasks, like medical diagnostics and treatments, educational diagnostics and intervention, organizational and political decision making, artificial intelligence knowledge based systems and neurocomputers, automatic control systems and similar devices for aiding people in the problem-solving process. In this paper, a heuristic framework for those non-conscious ways of reasoning is presented based on neurocognitive representations, heuristics, and fuzzy sets.

  20. Heuristic attacks against graphical password generators

    CSIR Research Space (South Africa)

    Peach, S

    2010-05-01

    Full Text Available In this paper the authors explore heuristic attacks against graphical password generators. A new trend is emerging to use user clickable pictures to generate passwords. This technique of authentication can be successfully used for - for example...

  1. A Direct Heuristic Algorithm for Linear Programming

    Indian Academy of Sciences (India)

    Abstract. An (3) mathematically non-iterative heuristic procedure that needs no artificial variable is presented for solving linear programming problems. An optimality test is included. Numerical experiments depict the utility/scope of such a procedure.

  2. Establishing usability heuristics for heuristics evaluation in a specific domain: Is there a consensus?

    Science.gov (United States)

    Hermawati, Setia; Lawson, Glyn

    2016-09-01

    Heuristics evaluation is frequently employed to evaluate usability. While general heuristics are suitable to evaluate most user interfaces, there is still a need to establish heuristics for specific domains to ensure that their specific usability issues are identified. This paper presents a comprehensive review of 70 studies related to usability heuristics for specific domains. The aim of this paper is to review the processes that were applied to establish heuristics in specific domains and identify gaps in order to provide recommendations for future research and area of improvements. The most urgent issue found is the deficiency of validation effort following heuristics proposition and the lack of robustness and rigour of validation method adopted. Whether domain specific heuristics perform better or worse than general ones is inconclusive due to lack of validation quality and clarity on how to assess the effectiveness of heuristics for specific domains. The lack of validation quality also affects effort in improving existing heuristics for specific domain as their weaknesses are not addressed. Copyright © 2016 Elsevier Ltd. All rights reserved.

  3. Development of a knowledge-based system for the design of composite automotive components

    Science.gov (United States)

    Moynihan, Gary P.; Stephens, J. Paul

    1997-01-01

    Composite materials are comprised of two or more constituents possessing significantly different physical properties. Due to their high strength and light weight, there is an emerging trend to utilize composites in the automotive industry. There is an inherent link between component design and the manufacturing processes necessary for fabrication. To many designers, this situation may be intimidating, since there is frequently little available understanding of composites and their processes. A direct results is high rates of product scrap and rework. Thus, there is a need to implement a systematic approach to composite material design. One such approach is quality function deployment (QFD). By translating customer requirements into design parameters, through the use of heuristics, QFD supports the improvement of product quality during the planning stages prior to actual production. The purpose of this research is to automate the use of knowledge pertaining to the design and application of composite materials within the automobile industry. This is being accomplished through the development of a prototype expert system incorporating a QFD approach. It will provide industry designers with access to knowledge of composite materials that might not be otherwise available.

  4. Put a limit on it: The protective effects of scarcity heuristics when self-control is low

    NARCIS (Netherlands)

    Cheung, T.T.L.; Kroese, F.M.; Fennis, Bob; de Ridder, D.T.D.

    2015-01-01

    Low self-control is a state in which consumers are assumed to be vulnerable to making impulsive choices that hurt long-term goals. Rather than increasing self-control, the current research exploits the tendency for heuristic-based thinking in low self-control by employing scarcity heuristics to

  5. RELAXATION HEURISTICS FOR THE SET COVERING PROBLEM

    OpenAIRE

    Umetani, Shunji; Yagiura, Mutsunori; 柳浦, 睦憲

    2007-01-01

    The set covering problem (SCP) is one of representative combinatorial optimization problems, which has many practical applications. The continuous development of mathematical programming has derived a number of impressive heuristic algorithms as well as exact branch-and-bound algorithms, which can solve huge SCP instances of bus, railway and airline crew scheduling problems. We survey heuristic algorithms for SCP focusing mainly on contributions of mathematical programming techniques to heuri...

  6. Psychology into economics: fast and frugal heuristics

    OpenAIRE

    Schilirò, Daniele

    2015-01-01

    The present essay focuses on the fast and frugal heuristics program set forth by Gerd Gigerenzer and his fellows. In particular it examines the contribution of Gigerenzer and Goldstein (1996) ‘Reasoning the Fast and Frugal Way: Models of Bounded Rationality’. This essay, following the theoretical propositions and the empirical evidence of Gigerenzer and Goldstein, points out that simple cognitive mechanisms such as fast and frugal heuristics can be capable of successful performance in real wo...

  7. Arational heuristic model of economic decision making

    OpenAIRE

    Grandori, Anna

    2010-01-01

    The article discuss the limits of both the rational actor and the behavioral paradigms in explaining and guiding innovative decision making and outlines a model of economic decision making that in the course of being 'heuristic' (research and discovery oriented) is also 'rational' (in the broad sense of following correct reasoning and scientific methods, non 'biasing'). The model specifies a set of 'rational heuristics' for innovative decision making, for the various sub-processes of problem ...

  8. Component-Based Approach for Educating Students in Bioinformatics

    Science.gov (United States)

    Poe, D.; Venkatraman, N.; Hansen, C.; Singh, G.

    2009-01-01

    There is an increasing need for an effective method of teaching bioinformatics. Increased progress and availability of computer-based tools for educating students have led to the implementation of a computer-based system for teaching bioinformatics as described in this paper. Bioinformatics is a recent, hybrid field of study combining elements of…

  9. Component Data Base for Space Station Resistojet Auxiliary Propulsion

    Science.gov (United States)

    Bader, Clayton H.

    1988-01-01

    The resistojet was baselined for Space Station auxiliary propulsion because of its operational versatility, efficiency, and durability. This report was conceived as a guide to designers and planners of the Space Station auxiliary propulsion system. It is directed to the low thrust resistojet concept, though it should have application to other station concepts or systems such as the Environmental Control and Life Support System (ECLSS), Manufacturing and Technology Laboratory (MTL), and the Waste Fluid Management System (WFMS). The information will likely be quite useful in the same capacity for other non-Space Station systems including satellite, freeflyers, explorers, and maneuvering vehicles. The report is a catalog of the most useful information for the most significant feed system components and is organized for the greatest convenience of the user.

  10. Social heuristics shape intuitive cooperation.

    Science.gov (United States)

    Rand, David G; Peysakhovich, Alexander; Kraft-Todd, Gordon T; Newman, George E; Wurzbacher, Owen; Nowak, Martin A; Greene, Joshua D

    2014-04-22

    Cooperation is central to human societies. Yet relatively little is known about the cognitive underpinnings of cooperative decision making. Does cooperation require deliberate self-restraint? Or is spontaneous prosociality reined in by calculating self-interest? Here we present a theory of why (and for whom) intuition favors cooperation: cooperation is typically advantageous in everyday life, leading to the formation of generalized cooperative intuitions. Deliberation, by contrast, adjusts behaviour towards the optimum for a given situation. Thus, in one-shot anonymous interactions where selfishness is optimal, intuitive responses tend to be more cooperative than deliberative responses. We test this 'social heuristics hypothesis' by aggregating across every cooperation experiment using time pressure that we conducted over a 2-year period (15 studies and 6,910 decisions), as well as performing a novel time pressure experiment. Doing so demonstrates a positive average effect of time pressure on cooperation. We also find substantial variation in this effect, and show that this variation is partly explained by previous experience with one-shot lab experiments.

  11. Usage of Modified Heuristic Model for Determination of Software Stability

    Directory of Open Access Journals (Sweden)

    Sergey Konstantinovich Marfenko

    2013-02-01

    Full Text Available The subject of this paper is analysis method for determining the stability of software against the attacks on its integrity. It is suggested to use the modified heuristic model of software reliability as mathematic basis of this method. This model is based on classic approach, but it takes into account impact levels of different software errors on system integrity. It allows to define critical characteristics of software: percentage of time in stable working, the possibility of failure.

  12. The use of meta-heuristics for airport gate assignment

    DEFF Research Database (Denmark)

    Cheng, Chun-Hung; Ho, Sin C.; Kwan, Cheuk-Lam

    2012-01-01

    proposed to generate good solutions within a reasonable timeframe. In this work, we attempt to assess the performance of three meta-heuristics, namely, genetic algorithm (GA), tabu search (TS), simulated annealing (SA) and a hybrid approach based on SA and TS. Flight data from Incheon International Airport...... are collected to carry out the computational comparison. Although the literature has documented these algorithms, this work may be a first attempt to evaluate their performance using a set of realistic flight data....

  13. Heuristic thinking makes a chemist smart.

    Science.gov (United States)

    Graulich, Nicole; Hopf, Henning; Schreiner, Peter R

    2010-05-01

    We focus on the virtually neglected use of heuristic principles in understanding and teaching of organic chemistry. As human thinking is not comparable to computer systems employing factual knowledge and algorithms--people rarely make decisions through careful considerations of every possible event and its probability, risks or usefulness--research in science and teaching must include psychological aspects of the human decision making processes. Intuitive analogical and associative reasoning and the ability to categorize unexpected findings typically demonstrated by experienced chemists should be made accessible to young learners through heuristic concepts. The psychology of cognition defines heuristics as strategies that guide human problem-solving and deciding procedures, for example with patterns, analogies, or prototypes. Since research in the field of artificial intelligence and current studies in the psychology of cognition have provided evidence for the usefulness of heuristics in discovery, the status of heuristics has grown into something useful and teachable. In this tutorial review, we present a heuristic analysis of a familiar fundamental process in organic chemistry--the cyclic six-electron case, and we show that this approach leads to a more conceptual insight in understanding, as well as in teaching and learning.

  14. Heuristics as Bayesian inference under extreme priors.

    Science.gov (United States)

    Parpart, Paula; Jones, Matt; Love, Bradley C

    2018-05-01

    Simple heuristics are often regarded as tractable decision strategies because they ignore a great deal of information in the input data. One puzzle is why heuristics can outperform full-information models, such as linear regression, which make full use of the available information. These "less-is-more" effects, in which a relatively simpler model outperforms a more complex model, are prevalent throughout cognitive science, and are frequently argued to demonstrate an inherent advantage of simplifying computation or ignoring information. In contrast, we show at the computational level (where algorithmic restrictions are set aside) that it is never optimal to discard information. Through a formal Bayesian analysis, we prove that popular heuristics, such as tallying and take-the-best, are formally equivalent to Bayesian inference under the limit of infinitely strong priors. Varying the strength of the prior yields a continuum of Bayesian models with the heuristics at one end and ordinary regression at the other. Critically, intermediate models perform better across all our simulations, suggesting that down-weighting information with the appropriate prior is preferable to entirely ignoring it. Rather than because of their simplicity, our analyses suggest heuristics perform well because they implement strong priors that approximate the actual structure of the environment. We end by considering how new heuristics could be derived by infinitely strengthening the priors of other Bayesian models. These formal results have implications for work in psychology, machine learning and economics. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.

  15. Environmental risk assessment of biocidal products: identification of relevant components and reliability of a component-based mixture assessment.

    Science.gov (United States)

    Coors, Anja; Vollmar, Pia; Heim, Jennifer; Sacher, Frank; Kehrer, Anja

    2018-01-01

    Biocidal products are mixtures of one or more active substances (a.s.) and a broad range of formulation additives. There is regulatory guidance currently under development that will specify how the combined effects of the a.s. and any relevant formulation additives shall be considered in the environmental risk assessment of biocidal products. The default option is a component-based approach (CBA) by which the toxicity of the product is predicted from the toxicity of 'relevant' components using concentration addition. Hence, unequivocal and practicable criteria are required for identifying the 'relevant' components to ensure protectiveness of the CBA, while avoiding unnecessary workload resulting from including by default components that do not significantly contribute to the product toxicity. The present study evaluated a set of different criteria for identifying 'relevant' components using confidential information on the composition of 21 wood preservative products. Theoretical approaches were complemented by experimentally testing the aquatic toxicity of seven selected products. For three of the seven tested products, the toxicity was underestimated for the most sensitive endpoint (green algae) by more than factor 2 if only the a.s. were considered in the CBA. This illustrated the necessity of including at least some additives along with the a.s. Considering additives that were deemed 'relevant' by the tentatively established criteria reduced the underestimation of toxicity for two of the three products. A lack of data for one specific additive was identified as the most likely reason for the remaining toxicity underestimation of the third product. In three other products, toxicity was overestimated by more than factor 2, while prediction and observation fitted well for the seventh product. Considering all additives in the prediction increased only the degree of overestimation. Supported by theoretical calculations and experimental verifications, the present

  16. Heuristics and stock buying decision: Evidence from Malaysian and Pakistani stock markets

    OpenAIRE

    Habib Hussain Khan; Iram Naz; Fiza Qureshi; Abdul Ghafoor

    2017-01-01

    Applying both qualitative and quantitative approaches, we examine whether or not investors fall prey to three heuristics; namely, anchoring and adjustment, representativeness, and availability, while investing in stocks. We also compare investors' vulnerability to these heuristics based on their economic association, their type and demographic factors such as income, education and experience. For the data collection, a self-constructed questionnaire was administered to investors in the Malays...

  17. PARTIAL TRAINING METHOD FOR HEURISTIC ALGORITHM OF POSSIBLE CLUSTERIZATION UNDER UNKNOWN NUMBER OF CLASSES

    Directory of Open Access Journals (Sweden)

    D. A. Viattchenin

    2009-01-01

    Full Text Available A method for constructing a subset of labeled objects which is used in a heuristic algorithm of possible  clusterization with partial  training is proposed in the  paper.  The  method  is  based  on  data preprocessing by the heuristic algorithm of possible clusterization using a transitive closure of a fuzzy tolerance. Method efficiency is demonstrated by way of an illustrative example.

  18. Critical components for diamond-based quantum coherent devices

    International Nuclear Information System (INIS)

    Greentree, Andrew D; Olivero, Paolo; Draganski, Martin; Trajkov, Elizabeth; Rabeau, James R; Reichart, Patrick; Gibson, Brant C; Rubanov, Sergey; Huntington, Shane T; Jamieson, David N; Prawer, Steven

    2006-01-01

    The necessary elements for practical devices exploiting quantum coherence in diamond materials are summarized, and progress towards their realization documented. A brief review of future prospects for diamond-based devices is also provided

  19. 78 FR 13895 - Certain Wireless Communications Base Stations and Components Thereof; Institution of...

    Science.gov (United States)

    2013-03-01

    ... the sale within the United States after importation of certain wireless communications base stations... United States after importation of certain wireless communications base stations and components thereof... INTERNATIONAL TRADE COMMISSION [Investigation No. 337-TA-871] Certain Wireless Communications Base...

  20. Component fragility data base for reliability and probability studies

    International Nuclear Information System (INIS)

    Bandyopadhyay, K.; Hofmayer, C.; Kassier, M.; Pepper, S.

    1989-01-01

    Safety-related equipment in a nuclear plant plays a vital role in its proper operation and control, and failure of such equipment due to an earthquake may pose a risk to the safe operation of the plant. Therefore, in order to assess the overall reliability of a plant, the reliability of performance of the equipment should be studied first. The success of a reliability or a probability study depends to a great extent on the data base. To meet this demand, Brookhaven National Laboratory (BNL) has formed a test data base relating the seismic capacity of equipment specimens to the earthquake levels. Subsequently, the test data have been analyzed for use in reliability and probability studies. This paper describes the data base and discusses the analysis methods. The final results that can be directly used in plant reliability and probability studies are also presented in this paper

  1. GENERIC, COMPONENT FAILURE DATA BASE FOR LIGHT WATER AND LIQUID SODIUM REACTOR PRAs

    Energy Technology Data Exchange (ETDEWEB)

    S. A. Eide; S. V. Chmielewski; T. D. Swantz

    1990-02-01

    A comprehensive generic component failure data base has been developed for light water and liquid sodium reactor probabilistic risk assessments (PRAs) . The Nuclear Computerized Library for Assessing Reactor Reliability (NUCLARR) and the Centralized Reliability Data Organization (CREDO) data bases were used to generate component failure rates . Using this approach, most of the failure rates are based on actual plant data rather than existing estimates .

  2. EEG-Based Emotion Recognition Using Deep Learning Network with Principal Component Based Covariate Shift Adaptation

    Directory of Open Access Journals (Sweden)

    Suwicha Jirayucharoensak

    2014-01-01

    Full Text Available Automatic emotion recognition is one of the most challenging tasks. To detect emotion from nonstationary EEG signals, a sophisticated learning algorithm that can represent high-level abstraction is required. This study proposes the utilization of a deep learning network (DLN to discover unknown feature correlation between input signals that is crucial for the learning task. The DLN is implemented with a stacked autoencoder (SAE using hierarchical feature learning approach. Input features of the network are power spectral densities of 32-channel EEG signals from 32 subjects. To alleviate overfitting problem, principal component analysis (PCA is applied to extract the most important components of initial input features. Furthermore, covariate shift adaptation of the principal components is implemented to minimize the nonstationary effect of EEG signals. Experimental results show that the DLN is capable of classifying three different levels of valence and arousal with accuracy of 49.52% and 46.03%, respectively. Principal component based covariate shift adaptation enhances the respective classification accuracy by 5.55% and 6.53%. Moreover, DLN provides better performance compared to SVM and naive Bayes classifiers.

  3. A practical implementation science heuristic for organizational readiness: R = MC2

    Science.gov (United States)

    Cook, Brittany S.; Lamont, Andrea; Wandersman, Abraham; Castellow, Jennifer; Katz, Jason; Beidas, Rinad S.

    2015-01-01

    There are many challenges when an innovation (i.e., a program, process, or policy that is new to an organization) is actively introduced into an organization. One critical component for successful implementation is the organization’s readiness for the innovation. In this article, we propose a practical implementation science heuristic, abbreviated as R= MC2. We propose that organizational readiness involves: 1) the motivation to implement an innovation, 2) the general capacities of an organization, and 3) the innovation-specific capacities needed for a particular innovation. Each of these components can be assessed independently and be used formatively. The heuristic can be used by organizations to assess readiness to implement and by training and technical assistance providers to help build organizational readiness. We present an illustration of the heuristic by showing how behavioral health organizations differ in readiness to implement a peer specialist initiative. Implications for research and practice of organizational readiness are discussed. PMID:26668443

  4. A novel heuristic algorithm for capacitated vehicle routing problem

    Science.gov (United States)

    Kır, Sena; Yazgan, Harun Reşit; Tüncel, Emre

    2017-09-01

    The vehicle routing problem with the capacity constraints was considered in this paper. It is quite difficult to achieve an optimal solution with traditional optimization methods by reason of the high computational complexity for large-scale problems. Consequently, new heuristic or metaheuristic approaches have been developed to solve this problem. In this paper, we constructed a new heuristic algorithm based on the tabu search and adaptive large neighborhood search (ALNS) with several specifically designed operators and features to solve the capacitated vehicle routing problem (CVRP). The effectiveness of the proposed algorithm was illustrated on the benchmark problems. The algorithm provides a better performance on large-scaled instances and gained advantage in terms of CPU time. In addition, we solved a real-life CVRP using the proposed algorithm and found the encouraging results by comparison with the current situation that the company is in.

  5. A Heuristics Approach for Classroom Scheduling Using Genetic Algorithm Technique

    Science.gov (United States)

    Ahmad, Izah R.; Sufahani, Suliadi; Ali, Maselan; Razali, Siti N. A. M.

    2018-04-01

    Reshuffling and arranging classroom based on the capacity of the audience, complete facilities, lecturing time and many more may lead to a complexity of classroom scheduling. While trying to enhance the productivity in classroom planning, this paper proposes a heuristic approach for timetabling optimization. A new algorithm was produced to take care of the timetabling problem in a university. The proposed of heuristics approach will prompt a superior utilization of the accessible classroom space for a given time table of courses at the university. Genetic Algorithm through Java programming languages were used in this study and aims at reducing the conflicts and optimizes the fitness. The algorithm considered the quantity of students in each class, class time, class size, time accessibility in each class and lecturer who in charge of the classes.

  6. Aristotle's 'completeness test' as Heuristics for an Account of Dynamicity

    DEFF Research Database (Denmark)

    Seibt, Johanna

    2015-01-01

    If being were ‘dynamic,’ would it be more amenable to a definition? In this paper I present a number of preliminary considerations for an exploration of this question. Working from the methodological stance of analytical ontology, I assume that the first task for an ontology of dynamic being...... heuristic leads to a class of inferential data (aspectual inferences) that analytical ontologists have all but overlooked so far. In addition, I suggest that the passage also can offer some ideas about how one might formulate, in mereological terms, a component of an implicit definition of dynamicity....

  7. A health literacy and usability heuristic evaluation of a mobile consumer health application.

    Science.gov (United States)

    Monkman, Helen; Kushniruk, Andre

    2013-01-01

    Usability and health literacy are two critical factors in the design and evaluation of consumer health information systems. However, methods for evaluating these two factors in conjunction remain limited. This study adapted a set of existing guidelines for the design of consumer health Web sites into evidence-based evaluation heuristics tailored specifically for mobile consumer health applications. In order to test the approach, a mobile consumer health application (app) was then evaluated using these heuristics. In addition to revealing ways to improve the usability of the system, this analysis identified opportunities to augment the content to make it more understandable by users with limited health literacy. This study successfully demonstrated the utility of converting existing design guidelines into heuristics for the evaluation of usability and health literacy. The heuristics generated could be applied for assessing and revising other existing consumer health information systems.

  8. Unified heuristics to solve routing problem of reverse logistics in sustainable supply chain

    Science.gov (United States)

    Anbuudayasankar, S. P.; Ganesh, K.; Lenny Koh, S. C.; Mohandas, K.

    2010-03-01

    A reverse logistics problem, motivated by many real-life applications, is examined where bottles/cans in which products are delivered from a processing depot to customers in one period are available for return to the depot in the following period. The picked-up bottles/cans need to be adjusted in the place of delivery load. This problem is termed as simultaneous delivery and pick-up problem with constrained capacity (SDPC). We develop three unified heuristics based on extended branch and bound heuristic, genetic algorithm and simulated annealing to solve SDPC. These heuristics are also designed to solve standard travelling salesman problem (TSP) and TSP with simultaneous delivery and pick-up (TSDP). We tested the heuristics on standard, derived and randomly generated datasets of TSP, TSDP and SDPC and obtained satisfying results with high convergence in reasonable time.

  9. An Efficient Heuristic Approach for Irregular Cutting Stock Problem in Ship Building Industry

    Directory of Open Access Journals (Sweden)

    Yan-xin Xu

    2016-01-01

    Full Text Available This paper presents an efficient approach for solving a real two-dimensional irregular cutting stock problem in ship building industry. Cutting stock problem is a common cutting and packing problem that arises in a variety of industrial applications. A modification of selection heuristic Exact Fit is applied in our research. In the case referring to irregular shapes, a placement heuristics is more important to construct a complete solution. A placement heuristic relating to bottom-left-fill is presented. We evaluate the proposed approach using generated instance only with convex shapes in literatures and some instances with nonconvex shapes based on real problem from ship building industry. The results demonstrate that the effectiveness and efficiency of the proposed approach are significantly better than some conventional heuristics.

  10. Performance comparison of heuristic algorithms for task scheduling in IaaS cloud computing environment

    Science.gov (United States)

    Madni, Syed Hamid Hussain; Abd Latiff, Muhammad Shafie; Abdullahi, Mohammed; Usman, Mohammed Joda

    2017-01-01

    Cloud computing infrastructure is suitable for meeting computational needs of large task sizes. Optimal scheduling of tasks in cloud computing environment has been proved to be an NP-complete problem, hence the need for the application of heuristic methods. Several heuristic algorithms have been developed and used in addressing this problem, but choosing the appropriate algorithm for solving task assignment problem of a particular nature is difficult since the methods are developed under different assumptions. Therefore, six rule based heuristic algorithms are implemented and used to schedule autonomous tasks in homogeneous and heterogeneous environments with the aim of comparing their performance in terms of cost, degree of imbalance, makespan and throughput. First Come First Serve (FCFS), Minimum Completion Time (MCT), Minimum Execution Time (MET), Max-min, Min-min and Sufferage are the heuristic algorithms considered for the performance comparison and analysis of task scheduling in cloud computing. PMID:28467505

  11. Response demands and the recruitment of heuristic strategies in syllogistic reasoning.

    Science.gov (United States)

    Reverberi, Carlo; Rusconi, Patrice; Paulesu, Eraldo; Cherubini, Paolo

    2009-03-01

    Two experiments investigated whether dealing with a homogeneous subset of syllogisms with time-constrained responses encouraged participants to develop and use heuristics for abstract (Experiment 1) and thematic (Experiment 2) syllogisms. An atmosphere-based heuristic accounted for most responses with both abstract and thematic syllogisms. With thematic syllogisms, a weaker effect of a belief heuristic was also observed, mainly where the correct response was inconsistent with the atmosphere of the premises. Analytic processes appear to have played little role in the time-constrained condition, whereas their involvement increased in a self-paced, unconstrained condition. From a dual-process perspective, the results further specify how task demands affect the recruitment of heuristic and analytic systems of reasoning. Because the syllogisms and experimental procedure were the same as those used in a previous neuroimaging study by Goel, Buchel, Frith, and Dolan (2000), the result also deepen our understanding of the cognitive processes investigated by that study.

  12. Performance comparison of heuristic algorithms for task scheduling in IaaS cloud computing environment.

    Science.gov (United States)

    Madni, Syed Hamid Hussain; Abd Latiff, Muhammad Shafie; Abdullahi, Mohammed; Abdulhamid, Shafi'i Muhammad; Usman, Mohammed Joda

    2017-01-01

    Cloud computing infrastructure is suitable for meeting computational needs of large task sizes. Optimal scheduling of tasks in cloud computing environment has been proved to be an NP-complete problem, hence the need for the application of heuristic methods. Several heuristic algorithms have been developed and used in addressing this problem, but choosing the appropriate algorithm for solving task assignment problem of a particular nature is difficult since the methods are developed under different assumptions. Therefore, six rule based heuristic algorithms are implemented and used to schedule autonomous tasks in homogeneous and heterogeneous environments with the aim of comparing their performance in terms of cost, degree of imbalance, makespan and throughput. First Come First Serve (FCFS), Minimum Completion Time (MCT), Minimum Execution Time (MET), Max-min, Min-min and Sufferage are the heuristic algorithms considered for the performance comparison and analysis of task scheduling in cloud computing.

  13. Modeling media as latent semantics based on cognitive components

    DEFF Research Database (Denmark)

    Petersen, Michael Kai

    as distinct states in the continuous ebb and flow of emotions underlying consciousness. Whether it being a soundscape of structured peaks or tiny black characters lined up across a page, we rely on syntax for parsing sequences of symbols, which based on hierarchically nested structures allow us to express...

  14. Selection of independent components based on cortical mapping of electromagnetic activity

    Science.gov (United States)

    Chan, Hui-Ling; Chen, Yong-Sheng; Chen, Li-Fen

    2012-10-01

    Independent component analysis (ICA) has been widely used to attenuate interference caused by noise components from the electromagnetic recordings of brain activity. However, the scalp topographies and associated temporal waveforms provided by ICA may be insufficient to distinguish functional components from artifactual ones. In this work, we proposed two component selection methods, both of which first estimate the cortical distribution of the brain activity for each component, and then determine the functional components based on the parcellation of brain activity mapped onto the cortical surface. Among all independent components, the first method can identify the dominant components, which have strong activity in the selected dominant brain regions, whereas the second method can identify those inter-regional associating components, which have similar component spectra between a pair of regions. For a targeted region, its component spectrum enumerates the amplitudes of its parceled brain activity across all components. The selected functional components can be remixed to reconstruct the focused electromagnetic signals for further analysis, such as source estimation. Moreover, the inter-regional associating components can be used to estimate the functional brain network. The accuracy of the cortical activation estimation was evaluated on the data from simulation studies, whereas the usefulness and feasibility of the component selection methods were demonstrated on the magnetoencephalography data recorded from a gender discrimination study.

  15. Knowledge Based Components of Expertise in Medical Diagnosis.

    Science.gov (United States)

    1981-09-01

    the ability to ac- cess and use knowledge that one "has" is situationally dependent (e.g., Melton, 1963)1 Tulving and Pearlstone , 1966). For example...Londons Wiley, 1976. Tulving , E., & Pearlstone , E. Availability versus accessibility of information in memory for words. Journal of Verbal Learning...encounter (c.f.Flexser’and Tulving , 19781 Tulving ., 1976). Expert-based instructional devices (computer assisted instruction or decision support sys

  16. Radiation induced defect flux behaviors at zirconium based component

    International Nuclear Information System (INIS)

    Choi, Sang Il; Kim, Ji Hyun; Kwon, Jun Hyun; Lee, Gyeong Geun

    2013-01-01

    In commercial reactor core, structure materials are located in high temperature and high pressure environment. Therefore, main concern of structure materials is corrosion and mechanical properties change than radiation effects on materials. However, radiation effects on materials become more important phenomena because research reactor condition is different from commercial reactor. The temperature is lower than 100 .deg. C and radiation dose is much higher than that of commercial reactor. Among the radiation effect on zirconium based metal, radiation induced growth (RIG), known as volume conservative distortion, is one of the most important phenomena. Recently, theoretical RIG modeling based on radiation damage theory (RDT) and balance equation are developed. However, these growth modeling have limited framework of single crystal and high temperature. To model theoretical RIG in research reactor, qualitative mechanism must be set up. Therefore, this paper intent is establishing defect flux mechanism of zirconium base metal in research reactor for RIG modeling. After than theoretical RIG work will be expanded to research reactor condition

  17. Analytic and heuristic processes in the detection and resolution of conflict.

    Science.gov (United States)

    Ferreira, Mário B; Mata, André; Donkin, Christopher; Sherman, Steven J; Ihmels, Max

    2016-10-01

    Previous research with the ratio-bias task found larger response latencies for conflict trials where the heuristic- and analytic-based responses are assumed to be in opposition (e.g., choosing between 1/10 and 9/100 ratios of success) when compared to no-conflict trials where both processes converge on the same response (e.g., choosing between 1/10 and 11/100). This pattern is consistent with parallel dual-process models, which assume that there is effective, rather than lax, monitoring of the output of heuristic processing. It is, however, unclear why conflict resolution sometimes fails. Ratio-biased choices may increase because of a decline in analytical reasoning (leaving heuristic-based responses unopposed) or to a rise in heuristic processing (making it more difficult for analytic processes to override the heuristic preferences). Using the process-dissociation procedure, we found that instructions to respond logically and response speed affected analytic (controlled) processing (C), leaving heuristic processing (H) unchanged, whereas the intuitive preference for large nominators (as assessed by responses to equal ratio trials) affected H but not C. These findings create new challenges to the debate between dual-process and single-process accounts, which are discussed.

  18. Low-power adaptive filter based on RNS components

    DEFF Research Database (Denmark)

    Bernocchi, Gian Luca; Cardarilli, Gian Carlo; Del Re, Andrea

    2007-01-01

    In this paper a low-power implementation of an adaptive FIR filter is presented. The filter is designed to meet the constraints of channel equalization for fixed wireless communications that typically requires a large number of taps, but a serial updating of the filter coefficients, based...... on the least mean squares (LMS) algorithm, is allowed. Previous work showed that the use of the residue number system (RNS) for the variable FIR filter grants advantages both in area and power consumption. On the other hand, the use of a binary serial implementation of the adaptation algorithm eliminates...... the need for complex scaling circuits in RNS. The advantages in terms of area and speed of the presented filter, with respect to its two's complement counterpart, are evaluated for implementations in standard cells....

  19. High performance coated board inspection system based on commercial components

    CERN Document Server

    Barjaktarovic, M; Radunovic, J

    2007-01-01

    This paper presents a vision system for defect (fault) detection on a coated board developed using three industrial firewire cameras and a PC. Application for image processing and system control was realized with the LabView software package. Software for defect detection is based on a variation of the image segmentation algorithm. Standard steps in image segmentation are modified to match the characteristics of defects. Software optimization was accomplished using SIMD (Single Instruction Multiple Data) technology available in the Intel Pentium 4 processors that provided real time inspection capability. System provides benefits such as: improvement in production process, higher quality of delivered coated board and reduction of waste. This was proven during successful exploitation of the system for more than a year.

  20. Multiobjective hyper heuristic scheme for system design and optimization

    Science.gov (United States)

    Rafique, Amer Farhan

    2012-11-01

    As system design is becoming more and more multifaceted, integrated, and complex, the traditional single objective optimization trends of optimal design are becoming less and less efficient and effective. Single objective optimization methods present a unique optimal solution whereas multiobjective methods present pareto front. The foremost intent is to predict a reasonable distributed pareto-optimal solution set independent of the problem instance through multiobjective scheme. Other objective of application of intended approach is to improve the worthiness of outputs of the complex engineering system design process at the conceptual design phase. The process is automated in order to provide the system designer with the leverage of the possibility of studying and analyzing a large multiple of possible solutions in a short time. This article presents Multiobjective Hyper Heuristic Optimization Scheme based on low level meta-heuristics developed for the application in engineering system design. Herein, we present a stochastic function to manage meta-heuristics (low-level) to augment surety of global optimum solution. Generic Algorithm, Simulated Annealing and Swarm Intelligence are used as low-level meta-heuristics in this study. Performance of the proposed scheme is investigated through a comprehensive empirical analysis yielding acceptable results. One of the primary motives for performing multiobjective optimization is that the current engineering systems require simultaneous optimization of conflicting and multiple. Random decision making makes the implementation of this scheme attractive and easy. Injecting feasible solutions significantly alters the search direction and also adds diversity of population resulting in accomplishment of pre-defined goals set in the proposed scheme.

  1. Recipient design in human communication: simple heuristics or perspective taking?

    Science.gov (United States)

    Blokpoel, Mark; van Kesteren, Marlieke; Stolk, Arjen; Haselager, Pim; Toni, Ivan; van Rooij, Iris

    2012-01-01

    Humans have a remarkable capacity for tuning their communicative behaviors to different addressees, a phenomenon also known as recipient design. It remains unclear how this tuning of communicative behavior is implemented during live human interactions. Classical theories of communication postulate that recipient design involves perspective taking, i.e., the communicator selects her behavior based on her hypotheses about beliefs and knowledge of the recipient. More recently, researchers have argued that perspective taking is computationally too costly to be a plausible mechanism in everyday human communication. These researchers propose that computationally simple mechanisms, or heuristics, are exploited to perform recipient design. Such heuristics may be able to adapt communicative behavior to an addressee with no consideration for the addressee's beliefs and knowledge. To test whether the simpler of the two mechanisms is sufficient for explaining the "how" of recipient design we studied communicators' behaviors in the context of a non-verbal communicative task (the Tacit Communication Game, TCG). We found that the specificity of the observed trial-by-trial adjustments made by communicators is parsimoniously explained by perspective taking, but not by simple heuristics. This finding is important as it suggests that humans do have a computationally efficient way of taking beliefs and knowledge of a recipient into account.

  2. Recipient design in human communication: Simple heuristics or perspective taking?

    Directory of Open Access Journals (Sweden)

    Mark eBlokpoel

    2012-09-01

    Full Text Available Humans have a remarkable capacity for tuning their communicative behaviors to different addressees, a phenomenon also known as recipient design. It remains unclear how this tuning of communicative behavior is implemented during live human interactions. Classical theories of communication postulate that recipient design involves perspective taking, i.e., the communicator selects her behavior based on her hypotheses about beliefs and knowledge of the recipient. More recently, researchers have argued that perspective taking is computationally too costly to be a plausible mechanism in everyday human communication. These researchers propose that computationally simple mechanisms, or heuristics, are exploited to perform recipient design. Such heuristics may be able to adapt communicative behavior to an addressee with no consideration for the addressee's beliefs and knowledge. To test whether the simpler of the two mechanisms is sufficient for explaining the `how' of recipient design we studied communicators' behaviors in the context of a non-verbal communicative task (the Tacit Communication Game, TCG. We found that the specificity of the observed trial-by-trial adjustments made by communicators is parsimoniously explained by perspective taking, but not by simple heuristics. This finding is important as it suggests that humans do have a computationally efficient way of taking beliefs and knowledge of a recipient into account.

  3. ITPI: Initial Transcription Process-Based Identification Method of Bioactive Components in Traditional Chinese Medicine Formula

    Directory of Open Access Journals (Sweden)

    Baixia Zhang

    2016-01-01

    Full Text Available Identification of bioactive components is an important area of research in traditional Chinese medicine (TCM formula. The reported identification methods only consider the interaction between the components and the target proteins, which is not sufficient to explain the influence of TCM on the gene expression. Here, we propose the Initial Transcription Process-based Identification (ITPI method for the discovery of bioactive components that influence transcription factors (TFs. In this method, genome-wide chip detection technology was used to identify differentially expressed genes (DEGs. The TFs of DEGs were derived from GeneCards. The components influencing the TFs were derived from STITCH. The bioactive components in the formula were identified by evaluating the molecular similarity between the components in formula and the components that influence the TF of DEGs. Using the formula of Tian-Zhu-San (TZS as an example, the reliability and limitation of ITPI were examined and 16 bioactive components that influence TFs were identified.

  4. Issues and approaches in risk-based aging analyses of passive components

    International Nuclear Information System (INIS)

    Uryasev, S.P.; Samanta, P.K.; Vesely, W.E.

    1994-01-01

    In previous NRC-sponsored work a general methodology was developed to quantify the risk contributions from aging components at nuclear plants. The methodology allowed Probabilistic Risk Analyses (PRAs) to be modified to incorporate the age-dependent component failure rates and also aging maintenance models to evaluate and prioritize the aging contributions from active components using the linear aging failure rate model and empirical components aging rates. In the present paper, this methodology is extended to passive components (for example, the pipes, heat exchangers, and the vessel). The analyses of passive components bring in issues different from active components. Here, we specifically focus on three aspects that need to be addressed in risk-based aging prioritization of passive components

  5. Calculations of atomic magnetic nuclear shielding constants based on the two-component normalized elimination of the small component method

    Science.gov (United States)

    Yoshizawa, Terutaka; Zou, Wenli; Cremer, Dieter

    2017-04-01

    A new method for calculating nuclear magnetic resonance shielding constants of relativistic atoms based on the two-component (2c), spin-orbit coupling including Dirac-exact NESC (Normalized Elimination of the Small Component) approach is developed where each term of the diamagnetic and paramagnetic contribution to the isotropic shielding constant σi s o is expressed in terms of analytical energy derivatives with regard to the magnetic field B and the nuclear magnetic moment 𝝁 . The picture change caused by renormalization of the wave function is correctly described. 2c-NESC/HF (Hartree-Fock) results for the σiso values of 13 atoms with a closed shell ground state reveal a deviation from 4c-DHF (Dirac-HF) values by 0.01%-0.76%. Since the 2-electron part is effectively calculated using a modified screened nuclear shielding approach, the calculation is efficient and based on a series of matrix manipulations scaling with (2M)3 (M: number of basis functions).

  6. Research on development model of nuclear component based on life cycle management

    International Nuclear Information System (INIS)

    Bao Shiyi; Zhou Yu; He Shuyan

    2005-01-01

    At present the development process of nuclear component, even nuclear component itself, is more and more supported by computer technology. This increasing utilization of the computer and software has led to the faster development of nuclear technology on one hand and also brought new problems on the other hand. Especially, the combination of hardware, software and humans has increased nuclear component system complexities to an unprecedented level. To solve this problem, Life Cycle Management technology is adopted in nuclear component system. Hence, an intensive discussion on the development process of a nuclear component is proposed. According to the characteristics of the nuclear component development, such as the complexities and strict safety requirements of the nuclear components, long-term design period, changeable design specifications and requirements, high capital investment, and satisfaction for engineering codes/standards, the development life-cycle model of nuclear component is presented. The development life-cycle model is classified at three levels, namely, component level development life-cycle, sub-component development life-cycle and component level verification/certification life-cycle. The purposes and outcomes of development processes are stated in detailed. A process framework for nuclear component based on system engineering and development environment of nuclear component is discussed for future research work. (authors)

  7. Novel Heuristics for Cell Radius Determination in WCDMA Systems and Their Application to Strategic Planning Studies

    Directory of Open Access Journals (Sweden)

    G. Esteve-Asensio

    2009-01-01

    Full Text Available We propose and compare three novel heuristics for the calculation of the optimal cell radius in mobile networks based on Wideband Code Division Multiple Access (WCDMA technology. The proposed heuristics solve the problem of the load assignment and cellular radius calculation. We have tested our approaches with experiments in multiservices scenarios showing that the proposed heuristics maximize the cell radius, providing the optimum load factor assignment. The main application of these algorithms is strategic planning studies, where an estimation of the number of Nodes B of the mobile operator, at a national level, is required for economic analysis. In this case due to the large number of different scenarios considered (cities, towns, and open areas other methods than simulation need to be considered. As far as we know, there is no other similar method in the literature and therefore these heuristics may represent a novelty in strategic network planning studies. The proposed heuristics are implemented in a strategic planning software tool and an example of their application for a case in Spain is presented. The proposed heuristics are used for telecommunications regulatory studies in several countries.

  8. Analysis of appraisal tool of system security engineering capability maturity based on component

    International Nuclear Information System (INIS)

    Liu Zhenghai; Yang Xiaohua; Zou Shuliang; Liu Yachun; Xiao Jiantian; Liu Zhiming

    2012-01-01

    Spent Fuel Reprocessing is a part of nuclear fuel cycle and is the inevitably choice of nuclear power sustainable development. Reprocessing needs to face with radiological, criticality, chemical hazards. Besides using the tradition appraisal methods based on the security goals, it is a beneficial supplement that using the appraisal method of system security engineering capability maturity model based on the process. Experts should check and approve large numbers of documents during the appraisal based on system security engineering capability maturity model, so it is necessary that developing a tool to assist the expert to complete the appraisal. The method of developing software based on component is highly effective, nimble and reliable. Component technology is analyzed, the methods of extraction model domain components and general components is introduced, and the appraisal system is developed based on component technology. (authors)

  9. Reliability-based sensitivity of mechanical components with arbitrary distribution parameters

    International Nuclear Information System (INIS)

    Zhang, Yi Min; Yang, Zhou; Wen, Bang Chun; He, Xiang Dong; Liu, Qiaoling

    2010-01-01

    This paper presents a reliability-based sensitivity method for mechanical components with arbitrary distribution parameters. Techniques from the perturbation method, the Edgeworth series, the reliability-based design theory, and the sensitivity analysis approach were employed directly to calculate the reliability-based sensitivity of mechanical components on the condition that the first four moments of the original random variables are known. The reliability-based sensitivity information of the mechanical components can be accurately and quickly obtained using a practical computer program. The effects of the design parameters on the reliability of mechanical components were studied. The method presented in this paper provides the theoretic basis for the reliability-based design of mechanical components

  10. Rapid Deployment of Optimal Control for Building HVAC Systems using Innovative Software Tools and a Hybrid Heuristic/Model Based Control Approach

    Science.gov (United States)

    2017-03-21

    Monitoring and Verification Protocol LCDR Lieutenant Commander ME Mechanical Engineering MPC Model Predictive Control: a model-based system...Activity PE Professional Engineer PG&E Pacific Gas & Electric : a California utility company PI Proportional + Integral: a common software element...INTRODUCTION The Department of Defense (DoD) spends approximately $4 billion per year on facility energy consumption to power and fuel over 500

  11. Can we trust module-respect heuristics?

    International Nuclear Information System (INIS)

    Mo, Yuchang

    2013-01-01

    BDD (Binary Decision Diagrams) have proven to be a very efficient tool to assess Fault Trees. However, the size of BDD, and therefore the efficiency of the whole methodology, depends dramatically on the choice of variable ordering. The determination of the best variable ordering is intractable. Therefore, heuristics have been designed to select reasonably good variable orderings. One very important common feature for good static heuristics is to respect modules. In this paper, the notion of module-respect is studied in a systematic way. It is proved that under certain condition there always exists an optimal ordering that respects modules. This condition is that for each module there is always a smallest module BDD and each included module variable appears only once. On the other hand, it is shown that for the trees not satisfying the above sufficient condition the optimal orderings may not be able to be directly generated using module-respect heuristics, even when the shuffling strategy is used.

  12. Judgment under Uncertainty: Heuristics and Biases.

    Science.gov (United States)

    Tversky, A; Kahneman, D

    1974-09-27

    This article described three heuristics that are employed in making judgements under uncertainty: (i) representativeness, which is usually employed when people are asked to judge the probability that an object or event A belongs to class or process B; (ii) availability of instances or scenarios, which is often employed when people are asked to assess the frequency of a class or the plausibility of a particular development; and (iii) adjustment from an anchor, which is usually employed in numerical prediction when a relevant value is available. These heuristics are highly economical and usually effective, but they lead to systematic and predictable errors. A better understanding of these heuristics and of the biases to which they lead could improve judgements and decisions in situations of uncertainty.

  13. Two efficient label-equivalence-based connected-component labeling algorithms for 3-D binary images.

    Science.gov (United States)

    He, Lifeng; Chao, Yuyan; Suzuki, Kenji

    2011-08-01

    Whenever one wants to distinguish, recognize, and/or measure objects (connected components) in binary images, labeling is required. This paper presents two efficient label-equivalence-based connected-component labeling algorithms for 3-D binary images. One is voxel based and the other is run based. For the voxel-based one, we present an efficient method of deciding the order for checking voxels in the mask. For the run-based one, instead of assigning each foreground voxel, we assign each run a provisional label. Moreover, we use run data to label foreground voxels without scanning any background voxel in the second scan. Experimental results have demonstrated that our voxel-based algorithm is efficient for 3-D binary images with complicated connected components, that our run-based one is efficient for those with simple connected components, and that both are much more efficient than conventional 3-D labeling algorithms.

  14. Intelligent System Design Using Hyper-Heuristics

    Directory of Open Access Journals (Sweden)

    Nelishia Pillay

    2015-07-01

    Full Text Available Determining the most appropriate search method or artificial intelligence technique to solve a problem is not always evident and usually requires implementation of the different approaches to ascertain this. In some instances a single approach may not be sufficient and hybridization of methods may be needed to find a solution. This process can be time consuming. The paper proposes the use of hyper-heuristics as a means of identifying which method or combination of approaches is needed to solve a problem. The research presented forms part of a larger initiative aimed at using hyper-heuristics to develop intelligent hybrid systems. As an initial step in this direction, this paper investigates this for classical artificial intelligence uninformed and informed search methods, namely depth first search, breadth first search, best first search, hill-climbing and the A* algorithm. The hyper-heuristic determines the search or combination of searches to use to solve the problem. An evolutionary algorithm hyper-heuristic is implemented for this purpose and its performance is evaluated in solving the 8-Puzzle, Towers of Hanoi and Blocks World problems. The hyper-heuristic employs a generational evolutionary algorithm which iteratively refines an initial population using tournament selection to select parents, which the mutation and crossover operators are applied to for regeneration. The hyper-heuristic was able to identify a search or combination of searches to produce solutions for the twenty 8-Puzzle, five Towers of Hanoi and five Blocks World problems. Furthermore, admissible solutions were produced for all problem instances.

  15. Design, construction, and technical implementation of a web-based interdisciplinary symptom evaluation (WISE) - a heuristic proposal for orofacial pain and temporomandibular disorders.

    Science.gov (United States)

    Ettlin, Dominik A; Sommer, Isabelle; Brönnimann, Ben; Maffioletti, Sergio; Scheidt, Jörg; Hou, Mei-Yin; Lukic, Nenad; Steiger, Beat

    2016-12-01

    Medical symptoms independent of body location burden individuals to varying degrees and may require care by more than one expert. Various paper and computer-based tools exist that aim to comprehensively capture data for optimal clinical management and research. A web-based interdisciplinary symptom evaluation (WISE) was newly designed, constructed, and technically implemented. For worldwide applicability and to avoid copyright infringements, open source software tools and free validated questionnaires available in multiple languages were used. Highly secure data storage limits access strictly to those who use the tool for collecting, storing, and evaluating their data. Concept and implementation is illustrated by a WISE sample tailored for the requirements of a single center in Switzerland providing interdisciplinary care to orofacial pain and temporomandibular disorder patients. By combining a symptom- burden checklist with in-depth questionnaires serving as case-finding instruments, an algorithm was developed that assists in clarifying case complexity and need for targeted expert evaluation. This novel modular approach provides a personalized, response-tailored instrument for the time- and cost-effective collection of symptom-burden focused quantitative data. The tool includes body drawing options and instructional videos. It is applicable for biopsychosocial evaluation in a variety of clinical settings and offers direct feedback by a case report summary. In clinical practice, the new instrument assists in clarifying case complexity and referral need, based on symptom burden and response -tailored case finding. It provides single-case summary reports from a biopsychosocial perspective and includes graphical symptom maps. Secure, centrally stored data collection of anonymous data is possible. The tool enables personalized medicine, facilitates interprofessional education and collaboration, and allows for multicenter patient-reported outcomes research.

  16. Feedback loops and temporal misalignment in component-based hydrologic modeling

    Science.gov (United States)

    Elag, Mostafa M.; Goodall, Jonathan L.; Castronova, Anthony M.

    2011-12-01

    In component-based modeling, a complex system is represented as a series of loosely integrated components with defined interfaces and data exchanges that allow the components to be coupled together through shared boundary conditions. Although the component-based paradigm is commonly used in software engineering, it has only recently been applied for modeling hydrologic and earth systems. As a result, research is needed to test and verify the applicability of the approach for modeling hydrologic systems. The objective of this work was therefore to investigate two aspects of using component-based software architecture for hydrologic modeling: (1) simulation of feedback loops between components that share a boundary condition and (2) data transfers between temporally misaligned model components. We investigated these topics using a simple case study where diffusion of mass is modeled across a water-sediment interface. We simulated the multimedia system using two model components, one for the water and one for the sediment, coupled using the Open Modeling Interface (OpenMI) standard. The results were compared with a more conventional numerical approach for solving the system where the domain is represented by a single multidimensional array. Results showed that the component-based approach was able to produce the same results obtained with the more conventional numerical approach. When the two components were temporally misaligned, we explored the use of different interpolation schemes to minimize mass balance error within the coupled system. The outcome of this work provides evidence that component-based modeling can be used to simulate complicated feedback loops between systems and guidance as to how different interpolation schemes minimize mass balance error introduced when components are temporally misaligned.

  17. Heuristics for container loading of furniture

    DEFF Research Database (Denmark)

    Egeblad, Jens; Garavelli, Claudio; Lisi, Stefano

    2010-01-01

    . In the studied company, the problem arises hundreds of times daily during transport planning. Instances may contain more than one hundred different items with irregular shapes. To solve this complex problem we apply a set of heuristics successively that each solve one part of the problem. Large items...... are combined in specific structures to ensure proper protection of the items during transportation and to simplify the problem. The solutions generated by the heuristic has an average loading utilization of 91.3% for the most general instances with average running times around 100 seconds....

  18. Heuristic extension of the Schwarzschild metric

    International Nuclear Information System (INIS)

    Espinosa, J.M.

    1982-01-01

    The Schwarzschild solution of Einstein's equations of gravitation has several singularities. It is known that the singularity at r = 2Gm/c 2 is only apparent, a result of the coordinates in which the solution was found. Paradoxical results occuring near the singularity show the system of coordinates is incomplete. We introduce a simple, two-dimensional metric with an apparent singularity that makes it incomplete. By a straightforward, heuristic procedure we extend and complete this simple metric. We then use the same procedure to give a heuristic derivation of the Kruskal system of coordinates, which is known to extend the Schwarzschild manifold past its apparent singularity and produce a complete manifold

  19. Age Effects and Heuristics in Decision Making.

    Science.gov (United States)

    Besedeš, Tibor; Deck, Cary; Sarangi, Sudipta; Shor, Mikhael

    2012-05-01

    Using controlled experiments, we examine how individuals make choices when faced with multiple options. Choice tasks are designed to mimic the selection of health insurance, prescription drug, or retirement savings plans. In our experiment, available options can be objectively ranked allowing us to examine optimal decision making. First, the probability of a person selecting the optimal option declines as the number of options increases, with the decline being more pronounced for older subjects. Second, heuristics differ by age with older subjects relying more on suboptimal decision rules. In a heuristics validation experiment, older subjects make worse decisions than younger subjects.

  20. Leveraging Existing Mission Tools in a Re-Usable, Component-Based Software Environment

    Science.gov (United States)

    Greene, Kevin; Grenander, Sven; Kurien, James; z,s (fshir. z[orttr); z,scer; O'Reilly, Taifun

    2006-01-01

    Emerging methods in component-based software development offer significant advantages but may seem incompatible with existing mission operations applications. In this paper we relate our positive experiences integrating existing mission applications into component-based tools we are delivering to three missions. In most operations environments, a number of software applications have been integrated together to form the mission operations software. In contrast, with component-based software development chunks of related functionality and data structures, referred to as components, can be individually delivered, integrated and re-used. With the advent of powerful tools for managing component-based development, complex software systems can potentially see significant benefits in ease of integration, testability and reusability from these techniques. These benefits motivate us to ask how component-based development techniques can be relevant in a mission operations environment, where there is significant investment in software tools that are not component-based and may not be written in languages for which component-based tools even exist. Trusted and complex software tools for sequencing, validation, navigation, and other vital functions cannot simply be re-written or abandoned in order to gain the advantages offered by emerging component-based software techniques. Thus some middle ground must be found. We have faced exactly this issue, and have found several solutions. Ensemble is an open platform for development, integration, and deployment of mission operations software that we are developing. Ensemble itself is an extension of an open source, component-based software development platform called Eclipse. Due to the advantages of component-based development, we have been able to vary rapidly develop mission operations tools for three surface missions by mixing and matching from a common set of mission operation components. We have also had to determine how to

  1. Association test based on SNP set: logistic kernel machine based test vs. principal component analysis.

    Directory of Open Access Journals (Sweden)

    Yang Zhao

    Full Text Available GWAS has facilitated greatly the discovery of risk SNPs associated with complex diseases. Traditional methods analyze SNP individually and are limited by low power and reproducibility since correction for multiple comparisons is necessary. Several methods have been proposed based on grouping SNPs into SNP sets using biological knowledge and/or genomic features. In this article, we compare the linear kernel machine based test (LKM and principal components analysis based approach (PCA using simulated datasets under the scenarios of 0 to 3 causal SNPs, as well as simple and complex linkage disequilibrium (LD structures of the simulated regions. Our simulation study demonstrates that both LKM and PCA can control the type I error at the significance level of 0.05. If the causal SNP is in strong LD with the genotyped SNPs, both the PCA with a small number of principal components (PCs and the LKM with kernel of linear or identical-by-state function are valid tests. However, if the LD structure is complex, such as several LD blocks in the SNP set, or when the causal SNP is not in the LD block in which most of the genotyped SNPs reside, more PCs should be included to capture the information of the causal SNP. Simulation studies also demonstrate the ability of LKM and PCA to combine information from multiple causal SNPs and to provide increased power over individual SNP analysis. We also apply LKM and PCA to analyze two SNP sets extracted from an actual GWAS dataset on non-small cell lung cancer.

  2. Microservices as an Evolutionary Architecture of Component-Based Development: A Think-aloud Study

    OpenAIRE

    Parizi, Reza M.

    2018-01-01

    Microservices become a fast growing and popular architectural style based on service-oriented development. One of the major advantages using component-based approaches is to support reuse. In this paper, we present a study of microservices and how these systems are related to the traditional abstract models of component-based systems. This research focuses on the core properties of microservices including their scalability, availability and resilience, consistency, coupling and cohesion, and ...

  3. Understanding place and health: a heuristic for using administrative data.

    Science.gov (United States)

    Frohlich, Katherine L; Dunn, James R; McLaren, Lindsay; Shiell, Alan; Potvin, Louise; Hawe, Penelope; Dassa, Clément; Thurston, Wilfreda E

    2007-06-01

    The increasing availability, use and limitations of administrative data for place-based population health research, and a lack of theory development, created the context for the current paper. We developed a heuristic to interrogate administrative data sets and to help us develop explanatory pathways for linking place and health. Guided by a worked example, we argue that some items in administrative data sets lend themselves to multiple theories, creating problems of inference owing to the implications of using inductive versus deductive reasoning during the research process, and that certain types of theories are privileged when used administrative data bases.

  4. Heuristics for multiobjective multiple sequence alignment.

    Science.gov (United States)

    Abbasi, Maryam; Paquete, Luís; Pereira, Francisco B

    2016-07-15

    Aligning multiple sequences arises in many tasks in Bioinformatics. However, the alignments produced by the current software packages are highly dependent on the parameters setting, such as the relative importance of opening gaps with respect to the increase of similarity. Choosing only one parameter setting may provide an undesirable bias in further steps of the analysis and give too simplistic interpretations. In this work, we reformulate multiple sequence alignment from a multiobjective point of view. The goal is to generate several sequence alignments that represent a trade-off between maximizing the substitution score and minimizing the number of indels/gaps in the sum-of-pairs score function. This trade-off gives to the practitioner further information about the similarity of the sequences, from which she could analyse and choose the most plausible alignment. We introduce several heuristic approaches, based on local search procedures, that compute a set of sequence alignments, which are representative of the trade-off between the two objectives (substitution score and indels). Several algorithm design options are discussed and analysed, with particular emphasis on the influence of the starting alignment and neighborhood search definitions on the overall performance. A perturbation technique is proposed to improve the local search, which provides a wide range of high-quality alignments. The proposed approach is tested experimentally on a wide range of instances. We performed several experiments with sequences obtained from the benchmark database BAliBASE 3.0. To evaluate the quality of the results, we calculate the hypervolume indicator of the set of score vectors returned by the algorithms. The results obtained allow us to identify reasonably good choices of parameters for our approach. Further, we compared our method in terms of correctly aligned pairs ratio and columns correctly aligned ratio with respect to reference alignments. Experimental results show

  5. An ontology for component-based models of water resource systems

    Science.gov (United States)

    Elag, Mostafa; Goodall, Jonathan L.

    2013-08-01

    Component-based modeling is an approach for simulating water resource systems where a model is composed of a set of components, each with a defined modeling objective, interlinked through data exchanges. Component-based modeling frameworks are used within the hydrologic, atmospheric, and earth surface dynamics modeling communities. While these efforts have been advancing, it has become clear that the water resources modeling community in particular, and arguably the larger earth science modeling community as well, faces a challenge of fully and precisely defining the metadata for model components. The lack of a unified framework for model component metadata limits interoperability between modeling communities and the reuse of models across modeling frameworks due to ambiguity about the model and its capabilities. To address this need, we propose an ontology for water resources model components that describes core concepts and relationships using the Web Ontology Language (OWL). The ontology that we present, which is termed the Water Resources Component (WRC) ontology, is meant to serve as a starting point that can be refined over time through engagement by the larger community until a robust knowledge framework for water resource model components is achieved. This paper presents the methodology used to arrive at the WRC ontology, the WRC ontology itself, and examples of how the ontology can aid in component-based water resources modeling by (i) assisting in identifying relevant models, (ii) encouraging proper model coupling, and (iii) facilitating interoperability across earth science modeling frameworks.

  6. Combining heuristic and statistical techniques in landslide hazard assessments

    Science.gov (United States)

    Cepeda, Jose; Schwendtner, Barbara; Quan, Byron; Nadim, Farrokh; Diaz, Manuel; Molina, Giovanni

    2014-05-01

    As a contribution to the Global Assessment Report 2013 - GAR2013, coordinated by the United Nations International Strategy for Disaster Reduction - UNISDR, a drill-down exercise for landslide hazard assessment was carried out by entering the results of both heuristic and statistical techniques into a new but simple combination rule. The data available for this evaluation included landslide inventories, both historical and event-based. In addition to the application of a heuristic method used in the previous editions of GAR, the availability of inventories motivated the use of statistical methods. The heuristic technique is largely based on the Mora & Vahrson method, which estimates hazard as the product of susceptibility and triggering factors, where classes are weighted based on expert judgment and experience. Two statistical methods were also applied: the landslide index method, which estimates weights of the classes for the susceptibility and triggering factors based on the evidence provided by the density of landslides in each class of the factors; and the weights of evidence method, which extends the previous technique to include both positive and negative evidence of landslide occurrence in the estimation of weights for the classes. One key aspect during the hazard evaluation was the decision on the methodology to be chosen for the final assessment. Instead of opting for a single methodology, it was decided to combine the results of the three implemented techniques using a combination rule based on a normalization of the results of each method. The hazard evaluation was performed for both earthquake- and rainfall-induced landslides. The country chosen for the drill-down exercise was El Salvador. The results indicate that highest hazard levels are concentrated along the central volcanic chain and at the centre of the northern mountains.

  7. A CORBA BASED ARCHITECTURE FOR ACCESSING REUSABLE SOFTWARE COMPONENTS ON THE WEB.

    Directory of Open Access Journals (Sweden)

    R. Cenk ERDUR

    2003-01-01

    Full Text Available In a very near future, as a result of the continious growth of Internet and advances in networking technologies, Internet will become the common software repository for people and organizations who employ component based reuse approach in their software development life cycles. In order to use the reusable components such as source codes, analysis, designs, design patterns during new software development processes, environments that support the identification of the components over Internet are needed. Basic elements of such an environment are the coordinator programs which deliver user requests to appropriate component libraries, user interfaces for querying, and programs that wrap the component libraries. First, a CORBA based architecture is proposed for such an environment. Then, an alternative architecture that is based on the Java 2 platform technologies is given for the same environment. Finally, the two architectures are compared.

  8. Methods of Si based ceramic components volatilization control in a gas turbine engine

    Science.gov (United States)

    Garcia-Crespo, Andres Jose; Delvaux, John; Dion Ouellet, Noemie

    2016-09-06

    A method of controlling volatilization of silicon based components in a gas turbine engine includes measuring, estimating and/or predicting a variable related to operation of the gas turbine engine; correlating the variable to determine an amount of silicon to control volatilization of the silicon based components in the gas turbine engine; and injecting silicon into the gas turbine engine to control volatilization of the silicon based components. A gas turbine with a compressor, combustion system, turbine section and silicon injection system may be controlled by a controller that implements the control method.

  9. Assessing Use of Cognitive Heuristic Representativeness in Clinical Reasoning

    OpenAIRE

    Payne, Velma L.; Crowley, Rebecca S.

    2008-01-01

    We performed a pilot study to investigate use of the cognitive heuristic Representativeness in clinical reasoning. We tested a set of tasks and assessments to determine whether subjects used the heuristics in reasoning, to obtain initial frequencies of heuristic use and related cognitive errors, and to collect cognitive process data using think-aloud techniques. The study investigates two aspects of the Representativeness heuristic - judging by perceived frequency and representativeness as ca...

  10. Automated generation of constructive ordering heuristics for educational timetabling

    OpenAIRE

    Pillay, Nelishia; Özcan, Ender

    2017-01-01

    Construction heuristics play an important role in solving combinatorial optimization problems. These heuristics are usually used to create an initial solution to the problem which is improved using optimization techniques such as metaheuristics. For examination timetabling and university course timetabling problems essentially graph colouring heuristics have been used for this purpose. The process of deriving heuristics manually for educational timetabling is a time consuming task. Furthermor...

  11. Adaptive selection of heuristics for improving exam timetables

    OpenAIRE

    Burke, Edmund; Qu, Rong; Soghier, Amr

    2014-01-01

    This paper presents a hyper-heuristic approach which hybridises low-level heuristic moves to improve timetables. Exams which cause a soft-constraint violation in the timetable are ordered and rescheduled to produce a better timetable. It is observed that both the order in which exams are rescheduled and the heuristic moves used to reschedule the exams and improve the timetable affect the quality of the solution produced. After testing different combinations in a hybrid hyper-heuristic approac...

  12. A Study on Components of Internal Control-Based Administrative System in Secondary Schools

    Science.gov (United States)

    Montri, Paitoon; Sirisuth, Chaiyuth; Lammana, Preeda

    2015-01-01

    The aim of this study was to study the components of the internal control-based administrative system in secondary schools, and make a Confirmatory Factor Analysis (CFA) to confirm the goodness of fit of empirical data and component model that resulted from the CFA. The study consisted of three steps: 1) studying of principles, ideas, and theories…

  13. Design and fabrication of a eccentric wheels based motorised alignment mechanism for cylindrical accelerator components

    International Nuclear Information System (INIS)

    Mundra, G.; Jain, V.; Karmarkar, Mangesh; Kotaiah, S.

    2006-01-01

    Precision alignment mechanisms with long term stability are required for accelerator components. For some of the components motorised and remotely operable alignment mechanism are required. An eccentric wheel mechanism based alignment system is very much suitable for such application. One such alignment system is designed, a prototype is machined/fabricated for SFDTL type accelerating structure and preliminary trial experiments have been done. (author)

  14. Prediction of Pure Component Adsorption Equilibria Using an Adsorption Isotherm Equation Based on Vacancy Solution Theory

    DEFF Research Database (Denmark)

    Marcussen, Lis; Aasberg-Petersen, K.; Krøll, Annette Elisabeth

    2000-01-01

    An adsorption isotherm equation for nonideal pure component adsorption based on vacancy solution theory and the Non-Random-Two-Liquid (NRTL) equation is found to be useful for predicting pure component adsorption equilibria at a variety of conditions. The isotherm equation is evaluated successfully...... adsorption systems, spreading pressure and isosteric heat of adsorption are also calculated....

  15. A Service Component-based Accounting and Charging Architecture to Support Interim Mechanisms across Multiple Domains

    NARCIS (Netherlands)

    Le, V.M.; van Beijnum, Bernhard J.F.; Huitema, G.B.

    Today, telematics services are o Aen compositions of different chargeable service components offered by different service providers. To enhance component-based accounting and charging, the service composition information is used to match with the corresponding charging structure of a service

  16. Design and Application of an Ontology for Component-Based Modeling of Water Systems

    Science.gov (United States)

    Elag, M.; Goodall, J. L.

    2012-12-01

    Many Earth system modeling frameworks have adopted an approach of componentizing models so that a large model can be assembled by linking a set of smaller model components. These model components can then be more easily reused, extended, and maintained by a large group of model developers and end users. While there has been a notable increase in component-based model frameworks in the Earth sciences in recent years, there has been less work on creating framework-agnostic metadata and ontologies for model components. Well defined model component metadata is needed, however, to facilitate sharing, reuse, and interoperability both within and across Earth system modeling frameworks. To address this need, we have designed an ontology for the water resources community named the Water Resources Component (WRC) ontology in order to advance the application of component-based modeling frameworks across water related disciplines. Here we present the design of the WRC ontology and demonstrate its application for integration of model components used in watershed management. First we show how the watershed modeling system Soil and Water Assessment Tool (SWAT) can be decomposed into a set of hydrological and ecological components that adopt the Open Modeling Interface (OpenMI) standard. Then we show how the components can be used to estimate nitrogen losses from land to surface water for the Baltimore Ecosystem study area. Results of this work are (i) a demonstration of how the WRC ontology advances the conceptual integration between components of water related disciplines by handling the semantic and syntactic heterogeneity present when describing components from different disciplines and (ii) an investigation of a methodology by which large models can be decomposed into a set of model components that can be well described by populating metadata according to the WRC ontology.

  17. Applying usability heuristics to radiotherapy systems

    International Nuclear Information System (INIS)

    Chan, Alvita J.; Islam, Mohammad K.; Rosewall, Tara; Jaffray, David A.; Easty, Anthony C.; Cafazzo, Joseph A.

    2012-01-01

    Background and purpose: Heuristic evaluations have been used to evaluate safety of medical devices by identifying and assessing usability issues. Since radiotherapy treatment delivery systems often consist of multiple complex user-interfaces, a heuristic evaluation was conducted to assess the potential safety issues of such a system. Material and methods: A heuristic evaluation was conducted to evaluate the treatment delivery system at Princess Margaret Hospital (Toronto, Canada). Two independent evaluators identified usability issues with the user-interfaces and rated the severity of each issue. Results: The evaluators identified 75 usability issues in total. Eighteen of them were rated as high severity, indicating the potential to have a major impact on patient safety. A majority of issues were found on the record and verify system, and many were associated with the patient setup process. While the hospital has processes in place to ensure patient safety, recommendations were developed to further mitigate the risks of potential consequences. Conclusions: Heuristic evaluation is an efficient and inexpensive method that can be successfully applied to radiotherapy delivery systems to identify usability issues and improve patient safety. Although this study was conducted only at one site, the findings may have broad implications for the design of these systems.

  18. Fast heuristics for a dynamic paratransit problem

    NARCIS (Netherlands)

    Cremers, M.L.A.G.; Klein Haneveld, W.K.; van der Vlerk, M.H.

    2008-01-01

    In a previous paper we developed a non-standard two-stage recourse model for the dynamic day-ahead paratransit planning problem. Two heuristics, which are frequently applied in the recourse model, contain many details which leads to large CPU times to solve instances of relatively small size. In

  19. A Heuristic for Improving Transmedia Exhibition Experience

    DEFF Research Database (Denmark)

    Selvadurai, Vashanth; Rosenstand, Claus Andreas Foss

    2017-01-01

    in the scientific field of designing transmedia experience in an exhibition context that links the pre- and post-activities to the actual visit (during-activities). The result of this study is a preliminary heuristic for establishing a relation between the platform and content complexity in transmedia exhibitions....

  20. Efficient heuristics for maximum common substructure search.

    Science.gov (United States)

    Englert, Péter; Kovács, Péter

    2015-05-26

    Maximum common substructure search is a computationally hard optimization problem with diverse applications in the field of cheminformatics, including similarity search, lead optimization, molecule alignment, and clustering. Most of these applications have strict constraints on running time, so heuristic methods are often preferred. However, the development of an algorithm that is both fast enough and accurate enough for most practical purposes is still a challenge. Moreover, in some applications, the quality of a common substructure depends not only on its size but also on various topological features of the one-to-one atom correspondence it defines. Two state-of-the-art heuristic algorithms for finding maximum common substructures have been implemented at ChemAxon Ltd., and effective heuristics have been developed to improve both their efficiency and the relevance of the atom mappings they provide. The implementations have been thoroughly evaluated and compared with existing solutions (KCOMBU and Indigo). The heuristics have been found to greatly improve the performance and applicability of the algorithms. The purpose of this paper is to introduce the applied methods and present the experimental results.

  1. Heuristics for speeding up gaze estimation

    DEFF Research Database (Denmark)

    Leimberg, Denis; Vester-Christensen, Martin; Ersbøll, Bjarne Kjær

    2005-01-01

    A deformable template method for eye tracking on full face images is presented. The strengths of the method are that it is fast and retains accuracy independently of the resolution. We compare the method with a state of the art active contour approach, showing that the heuristic method is more...

  2. The Heuristic Interpretation of Box Plots

    Science.gov (United States)

    Lem, Stephanie; Onghena, Patrick; Verschaffel, Lieven; Van Dooren, Wim

    2013-01-01

    Box plots are frequently used, but are often misinterpreted by students. Especially the area of the box in box plots is often misinterpreted as representing number or proportion of observations, while it actually represents their density. In a first study, reaction time evidence was used to test whether heuristic reasoning underlies this…

  3. Heuristic Classification. Technical Report Number 12.

    Science.gov (United States)

    Clancey, William J.

    A broad range of well-structured problems--embracing forms of diagnosis, catalog selection, and skeletal planning--are solved in expert computer systems by the method of heuristic classification. These programs have a characteristic inference structure that systematically relates data to a pre-enumerated set of solutions by abstraction, heuristic…

  4. Engineering applications of heuristic multilevel optimization methods

    Science.gov (United States)

    Barthelemy, Jean-Francois M.

    1989-01-01

    Some engineering applications of heuristic multilevel optimization methods are presented and the discussion focuses on the dependency matrix that indicates the relationship between problem functions and variables. Coordination of the subproblem optimizations is shown to be typically achieved through the use of exact or approximate sensitivity analysis. Areas for further development are identified.

  5. A Heuristic for the Teaching of Persuasion.

    Science.gov (United States)

    Schell, John F.

    Interpreting Aristotle's criteria for persuasive writing--ethos, logos, and pathos--as a concern for writer, language, and audience creates both an effective model for persuasive writing and a structure around which to organize discussions of relevant rhetorical issues. Use of this heuristic to analyze writing style, organization, and content…

  6. Heuristic Decision Making in Network Linking

    NARCIS (Netherlands)

    M.J.W. Harmsen - Van Hout (Marjolein); B.G.C. Dellaert (Benedict); P.J.J. Herings (Jean-Jacques)

    2015-01-01

    textabstractNetwork formation among individuals constitutes an important part of many OR processes, but relatively little is known about how individuals make their linking decisions in networks. This article provides an investigation of heuristic effects in individual linking decisions for

  7. Investigating Heuristic Evaluation: A Case Study.

    Science.gov (United States)

    Goldman, Kate Haley; Bendoly, Laura

    When museum professionals speak of evaluating a web site, they primarily mean formative evaluation, and by that they primarily mean testing the usability of the site. In the for-profit world, usability testing is a multi-million dollar industry, while non-profits often rely on far too few dollars to do too much. Hence, heuristic evaluation is one…

  8. Internal Medicine residents use heuristics to estimate disease probability

    Directory of Open Access Journals (Sweden)

    Sen Phang

    2015-12-01

    Conclusions: Our findings suggest that despite previous exposure to the use of Bayesian reasoning, residents use heuristics, such as the representative heuristic and anchoring with adjustment, to estimate probabilities. Potential reasons for attribute substitution include the relative cognitive ease of heuristics vs. Bayesian reasoning or perhaps residents in their clinical practice use gist traces rather than precise probability estimates when diagnosing.

  9. Heuristics Made Easy: An Effort-Reduction Framework

    Science.gov (United States)

    Shah, Anuj K.; Oppenheimer, Daniel M.

    2008-01-01

    In this article, the authors propose a new framework for understanding and studying heuristics. The authors posit that heuristics primarily serve the purpose of reducing the effort associated with a task. As such, the authors propose that heuristics can be classified according to a small set of effort-reduction principles. The authors use this…

  10. Heuristic Diagrams as a Tool to Teach History of Science

    Science.gov (United States)

    Chamizo, Jose A.

    2012-01-01

    The graphic organizer called here heuristic diagram as an improvement of Gowin's Vee heuristic is proposed as a tool to teach history of science. Heuristic diagrams have the purpose of helping students (or teachers, or researchers) to understand their own research considering that asks and problem-solving are central to scientific activity. The…

  11. Nuclear electronic components of surface contamination monitor based on multi-electrode proportional counter

    International Nuclear Information System (INIS)

    Du Xiangyang; Zhang Yong; Han Shuping; Rao Xianming; Fang Jintu

    2001-01-01

    The nuclear electronic components applying in Portal Monitor and Hands and Feet Surface Contamination Monitor were based on modern integrated circuit are introduced. The detailed points in circuit design and manufacturing technique are analyzed

  12. A component-based open hypermedia approach to integreting structure services

    DEFF Research Database (Denmark)

    Grønbæk, Kaj; Nürnberg, Peter J.; Bucka-Lassen, Dirk

    1999-01-01

    In this paper, we consider the issue of integrating different structure services within a component-based open hypermedia system. We do so by considering the task of collaborative editing, which calls for a variety of different structures traditionally supplied by different structure services. We...... discuss the nature of collaborative editing and how it can be supported by a combination of spatial and navigational hypermedia services. We then present a component-based open hypermedia system architecture and describe various methods of integrating different structure services provided within...... such an architecture. We show the advantages of integration within a component-based framework over other means of integration, highlighting some of the main advantages of the component-based approach to open hypermedia system design and implementation....

  13. Model-Based Design Tools for Extending COTS Components To Extreme Environments, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — The innovation in this project is model-based design (MBD) tools for predicting the performance and useful life of commercial-off-the-shelf (COTS) components and...

  14. RF Front End Based on MEMS Components for Miniaturized Digital EVA Radio, Phase II

    Data.gov (United States)

    National Aeronautics and Space Administration — In this SBIR project, AlphaSense, Inc. and the Carnegie Mellon University propose to develop a RF receiver front end based on CMOS-MEMS components for miniaturized...

  15. Uniframe: A Unified Framework for Developing Service-Oriented, Component-Based Distributed Software Systems

    National Research Council Canada - National Science Library

    Raje, Rajeev R; Olson, Andrew M; Bryant, Barrett R; Burt, Carol C; Auguston, Makhail

    2005-01-01

    .... It describes how this approach employs a unifying framework for specifying such systems to unite the concepts of service-oriented architectures, a component-based software engineering methodology...

  16. Adding a Performance-Based Component to Surface Warfare Officer Bonuses: Will it Affect Retention?

    National Research Council Canada - National Science Library

    Carman, Aron S; Mudd, Ryan M

    2008-01-01

    ... Authorization and the current officer inventory beginning at 9 years of commissioned service. The objective of this study was to analyze the 13-year retention effect of adding a performance-based component to the SWO Critical Skills Bonus (CSB...

  17. Acquisition and representation of knowledge related to crack formation in power plant components - practical implementation in a knowledge-based system (ESR)

    International Nuclear Information System (INIS)

    Jovanovic, A.; Kautz, H.R.; Raepke, H.; Witte, M. de

    1992-01-01

    The damage analysis in general and analysis of crack formation in particular, must rely on a careful balance of generic principles (e.g. generic knowledge about damage mechanisms) and information regarding the analyzed component (e.g. operating history, manufacturing data or construction records). Use of non-standard and/or advanced approaches (for damage parameters, monitoring techniques, etc.) can, on the one hand, enhance the solution by making it more comprehensive, but, on the other hand, it can also make the final decision more complex. In both cases, at the current state of practice, only heuristic knowledge can assure that all the important factors be properly considered. The examples shown in the paper illustrate the feasibility of obtaining useful interim results, which, through a continued long-term effort invested in knowledge acquisition and representation, and its integration into a knowledge-based system allow to achieve progressively the goals formulated at the beginning of the expert system for remaining life assessment project. (orig./DG)

  18. Heuristics structure and pervade formal risk assessment.

    Science.gov (United States)

    MacGillivray, Brian H

    2014-04-01

    Lay perceptions of risk appear rooted more in heuristics than in reason. A major concern of the risk regulation literature is that such "error-strewn" perceptions may be replicated in policy, as governments respond to the (mis)fears of the citizenry. This has led many to advocate a relatively technocratic approach to regulating risk, characterized by high reliance on formal risk and cost-benefit analysis. However, through two studies of chemicals regulation, we show that the formal assessment of risk is pervaded by its own set of heuristics. These include rules to categorize potential threats, define what constitutes valid data, guide causal inference, and to select and apply formal models. Some of these heuristics lay claim to theoretical or empirical justifications, others are more back-of-the-envelope calculations, while still more purport not to reflect some truth but simply to constrain discretion or perform a desk-clearing function. These heuristics can be understood as a way of authenticating or formalizing risk assessment as a scientific practice, representing a series of rules for bounding problems, collecting data, and interpreting evidence (a methodology). Heuristics are indispensable elements of induction. And so they are not problematic per se, but they can become so when treated as laws rather than as contingent and provisional rules. Pitfalls include the potential for systematic error, masking uncertainties, strategic manipulation, and entrenchment. Our central claim is that by studying the rules of risk assessment qua rules, we develop a novel representation of the methods, conventions, and biases of the prior art. © 2013 Society for Risk Analysis.

  19. Power Transformer Differential Protection Based on Neural Network Principal Component Analysis, Harmonic Restraint and Park's Plots

    OpenAIRE

    Tripathy, Manoj

    2012-01-01

    This paper describes a new approach for power transformer differential protection which is based on the wave-shape recognition technique. An algorithm based on neural network principal component analysis (NNPCA) with back-propagation learning is proposed for digital differential protection of power transformer. The principal component analysis is used to preprocess the data from power system in order to eliminate redundant information and enhance hidden pattern of differential current to disc...

  20. A Component-based Software Development and Execution Framework for CAx Applications

    Directory of Open Access Journals (Sweden)

    N. Matsuki

    2004-01-01

    Full Text Available Digitalization of the manufacturing process and technologies is regarded as the key to increased competitive ability. The MZ-Platform infrastructure is a component-based software development framework, designed for supporting enterprises to enhance digitalized technologies using software tools and CAx components in a self-innovative way. In the paper we show the algorithm, system architecture, and a CAx application example on MZ-Platform. We also propose a new parametric data structure based on MZ-Platform.

  1. Data base formation for important components of reactor TRIGA MARK II

    International Nuclear Information System (INIS)

    Jordan, R.; Mavko, B.; Kozuh, M.

    1992-01-01

    The paper represents specific data base formation for reactor TRIGA MARK II in Podgorica. Reactor operation data from year 1985 to 1990 were collected. Two groups of collected data were formed. The first group includes components data and the second group covers data of reactor scrams. Time related and demand related models were used for data evaluation. Parameters were estimated by classical method. Similar data bases are useful everywhere where components unavailabilities may have severe drawback. (author) [sl

  2. Software Component Clustering and Retrieval: An Entropy-based Fuzzy k-Modes Methodology

    OpenAIRE

    Stylianou, Constantinos; Andreou, Andreas S.

    2008-01-01

    The number of software houses attempting to adopt a component-based development approach is rapidly increasing. However many organisations still find it difficult to complete the shift as it requires them to alter their entire software development process and philosophy. Furthermore, to promote component-based software engineering, organisations must be ready to promote reusability and this can only be attained if the proper framework exists from which a developer can access, search and retri...

  3. Safety prediction for basic components of safety critical software based on static testing

    International Nuclear Information System (INIS)

    Son, H.S.; Seong, P.H.

    2001-01-01

    The purpose of this work is to develop a safety prediction method, with which we can predict the risk of software components based on static testing results at the early development stage. The predictive model combines the major factor with the quality factor for the components, both of which are calculated based on the measures proposed in this work. The application to a safety-critical software system demonstrates the feasibility of the safety prediction method. (authors)

  4. Safety prediction for basic components of safety-critical software based on static testing

    International Nuclear Information System (INIS)

    Son, H.S.; Seong, P.H.

    2000-01-01

    The purpose of this work is to develop a safety prediction method, with which we can predict the risk of software components based on static testing results at the early development stage. The predictive model combines the major factor with the quality factor for the components, which are calculated based on the measures proposed in this work. The application to a safety-critical software system demonstrates the feasibility of the safety prediction method. (authors)

  5. Heuristic Evaluation of E-Learning Courses: A Comparative Analysis of Two E-Learning Heuristic Sets

    Science.gov (United States)

    Zaharias, Panagiotis; Koutsabasis, Panayiotis

    2012-01-01

    Purpose: The purpose of this paper is to discuss heuristic evaluation as a method for evaluating e-learning courses and applications and more specifically to investigate the applicability and empirical use of two customized e-learning heuristic protocols. Design/methodology/approach: Two representative e-learning heuristic protocols were chosen…

  6. Smart strategies for doctors and doctors-in-training: heuristics in medicine.

    Science.gov (United States)

    Wegwarth, Odette; Gaissmaier, Wolfgang; Gigerenzer, Gerd

    2009-08-01

    How do doctors make sound decisions when confronted with probabilistic data, time pressures and a heavy workload? One theory that has been embraced by many researchers is based on optimisation, which emphasises the need to integrate all information in order to arrive at sound decisions. This notion makes heuristics, which use less than complete information, appear as second-best strategies. In this article, we challenge this pessimistic view of heuristics. We introduce two medical problems that involve decision making to the reader: one concerns coronary care issues and the other macrolide prescriptions. In both settings, decision-making tools grounded in the principles of optimisation and heuristics, respectively, have been developed to assist doctors in making decisions. We explain the structure of each of these tools and compare their performance in terms of their facilitation of correct predictions. For decisions concerning both the coronary care unit and the prescribing of macrolides, we demonstrate that sacrificing information does not necessarily imply a forfeiting of predictive accuracy, but can sometimes even lead to better decisions. Subsequently, we discuss common misconceptions about heuristics and explain when and why ignoring parts of the available information can lead to the making of more robust predictions. Heuristics are neither good nor bad per se, but, if applied in situations to which they have been adapted, can be helpful companions for doctors and doctors-in-training. This, however, requires that heuristics in medicine be openly discussed, criticised, refined and then taught to doctors-in-training rather than being simply dismissed as harmful or irrelevant. A more uniform use of explicit and accepted heuristics has the potential to reduce variations in diagnoses and to improve medical care for patients.

  7. A New Approach to Tuning Heuristic Parameters of Genetic Algorithms

    Czech Academy of Sciences Publication Activity Database

    Holeňa, Martin

    2006-01-01

    Roč. 3, č. 3 (2006), s. 562-569 ISSN 1790-0832. [AIKED'06. WSEAS International Conference on Artificial Intelligence , Knowledge Engineering and Data Bases. Madrid, 15.02.2006-17.02.2006] R&D Projects: GA ČR(CZ) GA201/05/0325; GA ČR(CZ) GA201/05/0557 Institutional research plan: CEZ:AV0Z10300504 Keywords : evolutionary optimization * genetic algorithms * heuristic parameters * parameter tuning * artificial neural networks * convergence speed * population diversity Subject RIV: IN - Informatics, Computer Science

  8. A NEW HEURISTIC ALGORITHM FOR MULTIPLE TRAVELING SALESMAN PROBLEM

    Directory of Open Access Journals (Sweden)

    F. NURIYEVA

    2017-06-01

    Full Text Available The Multiple Traveling Salesman Problem (mTSP is a combinatorial optimization problem in NP-hard class. The mTSP aims to acquire the minimum cost for traveling a given set of cities by assigning each of them to a different salesman in order to create m number of tours. This paper presents a new heuristic algorithm based on the shortest path algorithm to find a solution for the mTSP. The proposed method has been programmed in C language and its performance analysis has been carried out on the library instances. The computational results show the efficiency of this method.

  9. Four challenges for cognitive research on the recognition heuristic and a call for a research strategy shift

    Directory of Open Access Journals (Sweden)

    Tracy Tomlinson

    2011-02-01

    Full Text Available The recognition heuristic assumes that people make inferences based on the output of recognition memory. While much work has been devoted to establishing the recognition heuristic as a viable description of how people make inferences, more work is needed to fully integrate research on the recognition heuristic with research from the broader cognitive psychology literature. In this article, we outline four challenges that should be met for this integration to take place, and close with a call to address these four challenges collectively, rather than piecemeal.

  10. Identifying product development crises: The potential of adaptive heuristics

    DEFF Research Database (Denmark)

    Münzberger, C.; Stingl, Verena; Oehmen, Josef

    2017-01-01

    This paper introduces adaptive heuristics as a tool to identify crises in design projects and highlights potential applications of these heuristics as decision support tool for crisis identification. Crises may emerge slowly or suddenly, and often have ambiguous signals. Thus the identification...... for the application of heuristics in design sciences. To achieve this, the paper compares crises to 'business as usual', and presents sixteen indicators for emerging crises. These indicators are potential cues for adaptive heuristics. Specifically three adaptive heuristics, One-single-cue, Fast-and-Frugal-Trees...

  11. Learning process mapping heuristics under stochastic sampling overheads

    Science.gov (United States)

    Ieumwananonthachai, Arthur; Wah, Benjamin W.

    1991-01-01

    A statistical method was developed previously for improving process mapping heuristics. The method systematically explores the space of possible heuristics under a specified time constraint. Its goal is to get the best possible heuristics while trading between the solution quality of the process mapping heuristics and their execution time. The statistical selection method is extended to take into consideration the variations in the amount of time used to evaluate heuristics on a problem instance. The improvement in performance is presented using the more realistic assumption along with some methods that alleviate the additional complexity.

  12. Textile-Based Electronic Components for Energy Applications: Principles, Problems, and Perspective.

    Science.gov (United States)

    Kaushik, Vishakha; Lee, Jaehong; Hong, Juree; Lee, Seulah; Lee, Sanggeun; Seo, Jungmok; Mahata, Chandreswar; Lee, Taeyoon

    2015-09-07

    Textile-based electronic components have gained interest in the fields of science and technology. Recent developments in nanotechnology have enabled the integration of electronic components into textiles while retaining desirable characteristics such as flexibility, strength, and conductivity. Various materials were investigated in detail to obtain current conductive textile technology, and the integration of electronic components into these textiles shows great promise for common everyday applications. The harvest and storage of energy in textile electronics is a challenge that requires further attention in order to enable complete adoption of this technology in practical implementations. This review focuses on the various conductive textiles, their methods of preparation, and textile-based electronic components. We also focus on fabrication and the function of textile-based energy harvesting and storage devices, discuss their fundamental limitations, and suggest new areas of study.

  13. Textile-Based Electronic Components for Energy Applications: Principles, Problems, and Perspective

    Directory of Open Access Journals (Sweden)

    Vishakha Kaushik

    2015-09-01

    Full Text Available Textile-based electronic components have gained interest in the fields of science and technology. Recent developments in nanotechnology have enabled the integration of electronic components into textiles while retaining desirable characteristics such as flexibility, strength, and conductivity. Various materials were investigated in detail to obtain current conductive textile technology, and the integration of electronic components into these textiles shows great promise for common everyday applications. The harvest and storage of energy in textile electronics is a challenge that requires further attention in order to enable complete adoption of this technology in practical implementations. This review focuses on the various conductive textiles, their methods of preparation, and textile-based electronic components. We also focus on fabrication and the function of textile-based energy harvesting and storage devices, discuss their fundamental limitations, and suggest new areas of study.

  14. Fault Diagnosis Method Based on Information Entropy and Relative Principal Component Analysis

    Directory of Open Access Journals (Sweden)

    Xiaoming Xu

    2017-01-01

    Full Text Available In traditional principle component analysis (PCA, because of the neglect of the dimensions influence between different variables in the system, the selected principal components (PCs often fail to be representative. While the relative transformation PCA is able to solve the above problem, it is not easy to calculate the weight for each characteristic variable. In order to solve it, this paper proposes a kind of fault diagnosis method based on information entropy and Relative Principle Component Analysis. Firstly, the algorithm calculates the information entropy for each characteristic variable in the original dataset based on the information gain algorithm. Secondly, it standardizes every variable’s dimension in the dataset. And, then, according to the information entropy, it allocates the weight for each standardized characteristic variable. Finally, it utilizes the relative-principal-components model established for fault diagnosis. Furthermore, the simulation experiments based on Tennessee Eastman process and Wine datasets demonstrate the feasibility and effectiveness of the new method.

  15. Development of active-X component for use in web based thermal hydraulic data bank

    International Nuclear Information System (INIS)

    Lee, Y. J.; Chung, B. D.

    2003-01-01

    An active-X component to use as the engine for the web-based thermal hydraulic data bank has been developed. The development of the active-X component was carried out primarily for employment in the web-based thermal-hydraulic databank. The active-X component was developed with the objective to minimize the size of the component and the data traffic while maximizing the functionality. For this end, the data is downloaded in a compressed format to minimize the downloading time, and Delphi language is used in the efforts to minimize the size of the active-X component as well as for fast execution time. The functionality of active-X component was tested on ENCOUNTER data package by embedding the component in a prototype web-page under a server-client environment. The test demonstrated that the active-X component functions as intended and that it is capable of very easy data retrieval and display

  16. On the suitability of fast and frugal heuristics for designing values clarification methods in patient decision aids: a critical analysis.

    Science.gov (United States)

    Pieterse, Arwen H; de Vries, Marieke

    2013-09-01

    Increasingly, patient decision aids and values clarification methods (VCMs) are being developed to support patients in making preference-sensitive health-care decisions. Many VCMs encourage extensive deliberation about options, without solid theoretical or empirical evidence showing that deliberation is advantageous. Research suggests that simple, fast and frugal heuristic decision strategies sometimes result in better judgments and decisions. Durand et al. have developed two fast and frugal heuristic-based VCMs. To critically analyse the suitability of the 'take the best' (TTB) and 'tallying' fast and frugal heuristics in the context of patient decision making. Analysis of the structural similarities between the environments in which the TTB and tallying heuristics have been proven successful and the context of patient decision making and of the potential of these heuristic decision processes to support patient decision making. The specific nature of patient preference-sensitive decision making does not seem to resemble environments in which the TTB and tallying heuristics have proven successful. Encouraging patients to consider less rather than more relevant information potentially even deteriorates their values clarification process. Values clarification methods promoting the use of more intuitive decision strategies may sometimes be more effective. Nevertheless, we strongly recommend further theoretical thinking about the expected value of such heuristics and of other more intuitive decision strategies in this context, as well as empirical assessments of the mechanisms by which inducing such decision strategies may impact the quality and outcome of values clarification. © 2011 John Wiley & Sons Ltd.

  17. Reconsidering "evidence" for fast-and-frugal heuristics.

    Science.gov (United States)

    Hilbig, Benjamin E

    2010-12-01

    In several recent reviews, authors have argued for the pervasive use of fast-and-frugal heuristics in human judgment. They have provided an overview of heuristics and have reiterated findings corroborating that such heuristics can be very valid strategies leading to high accuracy. They also have reviewed previous work that implies that simple heuristics are actually used by decision makers. Unfortunately, concerning the latter point, these reviews appear to be somewhat incomplete. More important, previous conclusions have been derived from investigations that bear some noteworthy methodological limitations. I demonstrate these by proposing a new heuristic and provide some novel critical findings. Also, I review some of the relevant literature often not-or only partially-considered. Overall, although some fast-and-frugal heuristics indeed seem to predict behavior at times, there is little to no evidence for others. More generally, the empirical evidence available does not warrant the conclusion that heuristics are pervasively used.

  18. Modified meta-heuristics using random mutation for truss topology optimization with static and dynamic constraints

    Directory of Open Access Journals (Sweden)

    Vimal J. Savsani

    2017-04-01

    The static and dynamic responses to the TTO problems are challenging due to its search space, which is implicit, non-convex, non-linear, and often leading to divergence. Modified meta-heuristics are effective optimization methods to handle such problems in actual fact. In this paper, modified versions of Teaching–Learning-Based Optimization (TLBO, Heat Transfer Search (HTS, Water Wave Optimization (WWO, and Passing Vehicle Search (PVS are proposed by integrating the random mutation-based search technique with them. This paper compares the performance of four modified and four basic meta-heuristics to solve discrete TTO problems.

  19. A heuristic and hybrid method for the tank allocation problem in maritime bulk shipping

    DEFF Research Database (Denmark)

    Vilhelmsen, Charlotte; Larsen, Jesper; Lusby, Richard Martin

    2016-01-01

    In bulk shipping, ships often have multiple tanks and carry multiple inhomogeneous products at a time. When operating such ships it is therefore a major challenge to decide how to best allocate cargoes to available tanks while taking into account tank capacity, safety restrictions, ship stability...... finding a feasible solution. We have developed a heuristic that can efficiently find feasible cargo allocations. Computational results show that it can solve 99 % of the considered instances within 0.4 s and all of them if allowed longer time. We have also modified an optimality based method from...... the literature. The heuristic is much faster than this modified method on the vast majority of considered instances. However, the heuristic struggles on two instances which are relatively quickly solved by the modified optimality based method. These two methods therefore complement each other nicely and so, we...

  20. A Heuristic and Hybrid Method for the Tank Allocation Problem in Maritime Bulk Shipping

    DEFF Research Database (Denmark)

    Vilhelmsen, Charlotte; Larsen, Jesper; Lusby, Richard Martin

    In bulk shipping, ships often have multiple tanks and carry multiple inhomogeneous products at a time. When operating such ships it is therefore a major challenge to decide how to best allocate cargoes to available tanks while taking into account tank capacity, safety restrictions, ship stability...... finding a feasible solution. We have developed a heuristic that can efficiently find feasible cargo allocations. Computational results show that it can solve 99% of the considered instances within 0.4 seconds and all of them if allowed longer time. We have also modified an optimality based method from...... the literature. The heuristic is much faster than this modified method on the vast majority of considered instances. However, the heuristic struggles on two instances which are relatively quickly solved by the modified optimality based method. These two methods therefore complement each other nicely and so, we...

  1. Principal component analysis of tomato genotypes based on some morphological and biochemical quality indicators

    Directory of Open Access Journals (Sweden)

    Glogovac Svetlana

    2012-01-01

    Full Text Available This study investigates variability of tomato genotypes based on morphological and biochemical fruit traits. Experimental material is a part of tomato genetic collection from Institute of Filed and Vegetable Crops in Novi Sad, Serbia. Genotypes were analyzed for fruit mass, locule number, index of fruit shape, fruit colour, dry matter content, total sugars, total acidity, lycopene and vitamin C. Minimum, maximum and average values and main indicators of variability (CV and σ were calculated. Principal component analysis was performed to determinate variability source structure. Four principal components, which contribute 93.75% of the total variability, were selected for analysis. The first principal component is defined by vitamin C, locule number and index of fruit shape. The second component is determined by dry matter content, and total acidity, the third by lycopene, fruit mass and fruit colour. Total sugars had the greatest part in the fourth component.

  2. High Temperature Corrosion Problem of Boiler Components in presence of Sulfur and Alkali based Fuels

    Science.gov (United States)

    Ghosh, Debashis; Mitra, Swapan Kumar

    2011-04-01

    Material degradation and ageing is of particular concern for fossil fuel fired power plant components. New techniques/approaches have been explored in recent years for Residual Life assessment of aged components and material degradation due to different damage mechanism like creep, fatigue, corrosion and erosion etc. Apart from the creep, the high temperature corrosion problem in a fossil fuel fired boiler is a matter of great concern if the fuel contains sulfur, chlorine sodium, potassium and vanadium etc. This paper discusses the material degradation due to high temperature corrosion in different critical components of boiler like water wall, superheater and reheater tubes and also remedial measures to avoid the premature failure. This paper also high lights the Residual Life Assessment (RLA) methodology of the components based on high temperature fireside corrosion. of different critical components of boiler.

  3. Map matching and heuristic elimination of gyro drift for personal navigation systems in GPS-denied conditions

    International Nuclear Information System (INIS)

    Aggarwal, Priyanka; Thomas, David; Ojeda, Lauro; Borenstein, Johann

    2011-01-01

    This paper introduces a method for the substantial reduction of heading errors in inertial navigation systems used under GPS-denied conditions. Presumably, the method is applicable for both vehicle-based and personal navigation systems, but experiments were performed only with a personal navigation system called 'personal dead reckoning' (PDR). In order to work under GPS-denied conditions, the PDR system uses a foot-mounted inertial measurement unit (IMU). However, gyro drift in this IMU can cause large heading errors after just a few minutes of walking. To reduce these errors, the map-matched heuristic drift elimination (MAPHDE) method was developed, which estimates gyro drift errors by comparing IMU-derived heading to the direction of the nearest street segment in a database of street maps. A heuristic component in this method provides tolerance to short deviations from walking along the street, such as when crossing streets or intersections. MAPHDE keeps heading errors almost at zero, and, as a result, position errors are dramatically reduced. In this paper, MAPHDE was used in a variety of outdoor walks, without any use of GPS. This paper explains the MAPHDE method in detail and presents experimental results

  4. Application of heuristic and machine-learning approach to engine model calibration

    Science.gov (United States)

    Cheng, Jie; Ryu, Kwang R.; Newman, C. E.; Davis, George C.

    1993-03-01

    Automation of engine model calibration procedures is a very challenging task because (1) the calibration process searches for a goal state in a huge, continuous state space, (2) calibration is often a lengthy and frustrating task because of complicated mutual interference among the target parameters, and (3) the calibration problem is heuristic by nature, and often heuristic knowledge for constraining a search cannot be easily acquired from domain experts. A combined heuristic and machine learning approach has, therefore, been adopted to improve the efficiency of model calibration. We developed an intelligent calibration program called ICALIB. It has been used on a daily basis for engine model applications, and has reduced the time required for model calibrations from many hours to a few minutes on average. In this paper, we describe the heuristic control strategies employed in ICALIB such as a hill-climbing search based on a state distance estimation function, incremental problem solution refinement by using a dynamic tolerance window, and calibration target parameter ordering for guiding the search. In addition, we present the application of a machine learning program called GID3* for automatic acquisition of heuristic rules for ordering target parameters.

  5. Optimal condition-based maintenance decisions for systems with dependent stochastic degradation of components

    International Nuclear Information System (INIS)

    Hong, H.P.; Zhou, W.; Zhang, S.; Ye, W.

    2014-01-01

    Components in engineered systems are subjected to stochastic deterioration due to the operating environmental conditions, and the uncertainty in material properties. The components need to be inspected and possibly replaced based on preventive or failure replacement criteria to provide the intended and safe operation of the system. In the present study, we investigate the influence of dependent stochastic degradation of multiple components on the optimal maintenance decisions. We use copula to model the dependent stochastic degradation of components, and formulate the optimal decision problem based on the minimum expected cost rule and the stochastic dominance rules. The latter is used to cope with decision maker's risk attitude. We illustrate the developed probabilistic analysis approach and the influence of the dependency of the stochastic degradation on the preferred decisions through numerical examples

  6. A heuristic model for working memory deficit in schizophrenia.

    Science.gov (United States)

    Qi, Zhen; Yu, Gina P; Tretter, Felix; Pogarell, Oliver; Grace, Anthony A; Voit, Eberhard O

    2016-11-01

    The life of schizophrenia patients is severely affected by deficits in working memory. In various brain regions, the reciprocal interactions between excitatory glutamatergic neurons and inhibitory GABAergic neurons are crucial. Other neurotransmitters, in particular dopamine, serotonin, acetylcholine, and norepinephrine, modulate the local balance between glutamate and GABA and therefore regulate the function of brain regions. Persistent alterations in the balances between the neurotransmitters can result in working memory deficits. Here we present a heuristic computational model that accounts for interactions among neurotransmitters across various brain regions. The model is based on the concept of a neurochemical interaction matrix at the biochemical level and combines this matrix with a mobile model representing physiological dynamic balances among neurotransmitter systems associated with working memory. The comparison of clinical and simulation results demonstrates that the model output is qualitatively very consistent with the available data. In addition, the model captured how perturbations migrated through different neurotransmitters and brain regions. Results showed that chronic administration of ketamine can cause a variety of imbalances, and application of an antagonist of the D2 receptor in PFC can also induce imbalances but in a very different manner. The heuristic computational model permits a variety of assessments of genetic, biochemical, and pharmacological perturbations and serves as an intuitive tool for explaining clinical and biological observations. The heuristic model is more intuitive than biophysically detailed models. It can serve as an important tool for interdisciplinary communication and even for psychiatric education of patients and relatives. This article is part of a Special Issue entitled "System Genetics" Guest Editor: Dr. Yudong Cai and Dr. Tao Huang. Copyright © 2016 Elsevier B.V. All rights reserved.

  7. Color Independent Components Based SIFT Descriptors for Object/Scene Classification

    Science.gov (United States)

    Ai, Dan-Ni; Han, Xian-Hua; Ruan, Xiang; Chen, Yen-Wei

    In this paper, we present a novel color independent components based SIFT descriptor (termed CIC-SIFT) for object/scene classification. We first learn an efficient color transformation matrix based on independent component analysis (ICA), which is adaptive to each category in a database. The ICA-based color transformation can enhance contrast between the objects and the background in an image. Then we compute CIC-SIFT descriptors over all three transformed color independent components. Since the ICA-based color transformation can boost the objects and suppress the background, the proposed CIC-SIFT can extract more effective and discriminative local features for object/scene classification. The comparison is performed among seven SIFT descriptors, and the experimental classification results show that our proposed CIC-SIFT is superior to other conventional SIFT descriptors.

  8. Predictive based monitoring of nuclear plant component degradation using support vector regression

    International Nuclear Information System (INIS)

    Agarwal, Vivek; Alamaniotis, Miltiadis; Tsoukalas, Lefteri H.

    2015-01-01

    Nuclear power plants (NPPs) are large installations comprised of many active and passive assets. Degradation monitoring of all these assets is expensive (labor cost) and highly demanding task. In this paper a framework based on Support Vector Regression (SVR) for online surveillance of critical parameter degradation of NPP components is proposed. In this case, on time replacement or maintenance of components will prevent potential plant malfunctions, and reduce the overall operational cost. In the current work, we apply SVR equipped with a Gaussian kernel function to monitor components. Monitoring includes the one-step-ahead prediction of the component's respective operational quantity using the SVR model, while the SVR model is trained using a set of previous recorded degradation histories of similar components. Predictive capability of the model is evaluated upon arrival of a sensor measurement, which is compared to the component failure threshold. A maintenance decision is based on a fuzzy inference system that utilizes three parameters: (i) prediction evaluation in the previous steps, (ii) predicted value of the current step, (iii) and difference of current predicted value with components failure thresholds. The proposed framework will be tested on turbine blade degradation data.

  9. Reliability Evaluation of Machine Center Components Based on Cascading Failure Analysis

    Science.gov (United States)

    Zhang, Ying-Zhi; Liu, Jin-Tong; Shen, Gui-Xiang; Long, Zhe; Sun, Shu-Guang

    2017-07-01

    In order to rectify the problems that the component reliability model exhibits deviation, and the evaluation result is low due to the overlook of failure propagation in traditional reliability evaluation of machine center components, a new reliability evaluation method based on cascading failure analysis and the failure influenced degree assessment is proposed. A direct graph model of cascading failure among components is established according to cascading failure mechanism analysis and graph theory. The failure influenced degrees of the system components are assessed by the adjacency matrix and its transposition, combined with the Pagerank algorithm. Based on the comprehensive failure probability function and total probability formula, the inherent failure probability function is determined to realize the reliability evaluation of the system components. Finally, the method is applied to a machine center, it shows the following: 1) The reliability evaluation values of the proposed method are at least 2.5% higher than those of the traditional method; 2) The difference between the comprehensive and inherent reliability of the system component presents a positive correlation with the failure influenced degree of the system component, which provides a theoretical basis for reliability allocation of machine center system.

  10. A Natural Component-Based Oxygen Indicator with In-Pack Activation for Intelligent Food Packaging.

    Science.gov (United States)

    Won, Keehoon; Jang, Nan Young; Jeon, Junsu

    2016-12-28

    Intelligent food packaging can provide consumers with reliable and correct information on the quality and safety of packaged foods. One of the key constituents in intelligent packaging is a colorimetric oxygen indicator, which is widely used to detect oxygen gas involved in food spoilage by means of a color change. Traditional oxygen indicators consisting of redox dyes and strong reducing agents have two major problems: they must be manufactured and stored under anaerobic conditions because air depletes the reductant, and their components are synthetic and toxic. To address both of these serious problems, we have developed a natural component-based oxygen indicator characterized by in-pack activation. The conventional oxygen indicator composed of synthetic and artificial components was redesigned using naturally occurring compounds (laccase, guaiacol, and cysteine). These natural components were physically separated into two compartments by a fragile barrier. Only when the barrier was broken were all of the components mixed and the function as an oxygen indicator was begun (i.e., in-pack activation). Depending on the component concentrations, the natural component-based oxygen indicator exhibited different response times and color differences. The rate of the color change was proportional to the oxygen concentration. This novel colorimetric oxygen indicator will contribute greatly to intelligent packaging for healthier and safer foods.

  11. Built-In Data-Flow Integration Testing in Large-Scale Component-Based Systems

    Science.gov (United States)

    Piel, Éric; Gonzalez-Sanchez, Alberto; Gross, Hans-Gerhard

    Modern large-scale component-based applications and service ecosystems are built following a number of different component models and architectural styles, such as the data-flow architectural style. In this style, each building block receives data from a previous one in the flow and sends output data to other components. This organisation expresses information flows adequately, and also favours decoupling between the components, leading to easier maintenance and quicker evolution of the system. Integration testing is a major means to ensure the quality of large systems. Their size and complexity, together with the fact that they are developed and maintained by several stake holders, make Built-In Testing (BIT) an attractive approach to manage their integration testing. However, so far no technique has been proposed that combines BIT and data-flow integration testing. We have introduced the notion of a virtual component in order to realize such a combination. It permits to define the behaviour of several components assembled to process a flow of data, using BIT. Test-cases are defined in a way that they are simple to write and flexible to adapt. We present two implementations of our proposed virtual component integration testing technique, and we extend our previous proposal to detect and handle errors in the definition by the user. The evaluation of the virtual component testing approach suggests that more issues can be detected in systems with data-flows than through other integration testing approaches.

  12. A System for Automatically Generating Scheduling Heuristics

    Science.gov (United States)

    Morris, Robert

    1996-01-01

    The goal of this research is to improve the performance of automated schedulers by designing and implementing an algorithm by automatically generating heuristics by selecting a schedule. The particular application selected by applying this method solves the problem of scheduling telescope observations, and is called the Associate Principal Astronomer. The input to the APA scheduler is a set of observation requests submitted by one or more astronomers. Each observation request specifies an observation program as well as scheduling constraints and preferences associated with the program. The scheduler employs greedy heuristic search to synthesize a schedule that satisfies all hard constraints of the domain and achieves a good score with respect to soft constraints expressed as an objective function established by an astronomer-user.

  13. Heuristic program to design Relational Databases

    Directory of Open Access Journals (Sweden)

    Manuel Pereira Rosa

    2009-09-01

    Full Text Available The great development of today’s world determines that the world level of information increases day after day, however, the time allowed to transmit this information in the classrooms has not changed. Thus, the rational work in this respect is more than necessary. Besides, if for the solution of a given type of problem we do not have a working algorism, we have, first to look for a correct solution, then the heuristic programs are of paramount importance to succeed in these aspects. Having into consideration that the design of the database is, essentially, a process of problem resolution, this article aims at proposing a heuristic program for the design of the relating database.

  14. [Design of traditional Chinese medicines with antihypertensive components based on medicinal property combination modes].

    Science.gov (United States)

    Liao, Su-Fen; Yan, Su-Rong; Guo, Wei-Jia; Luo, Ji; Sun, Jing; Dong, Fang; Wang, Yun; Qiao, Yan-Jiang

    2014-07-01

    Multi-component traditional Chinese medicines are an innovative research mode for traditional Chinese medicines. Currently, there are many design methods for developing multi-component traditional Chinese medicines, but their common feature is the lack of effective connection of the traditional Chinese medicine theory. In this paper, the authors discussed the multi-component traditional Chinese medicine design methods based on medicinal property combination modes, provided the combination methods with the characteristics of traditional Chinese medicine for the prescription combinations, and proved its feasibly with hypertension cases.

  15. Optimal test intervals of standby components based on actual plant-specific data

    International Nuclear Information System (INIS)

    Jones, R.B.; Bickel, J.H.

    1987-01-01

    Based on standard reliability analysis techniques, both under testing and over testing affect the availability of standby components. If tests are performed too often, unavailability is increased since the equipment is being used excessively. Conversely if testing is performed too infrequently, the likelihood of component unavailability is also increased due to the formation of rust, heat or radiation damage, dirt infiltration, etc. Thus from a physical perspective, an optimal test interval should exist which minimizes unavailability. This paper illustrates the application of an unavailability model that calculates optimal testing intervals for components with a failure database. (orig./HSCH)

  16. Multi-Step Deep Reactive Ion Etching Fabrication Process for Silicon-Based Terahertz Components

    Science.gov (United States)

    Jung-Kubiak, Cecile (Inventor); Reck, Theodore (Inventor); Chattopadhyay, Goutam (Inventor); Perez, Jose Vicente Siles (Inventor); Lin, Robert H. (Inventor); Mehdi, Imran (Inventor); Lee, Choonsup (Inventor); Cooper, Ken B. (Inventor); Peralta, Alejandro (Inventor)

    2016-01-01

    A multi-step silicon etching process has been developed to fabricate silicon-based terahertz (THz) waveguide components. This technique provides precise dimensional control across multiple etch depths with batch processing capabilities. Nonlinear and passive components such as mixers and multipliers waveguides, hybrids, OMTs and twists have been fabricated and integrated into a small silicon package. This fabrication technique enables a wafer-stacking architecture to provide ultra-compact multi-pixel receiver front-ends in the THz range.

  17. Design of multi-tiered database application based on CORBA component

    International Nuclear Information System (INIS)

    Sun Xiaoying; Dai Zhimin

    2003-01-01

    As computer technology quickly developing, middleware technology changed traditional two-tier database system. The multi-tiered database system, consisting of client application program, application servers and database serves, is mainly applying. While building multi-tiered database system using CORBA component has become the mainstream technique. In this paper, an example of DUV-FEL database system is presented, and then discuss the realization of multi-tiered database based on CORBA component. (authors)

  18. Design of multi-tiered database application based on CORBA component in SDUV-FEL system

    International Nuclear Information System (INIS)

    Sun Xiaoying; Shen Liren; Dai Zhimin

    2004-01-01

    The drawback of usual two-tiered database architecture was analyzed and the Shanghai Deep Ultraviolet-Free Electron Laser database system under development was discussed. A project for realizing the multi-tiered database architecture based on common object request broker architecture (CORBA) component and middleware model constructed by C++ was presented. A magnet database was given to exhibit the design of the CORBA component. (authors)

  19. Heuristic Evaluation for Novice Programming Systems

    OpenAIRE

    Kölling, Michael; McKay, Fraser

    2016-01-01

    The past few years has seen a proliferation of novice programming tools. The availability of a large number of systems has made it difficult for many users to choose among them. Even for education researchers, comparing the relative quality of these tools, or judging their respective suitability for a given context, is hard in many instances. For designers of such systems, assessing the respective quality of competing design decisions can be equally difficult.\\ud Heuristic evaluation provides...

  20. Heuristics for the economic dispatch problem

    Energy Technology Data Exchange (ETDEWEB)

    Flores, Benjamin Carpio [Centro Nacional de Controle de Energia (CENACE), Mexico, D.F. (Mexico). Dept. de Planificacion Economica de Largo Plazo], E-mail: benjamin.carpo@cfe.gob.mx; Laureano Cruces, A L; Lopez Bracho, R; Ramirez Rodriguez, J. [Universidad Autonoma Metropolitana (UAM), Mexico, D.F. (Brazil). Dept. de Sistemas], Emails: clc@correo.azc.uam.mx, rlb@correo.azc.uam.mx, jararo@correo.azc.uam.mx

    2009-07-01

    This paper presents GRASP (Greedy Randomized Adaptive Search Procedure), Simulated Annealing (SAA), Genetic (GA), and Hybrid Genetic (HGA) Algorithms for the economic dispatch problem (EDP), considering non-convex cost functions and dead zones the only restrictions, showing the results obtained. We also present parameter settings that are specifically applicable to the EDP, and a comparative table of results for each heuristic. It is shown that these methods outperform the classical methods without the need to assume convexity of the target function. (author)

  1. Entrepreneurial Learning, Heuristics and Venture Creation

    OpenAIRE

    RAUF, MIAN SHAMS; ZAINULLAH, MOHAMMAD

    2009-01-01

    After rigorous criticism on trait approach and with the emergence of behavioral approach in entrepreneurship during 1980s, the researchers started to introduce learning and cognitive theories in entrepreneurship to describe and explain the dynamic nature of entrepreneurship. Many researchers have described venture creation as a core and the single most important element of entrepreneurship. This thesis will discuss and present the role of entrepreneurial learning and heuristics in venture cre...

  2. A general heuristic for genome rearrangement problems.

    Science.gov (United States)

    Dias, Ulisses; Galvão, Gustavo Rodrigues; Lintzmayer, Carla Négri; Dias, Zanoni

    2014-06-01

    In this paper, we present a general heuristic for several problems in the genome rearrangement field. Our heuristic does not solve any problem directly, it is rather used to improve the solutions provided by any non-optimal algorithm that solve them. Therefore, we have implemented several algorithms described in the literature and several algorithms developed by ourselves. As a whole, we implemented 23 algorithms for 9 well known problems in the genome rearrangement field. A total of 13 algorithms were implemented for problems that use the notions of prefix and suffix operations. In addition, we worked on 5 algorithms for the classic problem of sorting by transposition and we conclude the experiments by presenting results for 3 approximation algorithms for the sorting by reversals and transpositions problem and 2 approximation algorithms for the sorting by reversals problem. Another algorithm with better approximation ratio can be found for the last genome rearrangement problem, but it is purely theoretical with no practical implementation. The algorithms we implemented in addition to our heuristic lead to the best practical results in each case. In particular, we were able to improve results on the sorting by transpositions problem, which is a very special case because many efforts have been made to generate algorithms with good results in practice and some of these algorithms provide results that equal the optimum solutions in many cases. Our source codes and benchmarks are freely available upon request from the authors so that it will be easier to compare new approaches against our results.

  3. The recognition heuristic: A decade of research

    Directory of Open Access Journals (Sweden)

    Gerd Gigerenzer

    2011-02-01

    Full Text Available The recognition heuristic exploits the basic psychological capacity for recognition in order to make inferences about unknown quantities in the world. In this article, we review and clarify issues that emerged from our initial work (Goldstein and Gigerenzer, 1999, 2002, including the distinction between a recognition and an evaluation process. There is now considerable evidence that (i the recognition heuristic predicts the inferences of a substantial proportion of individuals consistently, even in the presence of one or more contradicting cues, (ii people are adaptive decision makers in that accordance increases with larger recognition validity and decreases in situations when the validity is low or wholly indeterminable, and (iii in the presence of contradicting cues, some individuals appear to select different strategies. Little is known about these individual differences, or how to precisely model the alternative strategies. Although some researchers have attributed judgments inconsistent with the use of the recognition heuristic to compensatory processing, little research on such compensatory models has been reported. We discuss extensions of the recognition model, open questions, unanticipated results, and the surprising predictive power of recognition in forecasting.

  4. Storage costs and heuristics interact to produce patterns of aphasic sentence comprehension performance.

    Science.gov (United States)

    Clark, David Glenn

    2012-01-01

    Despite general agreement that aphasic individuals exhibit difficulty understanding complex sentences, the nature of sentence complexity itself is unresolved. In addition, aphasic individuals appear to make use of heuristic strategies for understanding sentences. This research is a comparison of predictions derived from two approaches to the quantification of sentence complexity, one based on the hierarchical structure of sentences, and the other based on dependency locality theory (DLT). Complexity metrics derived from these theories are evaluated under various assumptions of heuristic use. A set of complexity metrics was derived from each general theory of sentence complexity and paired with assumptions of heuristic use. Probability spaces were generated that summarized the possible patterns of performance across 16 different sentence structures. The maximum likelihood of comprehension scores of 42 aphasic individuals was then computed for each probability space and the expected scores from the best-fitting points in the space were recorded for comparison to the actual scores. Predictions were then compared using measures of fit quality derived from linear mixed effects models. All three of the metrics that provide the most consistently accurate predictions of patient scores rely on storage costs based on the DLT. Patients appear to employ an Agent-Theme heuristic, but vary in their tendency to accept heuristically generated interpretations. Furthermore, the ability to apply the heuristic may be degraded in proportion to aphasia severity. DLT-derived storage costs provide the best prediction of sentence comprehension patterns in aphasia. Because these costs are estimated by counting incomplete syntactic dependencies at each point in a sentence, this finding suggests that aphasia is associated with reduced availability of cognitive resources for maintaining these dependencies.

  5. Improving performances of suboptimal greedy iterative biclustering heuristics via localization.

    Science.gov (United States)

    Erten, Cesim; Sözdinler, Melih

    2010-10-15

    criteria. The fact that the random extraction method based on localization REAL performs better than the representative greedy heuristic methods under same criteria also confirms the effectiveness of the suggested pre-processing method. Supplementary material including code implementations in LEDA C++ library, experimental data, and the results are available at http://code.google.com/p/biclustering/ cesim@khas.edu.tr; melihsozdinler@boun.edu.tr Supplementary data are available at Bioinformatics online.

  6. Heuristic for Task-Worker Assignment with Varying Learning Slopes

    Directory of Open Access Journals (Sweden)

    Wipawee Tharmmaphornphilas

    2010-04-01

    Full Text Available Fashion industry has variety products, so the multi-skilled workers are required to improve flexibility in production and assignment. Generally the supervisor will assign task to the workers based on skill and skill levels of worker. Since in fashion industry new product styles are launched more frequently and the order size tends to be smaller, the workers always learn when the raw material and the production process changes. Consequently they require less time to produce the succeeding units of a task based on their learning ability. Since the workers have both experience and inexperience workers, so each worker has different skill level and learning ability. Consequently, the assignment which assumed constant skill level is not proper to use. This paper proposes a task-worker assignment considering worker skill levels and learning abilities. Processing time of each worker changes along production period due to a worker learning ability. We focus on a task-worker assignment in a fashion industry where tasks are ordered in series; the number of tasks is greater than the number of workers. Therefore, workers can perform multiple assignments followed the precedence restriction as an assembly line balancing problem. The problem is formulated in an integer linear programming model with objective to minimize makespan. A heuristic is proposed to determine the lower bound (LB and the upper bound (UB of the problem and the best assignment is determined. The performance of the heuristic method is tested by comparing quality of solution and computational time to optimal solutions.

  7. The normalization heuristic: an untested hypothesis that may misguide medical decisions.

    Science.gov (United States)

    Aberegg, Scott K; O'Brien, James M

    2009-06-01

    Medical practice is increasingly informed by the evidence from randomized controlled trials. When such evidence is not available, clinical hypotheses based on pathophysiological reasoning and common sense guide clinical decision making. One commonly utilized general clinical hypothesis is the assumption that normalizing abnormal laboratory values and physiological parameters will lead to improved patient outcomes. We refer to the general use of this clinical hypothesis to guide medical therapeutics as the "normalization heuristic". In this paper, we operationally define this heuristic and discuss its limitations as a rule of thumb for clinical decision making. We review historical and contemporaneous examples of normalization practices as empirical evidence for the normalization heuristic and to highlight its frailty as a guide for clinical decision making.

  8. A nuclear heuristic for application to metaheuristics in-core fuel management optimization

    Energy Technology Data Exchange (ETDEWEB)

    Meneses, Anderson Alvarenga de Moura, E-mail: ameneses@lmp.ufrj.b [COPPE/Federal University of Rio de Janeiro, RJ (Brazil). Nuclear Engineering Program; Dalle Molle Institute for Artificial Intelligence (IDSIA), Manno-Lugano, TI (Switzerland); Gambardella, Luca Maria, E-mail: luca@idsia.c [Dalle Molle Institute for Artificial Intelligence (IDSIA), Manno-Lugano, TI (Switzerland); Schirru, Roberto, E-mail: schirru@lmp.ufrj.b [COPPE/Federal University of Rio de Janeiro, RJ (Brazil). Nuclear Engineering Program

    2009-07-01

    The In-Core Fuel Management Optimization (ICFMO) is a well-known problem of nuclear engineering whose features are complexity, high number of feasible solutions, and a complex evaluation process with high computational cost, thus it is prohibitive to have a great number of evaluations during an optimization process. Heuristics are criteria or principles for deciding which among several alternative courses of action are more effective with respect to some goal. In this paper, we propose a new approach for the use of relational heuristics for the search in the ICFMO. The Heuristic is based on the reactivity of the fuel assemblies and their position into the reactor core. It was applied to random search, resulting in less computational effort concerning the number of evaluations of loading patterns during the search. The experiments demonstrate that it is possible to achieve results comparable to results in the literature, for future application to metaheuristics in the ICFMO. (author)

  9. A nuclear heuristic for application to metaheuristics in-core fuel management optimization

    International Nuclear Information System (INIS)

    Meneses, Anderson Alvarenga de Moura; Gambardella, Luca Maria; Schirru, Roberto

    2009-01-01

    The In-Core Fuel Management Optimization (ICFMO) is a well-known problem of nuclear engineering whose features are complexity, high number of feasible solutions, and a complex evaluation process with high computational cost, thus it is prohibitive to have a great number of evaluations during an optimization process. Heuristics are criteria or principles for deciding which among several alternative courses of action are more effective with respect to some goal. In this paper, we propose a new approach for the use of relational heuristics for the search in the ICFMO. The Heuristic is based on the reactivity of the fuel assemblies and their position into the reactor core. It was applied to random search, resulting in less computational effort concerning the number of evaluations of loading patterns during the search. The experiments demonstrate that it is possible to achieve results comparable to results in the literature, for future application to metaheuristics in the ICFMO. (author)

  10. Constraint satisfaction adaptive neural network and heuristics combined approaches for generalized job-shop scheduling.

    Science.gov (United States)

    Yang, S; Wang, D

    2000-01-01

    This paper presents a constraint satisfaction adaptive neural network, together with several heuristics, to solve the generalized job-shop scheduling problem, one of NP-complete constraint satisfaction problems. The proposed neural network can be easily constructed and can adaptively adjust its weights of connections and biases of units based on the sequence and resource constraints of the job-shop scheduling problem during its processing. Several heuristics that can be combined with the neural network are also presented. In the combined approaches, the neural network is used to obtain feasible solutions, the heuristic algorithms are used to improve the performance of the neural network and the quality of the obtained solutions. Simulations have shown that the proposed neural network and its combined approaches are efficient with respect to the quality of solutions and the solving speed.

  11. "The Gaze Heuristic:" Biography of an Adaptively Rational Decision Process.

    Science.gov (United States)

    Hamlin, Robert P

    2017-04-01

    This article is a case study that describes the natural and human history of the gaze heuristic. The gaze heuristic is an interception heuristic that utilizes a single input (deviation from a constant angle of approach) repeatedly as a task is performed. Its architecture, advantages, and limitations are described in detail. A history of the gaze heuristic is then presented. In natural history, the gaze heuristic is the only known technique used by predators to intercept prey. In human history the gaze heuristic was discovered accidentally by Royal Air Force (RAF) fighter command just prior to World War II. As it was never discovered by the Luftwaffe, the technique conferred a decisive advantage upon the RAF throughout the war. After the end of the war in America, German technology was combined with the British heuristic to create the Sidewinder AIM9 missile, the most successful autonomous weapon ever built. There are no plans to withdraw it or replace its guiding gaze heuristic. The case study demonstrates that the gaze heuristic is a specific heuristic type that takes a single best input at the best time (take the best 2 ). Its use is an adaptively rational response to specific, rapidly evolving decision environments that has allowed those animals/humans/machines who use it to survive, prosper, and multiply relative to those who do not. Copyright © 2017 Cognitive Science Society, Inc.

  12. Feature selection for neural network based defect classification of ceramic components using high frequency ultrasound.

    Science.gov (United States)

    Kesharaju, Manasa; Nagarajah, Romesh

    2015-09-01

    The motivation for this research stems from a need for providing a non-destructive testing method capable of detecting and locating any defects and microstructural variations within armour ceramic components before issuing them to the soldiers who rely on them for their survival. The development of an automated ultrasonic inspection based classification system would make possible the checking of each ceramic component and immediately alert the operator about the presence of defects. Generally, in many classification problems a choice of features or dimensionality reduction is significant and simultaneously very difficult, as a substantial computational effort is required to evaluate possible feature subsets. In this research, a combination of artificial neural networks and genetic algorithms are used to optimize the feature subset used in classification of various defects in reaction-sintered silicon carbide ceramic components. Initially wavelet based feature extraction is implemented from the region of interest. An Artificial Neural Network classifier is employed to evaluate the performance of these features. Genetic Algorithm based feature selection is performed. Principal Component Analysis is a popular technique used for feature selection and is compared with the genetic algorithm based technique in terms of classification accuracy and selection of optimal number of features. The experimental results confirm that features identified by Principal Component Analysis lead to improved performance in terms of classification percentage with 96% than Genetic algorithm with 94%. Copyright © 2015 Elsevier B.V. All rights reserved.

  13. Exact and heuristic solutions to the Double TSP with Multiple Stacks

    DEFF Research Database (Denmark)

    Petersen, Hanne Løhmann; Archetti, Claudia; Madsen, Oli B.G.

    -pallet, which can be loaded in 3 stacks in a standard 40 foot container. Different exact and heuristic solution approaches to the DTSPMS have been implemented and tested. The exact approaches are based on different mathematical formulations of the problem which are solved using branch-and-cut. One formulation...... instances. The implemented heuristics include tabu search, simulated annealing and large neighbourhood search. Particularly the LNS approach shows promising results. It finds the known optimal solution of smaller instances (15 orders) within 10 seconds in most cases, and in 3 minutes it finds solutions...

  14. Wavelet decomposition based principal component analysis for face recognition using MATLAB

    Science.gov (United States)

    Sharma, Mahesh Kumar; Sharma, Shashikant; Leeprechanon, Nopbhorn; Ranjan, Aashish

    2016-03-01

    For the realization of face recognition systems in the static as well as in the real time frame, algorithms such as principal component analysis, independent component analysis, linear discriminate analysis, neural networks and genetic algorithms are used for decades. This paper discusses an approach which is a wavelet decomposition based principal component analysis for face recognition. Principal component analysis is chosen over other algorithms due to its relative simplicity, efficiency, and robustness features. The term face recognition stands for identifying a person from his facial gestures and having resemblance with factor analysis in some sense, i.e. extraction of the principal component of an image. Principal component analysis is subjected to some drawbacks, mainly the poor discriminatory power and the large computational load in finding eigenvectors, in particular. These drawbacks can be greatly reduced by combining both wavelet transform decomposition for feature extraction and principal component analysis for pattern representation and classification together, by analyzing the facial gestures into space and time domain, where, frequency and time are used interchangeably. From the experimental results, it is envisaged that this face recognition method has made a significant percentage improvement in recognition rate as well as having a better computational efficiency.

  15. Biochemical component identification by plasmonic improved whispering gallery mode optical resonance based sensor

    Science.gov (United States)

    Saetchnikov, Vladimir A.; Tcherniavskaia, Elina A.; Saetchnikov, Anton V.; Schweiger, Gustav; Ostendorf, Andreas

    2014-05-01

    Experimental data on detection and identification of variety of biochemical agents, such as proteins, microelements, antibiotic of different generation etc. in both single and multi component solutions under varied in wide range concentration analyzed on the light scattering parameters of whispering gallery mode optical resonance based sensor are represented. Multiplexing on parameters and components has been realized using developed fluidic sensor cell with fixed in adhesive layer dielectric microspheres and data processing. Biochemical component identification has been performed by developed network analysis techniques. Developed approach is demonstrated to be applicable both for single agent and for multi component biochemical analysis. Novel technique based on optical resonance on microring structures, plasmon resonance and identification tools has been developed. To improve a sensitivity of microring structures microspheres fixed by adhesive had been treated previously by gold nanoparticle solution. Another technique used thin film gold layers deposited on the substrate below adhesive. Both biomolecule and nanoparticle injections caused considerable changes of optical resonance spectra. Plasmonic gold layers under optimized thickness also improve parameters of optical resonance spectra. Biochemical component identification has been also performed by developed network analysis techniques both for single and for multi component solution. So advantages of plasmon enhancing optical microcavity resonance with multiparameter identification tools is used for development of a new platform for ultra sensitive label-free biomedical sensor.

  16. A comparative analysis of meta-heuristic methods for power management of a dual energy storage system for electric vehicles

    International Nuclear Information System (INIS)

    Trovão, João P.; Antunes, Carlos Henggeler

    2015-01-01

    Highlights: • Two meta-heuristic approaches are evaluated for multi-ESS management in electric vehicles. • An online global energy management strategy with two different layers is studied. • Meta-heuristic techniques are used to define optimized energy sharing mechanisms. • A comparative analysis for ARTEMIS driving cycle is addressed. • The effectiveness of the double-layer management with meta-heuristic is presented. - Abstract: This work is focused on the performance evaluation of two meta-heuristic approaches, simulated annealing and particle swarm optimization, to deal with power management of a dual energy storage system for electric vehicles. The proposed strategy is based on a global energy management system with two layers: long-term (energy) and short-term (power) management. A rule-based system deals with the long-term (strategic) layer and for the short-term (action) layer meta-heuristic techniques are developed to define optimized online energy sharing mechanisms. Simulations have been made for several driving cycles to validate the proposed strategy. A comparative analysis for ARTEMIS driving cycle is presented evaluating three performance indicators (computation time, final value of battery state of charge, and minimum value of supercapacitors state of charge) as a function of input parameters. The results show the effectiveness of an implementation based on a double-layer management system using meta-heuristic methods for online power management supported by a rule set that restricts the search space

  17. Comparing the performance of expert user heuristics and an integer linear program in aircraft carrier deck operations.

    Science.gov (United States)

    Ryan, Jason C; Banerjee, Ashis Gopal; Cummings, Mary L; Roy, Nicholas

    2014-06-01

    Planning operations across a number of domains can be considered as resource allocation problems with timing constraints. An unexplored instance of such a problem domain is the aircraft carrier flight deck, where, in current operations, replanning is done without the aid of any computerized decision support. Rather, veteran operators employ a set of experience-based heuristics to quickly generate new operating schedules. These expert user heuristics are neither codified nor evaluated by the United States Navy; they have grown solely from the convergent experiences of supervisory staff. As unmanned aerial vehicles (UAVs) are introduced in the aircraft carrier domain, these heuristics may require alterations due to differing capabilities. The inclusion of UAVs also allows for new opportunities for on-line planning and control, providing an alternative to the current heuristic-based replanning methodology. To investigate these issues formally, we have developed a decision support system for flight deck operations that utilizes a conventional integer linear program-based planning algorithm. In this system, a human operator sets both the goals and constraints for the algorithm, which then returns a proposed schedule for operator approval. As a part of validating this system, the performance of this collaborative human-automation planner was compared with that of the expert user heuristics over a set of test scenarios. The resulting analysis shows that human heuristics often outperform the plans produced by an optimization algorithm, but are also often more conservative.

  18. Two efficient heuristics to solve the integrated load distribution and production planning problem

    International Nuclear Information System (INIS)

    Gajpal, Yuvraj; Nourelfath, Mustapha

    2015-01-01

    This paper considers a multi-period production system where a set of machines are arranged in parallel. The machines are unreliable and the failure rate of machine depends on the load assigned to the machine. The expected production rate of the system is considered to be a non-monotonic function of its load. Because of the machine failure rate, the total production output depends on the combination of loads assigned to different machines. We consider the integration of load distribution decisions with production planning decision. The product demands are considered to be known in advance. The objective is to minimize the sum of holding costs, backorder costs, production costs, setup costs, capacity change costs and unused capacity costs while satisfying the demand over specified time horizon. The constraint is not to exceed available repair resources required to repair the machine breakdown. The paper develops two heuristics to solve the integrated load distribution and production planning problem. The first heuristic consists of a three-phase approach, while the second one is based on tabu search metaheuristic. The efficiency of the proposed heuristics is tested through the randomly generated problem instances. - Highlights: • The expected performance of the system is a non-monotonic function of its load. • We consider the integration of load distribution and production planning decisions. • The paper proposes three phase and tabu search based heuristics to solve the problem. • Lower bound has been developed for checking the effectiveness of the heuristics. • The efficiency of the heuristic is tested through randomly generated instances.

  19. Component-Based Development of Runtime Observers in the COMDES Framework

    DEFF Research Database (Denmark)

    Guan, Wei; Li, Gang; Angelov, Christo K.

    2013-01-01

    against formally specified properties. This paper presents a component-based design method for runtime observers in the context of COMDES framework—a component-based framework for distributed embedded system and its supporting tools. Therefore, runtime verification is facilitated by model......Formal verification methods, such as exhaustive model checking, are often infeasible because of high computational complexity. Runtime observers (monitors) provide an alternative, light-weight verification method, which offers a non-exhaustive but still feasible approach to monitor system behavior...

  20. Prognostic Health Monitoring System: Component Selection Based on Risk Criteria and Economic Benefit Assessment

    International Nuclear Information System (INIS)

    Pham, Binh T.; Agarwal, Vivek; Lybeck, Nancy J.; Tawfik, Magdy S.

    2012-01-01

    Prognostic health monitoring (PHM) is a proactive approach to monitor the ability of structures, systems, and components (SSCs) to withstand structural, thermal, and chemical loadings over the SSCs planned service lifespan. The current efforts to extend the operational license lifetime of the aging fleet of U.S. nuclear power plants from 40 to 60 years and beyond can benefit from a systematic application of PHM technology. Implementing a PHM system would strengthen the safety of nuclear power plants, reduce plant outage time, and reduce operation and maintenance costs. However, a nuclear power plant has thousands of SSCs, so implementing a PHM system that covers all SSCs requires careful planning and prioritization. This paper therefore focuses on a component selection that is based on the analysis of a component's failure probability, risk, and cost. Ultimately, the decision on component selection depends on the overall economical benefits arising from safety and operational considerations associated with implementing the PHM system. (author)