Directory of Open Access Journals (Sweden)
X. Wang
2012-10-01
Full Text Available To make first-order estimates of the probability of moraine-dammed lake outburst flood (MDLOF and prioritize the probabilities of breaching posed by potentially dangerous moraine-dammed lakes (PDMDLs in the Chinese Himalayas, an objective approach is presented. We first select five indicators to identify PDMDLs according to four predesigned criteria. The climatic background was regarded as the climatic precondition of the moraine-dam failure, and under different climatic preconditions, we distinguish the trigger mechanisms of MDLOFs and subdivide them into 17 possible breach modes, with each mode having three or four components; we combined the precondition, modes and components to construct a decision-making tree of moraine-dam failure. Conversion guidelines were established so as to quantify the probabilities of components of a breach mode employing the historic performance method combined with expert knowledge and experience. The region of the Chinese Himalayas was chosen as a study area where there have been frequent MDLOFs in recent decades. The results show that the breaching probabilities (P of 142 PDMDLs range from 0.037 to 0.345, and they can be further categorized as 43 lakes with very high breach probabilities (P ≥ 0.24, 47 lakes with high breach probabilities (0.18 ≤ P < 0.24, 24 lakes with mid-level breach probabilities (0.12 ≤ P < 0.18, 24 lakes with low breach probabilities (0.06 ≤ P < 0.12, and four lakes with very low breach probabilities (p < 0.06.
2014-06-30
precisely the content of the following result. The price we pay is that the assumption that A is a packing in (F, k ·k1) is too weak to make this happen...Regularité des trajectoires des fonctions aléatoires gaussiennes. In: École d’Été de Probabilités de Saint- Flour , IV-1974, pp. 1–96. Lecture Notes in...Lectures on probability theory and statistics (Saint- Flour , 1994), Lecture Notes in Math., vol. 1648, pp. 165–294. Springer, Berlin (1996) 50. Ledoux
A simplified physically-based breach model for a high concrete-faced rockfill dam: A case study
Qi-ming Zhong; Sheng-shui Chen; Zhao Deng
2018-01-01
A simplified physically-based model was developed to simulate the breaching process of the Gouhou concrete-faced rockfill dam (CFRD), which is the only breach case of a high CFRD in the world. Considering the dam height, a hydraulic method was chosen to simulate the initial scour position on the downstream slope, with the steepening of the downstream slope taken into account; a headcut erosion formula was adopted to simulate the backward erosion as well. The moment equilibrium method was util...
Energy Technology Data Exchange (ETDEWEB)
Belangia, David Warren [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2015-03-13
The Home Depot Data Breach is the second largest data breach on record. It has or will affect up to 56 million debit or credit cards. A trusted vendor account, coupled with the use of a previously unknown variant of malware that allowed the establishment of a foothold, was the entry point into the Home Depot network. Once inside the perimeter, privilege escalation provided an avenue to obtain the desired information. Home Depot did, however, learn some lessons from Target. Home Depot certainly communicated better than Target, procured insurance, and instituted as secure an environment as possible. There are specific measures an institution should undertake to prepare for a data breach, and everyone can learn from this breach. Publicly available information about the Home Depot Data Breach provides insight into the attack, an old malware variant with a new twist.While the malware was modified as to be unrecognizable with tools, it probably should have been detected. There are also concerns with Home Depot’s insurance and the insurance provider’s apparent lack of fully reimbursing Home Depot for their losses. The effect on shareholders and Home Depot’s stock price was short lived. This story is still evolving but provides interesting lessons learned concerning how an organization should prepare for it inevitable breach.
Gamble, Kate Huvane
2009-07-01
* The HITECH section of ARRA includes provisions relating to protected health information that could significantly alter the C-suite leader's strategy. * Patients will be entitled to request an accounting of disclosure for up to three years after the date of request. The onus will be on hospital leaders to put in place a process that makes accounting available without disrupting operations or patient care. * Because of the increased risks hospitals now face, it is critical that executives are aware of the new requirements, and are either involved in or have a solid understanding of the organization's breach notification policies.
High throughput nonparametric probability density estimation.
Farmer, Jenny; Jacobs, Donald
2018-01-01
In high throughput applications, such as those found in bioinformatics and finance, it is important to determine accurate probability distribution functions despite only minimal information about data characteristics, and without using human subjectivity. Such an automated process for univariate data is implemented to achieve this goal by merging the maximum entropy method with single order statistics and maximum likelihood. The only required properties of the random variables are that they are continuous and that they are, or can be approximated as, independent and identically distributed. A quasi-log-likelihood function based on single order statistics for sampled uniform random data is used to empirically construct a sample size invariant universal scoring function. Then a probability density estimate is determined by iteratively improving trial cumulative distribution functions, where better estimates are quantified by the scoring function that identifies atypical fluctuations. This criterion resists under and over fitting data as an alternative to employing the Bayesian or Akaike information criterion. Multiple estimates for the probability density reflect uncertainties due to statistical fluctuations in random samples. Scaled quantile residual plots are also introduced as an effective diagnostic to visualize the quality of the estimated probability densities. Benchmark tests show that estimates for the probability density function (PDF) converge to the true PDF as sample size increases on particularly difficult test probability densities that include cases with discontinuities, multi-resolution scales, heavy tails, and singularities. These results indicate the method has general applicability for high throughput statistical inference.
Intermittent ephemeral river-breaching
Reniers, A. J.; MacMahan, J. H.; Gallagher, E. L.; Shanks, A.; Morgan, S.; Jarvis, M.; Thornton, E. B.; Brown, J.; Fujimura, A.
2012-12-01
In the summer of 2011 we performed a field experiment in Carmel River State Beach, CA, at a time when the intermittent natural breaching of the ephemeral Carmel River occurred due to an unusually rainy period prior to the experiment associated with El Nino. At this time the river would fill the lagoon over the period of a number of days after which a breach would occur. This allowed us to document a number of breaches with unique pre- and post-breach topographic surveys, accompanying ocean and lagoon water elevations as well as extremely high flow (4m/s) velocities in the river mouth during the breaching event. The topographic surveys were obtained with a GPS-equipped backpack mounted on a walking human and show the evolution of the river breaching with a gradually widening and deepening river channel that cuts through the pre-existing beach and berm. The beach face is qualified as a steep with an average beach slope of 1:10 with significant reflection of the incident waves (MacMahan et al., 2012). The wave directions are generally shore normal as the waves refract over the deep canyon that is located offshore of the beach. The tide is mixed semi-diurnal with a range on the order of one meter. Breaching typically occurred during the low-low tide. Grain size is highly variable along the beach with layers of alternating fine and coarse material that could clearly be observed as the river exit channel was cutting through the beach. Large rocky outcroppings buried under the beach sand are also present along certain stretches of the beach controlling the depth of the breaching channel. The changes in the water level measured within the lagoon and the ocean side allows for an estimate of the volume flux associated with the breach as function of morphology, tidal elevation and wave conditions as well as an assessment of the conditions and mechanisms of breach closure, which occurred on the time scale of O(0.5 days). Exploratory model simulations will be presented at the
Psychological contract breach among allied health professionals.
Rodwell, John; Gulyas, Andre
2015-01-01
Allied health professionals are vital for effective healthcare yet there are continuing shortages of these employees. Building on work with other healthcare professionals, the purpose of this paper is to investigate the influence of psychological contract (PC) breach and types of organisational justice on variables important to retention among allied health professionals: mental health and organisational commitment. The potential effects of justice on the negative outcomes of breach were examined. Multiple regressions analysed data from 113 allied health professionals working in a medium-large Australian healthcare organisation. The main negative impacts on respondents' mental health and commitment were from high PC breach, low procedural and distributive justice and less respectful treatment from organisational representatives. The interaction between procedural justice and breach illustrates that breach may be forgivable if processes are fair. Surprisingly, a betrayal or "aggravated breach effect" may occur after a breach when interpersonal justice is high. Further, negative affectivity was negatively related to respondents' mental health (affective outcomes) but not commitment (work-related attitude). Healthcare organisations should ensure the fairness of decisions and avoid breaking promises within their control. If promises cannot reasonably be kept, transparency of processes behind the breach may allow allied health professionals to understand that the organisation did not purposefully fail to fulfil expectations. This study offers insights into how breach and four types of justice interact to influence employee mental health and work attitudes among allied health professionals.
Shiryaev, A N
1996-01-01
This book contains a systematic treatment of probability from the ground up, starting with intuitive ideas and gradually developing more sophisticated subjects, such as random walks, martingales, Markov chains, ergodic theory, weak convergence of probability measures, stationary stochastic processes, and the Kalman-Bucy filter Many examples are discussed in detail, and there are a large number of exercises The book is accessible to advanced undergraduates and can be used as a text for self-study This new edition contains substantial revisions and updated references The reader will find a deeper study of topics such as the distance between probability measures, metrization of weak convergence, and contiguity of probability measures Proofs for a number of some important results which were merely stated in the first edition have been added The author included new material on the probability of large deviations, and on the central limit theorem for sums of dependent random variables
Simulation of Breach Outflow for Earthfill Dam
International Nuclear Information System (INIS)
Razad, Azwin Zailti Abdul; Muda, Rahsidi Sabri; Sidek, Lariyah Mohd; Azia, Intan Shafilah Abdul; Mansor, Faezah Hanum; Yalit, Ruzaimei
2013-01-01
Dams have been built for many reasons such as irrigation, hydropower, flood mitigation, and water supply to support development for the benefit of human. However, the huge amount of water stored behind the dam can seriously pose adverse impacts to the downstream community should it be released due to unwanted dam break event. To minimise the potential loss of lives and property damages, a workable Emergency Response Plan is required to be developed. As part of a responsible dam owner and operator, TNB initiated a study on dam breach modelling for Cameron Highlands Hydroelectric Scheme to simulate the potential dam breach for Jor Dam. Prediction of dam breach parameters using the empirical equations of Froehlich and Macdonal-Langridge-Monopolis formed the basis of the modelling, coupled with MIKE 11 software to obtain the breach outflow due to Probable Maximum Flood (PMF). This paper will therefore discuss the model setup, simulation procedure and comparison of the prediction with existing equations.
2008-04-11
This document adopts, without change, the interim final rule that was published in the Federal Register on June 22, 2007, addressing data breaches of sensitive personal information that is processed or maintained by the Department of Veterans Affairs (VA). This final rule implements certain provisions of the Veterans Benefits, Health Care, and Information Technology Act of 2006. The regulations prescribe the mechanisms for taking action in response to a data breach of sensitive personal information.
International Nuclear Information System (INIS)
Walton, Z.P.; Kam, J.T.
2002-01-01
The Engineered Barrier System (EBS) represents the system of human engineered barriers in the isolation of high-level radioactive waste in the proposed repository at Yucca Mountain. It is designed to complement and enhance the natural barriers to isolate and prevent the transport of radionuclides into the surrounding environment. The transport mechanism most frequently postulated for radionuclides is liquid water flux that has penetrated the EBS through corrosion breaches in the drip shield and waste packages (WP). A water flux-splitting model is used to predict flow through WP and drip shield breaches and is documented in the ''EBS Radionuclide Transport Abstraction''. A future revision of the ''EBS Radionuclide Transport Abstraction'' will be one component of the total system performance assessment--license application (TSPA-LA) for the Yucca Mountain repository. The flux-splitting model is conservative based on the following assumptions: (1) Drip impact occurs without a loss of water mass. (2) Dripping flux falls exactly at the crown of the drip shield as opposed to different locations on the curved surface, which will effect splashing and flow patterns. (3) The flux passing through a drip shield patch is proportional to the ratio of the length of the penetration in the axial direction to the total axial length of the drip shield. In this assumption all fluid that drips and flows from the drip shield crown toward a penetration will be collected if the axial locations of the source and patch coincide. (4) The potential for evaporation is ignored. Because of these conservatisms, the current version of the flux-splitting model is incapable of accounting for water that has been splashed from the impact location, the deviation of water paths (rivulets) from the axis of impact, and water loss due to evaporation. This paper will present the results of a series of breached drip shield tests used to collect empirical data for the initial validation and further
Tapping Transaction Costs to Forecast Acquisition Cost Breaches
2016-01-01
experience a cost breach. In our medical example, we could use survival analysis to identify risk fac- tors, such as obesity , that might indicate a greater... exogenous variables on the probability of a dichotomous outcome, such as whether or not a cost breach occurs in any given program year. Logit is
Schaffhauser, Dian
2009-01-01
Will that data breach be the end of a chief security officer (CSO)? Managing information security in higher education requires more than just technical expertise, especially when the heat is cranked up. This article takes a look at how two CSOs deal with hack attacks at their universities. When Purdue University Chief Information Security Officer…
Experimental investigation of fluvial dike breaching due to flow overtopping
El Kadi Abderrezzak, K.; Rifai, I.; Erpicum, S.; Archambeau, P.; Violeau, D.; Pirotton, M.; Dewals, B.
2017-12-01
breaching. These specific features need to be incorporated in flood risk analyses involving fluvial dike breach and failure. In addition, a well-documented, reliable data set, with a continuous high resolution monitoring of the 3D breach evolution under various flow conditions, has been gathered, which can be used for validating numerical models.
Breached cylinder incident at the Portsmouth gaseous diffusion plant
Energy Technology Data Exchange (ETDEWEB)
Boelens, R.A. [Martin Marietta Energy Systems, Inc., Piketon, OH (United States)
1991-12-31
On June 16, 1990, during an inspection of valves on partially depleted product storage cylinders, a 14-ton partially depleted product cylinder was discovered breached. The cylinder had been placed in long-term storage in 1977 on the top row of Portsmouth`s (two rows high) storage area. The breach was observed when an inspector noticed a pile of green material along side of the cylinder. The breach was estimated to be approximately 8- inches wide and 16-inches long, and ran under the first stiffening ring of the cylinder. During the continuing inspection of the storage area, a second 14-ton product cylinder was discovered breached. This cylinder was stacked on the bottom row in the storage area in 1986. This breach was also located adjacent to a stiffening ring. This paper will discuss the contributing factors of the breaching of the cylinders, the immediate response, subsequent actions in support of the investigation, and corrective actions.
Focus in High School Mathematics: Statistics and Probability
National Council of Teachers of Mathematics, 2009
2009-01-01
Reasoning about and making sense of statistics and probability are essential to students' future success. This volume belongs to a series that supports National Council of Teachers of Mathematics' (NCTM's) "Focus in High School Mathematics: Reasoning and Sense Making" by providing additional guidance for making reasoning and sense making part of…
2010-07-01
... 38 Pensions, Bonuses, and Veterans' Relief 2 2010-07-01 2010-07-01 false Data breach. 75.113 Section 75.113 Pensions, Bonuses, and Veterans' Relief DEPARTMENT OF VETERANS AFFAIRS (CONTINUED) INFORMATION SECURITY MATTERS Data Breaches § 75.113 Data breach. Consistent with the definition of data breach in § 75.112 of this subpart, a data breach...
Seasonal breaching of coastal barriers
Tuan, Thieu Quang
2007-01-01
Natural or unintended breaching can be catastrophic, causing loss of human lives and damage to infrastructures, buildings and natural habitats. Quantitative understand-ing of coastal barrier breaching is therefore of great importance to vulnerability as-sessment of protection works as well as to
Domestic wells have high probability of pumping septic tank leachate
Bremer, J. E.; Harter, T.
2012-08-01
Onsite wastewater treatment systems are common in rural and semi-rural areas around the world; in the US, about 25-30% of households are served by a septic (onsite) wastewater treatment system, and many property owners also operate their own domestic well nearby. Site-specific conditions and local groundwater flow are often ignored when installing septic systems and wells. In areas with small lots (thus high spatial septic system densities), shallow domestic wells are prone to contamination by septic system leachate. Mass balance approaches have been used to determine a maximum septic system density that would prevent contamination of groundwater resources. In this study, a source area model based on detailed groundwater flow and transport modeling is applied for a stochastic analysis of domestic well contamination by septic leachate. Specifically, we determine the probability that a source area overlaps with a septic system drainfield as a function of aquifer properties, septic system density and drainfield size. We show that high spatial septic system density poses a high probability of pumping septic system leachate. The hydraulic conductivity of the aquifer has a strong influence on the intersection probability. We find that mass balance calculations applied on a regional scale underestimate the contamination risk of individual drinking water wells by septic systems. This is particularly relevant for contaminants released at high concentrations, for substances that experience limited attenuation, and those that are harmful even at low concentrations (e.g., pathogens).
Domestic wells have high probability of pumping septic tank leachate
Directory of Open Access Journals (Sweden)
J. E. Bremer
2012-08-01
Full Text Available Onsite wastewater treatment systems are common in rural and semi-rural areas around the world; in the US, about 25–30% of households are served by a septic (onsite wastewater treatment system, and many property owners also operate their own domestic well nearby. Site-specific conditions and local groundwater flow are often ignored when installing septic systems and wells. In areas with small lots (thus high spatial septic system densities, shallow domestic wells are prone to contamination by septic system leachate. Mass balance approaches have been used to determine a maximum septic system density that would prevent contamination of groundwater resources. In this study, a source area model based on detailed groundwater flow and transport modeling is applied for a stochastic analysis of domestic well contamination by septic leachate. Specifically, we determine the probability that a source area overlaps with a septic system drainfield as a function of aquifer properties, septic system density and drainfield size. We show that high spatial septic system density poses a high probability of pumping septic system leachate. The hydraulic conductivity of the aquifer has a strong influence on the intersection probability. We find that mass balance calculations applied on a regional scale underestimate the contamination risk of individual drinking water wells by septic systems. This is particularly relevant for contaminants released at high concentrations, for substances that experience limited attenuation, and those that are harmful even at low concentrations (e.g., pathogens.
Williams, S. Jeffress; Foley, Mary K.
2007-01-01
-control stabilization of the headlandds such as the Montauk Point headlands, and deepening of navigation channels by dredging through the tidal inlets and in the bays. Indirect impacts that have a bearing on decisions to deal with breaching are: high-risk development of the barrier islands and low-lying areas of the mainland vulnerable to flooding, and the dredging of nearshore sand shoals for beach nourishment. The NPS strives to employ a coastal management framework for decision making that is based on assessment of the physical and ecological properties of the shoreline as well as human welfare and property. In order to protect developed areas of Fire Island and the mainland from loss of life, flooding, and other economic and physical damage, the NPS will likely need to consider allowing artificial closure of some breaches within the FIIS under certain circumstances. The decision by the NPS to allow breaches to evolve naturally and possibly close or to allow artificially closing breaches is based on four criteria: 1. Volumes of sediment transported landward and exchange of water and nutrients;
High probability of disease in angina pectoris patients
DEFF Research Database (Denmark)
Høilund-Carlsen, Poul F.; Johansen, Allan; Vach, Werner
2007-01-01
BACKGROUND: According to most current guidelines, stable angina pectoris patients with a high probability of having coronary artery disease can be reliably identified clinically. OBJECTIVES: To examine the reliability of clinical evaluation with or without an at-rest electrocardiogram (ECG......) in patients with a high probability of coronary artery disease. PATIENTS AND METHODS: A prospective series of 357 patients referred for coronary angiography (CA) for suspected stable angina pectoris were examined by a trained physician who judged their type of pain and Canadian Cardiovascular Society grade...... on CA. Of the patients who had also an abnormal at-rest ECG, 14% to 21% of men and 42% to 57% of women had normal MPS. Sex-related differences were statistically significant. CONCLUSIONS: Clinical prediction appears to be unreliable. Addition of at-rest ECG data results in some improvement, particularly...
Applying Mechanistic Dam Breach Models to Historic Levee Breaches
Risher Paul; Gibson Stanford
2016-01-01
Hurricane Katrina elevated levee risk in the US national consciousness, motivating agencies to assess and improve their levee risk assessment methodology. Accurate computation of the flood flow magnitude and timing associated with a levee breach remains one of the most difficult and uncertain components of levee risk analysis. Contemporary methods are largely empirical and approximate, introducing substantial uncertainty to the damage and life loss models. Levee breach progressions are often ...
Probabilistic assessment of spent-fuel cladding breach
International Nuclear Information System (INIS)
Foadian, H.; Rashid, Y.R.; Seager, K.D.
1991-01-01
A methodology for determining the probability spent-fuel cladding breach due to normal and accident class B cask transport conditions is introduced. This technique uses deterministic stress analysis results as well as probabilistic cladding material properties, initial flaws, and breach criteria. Best estimates are presented for the probability distributions of irradiated Zircaloy properties such as ductility and fracture toughness, and for fuel rod initial conditions such as manufacturing flaws and PCI part-wall cracks. Example analyses are used to illustrate the implementation of this methodology for a BWR (GE 7 x 7) and a PWR (B ampersand W 15 x 15) assembly. The cladding breach probabilities for each assembly are tabulated for regulatory normal and accident transport conditions including fire
Probabilistic assessment of spent-fuel cladding breach
International Nuclear Information System (INIS)
Foadian, H.; Rashid, Y.R.; Seager, K.D.
1992-01-01
In this paper a methodology for determining the probability of spent-fuel cladding breach due to normal and accident class B cask transport conditions is introduced. This technique uses deterministic stress analysis results as well as probabilistic cladding material properties, initial flaws, and breach criteria. Best estimates are presented for the probability distributions of irradiated Zircaloy properties such as ductility and fracture toughness, and for fuel rod initial conditions such as manufacturing flaws and PCI part-wall cracks. Example analyses are used to illustrate the implementation of this methodology for a BWR (GE 7 x 7) and a PWR (B and W 15 x 15) assembly. The cladding breach probabilities for each assembly are tabulated for regulatory normal and accident transport conditions including fire
Shivers, Molly J.; Smith, S. Jerrod; Grout, Trevor S.; Lewis, Jason M.
2015-01-01
Dams provide beneficial functions such as flood control, recreation, and storage of water supplies, but they also entail risk; dam breaches and resultant floods can cause substantial property damage and loss of life. The State of Oklahoma requires each owner of a high-hazard dam, which the Federal Emergency Management Agency defines as dams for which failure or improper operation probably will cause loss of human life, to develop an emergency action plan specific to that dam. Components of an emergency action plan are to simulate a flood resulting from a possible dam breach and map the resulting downstream flood-inundation areas. The resulting flood-inundation maps can provide valuable information to city officials, emergency managers, and local residents for planning an emergency response if a dam breach occurs.
Dam Break Analysis of Embankment Dams Considering Breach Characteristics
Directory of Open Access Journals (Sweden)
Abolfazl Shamsaei
2004-05-01
Full Text Available The study of dam's break, needs the definition of various parameters such as the break cause, its type, its dimension and the duration of breach development. The precise forecast for different aspects of the breach is one of the most important factors for analyzing it in embankment dam. The characteristics of the breach and determination of their vulnerability has the most effect on the waves resulting from dam break. Investigating, about the parameters of the breach in "Silveh" earth dam have been determined using the suitable model. In Silve dam a trapezoid breach with side slope z=0.01m and the average base line b=80m was computed. The duration of the breaches development is 1.9 hour. Regarding the above results and the application of DAM Break software the consequences of the probable break of the dam was determined. The analysis of the results of water covering of the city of Piranshahr located 12km from silve dam confirms that in 3 hours the water will reach the height (level of 1425 meters.
Neutron emission probability at high excitation and isospin
International Nuclear Information System (INIS)
Aggarwal, Mamta
2005-01-01
One-neutron and two-neutron emission probability at different excitations and varying isospin have been studied. Several degrees of freedom like deformation, rotations, temperature, isospin fluctuations and shell structure are incorporated via statistical theory of hot rotating nuclei
High-resolution urban flood modelling - a joint probability approach
Hartnett, Michael; Olbert, Agnieszka; Nash, Stephen
2017-04-01
(Divoky et al., 2005). Nevertheless, such events occur and in Ireland alone there are several cases of serious damage due to flooding resulting from a combination of high sea water levels and river flows driven by the same meteorological conditions (e.g. Olbert et al. 2015). A November 2009 fluvial-coastal flooding of Cork City bringing €100m loss was one such incident. This event was used by Olbert et al. (2015) to determine processes controlling urban flooding and is further explored in this study to elaborate on coastal and fluvial flood mechanisms and their roles in controlling water levels. The objective of this research is to develop a methodology to assess combined effect of multiple source flooding on flood probability and severity in urban areas and to establish a set of conditions that dictate urban flooding due to extreme climatic events. These conditions broadly combine physical flood drivers (such as coastal and fluvial processes), their mechanisms and thresholds defining flood severity. The two main physical processes controlling urban flooding: high sea water levels (coastal flooding) and high river flows (fluvial flooding), and their threshold values for which flood is likely to occur, are considered in this study. Contribution of coastal and fluvial drivers to flooding and their impacts are assessed in a two-step process. The first step involves frequency analysis and extreme value statistical modelling of storm surges, tides and river flows and ultimately the application of joint probability method to estimate joint exceedence return periods for combination of surges, tide and river flows. In the second step, a numerical model of Cork Harbour MSN_Flood comprising a cascade of four nested high-resolution models is used to perform simulation of flood inundation under numerous hypothetical coastal and fluvial flood scenarios. The risk of flooding is quantified based on a range of physical aspects such as the extent and depth of inundation (Apel et al
Douglas County Dam Breach Inundation Areas
Kansas Data Access and Support Center — Dam breach analysis provides a prediction of the extent and timing of flooding from a catastrophic breach of the dams. These results are sufficient for developing...
13 CFR 115.69 - Imminent Breach.
2010-01-01
... an Imminent Breach of the terms of a Contract covered by an SBA guaranteed bond. The PSB Surety does... 13 Business Credit and Assistance 1 2010-01-01 2010-01-01 false Imminent Breach. 115.69 Section... Surety Bond (PSB) Guarantees § 115.69 Imminent Breach. (a) No prior approval requirement. SBA will...
Rendon, Samuel H.; Ashworth, Chad E.; Smith, S. Jerrod
2012-01-01
Dams provide beneficial functions such as flood control, recreation, and reliable water supplies, but they also entail risk: dam breaches and resultant floods can cause substantial property damage and loss of life. The State of Oklahoma requires each owner of a high-hazard dam, which the Federal Emergency Management Agency defines as dams for which failure or misoperation probably will cause loss of human life, to develop an emergency action plan specific to that dam. Components of an emergency action plan are to simulate a flood resulting from a possible dam breach and map the resulting downstream flood-inundation areas. The resulting flood-inundation maps can provide valuable information to city officials, emergency managers, and local residents for planning the emergency response if a dam breach occurs. Accurate topographic data are vital for developing flood-inundation maps. This report presents results of a cooperative study by the city of Lawton, Oklahoma, and the U.S. Geological Survey (USGS) to model dam-breach scenarios at Lakes Ellsworth and Lawtonka near Lawton and to map the potential flood-inundation areas of such dam breaches. To assist the city of Lawton with completion of the emergency action plans for Lakes Ellsworth and Lawtonka Dams, the USGS collected light detection and ranging (lidar) data that were used to develop a high-resolution digital elevation model and a 1-foot contour elevation map for the flood plains downstream from Lakes Ellsworth and Lawtonka. This digital elevation model and field measurements, streamflow-gaging station data (USGS streamflow-gaging station 07311000, East Cache Creek near Walters, Okla.), and hydraulic values were used as inputs for the dynamic (unsteady-flow) model, Hydrologic Engineering Center's River Analysis System (HEC-RAS). The modeled flood elevations were exported to a geographic information system to produce flood-inundation maps. Water-surface profiles were developed for a 75-percent probable maximum
The extreme risk of personal data breaches and the erosion of privacy
Wheatley, Spencer; Maillart, Thomas; Sornette, Didier
2016-01-01
Personal data breaches from organisations, enabling mass identity fraud, constitute an extreme risk. This risk worsens daily as an ever-growing amount of personal data are stored by organisations and on-line, and the attack surface surrounding this data becomes larger and harder to secure. Further, breached information is distributed and accumulates in the hands of cyber criminals, thus driving a cumulative erosion of privacy. Statistical modeling of breach data from 2000 through 2015 provides insights into this risk: A current maximum breach size of about 200 million is detected, and is expected to grow by fifty percent over the next five years. The breach sizes are found to be well modeled by an extremely heavy tailed truncated Pareto distribution, with tail exponent parameter decreasing linearly from 0.57 in 2007 to 0.37 in 2015. With this current model, given a breach contains above fifty thousand items, there is a ten percent probability of exceeding ten million. A size effect is unearthed where both the frequency and severity of breaches scale with organisation size like s0.6. Projections indicate that the total amount of breached information is expected to double from two to four billion items within the next five years, eclipsing the population of users of the Internet. This massive and uncontrolled dissemination of personal identities raises fundamental concerns about privacy.
High-resolution elastic recoil detection utilizing Bayesian probability theory
International Nuclear Information System (INIS)
Neumaier, P.; Dollinger, G.; Bergmaier, A.; Genchev, I.; Goergens, L.; Fischer, R.; Ronning, C.; Hofsaess, H.
2001-01-01
Elastic recoil detection (ERD) analysis is improved in view of depth resolution and the reliability of the measured spectra. Good statistics at even low ion fluences is obtained utilizing a large solid angle of 5 msr at the Munich Q3D magnetic spectrograph and using a 40 MeV 197 Au beam. In this way the elemental depth profiles are not essentially altered during analysis even if distributions with area densities below 1x10 14 atoms/cm 2 are measured. As the energy spread due to the angular acceptance is fully eliminated by ion-optical and numerical corrections, an accurate and reliable apparatus function is derived. It allows to deconvolute the measured spectra using the adaptive kernel method, a maximum entropy concept in the framework of Bayesian probability theory. In addition, the uncertainty of the reconstructed spectra is quantified. The concepts are demonstrated at 13 C depth profiles measured at ultra-thin films of tetrahedral amorphous carbon (ta-C). Depth scales of those profiles are given with an accuracy of 1.4x10 15 atoms/cm 2
Privacy Breach Analysis in Social Networks
Nagle, Frank
This chapter addresses various aspects of analyzing privacy breaches in social networks. We first review literature that defines three types of privacy breaches in social networks: interactive, active, and passive. We then survey the various network anonymization schemes that have been constructed to address these privacy breaches. After exploring these breaches and anonymization schemes, we evaluate a measure for determining the level of anonymity inherent in a network graph based on its topological structure. Finally, we close by emphasizing the difficulty of anonymizing social network data while maintaining usability for research purposes and offering areas for future work.
Probability based high temperature engineering creep and structural fire resistance
Razdolsky, Leo
2017-01-01
This volume on structural fire resistance is for aerospace, structural, and fire prevention engineers; architects, and educators. It bridges the gap between prescriptive- and performance-based methods and simplifies very complex and comprehensive computer analyses to the point that the structural fire resistance and high temperature creep deformations will have a simple, approximate analytical expression that can be used in structural analysis and design. The book emphasizes methods of the theory of engineering creep (stress-strain diagrams) and mathematical operations quite distinct from those of solid mechanics absent high-temperature creep deformations, in particular the classical theory of elasticity and structural engineering. Dr. Razdolsky’s previous books focused on methods of computing the ultimate structural design load to the different fire scenarios. The current work is devoted to the computing of the estimated ultimate resistance of the structure taking into account the effect of high temperatur...
Teaching Case: Security Breach at Target
Plachkinova, Miloslava; Maurer, Chris
2018-01-01
This case study follows the security breach that affected Target at the end of 2013 and resulted in the loss of financial data for over 70 million customers. The case provides an overview of the company and describes the reasons that led to one of the biggest security breaches in history. It offers a discussion on Target's vendor management…
High probability of comorbidities in bronchial asthma in Germany.
Heck, S; Al-Shobash, S; Rapp, D; Le, D D; Omlor, A; Bekhit, A; Flaig, M; Al-Kadah, B; Herian, W; Bals, R; Wagenpfeil, S; Dinh, Q T
2017-04-21
Clinical experience has shown that allergic and non-allergic respiratory, metabolic, mental, and cardiovascular disorders sometimes coexist with bronchial asthma. However, no study has been carried out that calculates the chance of manifestation of these disorders with bronchial asthma in Saarland and Rhineland-Palatinate, Germany. Using ICD10 diagnoses from health care institutions, the present study systematically analyzed the co-prevalence and odds ratios of comorbidities in the asthma population in Germany. The odds ratios were adjusted for age and sex for all comorbidities for patients with asthma vs. without asthma. Bronchial asthma was strongly associated with allergic and with a lesser extent to non-allergic comorbidities: OR 7.02 (95%CI:6.83-7.22) for allergic rhinitis; OR 4.98 (95%CI:4.67-5.32) allergic conjunctivitis; OR 2.41 (95%CI:2.33-2.52) atopic dermatitis; OR 2.47 (95%CI:2.16-2.82) food allergy, and OR 1.69 (95%CI:1.61-1.78) drug allergy. Interestingly, increased ORs were found for respiratory diseases: 2.06 (95%CI:1.64-2.58) vocal dysfunction; 1.83 (95%CI:1.74-1.92) pneumonia; 1.78 (95%CI:1.73-1.84) sinusitis; 1.71 (95%CI:1.65-1.78) rhinopharyngitis; 2.55 (95%CI:2.03-3.19) obstructive sleep apnea; 1.42 (95%CI:1.25-1.61) pulmonary embolism, and 3.75 (95%CI:1.64-8.53) bronchopulmonary aspergillosis. Asthmatics also suffer from psychiatric, metabolic, cardiac or other comorbidities. Myocardial infarction (OR 0.86, 95%CI:0.79-0.94) did not coexist with asthma. Based on the calculated chances of manifestation for these comorbidities, especially allergic and respiratory, to a lesser extent also metabolic, cardiovascular, and mental disorders should be taken into consideration in the diagnostic and treatment strategy of bronchial asthma. PREVALENCE OF CO-EXISTING DISEASES IN GERMANY: Patients in Germany with bronchial asthma are highly likely to suffer from co-existing diseases and their treatments should reflect this. Quoc Thai Dinh at Saarland
Making Sociology Relevant: The Assignment and Application of Breaching Experiments
Rafalovich, Adam
2006-01-01
Breaching experiments involve the conscious exhibition of "unexpected" behavior, an observation of the types of social reactions such behavioral violations engender, and an analysis of the social structure that makes these social reactions possible. The conscious violation of norms can be highly fruitful for sociology students, providing insights…
In-reactor cladding breach of EBR-II driver-fuel elements
International Nuclear Information System (INIS)
Seidel, B.R.; Einziger, R.E.
1977-01-01
Knowledge of performance and minimum useful element lifetime of Mark-II driver-fuel elements is required to maintain a high plant operating capacity factor with maximum fuel utilization. To obtain such knowledge, intentional cladding breach has been obtained in four run-to-cladding-breach Mark-II experimental driver-fuel subassemblies operating under normal conditions in EBR-II. Breach and subsequent fission-product release proved benign to reactor operations. The breaches originated on the outer surface of the cladding in the root of the restrainer dimples and were intergranular. The Weibull distribution of lifetime accurately predicts the observed minimum useful element lifetime of 10 at.% burnup, with breach ensuing shortly thereafter
Identifying Changes in the Probability of High Temperature, High Humidity Heat Wave Events
Ballard, T.; Diffenbaugh, N. S.
2016-12-01
Understanding how heat waves will respond to climate change is critical for adequate planning and adaptation. While temperature is the primary determinant of heat wave severity, humidity has been shown to play a key role in heat wave intensity with direct links to human health and safety. Here we investigate the individual contributions of temperature and specific humidity to extreme heat wave conditions in recent decades. Using global NCEP-DOE Reanalysis II daily data, we identify regional variability in the joint probability distribution of humidity and temperature. We also identify a statistically significant positive trend in humidity over the eastern U.S. during heat wave events, leading to an increased probability of high humidity, high temperature events. The extent to which we can expect this trend to continue under climate change is complicated due to variability between CMIP5 models, in particular among projections of humidity. However, our results support the notion that heat wave dynamics are characterized by more than high temperatures alone, and understanding and quantifying the various components of the heat wave system is crucial for forecasting future impacts.
What Caused the Breach? An Examination of Use of Information Technology and Health Data Breaches
Wikina, Suanu Bliss
2014-01-01
Data breaches arising from theft, loss, unauthorized access/disclosure, improper disclosure, or hacking incidents involving personal health information continue to increase every year. As of September 2013, reported breaches affecting individuals reached close to 27 million since 2009, when compilation of records on breaches began. These breaches, which involved 674 covered entities and 153 business associates, involved computer systems and networks, desktop computers, laptops, paper, e-mail, electronic health records, and removable/portable devices (CDs, USBs, x-ray films, backup tapes, etc.). Even with the increased use of health information technology by health institutions and allied businesses, theft and loss (not hacking) constitute the major types of data breaches encountered. Removable/portable devices, desktop computers, and laptops were the top sources or locations of the breached information, while the top six states—Virginia, Illinois, California, Florida, New York, and Tennessee—in terms of the number of reported breaches accounted for nearly 75 percent of the total individual breaches, 33 percent of breaches in covered entities, and about 30 percent of the total breaches involving business associates. PMID:25593574
What caused the breach? An examination of use of information technology and health data breaches.
Wikina, Suanu Bliss
2014-01-01
Data breaches arising from theft, loss, unauthorized access/disclosure, improper disclosure, or hacking incidents involving personal health information continue to increase every year. As of September 2013, reported breaches affecting individuals reached close to 27 million since 2009, when compilation of records on breaches began. These breaches, which involved 674 covered entities and 153 business associates, involved computer systems and networks, desktop computers, laptops, paper, e-mail, electronic health records, and removable/portable devices (CDs, USBs, x-ray films, backup tapes, etc.). Even with the increased use of health information technology by health institutions and allied businesses, theft and loss (not hacking) constitute the major types of data breaches encountered. Removable/portable devices, desktop computers, and laptops were the top sources or locations of the breached information, while the top six states-Virginia, Illinois, California, Florida, New York, and Tennessee-in terms of the number of reported breaches accounted for nearly 75 percent of the total individual breaches, 33 percent of breaches in covered entities, and about 30 percent of the total breaches involving business associates.
Data security breaches and privacy in Europe
Wong, Rebecca
2013-01-01
Data Security Breaches and Privacy in Europe aims to consider data protection and cybersecurity issues; more specifically, it aims to provide a fruitful discussion on data security breaches. A detailed analysis of the European Data Protection framework will be examined. In particular, the Data Protection Directive 95/45/EC, the Directive on Privacy and Electronic Communications and the proposed changes under the Data Protection Regulation (data breach notifications) and its implications are considered. This is followed by an examination of the Directive on Attacks against information systems a
Axelrod, Michael I.; Zank, Amber J.
2012-01-01
Noncompliance is one of the most problematic behaviors within the school setting. One strategy to increase compliance of noncompliant students is a high-probability command sequence (HPCS; i.e., a set of simple commands in which an individual is likely to comply immediately prior to the delivery of a command that has a lower probability of…
High But Not Low Probability of Gain Elicits a Positive Feeling Leading to the Framing Effect.
Gosling, Corentin J; Moutier, Sylvain
2017-01-01
Human risky decision-making is known to be highly susceptible to profit-motivated responses elicited by the way in which options are framed. In fact, studies investigating the framing effect have shown that the choice between sure and risky options depends on how these options are presented. Interestingly, the probability of gain of the risky option has been highlighted as one of the main factors causing variations in susceptibility to the framing effect. However, while it has been shown that high probabilities of gain of the risky option systematically lead to framing bias, questions remain about the influence of low probabilities of gain. Therefore, the first aim of this paper was to clarify the respective roles of high and low probabilities of gain in the framing effect. Due to the difference between studies using a within- or between-subjects design, we conducted a first study investigating the respective roles of these designs. For both designs, we showed that trials with a high probability of gain led to the framing effect whereas those with a low probability did not. Second, as emotions are known to play a key role in the framing effect, we sought to determine whether they are responsible for such a debiasing effect of the low probability of gain. Our second study thus investigated the relationship between emotion and the framing effect depending on high and low probabilities. Our results revealed that positive emotion was related to risk-seeking in the loss frame, but only for trials with a high probability of gain. Taken together, these results support the interpretation that low probabilities of gain suppress the framing effect because they prevent the positive emotion of gain anticipation.
High But Not Low Probability of Gain Elicits a Positive Feeling Leading to the Framing Effect
Gosling, Corentin J.; Moutier, Sylvain
2017-01-01
Human risky decision-making is known to be highly susceptible to profit-motivated responses elicited by the way in which options are framed. In fact, studies investigating the framing effect have shown that the choice between sure and risky options depends on how these options are presented. Interestingly, the probability of gain of the risky option has been highlighted as one of the main factors causing variations in susceptibility to the framing effect. However, while it has been shown that high probabilities of gain of the risky option systematically lead to framing bias, questions remain about the influence of low probabilities of gain. Therefore, the first aim of this paper was to clarify the respective roles of high and low probabilities of gain in the framing effect. Due to the difference between studies using a within- or between-subjects design, we conducted a first study investigating the respective roles of these designs. For both designs, we showed that trials with a high probability of gain led to the framing effect whereas those with a low probability did not. Second, as emotions are known to play a key role in the framing effect, we sought to determine whether they are responsible for such a debiasing effect of the low probability of gain. Our second study thus investigated the relationship between emotion and the framing effect depending on high and low probabilities. Our results revealed that positive emotion was related to risk-seeking in the loss frame, but only for trials with a high probability of gain. Taken together, these results support the interpretation that low probabilities of gain suppress the framing effect because they prevent the positive emotion of gain anticipation. PMID:28232808
Outcomes associated with breach and fulfillment of the psychological contract of safety.
Walker, Arlene
2013-12-01
The study investigated the outcomes associated with breach and fulfillment of the psychological contract of safety. The psychological contract of safety is defined as the beliefs of individuals about reciprocal employer and employee safety obligations inferred from implicit or explicit promises. When employees perceive that safety obligations promised by the employer have not been met, a breach of the psychological contract occurs, termed employer breach of obligations. The extent to which employees fulfill their safety obligations to the employer is termed employee fulfillment of obligations. Structural equation modeling was used to test a model of safety that investigated the positive and negative outcomes associated with breach and fulfillment of the psychological contract of safety. Participants were 424 health care workers recruited from two hospitals in the State of Victoria, Australia. Following slight modification of the hypothesized model, a good fitting model resulted. Being injured in the workplace was found to lower perceptions of trust in the employer and increase perceptions of employer breach of safety obligations. Trust in the employer significantly influenced perceived employer breach of safety obligations such that lowered trust resulted in higher perceptions of breach. Perceptions of employer breach significantly impacted employee fulfillment of safety obligations with high perceptions of breach resulting in low employee fulfillment of obligations. Trust and perceptions of breach significantly influenced safety attitudes, but not safety behavior. Fulfillment of employee safety obligations significantly impacted safety behavior, but not safety attitudes. Implications of these findings for safety and psychological contract research are explored. A positive emphasis on social exchange relationships in organizations will have positive outcomes for safety climate and safety behavior. © 2013.
Mitchell, Stewart
2009-01-01
This is the downloadable version of this new pocket guide which provides essential support for organisations who would like to have a tried and tested procedure in place for dealing with data breaches.
International Nuclear Information System (INIS)
Bickel, J.H.
1983-01-01
The quantification of the containment event trees in the Millstone Unit 3 Probabilistic Safety Study utilizes a conditional probability of failure given high pressure which is based on a new approach. The generation of this conditional probability was based on a weakest link failure mode model which considered contributions from a number of overlapping failure modes. This overlap effect was due to a number of failure modes whose mean failure pressures were clustered within a 5 psi range and which had uncertainties due to variances in material strengths and analytical uncertainties which were between 9 and 15 psi. Based on a review of possible probability laws to describe the failure probability of individual structural failure modes, it was determined that a Weibull probability law most adequately described the randomness in the physical process of interest. The resultant conditional probability of failure is found to have a median failure pressure of 132.4 psia. The corresponding 5-95 percentile values are 112 psia and 146.7 psia respectively. The skewed nature of the conditional probability of failure vs. pressure results in a lower overall containment failure probability for an appreciable number of the severe accident sequences of interest, but also probabilities which are more rigorously traceable from first principles
Incomplete Information, Renegotiation, and Breach of Contract
Jihong Lee
2005-01-01
Once a contract has been agreed by two agents, the problem of renegotiating breach under two-sided asymmetric information on the agents' outside options is equivalent to the problem of bilateral trade with uncertain gains. Thus, the theorem of Myerson and Satterthwaite (1983) implies the impossibility of efficient renegotiation. We also show that, assuming no renegotiation, the optimal breach mechanism in this setting corresponds to the expectation damage rule.
SWOT analysis of breach models for common dike failure mechanisms
Peeters, P.; Van Hoestenberghe, T.; Vincke, L.; Visser, P.J.
2011-01-01
The use of breach models includes two tasks: predicting breach characteristics and estimating flow through the breach. Strengths and weaknesses as well as opportunities and threats of different simplified and detailed physically-based breach models are listed following theoretical and practical
International Nuclear Information System (INIS)
Akashi, Masatsune; Kenjyo, Takao; Matsukura, Shinji; Kawamoto, Teruaki
1984-01-01
In order to discuss the probability distribution of intergranular stress corrsion carcking life for sensitized 304 stainless steels, a series of the creviced bent beem (CBB) and the uni-axial constant load tests were carried out in oxygenated high temperature, high purity water. The following concludions were resulted; (1) The initiation process of intergranular stress corrosion cracking has been assumed to be approximated by the Poisson stochastic process, based on the CBB test results. (2) The probability distribution of intergranular stress corrosion cracking life may consequently be approximated by the exponential probability distribution. (3) The experimental data could be fitted to the exponential probability distribution. (author)
Atkinson, Theresa P; Matthews, Russell A; Henderson, Alexandra A; Spitzmueller, Christiane
2018-01-30
Grounded in affective events theory, we investigated the effects of experimentally manipulated psychological contract breaches on participants' feelings of violation, subsequent perceptions of psychological contract strength, and organizational citizenship behaviours in a sample of working adults. Results support previous findings that pre-existing relational psychological contract strength interacts with severity of unmet promises or expectations. Specifically, individuals with high relational contracts who experience low severity of unmet promises/expectations have the lowest breach perceptions, whereas individuals with high relational contracts who experience more severe levels unmet promises/expectations experience the highest level of breach perceptions. Results also support the concept of a breach spiral in that prior perceptions of breach led to an increased likelihood of subsequent perceptions of breach following the experimental manipulation. Furthermore, consistent with affective events theory, results support the argument that a psychological contract breach's effect on specific organizational citizenship behaviours is mediated by feelings of violation and the reassessment of relational contracts. These effects were present even after controlling for the direct effects of the manipulated severity of unmet promises/expectations. Copyright © 2018 John Wiley & Sons, Ltd.
Decomposition of conditional probability for high-order symbolic Markov chains
Melnik, S. S.; Usatenko, O. V.
2017-07-01
The main goal of this paper is to develop an estimate for the conditional probability function of random stationary ergodic symbolic sequences with elements belonging to a finite alphabet. We elaborate on a decomposition procedure for the conditional probability function of sequences considered to be high-order Markov chains. We represent the conditional probability function as the sum of multilinear memory function monomials of different orders (from zero up to the chain order). This allows us to introduce a family of Markov chain models and to construct artificial sequences via a method of successive iterations, taking into account at each step increasingly high correlations among random elements. At weak correlations, the memory functions are uniquely expressed in terms of the high-order symbolic correlation functions. The proposed method fills the gap between two approaches, namely the likelihood estimation and the additive Markov chains. The obtained results may have applications for sequential approximation of artificial neural network training.
Burden of high fracture probability worldwide: secular increases 2010-2040.
Odén, A; McCloskey, E V; Kanis, J A; Harvey, N C; Johansson, H
2015-09-01
The number of individuals aged 50 years or more at high risk of osteoporotic fracture worldwide in 2010 was estimated at 158 million and is set to double by 2040. The aim of this study was to quantify the number of individuals worldwide aged 50 years or more at high risk of osteoporotic fracture in 2010 and 2040. A threshold of high fracture probability was set at the age-specific 10-year probability of a major fracture (clinical vertebral, forearm, humeral or hip fracture) which was equivalent to that of a woman with a BMI of 24 kg/m(2) and a prior fragility fracture but no other clinical risk factors. The prevalence of high risk was determined worldwide and by continent using all available country-specific FRAX models and applied the population demography for each country. Twenty-one million men and 137 million women had a fracture probability at or above the threshold in the world for the year 2010. The greatest number of men and women at high risk were from Asia (55 %). Worldwide, the number of high-risk individuals is expected to double over the next 40 years. We conclude that individuals with high probability of osteoporotic fractures comprise a very significant disease burden to society, particularly in Asia, and that this burden is set to increase markedly in the future. These analyses provide a platform for the evaluation of risk assessment and intervention strategies.
Moores, L; Kline, J; Portillo, A K; Resano, S; Vicente, A; Arrieta, P; Corres, J; Tapson, V; Yusen, R D; Jiménez, D
2016-01-01
ESSENTIALS: When high probability of pulmonary embolism (PE), sensitivity of computed tomography (CT) is unclear. We investigated the sensitivity of multidetector CT among 134 patients with a high probability of PE. A normal CT alone may not safely exclude PE in patients with a high clinical pretest probability. In patients with no clear alternative diagnosis after CTPA, further testing should be strongly considered. Whether patients with a negative multidetector computed tomographic pulmonary angiography (CTPA) result and a high clinical pretest probability of pulmonary embolism (PE) should be further investigated is controversial. This was a prospective investigation of the sensitivity of multidetector CTPA among patients with a priori clinical assessment of a high probability of PE according to the Wells criteria. Among patients with a negative CTPA result, the diagnosis of PE required at least one of the following conditions: ventilation/perfusion lung scan showing a high probability of PE in a patient with no history of PE, abnormal findings on venous ultrasonography in a patient without previous deep vein thrombosis at that site, or the occurrence of venous thromboembolism (VTE) in a 3-month follow-up period after anticoagulation was withheld because of a negative multidetector CTPA result. We identified 498 patients with a priori clinical assessment of a high probability of PE and a completed CTPA study. CTPA excluded PE in 134 patients; in these patients, the pooled incidence of VTE was 5.2% (seven of 134 patients; 95% confidence interval [CI] 1.5-9.0). Five patients had VTEs that were confirmed by an additional imaging test despite a negative CTPA result (five of 48 patients; 10.4%; 95% CI 1.8-19.1), and two patients had objectively confirmed VTEs that occurred during clinical follow-up of at least 3 months (two of 86 patients; 2.3%; 95% CI 0-5.5). None of the patients had a fatal PE during follow-up. A normal multidetector CTPA result alone may not safely
Antecedents of Psychological Contract Breach: The Role of Job Demands, Job Resources, and Affect.
Directory of Open Access Journals (Sweden)
Tim Vantilborgh
Full Text Available While it has been shown that psychological contract breach leads to detrimental outcomes, relatively little is known about factors leading to perceptions of breach. We examine if job demands and resources predict breach perceptions. We argue that perceiving high demands elicits negative affect, while perceiving high resources stimulates positive affect. Positive and negative affect, in turn, influence the likelihood that psychological contract breaches are perceived. We conducted two experience sampling studies to test our hypotheses: the first using daily surveys in a sample of volunteers, the second using weekly surveys in samples of volunteers and paid employees. Our results confirm that job demands and resources are associated with negative and positive affect respectively. Mediation analyses revealed that people who experienced high job resources were less likely to report psychological contract breach, because they experienced high levels of positive affect. The mediating role of negative affect was more complex, as it increased the likelihood to perceive psychological contract breach, but only in the short-term.
Antecedents of Psychological Contract Breach: The Role of Job Demands, Job Resources, and Affect.
Vantilborgh, Tim; Bidee, Jemima; Pepermans, Roland; Griep, Yannick; Hofmans, Joeri
2016-01-01
While it has been shown that psychological contract breach leads to detrimental outcomes, relatively little is known about factors leading to perceptions of breach. We examine if job demands and resources predict breach perceptions. We argue that perceiving high demands elicits negative affect, while perceiving high resources stimulates positive affect. Positive and negative affect, in turn, influence the likelihood that psychological contract breaches are perceived. We conducted two experience sampling studies to test our hypotheses: the first using daily surveys in a sample of volunteers, the second using weekly surveys in samples of volunteers and paid employees. Our results confirm that job demands and resources are associated with negative and positive affect respectively. Mediation analyses revealed that people who experienced high job resources were less likely to report psychological contract breach, because they experienced high levels of positive affect. The mediating role of negative affect was more complex, as it increased the likelihood to perceive psychological contract breach, but only in the short-term.
Mining of high utility-probability sequential patterns from uncertain databases.
Directory of Open Access Journals (Sweden)
Binbin Zhang
Full Text Available High-utility sequential pattern mining (HUSPM has become an important issue in the field of data mining. Several HUSPM algorithms have been designed to mine high-utility sequential patterns (HUPSPs. They have been applied in several real-life situations such as for consumer behavior analysis and event detection in sensor networks. Nonetheless, most studies on HUSPM have focused on mining HUPSPs in precise data. But in real-life, uncertainty is an important factor as data is collected using various types of sensors that are more or less accurate. Hence, data collected in a real-life database can be annotated with existing probabilities. This paper presents a novel pattern mining framework called high utility-probability sequential pattern mining (HUPSPM for mining high utility-probability sequential patterns (HUPSPs in uncertain sequence databases. A baseline algorithm with three optional pruning strategies is presented to mine HUPSPs. Moroever, to speed up the mining process, a projection mechanism is designed to create a database projection for each processed sequence, which is smaller than the original database. Thus, the number of unpromising candidates can be greatly reduced, as well as the execution time for mining HUPSPs. Substantial experiments both on real-life and synthetic datasets show that the designed algorithm performs well in terms of runtime, number of candidates, memory usage, and scalability for different minimum utility and minimum probability thresholds.
Directory of Open Access Journals (Sweden)
Christer Björkman
2011-01-01
Full Text Available It is anticipated that extreme population events, such as extinctions and outbreaks, will become more frequent as a consequence of climate change. To evaluate the increased probability of such events, it is crucial to understand the mechanisms involved. Variation between individuals in their response to climatic factors is an important consideration, especially if microevolution is expected to change the composition of populations.Here we present data of a willow leaf beetle species, showing high variation among individuals in oviposition rate at a high temperature (20 °C. It is particularly noteworthy that not all individuals responded to changes in temperature; individuals laying few eggs at 20 °C continued to do so when transferred to 12 °C, whereas individuals that laid many eggs at 20 °C reduced their oviposition and laid the same number of eggs as the others when transferred to 12 °C. When transferred back to 20 °C most individuals reverted to their original oviposition rate. Thus, high variation among individuals was only observed at the higher temperature. Using a simple population model and based on regional climate change scenarios we show that the probability of outbreaks increases if there is a realistic increase in the number of warm summers. The probability of outbreaks also increased with increasing heritability of the ability to respond to increased temperature.If climate becomes warmer and there is latent variation among individuals in their temperature response, the probability for outbreaks may increase. However, the likelihood for microevolution to play a role may be low. This conclusion is based on the fact that it has been difficult to show that microevolution affect the probability for extinctions. Our results highlight the urge for cautiousness when predicting the future concerning probabilities for extreme population events.
Probability of fracture and life extension estimate of the high-flux isotope reactor vessel
International Nuclear Information System (INIS)
Chang, S.J.
1998-01-01
The state of the vessel steel embrittlement as a result of neutron irradiation can be measured by its increase in ductile-brittle transition temperature (DBTT) for fracture, often denoted by RT NDT for carbon steel. This transition temperature can be calibrated by the drop-weight test and, sometimes, by the Charpy impact test. The life extension for the high-flux isotope reactor (HFIR) vessel is calculated by using the method of fracture mechanics that is incorporated with the effect of the DBTT change. The failure probability of the HFIR vessel is limited as the life of the vessel by the reactor core melt probability of 10 -4 . The operating safety of the reactor is ensured by periodic hydrostatic pressure test (hydrotest). The hydrotest is performed in order to determine a safe vessel static pressure. The fracture probability as a result of the hydrostatic pressure test is calculated and is used to determine the life of the vessel. Failure to perform hydrotest imposes the limit on the life of the vessel. The conventional method of fracture probability calculations such as that used by the NRC-sponsored PRAISE CODE and the FAVOR CODE developed in this Laboratory are based on the Monte Carlo simulation. Heavy computations are required. An alternative method of fracture probability calculation by direct probability integration is developed in this paper. The present approach offers simple and expedient ways to obtain numerical results without losing any generality. In this paper, numerical results on (1) the probability of vessel fracture, (2) the hydrotest time interval, and (3) the hydrotest pressure as a result of the DBTT increase are obtained
Davar, J; Roberts, E; Coghlan, J; Evans, T; Lipkin, D
2001-01-01
OBJECTIVE—To assess the prognostic significance of stress echocardiography in women with a high probability of coronary artery disease (CAD). SETTING—Secondary and tertiary cardiology unit at a university teaching hospital. PARTICIPANTS—A total of 135 women (mean (SD) age 63 (9) years) with pre-test probability of CAD ⩾80% were selected from a database of patients investigated by treadmill or dobutamine stress echocardiography between 1995 and 1998. MAIN OUTCOME MEASURES—Patients were followe...
On the importance of default breach remedies
Sloof, R.; Oosterbeek, H.; Sonnemans, J.
2007-01-01
Theory predicts that default breach remedies are immaterial whenever contracting costs are negligible. Some experimental studies, however, suggest that in practice default rules do matter, as they may affect parties' preferences over contract terms. This paper presents results from an experiment
Breached-pin testing in the US
International Nuclear Information System (INIS)
Mahagin, D.E.; Lambert, J.D.B.
1981-04-01
Experience gained at EBR-II by the late 1970's from a significant number of failures in experimental fuel-pin irradiations forms the basis of a program directed towards the characterization of breached pins. The questions to be answered and the issues raised by further testing are discussed
Data breach locations, types, and associated characteristics among US hospitals.
Gabriel, Meghan Hufstader; Noblin, Alice; Rutherford, Ashley; Walden, Amanda; Cortelyou-Ward, Kendall
2018-02-01
The objectives of this study were to describe the locations in hospitals where data are breached, the types of breaches that occur most often at hospitals, and hospital characteristics, including health information technology (IT) sophistication and biometric security capabilities, that may be predicting factors of large data breaches that affect 500 or more patients. The Office of Civil Rights breach data from healthcare providers regarding breaches that affected 500 or more individuals from 2009 to 2016 were linked with hospital characteristics from the Health Information Management Systems Society and the American Hospital Association Health IT Supplement databases. Descriptive statistics were used to characterize hospitals with and without breaches, data breach type, and location/mode of data breaches in hospitals. Multivariate logistic regression analysis explored hospital characteristics that were predicting factors of a data breach affecting at least 500 patients, including area characteristics, region, health system membership, size, type, biometric security use, health IT sophistication, and ownership. Of all types of healthcare providers, hospitals accounted for approximately one-third of all data breaches and hospital breaches affected the largest number of individuals. Paper and films were the most frequent location of breached data, occurring in 65 hospitals during the study period, whereas network servers were the least common location but their breaches affected the most patients overall. Adjusted multivariate results showed significant associations among data breach occurrences and some hospital characteristics, including type and size, but not others, including health IT sophistication or biometric use for security. Hospitals should conduct routine audits to allow them to see their vulnerabilities before a breach occurs. Additionally, information security systems should be implemented concurrently with health information technologies. Improving
Breached fuel pin contamination from Run Beyond Cladding Breach (RBCB) tests in EBR-II
International Nuclear Information System (INIS)
Colburn, R.P.; Strain, R.V.; Lambert, J.D.B.; Ukai, S.; Shibahara, I.
1988-09-01
Studies indicate there may be a large economic incentive to permit some continued reactor operation with breached fuel pin cladding. A major concern for this type of operation is the potential spread of contamination in the primary coolant system and its impact on plant maintenance. A study of the release and transport of contamination from naturally breached mixed oxide Liquid Metal Reactor (LMR) fuel pins was performed as part of the US Department of Energy/Power Reactor and Nuclear Fuel Development Corporation (DOE/PNC) Run Beyond Cladding Breach (RBCB) Program at EBR-II. The measurements were made using the Breached Fuel Test Facility (BFTF) at EBR-II with replaceable deposition samplers located approximately 1.5 meters from the breached fuel test assemblies. The effluent from the test assemblies containing the breached fuel pins was routed up through the samplers and past dedicated instrumentation in the BFTF before mixing with the main coolant flow stream. This paper discusses the first three contamination tests in this program. 2 refs., 5 figs., 2 tabs
Federal Information Security and Data Breach Notification Laws
2009-01-29
The following report describes information security and data breach notification requirements included in the Privacy Act, the Federal Information...information for unauthorized purposes. Data breach notification laws typically require covered entities to implement a breach notification policy, and...Feinstein), S. 495 (Leahy), and S. 1178 (Inouye)--were reported favorably out of Senate committees. Those bills include information security and data
Intelligent tutorial system for teaching of probability and statistics at high school in Mexico
Directory of Open Access Journals (Sweden)
Fernando Gudino Penaloza, Miguel Gonzalez Mendoza, Neil Hernandez Gress, Jaime Mora Vargas
2009-12-01
Full Text Available This paper describes the implementation of an intelligent tutoring system dedicated to teaching probability and statistics atthe preparatory school (or high school in Mexico. The system solution was used as a desktop computer and adapted tocarry a mobile environment for the implementation of mobile learning or m-learning. The system complies with the idea ofbeing adaptable to the needs of each student and is able to adapt to three different teaching models that meet the criteriaof three student profiles.
The Effect of High Frequency Pulse on the Discharge Probability in Micro EDM
Liu, Y.; Qu, Y.; Zhang, W.; Ma, F.; Sha, Z.; Wang, Y.; Rolfe, B.; Zhang, S.
2017-12-01
High frequency pulse improves the machining efficiency of micro electric discharge machining (micro EDM), while it also brings some changes in micro EDM process. This paper focuses on the influence of skin-effect under the high frequency pulse on energy distribution and transmission in micro EDM, based on which, the rules of discharge probability of electrode end face are also analysed. On the basis of the electrical discharge process under the condition of high frequency pulse in micro EDM, COMSOL Multiphysics software is used to establish energy transmission model in micro electrode. The discharge energy distribution and transmission within tool electrode under different pulse frequencies, electrical currents, and permeability situation are studied in order to get the distribution pattern of current density and electric field intensity in the electrode end face under the influence of electrical parameters change. The electric field intensity distribution is regarded as the influencing parameter of discharge probability on the electrode end. Finally, MATLAB is used to fit the curve and obtain the distribution of discharge probability of electrode end face.
Li, Zhanling; Li, Zhanjie; Li, Chengcheng
2014-05-01
Probability modeling of hydrological extremes is one of the major research areas in hydrological science. Most basins in humid and semi-humid south and east of China are concerned for probability modeling analysis of high flow extremes. While, for the inland river basin which occupies about 35% of the country area, there is a limited presence of such studies partly due to the limited data availability and a relatively low mean annual flow. The objective of this study is to carry out probability modeling of high flow extremes in the upper reach of Heihe River basin, the second largest inland river basin in China, by using the peak over threshold (POT) method and Generalized Pareto Distribution (GPD), in which the selection of threshold and inherent assumptions for POT series are elaborated in details. For comparison, other widely used probability distributions including generalized extreme value (GEV), Lognormal, Log-logistic and Gamma are employed as well. Maximum likelihood estimate is used for parameter estimations. Daily flow data at Yingluoxia station from 1978 to 2008 are used. Results show that, synthesizing the approaches of mean excess plot, stability features of model parameters, return level plot and the inherent independence assumption of POT series, an optimum threshold of 340m3/s is finally determined for high flow extremes in Yingluoxia watershed. The resulting POT series is proved to be stationary and independent based on Mann-Kendall test, Pettitt test and autocorrelation test. In terms of Kolmogorov-Smirnov test, Anderson-Darling test and several graphical diagnostics such as quantile and cumulative density function plots, GPD provides the best fit to high flow extremes in the study area. The estimated high flows for long return periods demonstrate that, as the return period increasing, the return level estimates are probably more uncertain. The frequency of high flow extremes exhibits a very slight but not significant decreasing trend from 1978 to
A prototype method for diagnosing high ice water content probability using satellite imager data
Yost, Christopher R.; Bedka, Kristopher M.; Minnis, Patrick; Nguyen, Louis; Strapp, J. Walter; Palikonda, Rabindra; Khlopenkov, Konstantin; Spangenberg, Douglas; Smith, William L., Jr.; Protat, Alain; Delanoe, Julien
2018-03-01
Recent studies have found that ingestion of high mass concentrations of ice particles in regions of deep convective storms, with radar reflectivity considered safe for aircraft penetration, can adversely impact aircraft engine performance. Previous aviation industry studies have used the term high ice water content (HIWC) to define such conditions. Three airborne field campaigns were conducted in 2014 and 2015 to better understand how HIWC is distributed in deep convection, both as a function of altitude and proximity to convective updraft regions, and to facilitate development of new methods for detecting HIWC conditions, in addition to many other research and regulatory goals. This paper describes a prototype method for detecting HIWC conditions using geostationary (GEO) satellite imager data coupled with in situ total water content (TWC) observations collected during the flight campaigns. Three satellite-derived parameters were determined to be most useful for determining HIWC probability: (1) the horizontal proximity of the aircraft to the nearest overshooting convective updraft or textured anvil cloud, (2) tropopause-relative infrared brightness temperature, and (3) daytime-only cloud optical depth. Statistical fits between collocated TWC and GEO satellite parameters were used to determine the membership functions for the fuzzy logic derivation of HIWC probability. The products were demonstrated using data from several campaign flights and validated using a subset of the satellite-aircraft collocation database. The daytime HIWC probability was found to agree quite well with TWC time trends and identified extreme TWC events with high probability. Discrimination of HIWC was more challenging at night with IR-only information. The products show the greatest capability for discriminating TWC ≥ 0.5 g m-3. Product validation remains challenging due to vertical TWC uncertainties and the typically coarse spatio-temporal resolution of the GEO data.
Breaching barriers to collaboration in public spaces
DEFF Research Database (Denmark)
Heinemann, Trine; Mitchell, Robb
2014-01-01
Technology provoking disparate individuals to collaborate or share experiences in the public space faces a difficult barrier, namely the ordinary social order of urban places. We employed the notion of the breaching experiment to explore how this barrier might be overcome. We analyse responses...... of life in public spaces. Arising from this, we argue for the importance of qualities such as availability, facilitation, perspicuous settings, and perspicuous participants to encourage and support co-located strangers to collaborate and share experiences....
Routes for breaching and protecting genetic privacy
Erlich, Yaniv; Narayanan, Arvind
2013-01-01
We are entering an era of ubiquitous genetic information for research, clinical care and personal curiosity. Sharing these datasets is vital for progress in biomedical research. However, one growing concern is the ability to protect the genetic privacy of the data originators. Here, we present an overview of genetic privacy breaching strategies. We outline the principles of each technique, point to the underlying assumptions, and assess its technological complexity and maturati...
Mass Transfer Model for a Breached Waste Package
International Nuclear Information System (INIS)
Hsu, C.; McClure, J.
2004-01-01
The degradation of waste packages, which are used for the disposal of spent nuclear fuel in the repository, can result in configurations that may increase the probability of criticality. A mass transfer model is developed for a breached waste package to account for the entrainment of insoluble particles. In combination with radionuclide decay, soluble advection, and colloidal transport, a complete mass balance of nuclides in the waste package becomes available. The entrainment equations are derived from dimensionless parameters such as drag coefficient and Reynolds number and based on the assumption that insoluble particles are subjected to buoyant force, gravitational force, and drag force only. Particle size distributions are utilized to calculate entrainment concentration along with geochemistry model abstraction to calculate soluble concentration, and colloid model abstraction to calculate colloid concentration and radionuclide sorption. Results are compared with base case geochemistry model, which only considers soluble advection loss
Preventing a data breach from becoming a disaster.
Goldberg, Ed
2013-01-01
Organisations have traditionally dealt with data breaches by investing in protective measures without a great deal of attention to mitigation of breach consequences and response. Conversely, business continuity (BC) planning has traditionally focused on mitigating disasters, not on preventing them. From a BC planning perspective, organisations need to assume that a data breach is inevitable and plan accordingly. The spate of data breaches in these past few years hit many organisations that were well protected. Those that suffered disastrous consequences as a result of a data breach lacked effective mitigation and response, not protection. The complexity and speed of an effective data breach response require that detailed planning takes place in advance of a breach.
Lattice Boltzmann Study on Seawall-Break Flows under the Influence of Breach and Buildings
Mei, Qiu-Ying; Zhang, Wen-Huan; Wang, Yi-Hang; Chen, Wen-Wen
2017-10-01
In the process of storm surge, the seawater often overflows and even destroys the seawall. The buildings near the shore are usually inundated by the seawater through the breach. However, at present, there is little study focusing on the effects of buildings and breach on the seawall-break flows. In this paper, the lattice Boltzmann (LB) model with nine velocities in two dimensions (D2Q9) for the shallow water equations is adopted to simulate the seawall-break flows. The flow patterns and water depth distributions for the seawall-break flows under various densities, layouts and shapes of buildings and different breach discharges, sizes and locations are investigated. It is found that when buildings with a high enough density are perpendicular to the main flow direction, an obvious backwater phenomenon appears near buildings while this phenomenon does not occur when buildings with the same density are parallel to the main flow direction. Moreover, it is observed that the occurrence of backwater phenomenon is independent of the building shape. As to the effects of breach on the seawall-break flows, it is found that only when the breach discharge is large enough or the breach size is small enough, the effects of asymmetric distribution of buildings on the seawall-break flows become important. The breach location only changes the flow pattern in the upstream area of the first building that seawater meets, but has little impact on the global water depth distribution. Supported by the National Natural Science Foundation of China under Grant No. 11502124, the Natural Science Foundation of Zhejiang Province under Grant No. LQ16A020001, the Scientific Research Fund of Zhejiang Provincial Education Department under Grant No. Y201533808, the Natural Science Foundation of Ningbo under Grant No. 2016A610075, and is sponsored by K.C. Wong Magna Fund in Ningbo University.
2002-02-01
This report examines the literature on involuntary, high-consequence, low-probability (IHL) events like nuclear power plant meltdowns to determine what can be applied to the problem of voluntary, low-consequence high-probability (VLH) events like tra...
Routes for breaching and protecting genetic privacy.
Erlich, Yaniv; Narayanan, Arvind
2014-06-01
We are entering an era of ubiquitous genetic information for research, clinical care and personal curiosity. Sharing these data sets is vital for progress in biomedical research. However, a growing concern is the ability to protect the genetic privacy of the data originators. Here, we present an overview of genetic privacy breaching strategies. We outline the principles of each technique, indicate the underlying assumptions, and assess their technological complexity and maturation. We then review potential mitigation methods for privacy-preserving dissemination of sensitive data and highlight different cases that are relevant to genetic applications.
The external costs of low probability-high consequence events: Ex ante damages and lay risks
International Nuclear Information System (INIS)
Krupnick, A.J.; Markandya, A.; Nickell, E.
1994-01-01
This paper provides an analytical basis for characterizing key differences between two perspectives on how to estimate the expected damages of low probability - high consequence events. One perspective is the conventional method used in the U.S.-EC fuel cycle reports [e.g., ORNL/RFF (1994a,b]. This paper articulates another perspective, using economic theory. The paper makes a strong case for considering this, approach as an alternative, or at least as a complement, to the conventional approach. This alternative approach is an important area for future research. I Interest has been growing worldwide in embedding the external costs of productive activities, particularly the fuel cycles resulting in electricity generation, into prices. In any attempt to internalize these costs, one must take into account explicitly the remote but real possibilities of accidents and the wide gap between lay perceptions and expert assessments of such risks. In our fuel cycle analyses, we estimate damages and benefits' by simply monetizing expected consequences, based on pollution dispersion models, exposure-response functions, and valuation functions. For accidents, such as mining and transportation accidents, natural gas pipeline accidents, and oil barge accidents, we use historical data to estimate the rates of these accidents. For extremely severe accidents--such as severe nuclear reactor accidents and catastrophic oil tanker spills--events are extremely rare and they do not offer a sufficient sample size to estimate their probabilities based on past occurrences. In those cases the conventional approach is to rely on expert judgments about both the probability of the consequences and their magnitude. As an example of standard practice, which we term here an expert expected damage (EED) approach to estimating damages, consider how evacuation costs are estimated in the nuclear fuel cycle report
The external costs of low probability-high consequence events: Ex ante damages and lay risks
Energy Technology Data Exchange (ETDEWEB)
Krupnick, A J; Markandya, A; Nickell, E
1994-07-01
This paper provides an analytical basis for characterizing key differences between two perspectives on how to estimate the expected damages of low probability - high consequence events. One perspective is the conventional method used in the U.S.-EC fuel cycle reports [e.g., ORNL/RFF (1994a,b]. This paper articulates another perspective, using economic theory. The paper makes a strong case for considering this, approach as an alternative, or at least as a complement, to the conventional approach. This alternative approach is an important area for future research. I Interest has been growing worldwide in embedding the external costs of productive activities, particularly the fuel cycles resulting in electricity generation, into prices. In any attempt to internalize these costs, one must take into account explicitly the remote but real possibilities of accidents and the wide gap between lay perceptions and expert assessments of such risks. In our fuel cycle analyses, we estimate damages and benefits' by simply monetizing expected consequences, based on pollution dispersion models, exposure-response functions, and valuation functions. For accidents, such as mining and transportation accidents, natural gas pipeline accidents, and oil barge accidents, we use historical data to estimate the rates of these accidents. For extremely severe accidents--such as severe nuclear reactor accidents and catastrophic oil tanker spills--events are extremely rare and they do not offer a sufficient sample size to estimate their probabilities based on past occurrences. In those cases the conventional approach is to rely on expert judgments about both the probability of the consequences and their magnitude. As an example of standard practice, which we term here an expert expected damage (EED) approach to estimating damages, consider how evacuation costs are estimated in the nuclear fuel cycle report.
The Significance of Mandatory Data Breach Warnings to Identity Crime
Eric Holm; Geraldine Mackenzie
2015-01-01
The relationship between data breaches and identity crime has been scarcely explored in current literature. However, there is an important relationship between the misuse of personal identification information and identity crime as the former is in many respects the catalyst for the latter. Data breaches are one of the ways in which this personal identification information is obtained by identity criminals, and thereby any response to data breaches is likely to impact the incidence of identit...
A 'new' Cromer-related high frequency antigen probably antithetical to WES.
Daniels, G L; Green, C A; Darr, F W; Anderson, H; Sistonen, P
1987-01-01
An antibody to a high frequency antigen, made in a WES+ Black antenatal patient (Wash.), failed to react with the red cells of a presumed WES+ homozygote and is, therefore, probably antithetical to anti-WES. Like anti-WES, it reacted with papain, ficin, trypsin or neuraminidase treated cells but not with alpha-chymotrypsin or pronase treated cells and was specifically inhibited by concentrated serum. It also reacted more strongly in titration with WES- cells than with WES+ cells. The antibody is Cromer-related as it failed to react with Inab phenotype (IFC-) cells and reacted only weakly with Dr(a-) cells. Wash. cells and those of the other possible WES+ homozygote are Cr(a+) Tc(a+b-c-) Dr(a+) IFC+ but reacted only very weakly with anti-Esa.
Venkatesh, Viswanath
2017-01-01
Drawing on expectation confirmation research, we develop hypotheses regarding the effect of compensation on key customer outcomes following a major data breach and consequent service recovery effort. Data were collected in a longitudinal field study of Sony customers during their data breach in 2011. A total of 144 customers participated in the two-phase data collection that began when the breach was announced and concluded after reparations were made. Using polynomial modeling an...
Bordia, Prashant; Restubog, Simon Lloyd D; Tang, Robert L
2008-09-01
In this article, psychological contract breach, revenge, and workplace deviance are brought together to identify the cognitive, affective, and motivational underpinnings of workplace deviance. On the basis of S. L. Robinson and R. J. Bennett's (1997) model of workplace deviance, the authors proposed that breach (a cognitive appraisal) and violation (an affective response) initiate revenge seeking. Motivated by revenge, employees then engage in workplace deviance. Three studies tested these ideas. All of the studies supported the hypothesized relationships. In addition, self-control was found to be a moderator of the relationship between revenge cognitions and deviant acts; the relationship was weaker for people high in self-control.
High-severity fire: evaluating its key drivers and mapping its probability across western US forests
Parks, Sean A.; Holsinger, Lisa M.; Panunto, Matthew H.; Jolly, W. Matt; Dobrowski, Solomon Z.; Dillon, Gregory K.
2018-04-01
Wildland fire is a critical process in forests of the western United States (US). Variation in fire behavior, which is heavily influenced by fuel loading, terrain, weather, and vegetation type, leads to heterogeneity in fire severity across landscapes. The relative influence of these factors in driving fire severity, however, is poorly understood. Here, we explore the drivers of high-severity fire for forested ecoregions in the western US over the period 2002–2015. Fire severity was quantified using a satellite-inferred index of severity, the relativized burn ratio. For each ecoregion, we used boosted regression trees to model high-severity fire as a function of live fuel, topography, climate, and fire weather. We found that live fuel, on average, was the most important factor driving high-severity fire among ecoregions (average relative influence = 53.1%) and was the most important factor in 14 of 19 ecoregions. Fire weather was the second most important factor among ecoregions (average relative influence = 22.9%) and was the most important factor in five ecoregions. Climate (13.7%) and topography (10.3%) were less influential. We also predicted the probability of high-severity fire, were a fire to occur, using recent (2016) satellite imagery to characterize live fuel for a subset of ecoregions in which the model skill was deemed acceptable (n = 13). These ‘wall-to-wall’ gridded ecoregional maps provide relevant and up-to-date information for scientists and managers who are tasked with managing fuel and wildland fire. Lastly, we provide an example of the predicted likelihood of high-severity fire under moderate and extreme fire weather before and after fuel reduction treatments, thereby demonstrating how our framework and model predictions can potentially serve as a performance metric for land management agencies tasked with reducing hazardous fuel across large landscapes.
Hydraulics of embankment-dam breaching
Walder, J. S.; Iverson, R. M.; Logan, M.; Godt, J. W.; Solovitz, S.
2012-12-01
Constructed or natural earthen dams can pose hazards to downstream communities. Experiments to date on earthen-dam breaching have focused on dam geometries relevant to engineering practice. We have begun experiments with dam geometries more like those of natural dams. Water was impounded behind dams constructed at the downstream end of the USGS debris-flow flume. Dams were made of compacted, well-sorted, moist beach sand (D50=0.21 mm), 3.5 m from toe to toe, but varying in height from 0.5 to 1 m; the lower the dam, the smaller the reservoir volume and the broader the initially flat crest. Breaching was started by cutting a slot 30-40 mm wide and deep in the dam crest after filling the reservoir. Water level and pore pressure within the dam were monitored. Experiments were also recorded by an array of still- and video cameras above the flume and a submerged video camera pointed at the upstream dam face. Photogrammetric software was used to create DEMs from stereo pairs, and particle-image velocimetry was used to compute the surface-velocity field from the motion of tracers scattered on the water surface. As noted by others, breaching involves formation and migration of a knickpoint (or several). Once the knickpoint reaches the upstream dam face, it takes on an arcuate form whose continued migration we determined by measuring the onset of motion of colored markers on the dam face. The arcuate feature, which can be considered the head of the "breach channel", is nearly coincident with the transition from subcritical to supercritical flow; that is, it acts as a weir that hydraulically controls reservoir emptying. Photogenic slope failures farther downstream, although the morphologically dominant process at work, play no role at all in hydraulic control aside from rare instances in which they extend upstream so far as to perturb the weir, where the flow cross section is nearly self-similar through time. The domain downstream of the critical-flow section does influence
Corrosion of breached UF6 storage cylinders
International Nuclear Information System (INIS)
Barber, E.J.; Taylor, M.S.; DeVan, J.H.
1993-01-01
This paper describes the corrosion processes that occurred following the mechanical failure of two steel 14-ton storage cylinders containing depleted UF 6 . The failures both were traced to small mechanical tears that occurred during stacking of the cylinders. Although subsequent corrosion processes greatly extended the openings in the wall. the reaction products formed were quite protective and prevented any significant environmental insult or loss of uranium. The relative sizes of the two holes correlated with the relative exposure times that had elapsed from the time of stacking. From the sizes and geometries of the two holes, together with analyses of the reaction products, it was possible to determine the chemical reactions that controlled the corrosion process and to develop a scenario for predicting the rate of hydrolysis of UF 6 , the loss rate of HF, and chemical attack of a breached UF 6 storage cylinder
Just in Time Research: Data Breaches in Higher Education
Grama, Joanna
2014-01-01
This "Just in Time" research is in response to recent discussions on the EDUCAUSE Higher Education Information Security Council (HEISC) discussion list about data breaches in higher education. Using data from the Privacy Rights Clearinghouse, this research analyzes data breaches attributed to higher education. The results from this…
"Financial Emergency" and the Faculty Furlough: A Breach of Contract.
Richards, Mary Sanders
1984-01-01
The power of the university to breach faculty contracts in order to meet its temporary cash-flow problems and the rights of faculty when this breach occurs are discussed. To avoid litigation, a university must have established internal guidelines which can be incorporated into an employment contract. (MLW)
An Analysis of Data Breach Notifications as Negative News
Veltsos, Jennifer R.
2012-01-01
Forty-six states require organizations to notify users when personally identifiable information has been exposed or when the organization's data security measures have been breached. This article describes a qualitative document analysis of 13 data breach notification templates from state and federal agencies. The results confirm much of the…
Do Data Breach Disclosure Laws Reduce Identity Theft?
Romanosky, Sasha; Telang, Rahul; Acquisti, Alessandro
2011-01-01
In the United States, identity theft resulted in corporate and consumer losses of $56 billion dollars in 2005, with up to 35 percent of known identity thefts caused by corporate data breaches. Many states have responded by adopting data breach disclosure laws that require firms to notify consumers if their personal information has been lost or…
The Cryptographic Implications of the LinkedIn Data Breach
Gune, Aditya
2017-01-01
Data security and personal privacy are difficult to maintain in the Internet age. In 2012, professional networking site LinkedIn suffered a breach, compromising the login of over 100 million accounts. The passwords were cracked and sold online, exposing the authentication credentials millions of users. This manuscript dissects the cryptographic failures implicated in the breach, and explores more secure methods of storing passwords.
Using Vignette Methodology to research the process of breach comparatively
Boone, M.M.; Beyens, K.; Maguire, N.; Laurinavicius, A.; Persson, A.
2015-01-01
Comparative research related to any aspect of the process of breach in either the pre-trial, sentencing or release phases is relatively rare. Comparative studies of decision making in the specific context of breach process are particularly lacking. One reason for the dearth of research in this area
Boistard, H.; Lopuhää, H.P.; Ruiz-Gazen, A.
2012-01-01
This paper is devoted to rejective sampling. We provide an expansion of joint inclusion probabilities of any order in terms of the inclusion probabilities of order one, extending previous results by Hájek (1964) and Hájek (1981) and making the remainder term more precise. Following Hájek (1981), the
Measurements of atomic transition probabilities in highly ionized atoms by fast ion beams
International Nuclear Information System (INIS)
Martinson, I.; Curtis, L.J.; Lindgaerd, A.
1977-01-01
A summary is given of the beam-foil method by which level lifetimes and transition probabilities can be determined in atoms and ions. Results are presented for systems of particular interest for fusion research, such as the Li, Be, Na, Mg, Cu and Zn isoelectronic sequences. The available experimental material is compared to theoretical transition probabilities. (author)
International Nuclear Information System (INIS)
Tercariol, Cesar Augusto Sangaletti; Kiipper, Felipe de Moura; Martinez, Alexandre Souto
2007-01-01
Consider that the coordinates of N points are randomly generated along the edges of a d-dimensional hypercube (random point problem). The probability P (d,N) m,n that an arbitrary point is the mth nearest neighbour to its own nth nearest neighbour (Cox probabilities) plays an important role in spatial statistics. Also, it has been useful in the description of physical processes in disordered media. Here we propose a simpler derivation of Cox probabilities, where we stress the role played by the system dimensionality d. In the limit d → ∞, the distances between pair of points become independent (random link model) and closed analytical forms for the neighbourhood probabilities are obtained both for the thermodynamic limit and finite-size system. Breaking the distance symmetry constraint drives us to the random map model, for which the Cox probabilities are obtained for two cases: whether a point is its own nearest neighbour or not
Morphologic evolution of the wilderness area breach at Fire Island, New York—2012–15
Hapke, Cheryl J.; Nelson, Timothy R.; Henderson, Rachel E.; Brenner, Owen T.; Miselis, Jennifer L.
2017-09-18
IntroductionHurricane Sandy, which made landfall on October 29, 2012, near Atlantic City, New Jersey, had a significant impact on the coastal system along the south shore of Long Island, New York. A record significant wave height of 9.6 meters (m) was measured at wave buoy 44025, approximately 48 kilometers offshore of Fire Island, New York. Surge and runup during the storm resulted in extensive beach and dune erosion and breaching of the Fire Island barrier island system at two locations, including a breach that formed within the Otis Pike Fire Island High Dune Wilderness area on the eastern side of Fire Island.The U.S. Geological Survey (USGS) has a long history of conducting morphologic change and processes research at Fire Island. One of the primary objectives of the current research effort is to understand the morphologic evolution of the barrier system on a variety of time scales (from storm scale to decade(s) to century). A number of studies that support the project objectives have been published. Prior to Hurricane Sandy, however, little information was available on specific storm-driven change in this region. The USGS received Hurricane Sandy supplemental funding (project GS2–2B: Linking Coastal Processes and Vulnerability, Fire Island, New York, Regional Study) to enhance existing research efforts at Fire Island. The existing research was greatly expanded to include inner continental shelf mapping and investigations of processes of inner shelf sediment transport; beach and dune response and recovery; and observation, analysis, and modeling of the newly formed breach in the Otis Pike High Dune Wilderness area, herein referred to as the wilderness breach. The breach formed at the site of Old Inlet, which was open from 1763 to 1825. The location of the initial island breaching does not directly correspond with topographic lows of the dunes, but instead the breach formed in the location of a cross-island boardwalk that was destroyed during Hurricane Sandy
Skin damage probabilities using fixation materials in high-energy photon beams
International Nuclear Information System (INIS)
Carl, J.; Vestergaard, A.
2000-01-01
Patient fixation, such as thermoplastic masks, carbon-fibre support plates and polystyrene bead vacuum cradles, is used to reproduce patient positioning in radiotherapy. Consequently low-density materials may be introduced in high-energy photon beams. The aim of the this study was to measure the increase in skin dose when low-density materials are present and calculate the radiobiological consequences in terms of probabilities of early and late skin damage. An experimental thin-windowed plane-parallel ion chamber was used. Skin doses were measured using various overlaying low-density fixation materials. A fixed geometry of a 10 x 10 cm field, a SSD = 100 cm and photon energies of 4, 6 and 10 MV on Varian Clinac 2100C accelerators were used for all measurements. Radiobiological consequences of introducing these materials into the high-energy photon beams were evaluated in terms of early and late damage of the skin based on the measured surface doses and the LQ-model. The experimental ion chamber save results consistent with other studies. A relationship between skin dose and material thickness in mg/cm 2 was established and used to calculate skin doses in scenarios assuming radiotherapy treatment with opposed fields. Conventional radiotherapy may apply mid-point doses up to 60-66 Gy in daily 2-Gy fractions opposed fields. Using thermoplastic fixation and high-energy photons as low as 4 MV do increase the dose to the skin considerably. However, using thermoplastic materials with thickness less than 100 mg/cm 2 skin doses are comparable with those produced by variation in source to skin distance, field size or blocking trays within clinical treatment set-ups. The use of polystyrene cradles and carbon-fibre materials with thickness less than 100 mg/cm 2 should be avoided at 4 MV at doses above 54-60 Gy. (author)
Jihadist Foreign Fighter Phenomenon in Western Europe: A Low-Probability, High-Impact Threat
Directory of Open Access Journals (Sweden)
Edwin Bakker
2015-11-01
Full Text Available The phenomenon of foreign fighters in Syria and Iraq is making headlines. Their involvement in the atrocities committed by terrorist groups such as the so-called “Islamic State” and Jabhat al-Nusra have caused grave concern and public outcry in the foreign fighters’ European countries of origin. While much has been written about these foreign fighters and the possible threat they pose, the impact of this phenomenon on Western European societies has yet to be documented. This Research Paper explores four particular areas where this impact is most visible: a violent incidents associated with (returned foreign fighters, b official and political responses linked to these incidents, c public opinion, and d anti-Islam reactions linked to these incidents. The authors conclude that the phenomenon of jihadist foreign fighters in European societies should be primarily regarded as a social and political threat, not a physical one. They consider the phenomenon of European jihadist foreign fighters a “low-probability, high-impact” threat.
Probable high prevalence of limb-girdle muscular dystrophy type 2D in Taiwan.
Liang, Wen-Chen; Chou, Po-Ching; Hung, Chia-Cheng; Su, Yi-Ning; Kan, Tsu-Min; Chen, Wan-Zi; Hayashi, Yukiko K; Nishino, Ichizo; Jong, Yuh-Jyh
2016-03-15
Limb-girdle muscular dystrophy type 2D (LGMD2D), an autosomal-recessive inherited LGMD, is caused by the mutations in SGCA. SGCA encodes alpha-sarcoglycan (SG) that forms a heterotetramer with other SGs in the sarcolemma, and comprises part of the dystrophin-glycoprotein complex. The frequency of LGMD2D is variable among different ethnic backgrounds, and so far only a few patients have been reported in Asia. We identified five patients with a novel homozygous mutation of c.101G>T (p.Arg34Leu) in SGCA from a big aboriginal family ethnically consisting of two tribes in Taiwan. Patient 3 is the maternal uncle of patients 1 and 2. All their parents, heterozygous for c.101G>T, denied consanguineous marriages although they were from the same tribe. The heterozygous parents of patients 4 and 5 were from two different tribes, originally residing in different geographic regions in Taiwan. Haplotype analysis showed that all five patients shared the same mutation-associated haplotype, indicating the probability of a founder effect and consanguinity. The results suggest that the carrier rate of c.101G>T in SGCA may be high in Taiwan, especially in the aboriginal population regardless of the tribes. It is important to investigate the prevalence of LGMD2D in Taiwan for early diagnosis and treatment. Copyright © 2016. Published by Elsevier B.V.
International Nuclear Information System (INIS)
Wayland, J.R.
1977-03-01
The overlap of the expanding plume of radioactive material from a hypothetical nuclear accident with rainstorms over dense population areas is considered. The conditional probability of the occurrence of hot spots from intense cellular rainfall is presented
Learning difficulties of senior high school students based on probability understanding levels
Anggara, B.; Priatna, N.; Juandi, D.
2018-05-01
Identifying students' difficulties in learning concept of probability is important for teachers to prepare the appropriate learning processes and can overcome obstacles that may arise in the next learning processes. This study revealed the level of students' understanding of the concept of probability and identified their difficulties as a part of the epistemological obstacles identification of the concept of probability. This study employed a qualitative approach that tends to be the character of descriptive research involving 55 students of class XII. In this case, the writer used the diagnostic test of probability concept learning difficulty, observation, and interview as the techniques to collect the data needed. The data was used to determine levels of understanding and the learning difficulties experienced by the students. From the result of students' test result and learning observation, it was found that the mean cognitive level was at level 2. The findings indicated that students had appropriate quantitative information of probability concept but it might be incomplete or incorrectly used. The difficulties found are the ones in arranging sample space, events, and mathematical models related to probability problems. Besides, students had difficulties in understanding the principles of events and prerequisite concept.
Kalyanapu, A. J.; Thames, B. A.
2013-12-01
Dam breach modeling often includes application of models that are sophisticated, yet computationally intensive to compute flood propagation at high temporal and spatial resolutions. This results in a significant need for computational capacity that requires development of newer flood models using multi-processor and graphics processing techniques. Recently, a comprehensive benchmark exercise titled the 12th Benchmark Workshop on Numerical Analysis of Dams, is organized by the International Commission on Large Dams (ICOLD) to evaluate the performance of these various tools used for dam break risk assessment. The ICOLD workshop is focused on estimating the consequences of failure of a hypothetical dam near a hypothetical populated area with complex demographics, and economic activity. The current study uses this hypothetical case study and focuses on evaluating the effects of dam breach methodologies on consequence estimation and analysis. The current study uses ICOLD hypothetical data including the topography, dam geometric and construction information, land use/land cover data along with socio-economic and demographic data. The objective of this study is to evaluate impacts of using four different dam breach methods on the consequence estimates used in the risk assessments. The four methodologies used are: i) Froehlich (1995), ii) MacDonald and Langridge-Monopolis 1984 (MLM), iii) Von Thun and Gillete 1990 (VTG), and iv) Froehlich (2008). To achieve this objective, three different modeling components were used. First, using the HEC-RAS v.4.1, dam breach discharge hydrographs are developed. These hydrographs are then provided as flow inputs into a two dimensional flood model named Flood2D-GPU, which leverages the computer's graphics card for much improved computational capabilities of the model input. Lastly, outputs from Flood2D-GPU, including inundated areas, depth grids, velocity grids, and flood wave arrival time grids, are input into HEC-FIA, which provides the
A laser profilometry technique for monitoring fluvial dike breaching in laboratory experiments
Dewals, Benjamin; Rifai, Ismail; Erpicum, Sébastien; Archambeau, Pierre; Violeau, Damien; Pirotton, Michel; El kadi Abderrezzak, Kamal
2017-04-01
A challenging aspect for experimental modelling of fluvial dike breaching is the continuous monitoring of the transient breach geometry. In dam breaching cases induced by flow overtopping over the whole breach crest (plane erosion), a side view through a glass wall is sufficient to monitor the breach formation. This approach can be extended for 3D dam breach tests (spatial erosion) if the glass wall is located along the breach centreline. In contrast, using a side view does not apply for monitoring fluvial dike breaching, because the breach is not symmetric in this case. We present a non-intrusive, high resolution technique to record the breach development in experimental models of fluvial dikes by means of a laser profilometry (Rifai et al. 2016). Most methods used for monitoring dam and dike breaching involve the projection of a pattern (fringes, grid) on the dam or dike body and the analysis of its deformation on images recorded during the breaching (e.g., Pickert et al. 2011, Frank and Hager 2014). A major limitation of these methods stems from reflection on the water surface, particularly in the vicinity of the breach where the free surface is irregular and rippled. This issue was addressed by Spinewine et al. (2004), who used a single laser sheet so that reflections on the water surface were strongly limited and did not hamper the accurate processing of each image. We have developed a similar laser profilometry technique tailored for laboratory experiments on fluvial dike breaching. The setup is simple and relatively low cost. It consists of a digital video camera (resolution of 1920 × 1080 pixels at 60 frames per second) and a swiping red diode 30 mW laser that enables the projection of a laser sheet over the dike body. The 2D image coordinates of each deformed laser profile incident on the dike are transformed into 3D object coordinates using the Direct Linear Transformation (DLT) algorithm. All 3D object coordinates computed over a swiping cycle of the
Run-beyond-clad-breach oxide testing in EBR-2
International Nuclear Information System (INIS)
Lambert, J.D.B.; Bottcher, J.H.; Strain, R.V.; Gross, K.C.; Lee, M.J.; Webb, J.P.; Colburn, R.P.; Ukai, S.; Nomura, S.; Odo, T.; Shikakura, S.
1990-01-01
Fourteen tests sponsored by the US and Japan were used to study reliability of breached LMR oxide fuel pins during continued operation in EBR-II for a range of conditions and parameters. The fuel-sodium reaction product governed pin behavior. It extended primary breaches by swelling and promoted secondary failures, yet it inhibited loss of fuel and fission products and enhanced release of delayed neutrons used in monitoring breach condition. Fission gas and cesium, the main contaminants from failures, could be adequately controlled. This positive EBR-II experience suggested that limited operation with failed fuel may be feasible in commercial LMR's. 16 refs., 14 figs., 4 tabs
Energy Technology Data Exchange (ETDEWEB)
Gotsman, E.; Maor, U. [Tel Aviv University, Department of Particle Physics, Raymond and Beverly Sackler Faculty of Exact Science, School of Physics and Astronomy, Tel Aviv (Israel); Levin, E. [Tel Aviv University, Department of Particle Physics, Raymond and Beverly Sackler Faculty of Exact Science, School of Physics and Astronomy, Tel Aviv (Israel); Universidad Tecnica Federico Santa Maria, Departemento de Fisica, Centro Cientifico-Tecnologico de Valparaiso, Valparaiso (Chile)
2016-04-15
We estimate the value of the survival probability for central exclusive production in a model which is based on the CGC/saturation approach. Hard and soft processes are described in the same framework. At LHC energies, we obtain a small value for the survival probability. The source of the small value is the impact parameter dependence of the hard amplitude. Our model has successfully described a large body of soft data: elastic, inelastic and diffractive cross sections, inclusive production and rapidity correlations, as well as the t-dependence of deep inelastic diffractive production of vector mesons. (orig.)
Hoogestraat, Galen K.
2011-01-01
Extensive information about the construction of dams or potential downstream hazards in the event of a dam breach is not available for many small reservoirs within the Black Hills National Forest. In 2009, the U.S. Forest Service identified the need for reconnaissance-level dam-breach assessments for four of these reservoirs within the Black Hills National Forest (Iron Creek, Horsethief, Lakota, and Mitchell Lakes) with the potential to flood downstream structures. Flood hydrology and dam-breach hydraulic analyses for the four selected reservoirs were conducted by the U.S. Geological Survey in cooperation with the U.S. Forest service to estimate the areal extent of downstream inundation. Three high-flow breach scenarios were considered for cases when the dam is in place (overtopped) and when a dam break (failure) occurs: the 100-year recurrence 24-hour precipitation, 500-year recurrence peak flow, and the probable maximum precipitation. Inundation maps were developed that show the estimated extent of downstream floodwaters from simulated scenarios. Simulation results were used to determine the hazard classification of a dam break (high, significant, or low), based primarily on the potential for loss of life or property damage resulting from downstream inundation because of the flood surge.The inflow design floods resulting from the two simulated storm events (100-year 24-hour and probable maximum precipitation) were determined using the U.S. Army Corps of Engineers Hydrologic Engineering Center Hydrologic Modeling System (HEC-HMS). The inflow design flood for the 500-year recurrence peak flow was determined by using regional regression equations developed for streamflow-gaging stations with similar watershed characteristics. The step-backwater hydraulic analysis model, Hydrologic Engineering Center's River Analysis System (HEC-RAS), was used to determine water-surface profiles of in-place and dam-break scenarios for the three inflow design floods that were
Influence of Personality on Perception of Psychological Contract Breach
Directory of Open Access Journals (Sweden)
Hassan Jafri
2014-10-01
Full Text Available The present research aimed to investigate the influence of personality (Five-Factor Model on Psychological Contract Breach. Using random sampling procedure, data were collected from 90 faculties of colleges of Royal University of Bhutan. Personality scales by John, Naumann, and Soto (2008 and Robinson and Morrison’s (2000 Psychological Contract Breach scale were used in this study. Correlation and regression analysis were carried out to analyze the obtained data. Results revealed that Extraversion and Neuroticism dimensions of the personality model have been found to be positively associated with the perception of breach. Employees who are by nature Agreeable and Conscientiousness are less likely to perceive breach in their psychological contract. Organization should look into the personality aspect while recruiting employees. If employees are hired with certain personality traits, they may focus on their performance and organizational growth.
Breach Risk Magnitude: A Quantitative Measure of Database Security.
Yasnoff, William A
2016-01-01
A quantitative methodology is described that provides objective evaluation of the potential for health record system breaches. It assumes that breach risk increases with the number of potential records that could be exposed, while it decreases when more authentication steps are required for access. The breach risk magnitude (BRM) is the maximum value for any system user of the common logarithm of the number of accessible database records divided by the number of authentication steps needed to achieve such access. For a one million record relational database, the BRM varies from 5.52 to 6 depending on authentication protocols. For an alternative data architecture designed specifically to increase security by separately storing and encrypting each patient record, the BRM ranges from 1.3 to 2.6. While the BRM only provides a limited quantitative assessment of breach risk, it may be useful to objectively evaluate the security implications of alternative database organization approaches.
Sodium erosion of boron carbide from breached absorber pins
International Nuclear Information System (INIS)
Basmajian, J.A.; Baker, D.E.
1981-03-01
The purpose of the irradiation experiment was to provide an engineering demonstration of the irradiation behavior of breached boron carbide absorber pins. By building defects into the cladding of prototypic absorber pins, and performing the irradiation under typical FFTF operating conditions, a qualitative assessment of the consequences of a breach was achieved. Additionally, a direct comparison of pin behavior with that of the ex-reactor test could be made
Genton, Marc G.
2017-09-07
We present a hierarchical decomposition scheme for computing the n-dimensional integral of multivariate normal probabilities that appear frequently in statistics. The scheme exploits the fact that the formally dense covariance matrix can be approximated by a matrix with a hierarchical low rank structure. It allows the reduction of the computational complexity per Monte Carlo sample from O(n2) to O(mn+knlog(n/m)), where k is the numerical rank of off-diagonal matrix blocks and m is the size of small diagonal blocks in the matrix that are not well-approximated by low rank factorizations and treated as dense submatrices. This hierarchical decomposition leads to substantial efficiencies in multivariate normal probability computations and allows integrations in thousands of dimensions to be practical on modern workstations.
Genton, Marc G.; Keyes, David E.; Turkiyyah, George
2017-01-01
We present a hierarchical decomposition scheme for computing the n-dimensional integral of multivariate normal probabilities that appear frequently in statistics. The scheme exploits the fact that the formally dense covariance matrix can be approximated by a matrix with a hierarchical low rank structure. It allows the reduction of the computational complexity per Monte Carlo sample from O(n2) to O(mn+knlog(n/m)), where k is the numerical rank of off-diagonal matrix blocks and m is the size of small diagonal blocks in the matrix that are not well-approximated by low rank factorizations and treated as dense submatrices. This hierarchical decomposition leads to substantial efficiencies in multivariate normal probability computations and allows integrations in thousands of dimensions to be practical on modern workstations.
Highly enhanced avalanche probability using sinusoidally-gated silicon avalanche photodiode
Energy Technology Data Exchange (ETDEWEB)
Suzuki, Shingo; Namekata, Naoto, E-mail: nnao@phys.cst.nihon-u.ac.jp; Inoue, Shuichiro [Institute of Quantum Science, Nihon University, 1-8-14 Kanda-Surugadai, Chiyoda-ku, Tokyo 101-8308 (Japan); Tsujino, Kenji [Tokyo Women' s Medical University, 8-1 Kawada-cho, Shinjuku-ku, Tokyo 162-8666 (Japan)
2014-01-27
We report on visible light single photon detection using a sinusoidally-gated silicon avalanche photodiode. Detection efficiency of 70.6% was achieved at a wavelength of 520 nm when an electrically cooled silicon avalanche photodiode with a quantum efficiency of 72.4% was used, which implies that a photo-excited single charge carrier in a silicon avalanche photodiode can trigger a detectable avalanche (charge) signal with a probability of 97.6%.
Sari, Dwi Ivayana; Hermanto, Didik
2017-08-01
This research is a developmental research of probabilistic thinking-oriented learning tools for probability materials at ninth grade students. This study is aimed to produce a good probabilistic thinking-oriented learning tools. The subjects were IX-A students of MTs Model Bangkalan. The stages of this development research used 4-D development model which has been modified into define, design and develop. Teaching learning tools consist of lesson plan, students' worksheet, learning teaching media and students' achievement test. The research instrument used was a sheet of learning tools validation, a sheet of teachers' activities, a sheet of students' activities, students' response questionnaire and students' achievement test. The result of those instruments were analyzed descriptively to answer research objectives. The result was teaching learning tools in which oriented to probabilistic thinking of probability at ninth grade students which has been valid. Since teaching and learning tools have been revised based on validation, and after experiment in class produced that teachers' ability in managing class was effective, students' activities were good, students' responses to the learning tools were positive and the validity, sensitivity and reliability category toward achievement test. In summary, this teaching learning tools can be used by teacher to teach probability for develop students' probabilistic thinking.
Directory of Open Access Journals (Sweden)
A. Emmer
2013-06-01
Full Text Available Glacial lake outburst floods (GLOFs and related debris flows represent a significant threat in high mountainous areas across the globe. It is necessary to quantify this threat so as to mitigate their catastrophic effects. Complete GLOF hazard assessment incorporates two phases: the probability of water release from a given glacial lake is estimated through lake and breach hazard assessment while the endangered areas are identified during downstream hazard assessment. This paper outlines a number of methods of lake and breach hazard assessment, which can be grouped into three categories: qualitative, of which we outline eight; semi-quantitative, of which we outline two; and quantitative, of which we outline three. It is considered that five groups of critical parameters are essential for an accurate regionally focused hazard assessment method for moraine-dammed lakes in the Cordillera Blanca. These comprise the possibility of dynamic slope movements into the lake, the possibility of a flood wave from a lake situated upstream, the possibility of dam rupture following a large earthquake, the size of the dam freeboard (or ratio of dam freeboard, and a distinction between natural dams and those with remedial work. It is shown that none of the summarised methods uses all these criteria with, at most, three of the five considered by the outlined methods. A number of these methods were used on six selected moraine-dammed lakes in the Cordillera Blanca: lakes Quitacocha, Checquiacocha, Palcacocha, Llaca, Rajucolta, and Tararhua. The results have been compared and show that each method has certain advantages and disadvantages when used in this region. These methods demonstrate that the most hazardous lake is Lake Palcacocha.
Directory of Open Access Journals (Sweden)
Tintswalo Brenda Mahlaola
2016-12-01
Conclusion: Our study supports previous findings that, in the absence of guidelines, most security breaches were non-intentional acts committed due to ignorance. Of concern are incidents in which sensitive information was intentionally shared via social media.
Directory of Open Access Journals (Sweden)
Tintswalo B. Mahlaola
2016-10-01
Objective: The purpose of this article is to explore the nature of and reasons for confidentiality breaches by PACS users in a South African context. Methods: A closed-ended questionnaire was used to collect quantitative data from 115 health professionals employed in a private hospital setting, including its radiology department and a second independent radiology department. The questionnaire sought to explore the attitudes of participants towards confidentiality breeches and reasons for suchbehaviour. Results: Breach incidences were expressed as percentage compliance and classified according to the nature and reasons provided by Sarkar's breach classification. Cross tabulations indicated a statistical significance (p < 0.00 between the expected and observed confidentiality practices of participants and also the adequacy of training, system knowledge and policy awareness. Conclusion: Our study supports previous findings that, in the absence of guidelines, most security breaches were non-intentional acts committed due to ignorance. Of concern are incidents in which sensitive information was intentionally shared via social media.
International Nuclear Information System (INIS)
Yotsumoto, Hiroki; Yoshida, Kikuo; Genchi, Hiroshi
2011-01-01
In order to control low probability-high consequence disaster which causes huge social and economic damage, it is necessary to develop simultaneous risk assessment simulation tool based on the scheme of disaster risk including diverse effects of primary disaster and secondary damages. We propose the scheme of this risk simulation tool. (author)
Recent trends in the probability of high out-of-pocket medical expenses in the United States
Directory of Open Access Journals (Sweden)
Katherine E Baird
2016-09-01
Full Text Available Objective: This article measures the probability that out-of-pocket expenses in the United States exceed a threshold share of income. It calculates this probability separately by individuals’ health condition, income, and elderly status and estimates changes occurring in these probabilities between 2010 and 2013. Data and Method: This article uses nationally representative household survey data on 344,000 individuals. Logistic regressions estimate the probabilities that out-of-pocket expenses exceed 5% and alternatively 10% of income in the two study years. These probabilities are calculated for individuals based on their income, health status, and elderly status. Results: Despite favorable changes in both health policy and the economy, large numbers of Americans continue to be exposed to high out-of-pocket expenditures. For instance, the results indicate that in 2013 over a quarter of nonelderly low-income citizens in poor health spent 10% or more of their income on out-of-pocket expenses, and over 40% of this group spent more than 5%. Moreover, for Americans as a whole, the probability of spending in excess of 5% of income on out-of-pocket costs increased by 1.4 percentage points between 2010 and 2013, with the largest increases occurring among low-income Americans; the probability of Americans spending more than 10% of income grew from 9.3% to 9.6%, with the largest increases also occurring among the poor. Conclusion: The magnitude of out-of-pocket’s financial burden and the most recent upward trends in it underscore a need to develop good measures of the degree to which health care policy exposes individuals to financial risk, and to closely monitor the Affordable Care Act’s success in reducing Americans’ exposure to large medical bills.
GIS inundation mapping and dam breach analysis of Woolwich Dam using HEC-geoRAS
Energy Technology Data Exchange (ETDEWEB)
Mocan, N. [Crozier and Associates Inc., Collingwood, ON (Canada); Joy, D.M. [Guelph Univ., ON (Canada); Rungis, G. [Grand River Conservation Authority, Cambridge, ON (Canada)
2006-07-01
A study was conducted to determine the extent of flood inundation given a hypothetical dam breach scenario of the Woolwich Dam located in the Grand River Watershed, 2.5 km north of the Town of Elmira, Ontario. The dam is operated by the Grand River Conservation Authority and was constructed to provide low-flow augmentation to Canagagigue Creek. Advances in the computational capabilities of numerical models along with the availability of fine resolution geospatial data has lead to significant advances in the evaluation of catastrophic consequences due to the ensuing flood waters when dams fail. The hydraulic models HEC-RAS and HEC-GeoRAS were used in this study along with GIS to produce high resolution spatial and temporal flood inundation mapping. Given the proximity to the Town of Elmira, the dam is classified as having a high hazard potential. The large size and high hazard potential of the dam suggests that the Inflow Design Flood (IDF) is the Probable Maximum Flood (PMF) event. The outlet structure of the spillway consists of 4 ogee-type concrete spillways equipped with radial gates. A low-level concrete pipe located within the spillway structure provides spillage for maintenance purposes. The full flow capacity of the spillway structure is 297 cubic metres per second at the full supply level of 364.8 metres. In addition to GIS flood inundation maps, this paper included the results of flood hydrographs, water surface profiles and peak flow data. It was concluded that techniques used in this analysis should be considered for use in the development of emergency management planning and dam safety assessments across Canada. 6 refs., 3 tabs., 4 figs.
International Nuclear Information System (INIS)
Cha, Eun Jeong; Ellingwood, Bruce R.
2012-01-01
Quantitative analysis and assessment of risk to civil infrastructure has two components: probability of a potentially damaging event and consequence of damage, measured in terms of financial or human losses. Decision models that have been utilized during the past three decades take into account the probabilistic component rationally, but address decision-maker attitudes toward consequences and risk only to a limited degree. The application of models reflecting these attitudes to decisions involving low-probability, high-consequence events that may impact civil infrastructure requires a fundamental understanding of risk acceptance attitudes and how they affect individual and group choices. In particular, the phenomenon of risk aversion may be a significant factor in decisions for civil infrastructure exposed to low-probability events with severe consequences, such as earthquakes, hurricanes or floods. This paper utilizes cumulative prospect theory to investigate the role and characteristics of risk-aversion in assurance of structural safety.
Values underlying perceptions of breach of the psychological contract
Directory of Open Access Journals (Sweden)
Leon Botha
2010-10-01
Research purpose: The study identifies the most important breaches and investigates which values underlie employee perceptions of breach of the psychological contract. It also addresses values that lead to employees interpreting incidents as breaches. Motivation for the study: The study calls on the fact that employees make inconsequential contributions to the terms of many formal employment contracts may imply that such contracts cannot be viewed as documents between equals. Research design, approach and method: The study identifies the most prominent breaches of the psychological contract and the values underlying the perceptions that violations have occurred. Main findings: The data revealed lack of promotion, poor interpersonal relations between colleagues and bad treatment by seniors as three main breaches of the contract, and social recognition, world of peace and sense of accomplishment as three dominant values that underlie perceptions of contract violation. Practical/managerial implications: The competent and intelligent manner in which lack of promotion is handled and communicated to employees is vital because it has implications for their willingness to contribute, their career prospects and their intention to stay in the organisation. Contribution/value-add: This research can serve as the basis for the development of survey or research instruments that are appropriate and relevant to the population.
International Nuclear Information System (INIS)
Venkatachalam, R.; Venugopal, M.; Prasad, T.
2007-01-01
Full text of publication follows: Suitability of computed radiography with Ir-192, Co-60 and up to 9 MeV x-rays for weld inspections is of importance to many heavy engineering and aerospace industries. CR is preferred because of lesser exposure and processing time as compared to film based radiography and also digital images offers other advantages such as image enhancements, quantitative measurements and easier archival. This paper describes systemic experimental approaches and image quality metrics to compare imaging performance of CR with film-based radiography. Experiments were designed using six-sigma methodology to validate performance of CR for steel thickness up to 160 mm with Ir- 192, Co-60 and x-ray energies varying from 100 kV up to 9 MeV. Weld specimens with defects such as lack of fusion, penetration, cracks, concavity, and porosities were studied for evaluating radiographic sensitivity and imaging performance of the system. Attempts were also made to quantify probability of detection using specimens with artificial and natural defects for various experimental conditions and were compared with film based systems. (authors)
Keyser, Alisa; Westerling, Anthony LeRoy
2017-05-01
A long history of fire suppression in the western United States has significantly changed forest structure and ecological function, leading to increasingly uncharacteristic fires in terms of size and severity. Prior analyses of fire severity in California forests showed that time since last fire and fire weather conditions predicted fire severity very well, while a larger regional analysis showed that topography and climate were important predictors of high severity fire. There has not yet been a large-scale study that incorporates topography, vegetation and fire-year climate to determine regional scale high severity fire occurrence. We developed models to predict the probability of high severity fire occurrence for the western US. We predict high severity fire occurrence with some accuracy, and identify the relative importance of predictor classes in determining the probability of high severity fire. The inclusion of both vegetation and fire-year climate predictors was critical for model skill in identifying fires with high fractional fire severity. The inclusion of fire-year climate variables allows this model to forecast inter-annual variability in areas at future risk of high severity fire, beyond what slower-changing fuel conditions alone can accomplish. This allows for more targeted land management, including resource allocation for fuels reduction treatments to decrease the risk of high severity fire.
Expert system for surveillance and diagnosis of breach fuel elements
Gross, Kenny C.
1989-01-01
An apparatus and method are disclosed for surveillance and diagnosis of breached fuel elements in a nuclear reactor. A delayed neutron monitoring system provides output signals indicating the delayed neutron activity and age and the equivalent recoil areas of a breached fuel element. Sensors are used to provide outputs indicating the status of each component of the delayed neutron monitoring system. Detectors also generate output signals indicating the reactor power level and the primary coolant flow rate of the reactor. The outputs from the detectors and sensors are interfaced with an artificial intelligence-based knowledge system which implements predetermined logic and generates output signals indicating the operability of the reactor.
Expert system for surveillance and diagnosis of breach fuel elements
International Nuclear Information System (INIS)
Gross, K.C.
1989-01-01
An apparatus and method are disclosed for surveillance and diagnosis of breached fuel elements in a nuclear reactor. A delayed neutron monitoring system provides output signals indicating the delayed neutron activity and age and the equivalent recoil areas of a breached fuel element. Sensors are used to provide outputs indicating the status of each component of the delayed neutron monitoring system. Detectors also generate output signals indicating the reactor power level and the primary coolant flow rate of the reactor. The outputs from the detectors and sensors are interfaced with an artificial intelligence-based knowledge system which implements predetermined logic and generates output signals indicating the operability of the reactor
The A to Z of healthcare data breaches.
Kobus, Theodore J
2012-01-01
There currently exists a myriad of privacy laws that impact a healthcare entity, including more than 47 notification laws that require notification when a data breach occurs, as well as the breach notification requirements of the Health Information Technology for Economic and Clinical Health Act. Given the plethora of issues a healthcare entity faces, there are certain principles that can be built into an organization's philosophy that will comply with the law and help protect it from reputational harm. © 2012 American Society for Healthcare Risk Management of the American Hospital Association.
International Nuclear Information System (INIS)
Spruch, L.; Shakeshaft, R.
1984-01-01
For asymptotically high incident velocities we provide simple, heuristic, almost classical, derivations of the cross section for forward charge transfer, and of the ratio of the cross section for capture to the elastic-scattering cross section for the projectile scattered through an angle close to π/3
Wang, Fuming; Hunsche, Stefan; Anunciado, Roy; Corradi, Antonio; Tien, Hung Yu; Tang, Peng; Wei, Junwei; Wang, Yongjun; Fang, Wei; Wong, Patrick; van Oosten, Anton; van Ingen Schenau, Koen; Slachter, Bram
2018-03-01
We present an experimental study of pattern variability and defectivity, based on a large data set with more than 112 million SEM measurements from an HMI high-throughput e-beam tool. The test case is a 10nm node SRAM via array patterned with a DUV immersion LELE process, where we see a variation in mean size and litho sensitivities between different unique via patterns that leads to a seemingly qualitative differences in defectivity. The large available data volume enables further analysis to reliably distinguish global and local CDU variations, including a breakdown into local systematics and stochastics. A closer inspection of the tail end of the distributions and estimation of defect probabilities concludes that there is a common defect mechanism and defect threshold despite the observed differences of specific pattern characteristics. We expect that the analysis methodology can be applied for defect probability modeling as well as general process qualification in the future.
The characterization and monitoring of metallic fuel breaches in EBR-2
International Nuclear Information System (INIS)
Pahl, R.G.; Batte, G.L.; Mikaili, R.; Lambert, J.D.B.; Hofman, G.L.
1991-01-01
This paper discusses the characterization and monitoring of metallic fuel breaches which is now a significant part of the Integral Fast Reactor fuel testing program at Argonne National Laboratory. Irradiation experience with failed metallic fuel now includes natural breaches in the plenum and fuel column regions in lead ''endurance'' tests as well as fuel column breaches in artificially-defected fuel which have operated for months in the run-beyond-cladding breach (RBCB) mode. Analyses of the fission gas (FG) release-to-birth (R/B) ratios of selected historical breaches have been completed and have proven to be very useful in differentiating between plenum and fuel column breaches
DEFF Research Database (Denmark)
Asmussen, Søren; Albrecher, Hansjörg
The book gives a comprehensive treatment of the classical and modern ruin probability theory. Some of the topics are Lundberg's inequality, the Cramér-Lundberg approximation, exact solutions, other approximations (e.g., for heavy-tailed claim size distributions), finite horizon ruin probabilities......, extensions of the classical compound Poisson model to allow for reserve-dependent premiums, Markov-modulation, periodicity, change of measure techniques, phase-type distributions as a computational vehicle and the connection to other applied probability areas, like queueing theory. In this substantially...... updated and extended second version, new topics include stochastic control, fluctuation theory for Levy processes, Gerber–Shiu functions and dependence....
Generalized Probability-Probability Plots
Mushkudiani, N.A.; Einmahl, J.H.J.
2004-01-01
We introduce generalized Probability-Probability (P-P) plots in order to study the one-sample goodness-of-fit problem and the two-sample problem, for real valued data.These plots, that are constructed by indexing with the class of closed intervals, globally preserve the properties of classical P-P
Shiryaev, Albert N
2016-01-01
This book contains a systematic treatment of probability from the ground up, starting with intuitive ideas and gradually developing more sophisticated subjects, such as random walks, martingales, Markov chains, the measure-theoretic foundations of probability theory, weak convergence of probability measures, and the central limit theorem. Many examples are discussed in detail, and there are a large number of exercises. The book is accessible to advanced undergraduates and can be used as a text for independent study. To accommodate the greatly expanded material in the third edition of Probability, the book is now divided into two volumes. This first volume contains updated references and substantial revisions of the first three chapters of the second edition. In particular, new material has been added on generating functions, the inclusion-exclusion principle, theorems on monotonic classes (relying on a detailed treatment of “π-λ” systems), and the fundamental theorems of mathematical statistics.
Earth Data Analysis Center, University of New Mexico — USFS, State Forestry, BLM, and DOI fire occurrence point locations from 1987 to 2008 were combined and converted into a fire occurrence probability or density grid...
Conclusion: probable and possible futures. MRI with ultra high magnetic field
International Nuclear Information System (INIS)
Le Bihan, D.
2009-01-01
MR neuroimaging does not interfere with brain function. Because it is safe, it can be used to study the brains of both patients and healthy volunteers. The tasks performed by neurons depend largely on their precise location, and high-field magnets have the potential to provide a 5- to 10-fold increase in spatio-temporal resolution. This should allow brain function to be studied on a scale of only a few thousand neurons, possibly at the intermediate scale of the 'neural code'. NeuroSpin, a new CEA research center, is dedicated to neuro-MRI at high magnetic field strengths. As a forum for dialogue between those developing and those using these instruments, it brings together researchers and engineers, technicians and medical doctors. NeuroSpin is one of the few institutions in Europe, if not the world, where these experts can come together in one place to design, construct and use machines equipped with ultra-strong magnets. The strongest 'routine' MR device currently operates at 3 Tesla (60 000 times the earth's magnetic field), whereas a first French system operating at 7 Tesla (140 000 times the earth's field) is now available for human studies, and another system operating at 11.7 Tesla (world record) should be delivered in 2011. Preclinical studies are also being conducted with magnets operating at 7 Tesla and, soon, 17.6 Tesla. (author)
International Nuclear Information System (INIS)
Turati, Pietro; Pedroni, Nicola; Zio, Enrico
2016-01-01
The efficient estimation of system reliability characteristics is of paramount importance for many engineering applications. Real world system reliability modeling calls for the capability of treating systems that are: i) dynamic, ii) complex, iii) hybrid and iv) highly reliable. Advanced Monte Carlo (MC) methods offer a way to solve these types of problems, which are feasible according to the potentially high computational costs. In this paper, the REpetitive Simulation Trials After Reaching Thresholds (RESTART) method is employed, extending it to hybrid systems for the first time (to the authors’ knowledge). The estimation accuracy and precision of RESTART highly depend on the choice of the Importance Function (IF) indicating how close the system is to failure: in this respect, proper IFs are here originally proposed to improve the performance of RESTART for the analysis of hybrid systems. The resulting overall simulation approach is applied to estimate the probability of failure of the control system of a liquid hold-up tank and of a pump-valve subsystem subject to degradation induced by fatigue. The results are compared to those obtained by standard MC simulation and by RESTART with classical IFs available in the literature. The comparison shows the improvement in the performance obtained by our approach. - Highlights: • We consider the issue of estimating small failure probabilities in dynamic systems. • We employ the RESTART method to estimate the failure probabilities. • New Importance Functions (IFs) are introduced to increase the method performance. • We adopt two dynamic, hybrid, highly reliable systems as case studies. • A comparison with literature IFs proves the effectiveness of the new IFs.
Energy Technology Data Exchange (ETDEWEB)
Moustafa, H; Elhaddad, SH; Wagih, SH; Ziada, G; Samy, A; Saber, R [Department of nuclear medicine and radiology, faculty of medicine, Cairo university, Cairo, (Egypt)
1995-10-01
138 patients proved with V/P scan to have different probabilities of pulmonary emboli event. Serial follow up scanning after 3 days, 2 weeks, 1 month and 3 months was done, with anticoagulant therapy. Out of the remaining 10 patients, 6 patients died with documented P.E. by P.M. study and lost follow up recorded in 4 patients. Complete response with disappearance of all perfusion defects after 2 weeks was detected in 37 patients (49.3%), partial improvement of lesions after 3 months was elicited in 32%. The overall incidence of response was (81.3%) such response was complete in low probability group (100%), (84.2%) in intermediate group and (79.3%) in high probability group with partial response in 45.3%. New lesions were evident in 18.7% of this series. To conclude that serial follow up V/P scan is mandatory for evaluation of response to anticoagulant therapy specially in first 3 months. 2 figs., 3 tabs.
A Simple model for breach formation by overtopping
Energy Technology Data Exchange (ETDEWEB)
Trousseau, P. [Hydro-Quebec, Montreal, PQ (Canada); Kahawita, R. [Ecole Polytechnique, Montreal, PQ (Canada)
2006-07-01
Failures in earth or rockfill dams are often caused by overtopping of the crest, leading to initiation and uncontrolled progression of a breach. Overtopping may occur because of large inflows into the reservoir caused by excessive rainfall or by the failure of an upstream dam that causes a large volume of water to suddenly arrive at the downstream reservoir thus rapidly exceeding the storage and spillway evacuation capacity. Breach formation in a rockfill or earthfill dike due to overtopping of the crest is a complex process as it involves interaction between the hydraulics of the flow and the erosion characteristics of the fill material. This paper presented a description and validation of a simple parametric model for breach formation due to overtopping. A study was conducted to model, as closely as possible, the physical processes involved within the restriction of the simplified analysis. The objective of the study was to predict the relevant timescales for the phenomenon leading to a prediction of the outflow hydrograph. The model has been validated on the Oros dam failure in Brazil as well as on embankment tests conducted at Rosvatn, Norway. It was concluded that the major impediment to the development of breach erosion models for use as predictive tools is in the characterization of the erosion behaviour. 19 refs., 2 tabs., 9 figs.
50 CFR 38.9 - Breach of the peace.
2010-10-01
... 50 Wildlife and Fisheries 6 2010-10-01 2010-10-01 false Breach of the peace. 38.9 Section 38.9... peace. No person on Midway Atoll National Wildlife Refuge will: (a) With intent to cause public..., gestures, or displays, or address abusive language to any person present; or create a hazardous or...
First HIV legal precedent in Kyrgyzstan: breach of medical privacy.
Iriskulbekov, Erik; Balybaeva, Asylgul
2007-12-01
A recent court case of a breach of the privacy rights of a person living with HIV/AIDS in Kyrgyzstan is the first of its kind in Central Asia, write Erik Iriskulbekov and Asylgul Balybaeva. ADILET, the NGO that brought the case to court, is one of only a few NGOs in Central Asia that provide legal assistance related to HIV and AIDS.
FFTF criteria for run to cladding breach experiments
International Nuclear Information System (INIS)
Van Keuren, J.C.; Heard, F.J.; Stepnewski, D.D.
1985-12-01
The review of experiments proposed for irradiation in FFTF resulted in the development of new criteria for run-to-cladding breach experiments. These criteria have allowed irradiation of aggressive experiments without compromising the safety bases for FFTF. This paper consisting of a set of narrated slides, discusses these criteria and related bases
Pinhole Breaches in Spent Fuel Containers: Some Modeling Considerations
International Nuclear Information System (INIS)
Casella, Andrew M.; Loyalka, Sudarsham K.; Hanson, Brady D.
2006-01-01
This paper replaces PNNL-SA-48024 and incorporates the ANS reviewer's comments, including the change in the title. Numerical methods to solve the equations for gas diffusion through very small breaches in spent fuel containers are presented and compared with previous literature results
(PACS) data security breaches: Intentional versus non-intentional ...
African Journals Online (AJOL)
Background: The Picture Archiving and Communication System (PACS) has led to an increase in breached health records and violation of patient confidentiality. The South African constitution makes provision for human dignity and privacy, virtues which confidentiality seeks to preserve. Confidentiality thus constitutes a ...
24 CFR 982.453 - Owner breach of contract.
2010-04-01
...) If the owner has committed fraud, bribery or any other corrupt or criminal act in connection with any..., bribery or any other corrupt or criminal act in connection with the mortgage or loan. (5) If the owner has... 24 Housing and Urban Development 4 2010-04-01 2010-04-01 false Owner breach of contract. 982.453...
''Spinolaminar breach'': an important sign in cervical spinous process fractures
International Nuclear Information System (INIS)
Matar, L.D.; Helms, C.A.
2000-01-01
Objective. To report the sign of ''spinolaminar breach'' and its likely importance in fractures of the cervical spinous processes.Design. Six cases of spinous process fractures demonstrating disruption of the spinolaminar line or ''spinolaminar breach'' were analyzed. Lateral and anteroposterior radiographs (n=6), CT scans (n=3) and MRI scans (n=1) were reviewed together by the authors, with consensus being reached as to the radiographic findings. Clinical records were also reviewed.Results. The levels of injury were C6 (n=5) and C5 (n=2). Injuries were associated with delayed anterior subluxation (n=4) and neurological deficit (n=2). Five patients were male and one was female with a mean age of 31 years (range 8-59 years). Injuries resulted from motor vehicle accidents (n=4), a motor cycle accident (n=1) and a fall (n=1).Conclusion. ''Spinolaminar breach'', or disruption of the spinolaminar line, indicates a complex spinous process fracture with extension into the lamina and spinal canal. Spinous process fractures with spinolaminar breach may have associated posterior ligamentous injury with potential for delayed instability and neurological deficit. It is important that radiologists and physicians caring for the trauma patient be aware of this sign in order to avoid misdiagnosis as a ''clay shoveler's fracture'', which can lead to adverse outcome. (orig.)
Decrease the Number of Glovebox Glove Breaches and Failures
Energy Technology Data Exchange (ETDEWEB)
Hurtle, Jackie C. [Los Alamos National Lab. (LANL), Los Alamos, NM (United States)
2013-12-24
Los Alamos National Laboratory (LANL) is committed to the protection of the workers, public, and environment while performing work and uses gloveboxes as engineered controls to protect workers from exposure to hazardous materials while performing plutonium operations. Glovebox gloves are a weak link in the engineered controls and are a major cause of radiation contamination events which can result in potential worker exposure and localized contamination making operational areas off-limits and putting programmatic work on hold. Each day of lost opportunity at Technical Area (TA) 55, Plutonium Facility (PF) 4 is estimated at $1.36 million. Between July 2011 and June 2013, TA-55-PF-4 had 65 glovebox glove breaches and failures with an average of 2.7 per month. The glovebox work follows the five step safety process promoted at LANL with a decision diamond interjected for whether or not a glove breach or failure event occurred in the course of performing glovebox work. In the event that no glove breach or failure is detected, there is an additional decision for whether or not contamination is detected. In the event that contamination is detected, the possibility for a glove breach or failure event is revisited.
Bolliger, Janine; Edwards, Thomas C.; Eggenberg, Stefan; Ismail, Sascha; Seidl, Irmi; Kienast, Felix
2011-01-01
Abandonment of agricultural land has resulted in forest regeneration in species-rich dry grasslands across European mountain regions and threatens conservation efforts in this vegetation type. To support national conservation strategies, we used a site-selection algorithm (MARXAN) to find optimum sets of floristic regions (reporting units) that contain grasslands of high conservation priority. We sought optimum sets that would accommodate 136 important dry-grassland species and that would minimize forest regeneration and costs of management needed to forestall predicted forest regeneration. We did not consider other conservation elements of dry grasslands, such as animal species richness, cultural heritage, and changes due to climate change. Optimal sets that included 95–100% of the dry grassland species encompassed an average of 56–59 floristic regions (standard deviation, SD 5). This is about 15% of approximately 400 floristic regions that contain dry-grassland sites and translates to 4800–5300 ha of dry grassland out of a total of approximately 23,000 ha for the entire study area. Projected costs to manage the grasslands in these optimum sets ranged from CHF (Swiss francs) 5.2 to 6.0 million/year. This is only 15–20% of the current total estimated cost of approximately CHF30–45 million/year required if all dry grasslands were to be protected. The grasslands of the optimal sets may be viewed as core sites in a national conservation strategy.
International Nuclear Information System (INIS)
Carausu, A.
1996-01-01
A method for the fragility estimation of seismically isolated nuclear power plant structure is proposed. The relationship between the ground motion intensity parameter (e.g. peak ground velocity or peak ground acceleration) and the response of isolated structures is expressed in terms of a bi-linear regression line, whose coefficients are estimated by the least-square method in terms of available data on seismic input and structural response. The notion of high confidence low probability of failure (HCLPF) value is also used for deriving compound fragility curves for coupled subsystems. (orig.)
Frisbee, Joseph H., Jr.
2015-01-01
Upper bounds on high speed satellite collision probability, PC †, have been investigated. Previous methods assume an individual position error covariance matrix is available for each object. The two matrices being combined into a single, relative position error covariance matrix. Components of the combined error covariance are then varied to obtain a maximum PC. If error covariance information for only one of the two objects was available, either some default shape has been used or nothing could be done. An alternative is presented that uses the known covariance information along with a critical value of the missing covariance to obtain an approximate but potentially useful Pc upper bound.
Directory of Open Access Journals (Sweden)
S. P. Zhemanyuk
2017-12-01
Full Text Available One of the topic issue of modern cardiology is factors of target blood pressure level non-achievement clarifying due to a better understanding how we can reduce cardiovascular complications. The aim of the study is to determine the factors of poor blood pressure control using the ambulatory blood pressure monitoring parameters and adenosine 5'-diphosphate-induced platelet aggregation parameters in patients with arterial hypertension. Material and methods. The study involved 153 patients with essential hypertension (EH stage II, II degree. The ambulatory blood pressure monitoring (ABPM was performed in patients during at least two of first-line antihypertensive drugs in optimal daily doses usage by the ABPM bifunctional device (Incart, S.-P., R.F.. Platelet aggregation was carried out using light transmittance aggregation by optical analyzer (Solar, R.B. with adenosine 5'-diphosphate (Sigma-Aldrich at final concentration of 10.0 × 10-6 mol / L. The first group were inadequately controlled essential hypertensive individuals with high systolic or/and diastolic BP level according to the ABPM results, and the second one were patients with adequately controlled EH. Groups of patients were comparable in age (60.39 ± 10.74 years vs. 62.80 ± 9.63; p = 0.181, respectively. In the group of EH patients who reached the target level of blood pressure, women predominated (60% vs. 39.81%; p = 0.021, respectively. We used the binary logistic regression analysis to determine the predictors of target blood pressure level poor reaching using ABPM and platelet aggregation parameters. Results According to the univariate logistic regression analysis, the dependent factors influencing the target blood pressure level poor reaching are the average diurnal diastolic blood pressure (DBP (OR = 44.8; diurnal variability of systolic blood pressure (SBP (OR = 4.4; square index of hypertension for diurnal periods SBP (OR = 318.9; square index of hypertension for diurnal
Quantum Probabilities as Behavioral Probabilities
Directory of Open Access Journals (Sweden)
Vyacheslav I. Yukalov
2017-03-01
Full Text Available We demonstrate that behavioral probabilities of human decision makers share many common features with quantum probabilities. This does not imply that humans are some quantum objects, but just shows that the mathematics of quantum theory is applicable to the description of human decision making. The applicability of quantum rules for describing decision making is connected with the nontrivial process of making decisions in the case of composite prospects under uncertainty. Such a process involves deliberations of a decision maker when making a choice. In addition to the evaluation of the utilities of considered prospects, real decision makers also appreciate their respective attractiveness. Therefore, human choice is not based solely on the utility of prospects, but includes the necessity of resolving the utility-attraction duality. In order to justify that human consciousness really functions similarly to the rules of quantum theory, we develop an approach defining human behavioral probabilities as the probabilities determined by quantum rules. We show that quantum behavioral probabilities of humans do not merely explain qualitatively how human decisions are made, but they predict quantitative values of the behavioral probabilities. Analyzing a large set of empirical data, we find good quantitative agreement between theoretical predictions and observed experimental data.
Patton, Madeline
2015-01-01
Data breach prevention is a battle, rarely plain and never simple. For higher education institutions, the Sisyphean aspects of the task are more complex than for industry and business. Two-year colleges have payrolls and vendor contracts like those enterprises. They also have public record and student confidentiality requirements. Colleges must…
DEFF Research Database (Denmark)
Rojas-Nandayapa, Leonardo
Tail probabilities of sums of heavy-tailed random variables are of a major importance in various branches of Applied Probability, such as Risk Theory, Queueing Theory, Financial Management, and are subject to intense research nowadays. To understand their relevance one just needs to think...... analytic expression for the distribution function of a sum of random variables. The presence of heavy-tailed random variables complicates the problem even more. The objective of this dissertation is to provide better approximations by means of sharp asymptotic expressions and Monte Carlo estimators...
48 CFR 52.233-4 - Applicable Law for Breach of Contract Claim.
2010-10-01
... Provisions and Clauses 52.233-4 Applicable Law for Breach of Contract Claim. As prescribed in 33.215(b), insert the following clause: Applicable Law for Breach of Contract Claim (OCT 2004) United States law... 48 Federal Acquisition Regulations System 2 2010-10-01 2010-10-01 false Applicable Law for Breach...
2013-01-25
... RIN 0945-AA03 Modifications to the HIPAA Privacy, Security, Enforcement, and Breach Notification Rules... HIPAA Privacy, Security, Breach Notification, and Enforcement Rules (the HIPAA Rules) to improve their... entities Total cost Notices of Privacy Practices.. 700,000 covered $55.9 million. entities. Breach...
25 CFR 163.42 - Obligated service and breach of contract.
2010-04-01
... 25 Indians 1 2010-04-01 2010-04-01 false Obligated service and breach of contract. 163.42 Section... breach of contract. (a) Obligated service. (1) Individuals completing forestry education programs with an... request for waiver. (b) Breach of contract. Any individual who has participated in and accepted financial...
A SWOT analysis of hydrodynamic models with respect to simulating breaching
van Damme, M.; Visser, P.J.
2015-01-01
Deriving the bed shear stresses from hydrodynamic models in breach models is challenging due to the continuous changing hydraulic head over the breach in combination with horizontal and vertical flow contractions, and the continuous rapidly changing breach geometry. Three stages can be distinguished
Quick Look Report for Chemical Reactivity Modeling of Various Multi-Canister Overpack Breaches
International Nuclear Information System (INIS)
Bratton, Robert Lawrence
2002-01-01
. A uranium oxide coating covers the exposed uranium metal, yet uranium hydride can still form under the protective oxide coating over the 40-year interim storage time span. The current treatment process at Hanford does not remove chemically bound water contained in the hydrates or in the waters of hydration. The chemically bound water is the source material for hydrogen production over the 40-year storage time. So, additional uranium hydride creates concerns that breaches of an MCO with the appropriate size openings could result in the onset of bulk uranium oxidation with the potential of a self-sustaining thermal excursion or pyrophoric event. For this analysis, the worst-case scenario appears to be the match head configuration in a vertically standing MCO, where all the reactive surface area is placed on the tips of the fuel elements. This configuration concentrates the heat-producing chemical reaction at the tips of the fuel elements. Because no mechanistic drop analysis has been performed at this time to determine the MCO failure modes, parametric breach configurations were chosen in this analysis to determine the MCOs external thermal response range. The first breach is a pair of holes that suddenly open in the MCO wall. This thermal excursion is controlled by the ''thermal chimney effect'' in the 4.27-m (14-ft) tall canisters caused by the multiple holes breach (one high and one low). A second breach where the MCO lid is suddenly removed and exposed to the ambient air environment is evaluated. This thermal excursion is controlled by the countercurrent flow through the top of the MCO. Computer models for these breach configurations were constructed and executed
Grinstead, Charles M; Snell, J Laurie
2011-01-01
This book explores four real-world topics through the lens of probability theory. It can be used to supplement a standard text in probability or statistics. Most elementary textbooks present the basic theory and then illustrate the ideas with some neatly packaged examples. Here the authors assume that the reader has seen, or is learning, the basic theory from another book and concentrate in some depth on the following topics: streaks, the stock market, lotteries, and fingerprints. This extended format allows the authors to present multiple approaches to problems and to pursue promising side discussions in ways that would not be possible in a book constrained to cover a fixed set of topics. To keep the main narrative accessible, the authors have placed the more technical mathematical details in appendices. The appendices can be understood by someone who has taken one or two semesters of calculus.
Dorogovtsev, A Ya; Skorokhod, A V; Silvestrov, D S; Skorokhod, A V
1997-01-01
This book of problems is intended for students in pure and applied mathematics. There are problems in traditional areas of probability theory and problems in the theory of stochastic processes, which has wide applications in the theory of automatic control, queuing and reliability theories, and in many other modern science and engineering fields. Answers to most of the problems are given, and the book provides hints and solutions for more complicated problems.
Directory of Open Access Journals (Sweden)
Xiao Liu
2016-01-01
Full Text Available Random disturbance factors would lead to the variation of target acquisition point during the long distance flight. To acquire a high target acquisition probability and improve the impact precision, missiles should be guided to an appropriate target acquisition position with certain attitude angles and line-of-sight (LOS angle rate. This paper has presented a new midcourse guidance law considering the influences of random disturbances, detection distance restraint, and target acquisition probability with Monte Carlo simulation. Detailed analyses of the impact points on the ground and the random distribution of the target acquisition position in the 3D space are given to get the appropriate attitude angles and the end position for the midcourse guidance. Then, a new formulation biased proportional navigation (BPN guidance law with angular constraint and LOS angle rate control has been derived to ensure the tracking ability when attacking the maneuvering target. Numerical simulations demonstrates that, compared with the proportional navigation guidance (PNG law and the near-optimal spatial midcourse guidance (NSMG law, BPN guidance law demonstrates satisfactory performances and can meet both the midcourse terminal angular constraint and the LOS angle rate requirement.
Consequences of and remedies for breach of natural gas contracts
International Nuclear Information System (INIS)
Gretener, N. M.; Evans, A.; Callihoo, M.
1999-01-01
A common clause in a gas purchase contract is one that provides for specific damages for the non-performance of an obligation. As a rule, damages will be calculated based on the loss in the value of the bargain plus those losses foreseeably caused by the breach of contract. the rationale being to put the non-breaching party in as good a position as it would have been had the contract been performed. This paper examines the complex issues involved assessing and measuring damages, the concept of injunctive relief in circumstances where damages will be inadequate or insufficient to prevent injustice, the doctrine of mitigation, the extent of the right of set-off between different contracts, and the impact of bankruptcy and insolvency laws on the exercise of remedies. Four case histories are presented to illustrate the Courts' treatment of gas purchase contracts in the context of bankruptcies and /or insolvencies. 36 refs
Consequences of and remedies for breach of natural gas contracts
Energy Technology Data Exchange (ETDEWEB)
Gretener, N. M.; Evans, A.; Callihoo, M. [Bennett Jones Law Group, Calgary, AB (Canada)
1999-07-01
A common clause in a gas purchase contract is one that provides for specific damages for the non-performance of an obligation. As a rule, damages will be calculated based on the loss in the value of the bargain plus those losses foreseeably caused by the breach of contract. the rationale being to put the non-breaching party in as good a position as it would have been had the contract been performed. This paper examines the complex issues involved assessing and measuring damages, the concept of injunctive relief in circumstances where damages will be inadequate or insufficient to prevent injustice, the doctrine of mitigation, the extent of the right of set-off between different contracts, and the impact of bankruptcy and insolvency laws on the exercise of remedies. Four case histories are presented to illustrate the Courts' treatment of gas purchase contracts in the context of bankruptcies and /or insolvencies. 36 refs.
Work engagement, psychological contract breach and job satisfaction
Rayton, Bruce A.; Yalabik, Zeynep Y.
2014-01-01
This study extends both Social Exchange Theory and the Job Demands-Resources model by examining the link between psychological contract breach (PCB) and work engagement, and by integrating job satisfaction into this exchange relationship. We argue that PCB reflects employees' feelings of resource loss, and that these feelings impact work engagement through their impact on job satisfaction. Levels of employee work engagement can therefore be viewed as reciprocation for the exchange content pro...
SECURITY BREACH IN TRADING SYSTEM-COUNTERMEASURE USING IPTRACEBACK
M. P. Rajakumar; V. Shanthi
2014-01-01
Recently, economic scenario is often facing security breach that has heavy impact on the financial soundness of a company particularly, stock prices on firms. The utmost consequence being the whole business comes to a standstill. From the estimates attributed by the financial sector, it has been inferred that the loss incurred on virus and worms attack is said to have the greatest impact that hampers the prosperity of a business entity. Thus, security strategies attempt on revolving around th...
International Nuclear Information System (INIS)
Hajnal, F.; Wilson, J.
1992-01-01
'Full Text:' Several High-altitude cosmic-ray neutron measurements were performed by the NASA Ames Laboratory in the mid-to late-1970s using airplanes flying at about 13km altitude along constant geomagnetic latitudes of 20, 44 and 51 degrees north. Bonner spheres and manganese, gold and aluminium foils were used in the measurements. In addition, large moderated BF-3 counters served as normalizing instruments. Data analyses performed at that time did not provide complete and unambiguous spectral information and field intensities. Recently, using our new unfolding methods and codes, and Bonner-sphere response function extensions for higher energies, 'new' neutron spectral intensities were obtained, which show progressive hardening of neutron spectra as a function of increasing geomagnetic latitude, with substantial increases in the energy region iron, 1 0 MeV to 10 GeV. For example, we found that the total neutron fluences at 20 and 51 degrees magnetic north are in the ratio of 1 to 5.2 and the 10 MeV to 10 GeV fluence ratio is 1 to 18. The magnitude of these ratios is quite remarkable. From the new results, the derived absolute neutron energy distribution is of the correct strength and shape for the albedo neutrons to be the main source of the high-energy protons trapped in the Earth's inner radiation belt. In addition, the results, depending on the extrapolation scheme used, indicate that the neutron dose equivalent rate may be as high as 0.1 mSv/h near the geomagnetic north pole and thus a significant contributor to the radiation exposures of pilots, flight attendants and the general public. (author)
Directory of Open Access Journals (Sweden)
Jiangpeng Chen
Full Text Available Influenza is a contagious disease with high transmissibility to spread around the world with considerable morbidity and mortality and presents an enormous burden on worldwide public health. Few mathematical models can be used because influenza incidence data are generally not normally distributed. We developed a mathematical model using Extreme Value Theory (EVT to forecast the probability of outbreak of highly pathogenic influenza.The incidence data of highly pathogenic influenza in Zhejiang province from April 2009 to November 2013 were retrieved from the website of Health and Family Planning Commission of Zhejiang Province. MATLAB "VIEM" toolbox was used to analyze data and modelling. In the present work, we used the Peak Over Threshold (POT model, assuming the frequency as a Poisson process and the intensity to be Pareto distributed, to characterize the temporal variability of the long-term extreme incidence of highly pathogenic influenza in Zhejiang, China.The skewness and kurtosis of the incidence of highly pathogenic influenza in Zhejiang between April 2009 and November 2013 were 4.49 and 21.12, which indicated a "fat tail" distribution. A QQ plot and a mean excess plot were used to further validate the features of the distribution. After determining the threshold, we modeled the extremes and estimated the shape parameter and scale parameter by the maximum likelihood method. The results showed that months in which the incidence of highly pathogenic influenza is about 4462/2286/1311/487 are predicted to occur once every five/three/two/one year, respectively.Despite the simplicity, the present study successfully offers the sound modeling strategy and a methodological avenue to implement forecasting of an epidemic in the midst of its course.
Sanz-Elorza, Mario; Dana, Elías D; González, Alberto; Sobrino, Eduardo
2003-08-01
Aerial images of the high summits of the Spanish Central Range reveal significant changes in vegetation over the period 1957 to 1991. These changes include the replacement of high-mountain grassland communities dominated by Festuca aragonensis, typical of the Cryoro-Mediterranean belt, by shrub patches of Juniperus communis ssp. alpina and Cytisus oromediterraneus from lower altitudes (Oro-Mediterranean belt). Climatic data indicate a shift towards warmer conditions in this mountainous region since the 1940s, with the shift being particularly marked from 1960. Changes include significantly higher minimum and maximum temperatures, fewer days with snow cover and a redistribution of monthly rainfall. Total yearly precipitation showed no significant variation. There were no marked changes in land use during the time frame considered, although there were minor changes in grazing species in the 19th century. It is hypothesized that the advance of woody species into higher altitudes is probably related to climate change, which could have acted in conjunction with discrete variations in landscape management. The pronounced changes observed in the plant communities of the area reflect the susceptibility of high-mountain Mediterranean species to environmental change.
Hazama, Gen-i; Inoue, Yuichi; Kojima, Kazushige; Ueta, Toshiyuki; Nakagome, Kazuyuki
2008-09-01
Delayed sleep phase syndrome (DSPS) is a circadian rhythm sleep disorder with a typical onset in the second decade of life. DSPS is characterized by the sleep-onset insomnia and the difficulty in waking at the desired time in the morning. Although DSPS is associated with inability to attend school, the prevalence has been controversial. To elucidate a change in the prevalence of DSPS among young population, epidemiological survey was conducted on Japanese students. A total of 4,971 students of junior high school, senior high school, and university were enrolled in this cross sectional study in Tottori Prefecture. They answered anonymous screening questionnaire regarding school schedule, sleep hygiene and symptomatic items of sleep disorders. The prevalence of probable DSPS was estimated at 0.48% among the total subject students without gender difference. In university, the prevalence of the last year students showed the highest value (1.66%), while that of the first year students showed the lowest value (0.09%) among all school years from junior high school to university. The prevalence increased with advancing university school years. Thus, a considerable number of Japanese students are affected with DSPS. Senior students of university are more vulnerable to the disorder than younger students. Appropriate school schedule may decrease the mismatch between the individual's sleep-wake cycle and the school schedule. Promotion of a regular sleep habit is necessary to prevent DSPS among this population.
Hui, Katrina; Patel, Kashyap; Kong, David C M; Kirkpatrick, Carl M J
2017-07-01
Clearance of small molecules such as amoxicillin and clavulanic acid is expected to increase during high-flux haemodialysis, which may result in lower concentrations and thus reduced efficacy. To date, clearance of amoxicillin/clavulanic acid (AMC) during high-flux haemodialysis remains largely unexplored. Using published pharmacokinetic parameters, a two-compartment model with first-order input was simulated to investigate the impact of high-flux haemodialysis on the probability of target attainment (PTA) of orally administered AMC combination therapy. The following pharmacokinetic/pharmacodynamic targets were used to calculate the PTA. For amoxicillin, the time that the free concentration remains above the minimum inhibitory concentration (MIC) of ≥50% of the dosing period (≥50%ƒT >MIC ) was used. For clavulanic acid, the time that the free concentration was >0.1 mg/L of ≥45% of the dosing period (≥45%ƒT >0.1 mg/L ) was used. Dialysis clearance reported in low-flux haemodialysis for both compounds was doubled to represent the likely clearance during high-flux haemodialysis. Monte Carlo simulations were performed to produce concentration-time profiles over 10 days in 1000 virtual patients. Seven different regimens commonly seen in clinical practice were explored. When AMC was dosed twice daily, the PTA was mostly ≥90% for both compounds regardless of when haemodialysis commenced. When administered once daily, the PTA was 20-30% for clavulanic acid and ≥90% for amoxicillin. The simulations suggest that once-daily orally administered AMC in patients receiving high-flux haemodialysis may result in insufficient concentrations of clavulanic acid to effectively treat infections, especially on days when haemodialysis occurs. Copyright © 2017 Elsevier B.V. and International Society of Chemotherapy. All rights reserved.
Quick Look Report for Chemical Reactivity Modeling of Various Multi-Canister Overpack Breaches
Energy Technology Data Exchange (ETDEWEB)
Bratton, Robert Lawrence
2002-04-01
uranium oxide coating covers the exposed uranium metal, yet uranium hydride can still form under the protective oxide coating over the 40-year interim storage time span. The current treatment process at Hanford does not remove chemically bound water contained in the hydrates or in the waters of hydration. The chemically bound water is the source material for hydrogen production over the 40-year storage time. So, additional uranium hydride creates concerns that breaches of an MCO with the appropriate size openings could result in the onset of bulk uranium oxidation with the potential of a self-sustaining thermal excursion or pyrophoric event. For this analysis, the worst-case scenario appears to be the match head configuration in a vertically standing MCO, where all the reactive surface area is placed on the tips of the fuel elements. This configuration concentrates the heat-producing chemical reaction at the tips of the fuel elements. Because no mechanistic drop analysis has been performed at this time to determine the MCO failure modes, parametric breach configurations were chosen in this analysis to determine the MCOs’ external thermal response range. The first breach is a pair of holes that suddenly open in the MCO wall. This thermal excursion is controlled by the “thermal chimney effect” in the 4.27-m (14-ft) tall canisters caused by the multiple holes breach (one high and one low). A second breach where the MCO lid is suddenly removed and exposed to the ambient air
CONTAINMENT EVALUATION OF BREACHED AL-SNF FOR CASK TRANSPORT
International Nuclear Information System (INIS)
Vinson, D. W.; Sindelar, R. L.; Iyer, N. C.
2005-01-01
Aluminum-based spent nuclear fuel (Al-SNF) from foreign and domestic research reactors (FRR/DRR) is being shipped to the Savannah River Site. To enter the U.S., the cask with loaded fuel must be certified to comply with the requirements in the Title 10 of the U.S. Code of Federal Regulations, Part 71. The requirements include demonstration of containment of the cask with its contents under normal and accident conditions. Al-SNF is subject to corrosion degradation in water storage, and many of the fuel assemblies are ''failed'' or have through-clad damage. A methodology has been developed with technical bases to show that Al-SNF with cladding breaches can be directly transported in standard casks and maintained within the allowable release rates. The approach to evaluate the limiting allowable leakage rate, L R , for a cask with breached Al-SNF for comparison to its test leakage rate could be extended to other nuclear material systems. The approach for containment analysis of Al-SNF follows calculations for commercial spent fuel as provided in NUREG/CR-6487 that adopts ANSI N14.5 as a methodology for containment analysis. The material-specific features and characteristics of damaged Al-SNF (fuel materials, fabrication techniques, microstructure, radionuclide inventory, and vapor corrosion rates) that were derived from literature sources and/or developed in laboratory testing are applied to generate the four containment source terms that yield four separate cask cavity activity densities; namely, those from fines; gaseous fission product species; volatile fission product species; and fuel assembly crud. The activity values, A 2 , are developed per the guidance of 10CFR71. The analysis is performed parametrically to evaluate maximum number of breached assemblies and exposed fuel area for a proposed shipment in a cask with a test leakage rate
Security breaches: tips for assessing and limiting your risks.
Coons, Leeanne R
2011-01-01
As part of their compliance planning, medical practices should undergo a risk assessment to determine any vulnerability within the practice relative to security breaches. Practices should also implement safeguards to limit their risks. Such safeguards include facility access controls, information and electronic media management, use of business associate agreements, and education and enforcement. Implementation of specific policies and procedures to address security incidents is another critical step that medical practices should take as part of their security incident prevention plan. Medical practices should not only develop policies and procedures to prevent, detect, contain, and correct security violations, but should make sure that such policies and procedures are actually implemented in their everyday operations.
Cameron, Eric; Dyke, Jonathan P; Hernandez, Nora; Louis, Elan D; Dydak, Ulrike
2018-03-10
Essential tremor (ET) is increasingly recognized as a multi-dimensional disorder with both motor and non-motor features. For this reason, imaging studies are more broadly examining regions outside the cerebellar motor loop. Reliable detection of cerebral gray matter (GM) atrophy requires optimized processing, adapted to high-resolution magnetic resonance imaging (MRI). We investigated cerebral GM volume loss in ET cases using automated segmentation of MRI T1-weighted images. MRI was acquired on 47 ET cases and 36 controls. Automated segmentation and voxel-wise comparisons of volume were performed using Statistical Parametric Mapping (SPM) software. To improve upon standard protocols, the high-resolution International Consortium for Brain Mapping (ICBM) 2009a atlas and tissue probability maps were used to process each subject image. Group comparisons were performed: all ET vs. Controls, ET with head tremor (ETH) vs. Controls, and severe ET vs. An analysis of variance (ANOVA) was performed between ET with and without head tremor and controls. Age, sex, and Montreal Cognitive Assessment (MoCA) score were regressed out from each comparison. We were able to consistently identify regions of cerebral GM volume loss in ET and in ET subgroups in the posterior insula, superior temporal gyri, cingulate cortex, inferior frontal gyri and other occipital and parietal regions. There were no significant increases in GM volume in ET in any comparisons with controls. This study, which uses improved methodologies, provides evidence that GM volume loss in ET is present beyond the cerebellum, and in fact, is widespread throughout the cerebrum as well. Copyright © 2018 Elsevier Ltd. All rights reserved.
Conscientiousness and reactions to psychological contract breach: a longitudinal field study.
Orvis, Karin A; Dudley, Nicole M; Cortina, Jose M
2008-09-01
The authors examined the role of employee conscientiousness as a moderator of the relationships between psychological contract breach and employee behavioral and attitudinal reactions to the breach. They collected data from 106 newly hired employees within the 1st month of employment (Time 1), 3 months later (Time 2), and 8 months after Time 1 (Time 3) to observe the progression through contract development, breach, and reaction. Results suggest that conscientiousness is a significant moderator for 4 of the 5 contract breach-employee reaction relationships examined (turnover intentions, organizational loyalty, job satisfaction, and 1 of 2 facets of job performance). Specifically, employees who were lower in conscientiousness had more negative reactions to perceived breach with respect to turnover intentions, organizational loyalty, and job satisfaction. In contrast, employees who were higher in conscientiousness reduced their job performance to a greater degree in response to contract breach. Future research directions are discussed.
Ng, Thomas W H; Feldman, Daniel C; Lam, Simon S K
2010-07-01
This study examined the relationships among psychological contract breaches, organizational commitment, and innovation-related behaviors (generating, spreading, implementing innovative ideas at work) over a 6-month period. Results indicate that the effects of psychological contract breaches on employees are not static. Specifically, perceptions of psychological contract breaches strengthened over time and were associated with decreased levels of affective commitment over time. Further, increased perceptions of psychological contract breaches were associated with decreases in innovation-related behaviors. We also found evidence that organizational commitment mediates the relationship between psychological contract breaches and innovation-related behaviors. These results highlight the importance of examining the nomological network of psychological contract breaches from a change perspective.
2016-01-01
ABP breaches Quantity changes > 5% PAUC growth Pre- Milestone C > 5% APUC growth 100 60 80 40 0 % 0 2012 ABP breaches Quantity changes > 5% PAUC growth...Pre- Milestone C > 5% APUC growth 100 60 80 40 % 2003 ABP breaches Nunn- McCurdy breaches Nunn- McCurdy breaches Nunn- McCurdy breaches Nunn- McCurdy...Quantity changes > 5% PAUC growth > 5% APUC growth Pre- Milestone C% 100 60 80 40 0 2004 ABP breaches Quantity changes > 5%
Tortious Interference with Contract versus "Efficient" Breach: Theory and Empirical Evidence.
McChesney, Fred S
1999-01-01
Tortious interference is bothersome, normatively and positively, to scholars espousing the economic model of "efficient breach" of contract because it penalizes third-party inducements to breach. Scholars nonetheless find innovative second-best arguments to justify the coexistence of tortious interference with "efficient" breach. This article shows normatively why tortious interference would be part of a first-best legal system. Tortious interference provides property protection to contract r...
Breaching confidentiality: medical mandatory reporting laws in Iran.
Milanifar, Alireza; Larijani, Bagher; Paykarzadeh, Parvaneh; Ashtari, Golanna; Mehdi Akhondi, Mohammad
2014-01-01
Medical ethics is a realm where four important subjects of philosophy, medicine, theology and law are covered. Physicians and philosophers cooperation in this area will have great efficiency in the respective ethical rules formation. In addition to respect the autonomy of the patient, physician's obligation is to ensure that the medical intervention has benefit for the patient and the harm is minimal. There is an obvious conflict between duty of confidentiality and duty of mandatory reporting. Professional confidentiality is one of the basic components in building a constant physician-patient relationship which nowadays, beside the novelty, it is the subject of discussion. Legal obligation of confidentiality is not absolute. In physician-patient relationship, keeping patient's secrets and maintaining confidentiality is a legal and ethical duty, and disclosure of such secrets is mainly through specific statutes. Thus, there are a number of situations where breach of confidentiality is permitted in different legal systems. One of the situations where breaching confidentiality is permitted is the medical mandatory reporting to the relevant authority which is in accordance with many countries' legal systems. Some situations are considered in many countries legal systems' such as notification of births and deaths, infectious diseases, child abuse, sport and relevant events, medical errors, drug side effects and dangerous pregnancies. In this paper, we will examine and discuss medical mandatory reporting and its ethical and legal aspects in the judicial and legal system of Iran and few other countries. Finally we will suggest making Medical Mandatory Reporting Law in Iran.
Burnik, Jelena
2012-01-01
In 2011 Sony suffered an extensive breach in its online game network that led to the theft of account data of 77 million users from all over the world. This was one of the largest internet security break-ins that resulted in a large scale personal data breach. As an answer to numerous incidents of security breaches where personal data have been compromised, an instrument of mandatory data breach notification is currently being implemented in the European Union that follows the approach taken ...
Gupta, Vishal; Agarwal, Upasna A; Khatri, Naresh
2016-11-01
This study examines the factors that mediate and moderate the relationships of perceived organizational support with work engagement and organization citizenship behaviour. Specifically, affective commitment is posited to mediate and psychological contract breach to moderate the above relationships. Nurses play a critical role in delivering exemplary health care. For nurses to perform at their best, they need to experience high engagement, which can be achieved by providing them necessary organizational support and proper working environment. Data were collected via a self-reported survey instrument. A questionnaire was administered to a random sample of 750 nurses in nine large hospitals in India during 2013-2014. Four hundred and seventy-five nurses (63%) responded to the survey. Hierarchical multiple regression was used for statistical analysis of the moderated-mediation model. Affective commitment was found to mediate the positive relationships between perceived organizational support and work outcomes (work engagement, organizational citizenship behaviour). The perception of unfulfilled expectations (psychological contract breach) was found to moderate the perceived organizational support-work outcome relationships adversely. The results of this study indicate that perceived organizational support exerts its influence on work-related outcomes and highlight the importance of taking organizational context, such as perceptions of psychological contract breach, into consideration when making sense of the influence of perceived organizational support on affective commitment, work engagement and citizenship behaviours of nurses. © 2016 John Wiley & Sons Ltd.
MINIMIZING GLOVEBOX GLOVE BREACHES, PART IV: CONTROL CHARTS
International Nuclear Information System (INIS)
Cournoyer, Michael E.; Lee, Michelle B.; Schreiber, Stephen B.
2007-01-01
At the Los Alamos National Laboratory (LANL) Plutonium Facility, plutonium. isotopes and other actinides are handled in a glovebox environment. The spread of radiological contamination, and excursions of contaminants into the worker's breathing zone, are minimized and/or prevented through the use of glovebox technology. Evaluating the glovebox configuration, the glovebo gloves are the most vulnerable part of this engineering control. Recognizing this vulnerability, the Glovebox Glove Integrity Program (GGIP) was developed to minimize and/or prevent unplanned openings in the glovebox environment, i.e., glove failures and breaches. In addition, LANL implement the 'Lean Six Sigma (LSS)' program that incorporates the practices of Lean Manufacturing and Six Sigma technologies and tools to effectively improve administrative and engineering controls and work processes. One tool used in LSS is the use of control charts, which is an effective way to characterize data collected from unplanned openings in the glovebox environment. The benefit management receives from using this tool is two-fold. First, control charts signal the absence or presence of systematic variations that result in process instability, in relation to glovebox glove breaches and failures. Second, these graphical representations of process variation detennine whether an improved process is under control. Further, control charts are used to identify statistically significant variations (trends) that can be used in decision making to improve processes. This paper discusses performance indicators assessed by the use control charts, provides examples of control charts, and shows how managers use the results to make decisions. This effort contributes to LANL Continuous Improvement Program by improving the efficiency, cost effectiveness, and formality of glovebox operations.
Bouncing back from psychological contract breach: How commitment recovers over time
Solinger, O.N.; Hofmans, J.; Bal, P.M.; Jansen, P.G.W.
2016-01-01
The post-violation model of the psychological contract outlines four ways in which a psychological contract may be resolved after breach (i.e., psychological contract thriving, reactivation, impairment, and dissolution). To explore the implications of this model for post-breach restoration of
A Longitudinal Study of Age-Related Differences in Reactions to Psychological Contract Breach
Bal, P.M.; Lange, A.H. de; Jansen, P.G.W.; Velde, M.E.G. van der
2013-01-01
The current paper investigated age-related differences in the relations of psychological contract breach with work outcomes over time. Based on affective events theory, we expected job satisfaction to mediate the longitudinal relationship of contract breach with changes in job performance. Moreover,
Bal, P.M.; Smit, P.
2012-01-01
The aim of this paper was to investigate the role of emotion regulation and age in reactions to psychological contract breach towards positive and negative affect. The authors expected that in the context of contract breach, reappraisal emotion regulation mitigate the negative relation with affect.
A longitudinal study of age-related differences in reactions to psychological contract breach
Bal, P.M.; de Lange, A.H.; Jansen, P.G.W.; van der Velde, E.G.
2013-01-01
The current paper investigated age-related differences in the relations of psychological contract breach with work outcomes over time. Based on affective events theory, we expected job satisfaction to mediate the longitudinal relationship of contract breach with changes in job performance. Moreover,
Testing the Differential Effects of Changes in Psychological Contract Breach and Fulfillment
Conway, Neil; Guest, David; Trenberth, Linda
2011-01-01
Rousseau (1989 and elsewhere) argued that a defining feature of psychological contract breach was that once a promise had been broken it could not easily be repaired and therefore that the effects of psychological contract breach outweighed those of psychological contract fulfillment. Using two independent longitudinal surveys, this paper…
Psychological Contract Breach and Job Attitudes: A Meta-Analysis of Age as a Moderator
Bal, P. Matthijs; De Lange, Annet H.; Jansen, Paul G. W.; Van Der Velde, Mandy E. G.
2008-01-01
The aim of this study was to examine the influence of age in the relation between psychological contract breach and the development of job attitudes. Based on affective events, social exchange, and lifespan theory, we hypothesized that (1) psychological contract breach would be related negatively to job attitudes, and (2) that age would moderate…
Psychological contract breach and job attitudes : A meta-analysis of age as a moderator
Bal, P. Matthijs; De lange, Annet H.; Jansen, Paul G. W.; Van der Velde, Mandy E. G.
The aim of this study was to examine the influence of age in the relation between psychological contract breach and the development of job attitudes. Based on affective events, social exchange, and lifespan theory, we hypothesized that (1) psychological contract breach would be related negatively to
Inversion Method for Early Detection of ARES-1 Case Breach Failure
Mackey, Ryan M.; Kulikov, Igor K.; Bajwa, Anupa; Berg, Peter; Smelyanskiy, Vadim
2010-01-01
A document describes research into the problem of detecting a case breach formation at an early stage of a rocket flight. An inversion algorithm for case breach allocation is proposed and analyzed. It is shown how the case breach can be allocated at an early stage of its development by using the rocket sensor data and the output data from the control block of the rocket navigation system. The results are simulated with MATLAB/Simulink software. The efficiency of an inversion algorithm for a case breach location is discussed. The research was devoted to the analysis of the ARES-l flight during the first 120 seconds after the launch and early prediction of case breach failure. During this time, the rocket is propelled by its first-stage Solid Rocket Booster (SRB). If a breach appears in SRB case, the gases escaping through it will produce the (side) thrust directed perpendicular to the rocket axis. The side thrust creates torque influencing the rocket attitude. The ARES-l control system will compensate for the side thrust until it reaches some critical value, after which the flight will be uncontrollable. The objective of this work was to obtain the start time of case breach development and its location using the rocket inertial navigation sensors and GNC data. The algorithm was effective for the detection and location of a breach in an SRB field joint at an early stage of its development.
An Examination of the Explicit Costs of Sensitive Information Security Breaches
Toe, Cleophas Adeodat
2013-01-01
Data security breaches are categorized as loss of information that is entrusted in an organization by its customers, partners, shareholders, and stakeholders. Data breaches are significant risk factors for companies that store, process, and transmit sensitive personal information. Sensitive information is defined as confidential or proprietary…
a longitudinal study of age-related differences in reactions to phsycological contract breach
Paul Jansen; Annet de Lange; Matthijs Bal; Mandy van der Velde
2013-01-01
The current paper investigated age‐related differences in the relations of psychological contract breach with work outcomes over time. Based on affective events theory, we expected job satisfaction to mediate the longitudinal relationship of contract breach with changes in job performance. Moreover,
Modelling dune erosion, overwash and breaching at Fire Island (NY) during hurricane Sandy
De Vet, P.L.M.; McCall, R.T.; Den Bieman, J.P.; Stive, M.J.F.; Van Ormondt, M.
2015-01-01
In 2012, Hurricane Sandy caused a breach at Fire Island (NY, USA), near Pelican Island. This paper aims at modelling dune erosion, overwash and breaching processes that occured during the hurricane event at this stretch of coast with the numerical model XBeach. By using the default settings, the
CSIR Research Space (South Africa)
Botha, J
2016-03-01
Full Text Available breaches but does not provide a clear indication of the level of personal information available on the internet since only reported incidents are taken into account. The possibility of pro-active automated breach detection has previously been discussed as a...
Peirce, Gretchen L; Desselle, Shane P; Draugalis, JoLaine R; Spies, Alan R; Davis, Tamra S; Bolino, Mark
2012-08-10
To identify pharmacy faculty members' perceptions of psychological contract breaches that can be used to guide improvements in faculty recruitment, retention, and development. A list of psychological contract breaches was developed using a Delphi procedure involving a panel of experts assembled through purposive sampling. The Delphi consisted of 4 rounds, the first of which elicited examples of psychological contract breaches in an open-ended format. The ensuing 3 rounds consisting of a survey and anonymous feedback on aggregated group responses. Usable responses were obtained from 11 of 12 faculty members who completed the Delphi procedure. The final list of psychological contract breaches included 27 items, after modifications based on participant feedback in subsequent rounds. The psychological contract breach items generated in this study provide guidance for colleges and schools of pharmacy regarding important aspects of faculty recruitment, retention, and development.
Probability Aggregates in Probability Answer Set Programming
Saad, Emad
2013-01-01
Probability answer set programming is a declarative programming that has been shown effective for representing and reasoning about a variety of probability reasoning tasks. However, the lack of probability aggregates, e.g. {\\em expected values}, in the language of disjunctive hybrid probability logic programs (DHPP) disallows the natural and concise representation of many interesting problems. In this paper, we extend DHPP to allow arbitrary probability aggregates. We introduce two types of p...
Introduction to imprecise probabilities
Augustin, Thomas; de Cooman, Gert; Troffaes, Matthias C M
2014-01-01
In recent years, the theory has become widely accepted and has been further developed, but a detailed introduction is needed in order to make the material available and accessible to a wide audience. This will be the first book providing such an introduction, covering core theory and recent developments which can be applied to many application areas. All authors of individual chapters are leading researchers on the specific topics, assuring high quality and up-to-date contents. An Introduction to Imprecise Probabilities provides a comprehensive introduction to imprecise probabilities, includin
Krofcheck, Daniel J; Hurteau, Matthew D; Scheller, Robert M; Loudermilk, E Louise
2018-02-01
In frequent fire forests of the western United States, a legacy of fire suppression coupled with increases in fire weather severity have altered fire regimes and vegetation dynamics. When coupled with projected climate change, these conditions have the potential to lead to vegetation type change and altered carbon (C) dynamics. In the Sierra Nevada, fuels reduction approaches that include mechanical thinning followed by regular prescribed fire are one approach to restore the ability of the ecosystem to tolerate episodic fire and still sequester C. Yet, the spatial extent of the area requiring treatment makes widespread treatment implementation unlikely. We sought to determine if a priori knowledge of where uncharacteristic wildfire is most probable could be used to optimize the placement of fuels treatments in a Sierra Nevada watershed. We developed two treatment placement strategies: the naive strategy, based on treating all operationally available area and the optimized strategy, which only treated areas where crown-killing fires were most probable. We ran forecast simulations using projected climate data through 2,100 to determine how the treatments differed in terms of C sequestration, fire severity, and C emissions relative to a no-management scenario. We found that in both the short (20 years) and long (100 years) term, both management scenarios increased C stability, reduced burn severity, and consequently emitted less C as a result of wildfires than no-management. Across all metrics, both scenarios performed the same, but the optimized treatment required significantly less C removal (naive=0.42 Tg C, optimized=0.25 Tg C) to achieve the same treatment efficacy. Given the extent of western forests in need of fire restoration, efficiently allocating treatments is a critical task if we are going to restore adaptive capacity in frequent-fire forests. © 2017 John Wiley & Sons Ltd.
I Am So Tired… How Fatigue May Exacerbate Stress Reactions to Psychological Contract Breach.
Achnak, Safâa; Griep, Yannick; Vantilborgh, Tim
2018-01-01
Previous research showed that perceptions of psychological contract (PC) breach have undesirable individual and organizational consequences. Surprisingly, the PC literature has paid little to no attention to the relationship between PC breach perceptions and stress. A better understanding of how PC breach may elicit stress seems crucial, given that stress plays a key role in employees' physical and mental well-being. Based on Conservation of Resources Theory, we suggest that PC breach perceptions represent a perceived loss of valued resources, subsequently leading employees to experience higher stress levels resulting from emerging negative emotions. Moreover, we suggest that this mediated relationship is moderated by initial levels of fatigue, due to fatigue lowering the personal resources necessary to cope with breach events. To tests our hypotheses, we analyzed the multilevel data we obtained from two experience sampling designs (Study 1: 51 Belgian employees; Study 2: 53 US employees). Note that the unit of analysis is "observations" rather than "respondents," resulting in an effective sample size of 730 (Study 1) and 374 (Study 2) observations. In both studies, we found evidence for the mediating role of negative emotions in the PC breach-stress relationship. In the second study, we also found evidence for the moderating role of fatigue in the mediated PC breach-stress relationship. Implications for research and practice are discussed.
Uncertainties and constraints on breaching and their implications for flood loss estimation.
Muir Wood, Robert; Bateman, William
2005-06-15
Around the coasts of the southern North Sea, flood risk is mediated everywhere by the performance of natural and man-made flood defences. Under the conditions of extreme surge with tide water levels, the performance of the defences determines the extent of inland flooding. Sensitivity tests reveal the enormous increase in the volume of water that can pass through a defence once breaching is initiated, with a 1m reduction in sill elevation doubling the loss. Empirical observations of defence performance in major storm surges around the North Sea reveal some of the principal controls on breaching. For the same defence type, the maximum size and depth of a breach is a function of the integral of the hydraulic gradient across the defence, which is in turn determined by the elevation of the floodplain and the degree to which water can continue to flow inland away from the breach. The most extensive and lowest floodplains thereby "generate" the largest breaches. For surges that approach the crest height, the weaker the protection of the defence, the greater the number of breaches. Defence reinforcement reduces both the number and size of the breaches.
Directory of Open Access Journals (Sweden)
Nils Ternès
2017-05-01
Full Text Available Abstract Background Thanks to the advances in genomics and targeted treatments, more and more prediction models based on biomarkers are being developed to predict potential benefit from treatments in a randomized clinical trial. Despite the methodological framework for the development and validation of prediction models in a high-dimensional setting is getting more and more established, no clear guidance exists yet on how to estimate expected survival probabilities in a penalized model with biomarker-by-treatment interactions. Methods Based on a parsimonious biomarker selection in a penalized high-dimensional Cox model (lasso or adaptive lasso, we propose a unified framework to: estimate internally the predictive accuracy metrics of the developed model (using double cross-validation; estimate the individual survival probabilities at a given timepoint; construct confidence intervals thereof (analytical or bootstrap; and visualize them graphically (pointwise or smoothed with spline. We compared these strategies through a simulation study covering scenarios with or without biomarker effects. We applied the strategies to a large randomized phase III clinical trial that evaluated the effect of adding trastuzumab to chemotherapy in 1574 early breast cancer patients, for which the expression of 462 genes was measured. Results In our simulations, penalized regression models using the adaptive lasso estimated the survival probability of new patients with low bias and standard error; bootstrapped confidence intervals had empirical coverage probability close to the nominal level across very different scenarios. The double cross-validation performed on the training data set closely mimicked the predictive accuracy of the selected models in external validation data. We also propose a useful visual representation of the expected survival probabilities using splines. In the breast cancer trial, the adaptive lasso penalty selected a prediction model with 4
Scaling Qualitative Probability
Burgin, Mark
2017-01-01
There are different approaches to qualitative probability, which includes subjective probability. We developed a representation of qualitative probability based on relational systems, which allows modeling uncertainty by probability structures and is more coherent than existing approaches. This setting makes it possible proving that any comparative probability is induced by some probability structure (Theorem 2.1), that classical probability is a probability structure (Theorem 2.2) and that i...
Breached fuel location in FFTF by delayed neutron monitor triangulation
International Nuclear Information System (INIS)
Bunch, W.L.; Tang, E.L.
1985-10-01
The Fast Flux Test Facility (FFTF) features a three-loop, sodium-cooled 400 MWt mixed oxide fueled reactor designed for the irradiation testing of fuels and materials for use in liquid metal cooled fast reactors. To establish the ultimate capability of a particular fuel design and thereby generate information that will lead to improvements, many of the fuel irradiations are continued until a loss of cladding integrity (failure) occurs. When the cladding fails, fission gas escapes from the fuel pin and enters the reactor cover gas system. If the cladding failure permits the primary sodium to come in contact with the fuel, recoil fission products can enter the sodium. The presence of recoil fission products in the sodium can be detected by monitoring for the presence of delayed neutrons in the coolant. It is the present philosophy to not operate FFTF when a failure has occurred that permits fission fragments to enter the sodium. Thus, it is important that the identity and location of the fuel assembly that contains the failed cladding be established in order that it might be removed from the core. This report discusses method of location of fuel element when cladding is breached
Baby milk companies accused of breaching marketing code.
Wise, J
1997-01-18
A consortium of 27 religious and health organizations has released a report entitled "Cracking the Code," which criticizes the bottle-feeding marketing techniques used by Nestle, Gerber, Mead Johnson, Wyeth, and Nutricia. Research for the report was carried out in Thailand, Bangladesh, South Africa, and Poland using a random sample of 800 mothers and 120 health workers in each country. In all 4 sites, women had received information that violated the World Health Organization's 1981 international code of marketing breast milk substitutes. Violations included promoting artificial feeding without recognizing breast feeding as the best source of infant nutrition. The investigation also found that women and health workers in all 4 sites received free samples of artificial milk. The report includes detailed examples of manufacturer representatives making unrequested visits to give product information to mothers, providing incentives to health workers to promote products, and promoting products outside of health care facilities. While the International Association of Infant Food Manufacturers condemned the study as biased, the Nestle company promised to review the allegations contained in the report and to deal with any breaches in the code. The Interagency Group on Breastfeeding Monitoring, which prepared the report, was created in 1994 to provide data to groups supporting a boycott of Nestle for code violations.
Ost, Suzanne
2016-01-01
In this article, I argue that sexual exploitation in the doctor–patient relationship would be dealt with more appropriately by the law in England and Wales on the basis of a breach of fiduciary duty. Three different types of sexual boundary breaches are discussed, and the particular focus is on breaches where the patient's consent is obtained through inducement. I contend that current avenues of redress do not clearly catch this behaviour and, moreover, they fail to capture the essence of the wrong committed by the doctor—the knowing breach of trust for self-gain—and the calculated way in which consent is induced. Finally, I demonstrate that the fiduciary approach is compatible with the contemporary pro-patient autonomy model of the doctor–patient relationship. PMID:26846652
Dimick, Chris
2010-04-01
Effective January 1, 2009, California healthcare providers were required to report every breach of patient information to the state. They have sent a flood of mishaps and a steady stream of malicious acts.
International Nuclear Information System (INIS)
Casella, Andrew M.; Loyalka, Sudarsham K.; Hanson, Brady D.
2007-01-01
By taking the differential forms of transport and equations of state, the equations describing aerosol transport in pinhole breaches can be solved directly using continuous models. The results are compared with discrete models.
Plugging Effects on Depressurization Time in Dry Storage Containers with Pinhole Breaches
International Nuclear Information System (INIS)
Casella, Andrew M.; LOYALKA, SUDARSHAN K.; Hanson, Brady D.
2006-01-01
As continuation on previous work, we now examine the effect that aerosol deposition may have on plugging pinhole breaches in spent fuel containers. A model is developed considering only diffusive settling
Deery, Stephen J; Iverson, Roderick D; Walsh, Janet T
2006-01-01
Experiences of psychological contract breach have been associated with a range of negative behavior. However, much of the research has focused on master of business administration alumni and managers and made use of self-reported outcomes. Studying a sample of customer service employees, the research found that psychological contract breach was related to lower organizational trust, which, in turn was associated with perceptions of less cooperative employment relations and higher levels of absenteeism. Furthermore, perceptions of external market pressures moderated the effect of psychological contract breach on absenteeism. The study indicated that psychological contract breach can arise when employees perceive discrepancies between an organization's espoused behavioral standards and its actual behavioral standards, and this can affect discretionary absence. (c) 2006 APA, all rights reserved.
Are Emotions Transmitted From Work to Family? A Crossover Model of Psychological Contract Breach.
Liang, Huai-Liang
2018-01-01
Based on affective events theory and the crossover model, this study examines the effect of psychological contract breach on employee dysfunctional behavior and partner family undermining and explores the crossover effect of employee dysfunctional behavior on partner family undermining in work-family issues. This study collected 370 employee-partner dyads (277 male employees, 93 female employees, M age = 43.59 years) from a large manufacturing organization. The results of this study support the conception that employees' psychological contract breach results in frustration in the workplace. In addition, mediation analysis results reveal that psychological contract breach relates to employee dysfunctional behavior in the workplace. The findings show that partners' psychological strain mediates the relationship between employee dysfunctional behavior and partner family undermining. Furthermore, these findings provide investigations for the crossover model to display the value of psychological contract breach in family issues.
International Nuclear Information System (INIS)
Roman, M.R.; Angelides, S.; Chen, N.
2000-01-01
Full text: Despite its limitations, V/Q scintigraphy remains the favoured non-invasive technique for the diagnosis of pulmonary embolism (PE). PE is present in 85-90% and 30-40% of high and intermediate probability V/Q studies respectively. The value of spiral CT (SCT), a newer imaging modality, has yet to be determined. The aims of this study were to determine the frequency of positive SCT for PE in high and intermediate probability V/Q studies performed within 24hr apart. 15 patients (6M, 9F, mean age - 70.2) with a high probability study were included. Six (40%) SCT were reported as positive (four with emboli present in the main pulmonary arteries), seven as negative, one equivocal and one was technically sub-optimal. Pulmonary angiography was not performed in any patient. In all seven negative studies, the SCT was performed before the V/Q study. Of these, two studies were revised to positive once the result of the V/Q study was known, while, three others had resolving mismatch V/Q defects on follow-up studies (performed 5-14 days later); two of these three also had a positive duplex scan of the lower limbs. One other was most likely due to chronic thromboembolic disease. Only three patients had a V/Q scan prior to the SCT; all were positive for PE on both imaging modalities. Of 26 patients (11M, 15F, mean age - 68.5) with an intermediate probability V/Q study, SCT was positive in only two (8%). Thus the low detection rate of PE by SCT in this albeit small series, raises doubts as to its role in the diagnosis of PE. Copyright (2000) The Australian and New Zealand Society of Nuclear Medicine Inc
Legal Effect of Breach of Warranty in Construction Insurance in Malaysia
Arazi Idrus; Mahmoud Sodangi; Jamaluddin Yaakob
2011-01-01
This study is aimed at analyzing the legal effect of breach of warranty in construction insurance contracts in Malaysia in light of the current developments in The English insurance law. The required data and information were collected from Malaysian and English court decisions dealing with breach of warranties in English marine insurance law from the online Malayan Law Journal published on the LexisNexis online database and from published textbooks related to insurance warranties. This study...
Cantisano, Gabriela Topa; Domínguez, J Francisco Morales; García, J Luis Caeiro
2007-05-01
This study focuses on the mediator role of social comparison in the relationship between perceived breach of psychological contract and burnout. A previous model showing the hypothesized effects of perceived breach on burnout, both direct and mediated, is proposed. The final model reached an optimal fit to the data and was confirmed through multigroup analysis using a sample of Spanish teachers (N = 401) belonging to preprimary, primary, and secondary schools. Multigroup analyses showed that the model fit all groups adequately.
Buyukyilmaz, Ozan; Cakmak, Ahmet F.
2013-01-01
This study aims to investigate the assumed direct and indirect relationships between psychological contract breach and turnover intention through psychological contract violation and perceived organizational support. Data for the sample was collected from 570 academicians from a variety of universities in Turkey. Hierarchical regression analyses were conducted to test the hypotheses. The results show that psychological contract breach was positively related to turnover intention and psycholog...
Durowoju, Olatunde; Chan, Hing; Wang, Xiaojun
2013-01-01
This study draws on information theory and aims to provide simulated evidence using real historical and statistical data to demonstrate how various levels of integration moderate the impact and uncertainties of information security breach on supply chain performance. We find that the supply chain behaves differently under various levels of integration when a security breach occurs. The entropy analysis revealed that the wholesaler experience the most uncertainty under system failure and data ...
Energy Technology Data Exchange (ETDEWEB)
Lee, Byung Chul; Hong, Soon Joon; Lee, Jin Yong; Lee, Kyung Jin; Lee, Kuh Hyung [FNC Tech. Co., Seoul (Korea, Republic of)
2008-04-15
Existing MELCOR 1.8.5 model was improved in view of severe accident natural circulation and MELCOR 1.8.6 input model was developed and calculation sheets for detailed MELCOR 1.8.6 model were produced. Effects of natural circulation modeling were found by simulating SBO accident by comparing existing model with detailed model. Major phenomenon and system operations which affect on natural circulation by high temperature and high pressure gas were investigated and representative accident sequences for creep rupture model of RCS pipeline and SG tube were selected.
Briggs, William M.
2012-01-01
The probability leakage of model M with respect to evidence E is defined. Probability leakage is a kind of model error. It occurs when M implies that events $y$, which are impossible given E, have positive probability. Leakage does not imply model falsification. Models with probability leakage cannot be calibrated empirically. Regression models, which are ubiquitous in statistical practice, often evince probability leakage.
Gordon, Allegra R.; Conron, Kerith J.; Calzo, Jerel P.; White, Matthew T.; Reisner, Sari L.; Austin, S. Bryn
2018-01-01
Background: Young people may experience school-based violence and bullying victimization related to their gender expression, independent of sexual orientation identity. However, the associations between gender expression and bullying and violence have not been examined in racially and ethnically diverse population-based samples of high school…
Ribeiro, Jose Pedro N.; Saggio, Ângelo; Lima, Maria Inês Salgueiro
2013-04-01
Artificial sandbar opening of intermittently open estuaries is a practice utilised worldwide to improve water quality, fishing, and recreational amenities and to prevent the flooding of adjacent properties. Breaching causes the water level to drop drastically, exposing plants to two water level extremes. With some exceptions, estuarine communities are adversely affected by this practice. Although breaching can happen naturally, artificial breaching is on the rise, and the impact of manipulating water levels on estuarine communities needs to be investigated. In this work, we described the breaching cycles of the Massaguaçu River Estuary and proposed flooding scenarios for the estuary's macrophyte banks based on our data. We calculated the relationship between plant distribution and flooding conditions and used our calculations to predict the estuary community's composition depending on the water level at breaching time. We discovered a strong relationship between plant distribution and flooding conditions, and we predicted that the estuarine community would be markedly different between flooding scenarios. Low frequency flooding scenarios would be related to submerged macrophytes and, as the flooding frequency increases, macrophytes would be replaced by amphibious plants, and eventually by the arboreal stratus. Therefore, we concluded that an increase in artificial breaching cycles would have a detrimental impact on the estuary community.
Galarraga, Ibon; Sainz de Murieta, Elisa; Markandya, Anil; María Abadie, Luis
2018-02-01
This addendum adds to the analysis presented in ‘Understanding risks in the light of uncertainty: low-probability, high-impact coastal events in cities’ Abadie et al (2017 Environ. Res. Lett. 12 014017). We propose to use the framework developed earlier to enhance communication and understanding of risks, with the aim of bridging the gap between highly technical risk management discussion to the public risk aversion debate. We also propose that the framework could be used for stress-testing resilience.
Probability of expected climate stresses in North America in the next one My
International Nuclear Information System (INIS)
Kukla, G.
1979-01-01
Climates one million years ahead were predicted upon the assumption that the natural climate variability during the past My will continue. Response of environment and climate in the Basin and Range province of the western USA to global fluctuations was reconstructed; the most remarkable change was the filling of closed basins with large freshwater lakes. Probabilities of permanent ice cover and floods are discussed. It is believed that a site with minimal probability of climate-related breach can be selected
Jia, Lei; Li, Lin; Gui, Tao; Liu, Siyang; Li, Hanping; Han, Jingwan; Guo, Wei; Liu, Yongjian; Li, Jingyun
2016-09-21
With increasing data on HIV-1, a more relevant molecular model describing mechanism details of HIV-1 genetic recombination usually requires upgrades. Currently an incomplete structural understanding of the copy choice mechanism along with several other issues in the field that lack elucidation led us to perform an analysis of the correlation between breakpoint distributions and (1) the probability of base pairing, and (2) intersubtype genetic similarity to further explore structural mechanisms. Near full length sequences of URFs from Asia, Europe, and Africa (one sequence/patient), and representative sequences of worldwide CRFs were retrieved from the Los Alamos HIV database. Their recombination patterns were analyzed by jpHMM in detail. Then the relationships between breakpoint distributions and (1) the probability of base pairing, and (2) intersubtype genetic similarities were investigated. Pearson correlation test showed that all URF groups and the CRF group exhibit the same breakpoint distribution pattern. Additionally, the Wilcoxon two-sample test indicated a significant and inexplicable limitation of recombination in regions with high pairing probability. These regions have been found to be strongly conserved across distinct biological states (i.e., strong intersubtype similarity), and genetic similarity has been determined to be a very important factor promoting recombination. Thus, the results revealed an unexpected disagreement between intersubtype similarity and breakpoint distribution, which were further confirmed by genetic similarity analysis. Our analysis reveals a critical conflict between results from natural HIV-1 isolates and those from HIV-1-based assay vectors in which genetic similarity has been shown to be a very critical factor promoting recombination. These results indicate the region with high-pairing probabilities may be a more fundamental factor affecting HIV-1 recombination than sequence similarity in natural HIV-1 infections. Our
Koo, Reginald; Jones, Martin L.
2011-01-01
Quite a number of interesting problems in probability feature an event with probability equal to 1/e. This article discusses three such problems and attempts to explain why this probability occurs with such frequency.
Goldberg, Samuel
1960-01-01
Excellent basic text covers set theory, probability theory for finite sample spaces, binomial theorem, probability distributions, means, standard deviations, probability function of binomial distribution, more. Includes 360 problems with answers for half.
Quantum probability measures and tomographic probability densities
Amosov, GG; Man'ko, [No Value
2004-01-01
Using a simple relation of the Dirac delta-function to generalized the theta-function, the relationship between the tomographic probability approach and the quantum probability measure approach with the description of quantum states is discussed. The quantum state tomogram expressed in terms of the
Applied probability and stochastic processes
Sumita, Ushio
1999-01-01
Applied Probability and Stochastic Processes is an edited work written in honor of Julien Keilson. This volume has attracted a host of scholars in applied probability, who have made major contributions to the field, and have written survey and state-of-the-art papers on a variety of applied probability topics, including, but not limited to: perturbation method, time reversible Markov chains, Poisson processes, Brownian techniques, Bayesian probability, optimal quality control, Markov decision processes, random matrices, queueing theory and a variety of applications of stochastic processes. The book has a mixture of theoretical, algorithmic, and application chapters providing examples of the cutting-edge work that Professor Keilson has done or influenced over the course of his highly-productive and energetic career in applied probability and stochastic processes. The book will be of interest to academic researchers, students, and industrial practitioners who seek to use the mathematics of applied probability i...
Toward a generalized probability theory: conditional probabilities
International Nuclear Information System (INIS)
Cassinelli, G.
1979-01-01
The main mathematical object of interest in the quantum logic approach to the foundations of quantum mechanics is the orthomodular lattice and a set of probability measures, or states, defined by the lattice. This mathematical structure is studied per se, independently from the intuitive or physical motivation of its definition, as a generalized probability theory. It is thought that the building-up of such a probability theory could eventually throw light on the mathematical structure of Hilbert-space quantum mechanics as a particular concrete model of the generalized theory. (Auth.)
Current status of the Run-Beyond-Cladding Breach (RBCB) tests for the Integral Fast Reactor (IFR)
International Nuclear Information System (INIS)
Batte, G.L.; Pahl, R.G.; Hofman, G.L.
1993-01-01
This paper describes the results from the Integral Fast Reactor (IFR) metallic fuel Run-Beyond-Cladding-Breach (RBCB) experiments conducted in the Experimental Breeder Reactor II (EBR-II). Included in the report are scoping test results and the data collected from the prototypical tests as well as the exam results and discussion from a naturally occurring breach of one of the lead IFR fuel tests. All results showed a characteristic delayed neutron and fission gas release pattern that readily allows for identification and evaluation of cladding breach events. Also, cladding breaches are very small and do not propagate during extensive post breach operation. Loss of fuel from breached cladding was found to be insignificant. The paper will conclude with a brief description of future RBCB experiments planned for irradiation in EBR-II
I Am So Tired… How Fatigue May Exacerbate Stress Reactions to Psychological Contract Breach
Achnak, Safâa; Griep, Yannick; Vantilborgh, Tim
2018-01-01
Previous research showed that perceptions of psychological contract (PC) breach have undesirable individual and organizational consequences. Surprisingly, the PC literature has paid little to no attention to the relationship between PC breach perceptions and stress. A better understanding of how PC breach may elicit stress seems crucial, given that stress plays a key role in employees' physical and mental well-being. Based on Conservation of Resources Theory, we suggest that PC breach perceptions represent a perceived loss of valued resources, subsequently leading employees to experience higher stress levels resulting from emerging negative emotions. Moreover, we suggest that this mediated relationship is moderated by initial levels of fatigue, due to fatigue lowering the personal resources necessary to cope with breach events. To tests our hypotheses, we analyzed the multilevel data we obtained from two experience sampling designs (Study 1: 51 Belgian employees; Study 2: 53 US employees). Note that the unit of analysis is “observations” rather than “respondents,” resulting in an effective sample size of 730 (Study 1) and 374 (Study 2) observations. In both studies, we found evidence for the mediating role of negative emotions in the PC breach—stress relationship. In the second study, we also found evidence for the moderating role of fatigue in the mediated PC breach—stress relationship. Implications for research and practice are discussed. PMID:29559935
I Am So Tired… How Fatigue May Exacerbate Stress Reactions to Psychological Contract Breach
Directory of Open Access Journals (Sweden)
Safâa Achnak
2018-03-01
Full Text Available Previous research showed that perceptions of psychological contract (PC breach have undesirable individual and organizational consequences. Surprisingly, the PC literature has paid little to no attention to the relationship between PC breach perceptions and stress. A better understanding of how PC breach may elicit stress seems crucial, given that stress plays a key role in employees' physical and mental well-being. Based on Conservation of Resources Theory, we suggest that PC breach perceptions represent a perceived loss of valued resources, subsequently leading employees to experience higher stress levels resulting from emerging negative emotions. Moreover, we suggest that this mediated relationship is moderated by initial levels of fatigue, due to fatigue lowering the personal resources necessary to cope with breach events. To tests our hypotheses, we analyzed the multilevel data we obtained from two experience sampling designs (Study 1: 51 Belgian employees; Study 2: 53 US employees. Note that the unit of analysis is “observations” rather than “respondents,” resulting in an effective sample size of 730 (Study 1 and 374 (Study 2 observations. In both studies, we found evidence for the mediating role of negative emotions in the PC breach—stress relationship. In the second study, we also found evidence for the moderating role of fatigue in the mediated PC breach—stress relationship. Implications for research and practice are discussed.
Do promises matter? An exploration of the role of promises in psychological contract breach.
Montes, Samantha D; Zweig, David
2009-09-01
Promises are positioned centrally in the study of psychological contract breach and are argued to distinguish psychological contracts from related constructs, such as employee expectations. However, because the effects of promises and delivered inducements are confounded in most research, the role of promises in perceptions of, and reactions to, breach remains unclear. If promises are not an important determinant of employee perceptions, emotions, and behavioral intentions, this would suggest that the psychological contract breach construct might lack utility. To assess the unique role of promises, the authors manipulated promises and delivered inducements separately in hypothetical scenarios in Studies 1 (558 undergraduates) and 2 (441 employees), and they measured them separately (longitudinally) in Study 3 (383 employees). The authors' results indicate that breach perceptions do not represent a discrepancy between what employees believe they were promised and were given. In fact, breach perceptions can exist in the absence of promises. Further, promises play a negligible role in predicting feelings of violation and behavioral intentions. Contrary to the extant literature, the authors' findings suggest that promises may matter little; employees are concerned primarily with what the organization delivers.
Directory of Open Access Journals (Sweden)
Kaushi S. T. Kanankege
2018-01-01
Full Text Available Zebra mussels (ZMs (Dreissena polymorpha and Eurasian watermilfoil (EWM (Myriophyllum spicatum are aggressive aquatic invasive species posing a conservation burden on Minnesota. Recognizing areas at high risk for invasion is a prerequisite for the implementation of risk-based prevention and mitigation management strategies. The early detection of invasion has been challenging, due in part to the imperfect observation process of invasions including the absence of a surveillance program, reliance on public reporting, and limited resource availability, which results in reporting bias. To predict the areas at high risk for invasions, while accounting for underreporting, we combined network analysis and probability co-kriging to estimate the risk of ZM and EWM invasions. We used network analysis to generate a waterbody-specific variable representing boater traffic, a known high risk activity for human-mediated transportation of invasive species. In addition, co-kriging was used to estimate the probability of species introduction, using waterbody-specific variables. A co-kriging model containing distance to the nearest ZM infested location, boater traffic, and road access was used to recognize the areas at high risk for ZM invasions (AUC = 0.78. The EWM co-kriging model included distance to the nearest EWM infested location, boater traffic, and connectivity to infested waterbodies (AUC = 0.76. Results suggested that, by 2015, nearly 20% of the waterbodies in Minnesota were at high risk of ZM (12.45% or EWM (12.43% invasions, whereas only 125/18,411 (0.67% and 304/18,411 (1.65% are currently infested, respectively. Prediction methods presented here can support decisions related to solving the problems of imperfect detection, which subsequently improve the early detection of biological invasions.
Kanankege, Kaushi S T; Alkhamis, Moh A; Phelps, Nicholas B D; Perez, Andres M
2017-01-01
Zebra mussels (ZMs) ( Dreissena polymorpha ) and Eurasian watermilfoil (EWM) ( Myriophyllum spicatum ) are aggressive aquatic invasive species posing a conservation burden on Minnesota. Recognizing areas at high risk for invasion is a prerequisite for the implementation of risk-based prevention and mitigation management strategies. The early detection of invasion has been challenging, due in part to the imperfect observation process of invasions including the absence of a surveillance program, reliance on public reporting, and limited resource availability, which results in reporting bias. To predict the areas at high risk for invasions, while accounting for underreporting, we combined network analysis and probability co-kriging to estimate the risk of ZM and EWM invasions. We used network analysis to generate a waterbody-specific variable representing boater traffic, a known high risk activity for human-mediated transportation of invasive species. In addition, co-kriging was used to estimate the probability of species introduction, using waterbody-specific variables. A co-kriging model containing distance to the nearest ZM infested location, boater traffic, and road access was used to recognize the areas at high risk for ZM invasions (AUC = 0.78). The EWM co-kriging model included distance to the nearest EWM infested location, boater traffic, and connectivity to infested waterbodies (AUC = 0.76). Results suggested that, by 2015, nearly 20% of the waterbodies in Minnesota were at high risk of ZM (12.45%) or EWM (12.43%) invasions, whereas only 125/18,411 (0.67%) and 304/18,411 (1.65%) are currently infested, respectively. Prediction methods presented here can support decisions related to solving the problems of imperfect detection, which subsequently improve the early detection of biological invasions.
International Nuclear Information System (INIS)
Batte, G.L.; Hoffman, G.L.
1990-01-01
In 1984 Argonne National Laboratory (ANL) began an aggressive program of research and development based on the concept of a closed system for fast-reactor power generation and on-site fuel reprocessing, exclusively designed around the use of metallic fuel. This is the Integral Fast Reactor (IFR). Although the Experimental Breeder Reactor-II (EBR-II) has used metallic fuel since its creation 25 yeas ago, in 1985 ANL began a study of the characteristics and behavior of an advanced-design metallic fuel based on uranium-zirconium (U-Zr) and uranium-plutonium-zirconium (U-Pu-Zr) alloys. During the past five years several areas were addressed concerning the performance of this fuel system. In all instances of testing the metallic fuel has demonstrated its ability to perform reliably to high burnups under varying design conditions. This paper will present one area of testing which concerns the fuel system's performance under breach conditions. It is the purpose of this paper to document the observed post-breach behavior of this advanced-design metallic fuel. 2 figs., 1 tab
Spering, Cynthia C; Hobson, Valerie; Lucas, John A; Menon, Chloe V; Hall, James R; O'Bryant, Sid E
2012-08-01
To validate and extend the findings of a raised cut score of O'Bryant and colleagues (O'Bryant SE, Humphreys JD, Smith GE, et al. Detecting dementia with the mini-mental state examination in highly educated individuals. Arch Neurol. 2008;65(7):963-967.) for the Mini-Mental State Examination in detecting cognitive dysfunction in a bilingual sample of highly educated ethnically diverse individuals. Archival data were reviewed from participants enrolled in the National Alzheimer's Coordinating Center minimum data set. Data on 7,093 individuals with 16 or more years of education were analyzed, including 2,337 cases with probable and possible Alzheimer's disease, 1,418 mild cognitive impairment patients, and 3,088 nondemented controls. Ethnic composition was characterized as follows: 6,296 Caucasians, 581 African Americans, 4 American Indians or Alaska natives, 2 native Hawaiians or Pacific Islanders, 149 Asians, 43 "Other," and 18 of unknown origin. Diagnostic accuracy estimates (sensitivity, specificity, and likelihood ratio) of Mini-Mental State Examination cut scores in detecting probable and possible Alzheimer's disease were examined. A standard Mini-Mental State Examination cut score of 24 (≤23) yielded a sensitivity of 0.58 and a specificity of 0.98 in detecting probable and possible Alzheimer's disease across ethnicities. A cut score of 27 (≤26) resulted in an improved balance of sensitivity and specificity (0.79 and 0.90, respectively). In the cognitively impaired group (mild cognitive impairment and probable and possible Alzheimer's disease), the standard cut score yielded a sensitivity of 0.38 and a specificity of 1.00 while raising the cut score to 27 resulted in an improved balance of 0.59 and 0.96 of sensitivity and specificity, respectively. These findings cross-validate our previous work and extend them to an ethnically diverse cohort. A higher cut score is needed to maximize diagnostic accuracy of the Mini-Mental State Examination in individuals
Wilcox, Andrew C.; O'Connor, James E.; Major, Jon J.
2014-01-01
Condit Dam on the White Salmon River, Washington, a 38 m high dam impounding a large volume (1.8 million m3) of fine-grained sediment (60% sand, 35% silt and clay, and 5% gravel), was rapidly breached in October 2011. This unique dam decommissioning produced dramatic upstream and downstream geomorphic responses in the hours and weeks following breaching. Blasting a 5 m wide hole into the base of the dam resulted in rapid reservoir drawdown, abruptly releasing ~1.6 million m3 of reservoir water, exposing reservoir sediment to erosion, and triggering mass failures of the thickly accumulated reservoir sediment. Within 90 min of breaching, the reservoir's water and ~10% of its sediment had evacuated. At a gauging station 2.3 km downstream, flow increased briefly by 400 m3 s−1during passage of the initial pulse of released reservoir water, followed by a highly concentrated flow phase—up to 32% sediment by volume—as landslide-generated slurries from the reservoir moved downstream. This hyperconcentrated flow, analogous to those following volcanic eruptions or large landslides, draped the downstream river with predominantly fine sand. During the ensuing weeks, suspended-sediment concentration declined and sand and gravel bed load derived from continued reservoir erosion aggraded the channel by >1 m at the gauging station, after which the river incised back to near its initial elevation at this site. Within 15 weeks after breaching, over 1 million m3 of suspended load is estimated to have passed the gauging station, consistent with estimates that >60% of the reservoir's sediment had eroded. This dam removal highlights the influence of interactions among reservoir erosion processes, sediment composition, and style of decommissioning on rate of reservoir erosion and consequent downstream behavior of released sediment.
Philosophical theories of probability
Gillies, Donald
2000-01-01
The Twentieth Century has seen a dramatic rise in the use of probability and statistics in almost all fields of research. This has stimulated many new philosophical ideas on probability. Philosophical Theories of Probability is the first book to present a clear, comprehensive and systematic account of these various theories and to explain how they relate to one another. Gillies also offers a distinctive version of the propensity theory of probability, and the intersubjective interpretation, which develops the subjective theory.
Jamil, Amber; Raja, Usman; Darr, Wendy
2013-01-01
This research examined the relationships between perceived psychological contract breach, felt violation, and burnout in a sample (n = 361) of employees from various organizations in Pakistan. The moderating role of contract types in these relationships was also tested. Findings supported a positive association between perceived psychological contract breach and felt violation and both were positively related to burnout. Transactional and relational contracts moderated the felt violation-burnout relationship. Scores on relational contract type tended to be higher than for transactional contract type showing some contextual influence.
Benci, Vieri; Horsten, Leon; Wenmackers, Sylvia
We propose an alternative approach to probability theory closely related to the framework of numerosity theory: non-Archimedean probability (NAP). In our approach, unlike in classical probability theory, all subsets of an infinite sample space are measurable and only the empty set gets assigned
Interpretations of probability
Khrennikov, Andrei
2009-01-01
This is the first fundamental book devoted to non-Kolmogorov probability models. It provides a mathematical theory of negative probabilities, with numerous applications to quantum physics, information theory, complexity, biology and psychology. The book also presents an interesting model of cognitive information reality with flows of information probabilities, describing the process of thinking, social, and psychological phenomena.
Di Marca, Salvatore; Cilia, Chiara; Campagna, Andrea; D'Arrigo, Graziella; Abd ElHafeez, Samar; Tripepi, Giovanni; Puccia, Giuseppe; Pisano, Marcella; Mastrosimone, Gianluca; Terranova, Valentina; Cardella, Antonella; Buonacera, Agata; Stancanelli, Benedetta; Zoccali, Carmine; Malatino, Lorenzo
2015-06-01
To assess and compare the diagnostic power for pulmonary embolism (PE) of Wells and revised Geneva scores in two independent cohorts (training and validation groups) of elderly adults hospitalized in a non-emergency department. Prospective clinical study, January 2011 to January 2013. Unit of Internal Medicine inpatients, University of Catania, Italy. Elderly adults (mean age 76 ± 12), presenting with dyspnea or chest pain and with high clinical probability of PE or D-dimer values greater than 500 ng/mL (N = 203), were enrolled and consecutively assigned to a training (n = 101) or a validation (n = 102) group. The clinical probability of PE was assessed using Wells and revised Geneva scores. Clinical examination, D-dimer test, and multidetector computed angiotomography were performed in all participants. The accuracy of the scores was assessed using receiver operating characteristic analyses. PE was confirmed in 46 participants (23%) (24 training group, 22 validation group). In the training group, the area under the receiver operating characteristic curve was 0.91 (95% confidence interval (CI) = 0.85-0.98) for the Wells score and 0.69 (95% CI = 0.56-0.82) for the revised Geneva score (P < .001). These results were confirmed in the validation group (P < .05). The positive (LR+) and negative likelihood ratios (LR-) (two indices combining sensitivity and specificity) of the Wells score were superior to those of the revised Geneva score in the training (LR+, 7.90 vs 1.34; LR-, 0.23 vs 0.66) and validation (LR+, 13.5 vs 1.46; LR-, 0.47 vs 0.54) groups. In high-risk elderly hospitalized adults, the Wells score is more accurate than the revised Geneva score for diagnosing PE. © 2015, Copyright the Authors Journal compilation © 2015, The American Geriatrics Society.
Introduction to probability theory with contemporary applications
Helms, Lester L
2010-01-01
This introduction to probability theory transforms a highly abstract subject into a series of coherent concepts. Its extensive discussions and clear examples, written in plain language, expose students to the rules and methods of probability. Suitable for an introductory probability course, this volume requires abstract and conceptual thinking skills and a background in calculus.Topics include classical probability, set theory, axioms, probability functions, random and independent random variables, expected values, and covariance and correlations. Additional subjects include stochastic process
Probability with applications and R
Dobrow, Robert P
2013-01-01
An introduction to probability at the undergraduate level Chance and randomness are encountered on a daily basis. Authored by a highly qualified professor in the field, Probability: With Applications and R delves into the theories and applications essential to obtaining a thorough understanding of probability. With real-life examples and thoughtful exercises from fields as diverse as biology, computer science, cryptology, ecology, public health, and sports, the book is accessible for a variety of readers. The book's emphasis on simulation through the use of the popular R software language c
De Ruiter, M.; Schaveling, J.; Schalk, R.; Gelder, van D.
2016-01-01
This study examined the impact of two types of psychological contract breach (organizational policies and social atmosphere breach) on resistance to change and engagement in the anticipatory phase of change and assessed whether supervisory informational justice mitigated the negative effects of
Matthijs Bal; Paul Jansen; Annet de Lange; Mandy van der Velde
2013-01-01
The current paper investigated age-related differences in the relations of psychological contract breach with work outcomes over time. Based on affective events theory, we expected job satisfaction to mediate the longitudinal relationship of contract breach with changes in job performance. Moreover,
de Ruiter, M.; Schalk, R.; Schaveling, Jaap; van Gelder, Daniel
This study examined the impact of two types of psychological contract breach (organizational policies and social atmosphere breach) on resistance to change and engagement in the anticipatory phase of change and assessed whether supervisory informational justice mitigated the negative effects of
Cassar, Vincent; Briner, Rob B.
2011-01-01
This study tested the mediating role of violation in the relationship between breach and both affective and continuance commitment and the extent to which this mediating role is moderated by exchange imbalance amongst a sample of 103 sales personnel. Results suggest that violation mediated the relationship between breach and commitment. Also,…
47 CFR 64.2011 - Notification of customer proprietary network information security breaches.
2010-10-01
... information security breaches. 64.2011 Section 64.2011 Telecommunication FEDERAL COMMUNICATIONS COMMISSION... Proprietary Network Information § 64.2011 Notification of customer proprietary network information security... criminal investigation or national security, such agency may direct the carrier not to so disclose or...
No harm done? Assessing risk of harm under the federal breach notification rule.
Dimick, Chris
2010-08-01
Provisions within the HITECH Act require that covered entities notify individuals if their protected health information is breached. However, the current regulation allows an exemption if the risk of harm is slight. Assessing risk can be subjective, and privacy officers have been working to create methods to conduct and document their analyses.
Coaches in the Courtroom: Recovery in Actions for Breach of Employment Contracts.
Graves, Judson
1986-01-01
The rapid hiring and firing of college athletic coaches, the litigation brought in breach of employment contracts, and the special problems presented by coaching contracts have raised hard legal questions about proper methods of contract enforcement and recovery of damages. (MSE)
Liability to disgorge profits upon breach of contract or a delict
Schrage, E.J.H.
2013-01-01
Remedies regarding contract and tort are, generally speaking, concerned with the incidence of liability for loss or damage suffered, whereas the claim in unjust enrichment is said to require that the enrichment has occurred at the expense of the creditor. Consequently claims for breach of contract
Suitable Penalty for Breach of Contract: AFROTC Cadets. A Research Report Submitted to the Faculty.
Reese, Robert D.
A legislative history of financial incentives in the Reserve Officer Training Corps gives perspective to an analysis of present law and policy concerning breach of contract for Air Force ROTC cadets. The changed environment, criticisms of the present law and policy, and the example of three other Western nations with all volunteer militaries are…
Psychological contract breach and work performance: Is social exchange a buffer or an intensifier?
Bal, P.M.; Chiaburu, D.S.; Jansen, P.G.W.
2010-01-01
Purpose: The aim of this paper is to investigate how social exchanges modify the relationship between psychological contract breach and work performance. It aims to present two concurrent hypotheses, based on theoretical interaction effects of social exchanges (conceptualized as social exchange
The Changing Contours of the Psychological Contract: Unpacking Context and Circumstances of Breach
Pate, Judy
2006-01-01
Purpose: The purpose of this paper is to propose a processual framework of psychological contract breach, which maps holistically the interactions among concepts drawn from the trust and justice literature. However, the price of a holistic picture is frequently a lack of depth of analysis of any single variable, and consequently the second part of…
Restubog, Simon Lloyd D.; Bordia, Prashant; Bordia, Sarbari
2011-01-01
The current study extends past research by examining leader-member exchange as a mediator of the relationship between employee reports of psychological contract breach and career success. In addition, we tested a competing perspective in which we proposed that performance mediators (i.e., in-role performance and organizational citizenship…
Chiu, Su-Fen; Peng, Jei-Chen
2008-01-01
This study investigated the main effects and the interaction effects of psychological contract breach and hostile attributional style on employee deviance (i.e., interpersonal deviance and organizational deviance). Data were collected from 233 employees and their supervisors in eight electronic companies in Taiwan. Results demonstrate that…
ECONOMIC AND LEGAL ASPECTS OF THE PLANNED DAMAGES ACTIONS FOR THE BREACHES OF EC ANTITRUST LAW
Directory of Open Access Journals (Sweden)
Elena Isac
2010-09-01
Full Text Available This paper investigates the planned damages actions for breaches of EC antitrust law in order to assess their impact on consumer welfare. It first examines the current legal situation and concurs that the European Union needs to regulate damages actions for breaches of EC antitrust law so that a higher number of consumers could be compensated for their losses. This paper then discusses the main legal provisions proposed by the Commission in the Green and in the White paper on damages actions for breaches of EC antitrust law. The analysis of these proposed legal provisions is done using arguments specific to the economic analysis of law. It is demonstrated that most of these proposed legal provisions will enhance consumer welfare but that there are also proposed legal provisions which will damage consumer welfare. The paper concludes that the planned damages actions for breaches of the EC law will be an improvement compared to the current situation. However, the Commission should amend some of the proposed legal provisions in order to help consumers further.
An inverse method to estimate the flow through a levee breach
D'Oria, Marco; Mignosa, Paolo; Tanda, Maria Giovanna
2015-08-01
We propose a procedure to estimate the flow through a levee breach based on water levels recorded in river stations downstream and/or upstream of the failure site. The inverse problem is solved using a Bayesian approach and requires the execution of several forward unsteady flow simulations. For this purpose, we have used the well-known 1-D HEC-RAS model, but any unsteady flow model could be adopted in the same way. The procedure has been tested using four synthetic examples. Levee breaches with different characteristics (free flow, flow with tailwater effects, etc.) have been simulated to collect the synthetic level data used at a later stage in the inverse procedure. The method was able to accurately reproduce the flow through the breach in all cases. The practicability of the procedure was then confirmed applying it to the inundation of the Polesine Region (Northern Italy) which occurred in 1951 and was caused by three contiguous and almost simultaneous breaches on the left embankment of the Po River.
Beckham, Joseph
1979-01-01
A cause of action for educational malpractice may well receive initial judicial recognition through successfully harmonizing allegations of breach of a statutory duty of care and acts of negligence of a type and magnitude that would distinguish a student-plaintiff's injuries from others for whose benefit the statutory duty was created. (Author)
One after the other : Effects of sequence patterns of breaches and overfulfilled obligations
de Jong, Jeroen; Rigotti, Thomas; Mulder, J.
2017-01-01
To date, the study of psychological contracts has primarily centred on the question how retrospective evaluations of the psychological contract impact employee attitudes and behaviours, and/or focus on individual coping processes in explaining responses to breached or overfulfilled obligations. In
DEFF Research Database (Denmark)
Lookofsky, Joseph
2011-01-01
For every breach of a binding contract, there must be some remedy. The gap-filling remedial structure of the 1980 Vienna Sales Convention (CISG) reflects the fact that all significant forms of remedial relief may be said to fall within three basic courses of action which modern legal systems make...
International Nuclear Information System (INIS)
Fraassen, B.C. van
1979-01-01
The interpretation of probabilities in physical theories are considered, whether quantum or classical. The following points are discussed 1) the functions P(μ, Q) in terms of which states and propositions can be represented, are classical (Kolmogoroff) probabilities, formally speaking, 2) these probabilities are generally interpreted as themselves conditional, and the conditions are mutually incompatible where the observables are maximal and 3) testing of the theory typically takes the form of confronting the expectation values of observable Q calculated with probability measures P(μ, Q) for states μ; hence, of comparing the probabilities P(μ, Q)(E) with the frequencies of occurrence of the corresponding events. It seems that even the interpretation of quantum mechanics, in so far as it concerns what the theory says about the empirical (i.e. actual, observable) phenomena, deals with the confrontation of classical probability measures with observable frequencies. This confrontation is studied. (Auth./C.F.)
Analytic Neutrino Oscillation Probabilities in Matter: Revisited
Energy Technology Data Exchange (ETDEWEB)
Parke, Stephen J. [Fermilab; Denton, Peter B. [Copenhagen U.; Minakata, Hisakazu [Madrid, IFT
2018-01-02
We summarize our recent paper on neutrino oscillation probabilities in matter, explaining the importance, relevance and need for simple, highly accurate approximations to the neutrino oscillation probabilities in matter.
Energy Technology Data Exchange (ETDEWEB)
Oldenburg, Curtis M. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States); Budnitz, Robert J. [Lawrence Berkeley National Lab. (LBNL), Berkeley, CA (United States)
2016-08-31
If Carbon dioxide Capture and Storage (CCS) is to be effective in mitigating climate change, it will need to be carried out on a very large scale. This will involve many thousands of miles of dedicated high-pressure pipelines in order to transport many millions of tonnes of CO_{2} annually, with the CO_{2} delivered to many thousands of wells that will inject the CO_{2} underground. The new CCS infrastructure could rival in size the current U.S. upstream natural gas pipeline and well infrastructure. This new infrastructure entails hazards for life, health, animals, the environment, and natural resources. Pipelines are known to rupture due to corrosion, from external forces such as impacts by vehicles or digging equipment, by defects in construction, or from the failure of valves and seals. Similarly, wells are vulnerable to catastrophic failure due to corrosion, cement degradation, or operational mistakes. While most accidents involving pipelines and wells will be minor, there is the inevitable possibility of accidents with very high consequences, especially to public health. The most important consequence of concern is CO_{2} release to the environment in concentrations sufficient to cause death by asphyxiation to nearby populations. Such accidents are thought to be very unlikely, but of course they cannot be excluded, even if major engineering effort is devoted (as it will be) to keeping their probability low and their consequences minimized. This project has developed a methodology for analyzing the risks of these rare but high-consequence accidents, using a step-by-step probabilistic methodology. A key difference between risks for pipelines and wells is that the former are spatially distributed along the pipe whereas the latter are confined to the vicinity of the well. Otherwise, the methodology we develop for risk assessment of pipeline and well failures is similar and provides an analysis both of the annual probabilities of
The quantum probability calculus
International Nuclear Information System (INIS)
Jauch, J.M.
1976-01-01
The Wigner anomaly (1932) for the joint distribution of noncompatible observables is an indication that the classical probability calculus is not applicable for quantum probabilities. It should, therefore, be replaced by another, more general calculus, which is specifically adapted to quantal systems. In this article this calculus is exhibited and its mathematical axioms and the definitions of the basic concepts such as probability field, random variable, and expectation values are given. (B.R.H)
Choice Probability Generating Functions
DEFF Research Database (Denmark)
Fosgerau, Mogens; McFadden, Daniel L; Bierlaire, Michel
This paper considers discrete choice, with choice probabilities coming from maximization of preferences from a random utility field perturbed by additive location shifters (ARUM). Any ARUM can be characterized by a choice-probability generating function (CPGF) whose gradient gives the choice...... probabilities, and every CPGF is consistent with an ARUM. We relate CPGF to multivariate extreme value distributions, and review and extend methods for constructing CPGF for applications....
Yaraghi, Niam; Gopal, Ram D
2018-03-01
Policy Points: Frequent data breaches in the US health care system undermine the privacy of millions of patients every year-a large number of which happen among business associates of the health care providers that continue to gain unprecedented access to patients' data as the US health care system becomes digitally integrated. Implementation of the HIPAA Omnibus Rules in 2013 has led to a significant decrease in the number of privacy breach incidents among business associates. Frequent data breaches in the US health care system undermine the privacy of millions of patients every year. A large number of such breaches happens among business associates of the health care providers that continue to gain unprecedented access to patients' data as the US health care system becomes digitally integrated. The Omnibus Rules of the Health Insurance Portability and Accountability Act (HIPAA), which were enacted in 2013, significantly increased the regulatory oversight and privacy protection requirements of business associates. The objective of this study is to empirically examine the effects of this shift in policy on the frequency of medical privacy breaches among business associates in the US health care system. The findings of this research shed light on how regulatory efforts can protect patients' privacy. Using publicly available data on breach incidents between October 2009 and August 2017 as reported by the Office for Civil Rights (OCR), we conducted an interrupted time-series analysis and a difference-in-differences analysis to examine the immediate and long-term effects of implementation of HIPAA omnibus rules on the frequency of medical privacy breaches. We show that implementation of the omnibus rules led to a significant reduction in the number of breaches among business associates and prevented 180 privacy breaches from happening, which could have affected nearly 18 million Americans. Implementation of HIPAA omnibus rules may have been a successful federal policy
Probability of satellite collision
Mccarter, J. W.
1972-01-01
A method is presented for computing the probability of a collision between a particular artificial earth satellite and any one of the total population of earth satellites. The collision hazard incurred by the proposed modular Space Station is assessed using the technique presented. The results of a parametric study to determine what type of satellite orbits produce the greatest contribution to the total collision probability are presented. Collision probability for the Space Station is given as a function of Space Station altitude and inclination. Collision probability was also parameterized over miss distance and mission duration.
Choice probability generating functions
DEFF Research Database (Denmark)
Fosgerau, Mogens; McFadden, Daniel; Bierlaire, Michel
2013-01-01
This paper considers discrete choice, with choice probabilities coming from maximization of preferences from a random utility field perturbed by additive location shifters (ARUM). Any ARUM can be characterized by a choice-probability generating function (CPGF) whose gradient gives the choice...... probabilities, and every CPGF is consistent with an ARUM. We relate CPGF to multivariate extreme value distributions, and review and extend methods for constructing CPGF for applications. The choice probabilities of any ARUM may be approximated by a cross-nested logit model. The results for ARUM are extended...
Florescu, Ionut
2013-01-01
THE COMPLETE COLLECTION NECESSARY FOR A CONCRETE UNDERSTANDING OF PROBABILITY Written in a clear, accessible, and comprehensive manner, the Handbook of Probability presents the fundamentals of probability with an emphasis on the balance of theory, application, and methodology. Utilizing basic examples throughout, the handbook expertly transitions between concepts and practice to allow readers an inclusive introduction to the field of probability. The book provides a useful format with self-contained chapters, allowing the reader easy and quick reference. Each chapter includes an introductio
Ash, Robert B; Lukacs, E
1972-01-01
Real Analysis and Probability provides the background in real analysis needed for the study of probability. Topics covered range from measure and integration theory to functional analysis and basic concepts of probability. The interplay between measure theory and topology is also discussed, along with conditional probability and expectation, the central limit theorem, and strong laws of large numbers with respect to martingale theory.Comprised of eight chapters, this volume begins with an overview of the basic concepts of the theory of measure and integration, followed by a presentation of var
International Nuclear Information System (INIS)
Ekonomou, L; Karampelas, P; Vita, V; Chatzarakis, G E
2011-01-01
One of the most popular methods of protecting high voltage transmission lines against lightning strikes and internal overvoltages is the use of arresters. The installation of arresters in high voltage transmission lines can prevent or even reduce the lines' failure rate. Several studies based on simulation tools have been presented in order to estimate the critical currents that exceed the arresters' rated energy stress and to specify the arresters' installation interval. In this work artificial intelligence, and more specifically a Q-learning artificial neural network (ANN) model, is addressed for evaluating the arresters' failure probability. The aims of the paper are to describe in detail the developed Q-learning ANN model and to compare the results obtained by its application in operating 150 kV Greek transmission lines with those produced using a simulation tool. The satisfactory and accurate results of the proposed ANN model can make it a valuable tool for designers of electrical power systems seeking more effective lightning protection, reducing operational costs and better continuity of service
Ekonomou, L.; Karampelas, P.; Vita, V.; Chatzarakis, G. E.
2011-04-01
One of the most popular methods of protecting high voltage transmission lines against lightning strikes and internal overvoltages is the use of arresters. The installation of arresters in high voltage transmission lines can prevent or even reduce the lines' failure rate. Several studies based on simulation tools have been presented in order to estimate the critical currents that exceed the arresters' rated energy stress and to specify the arresters' installation interval. In this work artificial intelligence, and more specifically a Q-learning artificial neural network (ANN) model, is addressed for evaluating the arresters' failure probability. The aims of the paper are to describe in detail the developed Q-learning ANN model and to compare the results obtained by its application in operating 150 kV Greek transmission lines with those produced using a simulation tool. The satisfactory and accurate results of the proposed ANN model can make it a valuable tool for designers of electrical power systems seeking more effective lightning protection, reducing operational costs and better continuity of service.
Kretch, Kari S.; Adolph, Karen E.
2013-01-01
Do infants, like adults, consider both the probability of falling and the severity of a potential fall when deciding whether to cross a bridge? Crawling and walking infants were encouraged to cross bridges varying in width over a small drop-off, a large drop-off, or no drop-off. Bridge width affects the probability of falling, whereas drop-off…
Investigation of breached depleted UF{sub 6} cylinders
Energy Technology Data Exchange (ETDEWEB)
DeVan, J.H. [Martin Marietta Energy Systems, Inc., Oak Ridge, TN (United States)
1991-12-31
In June 1990, during a three-site inspection of cylinders being used for long-term storage of solid depleted UF{sub 6}, two 14-ton cylinders at Portsmouth, Ohio, were discovered with holes in the barrel section of the cylinders. An investigation team was immediately formed to determine the cause of the failures and their impact on future storage procedures and to recommend corrective actions. Subsequent investigation showed that the failures most probably resulted from mechanical damage that occurred at the time that the cylinders had been placed in the storage yard. In both cylinders evidence pointed to the impact of a lifting lug of an adjacent cylinder near the front stiffening ring, where deflection of the cylinder could occur only by tearing the cylinder. The impacts appear to have punctured the cylinders and thereby set up corrosion processes that greatly extended the openings in the wall and obliterated the original crack. Fortunately, the reaction products formed by this process were relatively protective and prevented any large-scale loss of uranium. The main factors that precipitated the failures were inadequate spacing between cylinders and deviations in the orientations of lifting lugs from their intended horizontal position. After reviewing the causes and effects of the failures, the team`s principal recommendation for remedial action concerned improved cylinder handling and inspection procedures. Design modifications and supplementary mechanical tests were also recommended to improve the cylinder containment integrity during the stacking operation.
Freund, John E
1993-01-01
Thorough, lucid coverage of permutations and factorials, probabilities and odds, frequency interpretation, mathematical expectation, decision making, postulates of probability, rule of elimination, binomial distribution, geometric distribution, standard deviation, law of large numbers, and much more. Exercises with some solutions. Summary. Bibliography. Includes 42 black-and-white illustrations. 1973 edition.
Probability, Nondeterminism and Concurrency
DEFF Research Database (Denmark)
Varacca, Daniele
Nondeterminism is modelled in domain theory by the notion of a powerdomain, while probability is modelled by that of the probabilistic powerdomain. Some problems arise when we want to combine them in order to model computation in which both nondeterminism and probability are present. In particula...
Rocchi, Paolo
2014-01-01
The problem of probability interpretation was long overlooked before exploding in the 20th century, when the frequentist and subjectivist schools formalized two conflicting conceptions of probability. Beyond the radical followers of the two schools, a circle of pluralist thinkers tends to reconcile the opposing concepts. The author uses two theorems in order to prove that the various interpretations of probability do not come into opposition and can be used in different contexts. The goal here is to clarify the multifold nature of probability by means of a purely mathematical approach and to show how philosophical arguments can only serve to deepen actual intellectual contrasts. The book can be considered as one of the most important contributions in the analysis of probability interpretation in the last 10-15 years.
The disclosure of diagnosis codes can breach research participants' privacy.
Loukides, Grigorios; Denny, Joshua C; Malin, Bradley
2010-01-01
De-identified clinical data in standardized form (eg, diagnosis codes), derived from electronic medical records, are increasingly combined with research data (eg, DNA sequences) and disseminated to enable scientific investigations. This study examines whether released data can be linked with identified clinical records that are accessible via various resources to jeopardize patients' anonymity, and the ability of popular privacy protection methodologies to prevent such an attack. The study experimentally evaluates the re-identification risk of a de-identified sample of Vanderbilt's patient records involved in a genome-wide association study. It also measures the level of protection from re-identification, and data utility, provided by suppression and generalization. Privacy protection is quantified using the probability of re-identifying a patient in a larger population through diagnosis codes. Data utility is measured at a dataset level, using the percentage of retained information, as well as its description, and at a patient level, using two metrics based on the difference between the distribution of Internal Classification of Disease (ICD) version 9 codes before and after applying privacy protection. More than 96% of 2800 patients' records are shown to be uniquely identified by their diagnosis codes with respect to a population of 1.2 million patients. Generalization is shown to reduce further the percentage of de-identified records by less than 2%, and over 99% of the three-digit ICD-9 codes need to be suppressed to prevent re-identification. Popular privacy protection methods are inadequate to deliver a sufficiently protected and useful result when sharing data derived from complex clinical systems. The development of alternative privacy protection models is thus required.
Directory of Open Access Journals (Sweden)
Jian Wang
2014-01-01
Full Text Available The present study was to investigate whether a magnolia extract, named BL153, can prevent obesity-induced liver damage and identify the possible protective mechanism. To this end, obese mice were induced by feeding with high fat diet (HFD, 60% kcal as fat and the age-matched control mice were fed with control diet (10% kcal as fat for 6 months. Simultaneously these mice were treated with or without BL153 daily at 3 dose levels (2.5, 5, and 10 mg/kg by gavage. HFD feeding significantly increased the body weight and the liver weight. Administration of BL153 significantly reduced the liver weight but without effects on body weight. As a critical step of the development of NAFLD, hepatic fibrosis was induced in the mice fed with HFD, shown by upregulating the expression of connective tissue growth factor and transforming growth factor beta 1, which were significantly attenuated by BL153 in a dose-dependent manner. Mechanism study revealed that BL153 significantly suppressed HFD induced hepatic lipid accumulation and oxidative stress and slightly prevented liver inflammation. These results suggest that HFD induced fibrosis in the liver can be prevented partially by BL153, probably due to reduction of hepatic lipid accumulation, inflammation and oxidative stress.
Velasquez Montoya, Liliana; Sciaudone, Elizabeth J.; Mitasova, Helena; Overton, Margery F.
2018-03-01
The Outer Banks of North Carolina is a wave-dominated barrier island system that has experienced the opening and closure of numerous inlets in the last four centuries. The most recent of those inlets formed after the breaching of Pea Island during Hurricane Irene in 2011. The Pea Island Breach experienced a rapid evolution including episodic curvature of the main channel, rotation of the ebb channel, shoaling, widening by Hurricane Sandy in 2012, and finally closing before the summer of 2013. Studying the life cycle of Pea Island Breach contributes to understanding the behavior of ephemeral inlets in breaching-prone regions. This topic has gained relevance due to rising sea levels, a phenomenon that increases the chances of ephemeral inlet formation during extreme events. This study explores the spatiotemporal effects of tides, waves, and storms on flow velocities and morphology of the breach by means of remotely sensed data, geospatial metrics, and a numerical model. The combined use of observations and results from modeling experiments allowed building a conceptual model to explain the life cycle of Pea Island Breach. Wave seasonality dominated the morphological evolution of the inlet by controlling the magnitude and direction of the longshore current that continuously built transient spits at both sides of the breach. Sensitivity analysis to external forcings indicates that ocean waves can modify water levels and velocities in the back barrier. Sound-side storm surge regulates overall growth rate, duration, and decay of peak water levels entering the inlet during extreme events.
Billingsley, Patrick
2012-01-01
Praise for the Third Edition "It is, as far as I'm concerned, among the best books in math ever written....if you are a mathematician and want to have the top reference in probability, this is it." (Amazon.com, January 2006) A complete and comprehensive classic in probability and measure theory Probability and Measure, Anniversary Edition by Patrick Billingsley celebrates the achievements and advancements that have made this book a classic in its field for the past 35 years. Now re-issued in a new style and format, but with the reliable content that the third edition was revered for, this
International Nuclear Information System (INIS)
Bitsakis, E.I.; Nicolaides, C.A.
1989-01-01
The concept of probability is now, and always has been, central to the debate on the interpretation of quantum mechanics. Furthermore, probability permeates all of science, as well as our every day life. The papers included in this volume, written by leading proponents of the ideas expressed, embrace a broad spectrum of thought and results: mathematical, physical epistemological, and experimental, both specific and general. The contributions are arranged in parts under the following headings: Following Schroedinger's thoughts; Probability and quantum mechanics; Aspects of the arguments on nonlocality; Bell's theorem and EPR correlations; Real or Gedanken experiments and their interpretation; Questions about irreversibility and stochasticity; and Epistemology, interpretation and culture. (author). refs.; figs.; tabs
Shorack, Galen R
2017-01-01
This 2nd edition textbook offers a rigorous introduction to measure theoretic probability with particular attention to topics of interest to mathematical statisticians—a textbook for courses in probability for students in mathematical statistics. It is recommended to anyone interested in the probability underlying modern statistics, providing a solid grounding in the probabilistic tools and techniques necessary to do theoretical research in statistics. For the teaching of probability theory to post graduate statistics students, this is one of the most attractive books available. Of particular interest is a presentation of the major central limit theorems via Stein's method either prior to or alternative to a characteristic function presentation. Additionally, there is considerable emphasis placed on the quantile function as well as the distribution function. The bootstrap and trimming are both presented. Martingale coverage includes coverage of censored data martingales. The text includes measure theoretic...
Concepts of probability theory
Pfeiffer, Paul E
1979-01-01
Using the Kolmogorov model, this intermediate-level text discusses random variables, probability distributions, mathematical expectation, random processes, more. For advanced undergraduates students of science, engineering, or math. Includes problems with answers and six appendixes. 1965 edition.
Probability and Bayesian statistics
1987-01-01
This book contains selected and refereed contributions to the "Inter national Symposium on Probability and Bayesian Statistics" which was orga nized to celebrate the 80th birthday of Professor Bruno de Finetti at his birthplace Innsbruck in Austria. Since Professor de Finetti died in 1985 the symposium was dedicated to the memory of Bruno de Finetti and took place at Igls near Innsbruck from 23 to 26 September 1986. Some of the pa pers are published especially by the relationship to Bruno de Finetti's scientific work. The evolution of stochastics shows growing importance of probability as coherent assessment of numerical values as degrees of believe in certain events. This is the basis for Bayesian inference in the sense of modern statistics. The contributions in this volume cover a broad spectrum ranging from foundations of probability across psychological aspects of formulating sub jective probability statements, abstract measure theoretical considerations, contributions to theoretical statistics an...
Probability and Statistical Inference
Prosper, Harrison B.
2006-01-01
These lectures introduce key concepts in probability and statistical inference at a level suitable for graduate students in particle physics. Our goal is to paint as vivid a picture as possible of the concepts covered.
Hartmann, Stephan
2011-01-01
Many results of modern physics--those of quantum mechanics, for instance--come in a probabilistic guise. But what do probabilistic statements in physics mean? Are probabilities matters of objective fact and part of the furniture of the world, as objectivists think? Or do they only express ignorance or belief, as Bayesians suggest? And how are probabilistic hypotheses justified and supported by empirical evidence? Finally, what does the probabilistic nature of physics imply for our understanding of the world? This volume is the first to provide a philosophical appraisal of probabilities in all of physics. Its main aim is to make sense of probabilistic statements as they occur in the various physical theories and models and to provide a plausible epistemology and metaphysics of probabilities. The essays collected here consider statistical physics, probabilistic modelling, and quantum mechanics, and critically assess the merits and disadvantages of objectivist and subjectivist views of probabilities in these fie...
Grimmett, Geoffrey
2014-01-01
Probability is an area of mathematics of tremendous contemporary importance across all aspects of human endeavour. This book is a compact account of the basic features of probability and random processes at the level of first and second year mathematics undergraduates and Masters' students in cognate fields. It is suitable for a first course in probability, plus a follow-up course in random processes including Markov chains. A special feature is the authors' attention to rigorous mathematics: not everything is rigorous, but the need for rigour is explained at difficult junctures. The text is enriched by simple exercises, together with problems (with very brief hints) many of which are taken from final examinations at Cambridge and Oxford. The first eight chapters form a course in basic probability, being an account of events, random variables, and distributions - discrete and continuous random variables are treated separately - together with simple versions of the law of large numbers and the central limit th...
Hemmo, Meir
2012-01-01
What is the role and meaning of probability in physical theory, in particular in two of the most successful theories of our age, quantum physics and statistical mechanics? Laws once conceived as universal and deterministic, such as Newton‘s laws of motion, or the second law of thermodynamics, are replaced in these theories by inherently probabilistic laws. This collection of essays by some of the world‘s foremost experts presents an in-depth analysis of the meaning of probability in contemporary physics. Among the questions addressed are: How are probabilities defined? Are they objective or subjective? What is their explanatory value? What are the differences between quantum and classical probabilities? The result is an informative and thought-provoking book for the scientifically inquisitive.
Probability in quantum mechanics
Directory of Open Access Journals (Sweden)
J. G. Gilson
1982-01-01
Full Text Available By using a fluid theory which is an alternative to quantum theory but from which the latter can be deduced exactly, the long-standing problem of how quantum mechanics is related to stochastic processes is studied. It can be seen how the Schrödinger probability density has a relationship to time spent on small sections of an orbit, just as the probability density has in some classical contexts.
Quantum computing and probability.
Ferry, David K
2009-11-25
Over the past two decades, quantum computing has become a popular and promising approach to trying to solve computationally difficult problems. Missing in many descriptions of quantum computing is just how probability enters into the process. Here, we discuss some simple examples of how uncertainty and probability enter, and how this and the ideas of quantum computing challenge our interpretations of quantum mechanics. It is found that this uncertainty can lead to intrinsic decoherence, and this raises challenges for error correction.
Quantum computing and probability
International Nuclear Information System (INIS)
Ferry, David K
2009-01-01
Over the past two decades, quantum computing has become a popular and promising approach to trying to solve computationally difficult problems. Missing in many descriptions of quantum computing is just how probability enters into the process. Here, we discuss some simple examples of how uncertainty and probability enter, and how this and the ideas of quantum computing challenge our interpretations of quantum mechanics. It is found that this uncertainty can lead to intrinsic decoherence, and this raises challenges for error correction. (viewpoint)
Augustin, C. M.
2015-12-01
Carbon capture and storage (CCS) has been suggested by the Intergovernmental Panel on Climate Change as a partial solution to the greenhouse gas emissions problem. As CCS has become mainstream, researchers have raised multiple risk assessment issues typical of emerging technologies. In our research, we examine issues occuring when stored carbon dioxide (CO2) migrates to the near-surface or surface. We believe that both the public misperception and the physical reality of potential environmental, health, and commercial impacts of leak events from such subsurface sites have prevented widespread adoption of CCS. This paper is presented in three parts; the first is an evaluation of the systemic risk of a CCS site CO2 leak and models indicating potential likelihood of a leakage event. As the likelihood of a CCS site leak is stochastic and nonlinear, we present several Bayesian simulations for leak events based on research done with other low-probability, high-consequence gaseous pollutant releases. Though we found a large, acute leak to be exceptionally rare, we demonstrate potential for a localized, chronic leak at a CCS site. To that end, we present the second piece of this paper. Using a combination of spatio-temporal models and reaction-path models, we demonstrate the interplay between leak migrations, material interactions, and atmospheric dispersion for leaks of various duration and volume. These leak-event scenarios have implications for human, environmental, and economic health; they also have a significant impact on implementation support. Public acceptance of CCS is essential for a national low-carbon future, and this is what we address in the final part of this paper. We demonstrate that CCS remains unknown to the general public in the United States. Despite its unknown state, we provide survey findings -analyzed in Slovic and Weber's 2002 framework - that show a high unknown, high dread risk perception of leaks from a CCS site. Secondary findings are a
Knoblauch, Theresa A K; Stauffacher, Michael; Trutnevyte, Evelina
2018-04-01
Subsurface energy activities entail the risk of induced seismicity including low-probability high-consequence (LPHC) events. For designing respective risk communication, the scientific literature lacks empirical evidence of how the public reacts to different written risk communication formats about such LPHC events and to related uncertainty or expert confidence. This study presents findings from an online experiment (N = 590) that empirically tested the public's responses to risk communication about induced seismicity and to different technology frames, namely deep geothermal energy (DGE) and shale gas (between-subject design). Three incrementally different formats of written risk communication were tested: (i) qualitative, (ii) qualitative and quantitative, and (iii) qualitative and quantitative with risk comparison. Respondents found the latter two the easiest to understand, the most exact, and liked them the most. Adding uncertainty and expert confidence statements made the risk communication less clear, less easy to understand and increased concern. Above all, the technology for which risks are communicated and its acceptance mattered strongly: respondents in the shale gas condition found the identical risk communication less trustworthy and more concerning than in the DGE conditions. They also liked the risk communication overall less. For practitioners in DGE or shale gas projects, the study shows that the public would appreciate efforts in describing LPHC risks with numbers and optionally risk comparisons. However, there seems to be a trade-off between aiming for transparency by disclosing uncertainty and limited expert confidence, and thereby decreasing clarity and increasing concern in the view of the public. © 2017 Society for Risk Analysis.
Breach of duty: Power of shareholders to ratify directors fraudulent dealings
Directory of Open Access Journals (Sweden)
Anthony O. Nwafor
2014-07-01
Full Text Available Company directors owe duty of loyalty to the company which prohibits them from fraudulent dealings in the course of conducting the affairs of the company. Although the shareholders could, in the exercise of their voting powers, grant relieves to the directors from liabilities arising from a breach of duty that amounts to fraud, the extent and capacity in which the shareholders could exercise such powers is confounded by the elusive attempts by the courts in defining fraud. The paper argues that without a definite meaning ascribed to fraud, the power and capacity in which the shareholders could ratify a breach of duty arising from self-dealing and expropriation of corporate opportunities by directors cannot be predetermined, but that each case would be based on the peculiarities of its own facts.
Directory of Open Access Journals (Sweden)
Wendy Niesen
2018-03-01
Full Text Available This study investigates how quantitative and qualitative job insecurity relate to idea generation, a dimension of innovative work behaviour. We hypothesise that both types of job insecurity relate negatively to this type of innovative behaviour, and expect a stronger association between quantitative job insecurity and idea generation. Moreover, we argue that psychological contract breach mediates (‘explains’ these negative relationships. The hypotheses were tested in a sample of 1420 supervisors from a large Belgian organisation, using hierarchical regression analyses, bootstrapping analyses, and relative weight analysis. The results showed that both types of job insecurity are negatively associated with idea generation. Contrary to our expectations, the relationship between both forms of job insecurity and idea generation was equally strong. Psychological contract breach was found to mediate these relationships.
Electroconvulsive therapy, hypertensive surge, blood-brain barrier breach, and amnesia
DEFF Research Database (Denmark)
Andrade, Chittaranjan; Bolwig, Tom G
2014-01-01
Preclinical and clinical evidence show that electroconvulsive therapy (ECT)-induced intraictal surge in blood pressure may result in a small, transient breach in the blood-brain barrier, leading to mild cerebral edema and a possible leach of noxious substances from blood into brain tissues...... convincing evidence of benefits. It is concluded that there is insufficient support, at present, for the hypothesis that the hypertensive surge during ECT and the resultant blood-brain barrier breach contribute meaningfully to ECT-induced cognitive deficits. Future research should address the subset....... These changes may impair neuronal functioning and contribute to the mechanisms underlying ECT-induced cognitive deficits. Some but not all clinical data on the subject suggest that blood pressure changes during ECT correlate with indices of cognitive impairment. In animal models, pharmacological manipulations...
Crevasse Splays Versus Avulsions: A Recipe for Land Building With Levee Breaches
Nienhuis, Jaap H.; Törnqvist, Torbjörn E.; Esposito, Christopher R.
2018-05-01
Natural-levee breaches can not only initiate an avulsion but also, under the right circumstances, lead to crevasse splay formation and overbank sedimentation. The formative conditions for crevasse splays are not well understood, yet such river sediment diversions form an integral part of billion-dollar coastal restoration projects. Here we use Delft3D to investigate the influence of vegetation and soil consolidation on the evolution of a natural-levee breach. Model simulations show that crevasse splays heal because floodplain aggradation reduces the water surface slope, decreasing water discharge into the flood basin. Easily erodible and unvegetated floodplains increase the likelihood for channel avulsions. Denser vegetation and less potential for soil consolidation result in small crevasse splays that are not only efficient sediment traps but also short-lived. Successful crevasse splays that generate the largest land area gain for the imported sediment require a delicate balance between water and sediment discharge, vegetation root strength, and soil consolidation.
Establishing breach of the duty of care in the tort of negligence.
Tingle, John
This article, the third in a series on clinical negligence, looks at the law surrounding breach of the duty of care in negligence. It shows some of the principles that judges and lawyers use in order to decide whether a person has broken his/her duty of care in the tort of negligence. It will be seen that the principles are contained in decided court cases, some of which are quite old but are still relevant today. The focus of this article is on the rule that courts, in deciding the issue of a breach of duty of care, would judge the defendant's conduct by the standard of what the hypothetical, 'reasonable person' would have done in the circumstances of the case.
Medical negligence based on bad faith, breach of contract, or mental anguish.
Ficarra, B J
1980-01-01
Financial recovery owing to breach of contract is restricted to the pecuniary amount lost because of failure to perform on the stipulated contract. With the acquisition of newer knowledge, attorneys are now utilizing the weapon of contractual failure as applied to medical negligence. The impetus to this new weapon for the plaintiff has accrued because of the favorable verdicts rendered from positive decisions based upon bad faith.
Once more unto the breach managing information security in an uncertain world
Simmons, Andrea C
2012-01-01
In Once more unto the Breach, Andrea C Simmons speaks directly to information security managers and provides an insider's view of the role, offering priceless gems from her extensive experience and knowledge. Based on a typical year in the life of an information security manager, the book examines how the general principles can be applied to all situations and discusses the lessons learnt from a real project.
A method and programme (BREACH) for predicting the flow distribution in water cooled reactor cores
International Nuclear Information System (INIS)
Randles, J.; Roberts, H.A.
1961-03-01
The method presented here of evaluating the flow rate in individual reactor channels may be applied to any type of water cooled reactor in which boiling occurs The flow distribution is calculated with the aid of a MERCURY autocode programme, BREACH, which is described in detail. This programme computes the steady state longitudinal void distribution and pressure drop in a single channel on the basis of the homogeneous model of two phase flow. (author)
Operator decision aid for breached fuel operation in liquid metal cooled nuclear reactors
International Nuclear Information System (INIS)
Gross, K.C.; Hawkins, R.E.; Nickless, W.K.
1991-01-01
The purpose of this paper is to report the development of an expert system that provides continuous assessment of the safety significance and technical specification conformance of Delayed Neutron (DN) signals during breached fuel operation. The completed expert system has been parallelized on an innovative distributed-memory network-computing system that enables the computationally intensive kernel of the expert system to run in parallel on a group of low-cost Unix workstations. 1 ref
A method and programme (BREACH) for predicting the flow distribution in water cooled reactor cores
Energy Technology Data Exchange (ETDEWEB)
Randles, J; Roberts, H A [Technical Assessments and Services Division, Atomic Energy Establishment, Winfrith, Dorchester, Dorset (United Kingdom)
1961-03-15
The method presented here of evaluating the flow rate in individual reactor channels may be applied to any type of water cooled reactor in which boiling occurs The flow distribution is calculated with the aid of a MERCURY autocode programme, BREACH, which is described in detail. This programme computes the steady state longitudinal void distribution and pressure drop in a single channel on the basis of the homogeneous model of two phase flow. (author)
Psychological contract breach and work performance:Is social exchange a buffer or an intensifier?
Bal, P. Matthijs; Chiaburu, Dan S.; Jansen, Paul G. W.
2010-01-01
Purpose: The aim of this paper is to investigate how social exchanges modify the relationship between psychological contract breach and work performance. It aims to present two concurrent hypotheses, based on theoretical interaction effects of social exchanges (conceptualized as social exchange relationships, POS, and trust). Design/methodology/approach: Data were collected from a sample of 266 employees in a service sector company in the USA. Regression analysis was used to explore the moder...
A Depth-Averaged 2-D Simulation for Coastal Barrier Breaching Processes
2011-05-01
including bed change and variable flow density in the flow continuity and momentum equations. The model adopts the HLL approximate Riemann solver to handle...flow density in the flow continuity and momentum equations. The model adopts the HLL approximate Riemann solver to handle the mixed-regime flows near...18 547 Keulegan equation or the Bernoulli equation, and the breach morphological change is determined using simplified sediment transport models
VIBRATION ISOLATION SYSTEM PROBABILITY ANALYSIS
Directory of Open Access Journals (Sweden)
Smirnov Vladimir Alexandrovich
2012-10-01
Full Text Available The article deals with the probability analysis for a vibration isolation system of high-precision equipment, which is extremely sensitive to low-frequency oscillations even of submicron amplitude. The external sources of low-frequency vibrations may include the natural city background or internal low-frequency sources inside buildings (pedestrian activity, HVAC. Taking Gauss distribution into account, the author estimates the probability of the relative displacement of the isolated mass being still lower than the vibration criteria. This problem is being solved in the three dimensional space, evolved by the system parameters, including damping and natural frequency. According to this probability distribution, the chance of exceeding the vibration criteria for a vibration isolation system is evaluated. Optimal system parameters - damping and natural frequency - are being developed, thus the possibility of exceeding vibration criteria VC-E and VC-D is assumed to be less than 0.04.
The perception of probability.
Gallistel, C R; Krishan, Monika; Liu, Ye; Miller, Reilly; Latham, Peter E
2014-01-01
We present a computational model to explain the results from experiments in which subjects estimate the hidden probability parameter of a stepwise nonstationary Bernoulli process outcome by outcome. The model captures the following results qualitatively and quantitatively, with only 2 free parameters: (a) Subjects do not update their estimate after each outcome; they step from one estimate to another at irregular intervals. (b) The joint distribution of step widths and heights cannot be explained on the assumption that a threshold amount of change must be exceeded in order for them to indicate a change in their perception. (c) The mapping of observed probability to the median perceived probability is the identity function over the full range of probabilities. (d) Precision (how close estimates are to the best possible estimate) is good and constant over the full range. (e) Subjects quickly detect substantial changes in the hidden probability parameter. (f) The perceived probability sometimes changes dramatically from one observation to the next. (g) Subjects sometimes have second thoughts about a previous change perception, after observing further outcomes. (h) The frequency with which they perceive changes moves in the direction of the true frequency over sessions. (Explaining this finding requires 2 additional parametric assumptions.) The model treats the perception of the current probability as a by-product of the construction of a compact encoding of the experienced sequence in terms of its change points. It illustrates the why and the how of intermittent Bayesian belief updating and retrospective revision in simple perception. It suggests a reinterpretation of findings in the recent literature on the neurobiology of decision making. (PsycINFO Database Record (c) 2014 APA, all rights reserved).
Conditional Probability Modulates Visual Search Efficiency
Directory of Open Access Journals (Sweden)
Bryan eCort
2013-10-01
Full Text Available We investigated the effects of probability on visual search. Previous work has shown that people can utilize spatial and sequential probability information to improve target detection. We hypothesized that performance improvements from probability information would extend to the efficiency of visual search. Our task was a simple visual search in which the target was always present among a field of distractors, and could take one of two colors. The absolute probability of the target being either color was 0.5; however, the conditional probability – the likelihood of a particular color given a particular combination of two cues – varied from 0.1 to 0.9. We found that participants searched more efficiently for high conditional probability targets and less efficiently for low conditional probability targets, but only when they were explicitly informed of the probability relationship between cues and target color.
Psychological contract breach and outcomes: Combining meta-analysis and structural equation models.
Topa Cantisano, Gabriela; Morales Domínguez, J Francisco; Depolo, Marco
2008-08-01
In this study, meta-analytic procedures were used to examine the relationships between psychological contract perceived breach and certain outcome variables, such as organizational commitment, job satisfaction and organizational citizenship behaviours (OCB). Our review of the literature generated 41 independent samples in which perceived breach was used as a predictor of these personal and organizational outcomes. A medium effect size (ES) for desirable outcomes (job satisfaction, organizational commitment, organizational trust, OCB and performance) was obtained (r=-.35). For undesirable outcomes (neglect in role duties and intention to leave), ES were also medium (r=.31). When comparing attitudinal (job satisfaction, organizational commitment, organizational trust) and behavioural outcomes (OCB, neglect in role duties and performance), a stronger ES was found for attitudinal (r=-.24) than for behavioural outcomes (r=-.11). Potential moderator variables were examined, and it was found that they explained only a percentage of variability of primary studies. Structural equation analysis of the pooled meta-analytical correlation matrix indicated that the relationships of perceived breach with satisfaction, OCB, intention to leave and performance are fully mediated by organizational trust and commitment. Results are discussed in order to suggest theoretical and empirical implications.
Transient response and radiation dose estimates for breaches to a spent fuel processing facility
Energy Technology Data Exchange (ETDEWEB)
Solbrig, Charles W., E-mail: soltechco@aol.com; Pope, Chad; Andrus, Jason
2014-08-15
Highlights: • We model doses received from a nuclear fuel facility from boundary leaks due to an earthquake. • The supplemental exhaust system (SES) starts after breach causing air to be sucked into the cell. • Exposed metal fuel burns increasing pressure and release of radioactive contamination. • Facility releases are small and much less than the limits showing costly refits are unnecessary. • The method presented can be used in other nuclear fuel processing facilities. - Abstract: This paper describes the analysis of the design basis accident for Idaho National Laboratory Fuel Conditioning Facility (FCF). The facility is used to process spent metallic nuclear fuel. This analysis involves a model of the transient behavior of the FCF inert atmosphere hot cell following an earthquake initiated breach of pipes passing through the cell boundary. Such breaches allow the introduction of air and subsequent burning of pyrophoric metals. The model predicts the pressure, temperature, volumetric releases, cell heat transfer, metal fuel combustion, heat generation rates, radiological releases and other quantities. The results show that releases from the cell are minimal and satisfactory for safety. This analysis method should be useful in other facilities that have potential for damage from an earthquake and could eliminate the need to back fit facilities with earthquake proof boundaries or lessen the cost of new facilities.
After the data breach: Managing the crisis and mitigating the impact.
Brown, Hart S
2016-01-01
Historically, the unauthorised access and theft of information was a tactic used between countries as part of espionage campaigns, during times of conflict as well as for personal and criminal purposes. The consumers of the information were relatively isolated and specific. As information became stored and digitised in larger quantities in the 1980s the ability to access mass amounts of records at one time became possible. The expertise needed to remotely access and exfiltrate the data was not readily available and the number of markets to monetise the data was limited. Over the past ten years, shadow networks have been used by criminals to collaborate on hacking techniques, exchange hacking advice anonymously and commercialise data on the black market. The intersection of these networks along with the unintentional losses of information have resulted in 5,810 data breaches made public since 2005 (comprising some 847,807,830 records) and the velocity of these events is increasing. Organisations must be prepared for a potential breach event to maintain cyber resiliency. Proper management of a breach response can reduce response costs and can serve to mitigate potential reputational losses.
Transient response and radiation dose estimates for breaches to a spent fuel processing facility
International Nuclear Information System (INIS)
Solbrig, Charles W.; Pope, Chad; Andrus, Jason
2014-01-01
Highlights: • We model doses received from a nuclear fuel facility from boundary leaks due to an earthquake. • The supplemental exhaust system (SES) starts after breach causing air to be sucked into the cell. • Exposed metal fuel burns increasing pressure and release of radioactive contamination. • Facility releases are small and much less than the limits showing costly refits are unnecessary. • The method presented can be used in other nuclear fuel processing facilities. - Abstract: This paper describes the analysis of the design basis accident for Idaho National Laboratory Fuel Conditioning Facility (FCF). The facility is used to process spent metallic nuclear fuel. This analysis involves a model of the transient behavior of the FCF inert atmosphere hot cell following an earthquake initiated breach of pipes passing through the cell boundary. Such breaches allow the introduction of air and subsequent burning of pyrophoric metals. The model predicts the pressure, temperature, volumetric releases, cell heat transfer, metal fuel combustion, heat generation rates, radiological releases and other quantities. The results show that releases from the cell are minimal and satisfactory for safety. This analysis method should be useful in other facilities that have potential for damage from an earthquake and could eliminate the need to back fit facilities with earthquake proof boundaries or lessen the cost of new facilities
Directory of Open Access Journals (Sweden)
Zhiyong Zhu
2014-01-01
Full Text Available In 2004, the Ministry of Education, National Development and Reform Commission, State Ethnic Affairs Commission, Ministry of Finance, Ministry of Human Resources and Social Security jointly released a special document On Training High-Caliber Backbone Personnel from the Ethnic Minorities, as a supporting measure for the strategy of human resources in western development of China. The central government takes it as an important force in thriving and developing the western region, especially the region where ethnic minority groups reside in the west. However, more and more graduates through the Program have breached or intend to break the contract of the Program. Why do they not abide by the contract and return to their hometown province for the work after they have enjoyed the preferential Program? This case study focused on three graduates who broke the employment contract through the Program, traced the origin and development of their minds and action, and demonstrated more about what hid behind the phenomenon. According to the deep interaction with these ethnic minority students, the research found that the social background under which high education policies for the ethnic minorities had greatly changed. The changes have challenged the implementation of the Program. Furthermore, under the socialistic market economy system, the employment contract of the Program couldn’t be processed and secured as the involved actors have expected. Last but not the least, the graduates benefited from the Program have become more and more individualistic and diversified in their career development due to different micro and macro reasons.En el año 2004 y de manera conjunta, el Ministerio de Educación, la Comisión Nacional para el Desarrollo y las Reformas, la Comisión Estatal de Asuntos Étnicos, el Ministerio de Economía, el Ministerio de Recursos Humanos y la Seguridad Social, publicaron una normativa especial para la Formación de Alto Nivel de
Irreversibility and conditional probability
International Nuclear Information System (INIS)
Stuart, C.I.J.M.
1989-01-01
The mathematical entropy - unlike physical entropy - is simply a measure of uniformity for probability distributions in general. So understood, conditional entropies have the same logical structure as conditional probabilities. If, as is sometimes supposed, conditional probabilities are time-reversible, then so are conditional entropies and, paradoxically, both then share this symmetry with physical equations of motion. The paradox is, of course that probabilities yield a direction to time both in statistical mechanics and quantum mechanics, while the equations of motion do not. The supposed time-reversibility of both conditionals seems also to involve a form of retrocausality that is related to, but possibly not the same as, that described by Costa de Beaurgard. The retrocausality is paradoxically at odds with the generally presumed irreversibility of the quantum mechanical measurement process. Further paradox emerges if the supposed time-reversibility of the conditionals is linked with the idea that the thermodynamic entropy is the same thing as 'missing information' since this confounds the thermodynamic and mathematical entropies. However, it is shown that irreversibility is a formal consequence of conditional entropies and, hence, of conditional probabilities also. 8 refs. (Author)
Isaac, Richard
1995-01-01
The ideas of probability are all around us. Lotteries, casino gambling, the al most non-stop polling which seems to mold public policy more and more these are a few of the areas where principles of probability impinge in a direct way on the lives and fortunes of the general public. At a more re moved level there is modern science which uses probability and its offshoots like statistics and the theory of random processes to build mathematical descriptions of the real world. In fact, twentieth-century physics, in embrac ing quantum mechanics, has a world view that is at its core probabilistic in nature, contrary to the deterministic one of classical physics. In addition to all this muscular evidence of the importance of probability ideas it should also be said that probability can be lots of fun. It is a subject where you can start thinking about amusing, interesting, and often difficult problems with very little mathematical background. In this book, I wanted to introduce a reader with at least a fairl...
Experimental Probability in Elementary School
Andrew, Lane
2009-01-01
Concepts in probability can be more readily understood if students are first exposed to probability via experiment. Performing probability experiments encourages students to develop understandings of probability grounded in real events, as opposed to merely computing answers based on formulae.
International Nuclear Information System (INIS)
Suen, C.J.; Sullivan, T.M.
1990-01-01
This paper discusses the development of a source term model for low-level waste shallow land burial facilities and separates the problem into four individual compartments. These are water flow, corrosion and subsequent breaching of containers, leaching of the waste forms, and solute transport. For the first and the last compartments, we adopted the existing codes, FEMWATER and FEMWASTE, respectively. We wrote two new modules for the other two compartments in the form of two separate Fortran subroutines -- BREACH and LEACH. They were incorporated into a modified version of the transport code FEMWASTE. The resultant code, which contains all three modules of container breaching, waste form leaching, and solute transport, was renamed BLT (for Breach, Leach, and Transport). This paper summarizes the overall program structure and logistics, and presents two examples from the results of verification and sensitivity tests. 6 refs., 7 figs., 1 tab
Improving Ranking Using Quantum Probability
Melucci, Massimo
2011-01-01
The paper shows that ranking information units by quantum probability differs from ranking them by classical probability provided the same data used for parameter estimation. As probability of detection (also known as recall or power) and probability of false alarm (also known as fallout or size) measure the quality of ranking, we point out and show that ranking by quantum probability yields higher probability of detection than ranking by classical probability provided a given probability of ...
Choice probability generating functions
DEFF Research Database (Denmark)
Fosgerau, Mogens; McFadden, Daniel; Bierlaire, Michel
2010-01-01
This paper establishes that every random utility discrete choice model (RUM) has a representation that can be characterized by a choice-probability generating function (CPGF) with specific properties, and that every function with these specific properties is consistent with a RUM. The choice...... probabilities from the RUM are obtained from the gradient of the CPGF. Mixtures of RUM are characterized by logarithmic mixtures of their associated CPGF. The paper relates CPGF to multivariate extreme value distributions, and reviews and extends methods for constructing generating functions for applications....... The choice probabilities of any ARUM may be approximated by a cross-nested logit model. The results for ARUM are extended to competing risk survival models....
Probability and stochastic modeling
Rotar, Vladimir I
2012-01-01
Basic NotionsSample Space and EventsProbabilitiesCounting TechniquesIndependence and Conditional ProbabilityIndependenceConditioningThe Borel-Cantelli TheoremDiscrete Random VariablesRandom Variables and VectorsExpected ValueVariance and Other Moments. Inequalities for DeviationsSome Basic DistributionsConvergence of Random Variables. The Law of Large NumbersConditional ExpectationGenerating Functions. Branching Processes. Random Walk RevisitedBranching Processes Generating Functions Branching Processes Revisited More on Random WalkMarkov ChainsDefinitions and Examples. Probability Distributions of Markov ChainsThe First Step Analysis. Passage TimesVariables Defined on a Markov ChainErgodicity and Stationary DistributionsA Classification of States and ErgodicityContinuous Random VariablesContinuous DistributionsSome Basic Distributions Continuous Multivariate Distributions Sums of Independent Random Variables Conditional Distributions and ExpectationsDistributions in the General Case. SimulationDistribution F...
Collision Probability Analysis
DEFF Research Database (Denmark)
Hansen, Peter Friis; Pedersen, Preben Terndrup
1998-01-01
It is the purpose of this report to apply a rational model for prediction of ship-ship collision probabilities as function of the ship and the crew characteristics and the navigational environment for MS Dextra sailing on a route between Cadiz and the Canary Islands.The most important ship and crew...... characteristics are: ship speed, ship manoeuvrability, the layout of the navigational bridge, the radar system, the number and the training of navigators, the presence of a look out etc. The main parameters affecting the navigational environment are ship traffic density, probability distributions of wind speeds...... probability, i.e. a study of the navigator's role in resolving critical situations, a causation factor is derived as a second step.The report documents the first step in a probabilistic collision damage analysis. Future work will inlcude calculation of energy released for crushing of structures giving...
M.E.H. Bos (Marian); D.E. te Beest (Dennis); M. van Boven (Michiel); M.R.D.R.B. van Holle; A. Meijer (Adam); A. Bosman (Arnold); Y.M. Mulder (Yonne); M.P.G. Koopmans D.V.M. (Marion); A. Stegeman (Arjan)
2010-01-01
textabstractAn epizootic of avian influenza (H7N7) caused a large number of human infections in The Netherlands in 2003. We used data from this epizootic to estimate infection probabilities for persons involved in disease control on infected farms. Analyses were based on databases containing
Estimating Subjective Probabilities
DEFF Research Database (Denmark)
Andersen, Steffen; Fountain, John; Harrison, Glenn W.
2014-01-01
either construct elicitation mechanisms that control for risk aversion, or construct elicitation mechanisms which undertake 'calibrating adjustments' to elicited reports. We illustrate how the joint estimation of risk attitudes and subjective probabilities can provide the calibration adjustments...... that theory calls for. We illustrate this approach using data from a controlled experiment with real monetary consequences to the subjects. This allows the observer to make inferences about the latent subjective probability, under virtually any well-specified model of choice under subjective risk, while still...
Classic Problems of Probability
Gorroochurn, Prakash
2012-01-01
"A great book, one that I will certainly add to my personal library."—Paul J. Nahin, Professor Emeritus of Electrical Engineering, University of New Hampshire Classic Problems of Probability presents a lively account of the most intriguing aspects of statistics. The book features a large collection of more than thirty classic probability problems which have been carefully selected for their interesting history, the way they have shaped the field, and their counterintuitive nature. From Cardano's 1564 Games of Chance to Jacob Bernoulli's 1713 Golden Theorem to Parrondo's 1996 Perplexin
An Objective Theory of Probability (Routledge Revivals)
Gillies, Donald
2012-01-01
This reissue of D. A. Gillies highly influential work, first published in 1973, is a philosophical theory of probability which seeks to develop von Mises' views on the subject. In agreement with von Mises, the author regards probability theory as a mathematical science like mechanics or electrodynamics, and probability as an objective, measurable concept like force, mass or charge. On the other hand, Dr Gillies rejects von Mises' definition of probability in terms of limiting frequency and claims that probability should be taken as a primitive or undefined term in accordance with modern axioma
Gordon, V; Williams, D J; Donnelly, P D
2012-04-01
To explore the relationship between attention deficit hyperactivity disorder (ADHD) symptoms (inattention, hyperactivity and impulsivity) and violent and non-violent prison breaches of discipline in incarcerated male youths aged 18-21 years. A case-control study of 169 male youth offenders incarcerated in Scottish prisons and classified as 'symptomatic' or 'non-symptomatic' of inattentive and hyperactive/impulsive ADHD symptoms. ADHD symptoms were measured using the Conners' Adult ADHD Rating Scales-Self Report: Long Version, and prison breaches of discipline were gathered from the Scottish Prison Service's Prisoner Records System. Youths who were symptomatic of Diagnostic and Statistical Manual of Mental Disorders, 4th Edition (DSM-IV) ADHD total symptoms had a significantly higher number of prison breaches of discipline than those who were non-symptomatic. Youths who were symptomatic of DSM-IV hyperactive/impulsive symptoms had a significantly higher number of violent and non-violent prison breaches of discipline than those who were non-symptomatic. However, no such significant difference was found between youths who were symptomatic and non-symptomatic of DSM-IV inattentive symptoms. Young male offenders who are symptomatic of ADHD have a higher number of prison breaches of discipline. In particular, symptoms of hyperactivity/impulsivity are associated with breaches of both a violent and non-violent nature. Implications of such symptoms on rehabilitation and recidivism are discussed. Copyright © 2012 The Royal Society for Public Health. Published by Elsevier Ltd. All rights reserved.
Counterexamples in probability
Stoyanov, Jordan M
2013-01-01
While most mathematical examples illustrate the truth of a statement, counterexamples demonstrate a statement's falsity. Enjoyable topics of study, counterexamples are valuable tools for teaching and learning. The definitive book on the subject in regards to probability, this third edition features the author's revisions and corrections plus a substantial new appendix.
Plotnitsky, Arkady
2010-01-01
Offers an exploration of the relationships between epistemology and probability in the work of Niels Bohr, Werner Heisenberg, and Erwin Schrodinger; in quantum mechanics; and in modern physics. This book considers the implications of these relationships and of quantum theory for our understanding of the nature of thinking and knowledge in general
Transition probabilities for atoms
International Nuclear Information System (INIS)
Kim, Y.K.
1980-01-01
Current status of advanced theoretical methods for transition probabilities for atoms and ions is discussed. An experiment on the f values of the resonance transitions of the Kr and Xe isoelectronic sequences is suggested as a test for the theoretical methods
Negative probability in the framework of combined probability
Burgin, Mark
2013-01-01
Negative probability has found diverse applications in theoretical physics. Thus, construction of sound and rigorous mathematical foundations for negative probability is important for physics. There are different axiomatizations of conventional probability. So, it is natural that negative probability also has different axiomatic frameworks. In the previous publications (Burgin, 2009; 2010), negative probability was mathematically formalized and rigorously interpreted in the context of extende...
Contributions to quantum probability
International Nuclear Information System (INIS)
Fritz, Tobias
2010-01-01
Chapter 1: On the existence of quantum representations for two dichotomic measurements. Under which conditions do outcome probabilities of measurements possess a quantum-mechanical model? This kind of problem is solved here for the case of two dichotomic von Neumann measurements which can be applied repeatedly to a quantum system with trivial dynamics. The solution uses methods from the theory of operator algebras and the theory of moment problems. The ensuing conditions reveal surprisingly simple relations between certain quantum-mechanical probabilities. It also shown that generally, none of these relations holds in general probabilistic models. This result might facilitate further experimental discrimination between quantum mechanics and other general probabilistic theories. Chapter 2: Possibilistic Physics. I try to outline a framework for fundamental physics where the concept of probability gets replaced by the concept of possibility. Whereas a probabilistic theory assigns a state-dependent probability value to each outcome of each measurement, a possibilistic theory merely assigns one of the state-dependent labels ''possible to occur'' or ''impossible to occur'' to each outcome of each measurement. It is argued that Spekkens' combinatorial toy theory of quantum mechanics is inconsistent in a probabilistic framework, but can be regarded as possibilistic. Then, I introduce the concept of possibilistic local hidden variable models and derive a class of possibilistic Bell inequalities which are violated for the possibilistic Popescu-Rohrlich boxes. The chapter ends with a philosophical discussion on possibilistic vs. probabilistic. It can be argued that, due to better falsifiability properties, a possibilistic theory has higher predictive power than a probabilistic one. Chapter 3: The quantum region for von Neumann measurements with postselection. It is determined under which conditions a probability distribution on a finite set can occur as the outcome
von der Linden, Wolfgang; Dose, Volker; von Toussaint, Udo
2014-06-01
Preface; Part I. Introduction: 1. The meaning of probability; 2. Basic definitions; 3. Bayesian inference; 4. Combinatrics; 5. Random walks; 6. Limit theorems; 7. Continuous distributions; 8. The central limit theorem; 9. Poisson processes and waiting times; Part II. Assigning Probabilities: 10. Transformation invariance; 11. Maximum entropy; 12. Qualified maximum entropy; 13. Global smoothness; Part III. Parameter Estimation: 14. Bayesian parameter estimation; 15. Frequentist parameter estimation; 16. The Cramer-Rao inequality; Part IV. Testing Hypotheses: 17. The Bayesian way; 18. The frequentist way; 19. Sampling distributions; 20. Bayesian vs frequentist hypothesis tests; Part V. Real World Applications: 21. Regression; 22. Inconsistent data; 23. Unrecognized signal contributions; 24. Change point problems; 25. Function estimation; 26. Integral equations; 27. Model selection; 28. Bayesian experimental design; Part VI. Probabilistic Numerical Techniques: 29. Numerical integration; 30. Monte Carlo methods; 31. Nested sampling; Appendixes; References; Index.
Contributions to quantum probability
Energy Technology Data Exchange (ETDEWEB)
Fritz, Tobias
2010-06-25
Chapter 1: On the existence of quantum representations for two dichotomic measurements. Under which conditions do outcome probabilities of measurements possess a quantum-mechanical model? This kind of problem is solved here for the case of two dichotomic von Neumann measurements which can be applied repeatedly to a quantum system with trivial dynamics. The solution uses methods from the theory of operator algebras and the theory of moment problems. The ensuing conditions reveal surprisingly simple relations between certain quantum-mechanical probabilities. It also shown that generally, none of these relations holds in general probabilistic models. This result might facilitate further experimental discrimination between quantum mechanics and other general probabilistic theories. Chapter 2: Possibilistic Physics. I try to outline a framework for fundamental physics where the concept of probability gets replaced by the concept of possibility. Whereas a probabilistic theory assigns a state-dependent probability value to each outcome of each measurement, a possibilistic theory merely assigns one of the state-dependent labels ''possible to occur'' or ''impossible to occur'' to each outcome of each measurement. It is argued that Spekkens' combinatorial toy theory of quantum mechanics is inconsistent in a probabilistic framework, but can be regarded as possibilistic. Then, I introduce the concept of possibilistic local hidden variable models and derive a class of possibilistic Bell inequalities which are violated for the possibilistic Popescu-Rohrlich boxes. The chapter ends with a philosophical discussion on possibilistic vs. probabilistic. It can be argued that, due to better falsifiability properties, a possibilistic theory has higher predictive power than a probabilistic one. Chapter 3: The quantum region for von Neumann measurements with postselection. It is determined under which conditions a probability distribution on a
Waste Package Misload Probability
International Nuclear Information System (INIS)
Knudsen, J.K.
2001-01-01
The objective of this calculation is to calculate the probability of occurrence for fuel assembly (FA) misloads (i.e., Fa placed in the wrong location) and FA damage during FA movements. The scope of this calculation is provided by the information obtained from the Framatome ANP 2001a report. The first step in this calculation is to categorize each fuel-handling events that occurred at nuclear power plants. The different categories are based on FAs being damaged or misloaded. The next step is to determine the total number of FAs involved in the event. Using the information, a probability of occurrence will be calculated for FA misload and FA damage events. This calculation is an expansion of preliminary work performed by Framatome ANP 2001a
Probability theory and applications
Hsu, Elton P
1999-01-01
This volume, with contributions by leading experts in the field, is a collection of lecture notes of the six minicourses given at the IAS/Park City Summer Mathematics Institute. It introduces advanced graduates and researchers in probability theory to several of the currently active research areas in the field. Each course is self-contained with references and contains basic materials and recent results. Topics include interacting particle systems, percolation theory, analysis on path and loop spaces, and mathematical finance. The volume gives a balanced overview of the current status of probability theory. An extensive bibliography for further study and research is included. This unique collection presents several important areas of current research and a valuable survey reflecting the diversity of the field.
Paradoxes in probability theory
Eckhardt, William
2013-01-01
Paradoxes provide a vehicle for exposing misinterpretations and misapplications of accepted principles. This book discusses seven paradoxes surrounding probability theory. Some remain the focus of controversy; others have allegedly been solved, however the accepted solutions are demonstrably incorrect. Each paradox is shown to rest on one or more fallacies. Instead of the esoteric, idiosyncratic, and untested methods that have been brought to bear on these problems, the book invokes uncontroversial probability principles, acceptable both to frequentists and subjectivists. The philosophical disputation inspired by these paradoxes is shown to be misguided and unnecessary; for instance, startling claims concerning human destiny and the nature of reality are directly related to fallacious reasoning in a betting paradox, and a problem analyzed in philosophy journals is resolved by means of a computer program.
Measurement uncertainty and probability
Willink, Robin
2013-01-01
A measurement result is incomplete without a statement of its 'uncertainty' or 'margin of error'. But what does this statement actually tell us? By examining the practical meaning of probability, this book discusses what is meant by a '95 percent interval of measurement uncertainty', and how such an interval can be calculated. The book argues that the concept of an unknown 'target value' is essential if probability is to be used as a tool for evaluating measurement uncertainty. It uses statistical concepts, such as a conditional confidence interval, to present 'extended' classical methods for evaluating measurement uncertainty. The use of the Monte Carlo principle for the simulation of experiments is described. Useful for researchers and graduate students, the book also discusses other philosophies relating to the evaluation of measurement uncertainty. It employs clear notation and language to avoid the confusion that exists in this controversial field of science.
Model uncertainty and probability
International Nuclear Information System (INIS)
Parry, G.W.
1994-01-01
This paper discusses the issue of model uncertainty. The use of probability as a measure of an analyst's uncertainty as well as a means of describing random processes has caused some confusion, even though the two uses are representing different types of uncertainty with respect to modeling a system. The importance of maintaining the distinction between the two types is illustrated with a simple example
Retrocausality and conditional probability
International Nuclear Information System (INIS)
Stuart, C.I.J.M.
1989-01-01
Costa de Beauregard has proposed that physical causality be identified with conditional probability. The proposal is shown to be vulnerable on two accounts. The first, though mathematically trivial, seems to be decisive so far as the current formulation of the proposal is concerned. The second lies in a physical inconsistency which seems to have its source in a Copenhagenlike disavowal of realism in quantum mechanics. 6 refs. (Author)
Whittle, Peter
1992-01-01
This book is a complete revision of the earlier work Probability which ap peared in 1970. While revised so radically and incorporating so much new material as to amount to a new text, it preserves both the aim and the approach of the original. That aim was stated as the provision of a 'first text in probability, de manding a reasonable but not extensive knowledge of mathematics, and taking the reader to what one might describe as a good intermediate level'. In doing so it attempted to break away from stereotyped applications, and consider applications of a more novel and significant character. The particular novelty of the approach was that expectation was taken as the prime concept, and the concept of expectation axiomatized rather than that of a probability measure. In the preface to the original text of 1970 (reproduced below, together with that to the Russian edition of 1982) I listed what I saw as the advantages of the approach in as unlaboured a fashion as I could. I also took the view that the text...
Spatial probability aids visual stimulus discrimination
Directory of Open Access Journals (Sweden)
Michael Druker
2010-08-01
Full Text Available We investigated whether the statistical predictability of a target's location would influence how quickly and accurately it was classified. Recent results have suggested that spatial probability can be a cue for the allocation of attention in visual search. One explanation for probability cuing is spatial repetition priming. In our two experiments we used probability distributions that were continuous across the display rather than relying on a few arbitrary screen locations. This produced fewer spatial repeats and allowed us to dissociate the effect of a high probability location from that of short-term spatial repetition. The task required participants to quickly judge the color of a single dot presented on a computer screen. In Experiment 1, targets were more probable in an off-center hotspot of high probability that gradually declined to a background rate. Targets garnered faster responses if they were near earlier target locations (priming and if they were near the high probability hotspot (probability cuing. In Experiment 2, target locations were chosen on three concentric circles around fixation. One circle contained 80% of targets. The value of this ring distribution is that it allowed for a spatially restricted high probability zone in which sequentially repeated trials were not likely to be physically close. Participant performance was sensitive to the high-probability circle in addition to the expected effects of eccentricity and the distance to recent targets. These two experiments suggest that inhomogeneities in spatial probability can be learned and used by participants on-line and without prompting as an aid for visual stimulus discrimination and that spatial repetition priming is not a sufficient explanation for this effect. Future models of attention should consider explicitly incorporating the probabilities of targets locations and features.
Elger, Bernice Simone; Handtke, Violet; Wangmo, Tenzin
2015-06-01
This manuscript presents mental health practitioners' (MHPs) practice, attitudes and justifications for breaching confidentiality when imprisoned patients disclose suicidal thoughts or abuse by others. 24 MHPs working in Swiss prisons shared their experiences regarding confidentiality practices. The data were analysed qualitatively and MHPs' attitudes and course of action were identified. Analysis revealed paternalistic breaches of confidentiality. When patients reported suicidal thoughts and abuse, MHPs believed that forgoing confidentiality is necessary to protect patients, providing several justifications for it. Patients were informed that such information will be transmitted without their consent to medical and non-medical prison personnel. With reference to suicidal attempts, MHPs resorted to methods that may reduce suicidal attempts such as transfer to hospital or internal changes in living arrangements, which would require provision of certain information to prison guards. In cases of abuse, some MHPs convinced patients to accept intervention or sometimes overrode competent patients' refusals to report. Also in the case of abuse, provision of limited information to other prison personnel was seen as an acceptable method to protect patients from further harm. Breaches of confidentiality, whether limited or full, remain unethical, when used for competent patients based solely on paternalistic justifications. Institutionalising ethical and legal procedures to address suicidal and abuse situations would be helpful. Education and training to help both medical and prison personnel to respond to such situations in an appropriate manner that ensures confidentiality and protects patients from suicide and abuse are necessary. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.
Law society breaches competition rules over financial regulation training for conveyancers
Johnson, D.
2017-01-01
The article considers the impact of a competition law ruling against the Law Society of England and Wales. \\ud \\ud The Law Society was found to have breached UK competition law rules in relation to its provision of anti-money laundering and mortgage fraud training courses to law firms. The Law Society made it a condition of membership of its Quality Conveyancing Scheme that all law firm members must only receive this training from the Law Society. A competing provider of legal training course...
Breach of information duties in the B2C e-commerce: adequacy of available remedies
Directory of Open Access Journals (Sweden)
Zofia Bednarz
2016-07-01
Full Text Available
B2C e-commerce is characterised by the information asymmetry between the contracting parties. Various information duties are imposed on traders, both at the European and national level to correct this asymmetry and to ensure proper market functioning. The mandated disclosure is based on the assumption of consumers' rationality. However, developments of behavioural economics challenge this assumption. The utility of mandated disclosure in consumer contracts depends also on the remedies available to consumers in a case of breach of information duties. Those remedies are often heavily influenced by the national general private law applicable to the contractual relationship between the parties. Nevertheless, since the economics of general contract law differ importantly from principles of consumer e-commerce, various problems can be associated with the application of general law remedies to the breach of information duties in B2C contracts. The limited value of the majority of the online B2C transactions is incompatible with costly and lengthy court proceedings. Moreover, breach of information duties will often not produce enough material damage on the side of the consumer to make the remedies available. Different solutions are explored, from ADR, to the duty to advise, to non-legal mechanisms making the information easier to use for consumers throughlimiting disclosure. Finally, the right of withdrawal is analysed as an example of a specific remedy, adapted to the economics of the B2C electronic transactions, where the aims parties pursue through contracts are different than in commercial contracts, and their relationship is marked with the inequality of economic power and information asymmetry. However, the legally established cooling-off period is not free from limitations, and only a combination of various measures, including effective
Breach of Personal Security through Applicative use of Online Social Networks
Directory of Open Access Journals (Sweden)
Bojan Nikolovski
2013-11-01
Full Text Available Throughout this article there is an attempt to indicate the threats of potential to breach of personal security through applicative use of internet as well as applicative use of online social networks. In addition to many other ways of privacy protection applicative users of social network’s sites must take into considerations the risk of distributing private data. Through a series of actions and settings users can customize the security settings with the ultimate goal of reducing the risk of attack on their privacy.
Lessons learned from a privacy breach at an academic health science centre.
Malonda, Jacqueline; Campbell, Janice; Crivianu-Gaita, Daniela; Freedman, Melvin H; Stevens, Polly; Laxer, Ronald M
2009-01-01
In 2007, the Hospital for Sick Children experienced a serious privacy breach when a laptop computer containing the personal health information of approximately 3,000 patients and research subjects was stolen from a physician-researcher's vehicle. This incident was reported to the information and privacy commissioner of Ontario (IPC). The IPC issued an order that required the hospital to examine and revise its policies, practices and research protocols related to the protection of personal health information and to educate staff on privacy-related matters.
Pressure pulses generated by gas released from a breached fuel element
International Nuclear Information System (INIS)
Wu, T.S.
1979-01-01
In experimental measurements of liquid pressure pulses generated by rapid release of gas from breached fuel elements in a nuclear reactor, different peak pressures were observed at locations equidistant from the origin of the release. Using the model of a submerged spherical bubble with a nonstationary center, this analysis predicts not only that the peak pressure would be higher at a point in front of the advancing bubble than that at a point the same distance behind the bubble origin, but also that the pressure pulse in front of the bubble reaches its peak later than the pulse behind the origin
Establishing breach of the duty of care in the tort of negligence: 2.
Tingle, John
This article discusses the law surrounding breach of the duty of care in negligence. A mistake or error does not necessarily mean legal fault and negligence. Judges look at risks and benefits in determining what would have been the appropriate standard of care to be exercised in the circumstances and may decide that the defendant's conduct was reasonable. There are a number of interrelated factors which judges have to balance and these can be categorized as foreseeability of harm, magnitude of risk, burden of taking precautions, utility of the defendant's conduct and common practice.
Energy Technology Data Exchange (ETDEWEB)
Singh, Gyanender, E-mail: sing0550@umn.edu [Department of Mechanical Engineering, University of Minnesota, 111, Church St. SE, Minneapolis, MN 55455 (United States); Fok, Alex [Minnesota Dental Research in Biomaterials and Biomechanics, School of Dentistry, University of Minnesota, 515, Delaware St. SE, Minneapolis, MN 55455 (United States); Department of Mechanical Engineering, University of Minnesota, 111, Church St. SE, Minneapolis, MN 55455 (United States); Mantell, Susan [Department of Mechanical Engineering, University of Minnesota, 111, Church St. SE, Minneapolis, MN 55455 (United States)
2017-06-15
Highlights: • Failure probability of VHTR reflector bricks predicted though crack modeling. • Criterion chosen for defining failure strongly affects the predictions. • Breaching of the CRC could be significantly delayed through crack arrest. • Capability to predict crack initiation and propagation demonstrated. - Abstract: Graphite is used in nuclear reactor cores as a neutron moderator, reflector and structural material. The dimensions and physical properties of graphite change when it is exposed to neutron irradiation. The non-uniform changes in the dimensions and physical properties lead to the build-up of stresses over the course of time in the core components. When the stresses reach the critical limit, i.e. the strength of the material, cracking occurs and ultimately the components fail. In this paper, an explicit crack modeling approach to predict the probability of failure of a VHTR prismatic reactor core reflector brick is presented. Firstly, a constitutive model for graphite is constructed and used to predict the stress distribution in the reflector brick under in-reactor conditions of high temperature and irradiation. Fracture simulations are performed as part of a Monte Carlo analysis to predict the probability of failure. Failure probability is determined based on two different criteria for defining failure time: A) crack initiation and B) crack extension to near control rod channel. A significant difference is found between the failure probabilities based on the two criteria. It is predicted that the reflector bricks will start cracking during the time range of 5–9 years, while breaching of the control rod channels will occur during the period of 11–16 years. The results show that, due to crack arrest, there is a significantly delay between crack initiation and breaching of the control rod channel.
Quebras contratuais e dispersão de sentenças Contractual breaches and sentences dispersion
Directory of Open Access Journals (Sweden)
Christiane Leles Rezende
2011-06-01
Full Text Available O problema que motivou um estudo, o qual deu origem a este artigo, foram as quebras contratuais por parte dos produtores rurais geradas pela expressiva alta dos preços da soja e as consequentes disputas judiciais. Foram realizadas análises descritiva e econométrica utilizando 161 Apelações do Tribunal de Justiça de Goiás, e uma pesquisa quantitativa com 70 produtores rurais. O estudo considera a hipótese de que a instabilidade gerada a partir das decisões judiciais eleva os custos de transação e afeta as decisões dos agentes privados. Foi constatada larga dispersão entre decisões de primeira e segunda instâncias, bem como entre Câmaras Cíveis do TJ. Os agentes econômicos relataram que as alterações nas estratégias de suprimento foram centradas no aumento da exigência de garantias e redução do número de contratos. O conceito de função social do contrato está associado à elevação da instabilidade. Decorreram maiores custos de transação, bem como a adoção de sanções econômicas por parte dos agentes privados.The problems which have motivated a study and this article were the contractual breaches during an expressive increase of price and their judicial decisions. Descriptive and econometric analysis have been carried out on 161 appeal judicial decisions of Goiás Court of Justice (Brazil and a quantitative survey was done with 70 farmers. The study has supported the hypothesis that a weak judiciary increases transaction costs and decreases the economic development rate. A large dispersion of court decisions was found between the first and appeal decision, as well inside the same level. The use of the concept of “social role of contract” inserted a high degree of instability in contracts the effects of court decisions could be realized such as more requirements of warranties and the reduction in the number of contracts. Those soybean producers who did not breach their contracts also have been negatively
Probability mapping of contaminants
Energy Technology Data Exchange (ETDEWEB)
Rautman, C.A.; Kaplan, P.G. [Sandia National Labs., Albuquerque, NM (United States); McGraw, M.A. [Univ. of California, Berkeley, CA (United States); Istok, J.D. [Oregon State Univ., Corvallis, OR (United States); Sigda, J.M. [New Mexico Inst. of Mining and Technology, Socorro, NM (United States)
1994-04-01
Exhaustive characterization of a contaminated site is a physical and practical impossibility. Descriptions of the nature, extent, and level of contamination, as well as decisions regarding proposed remediation activities, must be made in a state of uncertainty based upon limited physical sampling. The probability mapping approach illustrated in this paper appears to offer site operators a reasonable, quantitative methodology for many environmental remediation decisions and allows evaluation of the risk associated with those decisions. For example, output from this approach can be used in quantitative, cost-based decision models for evaluating possible site characterization and/or remediation plans, resulting in selection of the risk-adjusted, least-cost alternative. The methodology is completely general, and the techniques are applicable to a wide variety of environmental restoration projects. The probability-mapping approach is illustrated by application to a contaminated site at the former DOE Feed Materials Production Center near Fernald, Ohio. Soil geochemical data, collected as part of the Uranium-in-Soils Integrated Demonstration Project, have been used to construct a number of geostatistical simulations of potential contamination for parcels approximately the size of a selective remediation unit (the 3-m width of a bulldozer blade). Each such simulation accurately reflects the actual measured sample values, and reproduces the univariate statistics and spatial character of the extant data. Post-processing of a large number of these equally likely statistically similar images produces maps directly showing the probability of exceeding specified levels of contamination (potential clean-up or personnel-hazard thresholds).
Probability mapping of contaminants
International Nuclear Information System (INIS)
Rautman, C.A.; Kaplan, P.G.; McGraw, M.A.; Istok, J.D.; Sigda, J.M.
1994-01-01
Exhaustive characterization of a contaminated site is a physical and practical impossibility. Descriptions of the nature, extent, and level of contamination, as well as decisions regarding proposed remediation activities, must be made in a state of uncertainty based upon limited physical sampling. The probability mapping approach illustrated in this paper appears to offer site operators a reasonable, quantitative methodology for many environmental remediation decisions and allows evaluation of the risk associated with those decisions. For example, output from this approach can be used in quantitative, cost-based decision models for evaluating possible site characterization and/or remediation plans, resulting in selection of the risk-adjusted, least-cost alternative. The methodology is completely general, and the techniques are applicable to a wide variety of environmental restoration projects. The probability-mapping approach is illustrated by application to a contaminated site at the former DOE Feed Materials Production Center near Fernald, Ohio. Soil geochemical data, collected as part of the Uranium-in-Soils Integrated Demonstration Project, have been used to construct a number of geostatistical simulations of potential contamination for parcels approximately the size of a selective remediation unit (the 3-m width of a bulldozer blade). Each such simulation accurately reflects the actual measured sample values, and reproduces the univariate statistics and spatial character of the extant data. Post-processing of a large number of these equally likely statistically similar images produces maps directly showing the probability of exceeding specified levels of contamination (potential clean-up or personnel-hazard thresholds)
Probability of causation approach
International Nuclear Information System (INIS)
Jose, D.E.
1988-01-01
Probability of causation (PC) is sometimes viewed as a great improvement by those persons who are not happy with the present rulings of courts in radiation cases. The author does not share that hope and expects that PC will not play a significant role in these issues for at least the next decade. If it is ever adopted in a legislative compensation scheme, it will be used in a way that is unlikely to please most scientists. Consequently, PC is a false hope for radiation scientists, and its best contribution may well lie in some of the spin-off effects, such as an influence on medical practice
Generalized Probability Functions
Directory of Open Access Journals (Sweden)
Alexandre Souto Martinez
2009-01-01
Full Text Available From the integration of nonsymmetrical hyperboles, a one-parameter generalization of the logarithmic function is obtained. Inverting this function, one obtains the generalized exponential function. Motivated by the mathematical curiosity, we show that these generalized functions are suitable to generalize some probability density functions (pdfs. A very reliable rank distribution can be conveniently described by the generalized exponential function. Finally, we turn the attention to the generalization of one- and two-tail stretched exponential functions. We obtain, as particular cases, the generalized error function, the Zipf-Mandelbrot pdf, the generalized Gaussian and Laplace pdf. Their cumulative functions and moments were also obtained analytically.
International Nuclear Information System (INIS)
Shimada, Yoshio
2000-01-01
It is anticipated that the change of frequency of surveillance tests, preventive maintenance or parts replacement of safety related components may cause the change of component failure probability and result in the change of core damage probability. It is also anticipated that the change is different depending on the initiating event frequency or the component types. This study assessed the change of core damage probability using simplified PSA model capable of calculating core damage probability in a short time period, which is developed by the US NRC to process accident sequence precursors, when various component's failure probability is changed between 0 and 1, or Japanese or American initiating event frequency data are used. As a result of the analysis, (1) It was clarified that frequency of surveillance test, preventive maintenance or parts replacement of motor driven pumps (high pressure injection pumps, residual heat removal pumps, auxiliary feedwater pumps) should be carefully changed, since the core damage probability's change is large, when the base failure probability changes toward increasing direction. (2) Core damage probability change is insensitive to surveillance test frequency change, since the core damage probability change is small, when motor operated valves and turbine driven auxiliary feed water pump failure probability changes around one figure. (3) Core damage probability change is small, when Japanese failure probability data are applied to emergency diesel generator, even if failure probability changes one figure from the base value. On the other hand, when American failure probability data is applied, core damage probability increase is large, even if failure probability changes toward increasing direction. Therefore, when Japanese failure probability data is applied, core damage probability change is insensitive to surveillance tests frequency change etc. (author)
Kogelenberg, Martijn
2012-01-01
textabstractThis thesis explores the notion of deliberate breach of contract and its potential remedial consequences. In the major jurisdictions in Europe and in the United States the notion of deliberate breach of contract is generally not coherently and officially defined and acknowledged as an independent legal phenomenon. The ultimate added value of this thesis intends to be a first coherent comparative research on deliberate breach of contract and its potential consequences for the core ...
Striatal activity is modulated by target probability.
Hon, Nicholas
2017-06-14
Target probability has well-known neural effects. In the brain, target probability is known to affect frontal activity, with lower probability targets producing more prefrontal activation than those that occur with higher probability. Although the effect of target probability on cortical activity is well specified, its effect on subcortical structures such as the striatum is less well understood. Here, I examined this issue and found that the striatum was highly responsive to target probability. This is consistent with its hypothesized role in the gating of salient information into higher-order task representations. The current data are interpreted in light of that fact that different components of the striatum are sensitive to different types of task-relevant information.
Collmann, Jeff; Cooper, Ted
2007-01-01
This case study describes and analyzes a breach of the confidentiality and integrity of personally identified health information (e.g. appointment details, answers to patients' questions, medical advice) for over 800 Kaiser Permanente (KP) members through KP Online, a web-enabled health care portal. The authors obtained and analyzed multiple types of qualitative data about this incident including interviews with KP staff, incident reports, root cause analyses, and media reports. Reasons at multiple levels account for the breach, including the architecture of the information system, the motivations of individual staff members, and differences among the subcultures of individual groups within as well as technical and social relations across the Kaiser IT program. None of these reasons could be classified, strictly speaking, as "security violations." This case study, thus, suggests that, to protect sensitive patient information, health care organizations should build safe organizational contexts for complex health information systems in addition to complying with good information security practice and regulations such as the Health Insurance Portability and Accountability Act (HIPAA) of 1996.
Probable maximum flood control
International Nuclear Information System (INIS)
DeGabriele, C.E.; Wu, C.L.
1991-11-01
This study proposes preliminary design concepts to protect the waste-handling facilities and all shaft and ramp entries to the underground from the probable maximum flood (PMF) in the current design configuration for the proposed Nevada Nuclear Waste Storage Investigation (NNWSI) repository protection provisions were furnished by the United States Bureau of Reclamation (USSR) or developed from USSR data. Proposed flood protection provisions include site grading, drainage channels, and diversion dikes. Figures are provided to show these proposed flood protection provisions at each area investigated. These areas are the central surface facilities (including the waste-handling building and waste treatment building), tuff ramp portal, waste ramp portal, men-and-materials shaft, emplacement exhaust shaft, and exploratory shafts facility
Nelson, Timothy R.; Miselis, Jennifer L.; Hapke, Cheryl J.; Brenner, Owen T.; Henderson, Rachel E.; Reynolds, Billy J.; Wilson, Kathleen E.
2017-05-12
Scientists from the U.S. Geological Survey (USGS) St. Petersburg Coastal and Marine Science Center in St. Petersburg, Florida, conducted a bathymetric survey of Fire Island from May 6-20, 2015. The USGS is involved in a post-Hurricane Sandy effort to map and monitor the morphologic evolution of the wilderness breach as a part of the Hurricane Sandy Supplemental Project GS2-2B. During this study, bathymetry data were collected with single-beam echo sounders and Global Positioning Systems, which were mounted to personal watercraft, along the Fire Island shoreface and within the wilderness breach. Additional bathymetry and elevation data were collected using backpack Global Positioning Systems on flood shoals and in shallow channels within the wilderness breach.
Muths, Erin L.; Scherer, R. D.; Amburgey, S. M.; Matthews, T.; Spencer, A. W.; Corn, P.S.
2016-01-01
In an era of shrinking budgets yet increasing demands for conservation, the value of existing (i.e., historical) data are elevated. Lengthy time series on common, or previously common, species are particularly valuable and may be available only through the use of historical information. We provide first estimates of the probability of survival and longevity (0.67–0.79 and 5–7 years, respectively) for a subalpine population of a small-bodied, ostensibly common amphibian, the Boreal Chorus Frog (Pseudacris maculata (Agassiz, 1850)), using historical data and contemporary, hypothesis-driven information–theoretic analyses. We also test a priori hypotheses about the effects of color morph (as suggested by early reports) and of drought (as suggested by recent climate predictions) on survival. Using robust mark–recapture models, we find some support for early hypotheses regarding the effect of color on survival, but we find no effect of drought. The congruence between early findings and our analyses highlights the usefulness of historical information in providing raw data for contemporary analyses and context for conservation and management decisions.
Probability concepts in quality risk management.
Claycamp, H Gregg
2012-01-01
Essentially any concept of risk is built on fundamental concepts of chance, likelihood, or probability. Although risk is generally a probability of loss of something of value, given that a risk-generating event will occur or has occurred, it is ironic that the quality risk management literature and guidelines on quality risk management tools are relatively silent on the meaning and uses of "probability." The probability concept is typically applied by risk managers as a combination of frequency-based calculation and a "degree of belief" meaning of probability. Probability as a concept that is crucial for understanding and managing risk is discussed through examples from the most general, scenario-defining and ranking tools that use probability implicitly to more specific probabilistic tools in risk management. A rich history of probability in risk management applied to other fields suggests that high-quality risk management decisions benefit from the implementation of more thoughtful probability concepts in both risk modeling and risk management. Essentially any concept of risk is built on fundamental concepts of chance, likelihood, or probability. Although "risk" generally describes a probability of loss of something of value, given that a risk-generating event will occur or has occurred, it is ironic that the quality risk management literature and guidelines on quality risk management methodologies and respective tools focus on managing severity but are relatively silent on the in-depth meaning and uses of "probability." Pharmaceutical manufacturers are expanding their use of quality risk management to identify and manage risks to the patient that might occur in phases of the pharmaceutical life cycle from drug development to manufacture, marketing to product discontinuation. A probability concept is typically applied by risk managers as a combination of data-based measures of probability and a subjective "degree of belief" meaning of probability. Probability as
Probability and rational choice
Directory of Open Access Journals (Sweden)
David Botting
2014-05-01
Full Text Available http://dx.doi.org/10.5007/1808-1711.2014v18n1p1 In this paper I will discuss the rationality of reasoning about the future. There are two things that we might like to know about the future: which hypotheses are true and what will happen next. To put it in philosophical language, I aim to show that there are methods by which inferring to a generalization (selecting a hypothesis and inferring to the next instance (singular predictive inference can be shown to be normative and the method itself shown to be rational, where this is due in part to being based on evidence (although not in the same way and in part on a prior rational choice. I will also argue that these two inferences have been confused, being distinct not only conceptually (as nobody disputes but also in their results (the value given to the probability of the hypothesis being not in general that given to the next instance and that methods that are adequate for one are not by themselves adequate for the other. A number of debates over method founder on this confusion and do not show what the debaters think they show.
COVAL, Compound Probability Distribution for Function of Probability Distribution
International Nuclear Information System (INIS)
Astolfi, M.; Elbaz, J.
1979-01-01
1 - Nature of the physical problem solved: Computation of the probability distribution of a function of variables, given the probability distribution of the variables themselves. 'COVAL' has been applied to reliability analysis of a structure subject to random loads. 2 - Method of solution: Numerical transformation of probability distributions
Results of transient overpower events on breached and unbreached fuel pins
International Nuclear Information System (INIS)
Strain, R.V.; Tsai, H.C.; Neimark, L.A.; Aratani, K.
1986-04-01
The objective of the extended overpower tests on intact pins was to determine the pin cladding breaching thresholds vis-a-vis the Plant Protection System (PPS) trip settings, typically at ∼10 to 15% overpower. These tests emphasize slow operational-type transients in light of earlier work which suggested that irradiated mixed-oxide fuel pins may be particularly vulnerable in the slow ramp-rate regime. An overview of the extended overpower test series was previously reported. More recent results on two of the tests in this series are included in this paper. These two tests, designated TOPI-1A and TOPI-1B, were each conducted on a 19-pin assembly with various pin design, operation and burnup variables. The overpower ramp rates for the TOPI-1A and -1B tests were 0.1%/s and 10%/s, respectively
Fuel-sodium reaction product formation in breached mixed-oxide fuel
International Nuclear Information System (INIS)
Bottcher, J.H.; Lambert, J.D.B.; Strain, R.V.; Ukai, S.; Shibahara, S.
1988-01-01
The run-beyond-cladding-breach (RBCB) operation of mixed-oxide LMR fuel pins has been studied for six years in the Experimental Breeder Reactor-II (EBR-II) as part of a joint program between the US Department of Energy and the Power Reactor and Nuclear Fuel Development Corporation of Japan. The formation of fuel-sodium reaction product (FSRP), Na 3 MO 4 , where M = U/sub 1-y/Pu/sub y/, in the outer fuel regions is the major phenomenon governing RBCB behavior. It increases fuel volume, decreases fuel stoichiometry, modifies fission-product distributions, and alters thermal performance of a pin. This paper describes the morphology of Na 3 MO 4 observed in 5.84-mm diameter pins covering a variety of conditions and RBCB times up to 150 EFPD's. 8 refs., 1 fig
Directory of Open Access Journals (Sweden)
Mareike Reimann
2017-04-01
Full Text Available This study examines the effects of psychological contract breach (PCB on employee mental and physical health (SF-12 using a sample of 3,870 employees derived from a German longitudinal linked employer-employee study across various industries. Results of multivariate regression models and mediation analysis suggest that PCB affects both the mental and the physical health of employees but is more threatening to employee mental health. In addition, mental health partly mediates the effects of PCB on physical health. Also, the findings of this study show that the relative importance of obligations not met by employers differs according to the specific contents of the psychological contract. In conclusion, the results of this study support the idea that PCB works as a psychosocial stressor at work that represents a crucial risk to employee health.
LMX, Breach Perceptions, Work-Family Conflict, and Well-Being: A Mediational Model.
Hill, Rachel T; Morganson, Valerie J; Matthews, Russell A; Atkinson, Theresa P
2016-01-01
Despite research advances, work-family scholars still lack an understanding of how leadership constructs relate to an employee's ability to effectively manage the work-family interface. In addition, there remains a need to examine the process through which leadership and work-family conflict influence well-being outcomes. Using a sample of 312 workers, a mediated process model grounded in social exchange theory is tested wherein the authors seek to explain how leaders shape employee perceptions, which, in turn, impact organizational fulfillment of expectations (i.e., psychological contract breach), work-family conflict, and well-being. A fully latent structural equation model was used to test study hypotheses, all of which were supported. Building on existing theory, findings suggest that the supervisor plays a critical role as a frontline representative for the organization and that work-family conflict is reduced and well-being enhanced through a process of social exchange between the supervisor and worker.
Consistency assessment of rating curve data in various locations using Bidirectional Reach (BReach)
Van Eerdenbrugh, Katrien; Van Hoey, Stijn; Coxon, Gemma; Freer, Jim; Verhoest, Niko E. C.
2017-10-01
When estimating discharges through rating curves, temporal data consistency is a critical issue. In this research, consistency in stage-discharge data is investigated using a methodology called Bidirectional Reach (BReach), which departs from a (in operational hydrology) commonly used definition of consistency. A period is considered to be consistent if no consecutive and systematic deviations from a current situation occur that exceed observational uncertainty. Therefore, the capability of a rating curve model to describe a subset of the (chronologically sorted) data is assessed in each observation by indicating the outermost data points for which the rating curve model behaves satisfactorily. These points are called the maximum left or right reach, depending on the direction of the investigation. This temporal reach should not be confused with a spatial reach (indicating a part of a river). Changes in these reaches throughout the data series indicate possible changes in data consistency and if not resolved could introduce additional errors and biases. In this research, various measurement stations in the UK, New Zealand and Belgium are selected based on their significant historical ratings information and their specific characteristics related to data consistency. For each country, regional information is maximally used to estimate observational uncertainty. Based on this uncertainty, a BReach analysis is performed and, subsequently, results are validated against available knowledge about the history and behavior of the site. For all investigated cases, the methodology provides results that appear to be consistent with this knowledge of historical changes and thus facilitates a reliable assessment of (in)consistent periods in stage-discharge measurements. This assessment is not only useful for the analysis and determination of discharge time series, but also to enhance applications based on these data (e.g., by informing hydrological and hydraulic model
Semantic-less Breach Detection of Polymorphic Malware in Federated Cloud
Directory of Open Access Journals (Sweden)
Yahav Biran
2017-06-01
Full Text Available Cloud computing is one of the largest emerging utility services that is expected to grow enormously over the next decade. Many organizations are moving into hybrid cloud/hosted computing models. Single cloud service provider introduces cost and environmental challenges. Also, multi-cloud solution implemented by the Cloud tenant is suboptimal as it requires expensive adaptation costs. Cloud Federation is a useful structure for aggregating cloud based services under a single umbrella to share resources and responsibilities for the benefit of the member cloud service providers. An efficient security model is crucial for successful cloud business. However, with the advent of large scale and multi-tenant environments, the traditional perimeter boundaries along with traditional security practices are changing. Defining and securing asset and enclave boundaries is more challenging, and system perimeter boundaries are more susceptible to breach. This paper to describe security best practices for Cloud Federation. The paper also describes a tool and technique for detecting anomalous behavior in resource usage across the federation participants. This is a particularly serious issue because of the possibility of an attacker potentially gaining access to more than one CSP federation member. Specifically, this technique is developed for Cloud Federations since they have to deal with heterogeneous multi-platform environments with a diverse mixture of data and security log schema, and it has to do this in real time. A Semantic-less Breach detection system that implements a self-learning system was prototyped and resulted in up to 87% True-Positive rate with 93% True-Negative.
M. van Kogelenberg (Martijn)
2012-01-01
textabstractThis thesis explores the notion of deliberate breach of contract and its potential remedial consequences. In the major jurisdictions in Europe and in the United States the notion of deliberate breach of contract is generally not coherently and officially defined and acknowledged as an
Data Quality Campaign, 2011
2011-01-01
Under security breach response laws, businesses--and sometimes state and governmental agencies--are required to inform individuals when the security, confidentiality or integrity of their personal information has been compromised. This resource provides a state-by-state analysis of security breach response laws. [The Data Quality Campaign has…
Coronado, Adolfo S.
2012-01-01
Using a sample of security and privacy breaches the present research examines the comparative announcement impact between the two types of events. The first part of the dissertation analyzes the impact of publicly announced security and privacy breaches on abnormal stock returns, the change in firm risk, and abnormal trading volume are measured.…
Ahmad, Altaf
2010-01-01
This dissertation involved the development of a new construct, perceived privacy breach (PPB), to evaluate how a person perceives breaches of privacy in terms of whether they perceive any exchange of information was fair or not and how they believe it will impact people whose information has been shared. . This instrument assists researchers to…
Assessing the clinical probability of pulmonary embolism
International Nuclear Information System (INIS)
Miniati, M.; Pistolesi, M.
2001-01-01
Clinical assessment is a cornerstone of the recently validated diagnostic strategies for pulmonary embolism (PE). Although the diagnostic yield of individual symptoms, signs, and common laboratory tests is limited, the combination of these variables, either by empirical assessment or by a prediction rule, can be used to express a clinical probability of PE. The latter may serve as pretest probability to predict the probability of PE after further objective testing (posterior or post-test probability). Over the last few years, attempts have been made to develop structured prediction models for PE. In a Canadian multicenter prospective study, the clinical probability of PE was rated as low, intermediate, or high according to a model which included assessment of presenting symptoms and signs, risk factors, and presence or absence of an alternative diagnosis at least as likely as PE. Recently, a simple clinical score was developed to stratify outpatients with suspected PE into groups with low, intermediate, or high clinical probability. Logistic regression was used to predict parameters associated with PE. A score ≤ 4 identified patients with low probability of whom 10% had PE. The prevalence of PE in patients with intermediate (score 5-8) and high probability (score ≥ 9) was 38 and 81%, respectively. As opposed to the Canadian model, this clinical score is standardized. The predictor variables identified in the model, however, were derived from a database of emergency ward patients. This model may, therefore, not be valid in assessing the clinical probability of PE in inpatients. In the PISA-PED study, a clinical diagnostic algorithm was developed which rests on the identification of three relevant clinical symptoms and on their association with electrocardiographic and/or radiographic abnormalities specific for PE. Among patients who, according to the model, had been rated as having a high clinical probability, the prevalence of proven PE was 97%, while it was 3
Falk, Ruma; Kendig, Keith
2013-01-01
Two contestants debate the notorious probability problem of the sex of the second child. The conclusions boil down to explication of the underlying scenarios and assumptions. Basic principles of probability theory are highlighted.
D'Silva, J.J.F.; Robinson, D.K.R.; Shelley Egan, Clare
2012-01-01
The potential benefits of nanotechnologies in healthcare are widely expected to be enormous and a considerable amount of investment is already pouring into public research in this area. These high expectations of benefits are coupled with uncertainty surrounding the potential risks of the
Yin, Bing; Li, Teng; Li, Jin-Feng; Yu, Yang; Li, Jian-Li; Wen, Zhen-Yi; Jiang, Zhen-Yi
2014-03-01
The first theoretical exploration of superhalogen properties of polynuclear structures based on pseudohalogen ligand is reported here via a case study on eight triply-bridged [Mg2(CN)5]- clusters. From our high-level ab initio results, all these clusters are superhalogens due to their high vertical electron detachment energies (VDE), of which the largest value is 8.67 eV at coupled-cluster single double triple (CCSD(T)) level. Although outer valence Green's function results are consistent with CCSD(T) in most cases, it overestimates the VDEs of three anions dramatically by more than 1 eV. Therefore, the combined usage of several theoretical methods is important for the accuracy of purely theoretical prediction of superhalogen properties of new structures. Spatial distribution of the extra electron of high-VDE anions here indicates two features: remarkable aggregation on bridging CN units and non-negligible distribution on every CN unit. These two features lower the potential and kinetic energies of the extra electron respectively and thus lead to high VDE. Besides superhalogen properties, the structures, relative stabilities and thermodynamic stabilities with respect to detachment of CN-1 were also investigated for these anions. The collection of these results indicates that polynuclear structures based on pseudohalogen ligand are promising candidates for new superhalogens with enhanced properties.
International Nuclear Information System (INIS)
Yin, Bing; Wen, Zhen-Yi; Li, Teng; Li, Jin-Feng; Yu, Yang; Li, Jian-Li; Jiang, Zhen-Yi
2014-01-01
The first theoretical exploration of superhalogen properties of polynuclear structures based on pseudohalogen ligand is reported here via a case study on eight triply-bridged [Mg 2 (CN) 5 ] − clusters. From our high-level ab initio results, all these clusters are superhalogens due to their high vertical electron detachment energies (VDE), of which the largest value is 8.67 eV at coupled-cluster single double triple (CCSD(T)) level. Although outer valence Green's function results are consistent with CCSD(T) in most cases, it overestimates the VDEs of three anions dramatically by more than 1 eV. Therefore, the combined usage of several theoretical methods is important for the accuracy of purely theoretical prediction of superhalogen properties of new structures. Spatial distribution of the extra electron of high-VDE anions here indicates two features: remarkable aggregation on bridging CN units and non-negligible distribution on every CN unit. These two features lower the potential and kinetic energies of the extra electron respectively and thus lead to high VDE. Besides superhalogen properties, the structures, relative stabilities and thermodynamic stabilities with respect to detachment of CN −1 were also investigated for these anions. The collection of these results indicates that polynuclear structures based on pseudohalogen ligand are promising candidates for new superhalogens with enhanced properties
Energy Technology Data Exchange (ETDEWEB)
Yin, Bing, E-mail: rayinyin@gmail.com; Wen, Zhen-Yi [MOE Key Laboratory of Synthetic and Natural Functional Molecule Chemistry, Shaanxi Key Laboratory of Physico-Inorganic Chemistry, College of Chemistry and Materials Science, Northwest University, Xi' an 710069 (China); Institute of Modern Physics, Northwest University, Xi' an 710069 (China); Li, Teng; Li, Jin-Feng; Yu, Yang; Li, Jian-Li [MOE Key Laboratory of Synthetic and Natural Functional Molecule Chemistry, Shaanxi Key Laboratory of Physico-Inorganic Chemistry, College of Chemistry and Materials Science, Northwest University, Xi' an 710069 (China); Jiang, Zhen-Yi [Institute of Modern Physics, Northwest University, Xi' an 710069 (China)
2014-03-07
The first theoretical exploration of superhalogen properties of polynuclear structures based on pseudohalogen ligand is reported here via a case study on eight triply-bridged [Mg{sub 2}(CN){sub 5}]{sup −} clusters. From our high-level ab initio results, all these clusters are superhalogens due to their high vertical electron detachment energies (VDE), of which the largest value is 8.67 eV at coupled-cluster single double triple (CCSD(T)) level. Although outer valence Green's function results are consistent with CCSD(T) in most cases, it overestimates the VDEs of three anions dramatically by more than 1 eV. Therefore, the combined usage of several theoretical methods is important for the accuracy of purely theoretical prediction of superhalogen properties of new structures. Spatial distribution of the extra electron of high-VDE anions here indicates two features: remarkable aggregation on bridging CN units and non-negligible distribution on every CN unit. These two features lower the potential and kinetic energies of the extra electron respectively and thus lead to high VDE. Besides superhalogen properties, the structures, relative stabilities and thermodynamic stabilities with respect to detachment of CN{sup −1} were also investigated for these anions. The collection of these results indicates that polynuclear structures based on pseudohalogen ligand are promising candidates for new superhalogens with enhanced properties.
Introduction to probability with R
Baclawski, Kenneth
2008-01-01
FOREWORD PREFACE Sets, Events, and Probability The Algebra of Sets The Bernoulli Sample Space The Algebra of Multisets The Concept of Probability Properties of Probability Measures Independent Events The Bernoulli Process The R Language Finite Processes The Basic Models Counting Rules Computing Factorials The Second Rule of Counting Computing Probabilities Discrete Random Variables The Bernoulli Process: Tossing a Coin The Bernoulli Process: Random Walk Independence and Joint Distributions Expectations The Inclusion-Exclusion Principle General Random Variable
Deliyianni, Eleni; Monoyiou, Annita; Elia, Iliada; Georgiou, Chryso; Zannettou, Eleni
2009-01-01
This study investigated the modes of representations generated by kindergarteners and first graders while solving standard and problematic problems in mathematics. Furthermore, it examined the influence of pupils' visual representations on the breach of the didactical contract rules in problem solving. The sample of the study consisted of 38…
Tuan, T.Q.; Stive, M.J.F.; Verhagen, H.J.
2006-01-01
Under effects of tropical cyclones, the coast is subjected to attack both by surge and wave from the sea and by flooding from the bay. These forces pose a serious breaching threat to natural sea-defence works such as barrier spits, barrier islands, lagoon barriers, etc. on the coast. Unintended
Becker, Cinda
2006-02-20
Consumers who claimed in recent surveys that they were "more afraid of cyber crimes than physical crimes" may have had reason for caution. A spate of well-publicized information thefts and security breaches at healthcare organizations have eroded trust in technology, says Carol Diamond, left, of the Markle Foundation, and that could have an adverse effect on acceptance of electronic medical records.
Ross, Sheldon
2014-01-01
A First Course in Probability, Ninth Edition, features clear and intuitive explanations of the mathematics of probability theory, outstanding problem sets, and a variety of diverse examples and applications. This book is ideal for an upper-level undergraduate or graduate level introduction to probability for math, science, engineering and business students. It assumes a background in elementary calculus.
Lectures on probability and statistics
International Nuclear Information System (INIS)
Yost, G.P.
1984-09-01
These notes are based on a set of statistics lectures delivered at Imperial College to the first-year postgraduate students in High Energy Physics. They are designed for the professional experimental scientist. We begin with the fundamentals of probability theory, in which one makes statements about the set of possible outcomes of an experiment, based upon a complete a priori understanding of the experiment. For example, in a roll of a set of (fair) dice, one understands a priori that any given side of each die is equally likely to turn up. From that, we can calculate the probability of any specified outcome. We finish with the inverse problem, statistics. Here, one begins with a set of actual data (e.g., the outcomes of a number of rolls of the dice), and attempts to make inferences about the state of nature which gave those data (e.g., the likelihood of seeing any given side of any given die turn up). This is a much more difficult problem, of course, and one's solutions often turn out to be unsatisfactory in one respect or another
Detection probability of Campylobacter
Evers, E.G.; Post, J.; Putirulan, F.F.; Wal, van der F.J.
2010-01-01
A rapid presence/absence test for Campylobacter in chicken faeces is being evaluated to support the scheduling of highly contaminated broiler flocks as a measure to reduce public health risks [Nauta, M. J., & Havelaar, A. H. (2008). Risk-based standards for Campylobacter in the broiler meat
Probability an introduction with statistical applications
Kinney, John J
2014-01-01
Praise for the First Edition""This is a well-written and impressively presented introduction to probability and statistics. The text throughout is highly readable, and the author makes liberal use of graphs and diagrams to clarify the theory."" - The StatisticianThoroughly updated, Probability: An Introduction with Statistical Applications, Second Edition features a comprehensive exploration of statistical data analysis as an application of probability. The new edition provides an introduction to statistics with accessible coverage of reliability, acceptance sampling, confidence intervals, h
Economic choices reveal probability distortion in macaque monkeys.
Stauffer, William R; Lak, Armin; Bossaerts, Peter; Schultz, Wolfram
2015-02-18
Economic choices are largely determined by two principal elements, reward value (utility) and probability. Although nonlinear utility functions have been acknowledged for centuries, nonlinear probability weighting (probability distortion) was only recently recognized as a ubiquitous aspect of real-world choice behavior. Even when outcome probabilities are known and acknowledged, human decision makers often overweight low probability outcomes and underweight high probability outcomes. Whereas recent studies measured utility functions and their corresponding neural correlates in monkeys, it is not known whether monkeys distort probability in a manner similar to humans. Therefore, we investigated economic choices in macaque monkeys for evidence of probability distortion. We trained two monkeys to predict reward from probabilistic gambles with constant outcome values (0.5 ml or nothing). The probability of winning was conveyed using explicit visual cues (sector stimuli). Choices between the gambles revealed that the monkeys used the explicit probability information to make meaningful decisions. Using these cues, we measured probability distortion from choices between the gambles and safe rewards. Parametric modeling of the choices revealed classic probability weighting functions with inverted-S shape. Therefore, the animals overweighted low probability rewards and underweighted high probability rewards. Empirical investigation of the behavior verified that the choices were best explained by a combination of nonlinear value and nonlinear probability distortion. Together, these results suggest that probability distortion may reflect evolutionarily preserved neuronal processing. Copyright © 2015 Stauffer et al.
A brief introduction to probability.
Di Paola, Gioacchino; Bertani, Alessandro; De Monte, Lavinia; Tuzzolino, Fabio
2018-02-01
The theory of probability has been debated for centuries: back in 1600, French mathematics used the rules of probability to place and win bets. Subsequently, the knowledge of probability has significantly evolved and is now an essential tool for statistics. In this paper, the basic theoretical principles of probability will be reviewed, with the aim of facilitating the comprehension of statistical inference. After a brief general introduction on probability, we will review the concept of the "probability distribution" that is a function providing the probabilities of occurrence of different possible outcomes of a categorical or continuous variable. Specific attention will be focused on normal distribution that is the most relevant distribution applied to statistical analysis.
K-forbidden transition probabilities
International Nuclear Information System (INIS)
Saitoh, T.R.; Sletten, G.; Bark, R.A.; Hagemann, G.B.; Herskind, B.; Saitoh-Hashimoto, N.; Tsukuba Univ., Ibaraki
2000-01-01
Reduced hindrance factors of K-forbidden transitions are compiled for nuclei with A∝180 where γ-vibrational states are observed. Correlations between these reduced hindrance factors and Coriolis forces, statistical level mixing and γ-softness have been studied. It is demonstrated that the K-forbidden transition probabilities are related to γ-softness. The decay of the high-K bandheads has been studied by means of the two-state mixing, which would be induced by the γ-softness, with the use of a number of K-forbidden transitions compiled in the present work, where high-K bandheads are depopulated by both E2 and ΔI=1 transitions. The validity of the two-state mixing scheme has been examined by using the proposed identity of the B(M1)/B(E2) ratios of transitions depopulating high-K bandheads and levels of low-K bands. A break down of the identity might indicate that other levels would mediate transitions between high- and low-K states. (orig.)
Propensity, Probability, and Quantum Theory
Ballentine, Leslie E.
2016-08-01
Quantum mechanics and probability theory share one peculiarity. Both have well established mathematical formalisms, yet both are subject to controversy about the meaning and interpretation of their basic concepts. Since probability plays a fundamental role in QM, the conceptual problems of one theory can affect the other. We first classify the interpretations of probability into three major classes: (a) inferential probability, (b) ensemble probability, and (c) propensity. Class (a) is the basis of inductive logic; (b) deals with the frequencies of events in repeatable experiments; (c) describes a form of causality that is weaker than determinism. An important, but neglected, paper by P. Humphreys demonstrated that propensity must differ mathematically, as well as conceptually, from probability, but he did not develop a theory of propensity. Such a theory is developed in this paper. Propensity theory shares many, but not all, of the axioms of probability theory. As a consequence, propensity supports the Law of Large Numbers from probability theory, but does not support Bayes theorem. Although there are particular problems within QM to which any of the classes of probability may be applied, it is argued that the intrinsic quantum probabilities (calculated from a state vector or density matrix) are most naturally interpreted as quantum propensities. This does not alter the familiar statistical interpretation of QM. But the interpretation of quantum states as representing knowledge is untenable. Examples show that a density matrix fails to represent knowledge.
Failure probability analysis of optical grid
Zhong, Yaoquan; Guo, Wei; Sun, Weiqiang; Jin, Yaohui; Hu, Weisheng
2008-11-01
Optical grid, the integrated computing environment based on optical network, is expected to be an efficient infrastructure to support advanced data-intensive grid applications. In optical grid, the faults of both computational and network resources are inevitable due to the large scale and high complexity of the system. With the optical network based distributed computing systems extensive applied in the processing of data, the requirement of the application failure probability have been an important indicator of the quality of application and an important aspect the operators consider. This paper will present a task-based analysis method of the application failure probability in optical grid. Then the failure probability of the entire application can be quantified, and the performance of reducing application failure probability in different backup strategies can be compared, so that the different requirements of different clients can be satisfied according to the application failure probability respectively. In optical grid, when the application based DAG (directed acyclic graph) is executed in different backup strategies, the application failure probability and the application complete time is different. This paper will propose new multi-objective differentiated services algorithm (MDSA). New application scheduling algorithm can guarantee the requirement of the failure probability and improve the network resource utilization, realize a compromise between the network operator and the application submission. Then differentiated services can be achieved in optical grid.
Prediction and probability in sciences
International Nuclear Information System (INIS)
Klein, E.; Sacquin, Y.
1998-01-01
This book reports the 7 presentations made at the third meeting 'physics and fundamental questions' whose theme was probability and prediction. The concept of probability that was invented to apprehend random phenomena has become an important branch of mathematics and its application range spreads from radioactivity to species evolution via cosmology or the management of very weak risks. The notion of probability is the basis of quantum mechanics and then is bound to the very nature of matter. The 7 topics are: - radioactivity and probability, - statistical and quantum fluctuations, - quantum mechanics as a generalized probability theory, - probability and the irrational efficiency of mathematics, - can we foresee the future of the universe?, - chance, eventuality and necessity in biology, - how to manage weak risks? (A.C.)
CONTAINMENT ANALYSIS METHODOLOGY FOR TRANSPORT OF BREACHED CLAD ALUMINUM SPENT FUEL
Energy Technology Data Exchange (ETDEWEB)
Vinson, D.
2010-07-11
Aluminum-clad, aluminum-based spent nuclear fuel (Al-SNF) from foreign and domestic research reactors (FRR/DRR) is being shipped to the Savannah River Site and placed in interim storage in a water basin. To enter the United States, a cask with loaded fuel must be certified to comply with the requirements in the Title 10 of the U.S. Code of Federal Regulations, Part 71. The requirements include demonstration of containment of the cask with its contents under normal and accident conditions. Many Al-SNF assemblies have suffered corrosion degradation in storage in poor quality water, and many of the fuel assemblies are 'failed' or have through-clad damage. A methodology was developed to evaluate containment of Al-SNF even with severe cladding breaches for transport in standard casks. The containment analysis methodology for Al-SNF is in accordance with the methodology provided in ANSI N14.5 and adopted by the U. S. Nuclear Regulatory Commission in NUREG/CR-6487 to meet the requirements of 10CFR71. The technical bases for the inputs and assumptions are specific to the attributes and characteristics of Al-SNF received from basin and dry storage systems and its subsequent performance under normal and postulated accident shipping conditions. The results of the calculations for a specific case of a cask loaded with breached fuel show that the fuel can be transported in standard shipping casks and maintained within the allowable release rates under normal and accident conditions. A sensitivity analysis has been conducted to evaluate the effects of modifying assumptions and to assess options for fuel at conditions that are not bounded by the present analysis. These options would include one or more of the following: reduce the fuel loading; increase fuel cooling time; reduce the degree of conservatism in the bounding assumptions; or measure the actual leak rate of the cask system. That is, containment analysis for alternative inputs at fuel-specific conditions and
Poisson Processes in Free Probability
An, Guimei; Gao, Mingchu
2015-01-01
We prove a multidimensional Poisson limit theorem in free probability, and define joint free Poisson distributions in a non-commutative probability space. We define (compound) free Poisson process explicitly, similar to the definitions of (compound) Poisson processes in classical probability. We proved that the sum of finitely many freely independent compound free Poisson processes is a compound free Poisson processes. We give a step by step procedure for constructing a (compound) free Poisso...
PROBABILITY SURVEYS , CONDITIONAL PROBABILITIES AND ECOLOGICAL RISK ASSESSMENT
We show that probability-based environmental resource monitoring programs, such as the U.S. Environmental Protection Agency's (U.S. EPA) Environmental Monitoring and Assessment Program, and conditional probability analysis can serve as a basis for estimating ecological risk over ...
Blanke, Sandra J; McGrady, Elizabeth
2016-07-01
Health care stakeholders are concerned about the growing risk of protecting sensitive patient health information from breaches. The Federal Emergency Management Agency (FEMA) has identified cyber attacks as an emerging concern, and regulations such as the Health Insurance Portability and Accountability Act (HIPAA) and the Health Information Technology for Economic and Clinical Health Act (HITECH) have increased security requirements and are enforcing compliance through stiff financial penalties. The purpose of this study is to describe health care breaches of protected information, analyze the hazards and vulnerabilities of reported breach cases, and prescribe best practices of managing risk through security controls and countermeasures. Prescriptive findings were used to construct a checklist tool to assess and monitor common risks. This research uses a case methodology to describe specific examples of the 3 major types of cyber breach hazards: portable device, insider, and physical breaches. We utilize a risk management framework to prescribe preventative actions that organizations can take to assess, analyze, and mitigate these risks. The health care sector has the largest number of reported breaches, with 3 major types: portable device, insider, and physical breaches. Analysis of actual cases indicates security gaps requiring prescriptive fixes based on "best practices." Our research culminates in a 25-item checklist that organizations can use to assess existing practices and identify security gaps requiring improvement. © 2016 American Society for Healthcare Risk Management of the American Hospital Association.
Determining probabilities of geologic events and processes
International Nuclear Information System (INIS)
Hunter, R.L.; Mann, C.J.; Cranwell, R.M.
1985-01-01
The Environmental Protection Agency has recently published a probabilistic standard for releases of high-level radioactive waste from a mined geologic repository. The standard sets limits for contaminant releases with more than one chance in 100 of occurring within 10,000 years, and less strict limits for releases of lower probability. The standard offers no methods for determining probabilities of geologic events and processes, and no consensus exists in the waste-management community on how to do this. Sandia National Laboratories is developing a general method for determining probabilities of a given set of geologic events and processes. In addition, we will develop a repeatable method for dealing with events and processes whose probability cannot be determined. 22 refs., 4 figs
Reimann, Mareike
2016-09-30
This study investigated whether the association between perceived psychological contract breach (PCB) and employee mental health is moderated by the cognitive-motivational pattern of overcommitment (OC). Linking the psychological contract approach to the effort-reward imbalance model, this study examines PCB as an imbalance in employment relationships that acts as a psychosocial stressor in the work environment and is associated with stress reactions that in turn negatively affect mental health. The analyses were based on a sample of 3,667 employees who participated in a longitudinal linked employer-employee survey representative of large organizations (with at least 500 employees who are subject so social security contributions) in Germany. Fixed-effects regression models, including PCB and OC, were estimated for employee mental health, and interaction effects between PCB and OC were assessed. The multivariate fixed-effects regression analyses showed a significant negative association between PCB and employee mental health. The results also confirmed that OC does indeed significantly increase the negative effect of PCB on mental health and that OC itself has a significant and negative effect on mental health. The results suggest that employees characterized by the cognitive-motivational pattern of OC are at an increased risk of developing poor mental health if they experience PCB compared with employees who are not overly committed to their work. The results of this study support the assumption that psychosocial work stressors play an important role in employee mental health.
Directory of Open Access Journals (Sweden)
Benjamin L. Franc
2003-10-01
Full Text Available The lipid bilayer of a cell presents a significant barrier for the delivery of many molecular imaging reagents into cells at target sites in the body. Protein translocation domains (PTDs are peptides that breach this barrier. Conjugation of PTDs to imaging agents can be utilized to facilitate the delivery of these agents through the cell wall, and in some cases, into the cell nucleus, and have potential for in vitro and in vivo applications. PTD imaging conjugates have included small molecules, peptides, proteins, DNA, metal chelates, and magnetic nanoparticles. The full potential of the use of PTDs in novel in vivo molecular probes is currently under investigation. Cells have been labeled in culture using magnetic nanoparticles derivatized with a PTD and monitored in vivo to assess trafficking patterns relative to cells expressing a target antigen. In vivo imaging of PTD-mediated gene transfer to cells of the skin has been demonstrated in living animals. Here we review several natural and synthetic PTDs that have evolved in the quest for easier translocation across biological barriers and the application of these peptide domains to in vivo delivery of imaging agents.
Directory of Open Access Journals (Sweden)
Joan Fontanet
2009-01-01
Full Text Available Development of HTRs requires the performance of a thorough safety study, which includes accident analyses. Confinement building performance is a key element of the system since the behaviour of aerosol and attached fission products within the building is of an utmost relevance in terms of the potential source term to the environment. This paper explores the available simulation capabilities (ASTEC and CONTAIN codes and illustrates the performance of a postulated HTR vented confinement under prototypical accident conditions by a scoping study based on two accident sequences characterized by Helium Pressure Boundary breaches, a small and a large break. The results obtained indicate that both codes predict very similar thermal-hydraulic responses of the confinement both in magnitude and timing. As for the aerosol behaviour, both codes predict that most of the inventory coming into the confinement is eventually depleted on the walls and only about 1% of the aerosol dust is released to the environment. The crosscomparison of codes states that largest differences are in the intercompartmental flows and the in-compartment gas composition.
Nifadkar, Sushil S; Bauer, Talya N
2016-01-01
Previous studies of newcomer socialization have underlined the importance of newcomers' information seeking for their adjustment to the organization, and the conflict literature has consistently reported negative effects of relationship conflict with coworkers. However, to date, no study has examined the consequences of relationship conflict on newcomers' information seeking. In this study, we examined newcomers' reactions when they have relationship conflict with their coworkers, and hence cannot obtain necessary information from them. Drawing upon belongingness theory, we propose a model that moves from breach of belongingness to its proximal and distal consequences, to newcomer information seeking, and then to task-related outcomes. In particular, we propose that second paths exist-first coworker-centric and the other supervisor-centric-that may have simultaneous yet contrasting influence on newcomer adjustment. To test our model, we employ a 3-wave data collection research design with egocentric and Likert-type multisource surveys among a sample of new software engineers and their supervisors working in India. This study contributes to the field by linking the literatures on relationship conflict and newcomer information seeking and suggesting that despite conflict with coworkers, newcomers may succeed in organizations by building relationships with and obtaining information from supervisors. (c) 2016 APA, all rights reserved).
Duty to warn of genetic harm in breach of patient confidentiality.
Keeling, Sharon L
2004-11-01
Harm caused by the failure of health professionals to warn an at-risk genetic relative of her or his risk is genetic harm. Genetic harm should be approached using the usual principles of negligence. When these principles are applied, it is shown that (a) genetic harm is foreseeable; (b) the salient features of vulnerability, the health professional's knowledge of the risk to the genetic relative and the determinancy of the affected class and individual result in a duty of care being owed to the genetic relative; (c) the standard of care required to fulfil the duty to warn should be the expectations of a reasonable person in the position of the relative; and (d) causation is satisfied as the harm is caused by the failure of intervention of the health professional. Legislation enacted subsequent to the Report of the Commonwealth of Australia, Panel of Eminent Persons (Chair D Ipp), Review of the Law of Negligence Report (2002) and relevant to a duty to warn of genetic harm is considered. The modes of regulation and penalties for breach of any future duty to warn of genetic harm are considered.
Probability inequalities for decomposition integrals
Czech Academy of Sciences Publication Activity Database
Agahi, H.; Mesiar, Radko
2017-01-01
Roč. 315, č. 1 (2017), s. 240-248 ISSN 0377-0427 Institutional support: RVO:67985556 Keywords : Decomposition integral * Superdecomposition integral * Probability inequalities Subject RIV: BA - General Mathematics OBOR OECD: Statistics and probability Impact factor: 1.357, year: 2016 http://library.utia.cas.cz/separaty/2017/E/mesiar-0470959.pdf
Expected utility with lower probabilities
DEFF Research Database (Denmark)
Hendon, Ebbe; Jacobsen, Hans Jørgen; Sloth, Birgitte
1994-01-01
An uncertain and not just risky situation may be modeled using so-called belief functions assigning lower probabilities to subsets of outcomes. In this article we extend the von Neumann-Morgenstern expected utility theory from probability measures to belief functions. We use this theory...
Invariant probabilities of transition functions
Zaharopol, Radu
2014-01-01
The structure of the set of all the invariant probabilities and the structure of various types of individual invariant probabilities of a transition function are two topics of significant interest in the theory of transition functions, and are studied in this book. The results obtained are useful in ergodic theory and the theory of dynamical systems, which, in turn, can be applied in various other areas (like number theory). They are illustrated using transition functions defined by flows, semiflows, and one-parameter convolution semigroups of probability measures. In this book, all results on transition probabilities that have been published by the author between 2004 and 2008 are extended to transition functions. The proofs of the results obtained are new. For transition functions that satisfy very general conditions the book describes an ergodic decomposition that provides relevant information on the structure of the corresponding set of invariant probabilities. Ergodic decomposition means a splitting of t...
Introduction to probability with Mathematica
Hastings, Kevin J
2009-01-01
Discrete ProbabilityThe Cast of Characters Properties of Probability Simulation Random SamplingConditional ProbabilityIndependenceDiscrete DistributionsDiscrete Random Variables, Distributions, and ExpectationsBernoulli and Binomial Random VariablesGeometric and Negative Binomial Random Variables Poisson DistributionJoint, Marginal, and Conditional Distributions More on ExpectationContinuous ProbabilityFrom the Finite to the (Very) Infinite Continuous Random Variables and DistributionsContinuous ExpectationContinuous DistributionsThe Normal Distribution Bivariate Normal DistributionNew Random Variables from OldOrder Statistics Gamma DistributionsChi-Square, Student's t, and F-DistributionsTransformations of Normal Random VariablesAsymptotic TheoryStrong and Weak Laws of Large Numbers Central Limit TheoremStochastic Processes and ApplicationsMarkov ChainsPoisson Processes QueuesBrownian MotionFinancial MathematicsAppendixIntroduction to Mathematica Glossary of Mathematica Commands for Probability Short Answers...
Linear positivity and virtual probability
International Nuclear Information System (INIS)
Hartle, James B.
2004-01-01
We investigate the quantum theory of closed systems based on the linear positivity decoherence condition of Goldstein and Page. The objective of any quantum theory of a closed system, most generally the universe, is the prediction of probabilities for the individual members of sets of alternative coarse-grained histories of the system. Quantum interference between members of a set of alternative histories is an obstacle to assigning probabilities that are consistent with the rules of probability theory. A quantum theory of closed systems therefore requires two elements: (1) a condition specifying which sets of histories may be assigned probabilities and (2) a rule for those probabilities. The linear positivity condition of Goldstein and Page is the weakest of the general conditions proposed so far. Its general properties relating to exact probability sum rules, time neutrality, and conservation laws are explored. Its inconsistency with the usual notion of independent subsystems in quantum mechanics is reviewed. Its relation to the stronger condition of medium decoherence necessary for classicality is discussed. The linear positivity of histories in a number of simple model systems is investigated with the aim of exhibiting linearly positive sets of histories that are not decoherent. The utility of extending the notion of probability to include values outside the range of 0-1 is described. Alternatives with such virtual probabilities cannot be measured or recorded, but can be used in the intermediate steps of calculations of real probabilities. Extended probabilities give a simple and general way of formulating quantum theory. The various decoherence conditions are compared in terms of their utility for characterizing classicality and the role they might play in further generalizations of quantum mechanics
Chen, Yong; Dong, Feng; Chen, Hong; Xu, Li
2016-08-01
The increase in globalization in the markets has driven firms to adopt online technologies and to cross-list their stocks. Recent studies have consistently found that the announcements of information security breaches (ISBs) are negatively associated with the market values of the announcing firms during the days surrounding the breach announcements. Given the improvement in firms’ information environments and the better protection for investors generated by cross-listing, does cross-listing help firms to reduce the negative impacts caused by their announcements of ISBs? This paper conducts an event study of 120 publicly traded firms (among which 25 cross-list and 95 do not), in order to explore the answer. The results indicate that the impact of ISB announcements on a firm's stock prices shows no difference between cross-listing firms and non-cross-listing firms. Cross-listing does not mitigate the impact of ISBs announcement on a firm's market value.
Vidmar, Ines; Ambrožič, Bojan; Debeljak, Barbara; Dolžan, Erazem; Gregorin, Špela; Grom, Nina; Herman, Polona; Keršmanc, Teja; Mencin, Eva; Mernik, Natalija; Švara, Astrid; Trobec, Ana; Turnšek, Anita; Vodeb, Petra; Torkar, Anja; Brenčič, Mihael
2013-04-01
On November 4-6 2012 heavy precipitation resulted in floods in the middle and lower course of Drava River in NE Slovenia causing damage to many properties in the flooded area. The meteorological situation that led to consequent floods was characterized by high precipitation, fast snowmelt, SW wind and relatively high air temperature. The weather event was part of a cyclone which was spreading over the area of North, West and Central Europe in the direction of Central Europe and carried with it the passing of a cold front through Slovenia on November 4 and 5. The flood wave travelled on the Drava River from Austria to Slovenia past the 11 hydroelectric power plants after eventually moving over the Slovenian-Croatian border. The river discharge increased in the early morning of November 5 reaching 3165 m3/s. This work focuses on a single event in the Ptujsko polje where among other damage caused by the flooding, the river broke through the drainage canal of the Formin hydroelectric power plant and changed its course. The Ptujsko polje contains two fluvial terraces. In the area of Formin HPP, the lower terrace is 1.5 km wide and the surface as well as the groundwater gradient shift from west to east with the groundwater flowing parallel to the river. These characteristics contributed to the flooding and consequential breach in the embankment of the drainage canal. Several aspects of the recent floods are discussed including a critical reflection of data accessibility, possible causes and mechanisms behind it as well as the possibility of its forecasting. Synthesis of accessible data from open domain sources is performed with emphasis on geological conditions. Discharge and precipitation data from the data base of Slovenian Environment Agency are collected, reviewed and analyzed. The flood event itself is analyzed and described in detail. It is determined that the flood wave was different from the ones regulated by natural processes which points to an anthropogenic
Johnson, Nicole Annette
2009-01-01
Especially during times of intense change, managers may negatively impact the quality of employee-manager relationships by breaching or violating psychological contract terms and exhibiting unfair treatment (i.e., interactional injustice) in the workplace. A psychological contract is conceptualized as an exmployee's perception or individualistic belief about the reciprocal and promissory nature of the employment relationship (Argyris, 1960; Levinson, Price, Munden, Mandl, & Solley, 1966; Rou...
Quasi 2D hydrodynamic modelling of the flooded hinterland due to dyke breaching on the Elbe River
Directory of Open Access Journals (Sweden)
S. Huang
2007-01-01
Full Text Available In flood modeling, many 1D and 2D combination and 2D models are used to simulate diversion of water from rivers through dyke breaches into the hinterland for extreme flood events. However, these models are too demanding in data requirements and computational resources which is an important consideration when uncertainty analysis using Monte Carlo techniques is used to complement the modeling exercise. The goal of this paper is to show the development of a quasi-2D modeling approach, which still calculates the dynamic wave in 1D but the discretisation of the computational units are in 2D, allowing a better spatial representation of the flow in the hinterland due to dyke breaching without a large additional expenditure on data pre-processing and computational time. A 2D representation of the flow and velocity fields is required to model sediment and micro-pollutant transport. The model DYNHYD (1D hydrodynamics from the WASP5 modeling package was used as a basis for the simulations. The model was extended to incorporate the quasi-2D approach and a Monte-Carlo Analysis was used to conduct a flood sensitivity analysis to determine the sensitivity of parameters and boundary conditions to the resulting water flow. An extreme flood event on the Elbe River, Germany, with a possible dyke breach area was used as a test case. The results show a good similarity with those obtained from another 1D/2D modeling study.
Probability Machines: Consistent Probability Estimation Using Nonparametric Learning Machines
Malley, J. D.; Kruppa, J.; Dasgupta, A.; Malley, K. G.; Ziegler, A.
2011-01-01
Summary Background Most machine learning approaches only provide a classification for binary responses. However, probabilities are required for risk estimation using individual patient characteristics. It has been shown recently that every statistical learning machine known to be consistent for a nonparametric regression problem is a probability machine that is provably consistent for this estimation problem. Objectives The aim of this paper is to show how random forests and nearest neighbors can be used for consistent estimation of individual probabilities. Methods Two random forest algorithms and two nearest neighbor algorithms are described in detail for estimation of individual probabilities. We discuss the consistency of random forests, nearest neighbors and other learning machines in detail. We conduct a simulation study to illustrate the validity of the methods. We exemplify the algorithms by analyzing two well-known data sets on the diagnosis of appendicitis and the diagnosis of diabetes in Pima Indians. Results Simulations demonstrate the validity of the method. With the real data application, we show the accuracy and practicality of this approach. We provide sample code from R packages in which the probability estimation is already available. This means that all calculations can be performed using existing software. Conclusions Random forest algorithms as well as nearest neighbor approaches are valid machine learning methods for estimating individual probabilities for binary responses. Freely available implementations are available in R and may be used for applications. PMID:21915433
Probable Inference and Quantum Mechanics
International Nuclear Information System (INIS)
Grandy, W. T. Jr.
2009-01-01
In its current very successful interpretation the quantum theory is fundamentally statistical in nature. Although commonly viewed as a probability amplitude whose (complex) square is a probability, the wavefunction or state vector continues to defy consensus as to its exact meaning, primarily because it is not a physical observable. Rather than approach this problem directly, it is suggested that it is first necessary to clarify the precise role of probability theory in quantum mechanics, either as applied to, or as an intrinsic part of the quantum theory. When all is said and done the unsurprising conclusion is that quantum mechanics does not constitute a logic and probability unto itself, but adheres to the long-established rules of classical probability theory while providing a means within itself for calculating the relevant probabilities. In addition, the wavefunction is seen to be a description of the quantum state assigned by an observer based on definite information, such that the same state must be assigned by any other observer based on the same information, in much the same way that probabilities are assigned.
Failure probability under parameter uncertainty.
Gerrard, R; Tsanakas, A
2011-05-01
In many problems of risk analysis, failure is equivalent to the event of a random risk factor exceeding a given threshold. Failure probabilities can be controlled if a decisionmaker is able to set the threshold at an appropriate level. This abstract situation applies, for example, to environmental risks with infrastructure controls; to supply chain risks with inventory controls; and to insurance solvency risks with capital controls. However, uncertainty around the distribution of the risk factor implies that parameter error will be present and the measures taken to control failure probabilities may not be effective. We show that parameter uncertainty increases the probability (understood as expected frequency) of failures. For a large class of loss distributions, arising from increasing transformations of location-scale families (including the log-normal, Weibull, and Pareto distributions), the article shows that failure probabilities can be exactly calculated, as they are independent of the true (but unknown) parameters. Hence it is possible to obtain an explicit measure of the effect of parameter uncertainty on failure probability. Failure probability can be controlled in two different ways: (1) by reducing the nominal required failure probability, depending on the size of the available data set, and (2) by modifying of the distribution itself that is used to calculate the risk control. Approach (1) corresponds to a frequentist/regulatory view of probability, while approach (2) is consistent with a Bayesian/personalistic view. We furthermore show that the two approaches are consistent in achieving the required failure probability. Finally, we briefly discuss the effects of data pooling and its systemic risk implications. © 2010 Society for Risk Analysis.
A philosophical essay on probabilities
Laplace, Marquis de
1996-01-01
A classic of science, this famous essay by ""the Newton of France"" introduces lay readers to the concepts and uses of probability theory. It is of especial interest today as an application of mathematical techniques to problems in social and biological sciences.Generally recognized as the founder of the modern phase of probability theory, Laplace here applies the principles and general results of his theory ""to the most important questions of life, which are, in effect, for the most part, problems in probability."" Thus, without the use of higher mathematics, he demonstrates the application
National Research Council Canada - National Science Library
Harvey, Kent
2006-01-01
...) Determine the difficulties and nuances of drilling behind wall test panels 3) Test different blast hole sizes, blast hole locations, and blasting sequences in an effort to identify the advantages and disadvantages of different breaching approaches...
Logic, probability, and human reasoning.
Johnson-Laird, P N; Khemlani, Sangeet S; Goodwin, Geoffrey P
2015-04-01
This review addresses the long-standing puzzle of how logic and probability fit together in human reasoning. Many cognitive scientists argue that conventional logic cannot underlie deductions, because it never requires valid conclusions to be withdrawn - not even if they are false; it treats conditional assertions implausibly; and it yields many vapid, although valid, conclusions. A new paradigm of probability logic allows conclusions to be withdrawn and treats conditionals more plausibly, although it does not address the problem of vapidity. The theory of mental models solves all of these problems. It explains how people reason about probabilities and postulates that the machinery for reasoning is itself probabilistic. Recent investigations accordingly suggest a way to integrate probability and deduction. Copyright © 2015 Elsevier Ltd. All rights reserved.
Free probability and random matrices
Mingo, James A
2017-01-01
This volume opens the world of free probability to a wide variety of readers. From its roots in the theory of operator algebras, free probability has intertwined with non-crossing partitions, random matrices, applications in wireless communications, representation theory of large groups, quantum groups, the invariant subspace problem, large deviations, subfactors, and beyond. This book puts a special emphasis on the relation of free probability to random matrices, but also touches upon the operator algebraic, combinatorial, and analytic aspects of the theory. The book serves as a combination textbook/research monograph, with self-contained chapters, exercises scattered throughout the text, and coverage of important ongoing progress of the theory. It will appeal to graduate students and all mathematicians interested in random matrices and free probability from the point of view of operator algebras, combinatorics, analytic functions, or applications in engineering and statistical physics.
Introduction to probability and measure
Parthasarathy, K R
2005-01-01
According to a remark attributed to Mark Kac 'Probability Theory is a measure theory with a soul'. This book with its choice of proofs, remarks, examples and exercises has been prepared taking both these aesthetic and practical aspects into account.
Joint probabilities and quantum cognition
International Nuclear Information System (INIS)
Acacio de Barros, J.
2012-01-01
In this paper we discuss the existence of joint probability distributions for quantumlike response computations in the brain. We do so by focusing on a contextual neural-oscillator model shown to reproduce the main features of behavioral stimulus-response theory. We then exhibit a simple example of contextual random variables not having a joint probability distribution, and describe how such variables can be obtained from neural oscillators, but not from a quantum observable algebra.
Joint probabilities and quantum cognition
Energy Technology Data Exchange (ETDEWEB)
Acacio de Barros, J. [Liberal Studies, 1600 Holloway Ave., San Francisco State University, San Francisco, CA 94132 (United States)
2012-12-18
In this paper we discuss the existence of joint probability distributions for quantumlike response computations in the brain. We do so by focusing on a contextual neural-oscillator model shown to reproduce the main features of behavioral stimulus-response theory. We then exhibit a simple example of contextual random variables not having a joint probability distribution, and describe how such variables can be obtained from neural oscillators, but not from a quantum observable algebra.
Default probabilities and default correlations
Erlenmaier, Ulrich; Gersbach, Hans
2001-01-01
Starting from the Merton framework for firm defaults, we provide the analytics and robustness of the relationship between default correlations. We show that loans with higher default probabilities will not only have higher variances but also higher correlations between loans. As a consequence, portfolio standard deviation can increase substantially when loan default probabilities rise. This result has two important implications. First, relative prices of loans with different default probabili...
The Probabilities of Unique Events
2012-08-30
Washington, DC USA Max Lotstein and Phil Johnson-Laird Department of Psychology Princeton University Princeton, NJ USA August 30th 2012...social justice and also participated in antinuclear demonstrations. The participants ranked the probability that Linda is a feminist bank teller as...retorted that such a flagrant violation of the probability calculus was a result of a psychological experiment that obscured the rationality of the
Probability Matching, Fast and Slow
Koehler, Derek J.; James, Greta
2014-01-01
A prominent point of contention among researchers regarding the interpretation of probability-matching behavior is whether it represents a cognitively sophisticated, adaptive response to the inherent uncertainty of the tasks or settings in which it is observed, or whether instead it represents a fundamental shortcoming in the heuristics that support and guide human decision making. Put crudely, researchers disagree on whether probability matching is "smart" or "dumb." Here, we consider eviden...
Probability & Statistics: Modular Learning Exercises. Teacher Edition
Actuarial Foundation, 2012
2012-01-01
The purpose of these modules is to provide an introduction to the world of probability and statistics to accelerated mathematics students at the high school level. The modules also introduce students to real world math concepts and problems that property and casualty actuaries come across in their work. They are designed to be used by teachers and…
Probability & Statistics: Modular Learning Exercises. Student Edition
Actuarial Foundation, 2012
2012-01-01
The purpose of these modules is to provide an introduction to the world of probability and statistics to accelerated mathematics students at the high school level. The materials are centered on the fictional town of Happy Shores, a coastal community which is at risk for hurricanes. Actuaries at an insurance company figure out the risks and…
Probably not future prediction using probability and statistical inference
Dworsky, Lawrence N
2008-01-01
An engaging, entertaining, and informative introduction to probability and prediction in our everyday lives Although Probably Not deals with probability and statistics, it is not heavily mathematical and is not filled with complex derivations, proofs, and theoretical problem sets. This book unveils the world of statistics through questions such as what is known based upon the information at hand and what can be expected to happen. While learning essential concepts including "the confidence factor" and "random walks," readers will be entertained and intrigued as they move from chapter to chapter. Moreover, the author provides a foundation of basic principles to guide decision making in almost all facets of life including playing games, developing winning business strategies, and managing personal finances. Much of the book is organized around easy-to-follow examples that address common, everyday issues such as: How travel time is affected by congestion, driving speed, and traffic lights Why different gambling ...
Normal probability plots with confidence.
Chantarangsi, Wanpen; Liu, Wei; Bretz, Frank; Kiatsupaibul, Seksan; Hayter, Anthony J; Wan, Fang
2015-01-01
Normal probability plots are widely used as a statistical tool for assessing whether an observed simple random sample is drawn from a normally distributed population. The users, however, have to judge subjectively, if no objective rule is provided, whether the plotted points fall close to a straight line. In this paper, we focus on how a normal probability plot can be augmented by intervals for all the points so that, if the population distribution is normal, then all the points should fall into the corresponding intervals simultaneously with probability 1-α. These simultaneous 1-α probability intervals provide therefore an objective mean to judge whether the plotted points fall close to the straight line: the plotted points fall close to the straight line if and only if all the points fall into the corresponding intervals. The powers of several normal probability plot based (graphical) tests and the most popular nongraphical Anderson-Darling and Shapiro-Wilk tests are compared by simulation. Based on this comparison, recommendations are given in Section 3 on which graphical tests should be used in what circumstances. An example is provided to illustrate the methods. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Probability theory a foundational course
Pakshirajan, R P
2013-01-01
This book shares the dictum of J. L. Doob in treating Probability Theory as a branch of Measure Theory and establishes this relation early. Probability measures in product spaces are introduced right at the start by way of laying the ground work to later claim the existence of stochastic processes with prescribed finite dimensional distributions. Other topics analysed in the book include supports of probability measures, zero-one laws in product measure spaces, Erdos-Kac invariance principle, functional central limit theorem and functional law of the iterated logarithm for independent variables, Skorohod embedding, and the use of analytic functions of a complex variable in the study of geometric ergodicity in Markov chains. This book is offered as a text book for students pursuing graduate programs in Mathematics and or Statistics. The book aims to help the teacher present the theory with ease, and to help the student sustain his interest and joy in learning the subject.
Approximation methods in probability theory
Čekanavičius, Vydas
2016-01-01
This book presents a wide range of well-known and less common methods used for estimating the accuracy of probabilistic approximations, including the Esseen type inversion formulas, the Stein method as well as the methods of convolutions and triangle function. Emphasising the correct usage of the methods presented, each step required for the proofs is examined in detail. As a result, this textbook provides valuable tools for proving approximation theorems. While Approximation Methods in Probability Theory will appeal to everyone interested in limit theorems of probability theory, the book is particularly aimed at graduate students who have completed a standard intermediate course in probability theory. Furthermore, experienced researchers wanting to enlarge their toolkit will also find this book useful.
Model uncertainty: Probabilities for models?
International Nuclear Information System (INIS)
Winkler, R.L.
1994-01-01
Like any other type of uncertainty, model uncertainty should be treated in terms of probabilities. The question is how to do this. The most commonly-used approach has a drawback related to the interpretation of the probabilities assigned to the models. If we step back and look at the big picture, asking what the appropriate focus of the model uncertainty question should be in the context of risk and decision analysis, we see that a different probabilistic approach makes more sense, although it raise some implementation questions. Current work that is underway to address these questions looks very promising
Knowledge typology for imprecise probabilities.
Energy Technology Data Exchange (ETDEWEB)
Wilson, G. D. (Gregory D.); Zucker, L. J. (Lauren J.)
2002-01-01
When characterizing the reliability of a complex system there are often gaps in the data available for specific subsystems or other factors influencing total system reliability. At Los Alamos National Laboratory we employ ethnographic methods to elicit expert knowledge when traditional data is scarce. Typically, we elicit expert knowledge in probabilistic terms. This paper will explore how we might approach elicitation if methods other than probability (i.e., Dempster-Shafer, or fuzzy sets) prove more useful for quantifying certain types of expert knowledge. Specifically, we will consider if experts have different types of knowledge that may be better characterized in ways other than standard probability theory.
Probability, Statistics, and Stochastic Processes
Olofsson, Peter
2011-01-01
A mathematical and intuitive approach to probability, statistics, and stochastic processes This textbook provides a unique, balanced approach to probability, statistics, and stochastic processes. Readers gain a solid foundation in all three fields that serves as a stepping stone to more advanced investigations into each area. This text combines a rigorous, calculus-based development of theory with a more intuitive approach that appeals to readers' sense of reason and logic, an approach developed through the author's many years of classroom experience. The text begins with three chapters that d
Statistical probability tables CALENDF program
International Nuclear Information System (INIS)
Ribon, P.
1989-01-01
The purpose of the probability tables is: - to obtain dense data representation - to calculate integrals by quadratures. They are mainly used in the USA for calculations by Monte Carlo and in the USSR and Europe for self-shielding calculations by the sub-group method. The moment probability tables, in addition to providing a more substantial mathematical basis and calculation methods, are adapted for condensation and mixture calculations, which are the crucial operations for reactor physics specialists. However, their extension is limited by the statistical hypothesis they imply. Efforts are being made to remove this obstacle, at the cost, it must be said, of greater complexity
Probability, statistics, and queueing theory
Allen, Arnold O
1990-01-01
This is a textbook on applied probability and statistics with computer science applications for students at the upper undergraduate level. It may also be used as a self study book for the practicing computer science professional. The successful first edition of this book proved extremely useful to students who need to use probability, statistics and queueing theory to solve problems in other fields, such as engineering, physics, operations research, and management science. The book has also been successfully used for courses in queueing theory for operations research students. This second edit
Probability and Statistics: 5 Questions
DEFF Research Database (Denmark)
Probability and Statistics: 5 Questions is a collection of short interviews based on 5 questions presented to some of the most influential and prominent scholars in probability and statistics. We hear their views on the fields, aims, scopes, the future direction of research and how their work fits...... in these respects. Interviews with Nick Bingham, Luc Bovens, Terrence L. Fine, Haim Gaifman, Donald Gillies, James Hawthorne, Carl Hoefer, James M. Joyce, Joseph B. Kadane Isaac Levi, D.H. Mellor, Patrick Suppes, Jan von Plato, Carl Wagner, Sandy Zabell...
Initial Impacts of the Mount Polley Tailings Pond Breach on Adjacent Aquatic Ecosystems
Petticrew, Ellen; Gantner, Nikolaus; Albers, Sam; Owens, Philip
2015-04-01
On August 4th 2014, the Mount Polley Tailings pond breach near Likely, B.C., released approximately 24 million cubic metres of tailings material into Polley Lake, Hazeltine Creek and Quesnel Lake. The discharge scoured and eroded a swath of soil and sediment delivering an unknown amount of metals and sediment into this tributary ecosystem of the Fraser River. Subsequent efforts by the mine operator to remediate by pumping tailings water from Polley Lake into Hazeltine Creek, which flows into Quesnel Lake, resulted in additional and continuous release of unknown volumes of contaminated water and sediments into the watershed. Heavy metals (e.g., selenium, copper, or mercury) reported as stored in the tailings pond entered the downstream aquatic environment and have been monitored in the water column of Quesnel Lake since August. These contaminants are likely particle-bound and thus subject to transport over long distances without appreciable degradation, resulting in the potential for chronic exposures and associated toxicological effects in exposed biota. While significant dilution is expected during aquatic transport, and the resulting concentrations in the water will likely be low, concentrations in exposed biota may become of concern over time. Metals such as mercury and selenium undergo bioaccumulation and biomagnification, once incorporated into the food chain/web. Thus, even small concentrations of such contaminants in water can lead to greater concentrations (~100 fold) in top predators. Over time, our predictions are that food web transfer will lead to an increase in concentrations from water (1-2 years)->invertebrates (1-2 yrs) ->fishes (2-5 yrs). Pacific salmon travel great distances in this watershed and may be exposed to contaminated water during their migrations. Resident species will be exposed to the contaminated waters and sediments in the study lakes year round. Little or no background/baseline data for metals in biota from Quesnel Lake exists
Coleman, Neil M; Kaktins, Uldis; Wojno, Stephanie
2016-06-01
In 1891 a report was published by an ASCE committee to investigate the cause of the Johnstown flood of 1889. They concluded that changes made to the dam by the South Fork Fishing and Hunting Club did not cause the disaster because the embankment would have been overflowed and breached if the changes were not made. We dispute that conclusion based on hydraulic analyses of the dam as originally built, estimates of the time of concentration and time to peak for the South Fork drainage basin, and reported conditions at the dam and in the watershed. We present a LiDAR-based volume of Lake Conemaugh at the time of dam failure (1.455 × 10(7) m(3)) and hydrographs of flood discharge and lake stage decline. Our analytical approach incorporates the complex shape of this dam breach. More than 65 min would have been needed to drain most of the lake, not the 45 min cited by most sources. Peak flood discharges were likely in the range 7200 to 8970 m(3) s(-1). The original dam design, with a crest ∼0.9 m higher and the added capacity of an auxiliary spillway and five discharge pipes, had a discharge capacity at overtopping more than twice that of the reconstructed dam. A properly rebuilt dam would not have overtopped and would likely have survived the runoff event, thereby saving thousands of lives. We believe the ASCE report represented state-of-the-art for 1891. However, the report contains discrepancies and lapses in key observations, and relied on excessive reservoir inflow estimates. The confidence they expressed that dam failure was inevitable was inconsistent with information available to the committee. Hydrodynamic erosion was a likely culprit in the 1862 dam failure that seriously damaged the embankment. The Club's substandard repair of this earlier breach sowed the seeds of its eventual destruction.
Dynamic SEP event probability forecasts
Kahler, S. W.; Ling, A.
2015-10-01
The forecasting of solar energetic particle (SEP) event probabilities at Earth has been based primarily on the estimates of magnetic free energy in active regions and on the observations of peak fluxes and fluences of large (≥ M2) solar X-ray flares. These forecasts are typically issued for the next 24 h or with no definite expiration time, which can be deficient for time-critical operations when no SEP event appears following a large X-ray flare. It is therefore important to decrease the event probability forecast with time as a SEP event fails to appear. We use the NOAA listing of major (≥10 pfu) SEP events from 1976 to 2014 to plot the delay times from X-ray peaks to SEP threshold onsets as a function of solar source longitude. An algorithm is derived to decrease the SEP event probabilities with time when no event is observed to reach the 10 pfu threshold. In addition, we use known SEP event size distributions to modify probability forecasts when SEP intensity increases occur below the 10 pfu event threshold. An algorithm to provide a dynamic SEP event forecast, Pd, for both situations of SEP intensities following a large flare is derived.
Conditional Independence in Applied Probability.
Pfeiffer, Paul E.
This material assumes the user has the background provided by a good undergraduate course in applied probability. It is felt that introductory courses in calculus, linear algebra, and perhaps some differential equations should provide the requisite experience and proficiency with mathematical concepts, notation, and argument. The document is…
Stretching Probability Explorations with Geoboards
Wheeler, Ann; Champion, Joe
2016-01-01
Students are faced with many transitions in their middle school mathematics classes. To build knowledge, skills, and confidence in the key areas of algebra and geometry, students often need to practice using numbers and polygons in a variety of contexts. Teachers also want students to explore ideas from probability and statistics. Teachers know…
GPS: Geometry, Probability, and Statistics
Field, Mike
2012-01-01
It might be said that for most occupations there is now less of a need for mathematics than there was say fifty years ago. But, the author argues, geometry, probability, and statistics constitute essential knowledge for everyone. Maybe not the geometry of Euclid, but certainly geometrical ways of thinking that might enable us to describe the world…
Swedish earthquakes and acceleration probabilities
International Nuclear Information System (INIS)
Slunga, R.
1979-03-01
A method to assign probabilities to ground accelerations for Swedish sites is described. As hardly any nearfield instrumental data is available we are left with the problem of interpreting macroseismic data in terms of acceleration. By theoretical wave propagation computations the relation between seismic strength of the earthquake, focal depth, distance and ground accelerations are calculated. We found that most Swedish earthquake of the area, the 1904 earthquake 100 km south of Oslo, is an exception and probably had a focal depth exceeding 25 km. For the nuclear power plant sites an annual probability of 10 -5 has been proposed as interesting. This probability gives ground accelerations in the range 5-20 % for the sites. This acceleration is for a free bedrock site. For consistency all acceleration results in this study are given for bedrock sites. When applicating our model to the 1904 earthquake and assuming the focal zone to be in the lower crust we get the epicentral acceleration of this earthquake to be 5-15 % g. The results above are based on an analyses of macrosismic data as relevant instrumental data is lacking. However, the macroseismic acceleration model deduced in this study gives epicentral ground acceleration of small Swedish earthquakes in agreement with existent distant instrumental data. (author)
DECOFF Probabilities of Failed Operations
DEFF Research Database (Denmark)
Gintautas, Tomas
2015-01-01
A statistical procedure of estimation of Probabilities of Failed Operations is described and exemplified using ECMWF weather forecasts and SIMO output from Rotor Lift test case models. Also safety factor influence is investigated. DECOFF statistical method is benchmarked against standard Alpha-factor...
Risk estimation using probability machines
2014-01-01
Background Logistic regression has been the de facto, and often the only, model used in the description and analysis of relationships between a binary outcome and observed features. It is widely used to obtain the conditional probabilities of the outcome given predictors, as well as predictor effect size estimates using conditional odds ratios. Results We show how statistical learning machines for binary outcomes, provably consistent for the nonparametric regression problem, can be used to provide both consistent conditional probability estimation and conditional effect size estimates. Effect size estimates from learning machines leverage our understanding of counterfactual arguments central to the interpretation of such estimates. We show that, if the data generating model is logistic, we can recover accurate probability predictions and effect size estimates with nearly the same efficiency as a correct logistic model, both for main effects and interactions. We also propose a method using learning machines to scan for possible interaction effects quickly and efficiently. Simulations using random forest probability machines are presented. Conclusions The models we propose make no assumptions about the data structure, and capture the patterns in the data by just specifying the predictors involved and not any particular model structure. So they do not run the same risks of model mis-specification and the resultant estimation biases as a logistic model. This methodology, which we call a “risk machine”, will share properties from the statistical machine that it is derived from. PMID:24581306
Probability and statistics: A reminder
International Nuclear Information System (INIS)
Clement, B.
2013-01-01
The main purpose of these lectures is to provide the reader with the tools needed to data analysis in the framework of physics experiments. Basic concepts are introduced together with examples of application in experimental physics. The lecture is divided into two parts: probability and statistics. It is build on the introduction from 'data analysis in experimental sciences' given in [1]. (authors)
Nash equilibrium with lower probabilities
DEFF Research Database (Denmark)
Groes, Ebbe; Jacobsen, Hans Jørgen; Sloth, Birgitte
1998-01-01
We generalize the concept of Nash equilibrium in mixed strategies for strategic form games to allow for ambiguity in the players' expectations. In contrast to other contributions, we model ambiguity by means of so-called lower probability measures or belief functions, which makes it possible...
On probability-possibility transformations
Klir, George J.; Parviz, Behzad
1992-01-01
Several probability-possibility transformations are compared in terms of the closeness of preserving second-order properties. The comparison is based on experimental results obtained by computer simulation. Two second-order properties are involved in this study: noninteraction of two distributions and projections of a joint distribution.
Directory of Open Access Journals (Sweden)
Álvaro Vidal Olivares
2016-12-01
Full Text Available This article refers to the regime risks in the CISG with the aim of showing that the regime that it is incorporated, is based on functional criteria to the interests of the parties to solve the problems that originated in the loss and prior to the transfer of risk to the buyer will have recourse to remedies system and thus achieve the connection with the breach of contract. In its development we used the dogmatic method, from the systematic analysis of the rules of the CISG, doctrine and case law.
Probability shapes perceptual precision: A study in orientation estimation.
Jabar, Syaheed B; Anderson, Britt
2015-12-01
Probability is known to affect perceptual estimations, but an understanding of mechanisms is lacking. Moving beyond binary classification tasks, we had naive participants report the orientation of briefly viewed gratings where we systematically manipulated contingent probability. Participants rapidly developed faster and more precise estimations for high-probability tilts. The shapes of their error distributions, as indexed by a kurtosis measure, also showed a distortion from Gaussian. This kurtosis metric was robust, capturing probability effects that were graded, contextual, and varying as a function of stimulus orientation. Our data can be understood as a probability-induced reduction in the variability or "shape" of estimation errors, as would be expected if probability affects the perceptual representations. As probability manipulations are an implicit component of many endogenous cuing paradigms, changes at the perceptual level could account for changes in performance that might have traditionally been ascribed to "attention." (c) 2015 APA, all rights reserved).
Probability of collective excited state decay
International Nuclear Information System (INIS)
Manykin, Eh.A.; Ozhovan, M.I.; Poluehktov, P.P.
1987-01-01
Decay mechanisms of condensed excited state formed of highly excited (Rydberg) atoms are considered, i.e. stability of so-called Rydberg substance is analyzed. It is shown that Auger recombination and radiation transitions are the basic processes. The corresponding probabilities are calculated and compared. It is ascertained that the ''Rydberg substance'' possesses macroscopic lifetime (several seconds) and in a sense it is metastable
Probability and statistics in particle physics
International Nuclear Information System (INIS)
Frodesen, A.G.; Skjeggestad, O.
1979-01-01
Probability theory is entered into at an elementary level and given a simple and detailed exposition. The material on statistics has been organised with an eye to the experimental physicist's practical need, which is likely to be statistical methods for estimation or decision-making. The book is intended for graduate students and research workers in experimental high energy and elementary particle physics, and numerous examples from these fields are presented. (JIW)
Vander Elst, Tinne; De Cuyper, Nele; Baillien, Elfi; Niesen, Wendy; De Witte, Hans
2016-04-01
This study aims to further knowledge on the mechanisms through which job insecurity is related to negative outcomes. Based on appraisal theory, two explanations-perceived control and psychological contract breach-were theoretically integrated in a comprehensive model and simultaneously examined as mediators of the job insecurity-outcome relationship. Different categories of outcomes were considered, namely work-related (i.e. vigour and need for recovery) and general strain (i.e. mental and physical health complaints), as well as psychological (i.e. job satisfaction and organizational commitment) and behavioural coping reactions (i.e. self-rated performance and innovative work behaviour). The hypotheses were tested using data of a heterogeneous sample of 2413 Flemish employees by means of both single and multiple mediator structural equation modelling analyses (bootstrapping method). Particularly, psychological contract breach accounted for the relationship between job insecurity and strain. Both perceived control and psychological contract breach mediated the relationships between job insecurity and psychological coping reactions, although the indirect effects were larger for psychological contract breach. Finally, perceived control was more important than psychological contract breach in mediating the relationships between job insecurity and behavioural coping reactions. This study meets previous calls for a theoretical integration regarding mediators of the job insecurity-outcome relationship. Copyright © 2014 John Wiley & Sons, Ltd.
Large deviations and idempotent probability
Puhalskii, Anatolii
2001-01-01
In the view of many probabilists, author Anatolii Puhalskii''s research results stand among the most significant achievements in the modern theory of large deviations. In fact, his work marked a turning point in the depth of our understanding of the connections between the large deviation principle (LDP) and well-known methods for establishing weak convergence results.Large Deviations and Idempotent Probability expounds upon the recent methodology of building large deviation theory along the lines of weak convergence theory. The author develops an idempotent (or maxitive) probability theory, introduces idempotent analogues of martingales (maxingales), Wiener and Poisson processes, and Ito differential equations, and studies their properties. The large deviation principle for stochastic processes is formulated as a certain type of convergence of stochastic processes to idempotent processes. The author calls this large deviation convergence.The approach to establishing large deviation convergence uses novel com...
Probability biases as Bayesian inference
Directory of Open Access Journals (Sweden)
Andre; C. R. Martins
2006-11-01
Full Text Available In this article, I will show how several observed biases in human probabilistic reasoning can be partially explained as good heuristics for making inferences in an environment where probabilities have uncertainties associated to them. Previous results show that the weight functions and the observed violations of coalescing and stochastic dominance can be understood from a Bayesian point of view. We will review those results and see that Bayesian methods should also be used as part of the explanation behind other known biases. That means that, although the observed errors are still errors under the be understood as adaptations to the solution of real life problems. Heuristics that allow fast evaluations and mimic a Bayesian inference would be an evolutionary advantage, since they would give us an efficient way of making decisions. %XX In that sense, it should be no surprise that humans reason with % probability as it has been observed.
Probability matching and strategy availability.
Koehler, Derek J; James, Greta
2010-09-01
Findings from two experiments indicate that probability matching in sequential choice arises from an asymmetry in strategy availability: The matching strategy comes readily to mind, whereas a superior alternative strategy, maximizing, does not. First, compared with the minority who spontaneously engage in maximizing, the majority of participants endorse maximizing as superior to matching in a direct comparison when both strategies are described. Second, when the maximizing strategy is brought to their attention, more participants subsequently engage in maximizing. Third, matchers are more likely than maximizers to base decisions in other tasks on their initial intuitions, suggesting that they are more inclined to use a choice strategy that comes to mind quickly. These results indicate that a substantial subset of probability matchers are victims of "underthinking" rather than "overthinking": They fail to engage in sufficient deliberation to generate a superior alternative to the matching strategy that comes so readily to mind.
Probability as a Physical Motive
Directory of Open Access Journals (Sweden)
Peter Martin
2007-04-01
Full Text Available Recent theoretical progress in nonequilibrium thermodynamics, linking thephysical principle of Maximum Entropy Production (Ã¢Â€ÂœMEPÃ¢Â€Â to the information-theoreticalÃ¢Â€ÂœMaxEntÃ¢Â€Â principle of scientific inference, together with conjectures from theoreticalphysics that there may be no fundamental causal laws but only probabilities for physicalprocesses, and from evolutionary theory that biological systems expand Ã¢Â€Âœthe adjacentpossibleÃ¢Â€Â as rapidly as possible, all lend credence to the proposition that probability shouldbe recognized as a fundamental physical motive. It is further proposed that spatial order andtemporal order are two aspects of the same thing, and that this is the essence of the secondlaw of thermodynamics.
Logic, Probability, and Human Reasoning
2015-01-01
accordingly suggest a way to integrate probability and deduction. The nature of deductive reasoning To be rational is to be able to make deductions...3–6] and they underlie mathematics, science, and tech- nology [7–10]. Plato claimed that emotions upset reason- ing. However, individuals in the grip...fundamental to human rationality . So, if counterexamples to its principal predictions occur, the theory will at least explain its own refutation
Probability Measures on Groups IX
1989-01-01
The latest in this series of Oberwolfach conferences focussed on the interplay between structural probability theory and various other areas of pure and applied mathematics such as Tauberian theory, infinite-dimensional rotation groups, central limit theorems, harmonizable processes, and spherical data. Thus it was attended by mathematicians whose research interests range from number theory to quantum physics in conjunction with structural properties of probabilistic phenomena. This volume contains 5 survey articles submitted on special invitation and 25 original research papers.
Probability matching and strategy availability
J. Koehler, Derek; Koehler, Derek J.; James, Greta
2010-01-01
Findings from two experiments indicate that probability matching in sequential choice arises from an asymmetry in strategy availability: The matching strategy comes readily to mind, whereas a superior alternative strategy, maximizing, does not. First, compared with the minority who spontaneously engage in maximizing, the majority of participants endorse maximizing as superior to matching in a direct comparison when both strategies are described. Second, when the maximizing strategy is brought...
The probability and severity of decompression sickness
Hada, Ethan A.; Vann, Richard D.; Denoble, Petar J.
2017-01-01
Decompression sickness (DCS), which is caused by inert gas bubbles in tissues, is an injury of concern for scuba divers, compressed air workers, astronauts, and aviators. Case reports for 3322 air and N2-O2 dives, resulting in 190 DCS events, were retrospectively analyzed and the outcomes were scored as (1) serious neurological, (2) cardiopulmonary, (3) mild neurological, (4) pain, (5) lymphatic or skin, and (6) constitutional or nonspecific manifestations. Following standard U.S. Navy medical definitions, the data were grouped into mild—Type I (manifestations 4–6)–and serious–Type II (manifestations 1–3). Additionally, we considered an alternative grouping of mild–Type A (manifestations 3–6)–and serious–Type B (manifestations 1 and 2). The current U.S. Navy guidance allows for a 2% probability of mild DCS and a 0.1% probability of serious DCS. We developed a hierarchical trinomial (3-state) probabilistic DCS model that simultaneously predicts the probability of mild and serious DCS given a dive exposure. Both the Type I/II and Type A/B discriminations of mild and serious DCS resulted in a highly significant (p probability of ‘mild’ DCS resulted in a longer allowable bottom time for the same 2% limit. However, for the 0.1% serious DCS limit, we found a vastly decreased allowable bottom dive time for all dive depths. If the Type A/B scoring was assigned to outcome severity, the no decompression limits (NDL) for air dives were still controlled by the acceptable serious DCS risk limit rather than the acceptable mild DCS risk limit. However, in this case, longer NDL limits were allowed than with the Type I/II scoring. The trinomial model mild and serious probabilities agree reasonably well with the current air NDL only with the Type A/B scoring and when 0.2% risk of serious DCS is allowed. PMID:28296928
Dependent Human Error Probability Assessment
International Nuclear Information System (INIS)
Simic, Z.; Mikulicic, V.; Vukovic, I.
2006-01-01
This paper presents an assessment of the dependence between dynamic operator actions modeled in a Nuclear Power Plant (NPP) PRA and estimate the associated impact on Core damage frequency (CDF). This assessment was done improve HEP dependencies implementation inside existing PRA. All of the dynamic operator actions modeled in the NPP PRA are included in this assessment. Determining the level of HEP dependence and the associated influence on CDF are the major steps of this assessment. A decision on how to apply the results, i.e., should permanent HEP model changes be made, is based on the resulting relative CDF increase. Some CDF increase was selected as a threshold based on the NPP base CDF value and acceptance guidelines from the Regulatory Guide 1.174. HEP dependence resulting in a CDF increase of > 5E-07 would be considered potential candidates for specific incorporation into the baseline model. The approach used to judge the level of dependence between operator actions is based on dependency level categories and conditional probabilities developed in the Handbook of Human Reliability Analysis with Emphasis on Nuclear Power Plant Applications NUREG/CR-1278. To simplify the process, NUREG/CR-1278 identifies five levels of dependence: ZD (zero dependence), LD (low dependence), MD (moderate dependence), HD (high dependence), and CD (complete dependence). NUREG/CR-1278 also identifies several qualitative factors that could be involved in determining the level of dependence. Based on the NUREG/CR-1278 information, Time, Function, and Spatial attributes were judged to be the most important considerations when determining the level of dependence between operator actions within an accident sequence. These attributes were used to develop qualitative criteria (rules) that were used to judge the level of dependence (CD, HD, MD, LD, ZD) between the operator actions. After the level of dependence between the various HEPs is judged, quantitative values associated with the
Internal Medicine residents use heuristics to estimate disease probability
Phang, Sen Han; Ravani, Pietro; Schaefer, Jeffrey; Wright, Bruce; McLaughlin, Kevin
2015-01-01
Background: Training in Bayesian reasoning may have limited impact on accuracy of probability estimates. In this study, our goal was to explore whether residents previously exposed to Bayesian reasoning use heuristics rather than Bayesian reasoning to estimate disease probabilities. We predicted that if residents use heuristics then post-test probability estimates would be increased by non-discriminating clinical features or a high anchor for a target condition. Method: We randomized 55 In...
[Biometric bases: basic concepts of probability calculation].
Dinya, E
1998-04-26
The author gives or outline of the basic concepts of probability theory. The bases of the event algebra, definition of the probability, the classical probability model and the random variable are presented.
International Nuclear Information System (INIS)
Fujii, Sadao; Shibazaki, Hiroaki; Takase, Kazuyuki; Kunugi, Tomoaki.
1997-01-01
Exchange flow rates through breaches of a vacuum vessel in a fusion reactor under the LOVA (Loss of VAcuum event) conditions were measured quantitatively by using a preliminary LOVA apparatus and exchange flow patterns over the breach were visualized qualitatively by smoke. Velocity distributions in the exchange flows were predicted from the observed flow patterns by using the correlation method in the flow visualization procedures. Mean velocities calculated from the predicted velocity distributions at the outside of the breach were in good agreement with the LOVA experimental results when the exchange flow velocities were low. It was found that the present flow visualization and the image processing system might be an useful procedure to evaluate the exchange flow rates. (author)
Probability and containment of turbine missiles
International Nuclear Information System (INIS)
Yeh, G.C.K.
1976-01-01
With the trend toward ever larger power generating plants with large high-speed turbines, an important plant design consideration is the potential for and consequences of mechanical failure of turbine rotors. Such rotor failure could result in high-velocity disc fragments (turbine missiles) perforating the turbine casing and jeopardizing vital plant systems. The designer must first estimate the probability of any turbine missile damaging any safety-related plant component for his turbine and his plant arrangement. If the probability is not low enough to be acceptable to the regulatory agency, he must design a shield to contain the postulated turbine missiles. Alternatively, the shield could be designed to retard (to reduce the velocity of) the missiles such that they would not damage any vital plant system. In this paper, some of the presently available references that can be used to evaluate the probability, containment and retardation of turbine missiles are reviewed; various alternative methods are compared; and subjects for future research are recommended. (Auth.)
Probability for Weather and Climate
Smith, L. A.
2013-12-01
Over the last 60 years, the availability of large-scale electronic computers has stimulated rapid and significant advances both in meteorology and in our understanding of the Earth System as a whole. The speed of these advances was due, in large part, to the sudden ability to explore nonlinear systems of equations. The computer allows the meteorologist to carry a physical argument to its conclusion; the time scales of weather phenomena then allow the refinement of physical theory, numerical approximation or both in light of new observations. Prior to this extension, as Charney noted, the practicing meteorologist could ignore the results of theory with good conscience. Today, neither the practicing meteorologist nor the practicing climatologist can do so, but to what extent, and in what contexts, should they place the insights of theory above quantitative simulation? And in what circumstances can one confidently estimate the probability of events in the world from model-based simulations? Despite solid advances of theory and insight made possible by the computer, the fidelity of our models of climate differs in kind from the fidelity of models of weather. While all prediction is extrapolation in time, weather resembles interpolation in state space, while climate change is fundamentally an extrapolation. The trichotomy of simulation, observation and theory which has proven essential in meteorology will remain incomplete in climate science. Operationally, the roles of probability, indeed the kinds of probability one has access too, are different in operational weather forecasting and climate services. Significant barriers to forming probability forecasts (which can be used rationally as probabilities) are identified. Monte Carlo ensembles can explore sensitivity, diversity, and (sometimes) the likely impact of measurement uncertainty and structural model error. The aims of different ensemble strategies, and fundamental differences in ensemble design to support of
Probability, Statistics, and Stochastic Processes
Olofsson, Peter
2012-01-01
This book provides a unique and balanced approach to probability, statistics, and stochastic processes. Readers gain a solid foundation in all three fields that serves as a stepping stone to more advanced investigations into each area. The Second Edition features new coverage of analysis of variance (ANOVA), consistency and efficiency of estimators, asymptotic theory for maximum likelihood estimators, empirical distribution function and the Kolmogorov-Smirnov test, general linear models, multiple comparisons, Markov chain Monte Carlo (MCMC), Brownian motion, martingales, and
Probability, statistics, and computational science.
Beerenwinkel, Niko; Siebourg, Juliane
2012-01-01
In this chapter, we review basic concepts from probability theory and computational statistics that are fundamental to evolutionary genomics. We provide a very basic introduction to statistical modeling and discuss general principles, including maximum likelihood and Bayesian inference. Markov chains, hidden Markov models, and Bayesian network models are introduced in more detail as they occur frequently and in many variations in genomics applications. In particular, we discuss efficient inference algorithms and methods for learning these models from partially observed data. Several simple examples are given throughout the text, some of which point to models that are discussed in more detail in subsequent chapters.
Sensitivity analysis using probability bounding
International Nuclear Information System (INIS)
Ferson, Scott; Troy Tucker, W.
2006-01-01
Probability bounds analysis (PBA) provides analysts a convenient means to characterize the neighborhood of possible results that would be obtained from plausible alternative inputs in probabilistic calculations. We show the relationship between PBA and the methods of interval analysis and probabilistic uncertainty analysis from which it is jointly derived, and indicate how the method can be used to assess the quality of probabilistic models such as those developed in Monte Carlo simulations for risk analyses. We also illustrate how a sensitivity analysis can be conducted within a PBA by pinching inputs to precise distributions or real values
Harper, Nigel M; Getty, Kelly J K; Schmidt, Karen A; Nutsch, Abbey L; Linton, Richard H
2011-03-01
The U.S. Food and Drug Administration's Bacteriological Analytical Manual recommends two enumeration methods for Bacillus cereus: (i) standard plate count method with mannitol-egg yolk-polymyxin (MYP) agar and (ii) a most-probable-number (MPN) method with tryptic soy broth (TSB) supplemented with 0.1% polymyxin sulfate. This study compared the effectiveness of MYP and MPN methods for detecting and enumerating B. cereus in raw and high-temperature, short-time pasteurized skim (0.5%), 2%, and whole (3.5%) bovine milk stored at 4°C for 96 h. Each milk sample was inoculated with B. cereus EZ-Spores and sampled at 0, 48, and 96 h after inoculation. There were no differences (P > 0.05) in B. cereus populations among sampling times for all milk types, so data were pooled to obtain overall mean values for each treatment. The overall B. cereus population mean of pooled sampling times for the MPN method (2.59 log CFU/ml) was greater (P milk samples ranged from 2.36 to 3.46 and 2.66 to 3.58 log CFU/ml for inoculated milk treatments for the MYP plate count and MPN methods, respectively, which is below the level necessary for toxin production. The MPN method recovered more B. cereus, which makes it useful for validation research. However, the MYP plate count method for enumeration of B. cereus also had advantages, including its ease of use and faster time to results (2 versus 5 days for MPN).
Trial type probability modulates the cost of antisaccades
Chiau, Hui-Yan; Tseng, Philip; Su, Jia-Han; Tzeng, Ovid J. L.; Hung, Daisy L.; Muggleton, Neil G.
2011-01-01
The antisaccade task, where eye movements are made away from a target, has been used to investigate the flexibility of cognitive control of behavior. Antisaccades usually have longer saccade latencies than prosaccades, the so-called antisaccade cost. Recent studies have shown that this antisaccade cost can be modulated by event probability. This may mean that the antisaccade cost can be reduced, or even reversed, if the probability of surrounding events favors the execution of antisaccades. The probabilities of prosaccades and antisaccades were systematically manipulated by changing the proportion of a certain type of trial in an interleaved pro/antisaccades task. We aimed to disentangle the intertwined relationship between trial type probabilities and the antisaccade cost with the ultimate goal of elucidating how probabilities of trial types modulate human flexible behaviors, as well as the characteristics of such modulation effects. To this end, we examined whether implicit trial type probability can influence saccade latencies and also manipulated the difficulty of cue discriminability to see how effects of trial type probability would change when the demand on visual perceptual analysis was high or low. A mixed-effects model was applied to the analysis to dissect the factors contributing to the modulation effects of trial type probabilities. Our results suggest that the trial type probability is one robust determinant of antisaccade cost. These findings highlight the importance of implicit probability in the flexibility of cognitive control of behavior. PMID:21543748
The Influence of Phonotactic Probability on Word Recognition in Toddlers
MacRoy-Higgins, Michelle; Shafer, Valerie L.; Schwartz, Richard G.; Marton, Klara
2014-01-01
This study examined the influence of phonotactic probability on word recognition in English-speaking toddlers. Typically developing toddlers completed a preferential looking paradigm using familiar words, which consisted of either high or low phonotactic probability sound sequences. The participants' looking behavior was recorded in response to…
Failure probability analysis on mercury target vessel
International Nuclear Information System (INIS)
Ishikura, Syuichi; Futakawa, Masatoshi; Kogawa, Hiroyuki; Sato, Hiroshi; Haga, Katsuhiro; Ikeda, Yujiro
2005-03-01
Failure probability analysis was carried out to estimate the lifetime of the mercury target which will be installed into the JSNS (Japan spallation neutron source) in J-PARC (Japan Proton Accelerator Research Complex). The lifetime was estimated as taking loading condition and materials degradation into account. Considered loads imposed on the target vessel were the static stresses due to thermal expansion and static pre-pressure on He-gas and mercury and the dynamic stresses due to the thermally shocked pressure waves generated repeatedly at 25 Hz. Materials used in target vessel will be degraded by the fatigue, neutron and proton irradiation, mercury immersion and pitting damages, etc. The imposed stresses were evaluated through static and dynamic structural analyses. The material-degradations were deduced based on published experimental data. As a result, it was quantitatively confirmed that the failure probability for the lifetime expected in the design is very much lower, 10 -11 in the safety hull, meaning that it will be hardly failed during the design lifetime. On the other hand, the beam window of mercury vessel suffered with high-pressure waves exhibits the failure probability of 12%. It was concluded, therefore, that the leaked mercury from the failed area at the beam window is adequately kept in the space between the safety hull and the mercury vessel by using mercury-leakage sensors. (author)
Probability theory a comprehensive course
Klenke, Achim
2014-01-01
This second edition of the popular textbook contains a comprehensive course in modern probability theory. Overall, probabilistic concepts play an increasingly important role in mathematics, physics, biology, financial engineering and computer science. They help us in understanding magnetism, amorphous media, genetic diversity and the perils of random developments at financial markets, and they guide us in constructing more efficient algorithms. To address these concepts, the title covers a wide variety of topics, many of which are not usually found in introductory textbooks, such as: • limit theorems for sums of random variables • martingales • percolation • Markov chains and electrical networks • construction of stochastic processes • Poisson point process and infinite divisibility • large deviation principles and statistical physics • Brownian motion • stochastic integral and stochastic differential equations. The theory is developed rigorously and in a self-contained way, with the c...
Excluding joint probabilities from quantum theory
Allahverdyan, Armen E.; Danageozian, Arshag
2018-03-01
Quantum theory does not provide a unique definition for the joint probability of two noncommuting observables, which is the next important question after the Born's probability for a single observable. Instead, various definitions were suggested, e.g., via quasiprobabilities or via hidden-variable theories. After reviewing open issues of the joint probability, we relate it to quantum imprecise probabilities, which are noncontextual and are consistent with all constraints expected from a quantum probability. We study two noncommuting observables in a two-dimensional Hilbert space and show that there is no precise joint probability that applies for any quantum state and is consistent with imprecise probabilities. This contrasts with theorems by Bell and Kochen-Specker that exclude joint probabilities for more than two noncommuting observables, in Hilbert space with dimension larger than two. If measurement contexts are included into the definition, joint probabilities are not excluded anymore, but they are still constrained by imprecise probabilities.
Use of an influence diagram and fuzzy probability for evaluating accident management in a BWR
International Nuclear Information System (INIS)
Yu, Donghan; Okrent, D.; Kastenberg, W.E.
1993-01-01
This paper develops a new approach for evaluating severe accident management strategies. At first, this approach considers accident management as a decision problem (i.e., ''applying a strategy'' vs. ''do nothing'') and uses influence diagrams. This approach introduces the concept of a ''fuzzy probability'' in the evaluation of an influence diagram. When fuzzy logic is applied, fuzzy probabilities in an influence diagram can be easily propagated to obtain results. In addition, the results obtained provide not only information similar to the classical approach using point-estimate values, but also additional information regarding the impact from imprecise input data. The proposed methodology is applied to the evaluation of the drywell flooding strategy for a long-term station blackout sequence in the Peach Bottom nuclear power plant. The results show that the drywell flooding strategy seems to be beneficial for preventing reactor vessel breach. It is also effective for reducing the probability of the containment failure for both liner melt-through and late overpressurization. Even though there exists uncertainty in the results, ''flooding'' is preferred to ''do nothing'' when evaluated in terms of expected consequences, i.e., early and late fatalities
Influence of the Probability Level on the Framing Effect
Directory of Open Access Journals (Sweden)
Kaja Damnjanovic
2016-11-01
Full Text Available Research of the framing effect of risky choice mostly applies to the tasks where the effect of only one probability or risk level on the choice of non-risky or risky options was examined. The conducted research was aimed to examine the framing effect in the function of probability level in the outcome of a risk option in three decision-making domains: health, money and human lives. It has been confirmed that the decision-making domain moderates the framing effect. In the monetary domain, the general risk aversion has been confirmed as registered in earlier research. At high probability levels, the framing effect is registered in both frames, while no framing effect is registered at lower probability levels. In the domain of decision-making about human lives, the framing effect is registered at medium high and medium low probability levels. In the domain of decision-making about health, the framing effect is registered almost in the entire probability range while this domain differs from the former two. The results show that the attitude to risk is not identical at different probability levels, that the dynamics of the attitude to risk influences the framing effect, and that the framing effect pattern is different in different decision-making domains. In other words, linguistic manipulation representing the frame in the tasks affects the change in the preference order only when the possibility of gain (expressed in probability is estimated as sufficiently high.
Probability theory and mathematical statistics for engineers
Pugachev, V S
1984-01-01
Probability Theory and Mathematical Statistics for Engineers focuses on the concepts of probability theory and mathematical statistics for finite-dimensional random variables.The publication first underscores the probabilities of events, random variables, and numerical characteristics of random variables. Discussions focus on canonical expansions of random vectors, second-order moments of random vectors, generalization of the density concept, entropy of a distribution, direct evaluation of probabilities, and conditional probabilities. The text then examines projections of random vector
Oil spill contamination probability in the southeastern Levantine basin.
Goldman, Ron; Biton, Eli; Brokovich, Eran; Kark, Salit; Levin, Noam
2015-02-15
Recent gas discoveries in the eastern Mediterranean Sea led to multiple operations with substantial economic interest, and with them there is a risk of oil spills and their potential environmental impacts. To examine the potential spatial distribution of this threat, we created seasonal maps of the probability of oil spill pollution reaching an area in the Israeli coastal and exclusive economic zones, given knowledge of its initial sources. We performed simulations of virtual oil spills using realistic atmospheric and oceanic conditions. The resulting maps show dominance of the alongshore northerly current, which causes the high probability areas to be stretched parallel to the coast, increasing contamination probability downstream of source points. The seasonal westerly wind forcing determines how wide the high probability areas are, and may also restrict these to a small coastal region near source points. Seasonal variability in probability distribution, oil state, and pollution time is also discussed. Copyright © 2014 Elsevier Ltd. All rights reserved.
Dinov, Ivo D; Siegrist, Kyle; Pearl, Dennis K; Kalinin, Alexandr; Christou, Nicolas
2016-06-01
Probability distributions are useful for modeling, simulation, analysis, and inference on varieties of natural processes and physical phenomena. There are uncountably many probability distributions. However, a few dozen families of distributions are commonly defined and are frequently used in practice for problem solving, experimental applications, and theoretical studies. In this paper, we present a new computational and graphical infrastructure, the Distributome , which facilitates the discovery, exploration and application of diverse spectra of probability distributions. The extensible Distributome infrastructure provides interfaces for (human and machine) traversal, search, and navigation of all common probability distributions. It also enables distribution modeling, applications, investigation of inter-distribution relations, as well as their analytical representations and computational utilization. The entire Distributome framework is designed and implemented as an open-source, community-built, and Internet-accessible infrastructure. It is portable, extensible and compatible with HTML5 and Web2.0 standards (http://Distributome.org). We demonstrate two types of applications of the probability Distributome resources: computational research and science education. The Distributome tools may be employed to address five complementary computational modeling applications (simulation, data-analysis and inference, model-fitting, examination of the analytical, mathematical and computational properties of specific probability distributions, and exploration of the inter-distributional relations). Many high school and college science, technology, engineering and mathematics (STEM) courses may be enriched by the use of modern pedagogical approaches and technology-enhanced methods. The Distributome resources provide enhancements for blended STEM education by improving student motivation, augmenting the classical curriculum with interactive webapps, and overhauling the
Mathijssen, J.W.
2007-01-01
The subject of "The Breach and the Observance" is retranslation for the theatre. Besides offering a model that incorporates the findings of previous scholarship, it casts new light on the motivation behind retranslation, using the case of translations of Shakespeare's Hamlet on the Dutch stage.
Lange, A.H. de; Bal, P.M.; Heijden, B.I.J.M. van der; Jong, N. de; Schaufeli, W.B.
2011-01-01
There is an increasing need for managers to understand what motivates younger versus older workers to continue work within their company. We believe that this two-wave study among 90 Dutch employees is the first to examine: (1) the cross-lagged relationships between breach of psychological contract
de Lange, Annet H.; Bal, P. Matthijs; Van der Heijden, Beatrice I. J. M.; de Jong, Nicole; Schaufeli, Wilmar B.
2011-01-01
There is an increasing need for managers to understand what motivates younger versus older workers to continue work within their company. We believe that this two-wave study among 90 Dutch employees is the first to examine: (1) the cross-lagged relationships between breach of psychological contract
Posterior Probability Matching and Human Perceptual Decision Making.
Directory of Open Access Journals (Sweden)
Richard F Murray
2015-06-01
Full Text Available Probability matching is a classic theory of decision making that was first developed in models of cognition. Posterior probability matching, a variant in which observers match their response probabilities to the posterior probability of each response being correct, is being used increasingly often in models of perception. However, little is known about whether posterior probability matching is consistent with the vast literature on vision and hearing that has developed within signal detection theory. Here we test posterior probability matching models using two tools from detection theory. First, we examine the models' performance in a two-pass experiment, where each block of trials is presented twice, and we measure the proportion of times that the model gives the same response twice to repeated stimuli. We show that at low performance levels, posterior probability matching models give highly inconsistent responses across repeated presentations of identical trials. We find that practised human observers are more consistent across repeated trials than these models predict, and we find some evidence that less practised observers more consistent as well. Second, we compare the performance of posterior probability matching models on a discrimination task to the performance of a theoretical ideal observer that achieves the best possible performance. We find that posterior probability matching is very inefficient at low-to-moderate performance levels, and that human observers can be more efficient than is ever possible according to posterior probability matching models. These findings support classic signal detection models, and rule out a broad class of posterior probability matching models for expert performance on perceptual tasks that range in complexity from contrast discrimination to symmetry detection. However, our findings leave open the possibility that inexperienced observers may show posterior probability matching behaviour, and our methods
Direct probability mapping of contaminants
International Nuclear Information System (INIS)
Rautman, C.A.
1993-01-01
Exhaustive characterization of a contaminated site is a physical and practical impossibility. Descriptions of the nature, extent, and level of contamination, as well as decisions regarding proposed remediation activities, must be made in a state of uncertainty based upon limited physical sampling. Geostatistical simulation provides powerful tools for investigating contaminant levels, and in particular, for identifying and using the spatial interrelationships among a set of isolated sample values. This additional information can be used to assess the likelihood of encountering contamination at unsampled locations and to evaluate the risk associated with decisions to remediate or not to remediate specific regions within a site. Past operation of the DOE Feed Materials Production Center has contaminated a site near Fernald, Ohio, with natural uranium. Soil geochemical data have been collected as part of the Uranium-in-Soils Integrated Demonstration Project. These data have been used to construct a number of stochastic images of potential contamination for parcels approximately the size of a selective remediation unit. Each such image accurately reflects the actual measured sample values, and reproduces the univariate statistics and spatial character of the extant data. Post-processing of a large number of these equally likely, statistically similar images produces maps directly showing the probability of exceeding specified levels of contamination. Evaluation of the geostatistical simulations can yield maps representing the expected magnitude of the contamination for various regions and other information that may be important in determining a suitable remediation process or in sizing equipment to accomplish the restoration
A Multidisciplinary Approach for Teaching Statistics and Probability
Rao, C. Radhakrishna
1971-01-01
The author presents a syllabus for an introductory (first year after high school) course in statistics and probability and some methods of teaching statistical techniques. The description comes basically from the procedures used at the Indian Statistical Institute, Calcutta. (JG)
Psychophysics of the probability weighting function
Takahashi, Taiki
2011-03-01
A probability weighting function w(p) for an objective probability p in decision under risk plays a pivotal role in Kahneman-Tversky prospect theory. Although recent studies in econophysics and neuroeconomics widely utilized probability weighting functions, psychophysical foundations of the probability weighting functions have been unknown. Notably, a behavioral economist Prelec (1998) [4] axiomatically derived the probability weighting function w(p)=exp(-() (01e)=1e,w(1)=1), which has extensively been studied in behavioral neuroeconomics. The present study utilizes psychophysical theory to derive Prelec's probability weighting function from psychophysical laws of perceived waiting time in probabilistic choices. Also, the relations between the parameters in the probability weighting function and the probability discounting function in behavioral psychology are derived. Future directions in the application of the psychophysical theory of the probability weighting function in econophysics and neuroeconomics are discussed.
Jones, Samantha K; Griep, Yannick
2018-01-01
Employees often draw meaning from personal experiences and contributions in their work, particularly when engaging in organizational activities that align with their personal identity or values. However, recent empirical findings have demonstrated how meaningful work can also have a negative effect on employee's well-being as employees feel so invested in their work, they push themselves beyond their limits resulting in strain and susceptibility to burnout. We develop a framework to understand this "double edged" role of meaningful work by drawing from ideological psychological contracts (iPCs), which are characterized by employees and their employer who are working to contribute to a shared ideology or set of values. Limited iPC research has demonstrated employees may actually work harder in response to an iPC breach. In light of these counterintuitive findings, we propose the following conceptual model to theoretically connect our understanding of iPCs, perceptions of breach, increases in work effort, and the potential "dark side" of repeated occurrences of iPC breach. We argue that time plays a central role in the unfolding process of employees' reactions to iPC breach over time. Further, we propose how perceptions of iPC breach relate to strain and, eventually, burnout. This model contributes to our understanding of the role of time in iPC development and maintenance, expands our exploration of ideology in the PC literature, and provides a framework to understanding why certain occupations are more susceptible to instances of strain and burnout. This framework has the potential to guide future employment interventions in ideology-infused organizations to help mitigate negative employee outcomes.
Jones, Samantha K.; Griep, Yannick
2018-01-01
Employees often draw meaning from personal experiences and contributions in their work, particularly when engaging in organizational activities that align with their personal identity or values. However, recent empirical findings have demonstrated how meaningful work can also have a negative effect on employee’s well-being as employees feel so invested in their work, they push themselves beyond their limits resulting in strain and susceptibility to burnout. We develop a framework to understand this “double edged” role of meaningful work by drawing from ideological psychological contracts (iPCs), which are characterized by employees and their employer who are working to contribute to a shared ideology or set of values. Limited iPC research has demonstrated employees may actually work harder in response to an iPC breach. In light of these counterintuitive findings, we propose the following conceptual model to theoretically connect our understanding of iPCs, perceptions of breach, increases in work effort, and the potential “dark side” of repeated occurrences of iPC breach. We argue that time plays a central role in the unfolding process of employees’ reactions to iPC breach over time. Further, we propose how perceptions of iPC breach relate to strain and, eventually, burnout. This model contributes to our understanding of the role of time in iPC development and maintenance, expands our exploration of ideology in the PC literature, and provides a framework to understanding why certain occupations are more susceptible to instances of strain and burnout. This framework has the potential to guide future employment interventions in ideology-infused organizations to help mitigate negative employee outcomes. PMID:29479334
Directory of Open Access Journals (Sweden)
Samantha K. Jones
2018-02-01
Full Text Available Employees often draw meaning from personal experiences and contributions in their work, particularly when engaging in organizational activities that align with their personal identity or values. However, recent empirical findings have demonstrated how meaningful work can also have a negative effect on employee’s well-being as employees feel so invested in their work, they push themselves beyond their limits resulting in strain and susceptibility to burnout. We develop a framework to understand this “double edged” role of meaningful work by drawing from ideological psychological contracts (iPCs, which are characterized by employees and their employer who are working to contribute to a shared ideology or set of values. Limited iPC research has demonstrated employees may actually work harder in response to an iPC breach. In light of these counterintuitive findings, we propose the following conceptual model to theoretically connect our understanding of iPCs, perceptions of breach, increases in work effort, and the potential “dark side” of repeated occurrences of iPC breach. We argue that time plays a central role in the unfolding process of employees’ reactions to iPC breach over time. Further, we propose how perceptions of iPC breach relate to strain and, eventually, burnout. This model contributes to our understanding of the role of time in iPC development and maintenance, expands our exploration of ideology in the PC literature, and provides a framework to understanding why certain occupations are more susceptible to instances of strain and burnout. This framework has the potential to guide future employment interventions in ideology-infused organizations to help mitigate negative employee outcomes.
THE BLACK HOLE FORMATION PROBABILITY
Energy Technology Data Exchange (ETDEWEB)
Clausen, Drew; Piro, Anthony L.; Ott, Christian D., E-mail: dclausen@tapir.caltech.edu [TAPIR, Walter Burke Institute for Theoretical Physics, California Institute of Technology, Mailcode 350-17, Pasadena, CA 91125 (United States)
2015-02-01
A longstanding question in stellar evolution is which massive stars produce black holes (BHs) rather than neutron stars (NSs) upon death. It has been common practice to assume that a given zero-age main sequence (ZAMS) mass star (and perhaps a given metallicity) simply produces either an NS or a BH, but this fails to account for a myriad of other variables that may effect this outcome, such as spin, binarity, or even stochastic differences in the stellar structure near core collapse. We argue that instead a probabilistic description of NS versus BH formation may be better suited to account for the current uncertainties in understanding how massive stars die. We present an initial exploration of the probability that a star will make a BH as a function of its ZAMS mass, P {sub BH}(M {sub ZAMS}). Although we find that it is difficult to derive a unique P {sub BH}(M {sub ZAMS}) using current measurements of both the BH mass distribution and the degree of chemical enrichment by massive stars, we demonstrate how P {sub BH}(M {sub ZAMS}) changes with these various observational and theoretical uncertainties. We anticipate that future studies of Galactic BHs and theoretical studies of core collapse will refine P {sub BH}(M {sub ZAMS}) and argue that this framework is an important new step toward better understanding BH formation. A probabilistic description of BH formation will be useful as input for future population synthesis studies that are interested in the formation of X-ray binaries, the nature and event rate of gravitational wave sources, and answering questions about chemical enrichment.
THE BLACK HOLE FORMATION PROBABILITY
International Nuclear Information System (INIS)
Clausen, Drew; Piro, Anthony L.; Ott, Christian D.
2015-01-01
A longstanding question in stellar evolution is which massive stars produce black holes (BHs) rather than neutron stars (NSs) upon death. It has been common practice to assume that a given zero-age main sequence (ZAMS) mass star (and perhaps a given metallicity) simply produces either an NS or a BH, but this fails to account for a myriad of other variables that may effect this outcome, such as spin, binarity, or even stochastic differences in the stellar structure near core collapse. We argue that instead a probabilistic description of NS versus BH formation may be better suited to account for the current uncertainties in understanding how massive stars die. We present an initial exploration of the probability that a star will make a BH as a function of its ZAMS mass, P BH (M ZAMS ). Although we find that it is difficult to derive a unique P BH (M ZAMS ) using current measurements of both the BH mass distribution and the degree of chemical enrichment by massive stars, we demonstrate how P BH (M ZAMS ) changes with these various observational and theoretical uncertainties. We anticipate that future studies of Galactic BHs and theoretical studies of core collapse will refine P BH (M ZAMS ) and argue that this framework is an important new step toward better understanding BH formation. A probabilistic description of BH formation will be useful as input for future population synthesis studies that are interested in the formation of X-ray binaries, the nature and event rate of gravitational wave sources, and answering questions about chemical enrichment
The Black Hole Formation Probability
Clausen, Drew; Piro, Anthony L.; Ott, Christian D.
2015-02-01
A longstanding question in stellar evolution is which massive stars produce black holes (BHs) rather than neutron stars (NSs) upon death. It has been common practice to assume that a given zero-age main sequence (ZAMS) mass star (and perhaps a given metallicity) simply produces either an NS or a BH, but this fails to account for a myriad of other variables that may effect this outcome, such as spin, binarity, or even stochastic differences in the stellar structure near core collapse. We argue that instead a probabilistic description of NS versus BH formation may be better suited to account for the current uncertainties in understanding how massive stars die. We present an initial exploration of the probability that a star will make a BH as a function of its ZAMS mass, P BH(M ZAMS). Although we find that it is difficult to derive a unique P BH(M ZAMS) using current measurements of both the BH mass distribution and the degree of chemical enrichment by massive stars, we demonstrate how P BH(M ZAMS) changes with these various observational and theoretical uncertainties. We anticipate that future studies of Galactic BHs and theoretical studies of core collapse will refine P BH(M ZAMS) and argue that this framework is an important new step toward better understanding BH formation. A probabilistic description of BH formation will be useful as input for future population synthesis studies that are interested in the formation of X-ray binaries, the nature and event rate of gravitational wave sources, and answering questions about chemical enrichment.