WorldWideScience

Sample records for projections represent average

  1. To quantum averages through asymptotic expansion of classical averages on infinite-dimensional space

    International Nuclear Information System (INIS)

    Khrennikov, Andrei

    2007-01-01

    We study asymptotic expansions of Gaussian integrals of analytic functionals on infinite-dimensional spaces (Hilbert and nuclear Frechet). We obtain an asymptotic equality coupling the Gaussian integral and the trace of the composition of scaling of the covariation operator of a Gaussian measure and the second (Frechet) derivative of a functional. In this way we couple classical average (given by an infinite-dimensional Gaussian integral) and quantum average (given by the von Neumann trace formula). We can interpret this mathematical construction as a procedure of 'dequantization' of quantum mechanics. We represent quantum mechanics as an asymptotic projection of classical statistical mechanics with infinite-dimensional phase space. This space can be represented as the space of classical fields, so quantum mechanics is represented as a projection of 'prequantum classical statistical field theory'

  2. 78 FR 17648 - Energy Conservation Program for Consumer Products: Representative Average Unit Costs of Energy

    Science.gov (United States)

    2013-03-22

    ... Conservation Program for Consumer Products: Representative Average Unit Costs of Energy'', dated April 26, 2012... DEPARTMENT OF ENERGY Office of Energy Efficiency and Renewable Energy Energy Conservation Program for Consumer Products: Representative Average Unit Costs of Energy AGENCY: Office of Energy Efficiency...

  3. Object detection by correlation coefficients using azimuthally averaged reference projections.

    Science.gov (United States)

    Nicholson, William V

    2004-11-01

    A method of computing correlation coefficients for object detection that takes advantage of using azimuthally averaged reference projections is described and compared with two alternative methods-computing a cross-correlation function or a local correlation coefficient versus the azimuthally averaged reference projections. Two examples of an application from structural biology involving the detection of projection views of biological macromolecules in electron micrographs are discussed. It is found that a novel approach to computing a local correlation coefficient versus azimuthally averaged reference projections, using a rotational correlation coefficient, outperforms using a cross-correlation function and a local correlation coefficient in object detection from simulated images with a range of levels of simulated additive noise. The three approaches perform similarly in detecting macromolecular views in electron microscope images of a globular macrolecular complex (the ribosome). The rotational correlation coefficient outperforms the other methods in detection of keyhole limpet hemocyanin macromolecular views in electron micrographs.

  4. 76 FR 13168 - Energy Conservation Program for Consumer Products: Representative Average Unit Costs of Energy

    Science.gov (United States)

    2011-03-10

    ... average unit costs of residential energy in a Federal Register notice entitled, ``Energy Conservation... DEPARTMENT OF ENERGY Office of Energy Efficiency and Renewable Energy Energy Conservation Program for Consumer Products: Representative Average Unit Costs of Energy AGENCY: Office of Energy Efficiency...

  5. Average regional end-use energy price projections to the year 2030

    International Nuclear Information System (INIS)

    1991-01-01

    The energy prices shown in this report cover the period from 1991 through 2030. These prices reflect sector/fuel price projections from the Annual Energy Outlook 1991 (AEO) base case, developed using the Energy Information Administration's (EIA) Intermediate Future Forecasting System (IFFS) forecasting model. Projections through 2010 are AEO base case forecasts. Projections for the period from 2011 through 2030 were developed separately from the AEO for this report, and the basis for these projections is described in Chapter 3. Projections in this report include average energy prices for each of four Census Regions for the residential, commercial, industrial, and transportation end-use sectors. Energy sources include electricity, distillate fuel oil, liquefied petroleum gas, motor gasoline, residual fuel oil, natural gas, and steam coal. (VC)

  6. The average carbon-stock approach for small-scale CDM AR projects

    Energy Technology Data Exchange (ETDEWEB)

    Garcia Quijano, J.F.; Muys, B. [Katholieke Universiteit Leuven, Laboratory for Forest, Nature and Landscape Research, Leuven (Belgium); Schlamadinger, B. [Joanneum Research Forschungsgesellschaft mbH, Institute for Energy Research, Graz (Austria); Emmer, I. [Face Foundation, Arnhem (Netherlands); Somogyi, Z. [Forest Research Institute, Budapest (Hungary); Bird, D.N. [Woodrising Consulting Inc., Belfountain, Ontario (Canada)

    2004-06-15

    In many afforestation and reforestation (AR) projects harvesting with stand regeneration forms an integral part of the silvicultural system and satisfies local timber and/or fuelwood demand. Especially clear-cut harvesting will lead to an abrupt and significant reduction of carbon stocks. The smaller the project, the more significant the fluctuations of the carbon stocks may be. In the extreme case a small-scale project could consist of a single forest stand. In such case, all accounted carbon may be removed during a harvesting operation and the time-path of carbon stocks will typically look as in the hypothetical example presented in the report. For the aggregate of many such small-scale projects there will be a constant benefit to the atmosphere during the projects, due to averaging effects.

  7. Adaptation of Regional Representative Soil Project and Soil Judging for Cameroon

    Science.gov (United States)

    Che, Celestine Akuma

    2013-01-01

    Representative regional soils have agricultural, cultural, economic, environmental, and historical importance to Cameroon. Twenty seven regional representative soils have been identified in Cameroon. A set of laboratory exercises, assignments and exam questions have been developed utilizing the Regional Representative Soil Project (RRSP) that…

  8. Dosimetric consequences of planning lung treatments on 4DCT average reconstruction to represent a moving tumour

    International Nuclear Information System (INIS)

    Dunn, L.F.; Taylor, M.L.; Kron, T.; Franich, R.

    2010-01-01

    Full text: Anatomic motion during a radiotherapy treatment is one of the more significant challenges in contemporary radiation therapy. For tumours of the lung, motion due to patient respiration makes both accurate planning and dose delivery difficult. One approach is to use the maximum intensity projection (MIP) obtained from a 40 computed tomography (CT) scan and then use this to determine the treatment volume. The treatment is then planned on a 4DCT average reco struction, rather than assuming the entire ITY has a uniform tumour density. This raises the question: how well does planning on a 'blurred' distribution of density with CT values greater than lung density but less than tumour density match the true case of a tumour moving within lung tissue? The aim of this study was to answer this question, determining the dosimetric impact of using a 4D-CT average reconstruction as the basis for a radiotherapy treatment plan. To achieve this, Monte-Carlo sim ulations were undertaken using GEANT4. The geometry consisted of a tumour (diameter 30 mm) moving with a sinusoidal pattern of amplitude = 20 mm. The tumour's excursion occurs within a lung equivalent volume beyond a chest wall interface. Motion was defined parallel to a 6 MY beam. This was then compared to a single oblate tumour of a magnitude determined by the extremes of the tumour motion. The variable density of the 4DCT average tumour is simulated by a time-weighted average, to achieve the observed density gradient. The generic moving tumour geometry is illustrated in the Figure.

  9. On Averaging Rotations

    DEFF Research Database (Denmark)

    Gramkow, Claus

    1999-01-01

    In this article two common approaches to averaging rotations are compared to a more advanced approach based on a Riemannian metric. Very offten the barycenter of the quaternions or matrices that represent the rotations are used as an estimate of the mean. These methods neglect that rotations belo...... approximations to the Riemannian metric, and that the subsequent corrections are inherient in the least squares estimation. Keywords: averaging rotations, Riemannian metric, matrix, quaternion......In this article two common approaches to averaging rotations are compared to a more advanced approach based on a Riemannian metric. Very offten the barycenter of the quaternions or matrices that represent the rotations are used as an estimate of the mean. These methods neglect that rotations belong...

  10. Characterisation of representative building typologies for social housing projects in Brazil and its energy performance

    International Nuclear Information System (INIS)

    Triana, Maria Andrea; Lamberts, Roberto; Sassi, Paola

    2015-01-01

    In Brazil the housing deficit is around 5.5 million houses. To address this need, the government created a programme called “My house, My life”. The main subsidies of the programme are for families earning up to three times the minimum wage. In order to formulate strategies for more energy efficiency buildings, it is necessary to understand the thermal and energy performance of what is being built. This article defines representative projects for typologies being built in the Brazilian social housing sector through the analysis of 108 projects considering two groups of income levels and investigates the thermal and energy performance of the representative projects in relation to the Regulation for Energy Efficiency Labelling of Residential Buildings in Brazil for two bioclimatic zones. Five representative building models were defined. Considering the most common features found on the sample, the study suggests the importance of addresing energy efficiency measures on the sector since current building techniques for social housing shows a tendency towards a low performance in relation to the thermal and energy performance criteria of the Energy Labelling especially for lower income projects. This provides a basis for future policy and allows for more in depth studies within the sector. - Highlights: • Characterisation of representative typologies built for social housing in Brazil. • More recurrent building physics characteristics considered. • Energy efficiency and thermal performance of Brazilian social housing analysed. • Regulation for Energy Efficiency Labelling of Residential Buildings in Brazil used for analysis.

  11. Multi-Repeated Projection Lithography for High-Precision Linear Scale Based on Average Homogenization Effect

    Directory of Open Access Journals (Sweden)

    Dongxu Ren

    2016-04-01

    Full Text Available A multi-repeated photolithography method for manufacturing an incremental linear scale using projection lithography is presented. The method is based on the average homogenization effect that periodically superposes the light intensity of different locations of pitches in the mask to make a consistent energy distribution at a specific wavelength, from which the accuracy of a linear scale can be improved precisely using the average pitch with different step distances. The method’s theoretical error is within 0.01 µm for a periodic mask with a 2-µm sine-wave error. The intensity error models in the focal plane include the rectangular grating error on the mask, static positioning error, and lithography lens focal plane alignment error, which affect pitch uniformity less than in the common linear scale projection lithography splicing process. It was analyzed and confirmed that increasing the repeat exposure number of a single stripe could improve accuracy, as could adjusting the exposure spacing to achieve a set proportion of black and white stripes. According to the experimental results, the effectiveness of the multi-repeated photolithography method is confirmed to easily realize a pitch accuracy of 43 nm in any 10 locations of 1 m, and the whole length accuracy of the linear scale is less than 1 µm/m.

  12. On Averaging Rotations

    DEFF Research Database (Denmark)

    Gramkow, Claus

    2001-01-01

    In this paper two common approaches to averaging rotations are compared to a more advanced approach based on a Riemannian metric. Very often the barycenter of the quaternions or matrices that represent the rotations are used as an estimate of the mean. These methods neglect that rotations belong ...... approximations to the Riemannian metric, and that the subsequent corrections are inherent in the least squares estimation.......In this paper two common approaches to averaging rotations are compared to a more advanced approach based on a Riemannian metric. Very often the barycenter of the quaternions or matrices that represent the rotations are used as an estimate of the mean. These methods neglect that rotations belong...

  13. An average-based accounting approach to capital asset investments: The case of project finance

    OpenAIRE

    Carlo Alberto Magni

    2014-01-01

    Literature and textbooks on capital budgeting endorse Net Present Value (NPV) and generally treat accounting rates of return as not being reliable tools. This paper shows that accounting numbers can be reconciled with NPV and fruitfully employed in real-life applications. Focusing on project finance transactions, an Average Return On Investment (AROI) is drawn from the pro forma financial statements, obtained as the ratio of aggregate income to aggregate book value. It is shown that such a me...

  14. A framework for extracting and representing project knowledge contexts using topic models and dynamic knowledge maps

    Science.gov (United States)

    Xu, Jin; Li, Zheng; Li, Shuliang; Zhang, Yanyan

    2015-07-01

    There is still a lack of effective paradigms and tools for analysing and discovering the contents and relationships of project knowledge contexts in the field of project management. In this paper, a new framework for extracting and representing project knowledge contexts using topic models and dynamic knowledge maps under big data environments is proposed and developed. The conceptual paradigm, theoretical underpinning, extended topic model, and illustration examples of the ontology model for project knowledge maps are presented, with further research work envisaged.

  15. Averaging models: parameters estimation with the R-Average procedure

    Directory of Open Access Journals (Sweden)

    S. Noventa

    2010-01-01

    Full Text Available The Functional Measurement approach, proposed within the theoretical framework of Information Integration Theory (Anderson, 1981, 1982, can be a useful multi-attribute analysis tool. Compared to the majority of statistical models, the averaging model can account for interaction effects without adding complexity. The R-Average method (Vidotto & Vicentini, 2007 can be used to estimate the parameters of these models. By the use of multiple information criteria in the model selection procedure, R-Average allows for the identification of the best subset of parameters that account for the data. After a review of the general method, we present an implementation of the procedure in the framework of R-project, followed by some experiments using a Monte Carlo method.

  16. Representativeness of single lidar stations for zonally averaged ozone profiles, their trends and attribution to proxies

    Directory of Open Access Journals (Sweden)

    C. Zerefos

    2018-05-01

    Full Text Available This paper is focusing on the representativeness of single lidar stations for zonally averaged ozone profile variations over the middle and upper stratosphere. From the lower to the upper stratosphere, ozone profiles from single or grouped lidar stations correlate well with zonal means calculated from the Solar Backscatter Ultraviolet Radiometer (SBUV satellite overpasses. The best representativeness with significant correlation coefficients is found within ±15° of latitude circles north or south of any lidar station. This paper also includes a multivariate linear regression (MLR analysis on the relative importance of proxy time series for explaining variations in the vertical ozone profiles. Studied proxies represent variability due to influences outside of the earth system (solar cycle and within the earth system, i.e. dynamic processes (the Quasi Biennial Oscillation, QBO; the Arctic Oscillation, AO; the Antarctic Oscillation, AAO; the El Niño Southern Oscillation, ENSO, those due to volcanic aerosol (aerosol optical depth, AOD, tropopause height changes (including global warming and those influences due to anthropogenic contributions to atmospheric chemistry (equivalent effective stratospheric chlorine, EESC. Ozone trends are estimated, with and without removal of proxies, from the total available 1980 to 2015 SBUV record. Except for the chemistry related proxy (EESC and its orthogonal function, the removal of the other proxies does not alter the significance of the estimated long-term trends. At heights above 15 hPa an inflection point between 1997 and 1999 marks the end of significant negative ozone trends, followed by a recent period between 1998 and 2015 with positive ozone trends. At heights between 15 and 40 hPa the pre-1998 negative ozone trends tend to become less significant as we move towards 2015, below which the lower stratosphere ozone decline continues in agreement with findings of recent literature.

  17. Representativeness of single lidar stations for zonally averaged ozone profiles, their trends and attribution to proxies

    Science.gov (United States)

    Zerefos, Christos; Kapsomenakis, John; Eleftheratos, Kostas; Tourpali, Kleareti; Petropavlovskikh, Irina; Hubert, Daan; Godin-Beekmann, Sophie; Steinbrecht, Wolfgang; Frith, Stacey; Sofieva, Viktoria; Hassler, Birgit

    2018-05-01

    This paper is focusing on the representativeness of single lidar stations for zonally averaged ozone profile variations over the middle and upper stratosphere. From the lower to the upper stratosphere, ozone profiles from single or grouped lidar stations correlate well with zonal means calculated from the Solar Backscatter Ultraviolet Radiometer (SBUV) satellite overpasses. The best representativeness with significant correlation coefficients is found within ±15° of latitude circles north or south of any lidar station. This paper also includes a multivariate linear regression (MLR) analysis on the relative importance of proxy time series for explaining variations in the vertical ozone profiles. Studied proxies represent variability due to influences outside of the earth system (solar cycle) and within the earth system, i.e. dynamic processes (the Quasi Biennial Oscillation, QBO; the Arctic Oscillation, AO; the Antarctic Oscillation, AAO; the El Niño Southern Oscillation, ENSO), those due to volcanic aerosol (aerosol optical depth, AOD), tropopause height changes (including global warming) and those influences due to anthropogenic contributions to atmospheric chemistry (equivalent effective stratospheric chlorine, EESC). Ozone trends are estimated, with and without removal of proxies, from the total available 1980 to 2015 SBUV record. Except for the chemistry related proxy (EESC) and its orthogonal function, the removal of the other proxies does not alter the significance of the estimated long-term trends. At heights above 15 hPa an inflection point between 1997 and 1999 marks the end of significant negative ozone trends, followed by a recent period between 1998 and 2015 with positive ozone trends. At heights between 15 and 40 hPa the pre-1998 negative ozone trends tend to become less significant as we move towards 2015, below which the lower stratosphere ozone decline continues in agreement with findings of recent literature.

  18. Averaged 30 year climate change projections mask opportunities for species establishment

    Science.gov (United States)

    Serra-Diaz, Josep M.; Franklin, Janet; Sweet, Lynn C.; McCullough, Ian M.; Syphard, Alexandra D.; Regan, Helen M.; Flint, Lorraine E.; Flint, Alan L.; Dingman, John; Moritz, Max A.; Redmond, Kelly T.; Hannah, Lee; Davis, Frank W.

    2016-01-01

    Survival of early life stages is key for population expansion into new locations and for persistence of current populations (Grubb 1977, Harper 1977). Relative to adults, these early life stages are very sensitive to climate fl uctuations (Ropert-Coudert et al. 2015), which often drive episodic or ‘event-limited’ regeneration (e.g. pulses) in long-lived plant species (Jackson et al. 2009). Th us, it is diffi cult to mechanistically associate 30-yr climate norms to dynamic processes involved in species range shifts (e.g. seedling survival). What are the consequences of temporal aggregation for estimating areas of potential establishment? We modeled seedling survival for three widespread tree species in California, USA ( Quercus douglasii, Q. kelloggii , Pinus sabiniana ) by coupling a large-scale, multi-year common garden experiment to high-resolution downscaled grids of climatic water defi cit and air temperature (Flint and Flint 2012, Supplementary material Appendix 1). We projected seedling survival for nine climate change projections in two mountain landscapes spanning wide elevation and moisture gradients. We compared areas with windows of opportunity for seedling survival – defi ned as three consecutive years of seedling survival in our species, a period selected based on studies of tree niche ontogeny (Supplementary material Appendix 1) – to areas of 30-yr averaged estimates of seedling survival. We found that temporal aggregation greatly underestimated the potential for species establishment (e.g. seedling survival) under climate change scenarios.

  19. Impact of Role Clarity and Strategic Fit on Average Project Success: Moderating Role of Market Turbulence on Telecom Companies of Pakistan

    Directory of Open Access Journals (Sweden)

    Najam UL MABOOD

    2017-06-01

    Full Text Available Advancement in technology has reshaped the businesses across the globe forcing companies to perform tasks and activities in the form of projects. Stakeholder behavior, stakeholder management, strategic fit, role and task clarity are some of the factors that redesign the project success. The current study examine the impact of strategic fit and role clarity on the Average project success and further it enlightens the moderating role of Market turbulence on the relationship between the aforementioned independent and dependent variables. The population of the study comprises of telecom sector of Pakistan. The Data was collected from 201 project team members working on diverse project in Telecom companies of Rawalpindi and Islamabad. The Data was gathered through a questionnaires measured on Likert scale adopted from the study of Beringer, Jonas & Kock (2013. Each Questionnaire comprises of 3 items to measure each variable. SPSS 20.0 Version was used to analyze the data by applying Pearson correlation and multiple regression analysis technique. Findings depicted that role clarity and strategic fit contributed significantly in enhancing success of a project. Results further evidenced that market turbulence negatively moderated the relationship of independent variables on Average project success. The study at the end highlights recommendations for the future researchers.

  20. Reliability ensemble averaging of 21st century projections of terrestrial net primary productivity reduces global and regional uncertainties

    Science.gov (United States)

    Exbrayat, Jean-François; Bloom, A. Anthony; Falloon, Pete; Ito, Akihiko; Smallman, T. Luke; Williams, Mathew

    2018-02-01

    Multi-model averaging techniques provide opportunities to extract additional information from large ensembles of simulations. In particular, present-day model skill can be used to evaluate their potential performance in future climate simulations. Multi-model averaging methods have been used extensively in climate and hydrological sciences, but they have not been used to constrain projected plant productivity responses to climate change, which is a major uncertainty in Earth system modelling. Here, we use three global observationally orientated estimates of current net primary productivity (NPP) to perform a reliability ensemble averaging (REA) method using 30 global simulations of the 21st century change in NPP based on the Inter-Sectoral Impact Model Intercomparison Project (ISIMIP) business as usual emissions scenario. We find that the three REA methods support an increase in global NPP by the end of the 21st century (2095-2099) compared to 2001-2005, which is 2-3 % stronger than the ensemble ISIMIP mean value of 24.2 Pg C y-1. Using REA also leads to a 45-68 % reduction in the global uncertainty of 21st century NPP projection, which strengthens confidence in the resilience of the CO2 fertilization effect to climate change. This reduction in uncertainty is especially clear for boreal ecosystems although it may be an artefact due to the lack of representation of nutrient limitations on NPP in most models. Conversely, the large uncertainty that remains on the sign of the response of NPP in semi-arid regions points to the need for better observations and model development in these regions.

  1. [Experience of knowledge translation in the ITSAL (immigration, work and health) research project with representatives of the target population].

    Science.gov (United States)

    Ronda, Elena; López-Jacob, M José; Paredes-Carbonell, Joan J; López, Pilar; Boix, Pere; García, Ana M

    2014-01-01

    This article describes the experience of knowledge translation between researchers of the ITSAL (immigration, work and health) project and representatives of organizations working with immigrants to discuss the results obtained in the project and future research lines. A meeting was held, attended by three researchers and 18 representatives from 11 institutions. Following a presentation of the methodology and results of the project, the participants discussed the results presented and research areas of interest, thus confirming matches between the two sides and obtaining proposals of interest for the ITSAL project. We understand the process described as an approach to social validation of some of the main results of this project. This experience has allowed us to open a channel of communication with the target population of the study, in line with the necessary two-way interaction between researchers and users. Copyright © 2013 SESPAS. Published by Elsevier Espana. All rights reserved.

  2. A workforce survey of Australian osteopathy: analysis of a nationally-representative sample of osteopaths from the Osteopathy Research and Innovation Network (ORION) project.

    Science.gov (United States)

    Adams, Jon; Sibbritt, David; Steel, Amie; Peng, Wenbo

    2018-05-10

    Limited information is available regarding the profile and clinical practice characteristics of the osteopathy workforce in Australia. This paper reports such information by analysing data from a nationally-representative sample of Australian osteopaths. Data was obtained from a workforce survey of Australian osteopathy, investigating the characteristics of the practitioner, their practice, clinical management features and perceptions regarding research. The survey questionnaire was distributed to all registered osteopaths across Australia in 2016 as part of the Osteopathy Research and Innovation Network (ORION) project. A total of 992 Australian osteopaths participated in this study representing a response rate of 49.1%. The average age of the participants was 38.0 years with 58.1% being female and the majority holding a Bachelor or higher degree qualification related to the osteopathy professional. Approximately 80.0% of the osteopaths were practicing in an urban area, with most osteopaths working in multi-practitioner locations, having referral relationships with a range of health care practitioners, managing patients a number of musculoskeletal disorders, and providing multi-model treatment options. A total of 3.9 million patients were estimated to consult with osteopaths every year and an average of approximate 3.0 million hours were spent delivering osteopathy services per year. Further research is required to provide rich, in-depth examination regarding a range of osteopathy workforce issues which will help ensure safe, effective patient care to all receiving and providing treatments as part of the broader Australian health system.

  3. Representing time

    Directory of Open Access Journals (Sweden)

    Luca Poncellini

    2010-06-01

    Full Text Available The analysis of natural phenomena applied to architectural planning and design is facing the most fascinating and elusive of the four dimensions through which man attempts to define life within the universe: time. We all know what time is, said St. Augustine, but nobody knows how to describe it. Within architectural projects and representations, time rarely appears in explicit form. This paper presents the results of a research conducted by students of NABA and of the Polytechnic of Milan with the purpose of representing time considered as a key element within architectural projects. Student investigated new approaches and methodologies to represent time using the two-dimensional support of a sheet of paper.

  4. Quality planning in Construction Project

    Science.gov (United States)

    Othman, I.; Shafiq, Nasir; Nuruddin, M. F.

    2017-12-01

    The purpose of this paper is to investigate deeper on the factors that contribute to the effectiveness of quality planning, identifying the common problems encountered in quality planning, practices and ways for improvements in quality planning for construction projects. This paper involves data collected from construction company representatives across Malaysia that are obtained through semi-structured interviews as well as questionnaire distributions. Results shows that design of experiments (average index: 4.61), inspection (average index: 4.45) and quality audit as well as other methods (average index: 4.26) rank first, second and third most important factors respectively.

  5. Improving consensus structure by eliminating averaging artifacts

    Directory of Open Access Journals (Sweden)

    KC Dukka B

    2009-03-01

    Full Text Available Abstract Background Common structural biology methods (i.e., NMR and molecular dynamics often produce ensembles of molecular structures. Consequently, averaging of 3D coordinates of molecular structures (proteins and RNA is a frequent approach to obtain a consensus structure that is representative of the ensemble. However, when the structures are averaged, artifacts can result in unrealistic local geometries, including unphysical bond lengths and angles. Results Herein, we describe a method to derive representative structures while limiting the number of artifacts. Our approach is based on a Monte Carlo simulation technique that drives a starting structure (an extended or a 'close-by' structure towards the 'averaged structure' using a harmonic pseudo energy function. To assess the performance of the algorithm, we applied our approach to Cα models of 1364 proteins generated by the TASSER structure prediction algorithm. The average RMSD of the refined model from the native structure for the set becomes worse by a mere 0.08 Å compared to the average RMSD of the averaged structures from the native structure (3.28 Å for refined structures and 3.36 A for the averaged structures. However, the percentage of atoms involved in clashes is greatly reduced (from 63% to 1%; in fact, the majority of the refined proteins had zero clashes. Moreover, a small number (38 of refined structures resulted in lower RMSD to the native protein versus the averaged structure. Finally, compared to PULCHRA 1, our approach produces representative structure of similar RMSD quality, but with much fewer clashes. Conclusion The benchmarking results demonstrate that our approach for removing averaging artifacts can be very beneficial for the structural biology community. Furthermore, the same approach can be applied to almost any problem where averaging of 3D coordinates is performed. Namely, structure averaging is also commonly performed in RNA secondary prediction 2, which

  6. SU-E-T-174: Evaluation of the Optimal Intensity Modulated Radiation Therapy Plans Done On the Maximum and Average Intensity Projection CTs

    Energy Technology Data Exchange (ETDEWEB)

    Jurkovic, I [University of Texas Health Science Center at San Antonio, San Antonio, TX (United States); Stathakis, S; Li, Y; Patel, A; Vincent, J; Papanikolaou, N; Mavroidis, P [Cancer Therapy and Research Center University of Texas Health Sciences Center at San Antonio, San Antonio, TX (United States)

    2014-06-01

    Purpose: To determine the difference in coverage between plans done on average intensity projection and maximum intensity projection CT data sets for lung patients and to establish correlations between different factors influencing the coverage. Methods: For six lung cancer patients, 10 phases of equal duration through the respiratory cycle, the maximum and average intensity projections (MIP and AIP) from their 4DCT datasets were obtained. MIP and AIP datasets had three GTVs delineated (GTVaip — delineated on AIP, GTVmip — delineated on MIP and GTVfus — delineated on each of the 10 phases and summed up). From the each GTV, planning target volumes (PTV) were then created by adding additional margins. For each of the PTVs an IMRT plan was developed on the AIP dataset. The plans were then copied to the MIP data set and were recalculated. Results: The effective depths in AIP cases were significantly smaller than in MIP (p < 0.001). The Pearson correlation coefficient of r = 0.839 indicates strong degree of positive linear relationship between the average percentage difference in effective depths and average PTV coverage on the MIP data set. The V2 0 Gy of involved lung depends on the PTV coverage. The relationship between PTVaip mean CT number difference and PTVaip coverage on MIP data set gives r = 0.830. When the plans are produced on MIP and copied to AIP, r equals −0.756. Conclusion: The correlation between the AIP and MIP data sets indicates that the selection of the data set for developing the treatment plan affects the final outcome (cases with high average percentage difference in effective depths between AIP and MIP should be calculated on AIP). The percentage of the lung volume receiving higher dose depends on how well PTV is covered, regardless of on which set plan is done.

  7. Overview of the HiLASE project: high average power pulsed DPSSL systems for research and industry

    Czech Academy of Sciences Publication Activity Database

    Divoký, Martin; Smrž, Martin; Chyla, Michal; Sikocinski, Pawel; Severová, Patricie; Novák, Ondřej; Huynh, Jaroslav; Nagisetty, Siva S.; Miura, Taisuke; Pilař, Jan; Slezák, Jiří; Sawicka, Magdalena; Jambunathan, Venkatesan; Vanda, Jan; Endo, Akira; Lucianetti, Antonio; Rostohar, Danijela; Mason, P.D.; Phillips, P.J.; Ertel, K.; Banerjee, S.; Hernandez-Gomez, C.; Collier, J.L.; Mocek, Tomáš

    2014-01-01

    Roč. 2, SI (2014), s. 1-10 ISSN 2095-4719 R&D Projects: GA MŠk ED2.1.00/01.0027; GA MŠk EE2.3.20.0143; GA MŠk EE2.3.30.0057 Grant - others:HILASE(XE) CZ.1.05/2.1.00/01.0027; OP VK 6(XE) CZ.1.07/2.3.00/20.0143; OP VK 4 POSTDOK(XE) CZ.1.07/2.3.00/30.0057 Institutional support: RVO:68378271 Keywords : DPSSL * Yb3C:YAG * thin-disk * multi-slab * pulsed high average power laser Subject RIV: BH - Optics, Masers, Lasers

  8. Future Projection of Droughts over South Korea Using Representative Concentration Pathways (RCPs

    Directory of Open Access Journals (Sweden)

    Byung Sik Kim

    2014-01-01

    Full Text Available The Standardized Precipitation Index (SPI, a method widely used to analyze droughts related to climate change, does not consider variables related to temperature and is limited because it cannot consider changes in hydrological balance, such as evapotranspiration from climate change. If we were to consider only the future increase in precipitation from climate change, droughts may decrease. However, because usable water can diminish from an increase in evapotranspiration, it is important to research on projected droughts considering the amount of evapotranspiration along with projecting and evaluating potential droughts considering the impact of climate change. As such, this study evaluated the occurrence of droughts using the Standardized Precipitation Evapotranspiration Index (SPEI as a newly conceptualized drought index that is similar to SPI but includes the temperature variability. We extracted simulated future precipitation and temperature data (2011 - 2099 from the Representative Concentration Pathway (RCP climate change scenario of IPCC AR5 to evaluate the impact of future climate change on the occurrence of droughts of South Korea. We analyzed the ratio of evapotranspiration to precipitation of meteorological observatories nationwide. In addition, we calculated the SPEI related to drought in the process to evaluate the future occurrence of droughts of South Korea. To confirm validity of SPEI results, extreme indices were analyzed. This resulted in the notion that as we go further into the future, the precipitation increases. But because of an increase in evapotranspiration also from a rise in temperature and continued dryness, the severity of droughts is projected to exacerbate.

  9. Greater-than-Class C low-level waste characterization. Appendix I: Impact of concentration averaging low-level radioactive waste volume projections

    International Nuclear Information System (INIS)

    Tuite, P.; Tuite, K.; O'Kelley, M.; Ely, P.

    1991-08-01

    This study provides a quantitative framework for bounding unpackaged greater-than-Class C low-level radioactive waste types as a function of concentration averaging. The study defines the three concentration averaging scenarios that lead to base, high, and low volumetric projections; identifies those waste types that could be greater-than-Class C under the high volume, or worst case, concentration averaging scenario; and quantifies the impact of these scenarios on identified waste types relative to the base case scenario. The base volume scenario was assumed to reflect current requirements at the disposal sites as well as the regulatory views. The high volume scenario was assumed to reflect the most conservative criteria as incorporated in some compact host state requirements. The low volume scenario was assumed to reflect the 10 CFR Part 61 criteria as applicable to both shallow land burial facilities and to practices that could be employed to reduce the generation of Class C waste types

  10. The average covering tree value for directed graph games

    NARCIS (Netherlands)

    Khmelnitskaya, Anna Borisovna; Selcuk, Özer; Talman, Dolf

    We introduce a single-valued solution concept, the so-called average covering tree value, for the class of transferable utility games with limited communication structure represented by a directed graph. The solution is the average of the marginal contribution vectors corresponding to all covering

  11. The Average Covering Tree Value for Directed Graph Games

    NARCIS (Netherlands)

    Khmelnitskaya, A.; Selcuk, O.; Talman, A.J.J.

    2012-01-01

    Abstract: We introduce a single-valued solution concept, the so-called average covering tree value, for the class of transferable utility games with limited communication structure represented by a directed graph. The solution is the average of the marginal contribution vectors corresponding to all

  12. How to average logarithmic retrievals?

    Directory of Open Access Journals (Sweden)

    B. Funke

    2012-04-01

    Full Text Available Calculation of mean trace gas contributions from profiles obtained by retrievals of the logarithm of the abundance rather than retrievals of the abundance itself are prone to biases. By means of a system simulator, biases of linear versus logarithmic averaging were evaluated for both maximum likelihood and maximum a priori retrievals, for various signal to noise ratios and atmospheric variabilities. These biases can easily reach ten percent or more. As a rule of thumb we found for maximum likelihood retrievals that linear averaging better represents the true mean value in cases of large local natural variability and high signal to noise ratios, while for small local natural variability logarithmic averaging often is superior. In the case of maximum a posteriori retrievals, the mean is dominated by the a priori information used in the retrievals and the method of averaging is of minor concern. For larger natural variabilities, the appropriateness of the one or the other method of averaging depends on the particular case because the various biasing mechanisms partly compensate in an unpredictable manner. This complication arises mainly because of the fact that in logarithmic retrievals the weight of the prior information depends on abundance of the gas itself. No simple rule was found on which kind of averaging is superior, and instead of suggesting simple recipes we cannot do much more than to create awareness of the traps related with averaging of mixing ratios obtained from logarithmic retrievals.

  13. Revealing, Reducing, and Representing Uncertainties in New Hydrologic Projections for Climate-changed Futures

    Science.gov (United States)

    Arnold, Jeffrey; Clark, Martyn; Gutmann, Ethan; Wood, Andy; Nijssen, Bart; Rasmussen, Roy

    2016-04-01

    The United States Army Corps of Engineers (USACE) has had primary responsibility for multi-purpose water resource operations on most of the major river systems in the U.S. for more than 200 years. In that time, the USACE projects and programs making up those operations have proved mostly robust against the range of natural climate variability encountered over their operating life spans. However, in some watersheds and for some variables, climate change now is known to be shifting the hydroclimatic baseline around which that natural variability occurs and changing the range of that variability as well. This makes historical stationarity an inappropriate basis for assessing continued project operations under climate-changed futures. That means new hydroclimatic projections are required at multiple scales to inform decisions about specific threats and impacts, and for possible adaptation responses to limit water-resource vulnerabilities and enhance operational resilience. However, projections of possible future hydroclimatologies have myriad complex uncertainties that require explicit guidance for interpreting and using them to inform those decisions about climate vulnerabilities and resilience. Moreover, many of these uncertainties overlap and interact. Recent work, for example, has shown the importance of assessing the uncertainties from multiple sources including: global model structure [Meehl et al., 2005; Knutti and Sedlacek, 2013]; internal climate variability [Deser et al., 2012; Kay et al., 2014]; climate downscaling methods [Gutmann et al., 2012; Mearns et al., 2013]; and hydrologic models [Addor et al., 2014; Vano et al., 2014; Mendoza et al., 2015]. Revealing, reducing, and representing these uncertainties is essential for defining the plausible quantitative climate change narratives required to inform water-resource decision-making. And to be useful, such quantitative narratives, or storylines, of climate change threats and hydrologic impacts must sample

  14. 77 FR 24940 - Energy Conservation Program for Consumer Products: Representative Average Unit Costs of Energy

    Science.gov (United States)

    2012-04-26

    ... 5-year average ratio with heating oil prices published in the Monthly Energy Review, but the propane... DEPARTMENT OF ENERGY Office of Energy Efficiency and Renewable Energy Energy Conservation Program... and Renewable Energy, Department of Energy. ACTION: Notice. SUMMARY: In this notice, the U.S...

  15. 75 FR 13123 - Energy Conservation Program for Consumer Products: Representative Average Unit Costs of Energy

    Science.gov (United States)

    2010-03-18

    ... that of heating oil, based on the 2004-2008 averages for these two fuels. The source for these price... DEPARTMENT OF ENERGY Office of Energy Efficiency and Renewable Energy Energy Conservation Program... and Renewable Energy, Department of Energy. ACTION: Notice. SUMMARY: In this notice, the U.S...

  16. Perceived Average Orientation Reflects Effective Gist of the Surface.

    Science.gov (United States)

    Cha, Oakyoon; Chong, Sang Chul

    2018-03-01

    The human ability to represent ensemble visual information, such as average orientation and size, has been suggested as the foundation of gist perception. To effectively summarize different groups of objects into the gist of a scene, observers should form ensembles separately for different groups, even when objects have similar visual features across groups. We hypothesized that the visual system utilizes perceptual groups characterized by spatial configuration and represents separate ensembles for different groups. Therefore, participants could not integrate ensembles of different perceptual groups on a task basis. We asked participants to determine the average orientation of visual elements comprising a surface with a contour situated inside. Although participants were asked to estimate the average orientation of all the elements, they ignored orientation signals embedded in the contour. This constraint may help the visual system to keep the visual features of occluding objects separate from those of the occluded objects.

  17. Gibbs equilibrium averages and Bogolyubov measure

    International Nuclear Information System (INIS)

    Sankovich, D.P.

    2011-01-01

    Application of the functional integration methods in equilibrium statistical mechanics of quantum Bose-systems is considered. We show that Gibbs equilibrium averages of Bose-operators can be represented as path integrals over a special Gauss measure defined in the corresponding space of continuous functions. We consider some problems related to integration with respect to this measure

  18. Representing Participation in ICT4D Projects

    DEFF Research Database (Denmark)

    Singh, J. P.; Flyverbom, Mikkel

    2016-01-01

    How do the discourses of participation inform deployment of information and communication technologies for development (ICT4D)? Discourses here mean narratives that assign roles to actors, and specify causes and outcomes for events. Based on the theory and practice of international development we......, depending on the context of their implementation, are permeated by multiple discourses about participation. Our four ideal types of participation discourses are, therefore, useful starting points to discuss the intricate dynamics of participation in ICT4D projects....

  19. Evaluation of Representative Smart Grid Investment Project Technologies: Demand Response

    Energy Technology Data Exchange (ETDEWEB)

    Fuller, Jason C.; Prakash Kumar, Nirupama; Bonebrake, Christopher A.

    2012-02-14

    This document is one of a series of reports estimating the benefits of deploying technologies similar to those implemented on the Smart Grid Investment Grant (SGIG) projects. Four technical reports cover the various types of technologies deployed in the SGIG projects, distribution automation, demand response, energy storage, and renewables integration. A fifth report in the series examines the benefits of deploying these technologies on a national level. This technical report examines the impacts of a limited number of demand response technologies and implementations deployed in the SGIG projects.

  20. Domain-averaged Fermi-hole Analysis for Solids

    Czech Academy of Sciences Publication Activity Database

    Baranov, A.; Ponec, Robert; Kohout, M.

    2012-01-01

    Roč. 137, č. 21 (2012), s. 214109 ISSN 0021-9606 R&D Projects: GA ČR GA203/09/0118 Institutional support: RVO:67985858 Keywords : bonding in solids * domain averaged fermi hole * natural orbitals Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 3.164, year: 2012

  1. Attractiveness of the female body: Preference for the average or the supernormal?

    Directory of Open Access Journals (Sweden)

    Marković Slobodan

    2017-01-01

    Full Text Available The main purpose of the present study was to contrast the two hypotheses of female body attractiveness. The first is the “preference-for-the average” hypothesis: the most attractive female body is the one that represents the average body proportions for a given population. The second is the “preference-for-the supernormal” hypothesis: according to the so-called “peak shift effect”, the most attractive female body is more feminine than the average. We investigated the preference for three female body characteristics: waist to hip ratio (WHR, buttocks and breasts. There were 456 participants of both genders. Using a program for computer animation (DAZ 3D three sets of stimuli were generated (WHR, buttocks and breasts. Each set included six stimuli ranked from the lowest to the highest femininity level. Participants were asked to choose the stimulus within each set which they found most attractive (task 1 and average (task 2. One group of participants judged the body parts that were presented in the global context (whole body, while the other group judged the stimuli in the local context (isolated body parts only. Analyses have shown that the most attractive WHR, buttocks and breasts are more feminine (meaning smaller for WHR and larger for breasts and buttocks than average ones, for both genders and in both presentation contexts. The effect of gender was obtained only for the most attractive breasts: males prefer larger breasts than females. Finally, most attractive and average WHR and breasts were less feminine in the local than in the global context. These results support the preference-for the supernormal hypothesis: all analyses have shown that both male and female participants preferred female body parts which are more feminine than those judged average. [Project of the Serbian Ministry of Education, Science and Technological Development, Grant no. 179033

  2. Average is Over

    Science.gov (United States)

    Eliazar, Iddo

    2018-02-01

    The popular perception of statistical distributions is depicted by the iconic bell curve which comprises of a massive bulk of 'middle-class' values, and two thin tails - one of small left-wing values, and one of large right-wing values. The shape of the bell curve is unimodal, and its peak represents both the mode and the mean. Thomas Friedman, the famous New York Times columnist, recently asserted that we have entered a human era in which "Average is Over" . In this paper we present mathematical models for the phenomenon that Friedman highlighted. While the models are derived via different modeling approaches, they share a common foundation. Inherent tipping points cause the models to phase-shift from a 'normal' bell-shape statistical behavior to an 'anomalous' statistical behavior: the unimodal shape changes to an unbounded monotone shape, the mode vanishes, and the mean diverges. Hence: (i) there is an explosion of small values; (ii) large values become super-large; (iii) 'middle-class' values are wiped out, leaving an infinite rift between the small and the super large values; and (iv) "Average is Over" indeed.

  3. On spectral averages in nuclear spectroscopy

    International Nuclear Information System (INIS)

    Verbaarschot, J.J.M.

    1982-01-01

    In nuclear spectroscopy one tries to obtain a description of systems of bound nucleons. By means of theoretical models one attemps to reproduce the eigenenergies and the corresponding wave functions which then enable the computation of, for example, the electromagnetic moments and the transition amplitudes. Statistical spectroscopy can be used for studying nuclear systems in large model spaces. In this thesis, methods are developed and applied which enable the determination of quantities in a finite part of the Hilbert space, which is defined by specific quantum values. In the case of averages in a space defined by a partition of the nucleons over the single-particle orbits, the propagation coefficients reduce to Legendre interpolation polynomials. In chapter 1 these polynomials are derived with the help of a generating function and a generalization of Wick's theorem. One can then deduce the centroid and the variance of the eigenvalue distribution in a straightforward way. The results are used to calculate the systematic energy difference between states of even and odd parity for nuclei in the mass region A=10-40. In chapter 2 an efficient method for transforming fixed angular momentum projection traces into fixed angular momentum for the configuration space traces is developed. In chapter 3 it is shown that the secular behaviour can be represented by a Gaussian function of the energies. (Auth.)

  4. GIS Tools to Estimate Average Annual Daily Traffic

    Science.gov (United States)

    2012-06-01

    This project presents five tools that were created for a geographical information system to estimate Annual Average Daily : Traffic using linear regression. Three of the tools can be used to prepare spatial data for linear regression. One tool can be...

  5. North Atlantic observations sharpen meridional overturning projections

    Science.gov (United States)

    Olson, R.; An, S.-I.; Fan, Y.; Evans, J. P.; Caesar, L.

    2018-06-01

    Atlantic Meridional Overturning Circulation (AMOC) projections are uncertain due to both model errors, as well as internal climate variability. An AMOC slowdown projected by many climate models is likely to have considerable effects on many aspects of global and North Atlantic climate. Previous studies to make probabilistic AMOC projections have broken new ground. However, they do not drift-correct or cross-validate the projections, and do not fully account for internal variability. Furthermore, they consider a limited subset of models, and ignore the skill of models at representing the temporal North Atlantic dynamics. We improve on previous work by applying Bayesian Model Averaging to weight 13 Coupled Model Intercomparison Project phase 5 models by their skill at modeling the AMOC strength, and its temporal dynamics, as approximated by the northern North-Atlantic temperature-based AMOC Index. We make drift-corrected projections accounting for structural model errors, and for the internal variability. Cross-validation experiments give approximately correct empirical coverage probabilities, which validates our method. Our results present more evidence that AMOC likely already started slowing down. While weighting considerably moderates and sharpens our projections, our results are at low end of previously published estimates. We project mean AMOC changes between periods 1960-1999 and 2060-2099 of -4.0 Sv and -6.8 Sv for RCP4.5 and RCP8.5 emissions scenarios respectively. The corresponding average 90% credible intervals for our weighted experiments are [-7.2, -1.2] and [-10.5, -3.7] Sv respectively for the two scenarios.

  6. Grade Point Average: Report of the GPA Pilot Project 2013-14

    Science.gov (United States)

    Higher Education Academy, 2015

    2015-01-01

    This report is published as the result of a range of investigations and debates involving many universities and colleges and a series of meetings, presentations, discussions and consultations. Interest in a grade point average (GPA) system was originally initiated by a group of interested universities, progressing to the systematic investigation…

  7. Evaluation of Representative Smart Grid Investment Grant Project Technologies: Distributed Generation

    Energy Technology Data Exchange (ETDEWEB)

    Singh, Ruchi; Vyakaranam, Bharat GNVSR

    2012-02-14

    This document is one of a series of reports estimating the benefits of deploying technologies similar to those implemented on the Smart Grid Investment Grant (SGIG) projects. Four technical reports cover the various types of technologies deployed in the SGIG projects, distribution automation, demand response, energy storage, and renewables integration. A fifth report in the series examines the benefits of deploying these technologies on a national level. This technical report examines the impacts of addition of renewable resources- solar and wind in the distribution system as deployed in the SGIG projects.

  8. Lateral dispersion coefficients as functions of averaging time

    International Nuclear Information System (INIS)

    Sheih, C.M.

    1980-01-01

    Plume dispersion coefficients are discussed in terms of single-particle and relative diffusion, and are investigated as functions of averaging time. To demonstrate the effects of averaging time on the relative importance of various dispersion processes, and observed lateral wind velocity spectrum is used to compute the lateral dispersion coefficients of total, single-particle and relative diffusion for various averaging times and plume travel times. The results indicate that for a 1 h averaging time the dispersion coefficient of a plume can be approximated by single-particle diffusion alone for travel times <250 s and by relative diffusion for longer travel times. Furthermore, it is shown that the power-law formula suggested by Turner for relating pollutant concentrations for other averaging times to the corresponding 15 min average is applicable to the present example only when the averaging time is less than 200 s and the tral time smaller than about 300 s. Since the turbulence spectrum used in the analysis is an observed one, it is hoped that the results could represent many conditions encountered in the atmosphere. However, as the results depend on the form of turbulence spectrum, the calculations are not for deriving a set of specific criteria but for demonstrating the need in discriminating various processes in studies of plume dispersion

  9. Weighted estimates for the averaging integral operator

    Czech Academy of Sciences Publication Activity Database

    Opic, Bohumír; Rákosník, Jiří

    2010-01-01

    Roč. 61, č. 3 (2010), s. 253-262 ISSN 0010-0757 R&D Projects: GA ČR GA201/05/2033; GA ČR GA201/08/0383 Institutional research plan: CEZ:AV0Z10190503 Keywords : averaging integral operator * weighted Lebesgue spaces * weights Subject RIV: BA - General Mathematics Impact factor: 0.474, year: 2010 http://link.springer.com/article/10.1007%2FBF03191231

  10. Unscrambling The "Average User" Of Habbo Hotel

    Directory of Open Access Journals (Sweden)

    Mikael Johnson

    2007-01-01

    Full Text Available The “user” is an ambiguous concept in human-computer interaction and information systems. Analyses of users as social actors, participants, or configured users delineate approaches to studying design-use relationships. Here, a developer’s reference to a figure of speech, termed the “average user,” is contrasted with design guidelines. The aim is to create an understanding about categorization practices in design through a case study about the virtual community, Habbo Hotel. A qualitative analysis highlighted not only the meaning of the “average user,” but also the work that both the developer and the category contribute to this meaning. The average user a represents the unknown, b influences the boundaries of the target user groups, c legitimizes the designer to disregard marginal user feedback, and d keeps the design space open, thus allowing for creativity. The analysis shows how design and use are intertwined and highlights the developers’ role in governing different users’ interests.

  11. Research & development and growth: A Bayesian model averaging analysis

    Czech Academy of Sciences Publication Activity Database

    Horváth, Roman

    2011-01-01

    Roč. 28, č. 6 (2011), s. 2669-2673 ISSN 0264-9993. [Society for Non-linear Dynamics and Econometrics Annual Conferencen. Washington DC, 16.03.2011-18.03.2011] R&D Projects: GA ČR GA402/09/0965 Institutional research plan: CEZ:AV0Z10750506 Keywords : Research and development * Growth * Bayesian model averaging Subject RIV: AH - Economic s Impact factor: 0.701, year: 2011 http://library.utia.cas.cz/separaty/2011/E/horvath-research & development and growth a bayesian model averaging analysis.pdf

  12. Implications of Methodist clergies' average lifespan and missional ...

    African Journals Online (AJOL)

    2015-06-09

    Jun 9, 2015 ... The author of Genesis 5 paid meticulous attention to the lifespan of several people ... of Southern Africa (MCSA), and to argue that memories of the ... average ages at death were added up and the sum was divided by 12 (which represents the 12 ..... not explicit in how the departed Methodist ministers were.

  13. Recent activities of the Seismology Division Early Career Representative(s)

    Science.gov (United States)

    Agius, Matthew; Van Noten, Koen; Ermert, Laura; Mai, P. Martin; Krawczyk, CharLotte

    2016-04-01

    The European Geosciences Union is a bottom-up-organisation, in which its members are represented by their respective scientific divisions, committees and council. In recent years, EGU has embarked on a mission to reach out for its numerous 'younger' members by giving awards to outstanding young scientists and the setting up of Early Career Scientists (ECS) representatives. The division representative's role is to engage in discussions that concern students and early career scientists. Several meetings between all the division representatives are held throughout the year to discuss ideas and Union-wide issues. One important impact ECS representatives have had on EGU is the increased number of short courses and workshops run by ECS during the annual General Assembly. Another important contribution of ECS representatives was redefining 'Young Scientist' to 'Early Career Scientist', which avoids discrimination due to age. Since 2014, the Seismology Division has its own ECS representative. In an effort to more effectively reach out for young seismologists, a blog and a social media page dedicated to seismology have been set up online. With this dedicated blog, we'd like to give more depth to the average browsing experience by enabling young researchers to explore various seismology topics in one place while making the field more exciting and accessible to the broader community. These pages are used to promote the latest research especially of young seismologists and to share interesting seismo-news. Over the months the pages proved to be popular, with hundreds of views every week and an increased number of followers. An online survey was conducted to learn more about the activities and needs of early career seismologists. We present the results from this survey, and the work that has been carried out over the last two years, including detail of what has been achieved so far, and what we would like the ECS representation for Seismology to achieve. Young seismologists are

  14. Approximations for transport parameters and self-averaging properties for point-like injections in heterogeneous media

    International Nuclear Information System (INIS)

    Eberhard, Jens

    2004-01-01

    We focus on transport parameters in heterogeneous media with a flow modelled by an ensemble of periodic and Gaussian random fields. The parameters are determined by ensemble averages. We study to what extent these averages represent the behaviour in a single realization. We calculate the centre-of-mass velocity and the dispersion coefficient using approximations based on a perturbative expansion for the transport equation, and on the iterative solution of the Langevin equation. Compared with simulations, the perturbation theory reproduces the numerical results only poorly, whereas the iterative solution yields good results. Using these approximations, we investigate the self-averaging properties. The ensemble average of the velocity characterizes the behaviour of a realization for large times in both ensembles. The dispersion coefficient is not self-averaging in the ensemble of periodic fields. For the Gaussian ensemble the asymptotic dispersion coefficient is self-averaging. For finite times, however, the fluctuations are so large that the average does not represent the behaviour in a single realization

  15. Low-level radioactive waste in the northeast: revised waste volume projections

    International Nuclear Information System (INIS)

    1984-06-01

    The volume of low-level radioactive waste generated in the eleven Northeast states has undergone significant change since the inital 1982 analysis and projection. These revised projections incorporate improved data reporting and evidence of sharp declines in certain categories of waste. Volumes in the 1982-1983 period reflect waste shipped for disposal as reported by disposal site operators. Projected waste volumes represent waste intended for disposal. The recent dramatic changes in source reduction and waste management practices underscore the need for annual review of waste volume projections. The volume of waste shipped for off-site disposal has declined approximately 12% in two years, from an average 1,092,500 ft 3 annually in 1979 to 1981 to an average annual 956,500 ft 3 in 1982 to 1983; reactor waste disposal volumes declined by about 39,000 ft 3 or 7% during this period. Non-reactor waste volumes shipped for disposal declined by over 70,000 ft 3 or 15% during this period. The data suggest that generators increased their use of such management practices as source reduction, compaction, or, for carbon-14 and tritium, temporary storage followed by disposal as non-radioactive waste under the NRC de minimus standard effective March 1981. Using the Technical Subcommittee projection methodology, the volume of low-level waste produced annually in the eleven states, individually and collectively, is expected to increase through the year 2000, but at a significantly lower rate of increase than initially projected. By the year 2000, the Northeast is projected to generate 1,137,600 ft 3 of waste annually, an increase of about 20% over 1982 to 1983 average volume

  16. "Intelligent Ensemble" Projections of Precipitation and Surface Radiation in Support of Agricultural Climate Change Adaptation

    Science.gov (United States)

    Taylor, Patrick C.; Baker, Noel C.

    2015-01-01

    Earth's climate is changing and will continue to change into the foreseeable future. Expected changes in the climatological distribution of precipitation, surface temperature, and surface solar radiation will significantly impact agriculture. Adaptation strategies are, therefore, required to reduce the agricultural impacts of climate change. Climate change projections of precipitation, surface temperature, and surface solar radiation distributions are necessary input for adaption planning studies. These projections are conventionally constructed from an ensemble of climate model simulations (e.g., the Coupled Model Intercomparison Project 5 (CMIP5)) as an equal weighted average, one model one vote. Each climate model, however, represents the array of climate-relevant physical processes with varying degrees of fidelity influencing the projection of individual climate variables differently. Presented here is a new approach, termed the "Intelligent Ensemble, that constructs climate variable projections by weighting each model according to its ability to represent key physical processes, e.g., precipitation probability distribution. This approach provides added value over the equal weighted average method. Physical process metrics applied in the "Intelligent Ensemble" method are created using a combination of NASA and NOAA satellite and surface-based cloud, radiation, temperature, and precipitation data sets. The "Intelligent Ensemble" method is applied to the RCP4.5 and RCP8.5 anthropogenic climate forcing simulations within the CMIP5 archive to develop a set of climate change scenarios for precipitation, temperature, and surface solar radiation in each USDA Farm Resource Region for use in climate change adaptation studies.

  17. National Health Expenditure Projections, 2017-26: Despite Uncertainty, Fundamentals Primarily Drive Spending Growth.

    Science.gov (United States)

    Cuckler, Gigi A; Sisko, Andrea M; Poisal, John A; Keehan, Sean P; Smith, Sheila D; Madison, Andrew J; Wolfe, Christian J; Hardesty, James C

    2018-03-01

    Under current law, national health spending is projected to grow 5.5 percent annually on average in 2017-26 and to represent 19.7 percent of the economy in 2026. Projected national health spending and enrollment growth over the next decade is largely driven by fundamental economic and demographic factors: changes in projected income growth, increases in prices for medical goods and services, and enrollment shifts from private health insurance to Medicare that are related to the aging of the population. The recent enactment of tax legislation that eliminated the individual mandate is expected to result in only a small reduction to insurance coverage trends.

  18. Evaluation of Representative Smart Grid Investment Grant Project Technologies: Thermal Energy Storage

    Energy Technology Data Exchange (ETDEWEB)

    Tuffner, Francis K.; Bonebrake, Christopher A.

    2012-02-14

    This document is one of a series of reports estimating the benefits of deploying technologies similar to those implemented on the Smart Grid Investment Grant (SGIG) projects. Four technical reports cover the various types of technologies deployed in the SGIG projects, distribution automation, demand response, energy storage, and renewables integration. A fifth report in the series examines the benefits of deploying these technologies on a national level. This technical report examines the impacts of energy storage technologies deployed in the SGIG projects.

  19. Yucca Mountain Climate Technical Support Representative

    International Nuclear Information System (INIS)

    Sharpe, Saxon E

    2007-01-01

    The primary objective of Project Activity ORD-FY04-012, 'Yucca Mountain Climate Technical Support Representative', was to provide the Office of Civilian Radioactive Waste Management (OCRWM) with expertise on past, present, and future climate scenarios and to support the technical elements of the Yucca Mountain Project (YMP) climate program. The Climate Technical Support Representative was to explain, defend, and interpret the YMP climate program to the various audiences during Site Recommendation and License Application. This technical support representative was to support DOE management in the preparation and review of documents, and to participate in comment response for the Final Environmental Impact Statement, the Site Recommendation Hearings, the NRC Sufficiency Comments, and other forums as designated by DOE management. Because the activity was terminated 12 months early and experience a 27% reduction in budget, it was not possible to complete all components of the tasks as originally envisioned. Activities not completed include the qualification of climate datasets and the production of a qualified technical report. The following final report is an unqualified summary of the activities that were completed given the reduced time and funding

  20. Characteristics of phase-averaged equations for modulated wave groups

    NARCIS (Netherlands)

    Klopman, G.; Petit, H.A.H.; Battjes, J.A.

    2000-01-01

    The project concerns the influence of long waves on coastal morphology. The modelling of the combined motion of the long waves and short waves in the horizontal plane is done by phase-averaging over the short wave motion and using intra-wave modelling for the long waves, see e.g. Roelvink (1993).

  1. Project-Based Emissions Trading. The Impact of Institutional Arrangements on Cost-Effectiveness

    International Nuclear Information System (INIS)

    Woerdman, E.; Van der Gaast, W.

    2001-01-01

    In this paper we demonstrate that the institutional arrangement (or: design) of Joint Implementation (JI) and the Clean Development Mechanism (CDM) has a decisive impact on their cost-effectiveness. We illustrate our arguments by statistically analyzing the costs from 94 Activities Implemented Jointly (AIJ) pilot phase projects as well as by adjusting these data on the basis of simple mathematical formulas. These calculations explicitly take into account the institutional differences between JI (sinks, no banking) and the CDM (banking, no sinks) under the Kyoto Protocol and also show the possible effects on credit costs of alternative design options. However, our numerical illustrations should be viewed with caution, because AIJ is only to a limited extent representative of potential future JI and CDM projects and because credit costs are not credit prices. Some of the main figures found in this study are: an average cost figure per unit of emission reduction for AIJ projects of 46 dollar per ton of carbon dioxide equivalent ($/Mg CO 2 -eq), an average potential JI credit cost figure which is lowered to 37 $/Mg CO 2 -eq by introducing banking and an average of 6 $/Mg CO 2 -eq per credit for potential low-cost CDM projects which includes sinks. However, at CoP6 in November 2000 in The Hague (The Netherlands), the Parties to the Framework Convention on Climate Change (FCCC) did not (yet) reach consensus on the institutional details of the project-based mechanisms, such as the possible arrangement of early JI action or the inclusion of sinks under the CDM. 55 refs

  2. Size and emotion averaging: costs of dividing attention after all.

    Science.gov (United States)

    Brand, John; Oriet, Chris; Tottenham, Laurie Sykes

    2012-03-01

    Perceptual averaging is a process by which sets of similar items are represented by summary statistics such as their average size, luminance, or orientation. Researchers have argued that this process is automatic, able to be carried out without interference from concurrent processing. Here, we challenge this conclusion and demonstrate a reliable cost of computing the mean size of circles distinguished by colour (Experiments 1 and 2) and the mean emotionality of faces distinguished by sex (Experiment 3). We also test the viability of two strategies that could have allowed observers to guess the correct response without computing the average size or emotionality of both sets concurrently. We conclude that although two means can be computed concurrently, doing so incurs a cost of dividing attention.

  3. [Project financing in public hospital trusts].

    Science.gov (United States)

    Contarino, F; Grosso, G; Mistretta, A

    2009-01-01

    The growing debate in recent years over how to finance public works through private capital has progressively highlighted the role of project finance (PF) and publicprivate partnerships (PPP) in general. More and more European countries are turning to PF to finance their public infrastructure development. The UK, which pioneered the adoption of project finance in this field, has been followed by Italy, Spain, France, Portugal and Germany and more recently by Greece, Czech Republic and Poland. Beginning in the late 1990's, Italy has steadily amplified its use of PF and PPPs in key sectors such as healthcare as an alternative way of funding the modernisation of its health facilities and hospitals. The trend reveal an average annual growth of 10.9% since 2002 with peaks of varying intensity over the five year period. Project finance and PPPs represent an effective response to the country's infrastructure gap and support the competitiveness of local systems and the quality of public services. None of this will transpire, however without energetic new planning efforts and adequate policy at the centre.

  4. A One-Electron Approximation to Domain Averaged Fermi hole Analysis

    Czech Academy of Sciences Publication Activity Database

    Cooper, D.L.; Ponec, Robert

    2008-01-01

    Roč. 10, č. 9 (2008), s. 1319-1329 ISSN 1463-9076 R&D Projects: GA AV ČR(CZ) IAA4072403 Institutional research plan: CEZ:AV0Z40720504 Keywords : domain-averaged fermi hole * comparisons Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 4.064, year: 2008

  5. Average configuration of the geomagnetic tail

    International Nuclear Information System (INIS)

    Fairfield, D.H.

    1979-01-01

    Over 3000 hours of Imp 6 magnetic field data obtained between 20 and 33 R/sub E/ in the geomagnetic tail have been used in a statistical study of the tail configuration. A distribution of 2.5-min averages of B/sub z/ as a function of position across the tail reveals that more flux crosses the equatorial plane near the dawn and dusk flanks (B-bar/sub z/=3.γ) than near midnight (B-bar/sub z/=1.8γ). The tail field projected in the solar magnetospheric equatorial plane deviates from the x axis due to flaring and solar wind aberration by an angle α=-0.9 Y/sub SM/-2.7, where Y/sub SM/ is in earth radii and α is in degrees. After removing these effects, the B/sub y/ component of the tail field is found to depend on interplanetary sector structure. During an 'away' sector the B/sub y/ component of the tail field is on average 0.5γ greater than that during a 'toward' sector, a result that is true in both tail lobes and is independent of location across the tail. This effect means the average field reversal between northern and southern lobes of the tail is more often 178 0 rather than the 180 0 that is generally supposed

  6. The performance indicators of model projects. A special evaluation

    International Nuclear Information System (INIS)

    1995-11-01

    As a result of the acknowledgment of the key role of the Model Project concept in the Agency's Technical Co-operation Programme, the present review of the objectives of the model projects which are now in operation, was undertaken, as recommended by the Board of Governors, to determine at an early stage: the extent to which the present objectives have been defined in a measurable way; whether objectively verifiable performance indicators and success criteria had been identified for each project; whether mechanisms to obtain feedback on the achievements had been foreseen. The overall budget for the 23 model projects, as approved from 1994 to 1998, amounts to $32,557,560, of which 45% is funded by Technical Co-operation Fund. This represents an average investment of about $8 million per year, that is over 15% of the annual TC budget. The conceptual importance of the Model Project initiative, as well as the significant funds allocated to them, led the Secretariat to plan the methods to be used to determine their socio-economic impact. 1 tab

  7. Modeling methane emission via the infinite moving average process

    Czech Academy of Sciences Publication Activity Database

    Jordanova, D.; Dušek, Jiří; Stehlík, M.

    2013-01-01

    Roč. 122, - (2013), s. 40-49 ISSN 0169-7439 R&D Projects: GA MŠk(CZ) ED1.1.00/02.0073; GA ČR(CZ) GAP504/11/1151 Institutional support: RVO:67179843 Keywords : Environmental chemistry * Pareto tails * t-Hill estimator * Weak consistency * Moving average process * Methane emission model Subject RIV: EH - Ecology, Behaviour Impact factor: 2.381, year: 2013

  8. Microscopic description of average level spacing in even-even nuclei

    International Nuclear Information System (INIS)

    Huong, Le Thi Quynh; Hung, Nguyen Quang; Phuc, Le Tan

    2017-01-01

    A microscopic theoretical approach to the average level spacing at the neutron binding energy in even-even nuclei is proposed. The approach is derived based on the Bardeen-Cooper-Schrieffer (BCS) theory at finite temperature and projection M of the total angular momentum J , which is often used to describe the superfluid properties of hot rotating nuclei. The exact relation of the J -dependent total level density to the M -dependent state densities, based on which the average level spacing is calculated, was employed. The numerical calculations carried out for several even-even nuclei have shown that in order to reproduce the experimental average level spacing, the M -dependent pairing gaps as well as the exact relation of the J -dependent total level density formula should be simultaneously used. (paper)

  9. Downscaled projections of Caribbean coral bleaching that can inform conservation planning.

    Science.gov (United States)

    van Hooidonk, Ruben; Maynard, Jeffrey Allen; Liu, Yanyun; Lee, Sang-Ki

    2015-09-01

    Projections of climate change impacts on coral reefs produced at the coarse resolution (~1°) of Global Climate Models (GCMs) have informed debate but have not helped target local management actions. Here, projections of the onset of annual coral bleaching conditions in the Caribbean under Representative Concentration Pathway (RCP) 8.5 are produced using an ensemble of 33 Coupled Model Intercomparison Project phase-5 models and via dynamical and statistical downscaling. A high-resolution (~11 km) regional ocean model (MOM4.1) is used for the dynamical downscaling. For statistical downscaling, sea surface temperature (SST) means and annual cycles in all the GCMs are replaced with observed data from the ~4-km NOAA Pathfinder SST dataset. Spatial patterns in all three projections are broadly similar; the average year for the onset of annual severe bleaching is 2040-2043 for all projections. However, downscaled projections show many locations where the onset of annual severe bleaching (ASB) varies 10 or more years within a single GCM grid cell. Managers in locations where this applies (e.g., Florida, Turks and Caicos, Puerto Rico, and the Dominican Republic, among others) can identify locations that represent relative albeit temporary refugia. Both downscaled projections are different for the Bahamas compared to the GCM projections. The dynamically downscaled projections suggest an earlier onset of ASB linked to projected changes in regional currents, a feature not resolved in GCMs. This result demonstrates the value of dynamical downscaling for this application and means statistically downscaled projections have to be interpreted with caution. However, aside from west of Andros Island, the projections for the two types of downscaling are mostly aligned; projected onset of ASB is within ±10 years for 72% of the reef locations. © 2015 The Authors. Global Change Biology Published by John Wiley & Sons Ltd.

  10. State Averages

    Data.gov (United States)

    U.S. Department of Health & Human Services — A list of a variety of averages for each state or territory as well as the national average, including each quality measure, staffing, fine amount and number of...

  11. [Administration of the "Healthy School" project].

    Science.gov (United States)

    Bjegović, V; Zivković, M; Marinković, J; Vuković, D; Legetić, B

    1996-01-01

    , the level of personal influence on different aspects of Project development and overall work motivation and satisfaction of all participants. (c) The outcomes of the given management attempts were analyzed by the following output variables: the number of different types of meetings held, the number of seminars, mass media presentation and articles, the amount of money raised and the number of questionnaires administered. Triangular method was used to gather the data: (1) direct observation, (2) four types of questionnaires and (3) project reports and documentation. Four types of specially designed questionnaires were used to examine four groups of participants (Project Coordinators, School Project Managers, Directors and Project Co-operators). The questionnaires were different in the questions concerning examinees' project tasks and types of external communication, while the questions referring to personal characteristics, general features of the project (goals, common jobs, participation in decision making, motivation and satisfaction) were the same for all groups. The average age of the project participants was 45.50 ranging from 25 to 60 years of age. The oldest group was the group of School Directors, while the youngest were School Co-operators. The project has been run mostly by women, while men were predominantly represented in the group of School Directors. The teaching occupation is presented by 61.8%, the rest being health professionals, mostly of preventive orientation. The analysis and classification of participants goals verify that the personal goals of all participants correspond with the main Project goals. Certain groups have also some additional motives which support their successful and affective movement towards the overall Project goals. The largest problem in all groups appear to be in the field of financing the Project activities (Figure 1). (ABSTRACT TRUNCATED)

  12. Twenty-sixth general conference meeting of representatives of RCA member states. Report

    International Nuclear Information System (INIS)

    1997-10-01

    The meeting discussed the following issues: RCA Annual Report 1996; Report of the 19th RCA Working Group Meeting, Yangon Myanmar, 10-14 March 1997; RCA in the next 25 years; Development finance and the RCA; locating a senior RCA representative in the region; RCA guidelines and operating rules; TCDC activities; terminal tripartite review meeting on UNDP/RCA/IAEA project; status of new joint UNDP/RCA/IAEA project; options for the new joint project; RCA activities and budget in 1997 and 1998; reformulation and health care project; RCA proposed programme for 1999-2000; country statements; extension of RCA agreement; RCA 25th anniversary; venue and date of the 20th regular meeting of national RCA representatives

  13. Twenty-sixth general conference meeting of representatives of RCA member states. Report

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1997-10-01

    The meeting discussed the following issues: RCA Annual Report 1996; Report of the 19th RCA Working Group Meeting, Yangon Myanmar, 10-14 March 1997; RCA in the next 25 years; Development finance and the RCA; locating a senior RCA representative in the region; RCA guidelines and operating rules; TCDC activities; terminal tripartite review meeting on UNDP/RCA/IAEA project; status of new joint UNDP/RCA/IAEA project; options for the new joint project; RCA activities and budget in 1997 and 1998; reformulation and health care project; RCA proposed programme for 1999-2000; country statements; extension of RCA agreement; RCA 25th anniversary; venue and date of the 20th regular meeting of national RCA representatives.

  14. Collaborating with consumer and community representatives in health and medical research in Australia: results from an evaluation

    Directory of Open Access Journals (Sweden)

    Bartu Anne E

    2011-05-01

    Full Text Available Abstract Objective To collaborate with consumer and community representatives in the Alcohol and Pregnancy Project from 2006-2008 http://www.ichr.uwa.edu.au/alcoholandpregnancy and evaluate researchers' and consumer and community representatives' perceptions of the process, context and impact of consumer and community participation in the project. Methods We formed two reference groups and sought consumer and community representatives' perspectives on all aspects of the project over a three year period. We developed an evaluation framework and asked consumer and community representatives and researchers to complete a self-administered questionnaire at the end of the project. Results Fifteen researchers (93.8% and seven (53.8% consumer and community representatives completed a questionnaire. Most consumer and community representatives agreed that the process and context measures of their participation had been achieved. Both researchers and consumer and community representatives identified areas for improvement and offered suggestions how these could be improved for future research. Researchers thought consumer and community participation contributed to project outputs and outcomes by enhancing scientific and ethical standards, providing legitimacy and authority, and increasing the project's credibility and participation. They saw it was fundamental to the research process and acknowledged consumer and community representatives for their excellent contribution. Consumer and community representatives were able to directly influence decisions about the research. They thought that consumer and community participation had significant influence on the success of project outputs and outcomes. Conclusions Consumer and community participation is an essential component of good research practice and contributed to the Alcohol and Pregnancy Project by enhancing research processes, outputs and outcomes, and this participation was valued by community and

  15. Collaborating with consumer and community representatives in health and medical research in Australia: results from an evaluation.

    Science.gov (United States)

    Payne, Janet M; D'Antoine, Heather A; France, Kathryn E; McKenzie, Anne E; Henley, Nadine; Bartu, Anne E; Elliott, Elizabeth J; Bower, Carol

    2011-05-14

    To collaborate with consumer and community representatives in the Alcohol and Pregnancy Project from 2006-2008 http://www.ichr.uwa.edu.au/alcoholandpregnancy and evaluate researchers' and consumer and community representatives' perceptions of the process, context and impact of consumer and community participation in the project. We formed two reference groups and sought consumer and community representatives' perspectives on all aspects of the project over a three year period. We developed an evaluation framework and asked consumer and community representatives and researchers to complete a self-administered questionnaire at the end of the project. Fifteen researchers (93.8%) and seven (53.8%) consumer and community representatives completed a questionnaire. Most consumer and community representatives agreed that the process and context measures of their participation had been achieved. Both researchers and consumer and community representatives identified areas for improvement and offered suggestions how these could be improved for future research. Researchers thought consumer and community participation contributed to project outputs and outcomes by enhancing scientific and ethical standards, providing legitimacy and authority, and increasing the project's credibility and participation. They saw it was fundamental to the research process and acknowledged consumer and community representatives for their excellent contribution. Consumer and community representatives were able to directly influence decisions about the research. They thought that consumer and community participation had significant influence on the success of project outputs and outcomes. Consumer and community participation is an essential component of good research practice and contributed to the Alcohol and Pregnancy Project by enhancing research processes, outputs and outcomes, and this participation was valued by community and consumer representatives and researchers. The National Health and

  16. Impact of connected vehicle guidance information on network-wide average travel time

    Directory of Open Access Journals (Sweden)

    Jiangfeng Wang

    2016-12-01

    Full Text Available With the emergence of connected vehicle technologies, the potential positive impact of connected vehicle guidance on mobility has become a research hotspot by data exchange among vehicles, infrastructure, and mobile devices. This study is focused on micro-modeling and quantitatively evaluating the impact of connected vehicle guidance on network-wide travel time by introducing various affecting factors. To evaluate the benefits of connected vehicle guidance, a simulation architecture based on one engine is proposed representing the connected vehicle–enabled virtual world, and connected vehicle route guidance scenario is established through the development of communication agent and intelligent transportation systems agents using connected vehicle application programming interface considering the communication properties, such as path loss and transmission power. The impact of connected vehicle guidance on network-wide travel time is analyzed by comparing with non-connected vehicle guidance in response to different market penetration rate, following rate, and congestion level. The simulation results explore that average network-wide travel time in connected vehicle guidance shows a significant reduction versus that in non–connected vehicle guidance. Average network-wide travel time in connected vehicle guidance have an increase of 42.23% comparing to that in non-connected vehicle guidance, and average travel time variability (represented by the coefficient of variance increases as the travel time increases. Other vital findings include that higher penetration rate and following rate generate bigger savings of average network-wide travel time. The savings of average network-wide travel time increase from 17% to 38% according to different congestion levels, and savings of average travel time in more serious congestion have a more obvious improvement for the same penetration rate or following rate.

  17. Demonstration of two-phase Direct Numerical Simulation (DNS) methods potentiality to give information to averaged models: application to bubbles column

    International Nuclear Information System (INIS)

    Magdeleine, S.

    2009-11-01

    This work is a part of a long term project that aims at using two-phase Direct Numerical Simulation (DNS) in order to give information to averaged models. For now, it is limited to isothermal bubbly flows with no phase change. It could be subdivided in two parts: Firstly, theoretical developments are made in order to build an equivalent of Large Eddy Simulation (LES) for two phase flows called Interfaces and Sub-grid Scales (ISS). After the implementation of the ISS model in our code called Trio U , a set of various cases is used to validate this model. Then, special test are made in order to optimize the model for our particular bubbly flows. Thus we showed the capacity of the ISS model to produce a cheap pertinent solution. Secondly, we use the ISS model to perform simulations of bubbly flows in column. Results of these simulations are averaged to obtain quantities that appear in mass, momentum and interfacial area density balances. Thus, we processed to an a priori test of a complete one dimensional averaged model.We showed that this model predicts well the simplest flows (laminar and monodisperse). Moreover, the hypothesis of one pressure, which is often made in averaged model like CATHARE, NEPTUNE and RELAP5, is satisfied in such flows. At the opposite, without a polydisperse model, the drag is over-predicted and the uncorrelated A i flux needs a closure law. Finally, we showed that in turbulent flows, fluctuations of velocity and pressure in the liquid phase are not represented by the tested averaged model. (author)

  18. A comparison of average wages with age-specific wages for assessing indirect productivity losses: analytic simplicity versus analytic precision.

    Science.gov (United States)

    Connolly, Mark P; Tashjian, Cole; Kotsopoulos, Nikolaos; Bhatt, Aomesh; Postma, Maarten J

    2017-07-01

    Numerous approaches are used to estimate indirect productivity losses using various wage estimates applied to poor health in working aged adults. Considering the different wage estimation approaches observed in the published literature, we sought to assess variation in productivity loss estimates when using average wages compared with age-specific wages. Published estimates for average and age-specific wages for combined male/female wages were obtained from the UK Office of National Statistics. A polynomial interpolation was used to convert 5-year age-banded wage data into annual age-specific wages estimates. To compare indirect cost estimates, average wages and age-specific wages were used to project productivity losses at various stages of life based on the human capital approach. Discount rates of 0, 3, and 6 % were applied to projected age-specific and average wage losses. Using average wages was found to overestimate lifetime wages in conditions afflicting those aged 1-27 and 57-67, while underestimating lifetime wages in those aged 27-57. The difference was most significant for children where average wage overestimated wages by 15 % and for 40-year-olds where it underestimated wages by 14 %. Large differences in projecting productivity losses exist when using the average wage applied over a lifetime. Specifically, use of average wages overestimates productivity losses between 8 and 15 % for childhood illnesses. Furthermore, during prime working years, use of average wages will underestimate productivity losses by 14 %. We suggest that to achieve more precise estimates of productivity losses, age-specific wages should become the standard analytic approach.

  19. ECLogger: Cross-Project Catch-Block Logging Prediction Using Ensemble of Classifiers

    Directory of Open Access Journals (Sweden)

    Sangeeta Lal

    2017-01-01

    Full Text Available Background: Software developers insert log statements in the source code to record program execution information. However, optimizing the number of log statements in the source code is challenging. Machine learning based within-project logging prediction tools, proposed in previous studies, may not be suitable for new or small software projects. For such software projects, we can use cross-project logging prediction. Aim: The aim of the study presented here is to investigate cross-project logging prediction methods and techniques. Method: The proposed method is ECLogger, which is a novel, ensemble-based, cross-project, catch-block logging prediction model. In the research We use 9 base classifiers were used and combined using ensemble techniques. The performance of ECLogger was evaluated on on three open-source Java projects: Tomcat, CloudStack and Hadoop. Results: ECLogger Bagging, ECLogger AverageVote, and ECLogger MajorityVote show a considerable improvement in the average Logged F-measure (LF on 3, 5, and 4 source -> target project pairs, respectively, compared to the baseline classifiers. ECLogger AverageVote performs best and shows improvements of 3.12% (average LF and 6.08% (average ACC – Accuracy. Conclusion: The classifier based on ensemble techniques, such as bagging, average vote, and majority vote outperforms the baseline classifier. Overall, the ECLogger AverageVote model performs best. The results show that the CloudStack project is more generalizable than the other projects.

  20. Relationships between average depth and number of misclassifications for decision trees

    KAUST Repository

    Chikalov, Igor

    2014-02-14

    This paper presents a new tool for the study of relationships between the total path length or the average depth and the number of misclassifications for decision trees. In addition to algorithm, the paper also presents the results of experiments with datasets from UCI ML Repository [9] and datasets representing Boolean functions with 10 variables.

  1. Relationships between average depth and number of misclassifications for decision trees

    KAUST Repository

    Chikalov, Igor; Hussain, Shahid; Moshkov, Mikhail

    2014-01-01

    This paper presents a new tool for the study of relationships between the total path length or the average depth and the number of misclassifications for decision trees. In addition to algorithm, the paper also presents the results of experiments with datasets from UCI ML Repository [9] and datasets representing Boolean functions with 10 variables.

  2. STUDY OF WITHERS HEIGHT AVERAGE PERFORMANCES IN HUCUL HORSE BREED – HROBY BLOODLINE

    Directory of Open Access Journals (Sweden)

    M. MAFTEI

    2008-10-01

    Full Text Available Study of average performances in a population have a huge importance because, regarding a population, the average of phenotypic value is equal with average of genotypic value. So, the studies of the average value of characters offer us an idea about the population genetic level. The biological material is represented by 177 hucul horse from Hroby bloodline divided in 6 stallion families (tab. 1 analyzed at 18, 30 and 42 months old, owned by Lucina hucul stood farm. The average performances for withers height are presented in tab. 2. We can observe here that the average performances of the character are between characteristic limits of the breed. Both sexes have a small grade of variability with a decreasing tendency in the same time with ageing. We can observe a normal evolution in time for growth process with significant differences only at age of 42 months. We can say in this condition that the average performances for withers height have different values, influenced by the age, with a decreasing tendency.

  3. STUDY OF WITHERS HEIGHT AVERAGE PERFORMANCES IN HUCUL HORSE BREED –GORAL BLOODLINE

    Directory of Open Access Journals (Sweden)

    M. MAFTEI

    2008-10-01

    Full Text Available Study of average performances in a population have a huge importance because, regarding a population, the average of phenotypic value is equal with average of genotypic value. So, the studies of the average value of characters offer us an idea about the population genetic level. The biological material is represented by 87 hucul horse from Goral bloodline divided in 5 stallion families (tab. 1 analyzed at 18, 30 and 42 months old, owned by Lucina hucul stood farm. The average performances for withers height are presented in tab. 2. We can observe here that the average performances of the character are between characteristic limits of the breed. Both sexes have a small grade of variability with a decreasing tendency in the same time with ageing. We can observe a normal evolution in time for growth process with significant differences only at age of 42 months. We can say in this condition that the average performances for withers height have different values, influenced by the age, with a decreasing tendency.

  4. A new market risk model for cogeneration project financing---combined heat and power development without a power purchase agreement

    Science.gov (United States)

    Lockwood, Timothy A.

    Federal legislative changes in 2006 no longer entitle cogeneration project financings by law to receive the benefit of a power purchase agreement underwritten by an investment-grade investor-owned utility. Consequently, this research explored the need for a new market-risk model for future cogeneration and combined heat and power (CHP) project financing. CHP project investment represents a potentially enormous energy efficiency benefit through its application by reducing fossil fuel use up to 55% when compared to traditional energy generation, and concurrently eliminates constituent air emissions up to 50%, including global warming gases. As a supplemental approach to a comprehensive technical analysis, a quantitative multivariate modeling was also used to test the statistical validity and reliability of host facility energy demand and CHP supply ratios in predicting the economic performance of CHP project financing. The resulting analytical models, although not statistically reliable at this time, suggest a radically simplified CHP design method for future profitable CHP investments using four easily attainable energy ratios. This design method shows that financially successful CHP adoption occurs when the average system heat-to-power-ratio supply is less than or equal to the average host-convertible-energy-ratio, and when the average nominally-rated capacity is less than average host facility-load-factor demands. New CHP investments can play a role in solving the world-wide problem of accommodating growing energy demand while preserving our precious and irreplaceable air quality for future generations.

  5. Diversity and representativeness: two key factors

    CERN Multimedia

    Staff Association

    2013-01-01

    In the past few weeks many of you have filled out the questionnaire for preparing the upcoming Five-yearly review. Similarly, Staff Association members have elected their delegates to the Staff Council for the coming two years. Once again we would like to thank all those who have taken the time and effort to make their voice heard on these two occasions. Elections to the Staff Council Below we publish the new Staff Council with its forty delegates who will represent in 2014 and 2015 all CERN staff in the discussions with Management and Member States in the various bodies and committees. Therefore it is important that the Staff Council represents as far as possible the diversity of the CERN population. By construction, the election process with its electoral colleges and two-step voting procedure guarantees that all Departments, even the small ones, and the various staff categories are correctly represented. Figure 1 shows the participation rate in the elections. The average rate is just above 52 %, with ...

  6. Database of average-power damage thresholds at 1064 nm

    International Nuclear Information System (INIS)

    Rainer, F.; Hildum, E.A.; Milam, D.

    1987-01-01

    We have completed a database of average-power, laser-induced, damage thresholds at 1064 nm on a variety of materials. Measurements were made with a newly constructed laser to provide design input for moderate and high average-power laser projects. The measurements were conducted with 16-ns pulses at pulse-repetition frequencies ranging from 6 to 120 Hz. Samples were typically irradiated for time ranging from a fraction of a second up to 5 minutes (36,000 shots). We tested seven categories of samples which included antireflective coatings, high reflectors, polarizers, single and multiple layers of the same material, bare and overcoated metal surfaces, bare polished surfaces, and bulk materials. The measured damage threshold ranged from 2 for some metals to > 46 J/cm 2 for a bare polished glass substrate. 4 refs., 7 figs., 1 tab

  7. Engaging Diverse Students in Statistical Inquiry: A Comparison of Learning Experiences and Outcomes of Under-Represented and Non-Underrepresented Students Enrolled in a Multidisciplinary Project-Based Statistics Course

    Science.gov (United States)

    Dierker, Lisa; Alexander, Jalen; Cooper, Jennifer L.; Selya, Arielle; Rose, Jennifer; Dasgupta, Nilanjana

    2016-01-01

    Introductory statistics needs innovative, evidence-based teaching practices that support and engage diverse students. To evaluate the success of a multidisciplinary, project-based course, we compared experiences of under-represented (URM) and non-underrepresented students in 4 years of the course. While URM students considered the material more…

  8. Do 16 Polycyclic Aromatic Hydrocarbons Represent PAH Air Toxicity?

    Science.gov (United States)

    Samburova, Vera; Zielinska, Barbara; Khlystov, Andrey

    2017-08-15

    Estimation of carcinogenic potency based on analysis of 16 polycyclic aromatic hydrocarbons (PAHs) ranked by U.S. Environmental Protection Agency (EPA) is the most popular approach within scientific and environmental air quality management communities. The majority of PAH monitoring projects have been focused on particle-bound PAHs, ignoring the contribution of gas-phase PAHs to the toxicity of PAH mixtures in air samples. In this study, we analyzed the results of 13 projects in which 88 PAHs in both gas and particle phases were collected from different sources (biomass burning, mining operation, and vehicle emissions), as well as in urban air. The aim was to investigate whether 16 particle-bound U.S. EPA priority PAHs adequately represented health risks of inhalation exposure to atmospheric PAH mixtures. PAH concentrations were converted to benzo(a)pyrene-equivalent (BaPeq) toxicity using the toxic equivalency factor (TEF) approach. TEFs of PAH compounds for which such data is not available were estimated using TEFs of close isomers. Total BaPeq toxicities (∑ 88 BaPeq) of gas- and particle-phase PAHs were compared with BaPeq toxicities calculated for the 16 particle-phase EPA PAH (∑ 16EPA BaPeq). The results showed that 16 EPA particle-bound PAHs underrepresented the carcinogenic potency on average by 85.6% relative to the total (gas and particle) BaPeq toxicity of 88 PAHs. Gas-phase PAHs, like methylnaphthalenes, may contribute up to 30% of ∑ 88 BaPeq. Accounting for other individual non-EPA PAHs (i.e., benzo(e)pyrene) and gas-phase PAHs (i.e., naphthalene, 1- and 2-methylnaphthalene) will make the risk assessment of PAH-containing air samples significantly more accurate.

  9. Deblurring of class-averaged images in single-particle electron microscopy

    International Nuclear Information System (INIS)

    Park, Wooram; Chirikjian, Gregory S; Madden, Dean R; Rockmore, Daniel N

    2010-01-01

    This paper proposes a method for the deblurring of class-averaged images in single-particle electron microscopy (EM). Since EM images of biological samples are very noisy, the images which are nominally identical projection images are often grouped, aligned and averaged in order to cancel or reduce the background noise. However, the noise in the individual EM images generates errors in the alignment process, which creates an inherent limit on the accuracy of the resulting class averages. This inaccurate class average due to the alignment errors can be viewed as the result of a convolution of an underlying clear image with a blurring function. In this work, we develop a deconvolution method that gives an estimate for the underlying clear image from a blurred class-averaged image using precomputed statistics of misalignment. Since this convolution is over the group of rigid-body motions of the plane, SE(2), we use the Fourier transform for SE(2) in order to convert the convolution into a matrix multiplication in the corresponding Fourier space. For practical implementation we use a Hermite-function-based image modeling technique, because Hermite expansions enable lossless Cartesian-polar coordinate conversion using the Laguerre–Fourier expansions, and Hermite expansion and Laguerre–Fourier expansion retain their structures under the Fourier transform. Based on these mathematical properties, we can obtain the deconvolution of the blurred class average using simple matrix multiplication. Tests of the proposed deconvolution method using synthetic and experimental EM images confirm the performance of our method

  10. Neutron resonance averaging

    International Nuclear Information System (INIS)

    Chrien, R.E.

    1986-10-01

    The principles of resonance averaging as applied to neutron capture reactions are described. Several illustrations of resonance averaging to problems of nuclear structure and the distribution of radiative strength in nuclei are provided. 30 refs., 12 figs

  11. Celebrating the Smithsonian Soils Exhibit in the Classroom with the State/Representative Soil Project

    Science.gov (United States)

    Mikhailova, E. A.; Post, C. J.; Koppenheffer, Andrea; Asbill, John

    2009-01-01

    State/representative soil is one of many symbols (e.g., tree, flower, bird, etc.) adopted by citizens to be recognized as an important item to their state. We have developed a set of laboratory exercises, assignments, and exam questions utilizing the state/representative soil that gives college students an opportunity to practice interpretation of…

  12. On the average configuration of the geomagnetic tail

    International Nuclear Information System (INIS)

    Fairfield, D.H.

    1978-03-01

    Over 3000 hours of IMP-6 magnetic field data obtained between 20 and 33 R sub E in the geomagnetic tail have been used in a statistical study of the tail configuration. A distribution of 2.5 minute averages of B sub Z as a function of position across the tail reveals that more flux crosses the equatorial plane near the dawn and dusk flanks than near midnight. The tail field projected in the solar magnetospheric equatorial plane deviates from the X axis due to flaring and solar wind aberration by an angle alpha = -0.9 y sub SM - 1.7, where y/sub SM/ is in earth radii and alpha is in degrees. After removing these effects the Y component of the tail field is found to depend on interplanetary sector structure. During an away sector the B/sub Y/ component of the tail field is on average 0.5 gamma greater than that during a toward sector, a result that is true in both tail lobes and is independent of location across the tail

  13. Compositional dependences of average positron lifetime in binary As-S/Se glasses

    International Nuclear Information System (INIS)

    Ingram, A.; Golovchak, R.; Kostrzewa, M.; Wacke, S.; Shpotyuk, M.; Shpotyuk, O.

    2012-01-01

    Compositional dependence of average positron lifetime is studied systematically in typical representatives of binary As-S and As-Se glasses. This dependence is shown to be in opposite with molar volume evolution. The origin of this anomaly is discussed in terms of bond free solid angle concept applied to different types of structurally-intrinsic nanovoids in a glass.

  14. Compositional dependences of average positron lifetime in binary As-S/Se glasses

    Energy Technology Data Exchange (ETDEWEB)

    Ingram, A. [Department of Physics of Opole University of Technology, 75 Ozimska str., Opole, PL-45370 (Poland); Golovchak, R., E-mail: roman_ya@yahoo.com [Department of Materials Science and Engineering, Lehigh University, 5 East Packer Avenue, Bethlehem, PA 18015-3195 (United States); Kostrzewa, M.; Wacke, S. [Department of Physics of Opole University of Technology, 75 Ozimska str., Opole, PL-45370 (Poland); Shpotyuk, M. [Lviv Polytechnic National University, 12, Bandery str., Lviv, UA-79013 (Ukraine); Shpotyuk, O. [Institute of Physics of Jan Dlugosz University, 13/15al. Armii Krajowej, Czestochowa, PL-42201 (Poland)

    2012-02-15

    Compositional dependence of average positron lifetime is studied systematically in typical representatives of binary As-S and As-Se glasses. This dependence is shown to be in opposite with molar volume evolution. The origin of this anomaly is discussed in terms of bond free solid angle concept applied to different types of structurally-intrinsic nanovoids in a glass.

  15. Translating Uncertain Sea Level Projections Into Infrastructure Impacts Using a Bayesian Framework

    Science.gov (United States)

    Moftakhari, Hamed; AghaKouchak, Amir; Sanders, Brett F.; Matthew, Richard A.; Mazdiyasni, Omid

    2017-12-01

    Climate change may affect ocean-driven coastal flooding regimes by both raising the mean sea level (msl) and altering ocean-atmosphere interactions. For reliable projections of coastal flood risk, information provided by different climate models must be considered in addition to associated uncertainties. In this paper, we propose a framework to project future coastal water levels and quantify the resulting flooding hazard to infrastructure. We use Bayesian Model Averaging to generate a weighted ensemble of storm surge predictions from eight climate models for two coastal counties in California. The resulting ensembles combined with msl projections, and predicted astronomical tides are then used to quantify changes in the likelihood of road flooding under representative concentration pathways 4.5 and 8.5 in the near-future (1998-2063) and mid-future (2018-2083). The results show that road flooding rates will be significantly higher in the near-future and mid-future compared to the recent past (1950-2015) if adaptation measures are not implemented.

  16. Global Annual Average PM2.5 Grids from MODIS and MISR Aerosol Optical Depth (AOD)

    Data.gov (United States)

    National Aeronautics and Space Administration — Global Annual PM2.5 Grids from MODIS and MISR Aerosol Optical Depth (AOD) data set represents a series of annual average grids (2001-2010) of fine particulate matter...

  17. Global Annual Average PM2.5 Grids from MODIS and MISR Aerosol Optical Depth (AOD)

    Data.gov (United States)

    National Aeronautics and Space Administration — Global Annual PM2.5 Grids from MODIS and MISR Aerosol Optical Depth (AOD) data sets represent a series of annual average grids (2001-2010) of fine particulate matter...

  18. Representing Boolean Functions by Decision Trees

    KAUST Repository

    Chikalov, Igor

    2011-01-01

    A Boolean or discrete function can be represented by a decision tree. A compact form of decision tree named binary decision diagram or branching program is widely known in logic design [2, 40]. This representation is equivalent to other forms, and in some cases it is more compact than values table or even the formula [44]. Representing a function in the form of decision tree allows applying graph algorithms for various transformations [10]. Decision trees and branching programs are used for effective hardware [15] and software [5] implementation of functions. For the implementation to be effective, the function representation should have minimal time and space complexity. The average depth of decision tree characterizes the expected computing time, and the number of nodes in branching program characterizes the number of functional elements required for implementation. Often these two criteria are incompatible, i.e. there is no solution that is optimal on both time and space complexity. © Springer-Verlag Berlin Heidelberg 2011.

  19. Calculating ensemble averaged descriptions of protein rigidity without sampling.

    Science.gov (United States)

    González, Luis C; Wang, Hui; Livesay, Dennis R; Jacobs, Donald J

    2012-01-01

    Previous works have demonstrated that protein rigidity is related to thermodynamic stability, especially under conditions that favor formation of native structure. Mechanical network rigidity properties of a single conformation are efficiently calculated using the integer body-bar Pebble Game (PG) algorithm. However, thermodynamic properties require averaging over many samples from the ensemble of accessible conformations to accurately account for fluctuations in network topology. We have developed a mean field Virtual Pebble Game (VPG) that represents the ensemble of networks by a single effective network. That is, all possible number of distance constraints (or bars) that can form between a pair of rigid bodies is replaced by the average number. The resulting effective network is viewed as having weighted edges, where the weight of an edge quantifies its capacity to absorb degrees of freedom. The VPG is interpreted as a flow problem on this effective network, which eliminates the need to sample. Across a nonredundant dataset of 272 protein structures, we apply the VPG to proteins for the first time. Our results show numerically and visually that the rigidity characterizations of the VPG accurately reflect the ensemble averaged [Formula: see text] properties. This result positions the VPG as an efficient alternative to understand the mechanical role that chemical interactions play in maintaining protein stability.

  20. Kumaraswamy autoregressive moving average models for double bounded environmental data

    Science.gov (United States)

    Bayer, Fábio Mariano; Bayer, Débora Missio; Pumi, Guilherme

    2017-12-01

    In this paper we introduce the Kumaraswamy autoregressive moving average models (KARMA), which is a dynamic class of models for time series taking values in the double bounded interval (a,b) following the Kumaraswamy distribution. The Kumaraswamy family of distribution is widely applied in many areas, especially hydrology and related fields. Classical examples are time series representing rates and proportions observed over time. In the proposed KARMA model, the median is modeled by a dynamic structure containing autoregressive and moving average terms, time-varying regressors, unknown parameters and a link function. We introduce the new class of models and discuss conditional maximum likelihood estimation, hypothesis testing inference, diagnostic analysis and forecasting. In particular, we provide closed-form expressions for the conditional score vector and conditional Fisher information matrix. An application to environmental real data is presented and discussed.

  1. Molecular basis sets - a general similarity-based approach for representing chemical spaces.

    Science.gov (United States)

    Raghavendra, Akshay S; Maggiora, Gerald M

    2007-01-01

    A new method, based on generalized Fourier analysis, is described that utilizes the concept of "molecular basis sets" to represent chemical space within an abstract vector space. The basis vectors in this space are abstract molecular vectors. Inner products among the basis vectors are determined using an ansatz that associates molecular similarities between pairs of molecules with their corresponding inner products. Moreover, the fact that similarities between pairs of molecules are, in essentially all cases, nonzero implies that the abstract molecular basis vectors are nonorthogonal, but since the similarity of a molecule with itself is unity, the molecular vectors are normalized to unity. A symmetric orthogonalization procedure, which optimally preserves the character of the original set of molecular basis vectors, is used to construct appropriate orthonormal basis sets. Molecules can then be represented, in general, by sets of orthonormal "molecule-like" basis vectors within a proper Euclidean vector space. However, the dimension of the space can become quite large. Thus, the work presented here assesses the effect of basis set size on a number of properties including the average squared error and average norm of molecular vectors represented in the space-the results clearly show the expected reduction in average squared error and increase in average norm as the basis set size is increased. Several distance-based statistics are also considered. These include the distribution of distances and their differences with respect to basis sets of differing size and several comparative distance measures such as Spearman rank correlation and Kruscal stress. All of the measures show that, even though the dimension can be high, the chemical spaces they represent, nonetheless, behave in a well-controlled and reasonable manner. Other abstract vector spaces analogous to that described here can also be constructed providing that the appropriate inner products can be directly

  2. Familiarity and Voice Representation: From Acoustic-Based Representation to Voice Averages

    Directory of Open Access Journals (Sweden)

    Maureen Fontaine

    2017-07-01

    Full Text Available The ability to recognize an individual from their voice is a widespread ability with a long evolutionary history. Yet, the perceptual representation of familiar voices is ill-defined. In two experiments, we explored the neuropsychological processes involved in the perception of voice identity. We specifically explored the hypothesis that familiar voices (trained-to-familiar (Experiment 1, and famous voices (Experiment 2 are represented as a whole complex pattern, well approximated by the average of multiple utterances produced by a single speaker. In experiment 1, participants learned three voices over several sessions, and performed a three-alternative forced-choice identification task on original voice samples and several “speaker averages,” created by morphing across varying numbers of different vowels (e.g., [a] and [i] produced by the same speaker. In experiment 2, the same participants performed the same task on voice samples produced by familiar speakers. The two experiments showed that for famous voices, but not for trained-to-familiar voices, identification performance increased and response times decreased as a function of the number of utterances in the averages. This study sheds light on the perceptual representation of familiar voices, and demonstrates the power of average in recognizing familiar voices. The speaker average captures the unique characteristics of a speaker, and thus retains the information essential for recognition; it acts as a prototype of the speaker.

  3. Approximate Dual Averaging Method for Multiagent Saddle-Point Problems with Stochastic Subgradients

    Directory of Open Access Journals (Sweden)

    Deming Yuan

    2014-01-01

    Full Text Available This paper considers the problem of solving the saddle-point problem over a network, which consists of multiple interacting agents. The global objective function of the problem is a combination of local convex-concave functions, each of which is only available to one agent. Our main focus is on the case where the projection steps are calculated approximately and the subgradients are corrupted by some stochastic noises. We propose an approximate version of the standard dual averaging method and show that the standard convergence rate is preserved, provided that the projection errors decrease at some appropriate rate and the noises are zero-mean and have bounded variance.

  4. IPAD project overview

    Science.gov (United States)

    Fulton, R. E.

    1980-01-01

    To respond to national needs for improved productivity in engineering design and manufacturing, a NASA supported joint industry/government project is underway denoted Integrated Programs for Aerospace-Vehicle Design (IPAD). The objective is to improve engineering productivity through better use of computer technology. It focuses on development of technology and associated software for integrated company-wide management of engineering information. The project has been underway since 1976 under the guidance of an Industry Technical Advisory Board (ITAB) composed of representatives of major engineering and computer companies and in close collaboration with the Air Force Integrated Computer-Aided Manufacturing (ICAM) program. Results to date on the IPAD project include an in-depth documentation of a representative design process for a large engineering project, the definition and design of computer-aided design software needed to support that process, and the release of prototype software to integrate selected design functions. Ongoing work concentrates on development of prototype software to manage engineering information, and initial software is nearing release.

  5. Model of averaged turbulent flow around cylindrical column for simulation of the saltation

    Czech Academy of Sciences Publication Activity Database

    Kharlamova, Irina; Kharlamov, Alexander; Vlasák, Pavel

    2014-01-01

    Roč. 21, č. 2 (2014), s. 103-110 ISSN 1802-1484 R&D Projects: GA ČR GA103/09/1718 Institutional research plan: CEZ:AV0Z20600510 Institutional support: RVO:67985874 Keywords : sediment transport * flow around cylinder * logarithmic profile * dipole line * averaged turbulent flow Subject RIV: BK - Fluid Dynamics

  6. A virtual pebble game to ensemble average graph rigidity.

    Science.gov (United States)

    González, Luis C; Wang, Hui; Livesay, Dennis R; Jacobs, Donald J

    2015-01-01

    The body-bar Pebble Game (PG) algorithm is commonly used to calculate network rigidity properties in proteins and polymeric materials. To account for fluctuating interactions such as hydrogen bonds, an ensemble of constraint topologies are sampled, and average network properties are obtained by averaging PG characterizations. At a simpler level of sophistication, Maxwell constraint counting (MCC) provides a rigorous lower bound for the number of internal degrees of freedom (DOF) within a body-bar network, and it is commonly employed to test if a molecular structure is globally under-constrained or over-constrained. MCC is a mean field approximation (MFA) that ignores spatial fluctuations of distance constraints by replacing the actual molecular structure by an effective medium that has distance constraints globally distributed with perfect uniform density. The Virtual Pebble Game (VPG) algorithm is a MFA that retains spatial inhomogeneity in the density of constraints on all length scales. Network fluctuations due to distance constraints that may be present or absent based on binary random dynamic variables are suppressed by replacing all possible constraint topology realizations with the probabilities that distance constraints are present. The VPG algorithm is isomorphic to the PG algorithm, where integers for counting "pebbles" placed on vertices or edges in the PG map to real numbers representing the probability to find a pebble. In the VPG, edges are assigned pebble capacities, and pebble movements become a continuous flow of probability within the network. Comparisons between the VPG and average PG results over a test set of proteins and disordered lattices demonstrate the VPG quantitatively estimates the ensemble average PG results well. The VPG performs about 20% faster than one PG, and it provides a pragmatic alternative to averaging PG rigidity characteristics over an ensemble of constraint topologies. The utility of the VPG falls in between the most

  7. 40 CFR 197.31 - What is a representative volume?

    Science.gov (United States)

    2010-07-01

    ... supply a given water demand. The DOE must project the concentration of radionuclides released from the... Public Health and Environmental Standards for Disposal Ground Water Protection Standards § 197.31 What is a representative volume? (a) It is the volume of ground water that would be withdrawn annually from...

  8. 28 May 2010 - Representatives of the Netherlands School of Public Administration guided in the ATLAS visitor centre by ATLAS Collaboration Member and NIKHEF G. Bobbink and ATLAS Magnet Project Leader H.ten Kate.

    CERN Document Server

    Maximilien Brice

    2010-01-01

    28 May 2010 - Representatives of the Netherlands School of Public Administration guided in the ATLAS visitor centre by ATLAS Collaboration Member and NIKHEF G. Bobbink and ATLAS Magnet Project Leader H.ten Kate.

  9. Average-energy games

    Directory of Open Access Journals (Sweden)

    Patricia Bouyer

    2015-09-01

    Full Text Available Two-player quantitative zero-sum games provide a natural framework to synthesize controllers with performance guarantees for reactive systems within an uncontrollable environment. Classical settings include mean-payoff games, where the objective is to optimize the long-run average gain per action, and energy games, where the system has to avoid running out of energy. We study average-energy games, where the goal is to optimize the long-run average of the accumulated energy. We show that this objective arises naturally in several applications, and that it yields interesting connections with previous concepts in the literature. We prove that deciding the winner in such games is in NP inter coNP and at least as hard as solving mean-payoff games, and we establish that memoryless strategies suffice to win. We also consider the case where the system has to minimize the average-energy while maintaining the accumulated energy within predefined bounds at all times: this corresponds to operating with a finite-capacity storage for energy. We give results for one-player and two-player games, and establish complexity bounds and memory requirements.

  10. Averaging of nonlinearity-managed pulses

    International Nuclear Information System (INIS)

    Zharnitsky, Vadim; Pelinovsky, Dmitry

    2005-01-01

    We consider the nonlinear Schroedinger equation with the nonlinearity management which describes Bose-Einstein condensates under Feshbach resonance. By using an averaging theory, we derive the Hamiltonian averaged equation and compare it with other averaging methods developed for this problem. The averaged equation is used for analytical approximations of nonlinearity-managed solitons

  11. National Health Expenditure Projections, 2015-25: Economy, Prices, And Aging Expected To Shape Spending And Enrollment.

    Science.gov (United States)

    Keehan, Sean P; Poisal, John A; Cuckler, Gigi A; Sisko, Andrea M; Smith, Sheila D; Madison, Andrew J; Stone, Devin A; Wolfe, Christian J; Lizonitz, Joseph M

    2016-08-01

    Health spending growth in the United States for 2015-25 is projected to average 5.8 percent-1.3 percentage points faster than growth in the gross domestic product-and to represent 20.1 percent of the total economy by 2025. As the initial impacts associated with the Affordable Care Act's coverage expansions fade, growth in health spending is expected to be influenced by changes in economic growth, faster growth in medical prices, and population aging. Projected national health spending growth, though faster than observed in the recent history, is slower than in the two decades before the recent Great Recession, in part because of trends such as increasing cost sharing in private health insurance plans and various Medicare payment update provisions. In addition, the share of total health expenditures paid for by federal, state, and local governments is projected to increase to 47 percent by 2025. Project HOPE—The People-to-People Health Foundation, Inc.

  12. Calculating ensemble averaged descriptions of protein rigidity without sampling.

    Directory of Open Access Journals (Sweden)

    Luis C González

    Full Text Available Previous works have demonstrated that protein rigidity is related to thermodynamic stability, especially under conditions that favor formation of native structure. Mechanical network rigidity properties of a single conformation are efficiently calculated using the integer body-bar Pebble Game (PG algorithm. However, thermodynamic properties require averaging over many samples from the ensemble of accessible conformations to accurately account for fluctuations in network topology. We have developed a mean field Virtual Pebble Game (VPG that represents the ensemble of networks by a single effective network. That is, all possible number of distance constraints (or bars that can form between a pair of rigid bodies is replaced by the average number. The resulting effective network is viewed as having weighted edges, where the weight of an edge quantifies its capacity to absorb degrees of freedom. The VPG is interpreted as a flow problem on this effective network, which eliminates the need to sample. Across a nonredundant dataset of 272 protein structures, we apply the VPG to proteins for the first time. Our results show numerically and visually that the rigidity characterizations of the VPG accurately reflect the ensemble averaged [Formula: see text] properties. This result positions the VPG as an efficient alternative to understand the mechanical role that chemical interactions play in maintaining protein stability.

  13. Thermal motion in proteins: Large effects on the time-averaged interaction energies

    International Nuclear Information System (INIS)

    Goethe, Martin; Rubi, J. Miguel; Fita, Ignacio

    2016-01-01

    As a consequence of thermal motion, inter-atomic distances in proteins fluctuate strongly around their average values, and hence, also interaction energies (i.e. the pair-potentials evaluated at the fluctuating distances) are not constant in time but exhibit pronounced fluctuations. These fluctuations cause that time-averaged interaction energies do generally not coincide with the energy values obtained by evaluating the pair-potentials at the average distances. More precisely, time-averaged interaction energies behave typically smoother in terms of the average distance than the corresponding pair-potentials. This averaging effect is referred to as the thermal smoothing effect. Here, we estimate the strength of the thermal smoothing effect on the Lennard-Jones pair-potential for globular proteins at ambient conditions using x-ray diffraction and simulation data of a representative set of proteins. For specific atom species, we find a significant smoothing effect where the time-averaged interaction energy of a single atom pair can differ by various tens of cal/mol from the Lennard-Jones potential at the average distance. Importantly, we observe a dependency of the effect on the local environment of the involved atoms. The effect is typically weaker for bulky backbone atoms in beta sheets than for side-chain atoms belonging to other secondary structure on the surface of the protein. The results of this work have important practical implications for protein software relying on free energy expressions. We show that the accuracy of free energy expressions can largely be increased by introducing environment specific Lennard-Jones parameters accounting for the fact that the typical thermal motion of protein atoms depends strongly on their local environment.

  14. Thermal motion in proteins: Large effects on the time-averaged interaction energies

    Energy Technology Data Exchange (ETDEWEB)

    Goethe, Martin, E-mail: martingoethe@ub.edu; Rubi, J. Miguel [Departament de Física Fonamental, Universitat de Barcelona, Martí i Franquès 1, 08028 Barcelona (Spain); Fita, Ignacio [Institut de Biologia Molecular de Barcelona, Baldiri Reixac 10, 08028 Barcelona (Spain)

    2016-03-15

    As a consequence of thermal motion, inter-atomic distances in proteins fluctuate strongly around their average values, and hence, also interaction energies (i.e. the pair-potentials evaluated at the fluctuating distances) are not constant in time but exhibit pronounced fluctuations. These fluctuations cause that time-averaged interaction energies do generally not coincide with the energy values obtained by evaluating the pair-potentials at the average distances. More precisely, time-averaged interaction energies behave typically smoother in terms of the average distance than the corresponding pair-potentials. This averaging effect is referred to as the thermal smoothing effect. Here, we estimate the strength of the thermal smoothing effect on the Lennard-Jones pair-potential for globular proteins at ambient conditions using x-ray diffraction and simulation data of a representative set of proteins. For specific atom species, we find a significant smoothing effect where the time-averaged interaction energy of a single atom pair can differ by various tens of cal/mol from the Lennard-Jones potential at the average distance. Importantly, we observe a dependency of the effect on the local environment of the involved atoms. The effect is typically weaker for bulky backbone atoms in beta sheets than for side-chain atoms belonging to other secondary structure on the surface of the protein. The results of this work have important practical implications for protein software relying on free energy expressions. We show that the accuracy of free energy expressions can largely be increased by introducing environment specific Lennard-Jones parameters accounting for the fact that the typical thermal motion of protein atoms depends strongly on their local environment.

  15. Different Multifractal Scaling of the 0 cm Average Ground Surface Temperature of Four Representative Weather Stations over China

    Directory of Open Access Journals (Sweden)

    Lei Jiang

    2013-01-01

    Full Text Available The temporal scaling properties of the daily 0 cm average ground surface temperature (AGST records obtained from four selected sites over China are investigated using multifractal detrended fluctuation analysis (MF-DFA method. Results show that the AGST records at all four locations exhibit strong persistence features and different scaling behaviors. The differences of the generalized Hurst exponents are very different for the AGST series of each site reflecting the different scaling behaviors of the fluctuation. Furthermore, the strengths of multifractal spectrum are different for different weather stations and indicate that the multifractal behaviors vary from station to station over China.

  16. Applications of ordered weighted averaging (OWA operators in environmental problems

    Directory of Open Access Journals (Sweden)

    Carlos Llopis-Albert

    2017-04-01

    Full Text Available This paper presents an application of a prioritized weighted aggregation operator based on ordered weighted averaging (OWA to deal with stakeholders' constructive participation in water resources projects. They have different degree of acceptance or preference regarding the measures and policies to be carried out, which lead to different environmental and socio-economic outcomes, and hence, to different levels of stakeholders’ satisfaction. The methodology establishes a prioritization relationship upon the stakeholders, which preferences are aggregated by means of weights depending on the satisfaction of the higher priority policy maker. The methodology establishes a prioritization relationship upon the stakeholders, which preferences are aggregated by means of weights depending on the satisfaction of the higher priority policy maker. The methodology has been successfully applied to a Public Participation Project (PPP in watershed management, thus obtaining efficient environmental measures in conflict resolution problems under actors’ preference uncertainties.

  17. The Use of the Nelder-Mead Method in Determining Projection Parameters for Globe Photographs

    Science.gov (United States)

    Gede, M.

    2009-04-01

    A photo of a terrestrial or celestial globe can be handled as a map. The only hard issue is its projection: the so-called Tilted Perspective Projection which, if the optical axis of the photo intersects the globe's centre, is simplified to the Vertical Near-Side Perspective Projection. When georeferencing such a photo, the exact parameters of the projections are also needed. These parameters depend on the position of the viewpoint of the camera. Several hundreds of globe photos had to be georeferenced during the Virtual Globes Museum project, which made necessary to automatize the calculation of the projection parameters. The author developed a program for this task which uses the Nelder-Mead Method in order to find the optimum parameters when a set of control points are given as input. The Nelder-Mead method is a numerical algorithm for minimizing a function in a many-dimensional space. The function in the present application is the average error of the control points calculated from the actual values of parameters. The parameters are the geographical coordinates of the projection centre, the image coordinates of the same point, the rotation of the projection, the height of the perspective point and the scale of the photo (calculated in pixels/km). The program reads the Global Mappers Ground Control Point (.GCP) file format as input and creates projection description files (.PRJ) for the same software. The initial values of the geographical coordinates of the projection centre are calculated as the average of the control points, while the other parameters are set to experimental values which represent the most common circumstances of taking a globe photograph. The algorithm runs until the change of the parameters sinks below a pre-defined limit. The minimum search can be refined by using the previous result parameter set as new initial values. This paper introduces the calculation mechanism and examples of the usage. Other possible other usages of the method are

  18. Image-processing of time-averaged interface distributions representing CCFL characteristics in a large scale model of a PWR hot-leg pipe geometry

    International Nuclear Information System (INIS)

    Al Issa, Suleiman; Macián-Juan, Rafael

    2017-01-01

    Highlights: • CCFL characteristics are investigated in PWR large-scale hot-leg pipe geometry. • Image processing of air-water interface produced time-averaged interface distributions. • Time-averages provide a comparative method of CCFL characteristics among different studies. • CCFL correlations depend upon the range of investigated water delivery for Dh ≫ 50 mm. • 1D codes are incapable of investigating CCFL because of lack of interface distribution. - Abstract: Countercurrent Flow Limitation (CCFL) was experimentally investigated in a 1/3.9 downscaled COLLIDER facility with a 190 mm pipe’s diameter using air/water at 1 atmospheric pressure. Previous investigations provided knowledge over the onset of CCFL mechanisms. In current article, CCFL characteristics at the COLLIDER facility are measured and discussed along with time-averaged distributions of the air/water interface for a selected matrix of liquid/gas velocities. The article demonstrates the time-averaged interface as a useful method to identify CCFL characteristics at quasi-stationary flow conditions eliminating variations that appears in single images, and showing essential comparative flow features such as: the degree of restriction at the bend, the extension and the intensity of the two-phase mixing zones, and the average water level within the horizontal part and the steam generator. Consequently, making it possible to compare interface distributions obtained at different investigations. The distributions are also beneficial for CFD validations of CCFL as the instant chaotic gas/liquid interface is impossible to reproduce in CFD simulations. The current study shows that final CCFL characteristics curve (and the corresponding CCFL correlation) depends upon the covered measuring range of water delivery. It also shows that a hydraulic diameter should be sufficiently larger than 50 mm in order to obtain CCFL characteristics comparable to the 1:1 scale data (namely the UPTF data). Finally

  19. Guidelines for Automation Project Execution

    OpenAIRE

    Takkinen, Heidi

    2011-01-01

    The purpose of this Master’s thesis was to create instructions for executing an automation project. Sarlin Oy Ab needed directions on how to execute an automation project. Sarlin is starting up a new business area offering total project solutions for customers. Sarlin focuses on small and minor automation projects on domestic markets. The thesis represents issues related to project execution starting from the theory of the project to its kick-off and termination. Site work is one importan...

  20. The difference between alternative averages

    Directory of Open Access Journals (Sweden)

    James Vaupel

    2012-09-01

    Full Text Available BACKGROUND Demographers have long been interested in how compositional change, e.g., change in age structure, affects population averages. OBJECTIVE We want to deepen understanding of how compositional change affects population averages. RESULTS The difference between two averages of a variable, calculated using alternative weighting functions, equals the covariance between the variable and the ratio of the weighting functions, divided by the average of the ratio. We compare weighted and unweighted averages and also provide examples of use of the relationship in analyses of fertility and mortality. COMMENTS Other uses of covariances in formal demography are worth exploring.

  1. Energy projections 1979

    International Nuclear Information System (INIS)

    1979-01-01

    The projections, prepared by Department of Energy officials, examine possible UK energy demand and supply prospects to the end of the century. They are based on certain broad long term assumptions about economic growth, technical improvements and movements in energy prices. The projections are intended to provide a broad quantitative framework for the consideration of possible energy futures and policy choices. Two cases are considered. In the first, the UK economy is assumed to grow at about 3 per cent to the end of the century and, in the second, at a lower level of about 2 per cent per annum. In both it is assumed that world oil prices will rise significantly above present levels (reaching some 30 dollars a barrel in terms of 1977 prices for Saudi Arabian marker crude by the end of the century). After incorporation of allowances for energy conservation which approximate to a reduction of some 20 per cent in demand, total primary fuel requirements in the year 2000 are estimated in the range 445 to 510 million tonnes of coal equivalent (mtce), representing an average rate of growth of 0.9 to 1.5 per cent a year. Potential indigenous energy supply by the end of the century is estimated in the range 390 to 410 mtce. This includes a possible installed nuclear capacity of up to 40 Gigawatts, approximately a fourfold increase on capacity already installed or under construction, and indigenous coal production of up to 155 million tonnes a year. The projections highlight the UK's prospective emergence during the later part of the century from a period from 1980 of energy surplus and the increasing roles which energy conservation, nuclear power and coal will be called upon to play as oil becomes scarcer and more expensive in the international market and as indigenous oil and gas production declines. (author)

  2. Opposite latitudinal gradients in projected ocean acidification and bleaching impacts on coral reefs.

    Science.gov (United States)

    van Hooidonk, Ruben; Maynard, Jeffrey Allen; Manzello, Derek; Planes, Serge

    2014-01-01

    Coral reefs and the services they provide are seriously threatened by ocean acidification and climate change impacts like coral bleaching. Here, we present updated global projections for these key threats to coral reefs based on ensembles of IPCC AR5 climate models using the new Representative Concentration Pathway (RCP) experiments. For all tropical reef locations, we project absolute and percentage changes in aragonite saturation state (Ωarag) for the period between 2006 and the onset of annual severe bleaching (thermal stress >8 degree heating weeks); a point at which it is difficult to believe reefs can persist as we know them. Severe annual bleaching is projected to start 10-15 years later at high-latitude reefs than for reefs in low latitudes under RCP8.5. In these 10-15 years, Ωarag keeps declining and thus any benefits for high-latitude reefs of later onset of annual bleaching may be negated by the effects of acidification. There are no long-term refugia from the effects of both acidification and bleaching. Of all reef locations, 90% are projected to experience severe bleaching annually by 2055. Furthermore, 5% declines in calcification are projected for all reef locations by 2034 under RCP8.5, assuming a 15% decline in calcification per unit of Ωarag. Drastic emissions cuts, such as those represented by RCP6.0, result in an average year for the onset of annual severe bleaching that is ~20 years later (2062 vs. 2044). However, global emissions are tracking above the current worst-case scenario devised by the scientific community, as has happened in previous generations of emission scenarios. The projections here for conditions on coral reefs are dire, but provide the most up-to-date assessment of what the changing climate and ocean acidification mean for the persistence of coral reefs. © 2013 John Wiley & Sons Ltd.

  3. Model parameters for representative wetland plant functional groups

    Science.gov (United States)

    Williams, Amber S.; Kiniry, James R.; Mushet, David M.; Smith, Loren M.; McMurry, Scott T.; Attebury, Kelly; Lang, Megan; McCarty, Gregory W.; Shaffer, Jill A.; Effland, William R.; Johnson, Mari-Vaughn V.

    2017-01-01

    Wetlands provide a wide variety of ecosystem services including water quality remediation, biodiversity refugia, groundwater recharge, and floodwater storage. Realistic estimation of ecosystem service benefits associated with wetlands requires reasonable simulation of the hydrology of each site and realistic simulation of the upland and wetland plant growth cycles. Objectives of this study were to quantify leaf area index (LAI), light extinction coefficient (k), and plant nitrogen (N), phosphorus (P), and potassium (K) concentrations in natural stands of representative plant species for some major plant functional groups in the United States. Functional groups in this study were based on these parameters and plant growth types to enable process-based modeling. We collected data at four locations representing some of the main wetland regions of the United States. At each site, we collected on-the-ground measurements of fraction of light intercepted, LAI, and dry matter within the 2013–2015 growing seasons. Maximum LAI and k variables showed noticeable variations among sites and years, while overall averages and functional group averages give useful estimates for multisite simulation modeling. Variation within each species gives an indication of what can be expected in such natural ecosystems. For P and K, the concentrations from highest to lowest were spikerush (Eleocharis macrostachya), reed canary grass (Phalaris arundinacea), smartweed (Polygonum spp.), cattail (Typha spp.), and hardstem bulrush (Schoenoplectus acutus). Spikerush had the highest N concentration, followed by smartweed, bulrush, reed canary grass, and then cattail. These parameters will be useful for the actual wetland species measured and for the wetland plant functional groups they represent. These parameters and the associated process-based models offer promise as valuable tools for evaluating environmental benefits of wetlands and for evaluating impacts of various agronomic practices in

  4. Reduced game property of the egalitarian non-k-averaged contribution value and the Shapley value

    NARCIS (Netherlands)

    Namekata, Tsuneyuki; Driessen, Theo; Namekata, T.

    2000-01-01

    The Egalitarian Non-k-Averaged Contribution (ENk AC-) value for TU-game represents the equal division of the surplus of the total profits, given that each player is already allocated his individual contribution specified by worths of coalitions of size k. This paper deals with the axiomatic

  5. SPATIAL DISTRIBUTION OF THE AVERAGE RUNOFF IN THE IZA AND VIȘEU WATERSHEDS

    Directory of Open Access Journals (Sweden)

    HORVÁTH CS.

    2015-03-01

    Full Text Available The average runoff represents the main parameter with which one can best evaluate an area’s water resources and it is also an important characteristic in al river runoff research. In this paper we choose a GIS methodology for assessing the spatial evolution of the average runoff, using validity curves we identifies three validity areas in which the runoff changes differently with altitude. The tree curves were charted using the average runoff values of 16 hydrometric stations from the area, eight in the Vișeu and eight in the Iza river catchment. Identifying the appropriate areas of the obtained correlations curves (between specific average runoff and catchments mean altitude allowed the assessment of potential runoff at catchment level and on altitudinal intervals. By integrating the curves functions in to GIS we created an average runoff map for the area; from which one can easily extract runoff data using GIS spatial analyst functions. The study shows that from the three areas the highest runoff corresponds with the third zone but because it’s small area the water volume is also minor. It is also shown that with the use of the created runoff map we can compute relatively quickly correct runoff values for areas without hydrologic control.

  6. How accurately are climatological characteristics and surface water and energy balances represented for the Colombian Caribbean Catchment Basin?

    Science.gov (United States)

    Hoyos, Isabel; Baquero-Bernal, Astrid; Hagemann, Stefan

    2013-09-01

    In Colombia, the access to climate related observational data is restricted and their quantity is limited. But information about the current climate is fundamental for studies on present and future climate changes and their impacts. In this respect, this information is especially important over the Colombian Caribbean Catchment Basin (CCCB) that comprises over 80 % of the population of Colombia and produces about 85 % of its GDP. Consequently, an ensemble of several datasets has been evaluated and compared with respect to their capability to represent the climate over the CCCB. The comparison includes observations, reconstructed data (CPC, Delaware), reanalyses (ERA-40, NCEP/NCAR), and simulated data produced with the regional climate model REMO. The capabilities to represent the average annual state, the seasonal cycle, and the interannual variability are investigated. The analyses focus on surface air temperature and precipitation as well as on surface water and energy balances. On one hand the CCCB characteristics poses some difficulties to the datasets as the CCCB includes a mountainous region with three mountain ranges, where the dynamical core of models and model parameterizations can fail. On the other hand, it has the most dense network of stations, with the longest records, in the country. The results can be summarised as follows: all of the datasets demonstrate a cold bias in the average temperature of CCCB. However, the variability of the average temperature of CCCB is most poorly represented by the NCEP/NCAR dataset. The average precipitation in CCCB is overestimated by all datasets. For the ERA-40, NCEP/NCAR, and REMO datasets, the amplitude of the annual cycle is extremely high. The variability of the average precipitation in CCCB is better represented by the reconstructed data of CPC and Delaware, as well as by NCEP/NCAR. Regarding the capability to represent the spatial behaviour of CCCB, temperature is better represented by Delaware and REMO, while

  7. Web Based Project Management System

    OpenAIRE

    Aadamsoo, Anne-Mai

    2010-01-01

    To increase an efficiency of a product, nowadays many web development companies are using different project management systems. A company may run a number of projects at a time, and requires input from a number of individuals, or teams for a multi level development plan, whereby a good project management system is needed. Project management systems represent a rapidly growing technology in IT industry. As the number of users, who utilize project management applications continues to grow, w...

  8. Simulated effects of projected ground-water withdrawals in the Floridan aquifer system, greater Orlando metropolitan area, east-central Florida

    Science.gov (United States)

    Murray, Louis C.; Halford, Keith J.

    1999-01-01

    Ground-water levels in the Floridan aquifer system within the greater Orlando metropolitan area are expected to decline because of a projected increase in the average pumpage rate from 410 million gallons per day in 1995 to 576 million gallons per day in 2020. The potential decline in ground-water levels and spring discharge within the area was investigated with a calibrated, steady-state, ground-water flow model. A wetter-than-average condition scenario and a drought-condition scenario were simulated to bracket the range of water-levels and springflow that may occur in 2020 under average rainfall conditions. Pumpage used to represent the drought-condition scenario totaled 865 million gallons per day, about 50 percent greater than the projected average pumpage rate in 2020. Relative to average 1995 steady-state conditions, drawdowns simulated in the Upper Floridan aquifer exceeded 10 and 25 feet for wet and dry conditions, respectively, in parts of central and southwest Orange County and in north Osceola County. In Seminole County, drawdowns of up to 20 feet were simulated for dry conditions, compared with 5 to 10 feet simulated for wet conditions. Computed springflow was reduced by 10 percent for wet conditions and by 38 percent for dry conditions, with the largest reductions (28 and 76 percent) occurring at the Sanlando Springs group. In the Lower Floridan aquifer, drawdowns simulated in southwest Orange County exceeded 20 and 40 feet for wet and dry conditions, respectively.

  9. A Boy with a Mild Case of Cornelia de Lange Syndrome with Above Average Intelligence.

    Science.gov (United States)

    Lacassie, Yves; Bobadilla, Olga; Cambias, Ron D., Jr.

    1997-01-01

    Describes the characteristics of an 11-year-old boy who represents the only documented case of an individual with Cornelia de Lange syndrome who also has above average cognitive functioning. Major diagnostic criteria for de Lange syndrome and comparisons with other severe and mild cases are discussed. (Author/CR)

  10. Bayesian model averaging and weighted average least squares : Equivariance, stability, and numerical issues

    NARCIS (Netherlands)

    De Luca, G.; Magnus, J.R.

    2011-01-01

    In this article, we describe the estimation of linear regression models with uncertainty about the choice of the explanatory variables. We introduce the Stata commands bma and wals, which implement, respectively, the exact Bayesian model-averaging estimator and the weighted-average least-squares

  11. Implementation of lean construction techniques for minimizing the risks effect on project construction time

    Directory of Open Access Journals (Sweden)

    Usama Hamed Issa

    2013-12-01

    Full Text Available The construction projects involve various risk factors which have various impacts on time objective that may lead to time-overrun. This study suggests and applies a new technique for minimizing risk factors effect on time using lean construction principles. The lean construction is implemented in this study using the last planner system through execution of an industrial project in Egypt. Evaluating the effect of using the new tool is described in terms of two measurements: Percent Expected Time-overrun (PET and Percent Plan Completed (PPC. The most important risk factors are identified and assessed, while PET is quantified at the project start and during the project execution using a model for time-overrun quantification. The results showed that total project time is reduced by 15.57% due to decreasing PET values, while PPC values improved. This is due to minimizing and mitigating the effect of most of the risk factors in this project due to implementing lean construction techniques. The results proved that the quantification model is suitable for evaluating the effect of using lean construction techniques. In addition, the results showed that average value of PET due to factors affected by lean techniques represents 67% from PET values due to all minimized risk factors.

  12. Northern Winter Climate Change: Assessment of Uncertainty in CMIP5 Projections Related to Stratosphere-Troposphere Coupling

    Science.gov (United States)

    Manzini, E.; Karpechko, A.Yu.; Anstey, J.; Shindell, Drew Todd; Baldwin, M.P.; Black, R.X.; Cagnazzo, C.; Calvo, N.; Charlton-Perez, A.; Christiansen, B.; hide

    2014-01-01

    Future changes in the stratospheric circulation could have an important impact on northern winter tropospheric climate change, given that sea level pressure (SLP) responds not only to tropospheric circulation variations but also to vertically coherent variations in troposphere-stratosphere circulation. Here we assess northern winter stratospheric change and its potential to influence surface climate change in the Coupled Model Intercomparison Project-Phase 5 (CMIP5) multimodel ensemble. In the stratosphere at high latitudes, an easterly change in zonally averaged zonal wind is found for the majority of the CMIP5 models, under the Representative Concentration Pathway 8.5 scenario. Comparable results are also found in the 1% CO2 increase per year projections, indicating that the stratospheric easterly change is common feature in future climate projections. This stratospheric wind change, however, shows a significant spread among the models. By using linear regression, we quantify the impact of tropical upper troposphere warming, polar amplification, and the stratospheric wind change on SLP. We find that the intermodel spread in stratospheric wind change contributes substantially to the intermodel spread in Arctic SLP change. The role of the stratosphere in determining part of the spread in SLP change is supported by the fact that the SLP change lags the stratospheric zonally averaged wind change. Taken together, these findings provide further support for the importance of simulating the coupling between the stratosphere and the troposphere, to narrow the uncertainty in the future projection of tropospheric circulation changes.

  13. Averaged head phantoms from magnetic resonance images of Korean children and young adults

    Science.gov (United States)

    Han, Miran; Lee, Ae-Kyoung; Choi, Hyung-Do; Jung, Yong Wook; Park, Jin Seo

    2018-02-01

    Increased use of mobile phones raises concerns about the health risks of electromagnetic radiation. Phantom heads are routinely used for radiofrequency dosimetry simulations, and the purpose of this study was to construct averaged phantom heads for children and young adults. Using magnetic resonance images (MRI), sectioned cadaver images, and a hybrid approach, we initially built template phantoms representing 6-, 9-, 12-, 15-year-old children and young adults. Our subsequent approach revised the template phantoms using 29 averaged items that were identified by averaging the MRI data from 500 children and young adults. In females, the brain size and cranium thickness peaked in the early teens and then decreased. This is contrary to what was observed in males, where brain size and cranium thicknesses either plateaued or grew continuously. The overall shape of brains was spherical in children and became ellipsoidal by adulthood. In this study, we devised a method to build averaged phantom heads by constructing surface and voxel models. The surface model could be used for phantom manipulation, whereas the voxel model could be used for compliance test of specific absorption rate (SAR) for users of mobile phones or other electronic devices.

  14. An Invariance Property for the Maximum Likelihood Estimator of the Parameters of a Gaussian Moving Average Process

    OpenAIRE

    Godolphin, E. J.

    1980-01-01

    It is shown that the estimation procedure of Walker leads to estimates of the parameters of a Gaussian moving average process which are asymptotically equivalent to the maximum likelihood estimates proposed by Whittle and represented by Godolphin.

  15. The database of the PREDICTS (Projecting Responses of Ecological Diversity In Changing Terrestrial Systems) project

    DEFF Research Database (Denmark)

    Hudson, Lawrence N; Newbold, Tim; Contu, Sara

    2017-01-01

    The PREDICTS project-Projecting Responses of Ecological Diversity In Changing Terrestrial Systems (www.predicts.org.uk)-has collated from published studies a large, reasonably representative database of comparable samples of biodiversity from multiple sites that differ in the nature or intensity ...

  16. Representing with Light. Video Projection Mapping for Cultural Heritage

    Science.gov (United States)

    Barbiani, C.; Guerra, F.; Pasini, T.; Visonà, M.

    2018-05-01

    In this paper, we describe a cross-disciplinary process that uses photogrammetric surveys as a precise basis for video projection mapping techniques. Beginning with a solid basis that uses geoinformatics technologies, such as laser scanning and photogrammetric survey, the method sets, as a first step, the physical and geometrical acquisition of the object. Precision and accuracy are the basics that allow the analysis of the artwork, both at a small or large scale, to evaluate details and correspondences. Testing contents at different scales of the object, using 3D printed replicas or real architectures is the second step of the investigation.The core of the process is the use of equations of collinearity into an interactive system such as Max 7, a visual programming language for music and multimedia, in order to facilitate operators to have a fast image correction, directly inside the interactive software. Interactivity gives also the opportunity to easily configure a set of actions to let the spectators to directly change and control the animation content. The paper goes through the different phases of the research, analysing the results and the progress through a series of events on real architecture and experiments on 3d printed models to test the level of involvement of the audience and the flexibility of the system in terms of content.The idea of using the collinearity equation inside da software Max 7 was developed for the M.Arch final Thesis by Massimo Visonà and Tommaso Pasini of the University of Venice (IUAV) in collaboration with the Digital Exhibit Postgraduate Master Course (MDE Iuav).

  17. Application of average adult Japanese voxel phantoms to evaluation of photon specific absorbed fractions

    International Nuclear Information System (INIS)

    Sato, Kaoru; Manabe, Kentaro; Endo, Akira

    2012-01-01

    Average adult Japanese male (JM-103) and female (JF-103) voxel (volume pixel) phantoms newly constructed at the Japan Atomic Energy Agency (JAEA) have average characteristics of body sizes and organ masses in adult Japanese. In JM-103 and JF-103, several organs and tissues were newly modeled for dose assessments based on tissue weighting factors of the 2007 Recommendations of the International Commission on Radiological Protection(ICRP). In this study, SAFs for thyroid, stomach, lungs and lymphatic nodes of JM-103 and JF-103 phantoms were calculated, and were compared with those of other adult Japanese phantoms based on individual medical images. In most cases, differences in SAFs between JM-103, JF-103 and other phantoms were about several tens percent, and was mainly attributed to mass differences of organs, tissues and contents. Therefore, it was concluded that SAFs of JM-103 and JF-103 represent those of average adult Japanese and that the two phantoms are applied to dose assessment for average adult Japanese on the basis of the 2007 Recommendations. (author)

  18. Evaluating the US government's crude oil price projections

    International Nuclear Information System (INIS)

    Williams, M.D.

    1992-01-01

    The U.S. Department of Energy's (DOE) 1991 official long run crude oil price projections are evaluated by comparing parameter averages for the forecast period (1991-2010) to parameter averages from crude oil price history (1859-1990). The parameters used in the evaluation are average price, average annual price changes, and average cycle duration (in years). All prices used in the analysis are annual prices in constant 1990 dollars per barrel. 13 figs

  19. Nine-year-old children use norm-based coding to visually represent facial expression.

    Science.gov (United States)

    Burton, Nichola; Jeffery, Linda; Skinner, Andrew L; Benton, Christopher P; Rhodes, Gillian

    2013-10-01

    Children are less skilled than adults at making judgments about facial expression. This could be because they have not yet developed adult-like mechanisms for visually representing faces. Adults are thought to represent faces in a multidimensional face-space, and have been shown to code the expression of a face relative to the norm or average face in face-space. Norm-based coding is economical and adaptive, and may be what makes adults more sensitive to facial expression than children. This study investigated the coding system that children use to represent facial expression. An adaptation aftereffect paradigm was used to test 24 adults and 18 children (9 years 2 months to 9 years 11 months old). Participants adapted to weak and strong antiexpressions. They then judged the expression of an average expression. Adaptation created aftereffects that made the test face look like the expression opposite that of the adaptor. Consistent with the predictions of norm-based but not exemplar-based coding, aftereffects were larger for strong than weak adaptors for both age groups. Results indicate that, like adults, children's coding of facial expressions is norm-based. PsycINFO Database Record (c) 2013 APA, all rights reserved.

  20. Hanford Environmental Dose Reconstruction Project

    International Nuclear Information System (INIS)

    McMakin, A.H.; Cannon, S.D.; Finch, S.M.

    1992-07-01

    The objective of the Hanford Environmental Dose Reconstruction (HEDR) Project is to estimate the radiation doses that individuals and populations could have received from nuclear operations at Hanford since 1944. The TSP consists of experts in environmental pathways, epidemiology, surface-water transport, ground-water transport, statistics, demography, agriculture, meteorology, nuclear engineering, radiation dosimetry, and cultural anthropology. Included are appointed technical members representing the states of Oregon, Washington, and Idaho, a representative of Native American tribes, and an individual representing the public. The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed from release to impact on humans (dose estimates): Source terms, environmental transport, environmental monitoring data, demography, food consumption, and agriculture, and environmental pathways and dose estimates. Progress is discussed

  1. Economics of biofiltration for remediation projects

    International Nuclear Information System (INIS)

    Yudelson, J.M.; Tinari, P.D.

    1995-01-01

    Biofilters with granular activated carbon (GAC) filter backup units offer substantial savings compared to conventional GAC filters and catalytic/thermal oxidation (Catox) units in controlling emissions of volatile organic compounds (VOCs) from petroleum remediation projects. Provided that the biofilter supplier is willing to satisfy the client's and consultant's risk-management concerns, biofilters offer anew method for reducing the cost of remediation projects, with savings of up to $10,000 (24%) per facility in 24-month projects and up to $16,000 (32%) per facility in 36-month projects for simple gas station remediation projects. Savings will be greater for longer projects and projects with higher average contaminant loadings

  2. Projections of Temperature-Attributable Premature Deaths in 209 U.S. Cities Using a Cluster-Based Poisson Approach

    Science.gov (United States)

    Schwartz, Joel D.; Lee, Mihye; Kinney, Patrick L.; Yang, Suijia; Mills, David; Sarofim, Marcus C.; Jones, Russell; Streeter, Richard; St. Juliana, Alexis; Peers, Jennifer; hide

    2015-01-01

    Background: A warming climate will affect future temperature-attributable premature deaths. This analysis is the first to project these deaths at a near national scale for the United States using city and month-specific temperature-mortality relationships. Methods: We used Poisson regressions to model temperature-attributable premature mortality as a function of daily average temperature in 209 U.S. cities by month. We used climate data to group cities into clusters and applied an Empirical Bayes adjustment to improve model stability and calculate cluster-based month-specific temperature-mortality functions. Using data from two climate models, we calculated future daily average temperatures in each city under Representative Concentration Pathway 6.0. Holding population constant at 2010 levels, we combined the temperature data and cluster-based temperature-mortality functions to project city-specific temperature-attributable premature deaths for multiple future years which correspond to a single reporting year. Results within the reporting periods are then averaged to account for potential climate variability and reported as a change from a 1990 baseline in the future reporting years of 2030, 2050 and 2100. Results: We found temperature-mortality relationships that vary by location and time of year. In general, the largest mortality response during hotter months (April - September) was in July in cities with cooler average conditions. The largest mortality response during colder months (October-March) was at the beginning (October) and end (March) of the period. Using data from two global climate models, we projected a net increase in premature deaths, aggregated across all 209 cities, in all future periods compared to 1990. However, the magnitude and sign of the change varied by cluster and city. Conclusions: We found increasing future premature deaths across the 209 modeled U.S. cities using two climate model projections, based on constant temperature

  3. Estimate of average glandular dose (AGD) in national clinics of mammography

    International Nuclear Information System (INIS)

    Mora, Patricia; Segura, Helena

    2004-01-01

    The breast cancer represents the second cause of death by cancer in the femme population of our country. The specialized equipment for the obtaining of the mammographic images is higher every day and its use increases daily. The quality of the radiographic study is linked to the dose that this tissue intrinsically sensible receives to the ionizing radiations. The present work makes the first national study to quantify the average glandular doses and to connect them with the diagnostic quality and the recommendations to international scale. (Author) [es

  4. Averaging in spherically symmetric cosmology

    International Nuclear Information System (INIS)

    Coley, A. A.; Pelavas, N.

    2007-01-01

    The averaging problem in cosmology is of fundamental importance. When applied to study cosmological evolution, the theory of macroscopic gravity (MG) can be regarded as a long-distance modification of general relativity. In the MG approach to the averaging problem in cosmology, the Einstein field equations on cosmological scales are modified by appropriate gravitational correlation terms. We study the averaging problem within the class of spherically symmetric cosmological models. That is, we shall take the microscopic equations and effect the averaging procedure to determine the precise form of the correlation tensor in this case. In particular, by working in volume-preserving coordinates, we calculate the form of the correlation tensor under some reasonable assumptions on the form for the inhomogeneous gravitational field and matter distribution. We find that the correlation tensor in a Friedmann-Lemaitre-Robertson-Walker (FLRW) background must be of the form of a spatial curvature. Inhomogeneities and spatial averaging, through this spatial curvature correction term, can have a very significant dynamical effect on the dynamics of the Universe and cosmological observations; in particular, we discuss whether spatial averaging might lead to a more conservative explanation of the observed acceleration of the Universe (without the introduction of exotic dark matter fields). We also find that the correlation tensor for a non-FLRW background can be interpreted as the sum of a spatial curvature and an anisotropic fluid. This may lead to interesting effects of averaging on astrophysical scales. We also discuss the results of averaging an inhomogeneous Lemaitre-Tolman-Bondi solution as well as calculations of linear perturbations (that is, the backreaction) in an FLRW background, which support the main conclusions of the analysis

  5. The database of the PREDICTS (Projecting Responses of Ecological Diversity In Changing Terrestrial Systems) project

    NARCIS (Netherlands)

    Hudson, Lawrence N; Newbold, Tim; Contu, Sara; Hill, Samantha L L; Lysenko, Igor; De Palma, Adriana; Phillips, Helen R P; Alhusseini, Tamera I; Bedford, Felicity E; Bennett, Dominic J; Booth, Hollie; Burton, Victoria J; Chng, Charlotte W T; Choimes, Argyrios; Correia, David L P; Day, Julie; Echeverría-Londoño, Susy; Emerson, Susan R; Gao, Di; Garon, Morgan; Harrison, Michelle L K; Ingram, Daniel J; Jung, Martin; Kemp, Victoria; Kirkpatrick, Lucinda; Martin, Callum D; Pan, Yuan; Pask-Hale, Gwilym D; Pynegar, Edwin L; Robinson, Alexandra N; Sanchez-Ortiz, Katia; Senior, Rebecca A; Simmons, Benno I; White, Hannah J; Zhang, Hanbin; Aben, Job; Abrahamczyk, Stefan; Adum, Gilbert B; Aguilar-Barquero, Virginia; Aizen, Marcelo A; Albertos, Belén; Alcala, E L; Del Mar Alguacil, Maria; Alignier, Audrey; Ancrenaz, Marc; Andersen, Alan N; Arbeláez-Cortés, Enrique; Armbrecht, Inge; Arroyo-Rodríguez, Víctor; Aumann, Tom; Axmacher, Jan C; Azhar, Badrul; Azpiroz, Adrián B; Baeten, Lander; Bakayoko, Adama; Báldi, András; Banks, John E; Baral, Sharad K; Barlow, Jos; Barratt, Barbara I P; Barrico, Lurdes; Bartolommei, Paola; Barton, Diane M; Basset, Yves; Batáry, Péter; Bates, Adam J; Baur, Bruno; Bayne, Erin M; Beja, Pedro; Benedick, Suzan; Berg, Åke; Bernard, Henry; Berry, Nicholas J; Bhatt, Dinesh; Bicknell, Jake E; Bihn, Jochen H; Blake, Robin J; Bobo, Kadiri S; Bóçon, Roberto; Boekhout, Teun; Böhning-Gaese, Katrin; Bonham, Kevin J; Borges, Paulo A V; Borges, Sérgio H; Boutin, Céline; Bouyer, Jérémy; Bragagnolo, Cibele; Brandt, Jodi S; Brearley, Francis Q; Brito, Isabel; Bros, Vicenç; Brunet, Jörg; Buczkowski, Grzegorz; Buddle, Christopher M; Bugter, Rob; Buscardo, Erika; Buse, Jörn; Cabra-García, Jimmy; Cáceres, Nilton C; Cagle, Nicolette L; Calviño-Cancela, María; Cameron, Sydney A; Cancello, Eliana M; Caparrós, Rut; Cardoso, Pedro; Carpenter, Dan; Carrijo, Tiago F; Carvalho, Anelena L; Cassano, Camila R; Castro, Helena; Castro-Luna, Alejandro A; Rolando, Cerda B; Cerezo, Alexis; Chapman, Kim Alan; Chauvat, Matthieu; Christensen, Morten; Clarke, Francis M; Cleary, Daniel F R; Colombo, Giorgio; Connop, Stuart P; Craig, Michael D; Cruz-López, Leopoldo; Cunningham, Saul A; D'Aniello, Biagio; D'Cruze, Neil; da Silva, Pedro Giovâni; Dallimer, Martin; Danquah, Emmanuel; Darvill, Ben; Dauber, Jens; Davis, Adrian L V; Dawson, Jeff; de Sassi, Claudio; de Thoisy, Benoit; Deheuvels, Olivier; Dejean, Alain; Devineau, Jean-Louis; Diekötter, Tim; Dolia, Jignasu V; Domínguez, Erwin; Dominguez-Haydar, Yamileth; Dorn, Silvia; Draper, Isabel; Dreber, Niels; Dumont, Bertrand; Dures, Simon G; Dynesius, Mats; Edenius, Lars; Eggleton, Paul; Eigenbrod, Felix; Elek, Zoltán; Entling, Martin H; Esler, Karen J; de Lima, Ricardo F; Faruk, Aisyah; Farwig, Nina; Fayle, Tom M; Felicioli, Antonio; Felton, Annika M; Fensham, Roderick J; Fernandez, Ignacio C; Ferreira, Catarina C; Ficetola, Gentile F; Fiera, Cristina; Filgueiras, Bruno K C; Fırıncıoğlu, Hüseyin K; Flaspohler, David; Floren, Andreas; Fonte, Steven J; Fournier, Anne; Fowler, Robert E; Franzén, Markus; Fraser, Lauchlan H; Fredriksson, Gabriella M; Freire, Geraldo B; Frizzo, Tiago L M; Fukuda, Daisuke; Furlani, Dario; Gaigher, René; Ganzhorn, Jörg U; García, Karla P; Garcia-R, Juan C; Garden, Jenni G; Garilleti, Ricardo; Ge, Bao-Ming; Gendreau-Berthiaume, Benoit; Gerard, Philippa J; Gheler-Costa, Carla; Gilbert, Benjamin; Giordani, Paolo; Giordano, Simonetta; Golodets, Carly; Gomes, Laurens G L; Gould, Rachelle K; Goulson, Dave; Gove, Aaron D; Granjon, Laurent; Grass, Ingo; Gray, Claudia L; Grogan, James; Gu, Weibin; Guardiola, Moisès; Gunawardene, Nihara R; Gutierrez, Alvaro G; Gutiérrez-Lamus, Doris L; Haarmeyer, Daniela H; Hanley, Mick E; Hanson, Thor; Hashim, Nor R; Hassan, Shombe N; Hatfield, Richard G; Hawes, Joseph E; Hayward, Matt W; Hébert, Christian; Helden, Alvin J; Henden, John-André; Henschel, Philipp; Hernández, Lionel; Herrera, James P; Herrmann, Farina; Herzog, Felix; Higuera-Diaz, Diego; Hilje, Branko; Höfer, Hubert; Hoffmann, Anke; Horgan, Finbarr G; Hornung, Elisabeth; Horváth, Roland; Hylander, Kristoffer; Isaacs-Cubides, Paola; Ishida, Hiroaki; Ishitani, Masahiro; Jacobs, Carmen T; Jaramillo, Víctor J; Jauker, Birgit; Hernández, F Jiménez; Johnson, McKenzie F; Jolli, Virat; Jonsell, Mats; Juliani, S Nur; Jung, Thomas S; Kapoor, Vena; Kappes, Heike; Kati, Vassiliki; Katovai, Eric; Kellner, Klaus; Kessler, Michael; Kirby, Kathryn R; Kittle, Andrew M; Knight, Mairi E; Knop, Eva; Kohler, Florian; Koivula, Matti; Kolb, Annette; Kone, Mouhamadou; Kőrösi, Ádám; Krauss, Jochen; Kumar, Ajith; Kumar, Raman; Kurz, David J; Kutt, Alex S; Lachat, Thibault; Lantschner, Victoria; Lara, Francisco; Lasky, Jesse R; Latta, Steven C; Laurance, William F; Lavelle, Patrick; Le Féon, Violette; LeBuhn, Gretchen; Légaré, Jean-Philippe; Lehouck, Valérie; Lencinas, María V; Lentini, Pia E; Letcher, Susan G; Li, Qi; Litchwark, Simon A; Littlewood, Nick A; Liu, Yunhui; Lo-Man-Hung, Nancy; López-Quintero, Carlos A; Louhaichi, Mounir; Lövei, Gabor L; Lucas-Borja, Manuel Esteban; Luja, Victor H; Luskin, Matthew S; MacSwiney G, M Cristina; Maeto, Kaoru; Magura, Tibor; Mallari, Neil Aldrin; Malone, Louise A; Malonza, Patrick K; Malumbres-Olarte, Jagoba; Mandujano, Salvador; Måren, Inger E; Marin-Spiotta, Erika; Marsh, Charles J; Marshall, E J P; Martínez, Eliana; Martínez Pastur, Guillermo; Moreno Mateos, David; Mayfield, Margaret M; Mazimpaka, Vicente; McCarthy, Jennifer L; McCarthy, Kyle P; McFrederick, Quinn S; McNamara, Sean; Medina, Nagore G; Medina, Rafael; Mena, Jose L; Mico, Estefania; Mikusinski, Grzegorz; Milder, Jeffrey C; Miller, James R; Miranda-Esquivel, Daniel R; Moir, Melinda L; Morales, Carolina L; Muchane, Mary N; Muchane, Muchai; Mudri-Stojnic, Sonja; Munira, A Nur; Muoñz-Alonso, Antonio; Munyekenye, B F; Naidoo, Robin; Naithani, A; Nakagawa, Michiko; Nakamura, Akihiro; Nakashima, Yoshihiro; Naoe, Shoji; Nates-Parra, Guiomar; Navarrete Gutierrez, Dario A; Navarro-Iriarte, Luis; Ndang'ang'a, Paul K; Neuschulz, Eike L; Ngai, Jacqueline T; Nicolas, Violaine; Nilsson, Sven G; Noreika, Norbertas; Norfolk, Olivia; Noriega, Jorge Ari; Norton, David A; Nöske, Nicole M; Nowakowski, A Justin; Numa, Catherine; O'Dea, Niall; O'Farrell, Patrick J; Oduro, William; Oertli, Sabine; Ofori-Boateng, Caleb; Oke, Christopher Omamoke; Oostra, Vicencio; Osgathorpe, Lynne M; Otavo, Samuel Eduardo; Page, Navendu V; Paritsis, Juan; Parra-H, Alejandro; Parry, Luke; Pe'er, Guy; Pearman, Peter B; Pelegrin, Nicolás; Pélissier, Raphaël; Peres, Carlos A; Peri, Pablo L; Persson, Anna S; Petanidou, Theodora; Peters, Marcell K; Pethiyagoda, Rohan S; Phalan, Ben; Philips, T Keith; Pillsbury, Finn C; Pincheira-Ulbrich, Jimmy; Pineda, Eduardo; Pino, Joan; Pizarro-Araya, Jaime; Plumptre, A J; Poggio, Santiago L; Politi, Natalia; Pons, Pere; Poveda, Katja; Power, Eileen F; Presley, Steven J; Proença, Vânia; Quaranta, Marino; Quintero, Carolina; Rader, Romina; Ramesh, B R; Ramirez-Pinilla, Martha P; Ranganathan, Jai; Rasmussen, Claus; Redpath-Downing, Nicola A; Reid, J Leighton; Reis, Yana T; Rey Benayas, José M; Rey-Velasco, Juan Carlos; Reynolds, Chevonne; Ribeiro, Danilo Bandini; Richards, Miriam H; Richardson, Barbara A; Richardson, Michael J; Ríos, Rodrigo Macip; Robinson, Richard; Robles, Carolina A; Römbke, Jörg; Romero-Duque, Luz Piedad; Rös, Matthias; Rosselli, Loreta; Rossiter, Stephen J; Roth, Dana S; Roulston, T'ai H; Rousseau, Laurent; Rubio, André V; Ruel, Jean-Claude; Sadler, Jonathan P; Sáfián, Szabolcs; Saldaña-Vázquez, Romeo A; Sam, Katerina; Samnegård, Ulrika; Santana, Joana; Santos, Xavier; Savage, Jade; Schellhorn, Nancy A; Schilthuizen, Menno; Schmiedel, Ute; Schmitt, Christine B; Schon, Nicole L; Schüepp, Christof; Schumann, Katharina; Schweiger, Oliver; Scott, Dawn M; Scott, Kenneth A; Sedlock, Jodi L; Seefeldt, Steven S; Shahabuddin, Ghazala; Shannon, Graeme; Sheil, Douglas; Sheldon, Frederick H; Shochat, Eyal; Siebert, Stefan J; Silva, Fernando A B; Simonetti, Javier A; Slade, Eleanor M; Smith, Jo; Smith-Pardo, Allan H; Sodhi, Navjot S; Somarriba, Eduardo J; Sosa, Ramón A; Soto Quiroga, Grimaldo; St-Laurent, Martin-Hugues; Starzomski, Brian M; Stefanescu, Constanti; Steffan-Dewenter, Ingolf; Stouffer, Philip C; Stout, Jane C; Strauch, Ayron M; Struebig, Matthew J; Su, Zhimin; Suarez-Rubio, Marcela; Sugiura, Shinji; Summerville, Keith S; Sung, Yik-Hei; Sutrisno, Hari; Svenning, Jens-Christian; Teder, Tiit; Threlfall, Caragh G; Tiitsaar, Anu; Todd, Jacqui H; Tonietto, Rebecca K; Torre, Ignasi; Tóthmérész, Béla; Tscharntke, Teja; Turner, Edgar C; Tylianakis, Jason M; Uehara-Prado, Marcio; Urbina-Cardona, Nicolas; Vallan, Denis; Vanbergen, Adam J; Vasconcelos, Heraldo L; Vassilev, Kiril; Verboven, Hans A F; Verdasca, Maria João; Verdú, José R; Vergara, Carlos H; Vergara, Pablo M; Verhulst, Jort; Virgilio, Massimiliano; Vu, Lien Van; Waite, Edward M; Walker, Tony R; Wang, Hua-Feng; Wang, Yanping; Watling, James I; Weller, Britta; Wells, Konstans; Westphal, Catrin; Wiafe, Edward D; Williams, Christopher D; Willig, Michael R; Woinarski, John C Z; Wolf, Jan H D; Wolters, Volkmar; Woodcock, Ben A; Wu, Jihua; Wunderle, Joseph M; Yamaura, Yuichi; Yoshikura, Satoko; Yu, Douglas W; Zaitsev, Andrey S; Zeidler, Juliane; Zou, Fasheng; Collen, Ben; Ewers, Rob M; Mace, Georgina M; Purves, Drew W; Scharlemann, Jörn P W; Purvis, Andy

    The PREDICTS project-Projecting Responses of Ecological Diversity In Changing Terrestrial Systems (www.predicts.org.uk)-has collated from published studies a large, reasonably representative database of comparable samples of biodiversity from multiple sites that differ in the nature or intensity of

  6. ANG coal gasification project management control system report. [Great Plains project

    Energy Technology Data Exchange (ETDEWEB)

    1981-01-01

    Much time, money and effort has been spent in the forefront of this project for project controls. The work breakdown structure for the systems has been custom designed. The systems, both manual and computerized, have been well scrutinized and chosen by ANG to represent the most cost effective and efficient way of controlling a project the magnitude of $1.5 billion. These systems have been developed in a manner so that information can be gathered as detailed or as summarized as necessary, and in the most timely and expeditious ways.

  7. The heterogeneous response method applied to couple the average pin cell and bulk moderator in cluster geometry

    International Nuclear Information System (INIS)

    Lerner, A.M.

    1986-01-01

    The first step towards evaluation of the neutron flux throughout a fuel cluster usually consists of obtaining the multigroup flux distribution in the average pin cell and in the circular outside system of shroud and bulk moderator. Here, an application of the so-called heterogeneous response method (HRM) is described to find this multigroup flux. The rather complex geometry is reduced to a microsystem, the average pin cell, and the outside or macrosystem of shroud and bulk moderator. In each of these systems, collision probabilities are used to obtain their response fluxes caused by sources and in-currents. The two systems are then coupled by cosine currents across that fraction of the average pin-cell boundary, called 'window', that represents the average common boundary between pin cells and the outside system. (author)

  8. Local-scale projections of coral reef futures and implications of the Paris Agreement.

    Science.gov (United States)

    van Hooidonk, Ruben; Maynard, Jeffrey; Tamelander, Jerker; Gove, Jamison; Ahmadia, Gabby; Raymundo, Laurie; Williams, Gareth; Heron, Scott F; Planes, Serge

    2016-12-21

    Increasingly frequent severe coral bleaching is among the greatest threats to coral reefs posed by climate change. Global climate models (GCMs) project great spatial variation in the timing of annual severe bleaching (ASB) conditions; a point at which reefs are certain to change and recovery will be limited. However, previous model-resolution projections (~1 × 1°) are too coarse to inform conservation planning. To meet the need for higher-resolution projections, we generated statistically downscaled projections (4-km resolution) for all coral reefs; these projections reveal high local-scale variation in ASB. Timing of ASB varies >10 years in 71 of the 87 countries and territories with >500 km 2 of reef area. Emissions scenario RCP4.5 represents lower emissions mid-century than will eventuate if pledges made following the 2015 Paris Climate Change Conference (COP21) become reality. These pledges do little to provide reefs with more time to adapt and acclimate prior to severe bleaching conditions occurring annually. RCP4.5 adds 11 years to the global average ASB timing when compared to RCP8.5; however, >75% of reefs still experience ASB before 2070 under RCP4.5. Coral reef futures clearly vary greatly among and within countries, indicating the projections warrant consideration in most reef areas during conservation and management planning.

  9. THE VALUE OF THE PROJECT, ITS HISTORICAL CONTEXT AND MEANINGFUL

    Directory of Open Access Journals (Sweden)

    Anastasiia Liezina

    2017-03-01

    Full Text Available In the article the historical stages of the category of “value” in various sciences, particularly in project management were considered. Chronology of evolutionary change and borrowing the term “value” in various sciences was developed. Existing approaches to determining the value of project management were analyzed. Author's interpretation of value projects was represented. Author's vision “iron” triangle of project management criteria based on the transformation of the classical approach was represented. Substance of the value of projects, taking into account the interests of all stakeholders, in terms of their value expectations was revealed. Key words: value, project management, project value, project stakeholders, the criteria for success of the project, “Iron” triangle.

  10. Projections of rapidly rising surface temperatures over Africa under low mitigation

    International Nuclear Information System (INIS)

    Engelbrecht, Francois; Bopape, Mary-Jane; Naidoo, Mogesh; Garland, Rebecca; Adegoke, Jimmy; Thatcher, Marcus; McGregor, John; Katzfey, Jack; Werner, Micha; Ichoku, Charles; Gatebe, Charles

    2015-01-01

    An analysis of observed trends in African annual-average near-surface temperatures over the last five decades reveals drastic increases, particularly over parts of the subtropics and central tropical Africa. Over these regions, temperatures have been rising at more than twice the global rate of temperature increase. An ensemble of high-resolution downscalings, obtained using a single regional climate model forced with the sea-surface temperatures and sea-ice fields of an ensemble of global circulation model (GCM) simulations, is shown to realistically represent the relatively strong temperature increases observed in subtropical southern and northern Africa. The amplitudes of warming are generally underestimated, however. Further warming is projected to occur during the 21st century, with plausible increases of 4–6 °C over the subtropics and 3–5 °C over the tropics by the end of the century relative to present-day climate under the A2 (a low mitigation) scenario of the Special Report on Emission Scenarios. High impact climate events such as heat-wave days and high fire-danger days are consistently projected to increase drastically in their frequency of occurrence. General decreases in soil-moisture availability are projected, even for regions where increases in rainfall are plausible, due to enhanced levels of evaporation. The regional dowscalings presented here, and recent GCM projections obtained for Africa, indicate that African annual-averaged temperatures may plausibly rise at about 1.5 times the global rate of temperature increase in the subtropics, and at a somewhat lower rate in the tropics. These projected increases although drastic, may be conservative given the model underestimations of observed temperature trends. The relatively strong rate of warming over Africa, in combination with the associated increases in extreme temperature events, may be key factors to consider when interpreting the suitability of global mitigation targets in terms of

  11. Hybrid Reynolds-Averaged/Large-Eddy Simulations of a Coaxial Supersonic Free-Jet Experiment

    Science.gov (United States)

    Baurle, Robert A.; Edwards, Jack R.

    2010-01-01

    Reynolds-averaged and hybrid Reynolds-averaged/large-eddy simulations have been applied to a supersonic coaxial jet flow experiment. The experiment was designed to study compressible mixing flow phenomenon under conditions that are representative of those encountered in scramjet combustors. The experiment utilized either helium or argon as the inner jet nozzle fluid, and the outer jet nozzle fluid consisted of laboratory air. The inner and outer nozzles were designed and operated to produce nearly pressure-matched Mach 1.8 flow conditions at the jet exit. The purpose of the computational effort was to assess the state-of-the-art for each modeling approach, and to use the hybrid Reynolds-averaged/large-eddy simulations to gather insight into the deficiencies of the Reynolds-averaged closure models. The Reynolds-averaged simulations displayed a strong sensitivity to choice of turbulent Schmidt number. The initial value chosen for this parameter resulted in an over-prediction of the mixing layer spreading rate for the helium case, but the opposite trend was observed when argon was used as the injectant. A larger turbulent Schmidt number greatly improved the comparison of the results with measurements for the helium simulations, but variations in the Schmidt number did not improve the argon comparisons. The hybrid Reynolds-averaged/large-eddy simulations also over-predicted the mixing layer spreading rate for the helium case, while under-predicting the rate of mixing when argon was used as the injectant. The primary reason conjectured for the discrepancy between the hybrid simulation results and the measurements centered around issues related to the transition from a Reynolds-averaged state to one with resolved turbulent content. Improvements to the inflow conditions were suggested as a remedy to this dilemma. Second-order turbulence statistics were also compared to their modeled Reynolds-averaged counterparts to evaluate the effectiveness of common turbulence closure

  12. Average cross sections calculated in various neutron fields

    International Nuclear Information System (INIS)

    Shibata, Keiichi

    2002-01-01

    Average cross sections have been calculated for the reactions contained in the dosimetry files, JENDL/D-99, IRDF-90V2, and RRDF-98 in order to select the best data for the new library IRDF-2002. The neutron spectra used in the calculations are as follows: 1) 252 Cf spontaneous fission spectrum (NBS evaluation), 2) 235 U thermal fission spectrum (NBS evaluation), 3) Intermediate-energy Standard Neutron Field (ISNF), 4) Coupled Fast Reactivity Measurement Facility (CFRMF), 5) Coupled thermal/fast uranium and boron carbide spherical assembly (ΣΣ), 6) Fast neutron source reactor (YAYOI), 7) Experimental fast reactor (JOYO), 8) Japan Material Testing Reactor (JMTR), 9) d-Li neutron spectrum with a 2-MeV deuteron beam. The items 3)-7) represent fast neutron spectra, while JMTR is a light water reactor. The Q-value for the d-Li reaction mentioned above is 15.02 MeV. Therefore, neutrons with energies up to 17 MeV can be produced in the d-Li reaction. The calculated average cross sections were compared with the measurements. Figures 1-9 show the ratios of the calculations to the experimental data which are given. It is found from these figures that the 58 Fe(n, γ) cross section in JENDL/D-99 reproduces the measurements in the thermal and fast reactor spectra better than that in IRDF-90V2. (author)

  13. A projection-free method for representing plane-wave DFT results in an atom-centered basis

    International Nuclear Information System (INIS)

    Dunnington, Benjamin D.; Schmidt, J. R.

    2015-01-01

    Plane wave density functional theory (DFT) is a powerful tool for gaining accurate, atomic level insight into bulk and surface structures. Yet, the delocalized nature of the plane wave basis set hinders the application of many powerful post-computation analysis approaches, many of which rely on localized atom-centered basis sets. Traditionally, this gap has been bridged via projection-based techniques from a plane wave to atom-centered basis. We instead propose an alternative projection-free approach utilizing direct calculation of matrix elements of the converged plane wave DFT Hamiltonian in an atom-centered basis. This projection-free approach yields a number of compelling advantages, including strict orthonormality of the resulting bands without artificial band mixing and access to the Hamiltonian matrix elements, while faithfully preserving the underlying DFT band structure. The resulting atomic orbital representation of the Kohn-Sham wavefunction and Hamiltonian provides a gateway to a wide variety of analysis approaches. We demonstrate the utility of the approach for a diverse set of chemical systems and example analysis approaches

  14. Fiscal years 2007 and 2008 : representative examples of completed RD&T projects and activities

    Science.gov (United States)

    2009-06-01

    The tables in this document correspond to major offices and initiatives at the Federal Highway Administration's Office of Research, Development, and Technology (RD&T), located at the Turner-Fairbank Highway Research Center. The tables provide represe...

  15. Evaluation on Collaborative Satisfaction for Project Management Team in Integrated Project Delivery Mode

    Science.gov (United States)

    Zhang, L.; Li, Y.; Wu, Q.

    2013-05-01

    Integrated Project Delivery (IPD) is a newly-developed project delivery approach for construction projects, and the level of collaboration of project management team is crucial to the success of its implementation. Existing research has shown that collaborative satisfaction is one of the key indicators of team collaboration. By reviewing the literature on team collaborative satisfaction and taking into consideration the characteristics of IPD projects, this paper summarizes the factors that influence collaborative satisfaction of IPD project management team. Based on these factors, this research develops a fuzzy linguistic method to effectively evaluate the level of team collaborative satisfaction, in which the authors adopted the 2-tuple linguistic variables and 2-tuple linguistic hybrid average operators to enhance the objectivity and accuracy of the evaluation. The paper demonstrates the practicality and effectiveness of the method through carrying out a case study with the method.

  16. A Prospective Analysis of Dynamic Loss of Breast Projection in Tissue Expander-Implant Reconstruction

    Directory of Open Access Journals (Sweden)

    Lauren M Mioton

    2015-05-01

    Full Text Available BackgroundBreast projection is a critical element of breast reconstruction aesthetics, but little has been published regarding breast projection as the firm expander is changed to a softer implant. Quantitative data representing this loss in projection may enhance patient education and improve our management of patient expectations.MethodsFemale patients who were undergoing immediate tissue-expander breast reconstruction with the senior author were enrolled in this prospective study. Three-dimensional camera software was used for all patient photographs and data analysis. Projection was calculated as the distance between the chest wall and the point of maximal projection of the breast form. Values were calculated for final tissue expander expansion and at varying intervals 3, 6, and 12 months after implant placement.ResultsFourteen breasts from 12 patients were included in the final analysis. Twelve of the 14 breasts had a loss of projection at three months following the implant placement or beyond. The percentage of projection lost in these 12 breasts ranged from 6.30% to 43.4%, with an average loss of projection of 21.05%.ConclusionsThis study is the first prospective quantitative analysis of temporal changes in breast projection after expander-implant reconstruction. By prospectively capturing projection data with three-dimensional photographic software, we reveal a loss of projection in this population by three months post-implant exchange. These findings will not only aid in managing patient expectations, but our methodology provides a foundation for future objective studies of the breast form.

  17. Organization of project management

    International Nuclear Information System (INIS)

    Schmidt, R.

    1975-01-01

    When speaking about interfaces within a project and their management, one has to understand and define what an interface is. In general, each component facing another one and each person working on a project with another person represents an interface. Therefore a project will consist practically in its entirety of interfaces with components and people sandwiched between them. This paper is limited to the most important interfaces with a focus on the problems occuring at them and their resolution. (orig.) [de

  18. The Average IQ of Sub-Saharan Africans: Comments on Wicherts, Dolan, and van der Maas

    Science.gov (United States)

    Lynn, Richard; Meisenberg, Gerhard

    2010-01-01

    Wicherts, Dolan, and van der Maas (2009) contend that the average IQ of sub-Saharan Africans is about 80. A critical evaluation of the studies presented by WDM shows that many of these are based on unrepresentative elite samples. We show that studies of 29 acceptably representative samples on tests other than the Progressive Matrices give a…

  19. Bond Formation in Diatomic Transition Metal Hydrides: Insights from the Analysis of Domain-Averaged Fermi Holes

    Czech Academy of Sciences Publication Activity Database

    Cooper, D.L.; Ponec, Robert

    2013-01-01

    Roč. 113, č. 2 (2013), s. 102-111 ISSN 0020-7608 R&D Projects: GA ČR GA203/09/0118 Institutional support: RVO:67985858 Keywords : transition metal hydrides * bond formation * analysis of domain averaged Fermi holes Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 1.166, year: 2013

  20. Assessing trends in observed and modelled climate extremes over Australia in relation to future projections

    International Nuclear Information System (INIS)

    Alexander, Lisa

    2007-01-01

    Full text: Nine global coupled climate models were assessed for their ability to reproduce observed trends in a set of indices representing temperature and precipitation extremes over Australia. Observed trends for 1957-1999 were compared with individual and multi-modelled trends calculated over the same period. When averaged across Australia the magnitude of trends and interannual variability of temperature extremes were well simulated by most models, particularly for the warm nights index. Except for consecutive dry days, the majority of models also reproduced the correct sign of trend for precipitation extremes. A bootstrapping technique was used to show that most models produce plausible trends when averaged over Australia, although only heavy precipitation days simulated from the multi-model ensemble showed significant skill at reproducing the observed spatial pattern of trends. Two of the models with output from different forcings showed that only with anthropogenic forcing included could the models capture the observed areally averaged trend for some of the temperature indices, but the forcing made little difference to the models' ability to reproduce the spatial pattern of trends over Australia. Future projected changes in extremes using three emissions scenarios were also analysed. Australia shows a shift towards significant warming of temperature extremes with much longer dry spells interspersed with periods of increased extreme precipitation irrespective of the scenario used. More work is required to determine whether regional projected changes over Australia are robust

  1. Overview of research in the ADVANTAGE project

    DEFF Research Database (Denmark)

    Angjelichinoski, Marko; Cosovic, Mirsad; Kalalas, Charalampos

    2016-01-01

    The European Marie Curie Project ADVANTAGE (Advanced Communications and Information processing in smart grid systems) was launched in 2014. It represents a major inter-disciplinary research project into the topic of Smart Grid technology. A key aspect of the project is to bring together and train...

  2. Lagrangian averaging with geodesic mean.

    Science.gov (United States)

    Oliver, Marcel

    2017-11-01

    This paper revisits the derivation of the Lagrangian averaged Euler (LAE), or Euler- α equations in the light of an intrinsic definition of the averaged flow map as the geodesic mean on the volume-preserving diffeomorphism group. Under the additional assumption that first-order fluctuations are statistically isotropic and transported by the mean flow as a vector field, averaging of the kinetic energy Lagrangian of an ideal fluid yields the LAE Lagrangian. The derivation presented here assumes a Euclidean spatial domain without boundaries.

  3. Modeling an Application's Theoretical Minimum and Average Transactional Response Times

    Energy Technology Data Exchange (ETDEWEB)

    Paiz, Mary Rose [Sandia National Lab. (SNL-NM), Albuquerque, NM (United States)

    2015-04-01

    The theoretical minimum transactional response time of an application serves as a ba- sis for the expected response time. The lower threshold for the minimum response time represents the minimum amount of time that the application should take to complete a transaction. Knowing the lower threshold is beneficial in detecting anomalies that are re- sults of unsuccessful transactions. On the converse, when an application's response time falls above an upper threshold, there is likely an anomaly in the application that is causing unusual performance issues in the transaction. This report explains how the non-stationary Generalized Extreme Value distribution is used to estimate the lower threshold of an ap- plication's daily minimum transactional response time. It also explains how the seasonal Autoregressive Integrated Moving Average time series model is used to estimate the upper threshold for an application's average transactional response time.

  4. Average thermal stress in the Al+SiC composite due to its manufacturing process

    International Nuclear Information System (INIS)

    Miranda, Carlos A.J.; Libardi, Rosani M.P.; Marcelino, Sergio; Boari, Zoroastro M.

    2013-01-01

    The numerical analyses framework to obtain the average thermal stress in the Al+SiC Composite due to its manufacturing process is presented along with the obtained results. The mixing of Aluminum and SiC powders is done at elevated temperature and the usage is at room temperature. A thermal stress state arises in the composite due to the different thermal expansion coefficients of the materials. Due to the particles size and randomness in the SiC distribution, some sets of models were analyzed and a statistical procedure used to evaluate the average stress state in the composite. In each model the particles position, form and size are randomly generated considering a volumetric ratio (VR) between 20% and 25%, close to an actual composite. The obtained stress field is represented by a certain number of iso stress curves, each one weighted by the area it represents. Systematically it was investigated the influence of: (a) the material behavior: linear x non-linear; (b) the carbide particles form: circular x quadrilateral; (c) the number of iso stress curves considered in each analysis; and (e) the model size (the number of particles). Each of above analyzed condition produced conclusions to guide the next step. Considering a confidence level of 95%, the average thermal stress value in the studied composite (20% ≤ VR ≤ 25%) is 175 MPa with a standard deviation of 10 MPa. Depending on its usage, this value should be taken into account when evaluating the material strength. (author)

  5. Systems Performance Analyses of Alaska Wind-Diesel Projects; Kotzebue, Alaska (Fact Sheet)

    Energy Technology Data Exchange (ETDEWEB)

    Baring-Gould, I.

    2009-04-01

    This fact sheet summarizes a systems performance analysis of the wind-diesel project in Kotzebue, Alaska. Data provided for this project include wind turbine output, average wind speed, average net capacity factor, and optimal net capacity factor based on Alaska Energy Authority wind data, estimated fuel savings, and wind system availability.

  6. Utilization of critical group and representative person methodologies: differences and difficulties

    International Nuclear Information System (INIS)

    Ferreira, Nelson L.D.; Rochedo, Elaine R.R.; Mazzilli, Barbara P.

    2013-01-01

    In Brazil, the assessment of the environmental impact due to routine discharges of radionuclides, which is used to the public protection, normally is based on the determination of the so-called 'critical group'. For the same purpose, the ICRP (2007) proposed the adoption of the 'representative person', defined as the individual receiving a dose representative of the members of the population who are subject to the higher exposures. In this work, are discussed, basically, the different characteristics of each one (critical group and representative person), related, mainly, to its methodologies and the necessary data demanded. Some difficulties to obtain site specific data, mainly habit data, as well as the way they are used, are discussed too. The critical group methodology uses, basically, average values, while the representative person methodology performs deterministic or probabilistic analysis using values obtained from distributions. As reference, it was considered the predicted effluents releases from Uranium Hexafluoride Production Plant (USEXA) and the effective doses calculated to the members of the previously defined critical group of Centro Experimental Aramar (CEA). (author)

  7. The average Indian female nose.

    Science.gov (United States)

    Patil, Surendra B; Kale, Satish M; Jaiswal, Sumeet; Khare, Nishant; Math, Mahantesh

    2011-12-01

    This study aimed to delineate the anthropometric measurements of the noses of young women of an Indian population and to compare them with the published ideals and average measurements for white women. This anthropometric survey included a volunteer sample of 100 young Indian women ages 18 to 35 years with Indian parents and no history of previous surgery or trauma to the nose. Standardized frontal, lateral, oblique, and basal photographs of the subjects' noses were taken, and 12 standard anthropometric measurements of the nose were determined. The results were compared with published standards for North American white women. In addition, nine nasal indices were calculated and compared with the standards for North American white women. The nose of Indian women differs significantly from the white nose. All the nasal measurements for the Indian women were found to be significantly different from those for North American white women. Seven of the nine nasal indices also differed significantly. Anthropometric analysis suggests differences between the Indian female nose and the North American white nose. Thus, a single aesthetic ideal is inadequate. Noses of Indian women are smaller and wider, with a less projected and rounded tip than the noses of white women. This study established the nasal anthropometric norms for nasal parameters, which will serve as a guide for cosmetic and reconstructive surgery in Indian women.

  8. Projected Heat Wave Characteristics over the Korean Peninsula During the Twenty-First Century

    Science.gov (United States)

    Shin, Jongsoo; Olson, Roman; An, Soon-Il

    2018-02-01

    Climate change is expected to increase temperatures globally, and consequently more frequent, longer, and hotter heat waves are likely to occur. Ambiguity in defining heat waves appropriately makes it difficult to compare changes in heat wave events over time. This study provides a quantitative definition of a heat wave and makes probabilistic heat wave projections for the Korean Peninsula under two global warming scenarios. Changes to heat waves under global warming are investigated using the representative concentration pathway 4.5 (RCP4.5) and 8.5 (RCP8.5) experiments from 30 coupled models participating in phase five of the Coupled Model Inter-comparison Project. Probabilistic climate projections from multi-model ensembles have been constructed using both simple and weighted averaging. Results from both methods are similar and show that heat waves will be more intense, frequent, and longer lasting. These trends are more apparent under the RCP8.5 scenario as compared to the RCP4.5 scenario. Under the RCP8.5 scenario, typical heat waves are projected to become stronger than any heat wave experienced in the recent measurement record. Furthermore, under this scenario, it cannot be ruled out that Korea will experience heat wave conditions spanning almost an entire summer before the end of the 21st century.

  9. Projected impacts of climate change on regional capacities for global plant species richness.

    Science.gov (United States)

    Sommer, Jan Henning; Kreft, Holger; Kier, Gerold; Jetz, Walter; Mutke, Jens; Barthlott, Wilhelm

    2010-08-07

    Climate change represents a major challenge to the maintenance of global biodiversity. To date, the direction and magnitude of net changes in the global distribution of plant diversity remain elusive. We use the empirical multi-variate relationships between contemporary water-energy dynamics and other non-climatic predictor variables to model the regional capacity for plant species richness (CSR) and its projected future changes. We find that across all analysed Intergovernmental Panel on Climate Change emission scenarios, relative changes in CSR increase with increased projected temperature rise. Between now and 2100, global average CSR is projected to remain similar to today (+0.3%) under the optimistic B1/+1.8 degrees C scenario, but to decrease significantly (-9.4%) under the 'business as usual' A1FI/+4.0 degrees C scenario. Across all modelled scenarios, the magnitude and direction of CSR change are geographically highly non-uniform. While in most temperate and arctic regions, a CSR increase is expected, the projections indicate a strong decline in most tropical and subtropical regions. Countries least responsible for past and present greenhouse gas emissions are likely to incur disproportionately large future losses in CSR, whereas industrialized countries have projected moderate increases. Independent of direction, we infer that all changes in regional CSR will probably induce on-site species turnover and thereby be a threat to native floras.

  10. USEPAM Project: The State of Progress

    DEFF Research Database (Denmark)

    Lund, Søren

    The paper represents the introductory speech at the 2nd regional workshop of the USEPAM project in Phnom Penh March 2005. It gives a review of the project's activities and achievements, and an introduction to the objectives, structure and programme of the workshop...

  11. Integral transforms of the quantum mechanical path integral: Hit function and path-averaged potential

    Science.gov (United States)

    Edwards, James P.; Gerber, Urs; Schubert, Christian; Trejo, Maria Anabel; Weber, Axel

    2018-04-01

    We introduce two integral transforms of the quantum mechanical transition kernel that represent physical information about the path integral. These transforms can be interpreted as probability distributions on particle trajectories measuring respectively the relative contribution to the path integral from paths crossing a given spatial point (the hit function) and the likelihood of values of the line integral of the potential along a path in the ensemble (the path-averaged potential).

  12. Distribution of average, marginal, and participation tax rates among Czech taxpayers: results from a TAXBEN model

    Czech Academy of Sciences Publication Activity Database

    Dušek, Libor; Kalíšková, Klára; Münich, Daniel

    2013-01-01

    Roč. 63, č. 6 (2013), s. 474-504 ISSN 0015-1920 R&D Projects: GA TA ČR(CZ) TD010033 Institutional support: RVO:67985998 Keywords : TAXBEN models * average tax rates * marginal tax rates Subject RIV: AH - Economics Impact factor: 0.358, year: 2013 http://journal.fsv.cuni.cz/storage/1287_dusek.pdf

  13. Distribution of average, marginal, and participation tax rates among Czech taxpayers: results from a TAXBEN model

    Czech Academy of Sciences Publication Activity Database

    Dušek, Libor; Kalíšková, Klára; Münich, Daniel

    2013-01-01

    Roč. 63, č. 6 (2013), s. 474-504 ISSN 0015-1920 R&D Projects: GA MŠk(CZ) SVV 267801/2013 Institutional support: PRVOUK-P23 Keywords : TAXBEN models * average tax rates * marginal tax rates Subject RIV: AH - Economics Impact factor: 0.358, year: 2013 http://journal.fsv.cuni.cz/storage/1287_dusek.pdf

  14. Spacetime averaging of exotic singularity universes

    International Nuclear Information System (INIS)

    Dabrowski, Mariusz P.

    2011-01-01

    Taking a spacetime average as a measure of the strength of singularities we show that big-rips (type I) are stronger than big-bangs. The former have infinite spacetime averages while the latter have them equal to zero. The sudden future singularities (type II) and w-singularities (type V) have finite spacetime averages. The finite scale factor (type III) singularities for some values of the parameters may have an infinite average and in that sense they may be considered stronger than big-bangs.

  15. Raspberry Pi projects

    CERN Document Server

    Robinson, Andrew

    2013-01-01

    Learn to build software and hardware projects featuring the Raspberry Pi! Raspberry Pi represents a new generation of computers that encourages the user to play and to learn and this unique book is aimed at the beginner Raspberry Pi user who is eager to get started creating real-world projects. Taking you on a journey of creating 15 practical projects, this fun and informative resource introduces you to the skills you need to have in order to make the most of the Pi. The book begins with a quick look at how to get the Pi up and running and then encourages you to dive into the array of exciti

  16. 30 March 2009 - Representatives of the Danish Council for Independent Research Natural Sciences visiting the LHC tunnel at Point 1 with Collaboration Spokesperson F. Gianotti, Former Spokesperson P. Jenni and Transition Radiation Tracker Project Leader C. Rembser.

    CERN Document Server

    Maximilien Brice

    2009-01-01

    30 March 2009 - Representatives of the Danish Council for Independent Research Natural Sciences visiting the LHC tunnel at Point 1 with Collaboration Spokesperson F. Gianotti, Former Spokesperson P. Jenni and Transition Radiation Tracker Project Leader C. Rembser.

  17. A comparative analysis of 9 multi-model averaging approaches in hydrological continuous streamflow simulation

    Science.gov (United States)

    Arsenault, Richard; Gatien, Philippe; Renaud, Benoit; Brissette, François; Martel, Jean-Luc

    2015-10-01

    This study aims to test whether a weighted combination of several hydrological models can simulate flows more accurately than the models taken individually. In addition, the project attempts to identify the most efficient model averaging method and the optimal number of models to include in the weighting scheme. In order to address the first objective, streamflow was simulated using four lumped hydrological models (HSAMI, HMETS, MOHYSE and GR4J-6), each of which were calibrated with three different objective functions on 429 watersheds. The resulting 12 hydrographs (4 models × 3 metrics) were weighted and combined with the help of 9 averaging methods which are the simple arithmetic mean (SAM), Akaike information criterion (AICA), Bates-Granger (BGA), Bayes information criterion (BICA), Bayesian model averaging (BMA), Granger-Ramanathan average variant A, B and C (GRA, GRB and GRC) and the average by SCE-UA optimization (SCA). The same weights were then applied to the hydrographs in validation mode, and the Nash-Sutcliffe Efficiency metric was measured between the averaged and observed hydrographs. Statistical analyses were performed to compare the accuracy of weighted methods to that of individual models. A Kruskal-Wallis test and a multi-objective optimization algorithm were then used to identify the most efficient weighted method and the optimal number of models to integrate. Results suggest that the GRA, GRB, GRC and SCA weighted methods perform better than the individual members. Model averaging from these four methods were superior to the best of the individual members in 76% of the cases. Optimal combinations on all watersheds included at least one of each of the four hydrological models. None of the optimal combinations included all members of the ensemble of 12 hydrographs. The Granger-Ramanathan average variant C (GRC) is recommended as the best compromise between accuracy, speed of execution, and simplicity.

  18. Real-time volumetric image reconstruction and 3D tumor localization based on a single x-ray projection image for lung cancer radiotherapy.

    Science.gov (United States)

    Li, Ruijiang; Jia, Xun; Lewis, John H; Gu, Xuejun; Folkerts, Michael; Men, Chunhua; Jiang, Steve B

    2010-06-01

    To develop an algorithm for real-time volumetric image reconstruction and 3D tumor localization based on a single x-ray projection image for lung cancer radiotherapy. Given a set of volumetric images of a patient at N breathing phases as the training data, deformable image registration was performed between a reference phase and the other N-1 phases, resulting in N-1 deformation vector fields (DVFs). These DVFs can be represented efficiently by a few eigenvectors and coefficients obtained from principal component analysis (PCA). By varying the PCA coefficients, new DVFs can be generated, which, when applied on the reference image, lead to new volumetric images. A volumetric image can then be reconstructed from a single projection image by optimizing the PCA coefficients such that its computed projection matches the measured one. The 3D location of the tumor can be derived by applying the inverted DVF on its position in the reference image. The algorithm was implemented on graphics processing units (GPUs) to achieve real-time efficiency. The training data were generated using a realistic and dynamic mathematical phantom with ten breathing phases. The testing data were 360 cone beam projections corresponding to one gantry rotation, simulated using the same phantom with a 50% increase in breathing amplitude. The average relative image intensity error of the reconstructed volumetric images is 6.9% +/- 2.4%. The average 3D tumor localization error is 0.8 +/- 0.5 mm. On an NVIDIA Tesla C1060 GPU card, the average computation time for reconstructing a volumetric image from each projection is 0.24 s (range: 0.17 and 0.35 s). The authors have shown the feasibility of reconstructing volumetric images and localizing tumor positions in 3D in near real-time from a single x-ray image.

  19. Novel relations between the ergodic capacity and the average bit error rate

    KAUST Repository

    Yilmaz, Ferkan

    2011-11-01

    Ergodic capacity and average bit error rate have been widely used to compare the performance of different wireless communication systems. As such recent scientific research and studies revealed strong impact of designing and implementing wireless technologies based on these two performance indicators. However and to the best of our knowledge, the direct links between these two performance indicators have not been explicitly proposed in the literature so far. In this paper, we propose novel relations between the ergodic capacity and the average bit error rate of an overall communication system using binary modulation schemes for signaling with a limited bandwidth and operating over generalized fading channels. More specifically, we show that these two performance measures can be represented in terms of each other, without the need to know the exact end-to-end statistical characterization of the communication channel. We validate the correctness and accuracy of our newly proposed relations and illustrated their usefulness by considering some classical examples. © 2011 IEEE.

  20. Changing the Project Execution Culture at NASA Dryden

    Science.gov (United States)

    Horn, Thomas J.

    2012-01-01

    Dryden has embarked on implementing Critical Chain Project Management (CCPM) philosophies and tools to reduce workforce stress and increase the centers work throughput. This effort has been under way for over one year and represents a fundamental state change in how various projects are planned and executed at the center. The implementation of CCPM philosophies and the required cultural changes represent the most difficult aspects of the implementation.

  1. Status of CSNS project

    International Nuclear Information System (INIS)

    Zhang, J.; Fu, S.N.; Chen, H.S.

    2015-01-01

    The China Spallation Neutron Source (CSNS) accelerator is designed to accelerate proton beam pulses to 1.6 GeV at 25 Hz repetition rate, striking a solid metal target to produce spallation neutrons. The accelerator provides a beam power of 100 kW on the target in the first phase and then 500 kW in the second phase by increasing the average beam intensity 5 times while raising the linac output energy. The project construction has been formally launched in 2011 and it is planned to complete the project in March 2018. It is one of the high intensity proton accelerator projects in the world and it imposes a great challenge to Chinese accelerator community. This presentation will cover the status and challenges of the CSNS project. (author)

  2. The Stripa project. Annual report 1990

    International Nuclear Information System (INIS)

    1991-07-01

    The Stripa project is an international project being performed under the sponsorship of the OECD Nuclear Energy Agency (NEA). The project concerns research related to the disposal of highly radioactive waste in crystalline rock. The research and development division of the Swedish Nuclear Fuel and Waste Management Co. (SKB) has been entrusted with the management of the project, under the direction of representatives from each participating country. The aim of this report is to inform the OECD Nuclear Energy Agency and the participants in the project about the general progress of work during 1990

  3. The Stripa project annual report 1991

    International Nuclear Information System (INIS)

    1992-05-01

    The Stripa project is an international project being performed under the sponsorship of the OECD Nuclear Energy Agency (NEA). The project concerns research related to the disposal of highly radioactive waste in crystalline rock. The Research and Development Division of the Swedish Nuclear Fuel and Waste Management Company (SKB) has been entrusted with the management of the project, under the direction of representatives from each participating country. The aim of this report is to inform the OECD Nuclear Energy Agency and the participants in the project about the general progress of work during 1991

  4. Coordinated research projects (CRP). Coordinated research project (CRP)

    International Nuclear Information System (INIS)

    Takagi, Hidekazu; Koike, Fumihiro; Nakamura, Nobuyuki

    2013-01-01

    In the present paper, the contribution of Japanese scientists in coordinated research projects on thermonuclear fusion. Representative subjects taken in seven projects are the precise computation of theoretical data on electron-molecule collisions in the peripheral plasma, the computation of spectroscopic data of multi-charged tungsten ions, the spectroscopic measurement of multi-charged tungsten ions using an ion trap device, the development of collisional-radiative model for plasmas including hydrogen and helium, the computational and theoretical studies on the behavior of tungsten and beryllium in the plasma-wall interaction, the study on the property of dusts generated in fusion devices. These subjects are those of most important issues in ITER. (author)

  5. Permitted Marine Hydrokinetic Projects

    Data.gov (United States)

    National Oceanic and Atmospheric Administration, Department of Commerce — This data represents pending or issued preliminary permits or issued licenses for marine hydrokinetic projects that produce energy from waves or directly from the...

  6. Practical considerations of reservoir heterogeneities on SAGD projects

    Energy Technology Data Exchange (ETDEWEB)

    Baker, R.; Fong, C.; Li, T. [Epic Consulting Services Ltd., Calgary, AB (Canada); Bowes, C.; Toews, M. [Calgary Univ., AB (Canada)

    2008-10-15

    Significant emphasis has been placed on developing cost-effective strategies for the production of large heavy oil and bitumen reserves located in western Canada and around the world. An effective method that has been proven to be effective in this regard is steam-assisted gravity drainage (SAGD). However, determining the optimum and cost-effective strategy is a challenge to any SAGD reservoir. Average rock quality and reservoir heterogeneities have a significant impact on steam chamber development and the overall volumetric sweep. As well, the approach to SAGD simulation varies as heterogeneity changes. This paper examined two well pairs with different degrees of heterogeneity in the Surmont pilot project. The paper also addressed potential geological risk through analogy and the amount of heterogeneity that must be accounted for when developing a representative simulation. The paper provided background information on the Surmont pilot project, which consists of three horizontal SAGD well pairs in the Athabasca oil sands of northeast Alberta. The reservoir simulation model was then described. Results and conclusions were offered. It was concluded that careful production controls and strategy must be applied particular to the reservoir to ensure that the SAGD well pairs were capable of draining the mobilized oil. 5 refs., 1 tab., 25 figs.

  7. The true bladder dose: on average thrice higher than the ICRU reference

    International Nuclear Information System (INIS)

    Barillot, I.; Horiot, J.C.; Maingon, P.; Bone-Lepinoy, M.C.; D'Hombres, A.; Comte, J.; Delignette, A.; Feutray, S.; Vaillant, D.

    1996-01-01

    The aim of this study is to compare ICRU dose to doses at the bladder base located from ultrasonography measurements. Since 1990, the dose delivered to the bladder during utero-vaginal brachytherapy was systematically calculated at 3 or 4 points representative of bladder base determined with ultrasonography. The ICRU Reference Dose (IRD) from films, the Maximum Dose (Dmax), the Mean Dose (Dmean) representative of the dose received by a large area of bladder mucosa, the Reference Dose Rate (RDR) and the Mean Dose Rate (MDR) were recorded. Material: from 1990 to 1994, 198 measurements were performed in 152 patients. 98 patients were treated for cervix carcinomas, 54 for endometrial carcinomas. Methods: Bladder complications were classified using French Italian Syllabus. The influence of doses and dose rates on complications were tested using non parametric t test. Results: On average IRD is 21 Gy +/- 12 Gy, Dmax is 51Gy +/- 21Gy, Dmean is 40 Gy +/16 Gy. On average Dmax is thrice higher than IRD and Dmean twice higher than IRD. The same results are obtained for cervix and endometrium. Comparisons on dose rates were also performed: MDR is on average twice higher than RDR (RDR 48 cGy/h vs MDR 88 cGy/h). The five observed complications consist of incontinence only (3 G1, 1G2, 1G3). They are only statistically correlated with RDR p=0.01 (46 cGy/h in patients without complications vs 74 cGy/h in patients with complications). However the full responsibility of RT remains doubtful and should be shared with surgery in all cases. In summary: Bladder mucosa seems to tolerate well much higher doses than previous recorded without increased risk of severe sequelae. However this finding is probably explained by our efforts to spare most of bladder mucosa by 1 deg. ) customised external irradiation therapy (4 fields, full bladder) 2 deg. ) reproduction of physiologic bladder filling during brachytherapy by intermittent clamping of the Foley catheter

  8. Projected 21st century decrease in marine productivity: a multi-model analysis

    Directory of Open Access Journals (Sweden)

    M. Steinacher

    2010-03-01

    Full Text Available Changes in marine net primary productivity (PP and export of particulate organic carbon (EP are projected over the 21st century with four global coupled carbon cycle-climate models. These include representations of marine ecosystems and the carbon cycle of different structure and complexity. All four models show a decrease in global mean PP and EP between 2 and 20% by 2100 relative to preindustrial conditions, for the SRES A2 emission scenario. Two different regimes for productivity changes are consistently identified in all models. The first chain of mechanisms is dominant in the low- and mid-latitude ocean and in the North Atlantic: reduced input of macro-nutrients into the euphotic zone related to enhanced stratification, reduced mixed layer depth, and slowed circulation causes a decrease in macro-nutrient concentrations and in PP and EP. The second regime is projected for parts of the Southern Ocean: an alleviation of light and/or temperature limitation leads to an increase in PP and EP as productivity is fueled by a sustained nutrient input. A region of disagreement among the models is the Arctic, where three models project an increase in PP while one model projects a decrease. Projected changes in seasonal and interannual variability are modest in most regions. Regional model skill metrics are proposed to generate multi-model mean fields that show an improved skill in representing observation-based estimates compared to a simple multi-model average. Model results are compared to recent productivity projections with three different algorithms, usually applied to infer net primary production from satellite observations.

  9. Average gluon and quark jet multiplicities at higher orders

    Energy Technology Data Exchange (ETDEWEB)

    Bolzoni, Paolo; Kniehl, Bernd A. [Hamburg Univ. (Germany). 2. Inst. fuer Theoretische Physik; Kotikov, Anatoly V. [Hamburg Univ. (Germany). 2. Inst. fuer Theoretische Physik; Joint Institute of Nuclear Research, Moscow (Russian Federation). Bogoliubov Lab. of Theoretical Physics

    2013-05-15

    We develop a new formalism for computing and including both the perturbative and nonperturbative QCD contributions to the scale evolution of average gluon and quark jet multiplicities. The new method is motivated by recent progress in timelike small-x resummation obtained in the MS factorization scheme. We obtain next-to-next-to-leading-logarithmic (NNLL) resummed expressions, which represent generalizations of previous analytic results. Our expressions depend on two nonperturbative parameters with clear and simple physical interpretations. A global fit of these two quantities to all available experimental data sets that are compatible with regard to the jet algorithms demonstrates by its goodness how our results solve a longstanding problem of QCD. We show that the statistical and theoretical uncertainties both do not exceed 5% for scales above 10 GeV. We finally propose to use the jet multiplicity data as a new way to extract the strong-coupling constant. Including all the available theoretical input within our approach, we obtain {alpha}{sub s}{sup (5)}(M{sub Z})=0.1199{+-}0.0026 in the MS scheme in an approximation equivalent to next-to-next-to-leading order enhanced by the resummations of ln(x) terms through the NNLL level and of ln Q{sup 2} terms by the renormalization group, in excellent agreement with the present world average.

  10. The Challenges and Enhancing Opportunities of Global Project Management: Evidence from Chinese and Dutch Cross-Cultural Project Management

    OpenAIRE

    Zhang, Ying; Marquis, Christopher G; Filippov, Sergey; Haasnoot, Henk-Jan; van der Steen, Martijn

    2015-01-01

    This study investigates the role of national and organisational culture in day-to-day activities of multinational project teams, specifically focusing on differences between Chinese and Dutch project managers. We rely on fieldwork observation and interviews with representatives from a diverse set of organizations in China and the Netherlands. Analyses focus on the impact of cultural differences on five project management processes – (1) project planning, (2) cost and quality management, (3) r...

  11. Space-Varying Iterative Restoration of Diffuse Optical Tomograms Reconstructed by the Photon Average Trajectories Method

    Directory of Open Access Journals (Sweden)

    Kravtsenyuk Olga V

    2007-01-01

    Full Text Available The possibility of improving the spatial resolution of diffuse optical tomograms reconstructed by the photon average trajectories (PAT method is substantiated. The PAT method recently presented by us is based on a concept of an average statistical trajectory for transfer of light energy, the photon average trajectory (PAT. The inverse problem of diffuse optical tomography is reduced to a solution of an integral equation with integration along a conditional PAT. As a result, the conventional algorithms of projection computed tomography can be used for fast reconstruction of diffuse optical images. The shortcoming of the PAT method is that it reconstructs the images blurred due to averaging over spatial distributions of photons which form the signal measured by the receiver. To improve the resolution, we apply a spatially variant blur model based on an interpolation of the spatially invariant point spread functions simulated for the different small subregions of the image domain. Two iterative algorithms for solving a system of linear algebraic equations, the conjugate gradient algorithm for least squares problem and the modified residual norm steepest descent algorithm, are used for deblurring. It is shown that a gain in spatial resolution can be obtained.

  12. Space-Varying Iterative Restoration of Diffuse Optical Tomograms Reconstructed by the Photon Average Trajectories Method

    Directory of Open Access Journals (Sweden)

    Vladimir V. Lyubimov

    2007-01-01

    Full Text Available The possibility of improving the spatial resolution of diffuse optical tomograms reconstructed by the photon average trajectories (PAT method is substantiated. The PAT method recently presented by us is based on a concept of an average statistical trajectory for transfer of light energy, the photon average trajectory (PAT. The inverse problem of diffuse optical tomography is reduced to a solution of an integral equation with integration along a conditional PAT. As a result, the conventional algorithms of projection computed tomography can be used for fast reconstruction of diffuse optical images. The shortcoming of the PAT method is that it reconstructs the images blurred due to averaging over spatial distributions of photons which form the signal measured by the receiver. To improve the resolution, we apply a spatially variant blur model based on an interpolation of the spatially invariant point spread functions simulated for the different small subregions of the image domain. Two iterative algorithms for solving a system of linear algebraic equations, the conjugate gradient algorithm for least squares problem and the modified residual norm steepest descent algorithm, are used for deblurring. It is shown that a 27% gain in spatial resolution can be obtained.

  13. Integrated Project Teams - An Essential Element of Project Management during Project Planning and Execution - 12155

    Energy Technology Data Exchange (ETDEWEB)

    Burritt, James G.; Berkey, Edgar [Longenecker and Associates, Las Vegas, NV 89135 (United States)

    2012-07-01

    Managing complex projects requires a capable, effective project manager to be in place, who is assisted by a team of competent assistants in various relevant disciplines. This team of assistants is known as the Integrated Project Team (IPT). he IPT is composed of a multidisciplinary group of people who are collectively responsible for delivering a defined project outcome and who plan, execute, and implement over the entire life-cycle of a project, which can be a facility being constructed or a system being acquired. An ideal IPT includes empowered representatives from all functional areas involved with a project-such as engineering design, technology, manufacturing, test and evaluation, contracts, legal, logistics, and especially, the customer. Effective IPTs are an essential element of scope, cost, and schedule control for any complex, large construction project, whether funded by DOE or another organization. By recently assessing a number of major, on-going DOE waste management projects, the characteristics of high performing IPTs have been defined as well as the reasons for potential IPT failure. Project managers should use IPTs to plan and execute projects, but the IPTs must be properly constituted and the members capable and empowered. For them to be effective, the project manager must select the right team, and provide them with the training and guidance for them to be effective. IPT members must treat their IPT assignment as a primary duty, not some ancillary function. All team members must have an understanding of the factors associated with successful IPTs, and the reasons that some IPTs fail. Integrated Project Teams should be used by both government and industry. (authors)

  14. Major Events Coordinated Security Solutions: The Application of the Project Management Body of Knowledge for Managing a Science and Technology Project

    Science.gov (United States)

    2011-02-01

    represents the sum of knowledge related to project management and includes best practises and techniques generally accepted by the project...project has demonstrated that the principles of project management can be applied even to those activities that naturally resist the structure and...team and to engender trust in the output. DRDC believes strongly in the application of sound project management principles and uses an abbreviated

  15. Design and evaluation of representative indoor radon surveys

    International Nuclear Information System (INIS)

    Csige, I.; Csegzi, S.

    2004-01-01

    We have developed a procedure to design and evaluate representative indoor radon surveys. The procedure is based on random sampling of a population of houses and careful statistical analysis of measured indoor radon concentrations. The method is designed to estimate the fraction of houses in which annual average 222 Rn activity concentration may exceed a certain reference level. Measurements of annual average indoor 222 Rn activity concentration were done in sleeping rooms at pillow level using etched track type radon detectors. We applied the above procedure in an old fashioned village and in a fast developing small city in Transylvania, Romania. In the village almost all houses were single floor wooden made houses without cellar built with traditional technology on a geologically uniform area. The distribution of indoor 222 Rn activity concentration in a sample of 115 houses can almost perfectly be fitted with log-normal probability density function. The correlation coefficient of linear fitting on linearized scales was k = -0.9980. The percentages of houses expected to have annual average 222 Rn activity concentration higher than 400 Bq m -3 is less than 1 %, and of those higher than 600 Bq m -3 can be estimated to be around 0.1 %. The small city, on the other hand lies on a geologically inhomogeneous area, and house construction technology has also changed dramatically in past decades. The resulting distribution of measured indoor 222 Rn activity concentration in a sample of 116 houses cannot be fitted with any simple probability density function. Therefore the prediction of the fraction of houses in which the annual average 222 Rn activity concentration may exceed a certain reference level could not be done adequately. With certain assumptions we estimated that the percentages of houses expected to have annual average 222 Rn activity concentration higher than 400 Bq m -3 is between 3 and 7 %, and of those higher than 600 Bq m -3 can be estimated to be between

  16. Using SNOMED CT to represent two interface terminologies.

    Science.gov (United States)

    Rosenbloom, S Trent; Brown, Steven H; Froehling, David; Bauer, Brent A; Wahner-Roedler, Dietlind L; Gregg, William M; Elkin, Peter L

    2009-01-01

    Interface terminologies are designed to support interactions between humans and structured medical information. In particular, many interface terminologies have been developed for structured computer based documentation systems. Experts and policy-makers have recommended that interface terminologies be mapped to reference terminologies. The goal of the current study was to evaluate how well the reference terminology SNOMED CT could map to and represent two interface terminologies, MEDCIN and the Categorical Health Information Structured Lexicon (CHISL). Automated mappings between SNOMED CT and 500 terms from each of the two interface terminologies were evaluated by human reviewers, who also searched SNOMED CT to identify better mappings when this was judged to be necessary. Reviewers judged whether they believed the interface terms to be clinically appropriate, whether the terms were covered by SNOMED CT concepts and whether the terms' implied semantic structure could be represented by SNOMED CT. Outcomes included concept coverage by SNOMED CT for study terms and their implied semantics. Agreement statistics and compositionality measures were calculated. The SNOMED CT terminology contained concepts to represent 92.4% of MEDCIN and 95.9% of CHISL terms. Semantic structures implied by study terms were less well covered, with some complex compositional expressions requiring semantics not present in SNOMED CT. Among sampled terms, those from MEDCIN were more complex than those from CHISL, containing an average 3.8 versus 1.8 atomic concepts respectively, pterms.

  17. Research Planning and Evaluation Project

    International Nuclear Information System (INIS)

    Song, Seunghyun; Kim, Doyang; Ryu, Byunghoon; Lim, Chaeyoung; Song, Leeyoung; Lee, Youngchul; Han, Changsun; Kim, Hackchoon

    2011-12-01

    - To activate R and D through a systematic and impartial evaluation by using information on efficient distribution of research resource, setting project priorities, and measuring achievement against goals produced after research on planning and evaluation system for the government-funded project for KAERI was conducted. - Nuclear R and D project is the representative national R and D project which has been implemented in Korea. For the sustainable development of nuclear energy which supplies about 40% of total electricity generation and the enhancement of it innovative ability in the future, a systematic and efficient strategy in the planning stage is required

  18. Understanding errors in EIA projections of energy demand

    Energy Technology Data Exchange (ETDEWEB)

    Fischer, Carolyn; Herrnstadt, Evan; Morgenstern, Richard [Resources for the Future, 1616 P St. NW, Washington, DC 20036 (United States)

    2009-08-15

    This paper investigates the potential for systematic errors in the Energy Information Administration's (EIA) widely used Annual Energy Outlook, focusing on the near- to mid-term projections of energy demand. Based on analysis of the EIA's 22-year projection record, we find a fairly modest but persistent tendency to underestimate total energy demand by an average of 2 percent per year after controlling for projection errors in gross domestic product, oil prices, and heating/cooling degree days. For 14 individual fuels/consuming sectors routinely reported by the EIA, we observe a great deal of directional consistency in the errors over time, ranging up to 7 percent per year. Electric utility renewables, electric utility natural gas, transportation distillate, and residential electricity show significant biases on average. Projections for certain other sectors have significant unexplained errors for selected time horizons. Such independent evaluation can be useful for validating analytic efforts and for prioritizing future model revisions. (author)

  19. Shape and depth determinations from second moving average residual self-potential anomalies

    International Nuclear Information System (INIS)

    Abdelrahman, E M; El-Araby, T M; Essa, K S

    2009-01-01

    We have developed a semi-automatic method to determine the depth and shape (shape factor) of a buried structure from second moving average residual self-potential anomalies obtained from observed data using filters of successive window lengths. The method involves using a relationship between the depth and the shape to source and a combination of windowed observations. The relationship represents a parametric family of curves (window curves). For a fixed window length, the depth is determined for each shape factor. The computed depths are plotted against the shape factors, representing a continuous monotonically increasing curve. The solution for the shape and depth is read at the common intersection of the window curves. The validity of the method is tested on a synthetic example with and without random errors and on two field examples from Turkey and Germany. In all cases examined, the depth and the shape solutions obtained are in very good agreement with the true ones

  20. Project control - the next generation

    International Nuclear Information System (INIS)

    Iorii, V.F.; McKinnon, B.L.

    1993-01-01

    The Yucca Mountain Site Characterization Project (YMP) is the U.S. Department of Energy's (DOE) second largest Major System Acquisition Project. We have developed an integrated planning and control system (called PACS) that we believe represents the 'Next Generation' in project control. PACS integrates technical scope, cost, and schedule information for over 50 participating organizations and produces performances measurement reports for science and engineering managers at all levels. Our 'Next Generation' project control too, PACS, has been found to be in compliance with the new DOE Project Control System Guidelines. Additionally, the nuclear utility oversight group of the Edison Electric Institute has suggested PACS be used as a model for other civilian radioactive waste management projects. A 'Next Generation' project control tool will be necessary to do science in the 21st century

  1. SIYAZAMA ENTREPRENEURIAL DEVELOPMENT PROJECT ...

    African Journals Online (AJOL)

    course of planning, implementation and evaluation are presented. Identification of challenges ..... 2012. 113. Most participants in the Siyazama project live in informal housing with an average of six ..... The contributions of educational psychology to school psychology. In: T.B. Gutkin & C.R. Reynolds (eds.), The handbook of ...

  2. Parents' Reactions to Finding Out That Their Children Have Average or above Average IQ Scores.

    Science.gov (United States)

    Dirks, Jean; And Others

    1983-01-01

    Parents of 41 children who had been given an individually-administered intelligence test were contacted 19 months after testing. Parents of average IQ children were less accurate in their memory of test results. Children with above average IQ experienced extremely low frequencies of sibling rivalry, conceit or pressure. (Author/HLM)

  3. Evaluations of average level spacings

    International Nuclear Information System (INIS)

    Liou, H.I.

    1980-01-01

    The average level spacing for highly excited nuclei is a key parameter in cross section formulas based on statistical nuclear models, and also plays an important role in determining many physics quantities. Various methods to evaluate average level spacings are reviewed. Because of the finite experimental resolution, to detect a complete sequence of levels without mixing other parities is extremely difficult, if not totally impossible. Most methods derive the average level spacings by applying a fit, with different degrees of generality, to the truncated Porter-Thomas distribution for reduced neutron widths. A method that tests both distributions of level widths and positions is discussed extensivey with an example of 168 Er data. 19 figures, 2 tables

  4. A transportation-scheduling system for managing silvicultural projects

    Science.gov (United States)

    Jorge F. Valenzuela; H. Hakan Balci; Timothy McDonald

    2005-01-01

    A silvicultural project encompasses tasks such as sitelevel planning, regeneration, harvestin, and stand-tending treatments. an essential problem in managing silvicultural projects is to efficiently schedule the operations while considering project task due dates and costs of moving scarce resources to specific job locations. Transportation costs represent a...

  5. 53 W average power few-cycle fiber laser system generating soft x rays up to the water window.

    Science.gov (United States)

    Rothhardt, Jan; Hädrich, Steffen; Klenke, Arno; Demmler, Stefan; Hoffmann, Armin; Gotschall, Thomas; Eidam, Tino; Krebs, Manuel; Limpert, Jens; Tünnermann, Andreas

    2014-09-01

    We report on a few-cycle laser system delivering sub-8-fs pulses with 353 μJ pulse energy and 25 GW of peak power at up to 150 kHz repetition rate. The corresponding average output power is as high as 53 W, which represents the highest average power obtained from any few-cycle laser architecture so far. The combination of both high average and high peak power provides unique opportunities for applications. We demonstrate high harmonic generation up to the water window and record-high photon flux in the soft x-ray spectral region. This tabletop source of high-photon flux soft x rays will, for example, enable coherent diffractive imaging with sub-10-nm resolution in the near future.

  6. Implementation, Comparison and Application of an Average Simulation Model of a Wind Turbine Driven Doubly Fed Induction Generator

    Directory of Open Access Journals (Sweden)

    Lidula N. Widanagama Arachchige

    2017-10-01

    Full Text Available Wind turbine driven doubly-fed induction generators (DFIGs are widely used in the wind power industry. With the increasing penetration of wind farms, analysis of their effect on power systems has become a critical requirement. This paper presents the modeling of wind turbine driven DFIGs using the conventional vector controls in a detailed model of a DFIG that represents power electronics (PE converters with device level models and proposes an average model eliminating the PE converters. The PSCAD/EMTDC™ (4.6 electromagnetic transient simulation software is used to develop the detailed and the proposing average model of a DFIG. The comparison of the two models reveals that the designed average DFIG model is adequate for simulating and analyzing most of the transient conditions.

  7. The use of induction linacs with nonlinear magnetic drive as high average power accelerators

    International Nuclear Information System (INIS)

    Birx, D.L.; Cook, E.G.; Hawkins, S.A.; Newton, M.A.; Poor, S.E.; Reginato, L.L.; Schmidt, J.A.; Smith, M.W.

    1985-01-01

    The marriage of induction linac technology with Nonlinear Magnetic Modulators has produced some unique capabilities. It appears possible to produce electron beams with average currents measured in amperes, at gradients exceeding 1 MeV/m, and with power efficiences approaching 50%. A 2 MeV, 5 kA electron accelerator is under construction at Lawrence Livermore National Laboratory (LLNL) to allow us to demonstrate some of these concepts. Progress on this project is reported here. (orig.)

  8. Quantum canonical ensemble: A projection operator approach

    Science.gov (United States)

    Magnus, Wim; Lemmens, Lucien; Brosens, Fons

    2017-09-01

    Knowing the exact number of particles N, and taking this knowledge into account, the quantum canonical ensemble imposes a constraint on the occupation number operators. The constraint particularly hampers the systematic calculation of the partition function and any relevant thermodynamic expectation value for arbitrary but fixed N. On the other hand, fixing only the average number of particles, one may remove the above constraint and simply factorize the traces in Fock space into traces over single-particle states. As is well known, that would be the strategy of the grand-canonical ensemble which, however, comes with an additional Lagrange multiplier to impose the average number of particles. The appearance of this multiplier can be avoided by invoking a projection operator that enables a constraint-free computation of the partition function and its derived quantities in the canonical ensemble, at the price of an angular or contour integration. Introduced in the recent past to handle various issues related to particle-number projected statistics, the projection operator approach proves beneficial to a wide variety of problems in condensed matter physics for which the canonical ensemble offers a natural and appropriate environment. In this light, we present a systematic treatment of the canonical ensemble that embeds the projection operator into the formalism of second quantization while explicitly fixing N, the very number of particles rather than the average. Being applicable to both bosonic and fermionic systems in arbitrary dimensions, transparent integral representations are provided for the partition function ZN and the Helmholtz free energy FN as well as for two- and four-point correlation functions. The chemical potential is not a Lagrange multiplier regulating the average particle number but can be extracted from FN+1 -FN, as illustrated for a two-dimensional fermion gas.

  9. Characterizing individual painDETECT symptoms by average pain severity

    Directory of Open Access Journals (Sweden)

    Sadosky A

    2016-07-01

    Full Text Available Alesia Sadosky,1 Vijaya Koduru,2 E Jay Bienen,3 Joseph C Cappelleri4 1Pfizer Inc, New York, NY, 2Eliassen Group, New London, CT, 3Outcomes Research Consultant, New York, NY, 4Pfizer Inc, Groton, CT, USA Background: painDETECT is a screening measure for neuropathic pain. The nine-item version consists of seven sensory items (burning, tingling/prickling, light touching, sudden pain attacks/electric shock-type pain, cold/heat, numbness, and slight pressure, a pain course pattern item, and a pain radiation item. The seven-item version consists only of the sensory items. Total scores of both versions discriminate average pain-severity levels (mild, moderate, and severe, but their ability to discriminate individual item severity has not been evaluated.Methods: Data were from a cross-sectional, observational study of six neuropathic pain conditions (N=624. Average pain severity was evaluated using the Brief Pain Inventory-Short Form, with severity levels defined using established cut points for distinguishing mild, moderate, and severe pain. The Wilcoxon rank sum test was followed by ridit analysis to represent the probability that a randomly selected subject from one average pain-severity level had a more favorable outcome on the specific painDETECT item relative to a randomly selected subject from a comparator severity level.Results: A probability >50% for a better outcome (less severe pain was significantly observed for each pain symptom item. The lowest probability was 56.3% (on numbness for mild vs moderate pain and highest probability was 76.4% (on cold/heat for mild vs severe pain. The pain radiation item was significant (P<0.05 and consistent with pain symptoms, as well as with total scores for both painDETECT versions; only the pain course item did not differ.Conclusion: painDETECT differentiates severity such that the ability to discriminate average pain also distinguishes individual pain item severity in an interpretable manner. Pain

  10. Averaged RMHD equations

    International Nuclear Information System (INIS)

    Ichiguchi, Katsuji

    1998-01-01

    A new reduced set of resistive MHD equations is derived by averaging the full MHD equations on specified flux coordinates, which is consistent with 3D equilibria. It is confirmed that the total energy is conserved and the linearized equations for ideal modes are self-adjoint. (author)

  11. Gamma-Weighted Discrete Ordinate Two-Stream Approximation for Computation of Domain Averaged Solar Irradiance

    Science.gov (United States)

    Kato, S.; Smith, G. L.; Barker, H. W.

    2001-01-01

    An algorithm is developed for the gamma-weighted discrete ordinate two-stream approximation that computes profiles of domain-averaged shortwave irradiances for horizontally inhomogeneous cloudy atmospheres. The algorithm assumes that frequency distributions of cloud optical depth at unresolved scales can be represented by a gamma distribution though it neglects net horizontal transport of radiation. This algorithm is an alternative to the one used in earlier studies that adopted the adding method. At present, only overcast cloudy layers are permitted.

  12. Cernavoda - Unit 2. A strategic project for Romania

    International Nuclear Information System (INIS)

    Saroudis, J.I.; Chirica, T.; Villabruna, G.

    1999-01-01

    The paper presents some key aspects of the Romanian nuclear program, focusing on Cernavoda NPP Unit No. 2 and the partnership with Atomic Energy of Canada Ltd. (AECL) and ANSALDO Italy for completion of this project. A brief historical presentation of the Romanian nuclear program is included. The success of Cernavoda NPP Unit 1 represents an important element in finalizing Unit 2 in an advanced state of equipment installation and more than 40 % complete. Also, the national infrastructure, including the legal framework and new Electricity Law represents a positive element for Project completion. The Romanian Nuclear Regulatory Body represents the guarantee for the safe operation of CANDU reactors in Romania.(author)

  13. The Influence of Atoms-in Molecules Methods on Shared Electron Distribution Indices and Domain Averaged Fermi Holes

    Czech Academy of Sciences Publication Activity Database

    Bultinck, P.; Cooper, D.L.; Ponec, Robert

    2010-01-01

    Roč. 114, č. 33 (2010), s. 8754-8763 ISSN 1089-5639 R&D Projects: GA ČR GA203/09/0118 Institutional research plan: CEZ:AV0Z40720504 Keywords : shared electron distribution index * domain averaged fermi holes * atoms in molecules Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 2.732, year: 2010

  14. The Stripa project. Annual report 1988

    International Nuclear Information System (INIS)

    1989-05-01

    The Stripa project is an international project being performed under the sponsorship of the OECD Nuclear Energy Agency (NEA). The project concerns research related to the disposal of highly radioactive waste in crystalline rock. The Research and Development Division of the Swedish Nuclear Fuel and Waste Management Company (SKB) hsa been enstrusted with the management of the project, under the direction of representatives from each participating country. The aim of this report is to inform the OECD Nuclear Energy Agency and the participants in the project about the general progress of work during 1988. (36 figs., 4 tabs.)

  15. Reproducing multi-model ensemble average with Ensemble-averaged Reconstructed Forcings (ERF) in regional climate modeling

    Science.gov (United States)

    Erfanian, A.; Fomenko, L.; Wang, G.

    2016-12-01

    Multi-model ensemble (MME) average is considered the most reliable for simulating both present-day and future climates. It has been a primary reference for making conclusions in major coordinated studies i.e. IPCC Assessment Reports and CORDEX. The biases of individual models cancel out each other in MME average, enabling the ensemble mean to outperform individual members in simulating the mean climate. This enhancement however comes with tremendous computational cost, which is especially inhibiting for regional climate modeling as model uncertainties can originate from both RCMs and the driving GCMs. Here we propose the Ensemble-based Reconstructed Forcings (ERF) approach to regional climate modeling that achieves a similar level of bias reduction at a fraction of cost compared with the conventional MME approach. The new method constructs a single set of initial and boundary conditions (IBCs) by averaging the IBCs of multiple GCMs, and drives the RCM with this ensemble average of IBCs to conduct a single run. Using a regional climate model (RegCM4.3.4-CLM4.5), we tested the method over West Africa for multiple combination of (up to six) GCMs. Our results indicate that the performance of the ERF method is comparable to that of the MME average in simulating the mean climate. The bias reduction seen in ERF simulations is achieved by using more realistic IBCs in solving the system of equations underlying the RCM physics and dynamics. This endows the new method with a theoretical advantage in addition to reducing computational cost. The ERF output is an unaltered solution of the RCM as opposed to a climate state that might not be physically plausible due to the averaging of multiple solutions with the conventional MME approach. The ERF approach should be considered for use in major international efforts such as CORDEX. Key words: Multi-model ensemble, ensemble analysis, ERF, regional climate modeling

  16. Averaging for solitons with nonlinearity management

    International Nuclear Information System (INIS)

    Pelinovsky, D.E.; Kevrekidis, P.G.; Frantzeskakis, D.J.

    2003-01-01

    We develop an averaging method for solitons of the nonlinear Schroedinger equation with a periodically varying nonlinearity coefficient, which is used to effectively describe solitons in Bose-Einstein condensates, in the context of the recently proposed technique of Feshbach resonance management. Using the derived local averaged equation, we study matter-wave bright and dark solitons and demonstrate a very good agreement between solutions of the averaged and full equations

  17. The database of the PREDICTS (Projecting Responses of Ecological Diversity In Changing Terrestrial Systems) project

    Science.gov (United States)

    Lawrence N. Hudson; Joseph Wunderle M.; And Others

    2016-01-01

    The PREDICTS project—Projecting Responses of Ecological Diversity In Changing Terrestrial Systems (www.predicts.org.uk)—has collated from published studies a large, reasonably representative database of comparable samples of biodiversity from multiple sites that differ in the nature or intensity of human impacts relating to land use. We have used this evidence base to...

  18. Impact of internal variability on projections of Sahel precipitation change

    Science.gov (United States)

    Monerie, Paul-Arthur; Sanchez-Gomez, Emilia; Pohl, Benjamin; Robson, Jon; Dong, Buwen

    2017-11-01

    The impact of the increase of greenhouse gases on Sahelian precipitation is very uncertain in both its spatial pattern and magnitude. In particular, the relative importance of internal variability versus external forcings depends on the time horizon considered in the climate projection. In this study we address the respective roles of the internal climate variability versus external forcings on Sahelian precipitation by using the data from the CESM Large Ensemble Project, which consists of a 40 member ensemble performed with the CESM1-CAM5 coupled model for the period 1920-2100. We show that CESM1-CAM5 is able to simulate the mean and interannual variability of Sahel precipitation, and is representative of a CMIP5 ensemble of simulations (i.e. it simulates the same pattern of precipitation change along with equivalent magnitude and seasonal cycle changes as the CMIP5 ensemble mean). However, CESM1-CAM5 underestimates the long-term decadal variability in Sahel precipitation. For short-term (2010-2049) and mid-term (2030-2069) projections the simulated internal variability component is able to obscure the projected impact of the external forcing. For long-term (2060-2099) projections external forcing induced change becomes stronger than simulated internal variability. Precipitation changes are found to be more robust over the central Sahel than over the western Sahel, where climate change effects struggle to emerge. Ten (thirty) members are needed to separate the 10 year averaged forced response from climate internal variability response in the western Sahel for a long-term (short-term) horizon. Over the central Sahel two members (ten members) are needed for a long-term (short-term) horizon.

  19. Balancing Design Project Supervision and Learning Facilitation

    DEFF Research Database (Denmark)

    Nielsen, Louise Møller

    2012-01-01

    experiences and expertise to guide the students’ decisions in relation to the design project. This paper focuses on project supervision in the context of design education – and more specifically on how this supervision is unfolded in a Problem Based Learning culture. The paper explores the supervisor......’s balance between the roles: 1) Design Project Supervisor – and 2) Learning Facilitator – with the aim to understand when to apply the different roles, and what to be aware of when doing so. This paper represents the first pilot-study of a larger research effort. It is based on a Lego Serious Play workshop......In design there is a long tradition for apprenticeship, as well as tradition for learning through design projects. Today many design educations are positioned within the University context, and have to be aligned with the learning culture and structure, which they represent. This raises a specific...

  20. Hanford Environmental Dose Reconstruction Project monthly report

    International Nuclear Information System (INIS)

    McMakin, A.H., Cannon, S.D.; Finch, S.M.

    1992-09-01

    The objective of the Hanford Environmental Dose Reconstruction MDR) Project is to estimate the radiation doses that individuals and populations could have received from nuclear operations at Hanford since 1944. The TSP consists of experts in envirorunental pathways. epidemiology, surface-water transport, ground-water transport, statistics, demography, agriculture, meteorology, nuclear engineering. radiation dosimetry. and cultural anthropology. Included are appointed members representing the states of Oregon, Washington, and Idaho, a representative of Native American tribes, and an individual representing the public. The project is divided into the following technical tasks. These tasks correspond to the path radionuclides followed from release to impact on humans (dose estimates): Source Terms; Environmental Transport; Environmental Monitoring Data Demography, Food Consumption, and Agriculture; and Environmental Pathways and Dose Estimates

  1. The database of the PREDICTS (Projecting Responses of Ecological Diversity In Changing Terrestrial Systems) project.

    Science.gov (United States)

    Hudson, Lawrence N; Newbold, Tim; Contu, Sara; Hill, Samantha L L; Lysenko, Igor; De Palma, Adriana; Phillips, Helen R P; Alhusseini, Tamera I; Bedford, Felicity E; Bennett, Dominic J; Booth, Hollie; Burton, Victoria J; Chng, Charlotte W T; Choimes, Argyrios; Correia, David L P; Day, Julie; Echeverría-Londoño, Susy; Emerson, Susan R; Gao, Di; Garon, Morgan; Harrison, Michelle L K; Ingram, Daniel J; Jung, Martin; Kemp, Victoria; Kirkpatrick, Lucinda; Martin, Callum D; Pan, Yuan; Pask-Hale, Gwilym D; Pynegar, Edwin L; Robinson, Alexandra N; Sanchez-Ortiz, Katia; Senior, Rebecca A; Simmons, Benno I; White, Hannah J; Zhang, Hanbin; Aben, Job; Abrahamczyk, Stefan; Adum, Gilbert B; Aguilar-Barquero, Virginia; Aizen, Marcelo A; Albertos, Belén; Alcala, E L; Del Mar Alguacil, Maria; Alignier, Audrey; Ancrenaz, Marc; Andersen, Alan N; Arbeláez-Cortés, Enrique; Armbrecht, Inge; Arroyo-Rodríguez, Víctor; Aumann, Tom; Axmacher, Jan C; Azhar, Badrul; Azpiroz, Adrián B; Baeten, Lander; Bakayoko, Adama; Báldi, András; Banks, John E; Baral, Sharad K; Barlow, Jos; Barratt, Barbara I P; Barrico, Lurdes; Bartolommei, Paola; Barton, Diane M; Basset, Yves; Batáry, Péter; Bates, Adam J; Baur, Bruno; Bayne, Erin M; Beja, Pedro; Benedick, Suzan; Berg, Åke; Bernard, Henry; Berry, Nicholas J; Bhatt, Dinesh; Bicknell, Jake E; Bihn, Jochen H; Blake, Robin J; Bobo, Kadiri S; Bóçon, Roberto; Boekhout, Teun; Böhning-Gaese, Katrin; Bonham, Kevin J; Borges, Paulo A V; Borges, Sérgio H; Boutin, Céline; Bouyer, Jérémy; Bragagnolo, Cibele; Brandt, Jodi S; Brearley, Francis Q; Brito, Isabel; Bros, Vicenç; Brunet, Jörg; Buczkowski, Grzegorz; Buddle, Christopher M; Bugter, Rob; Buscardo, Erika; Buse, Jörn; Cabra-García, Jimmy; Cáceres, Nilton C; Cagle, Nicolette L; Calviño-Cancela, María; Cameron, Sydney A; Cancello, Eliana M; Caparrós, Rut; Cardoso, Pedro; Carpenter, Dan; Carrijo, Tiago F; Carvalho, Anelena L; Cassano, Camila R; Castro, Helena; Castro-Luna, Alejandro A; Rolando, Cerda B; Cerezo, Alexis; Chapman, Kim Alan; Chauvat, Matthieu; Christensen, Morten; Clarke, Francis M; Cleary, Daniel F R; Colombo, Giorgio; Connop, Stuart P; Craig, Michael D; Cruz-López, Leopoldo; Cunningham, Saul A; D'Aniello, Biagio; D'Cruze, Neil; da Silva, Pedro Giovâni; Dallimer, Martin; Danquah, Emmanuel; Darvill, Ben; Dauber, Jens; Davis, Adrian L V; Dawson, Jeff; de Sassi, Claudio; de Thoisy, Benoit; Deheuvels, Olivier; Dejean, Alain; Devineau, Jean-Louis; Diekötter, Tim; Dolia, Jignasu V; Domínguez, Erwin; Dominguez-Haydar, Yamileth; Dorn, Silvia; Draper, Isabel; Dreber, Niels; Dumont, Bertrand; Dures, Simon G; Dynesius, Mats; Edenius, Lars; Eggleton, Paul; Eigenbrod, Felix; Elek, Zoltán; Entling, Martin H; Esler, Karen J; de Lima, Ricardo F; Faruk, Aisyah; Farwig, Nina; Fayle, Tom M; Felicioli, Antonio; Felton, Annika M; Fensham, Roderick J; Fernandez, Ignacio C; Ferreira, Catarina C; Ficetola, Gentile F; Fiera, Cristina; Filgueiras, Bruno K C; Fırıncıoğlu, Hüseyin K; Flaspohler, David; Floren, Andreas; Fonte, Steven J; Fournier, Anne; Fowler, Robert E; Franzén, Markus; Fraser, Lauchlan H; Fredriksson, Gabriella M; Freire, Geraldo B; Frizzo, Tiago L M; Fukuda, Daisuke; Furlani, Dario; Gaigher, René; Ganzhorn, Jörg U; García, Karla P; Garcia-R, Juan C; Garden, Jenni G; Garilleti, Ricardo; Ge, Bao-Ming; Gendreau-Berthiaume, Benoit; Gerard, Philippa J; Gheler-Costa, Carla; Gilbert, Benjamin; Giordani, Paolo; Giordano, Simonetta; Golodets, Carly; Gomes, Laurens G L; Gould, Rachelle K; Goulson, Dave; Gove, Aaron D; Granjon, Laurent; Grass, Ingo; Gray, Claudia L; Grogan, James; Gu, Weibin; Guardiola, Moisès; Gunawardene, Nihara R; Gutierrez, Alvaro G; Gutiérrez-Lamus, Doris L; Haarmeyer, Daniela H; Hanley, Mick E; Hanson, Thor; Hashim, Nor R; Hassan, Shombe N; Hatfield, Richard G; Hawes, Joseph E; Hayward, Matt W; Hébert, Christian; Helden, Alvin J; Henden, John-André; Henschel, Philipp; Hernández, Lionel; Herrera, James P; Herrmann, Farina; Herzog, Felix; Higuera-Diaz, Diego; Hilje, Branko; Höfer, Hubert; Hoffmann, Anke; Horgan, Finbarr G; Hornung, Elisabeth; Horváth, Roland; Hylander, Kristoffer; Isaacs-Cubides, Paola; Ishida, Hiroaki; Ishitani, Masahiro; Jacobs, Carmen T; Jaramillo, Víctor J; Jauker, Birgit; Hernández, F Jiménez; Johnson, McKenzie F; Jolli, Virat; Jonsell, Mats; Juliani, S Nur; Jung, Thomas S; Kapoor, Vena; Kappes, Heike; Kati, Vassiliki; Katovai, Eric; Kellner, Klaus; Kessler, Michael; Kirby, Kathryn R; Kittle, Andrew M; Knight, Mairi E; Knop, Eva; Kohler, Florian; Koivula, Matti; Kolb, Annette; Kone, Mouhamadou; Kőrösi, Ádám; Krauss, Jochen; Kumar, Ajith; Kumar, Raman; Kurz, David J; Kutt, Alex S; Lachat, Thibault; Lantschner, Victoria; Lara, Francisco; Lasky, Jesse R; Latta, Steven C; Laurance, William F; Lavelle, Patrick; Le Féon, Violette; LeBuhn, Gretchen; Légaré, Jean-Philippe; Lehouck, Valérie; Lencinas, María V; Lentini, Pia E; Letcher, Susan G; Li, Qi; Litchwark, Simon A; Littlewood, Nick A; Liu, Yunhui; Lo-Man-Hung, Nancy; López-Quintero, Carlos A; Louhaichi, Mounir; Lövei, Gabor L; Lucas-Borja, Manuel Esteban; Luja, Victor H; Luskin, Matthew S; MacSwiney G, M Cristina; Maeto, Kaoru; Magura, Tibor; Mallari, Neil Aldrin; Malone, Louise A; Malonza, Patrick K; Malumbres-Olarte, Jagoba; Mandujano, Salvador; Måren, Inger E; Marin-Spiotta, Erika; Marsh, Charles J; Marshall, E J P; Martínez, Eliana; Martínez Pastur, Guillermo; Moreno Mateos, David; Mayfield, Margaret M; Mazimpaka, Vicente; McCarthy, Jennifer L; McCarthy, Kyle P; McFrederick, Quinn S; McNamara, Sean; Medina, Nagore G; Medina, Rafael; Mena, Jose L; Mico, Estefania; Mikusinski, Grzegorz; Milder, Jeffrey C; Miller, James R; Miranda-Esquivel, Daniel R; Moir, Melinda L; Morales, Carolina L; Muchane, Mary N; Muchane, Muchai; Mudri-Stojnic, Sonja; Munira, A Nur; Muoñz-Alonso, Antonio; Munyekenye, B F; Naidoo, Robin; Naithani, A; Nakagawa, Michiko; Nakamura, Akihiro; Nakashima, Yoshihiro; Naoe, Shoji; Nates-Parra, Guiomar; Navarrete Gutierrez, Dario A; Navarro-Iriarte, Luis; Ndang'ang'a, Paul K; Neuschulz, Eike L; Ngai, Jacqueline T; Nicolas, Violaine; Nilsson, Sven G; Noreika, Norbertas; Norfolk, Olivia; Noriega, Jorge Ari; Norton, David A; Nöske, Nicole M; Nowakowski, A Justin; Numa, Catherine; O'Dea, Niall; O'Farrell, Patrick J; Oduro, William; Oertli, Sabine; Ofori-Boateng, Caleb; Oke, Christopher Omamoke; Oostra, Vicencio; Osgathorpe, Lynne M; Otavo, Samuel Eduardo; Page, Navendu V; Paritsis, Juan; Parra-H, Alejandro; Parry, Luke; Pe'er, Guy; Pearman, Peter B; Pelegrin, Nicolás; Pélissier, Raphaël; Peres, Carlos A; Peri, Pablo L; Persson, Anna S; Petanidou, Theodora; Peters, Marcell K; Pethiyagoda, Rohan S; Phalan, Ben; Philips, T Keith; Pillsbury, Finn C; Pincheira-Ulbrich, Jimmy; Pineda, Eduardo; Pino, Joan; Pizarro-Araya, Jaime; Plumptre, A J; Poggio, Santiago L; Politi, Natalia; Pons, Pere; Poveda, Katja; Power, Eileen F; Presley, Steven J; Proença, Vânia; Quaranta, Marino; Quintero, Carolina; Rader, Romina; Ramesh, B R; Ramirez-Pinilla, Martha P; Ranganathan, Jai; Rasmussen, Claus; Redpath-Downing, Nicola A; Reid, J Leighton; Reis, Yana T; Rey Benayas, José M; Rey-Velasco, Juan Carlos; Reynolds, Chevonne; Ribeiro, Danilo Bandini; Richards, Miriam H; Richardson, Barbara A; Richardson, Michael J; Ríos, Rodrigo Macip; Robinson, Richard; Robles, Carolina A; Römbke, Jörg; Romero-Duque, Luz Piedad; Rös, Matthias; Rosselli, Loreta; Rossiter, Stephen J; Roth, Dana S; Roulston, T'ai H; Rousseau, Laurent; Rubio, André V; Ruel, Jean-Claude; Sadler, Jonathan P; Sáfián, Szabolcs; Saldaña-Vázquez, Romeo A; Sam, Katerina; Samnegård, Ulrika; Santana, Joana; Santos, Xavier; Savage, Jade; Schellhorn, Nancy A; Schilthuizen, Menno; Schmiedel, Ute; Schmitt, Christine B; Schon, Nicole L; Schüepp, Christof; Schumann, Katharina; Schweiger, Oliver; Scott, Dawn M; Scott, Kenneth A; Sedlock, Jodi L; Seefeldt, Steven S; Shahabuddin, Ghazala; Shannon, Graeme; Sheil, Douglas; Sheldon, Frederick H; Shochat, Eyal; Siebert, Stefan J; Silva, Fernando A B; Simonetti, Javier A; Slade, Eleanor M; Smith, Jo; Smith-Pardo, Allan H; Sodhi, Navjot S; Somarriba, Eduardo J; Sosa, Ramón A; Soto Quiroga, Grimaldo; St-Laurent, Martin-Hugues; Starzomski, Brian M; Stefanescu, Constanti; Steffan-Dewenter, Ingolf; Stouffer, Philip C; Stout, Jane C; Strauch, Ayron M; Struebig, Matthew J; Su, Zhimin; Suarez-Rubio, Marcela; Sugiura, Shinji; Summerville, Keith S; Sung, Yik-Hei; Sutrisno, Hari; Svenning, Jens-Christian; Teder, Tiit; Threlfall, Caragh G; Tiitsaar, Anu; Todd, Jacqui H; Tonietto, Rebecca K; Torre, Ignasi; Tóthmérész, Béla; Tscharntke, Teja; Turner, Edgar C; Tylianakis, Jason M; Uehara-Prado, Marcio; Urbina-Cardona, Nicolas; Vallan, Denis; Vanbergen, Adam J; Vasconcelos, Heraldo L; Vassilev, Kiril; Verboven, Hans A F; Verdasca, Maria João; Verdú, José R; Vergara, Carlos H; Vergara, Pablo M; Verhulst, Jort; Virgilio, Massimiliano; Vu, Lien Van; Waite, Edward M; Walker, Tony R; Wang, Hua-Feng; Wang, Yanping; Watling, James I; Weller, Britta; Wells, Konstans; Westphal, Catrin; Wiafe, Edward D; Williams, Christopher D; Willig, Michael R; Woinarski, John C Z; Wolf, Jan H D; Wolters, Volkmar; Woodcock, Ben A; Wu, Jihua; Wunderle, Joseph M; Yamaura, Yuichi; Yoshikura, Satoko; Yu, Douglas W; Zaitsev, Andrey S; Zeidler, Juliane; Zou, Fasheng; Collen, Ben; Ewers, Rob M; Mace, Georgina M; Purves, Drew W; Scharlemann, Jörn P W; Purvis, Andy

    2017-01-01

    The PREDICTS project-Projecting Responses of Ecological Diversity In Changing Terrestrial Systems (www.predicts.org.uk)-has collated from published studies a large, reasonably representative database of comparable samples of biodiversity from multiple sites that differ in the nature or intensity of human impacts relating to land use. We have used this evidence base to develop global and regional statistical models of how local biodiversity responds to these measures. We describe and make freely available this 2016 release of the database, containing more than 3.2 million records sampled at over 26,000 locations and representing over 47,000 species. We outline how the database can help in answering a range of questions in ecology and conservation biology. To our knowledge, this is the largest and most geographically and taxonomically representative database of spatial comparisons of biodiversity that has been collated to date; it will be useful to researchers and international efforts wishing to model and understand the global status of biodiversity.

  2. Patch-based visual tracking with online representative sample selection

    Science.gov (United States)

    Ou, Weihua; Yuan, Di; Li, Donghao; Liu, Bin; Xia, Daoxun; Zeng, Wu

    2017-05-01

    Occlusion is one of the most challenging problems in visual object tracking. Recently, a lot of discriminative methods have been proposed to deal with this problem. For the discriminative methods, it is difficult to select the representative samples for the target template updating. In general, the holistic bounding boxes that contain tracked results are selected as the positive samples. However, when the objects are occluded, this simple strategy easily introduces the noises into the training data set and the target template and then leads the tracker to drift away from the target seriously. To address this problem, we propose a robust patch-based visual tracker with online representative sample selection. Different from previous works, we divide the object and the candidates into several patches uniformly and propose a score function to calculate the score of each patch independently. Then, the average score is adopted to determine the optimal candidate. Finally, we utilize the non-negative least square method to find the representative samples, which are used to update the target template. The experimental results on the object tracking benchmark 2013 and on the 13 challenging sequences show that the proposed method is robust to the occlusion and achieves promising results.

  3. Temperatures and heating energy in New Zealand houses from a nationally representative study - HEEP

    Energy Technology Data Exchange (ETDEWEB)

    French, L.J.; Camilleri, M.J.; Isaacs, N.P.; Pollard, A.R. [BRANZ Ltd., Private Bag 50 908, Porirua City (New Zealand)

    2007-07-15

    The household energy end-use project (HEEP) has collected energy and temperature data from a randomly selected, nationally representative sample of about 400 houses throughout New Zealand. This database has been used to explore the drivers of indoor temperatures and heating energy. Initial analysis of the winter living room temperatures shows that heating type, climate and house age are the key drivers. On average, houses heated by solid fuel are the warmest, with houses heated by portable LPG and electric heaters the coldest. Over the three winter months, living rooms are below 20 {sup o}C for 83% of the time - and the living room is typically the warmest room. Central heating is in only 5% of houses. Solid fuel is the dominant heating fuel in houses. The lack of air conditioning means that summer temperatures are affected by passive influences (e.g. house design, construction). Summer temperatures are strongly influenced by the house age and the local climate - together these variables explain 69% of the variation in daytime (9 a.m. to 5 p.m.) living room temperatures. In both summer and winter newer (post-1978) houses are warmer - this is beneficial in winter, but the high temperatures in summer are potentially uncomfortable. (author)

  4. Nodal O(h4)-superconvergence in 3D by averaging piecewise linear, bilinear, and trilinear FE approximations

    Czech Academy of Sciences Publication Activity Database

    Hannukainen, A.; Korotov, S.; Křížek, Michal

    2010-01-01

    Roč. 28, č. 1 (2010), s. 1-10 ISSN 0254-9409 R&D Projects: GA AV ČR(CZ) IAA100190803 Institutional research plan: CEZ:AV0Z10190503 Keywords : higher order error estimates * tetrahedral and prismatic elements * superconvergence * averaging operators Subject RIV: BA - General Mathematics Impact factor: 0.760, year: 2010 http://www.jstor.org/stable/43693564

  5. Scheduling of resource-constrained projects

    CERN Document Server

    Klein, Robert

    2000-01-01

    Project management has become a widespread instrument enabling organizations to efficiently master the challenges of steadily shortening product life cycles, global markets and decreasing profit margins. With projects increasing in size and complexity, their planning and control represents one of the most crucial management tasks. This is especially true for scheduling, which is concerned with establishing execution dates for the sub-activities to be performed in order to complete the project. The ability to manage projects where resources must be allocated between concurrent projects or even sub-activities of a single project requires the use of commercial project management software packages. However, the results yielded by the solution procedures included are often rather unsatisfactory. Scheduling of Resource-Constrained Projects develops more efficient procedures, which can easily be integrated into software packages by incorporated programming languages, and thus should be of great interest for practiti...

  6. Average and dispersion of the luminosity-redshift relation in the concordance model

    Energy Technology Data Exchange (ETDEWEB)

    Ben-Dayan, I. [DESY Hamburg (Germany). Theory Group; Gasperini, M. [Bari Univ. (Italy). Dipt. di Fisica; Istituto Nazionale di Fisica Nucleare, Bari (Italy); Marozzi, G. [College de France, 75 - Paris (France); Geneve Univ. (Switzerland). Dept. de Physique Theorique and CAP; Nugier, F. [Ecole Normale Superieure CNRS, Paris (France). Laboratoire de Physique Theorique; Veneziano, G. [College de France, 75 - Paris (France); CERN, Geneva (Switzerland). Physics Dept.; New York Univ., NY (United States). Dept. of Physics

    2013-03-15

    Starting from the luminosity-redshift relation recently given up to second order in the Poisson gauge, we calculate the effects of the realistic stochastic background of perturbations of the so-called concordance model on the combined light-cone and ensemble average of various functions of the luminosity distance, and on their variance, as functions of redshift. We apply a gauge-invariant light-cone averaging prescription which is free from infrared and ultraviolet divergences, making our results robust with respect to changes of the corresponding cutoffs. Our main conclusions, in part already anticipated in a recent letter for the case of a perturbation spectrum computed in the linear regime, are that such inhomogeneities not only cannot avoid the need for dark energy, but also cannot prevent, in principle, the determination of its parameters down to an accuracy of order 10{sup -3} - 10{sup -5}, depending on the averaged observable and on the regime considered for the power spectrum. However, taking into account the appropriate corrections arising in the non-linear regime, we predict an irreducible scatter of the data approaching the 10% level which, for limited statistics, will necessarily limit the attainable precision. The predicted dispersion appears to be in good agreement with current observational estimates of the distance-modulus variance due to Doppler and lensing effects (at low and high redshifts, respectively), and represents a challenge for future precision measurements.

  7. The Visible Human Project of the National Library of Medicine: Remote access and distribution of a multi-gigabyte data set

    Science.gov (United States)

    Ackerman, Michael J.

    1993-01-01

    As part of the 1986 Long-Range Plan for the National Library of Medicine (NLM), the Planning Panel on Medical Education wrote that NLM should '...thoroughly and systematically investigate the technical requirements for and feasibility of instituting a biomedical images library.' The panel noted the increasing use of images in clinical practice and biomedical research. An image library would complement NLM's existing bibliographic and factual database services and would ideally be available through the same computer networks as are these current NLM services. Early in 1989, NLM's Board of Regents convened an ad hoc planning panel to explore possible roles for the NLM in the area of electronic image libraries. In its report to the Board of Regents, the NLM Planning Panel on Electronic Image Libraries recommended that 'NLM should undertake a first project building a digital image library of volumetric data representing a complete, normal adult male and female. This Visible Human Project will include digitized photographic images for cryosectioning, digital images derived from computerized tomography, and digital magnetic resonance images of cadavers.' The technologies needed to support digital high resolution image libraries, including rapid development; and that NLM encourage investigator-initiated research into methods for representing and linking spatial and textual information, structural informatics. The first part of the Visible Human Project is the acquisition of cross-sectional CT and MRI digital images and cross-sectional cryosectional photographic images of a representative male and female cadaver at an average of one millimeter intervals. The corresponding cross-sections in each of the three modalities are to be registerable with one another.

  8. Projects of SOS FAIM

    Directory of Open Access Journals (Sweden)

    Mees, M.

    1985-01-01

    Full Text Available In Ivory Coast, the freshwater fishculture in rural areas is mainly on a small scale. This type of breeding in ponds (2 to 4 ares yields on an average 3 metric tons of fish/ha/year and represents only an activity with self-consumption of products. The yield in intensive pond fishculture Tilapia nilotica is on an average 6 to 7 metric tons/ha/year but yields bigger than 10 metric tons/ha/year are not uncommon. The intensive fishculture in floating cages, requiring a minor investment but a more improved formation than in fischculture, yields on an average about 30 to 40 kg/m3/year. However the effective development of this activity rests on the resolution of problems like the sufficient fry production, the feeding and the commercialization.

  9. Econometric modelling of Serbian current account determinants: Jackknife Model Averaging approach

    Directory of Open Access Journals (Sweden)

    Petrović Predrag

    2014-01-01

    Full Text Available This research aims to model Serbian current account determinants for the period Q1 2002 - Q4 2012. Taking into account the majority of relevant determinants, using the Jackknife Model Averaging approach, 48 different models have been estimated, where 1254 equations needed to be estimated and averaged for each of the models. The results of selected representative models indicate moderate persistence of the CA and positive influence of: fiscal balance, oil trade balance, terms of trade, relative income and real effective exchange rates, where we should emphasise: (i a rather strong influence of relative income, (ii the fact that the worsening of oil trade balance results in worsening of other components (probably non-oil trade balance of CA and (iii that the positive influence of terms of trade reveals functionality of the Harberger-Laursen-Metzler effect in Serbia. On the other hand, negative influence is evident in case of: relative economic growth, gross fixed capital formation, net foreign assets and trade openness. What particularly stands out is the strong effect of relative economic growth that, most likely, reveals high citizens' future income growth expectations, which has negative impact on the CA.

  10. A scaling approach to project regional sea level rise and its uncertainties

    Directory of Open Access Journals (Sweden)

    M. Perrette

    2013-01-01

    Full Text Available Climate change causes global mean sea level to rise due to thermal expansion of seawater and loss of land ice from mountain glaciers, ice caps and ice sheets. Locally, sea level can strongly deviate from the global mean rise due to changes in wind and ocean currents. In addition, gravitational adjustments redistribute seawater away from shrinking ice masses. However, the land ice contribution to sea level rise (SLR remains very challenging to model, and comprehensive regional sea level projections, which include appropriate gravitational adjustments, are still a nascent field (Katsman et al., 2011; Slangen et al., 2011. Here, we present an alternative approach to derive regional sea level changes for a range of emission and land ice melt scenarios, combining probabilistic forecasts of a simple climate model (MAGICC6 with the new CMIP5 general circulation models. The contribution from ice sheets varies considerably depending on the assumptions for the ice sheet projections, and thus represents sizeable uncertainties for future sea level rise. However, several consistent and robust patterns emerge from our analysis: at low latitudes, especially in the Indian Ocean and Western Pacific, sea level will likely rise more than the global mean (mostly by 10–20%. Around the northeastern Atlantic and the northeastern Pacific coasts, sea level will rise less than the global average or, in some rare cases, even fall. In the northwestern Atlantic, along the American coast, a strong dynamic sea level rise is counteracted by gravitational depression due to Greenland ice melt; whether sea level will be above- or below-average will depend on the relative contribution of these two factors. Our regional sea level projections and the diagnosed uncertainties provide an improved basis for coastal impact analysis and infrastructure planning for adaptation to climate change.

  11. Project Design Concept for Monitoring and Control System

    International Nuclear Information System (INIS)

    MCGREW, D.L.

    2000-01-01

    This Project Design Concept represents operational requirements established for use in design the tank farm Monitoring and Control System. These upgrades are included within the scope of Project W-314, Tank Farm Restoration and Safe Operations

  12. Ecological effects and potential risks of the water diversion project in the Heihe River Basin.

    Science.gov (United States)

    Zhang, Mengmeng; Wang, Shuai; Fu, Bojie; Gao, Guangyao; Shen, Qin

    2018-04-01

    To curb the severe ecological deterioration in the lower Heihe River Basin (HRB) in northwest China, a water diversion project was initiated in 2000. A comprehensive analysis of the ecological effects and potential risks associated with the project is needed. We assessed the hydrological and ecological achievements, and also analyzed the potential problems after the project was completed. We found that since the project began the hydrological regime has changed, with more than 57.82% of the upstream water being discharged to the lower reaches on average. As a result, the groundwater level in the lower reaches has risen; the terminal lake has gradually expanded to a maximum area in excess of 50km 2 since 2010, and there has been a significant recovery of vegetation in the riparian zone and the Ejin core oases, which represents the initial rehabilitation of the degraded downstream environment. Additionally, the economy of Ejin has developed spectacularly, with an annual growth rate of 28.06%. However, in the middle reaches, the average groundwater level has continuously declined by a total of 5.8m and significant degradation of the vegetation has occurred along the river course. The discrepancy in the water allocation between the middle and lower reaches has intensified. This highlights the inability of the current water diversion scheme to realize further ecological restoration and achieve sustainable development throughout the whole basin. In future water management programs, we recommend that water allocation is coordinated by considering the basin as an integrated entity and to scientifically determine the size of the midstream farmland and downstream oasis; restrict non-ecological water use in the lower reaches, and jointly dispatch the surface water and groundwater. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Managing projects a team-based approach

    CERN Document Server

    Brown, Karen A

    2010-01-01

    Students today are likely to be assigned to project teams or to be project managers almost immediately in their first job. Managing Projects: A Team-Based Approach was written for a wide range of stakeholders, including project managers, project team members, support personnel, functional mangers who provide resources for projects, project customers (and customer representatives), project sponsors, project subcontractors, and anyone who plays a role in the project delivery process. The need for project management is on the rise as product life cycles compress, demand for IT systems increases, and business takes on an increasingly global character. This book adds to the project management knowledge base in a way that fills an unmet need—it shows how teams can apply many of the standard project management tools, as well as several tools that are relatively new to the field. Managing Projects: A Team-Based Approach offers the academic rigor found in most textbooks along with the practical attributes often foun...

  14. Comparison of power spectra for tomosynthesis projections and reconstructed images

    International Nuclear Information System (INIS)

    Engstrom, Emma; Reiser, Ingrid; Nishikawa, Robert

    2009-01-01

    Burgess et al. [Med. Phys. 28, 419-437 (2001)] showed that the power spectrum of mammographic breast background follows a power law and that lesion detectability is affected by the power-law exponent β which measures the amount of structure in the background. Following the study of Burgess et al., the authors measured and compared the power-law exponent of mammographic backgrounds in tomosynthesis projections and reconstructed slices to investigate the effect of tomosynthesis imaging on background structure. Our data set consisted of 55 patient cases. For each case, regions of interest (ROIs) were extracted from both projection images and reconstructed slices. The periodogram of each ROI was computed by taking the squared modulus of the Fourier transform of the ROI. The power-law exponent was determined for each periodogram and averaged across all ROIs extracted from all projections or reconstructed slices for each patient data set. For the projections, the mean β averaged across the 55 cases was 3.06 (standard deviation of 0.21), while it was 2.87 (0.24) for the corresponding reconstructions. The difference in β for a given patient between the projection ROIs and the reconstructed ROIs averaged across the 55 cases was 0.194, which was statistically significant (p<0.001). The 95% CI for the difference between the mean value of β for the projections and reconstructions was [0.170, 0.218]. The results are consistent with the observation that the amount of breast structure in the tomosynthesis slice is reduced compared to projection mammography and that this may lead to improved lesion detectability.

  15. 40 CFR 600.510-12 - Calculation of average fuel economy and average carbon-related exhaust emissions.

    Science.gov (United States)

    2010-07-01

    ... and average carbon-related exhaust emissions. 600.510-12 Section 600.510-12 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) ENERGY POLICY FUEL ECONOMY AND CARBON-RELATED EXHAUST EMISSIONS OF... Transportation. (iv) [Reserved] (2) Average carbon-related exhaust emissions will be calculated to the nearest...

  16. Improved averaging for non-null interferometry

    Science.gov (United States)

    Fleig, Jon F.; Murphy, Paul E.

    2013-09-01

    Arithmetic averaging of interferometric phase measurements is a well-established method for reducing the effects of time varying disturbances, such as air turbulence and vibration. Calculating a map of the standard deviation for each pixel in the average map can provide a useful estimate of its variability. However, phase maps of complex and/or high density fringe fields frequently contain defects that severely impair the effectiveness of simple phase averaging and bias the variability estimate. These defects include large or small-area phase unwrapping artifacts, large alignment components, and voids that change in number, location, or size. Inclusion of a single phase map with a large area defect into the average is usually sufficient to spoil the entire result. Small-area phase unwrapping and void defects may not render the average map metrologically useless, but they pessimistically bias the variance estimate for the overwhelming majority of the data. We present an algorithm that obtains phase average and variance estimates that are robust against both large and small-area phase defects. It identifies and rejects phase maps containing large area voids or unwrapping artifacts. It also identifies and prunes the unreliable areas of otherwise useful phase maps, and removes the effect of alignment drift from the variance estimate. The algorithm has several run-time adjustable parameters to adjust the rejection criteria for bad data. However, a single nominal setting has been effective over a wide range of conditions. This enhanced averaging algorithm can be efficiently integrated with the phase map acquisition process to minimize the number of phase samples required to approach the practical noise floor of the metrology environment.

  17. Smoothing and projecting age-specific probabilities of death by TOPALS

    Directory of Open Access Journals (Sweden)

    Joop de Beer

    2012-10-01

    Full Text Available BACKGROUND TOPALS is a new relational model for smoothing and projecting age schedules. The model is operationally simple, flexible, and transparent. OBJECTIVE This article demonstrates how TOPALS can be used for both smoothing and projecting age-specific mortality for 26 European countries and compares the results of TOPALS with those of other smoothing and projection methods. METHODS TOPALS uses a linear spline to describe the ratios between the age-specific death probabilities of a given country and a standard age schedule. For smoothing purposes I use the average of death probabilities over 15 Western European countries as standard, whereas for projection purposes I use an age schedule of 'best practice' mortality. A partial adjustment model projects how quickly the death probabilities move in the direction of the best-practice level of mortality. RESULTS On average, TOPALS performs better than the Heligman-Pollard model and the Brass relational method in smoothing mortality age schedules. TOPALS can produce projections that are similar to those of the Lee-Carter method, but can easily be used to produce alternative scenarios as well. This article presents three projections of life expectancy at birth for the year 2060 for 26 European countries. The Baseline scenario assumes a continuation of the past trend in each country, the Convergence scenario assumes that there is a common trend across European countries, and the Acceleration scenario assumes that the future decline of death probabilities will exceed that in the past. The Baseline scenario projects that average European life expectancy at birth will increase to 80 years for men and 87 years for women in 2060, whereas the Acceleration scenario projects an increase to 90 and 93 years respectively. CONCLUSIONS TOPALS is a useful new tool for demographers for both smoothing age schedules and making scenarios.

  18. 14 CFR 1274.906 - Designation of New Technology Representative and Patent Representative.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 5 2010-01-01 2010-01-01 false Designation of New Technology... Conditions § 1274.906 Designation of New Technology Representative and Patent Representative. Designation of New Technology Representative and Patent Representative July 2002 (a) For purposes of administration...

  19. A time-averaged cosmic ray propagation theory

    International Nuclear Information System (INIS)

    Klimas, A.J.

    1975-01-01

    An argument is presented, which casts doubt on our ability to choose an appropriate magnetic field ensemble for computing the average behavior of cosmic ray particles. An alternate procedure, using time-averages rather than ensemble-averages, is presented. (orig.) [de

  20. Using SNOMED CT to Represent Two Interface Terminologies

    Science.gov (United States)

    Rosenbloom, S. Trent; Brown, Steven H.; Froehling, David; Bauer, Brent A.; Wahner-Roedler, Dietlind L.; Gregg, William M.; Elkin, Peter L.

    2009-01-01

    Objective Interface terminologies are designed to support interactions between humans and structured medical information. In particular, many interface terminologies have been developed for structured computer based documentation systems. Experts and policy-makers have recommended that interface terminologies be mapped to reference terminologies. The goal of the current study was to evaluate how well the reference terminology SNOMED CT could map to and represent two interface terminologies, MEDCIN and the Categorical Health Information Structured Lexicon (CHISL). Design Automated mappings between SNOMED CT and 500 terms from each of the two interface terminologies were evaluated by human reviewers, who also searched SNOMED CT to identify better mappings when this was judged to be necessary. Reviewers judged whether they believed the interface terms to be clinically appropriate, whether the terms were covered by SNOMED CT concepts and whether the terms' implied semantic structure could be represented by SNOMED CT. Measurements Outcomes included concept coverage by SNOMED CT for study terms and their implied semantics. Agreement statistics and compositionality measures were calculated. Results The SNOMED CT terminology contained concepts to represent 92.4% of MEDCIN and 95.9% of CHISL terms. Semantic structures implied by study terms were less well covered, with some complex compositional expressions requiring semantics not present in SNOMED CT. Among sampled terms, those from MEDCIN were more complex than those from CHISL, containing an average 3.8 versus 1.8 atomic concepts respectively, pterms. PMID:18952944

  1. FLORAM project

    Energy Technology Data Exchange (ETDEWEB)

    Zulauf, W.E. [Sao Paolos Environmental Secretariat, Sao Paolo (Brazil); Goelho, A.S.R. [Riocell, S.A. (Brazil); Saber, A. [IEA-Instituto de Estudos Avancados (Brazil)] [and others

    1995-12-31

    The project FLORAM was formulated at the `Institute for Advanced Studies` of the University of Sao Paulo. It aims at decreasing the level of carbon dioxide in the atmosphere and thus curbing the green-house effect by way of a huge effort of forestation and reforestation. The resulting forests when the trees mature, will be responsible for the absorption of about 6 billion tons of excess carbon. It represents 5 % of the total amount of CO{sub 2} which is in excess in the earth`s atmosphere and represents 5 % of the available continental surfaces which can be forested as well. Therefore, if similar projects are implemented throughout the world, in theory all the exceeding CO{sub 2}, responsible for the `greenhouse effect`, (27 % or 115 billion tons of carbon) would be absorbed. Regarding this fact, there would be a 400 million hectar increase of growing forests. FLORAM in Brazil aims to plant 20.000.000 ha in 2 years at a cost of 20 billion dollars. If it reaches its goals that will mean that Brazil will have reforested an area almost half as big as France. (author)

  2. FLORAM project

    Energy Technology Data Exchange (ETDEWEB)

    Zulauf, W E [Sao Paolos Environmental Secretariat, Sao Paolo (Brazil); Goelho, A S.R. [Riocell, S.A. (Brazil); Saber, A [IEA-Instituto de Estudos Avancados (Brazil); and others

    1996-12-31

    The project FLORAM was formulated at the `Institute for Advanced Studies` of the University of Sao Paulo. It aims at decreasing the level of carbon dioxide in the atmosphere and thus curbing the green-house effect by way of a huge effort of forestation and reforestation. The resulting forests when the trees mature, will be responsible for the absorption of about 6 billion tons of excess carbon. It represents 5 % of the total amount of CO{sub 2} which is in excess in the earth`s atmosphere and represents 5 % of the available continental surfaces which can be forested as well. Therefore, if similar projects are implemented throughout the world, in theory all the exceeding CO{sub 2}, responsible for the `greenhouse effect`, (27 % or 115 billion tons of carbon) would be absorbed. Regarding this fact, there would be a 400 million hectar increase of growing forests. FLORAM in Brazil aims to plant 20.000.000 ha in 2 years at a cost of 20 billion dollars. If it reaches its goals that will mean that Brazil will have reforested an area almost half as big as France. (author)

  3. Canadian fusion fuels technology project

    International Nuclear Information System (INIS)

    1986-01-01

    The Canadian Fusion Fuels Technology Project was launched in 1982 to coordinate Canada's provision of fusion fuels technology to international fusion power development programs. The project has a mandate to extend and adapt existing Canadian tritium technologies for use in international fusion power development programs. 1985-86 represents the fourth year of the first five-year term of the Canadian Fusion Fuels Technology Project (CFFTP). This reporting period coincides with an increasing trend in global fusion R and D to direct more effort towards the management of tritium. This has resulted in an increased linking of CFFTP activities and objectives with those of facilities abroad. In this way there has been a continuing achievement resulting from CFFTP efforts to have cooperative R and D and service activities with organizations abroad. All of this is aided by the cooperative international atmosphere within the fusion community. This report summarizes our past year and provides some highlights of the upcoming year 1986/87, which is the final year of the first five-year phase of the program. AECL (representing the Federal Government), the Ministry of Energy (representing Ontario) and Ontario Hydro, have given formal indication of their intent to continue with a second five-year program. Plans for the second phase will continue to emphasize tritium technology and remote handling

  4. Averaged emission factors for the Hungarian car fleet

    Energy Technology Data Exchange (ETDEWEB)

    Haszpra, L. [Inst. for Atmospheric Physics, Budapest (Hungary); Szilagyi, I. [Central Research Inst. for Chemistry, Budapest (Hungary)

    1995-12-31

    The vehicular emission of non-methane hydrocarbon (NMHC) is one of the largest anthropogenic sources of NMHC in Hungary and in most of the industrialized countries. Non-methane hydrocarbon plays key role in the formation of photo-chemical air pollution, usually characterized by the ozone concentration, which seriously endangers the environment and human health. The ozone forming potential of the different NMHCs differs from each other significantly, while the NMHC composition of the car exhaust is influenced by the fuel and engine type, technical condition of the vehicle, vehicle speed and several other factors. In Hungary the majority of the cars are still of Eastern European origin. They represent the technological standard of the 70`s, although there are changes recently. Due to the long-term economical decline in Hungary the average age of the cars was about 9 years in 1990 and reached 10 years by 1993. The condition of the majority of the cars is poor. In addition, almost one third (31.2 %) of the cars are equipped with two-stroke engines which emit less NO{sub x} but much more hydrocarbon. The number of cars equipped with catalytic converter was negligible in 1990 and is slowly increasing only recently. As a consequence of these facts the traffic emission in Hungary may differ from that measured in or estimated for the Western European countries and the differences should be taken into account in the air pollution models. For the estimation of the average emission of the Hungarian car fleet a one-day roadway tunnel experiment was performed in the downtown of Budapest in summer, 1991. (orig.)

  5. Severn Barrage project. Detailed report - V. 5

    Energy Technology Data Exchange (ETDEWEB)

    1989-01-01

    Prior to the present programme of work, the effects which a tidal power barrage would have on the region, during both construction and operation, had not been studied in detail. This volume of the Detailed Report therefore represents a significant extension of work into these aspects of the Severn Barrage Project. In the Regional Study, a number of benefits have been identified, some of which may represent net benefits nationally. The economic assessment of both regional and national benefits and costs is presented. The second part of this volume reports on the work done on the Legal Background for the Project. (author).

  6. A method for projecting full-scale performance of CO/sub 2/ flooding in the Willard Unit

    International Nuclear Information System (INIS)

    Bilhartz, H.L.; Charlson, G.S.; Stalkup, F.I.; Miller, C.C.

    1985-01-01

    A non-producing CO/sub 2/ flood tertiary recovery test was recently completed in the Willard Unit of Wasson Field. Flood responses during waterflood and alternate injection of CO/sub 2/ and water were monitored at a logging observation well using compensated neutron and pulsed neutron logs. A pressure core was taken to measure residual oil saturations at the test conclusion. The overall objective of the testing was to obtain information for evaluating the potential for full-scale CO/sub 2/ flooding in the unit. Out method for making this evaluation involves: (1) defining CO/sub 2/ flood displacement efficiency and representing this efficiency in a miscible flood reservoir simulator; (2) defining a representative average reservoir description; and (3) projecting full-scale CO/sub 2/ flood performance with the simulator. The paper provides a status report on progress to assess CO/sub 2/ flooding potential for the Willard Unit in this manner

  7. CLIC project timeline

    CERN Multimedia

    CLIC, Compact Linear Collider Project

    2018-01-01

    The CLIC project timeline. Current plan is to start at sqrt(s)=380 GeV for Higgs and top quark precision physics and upgrade up to 3 TeV. This timeline represent a purely technical schedule and assumes support at the European Strategy for Particle Physics (ESPP) in 2020 and available funding.

  8. High-Average-Power Diffraction Pulse-Compression Gratings Enabling Next-Generation Ultrafast Laser Systems

    Energy Technology Data Exchange (ETDEWEB)

    Alessi, D. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2016-11-01

    Pulse compressors for ultrafast lasers have been identified as a technology gap in the push towards high peak power systems with high average powers for industrial and scientific applications. Gratings for ultrashort (sub-150fs) pulse compressors are metallic and can absorb a significant percentage of laser energy resulting in up to 40% loss as well as thermal issues which degrade on-target performance. We have developed a next generation gold grating technology which we have scaled to the petawatt-size. This resulted in improvements in efficiency, uniformity and processing as compared to previous substrate etched gratings for high average power. This new design has a deposited dielectric material for the grating ridge rather than etching directly into the glass substrate. It has been observed that average powers as low as 1W in a compressor can cause distortions in the on-target beam. We have developed and tested a method of actively cooling diffraction gratings which, in the case of gold gratings, can support a petawatt peak power laser with up to 600W average power. We demonstrated thermo-mechanical modeling of a grating in its use environment and benchmarked with experimental measurement. Multilayer dielectric (MLD) gratings are not yet used for these high peak power, ultrashort pulse durations due to their design challenges. We have designed and fabricated broad bandwidth, low dispersion MLD gratings suitable for delivering 30 fs pulses at high average power. This new grating design requires the use of a novel Out Of Plane (OOP) compressor, which we have modeled, designed, built and tested. This prototype compressor yielded a transmission of 90% for a pulse with 45 nm bandwidth, and free of spatial and angular chirp. In order to evaluate gratings and compressors built in this project we have commissioned a joule-class ultrafast Ti:Sapphire laser system. Combining the grating cooling and MLD technologies developed here could enable petawatt laser systems to

  9. Non-Relative Value Unit-Generating Activities Represent One-Fifth of Academic Neuroradiologist Productivity.

    Science.gov (United States)

    Wintermark, M; Zeineh, M; Zaharchuk, G; Srivastava, A; Fischbein, N

    2016-07-01

    A neuroradiologist's activity includes many tasks beyond interpreting relative value unit-generating imaging studies. Our aim was to test a simple method to record and quantify the non-relative value unit-generating clinical activity represented by consults and clinical conferences, including tumor boards. Four full-time neuroradiologists, working an average of 50% clinical and 50% academic activity, systematically recorded all the non-relative value unit-generating consults and conferences in which they were involved during 3 months by using a simple, Web-based, computer-based application accessible from smartphones, tablets, or computers. The number and type of imaging studies they interpreted during the same period and the associated relative value units were extracted from our billing system. During 3 months, the 4 neuroradiologists working an average of 50% clinical activity interpreted 4241 relative value unit-generating imaging studies, representing 8152 work relative value units. During the same period, they recorded 792 non-relative value unit-generating study reviews as part of consults and conferences (not including reading room consults), representing 19% of the interpreted relative value unit-generating imaging studies. We propose a simple Web-based smartphone app to record and quantify non-relative value unit-generating activities including consults, clinical conferences, and tumor boards. The quantification of non-relative value unit-generating activities is paramount in this time of a paradigm shift from volume to value. It also represents an important tool for determining staffing levels, which cannot be performed on the basis of relative value unit only, considering the importance of time spent by radiologists on non-relative value unit-generating activities. It may also influence payment models from medical centers to radiology departments or practices. © 2016 by American Journal of Neuroradiology.

  10. 40 CFR 426.122 - Effluent limitations guidelines representing the degree of effluent reduction attainable by the...

    Science.gov (United States)

    2010-07-01

    ... Incandescent Lamp Envelope Manufacturing Subcategory § 426.122 Effluent limitations guidelines representing the...): (a) Any manufacturing plant which produces incandescent lamp envelopes shall meet the following... any 1 day Average of daily values for 30 consecutive days shall not exceed— Metric units (g/kkg of...

  11. 40 CFR 76.11 - Emissions averaging.

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 16 2010-07-01 2010-07-01 false Emissions averaging. 76.11 Section 76.11 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) AIR PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.11 Emissions averaging. (a) General...

  12. Average spectral efficiency analysis of FSO links over turbulence channel with adaptive transmissions and aperture averaging

    Science.gov (United States)

    Aarthi, G.; Ramachandra Reddy, G.

    2018-03-01

    In our paper, the impact of adaptive transmission schemes: (i) optimal rate adaptation (ORA) and (ii) channel inversion with fixed rate (CIFR) on the average spectral efficiency (ASE) are explored for free-space optical (FSO) communications with On-Off Keying (OOK), Polarization shift keying (POLSK), and Coherent optical wireless communication (Coherent OWC) systems under different turbulence regimes. Further to enhance the ASE we have incorporated aperture averaging effects along with the above adaptive schemes. The results indicate that ORA adaptation scheme has the advantage of improving the ASE performance compared with CIFR under moderate and strong turbulence regime. The coherent OWC system with ORA excels the other modulation schemes and could achieve ASE performance of 49.8 bits/s/Hz at the average transmitted optical power of 6 dBm under strong turbulence. By adding aperture averaging effect we could achieve an ASE of 50.5 bits/s/Hz under the same conditions. This makes ORA with Coherent OWC modulation as a favorable candidate for improving the ASE of the FSO communication system.

  13. The LHC Computing Grid Project

    CERN Multimedia

    Åkesson, T

    In the last ATLAS eNews I reported on the preparations for the LHC Computing Grid Project (LCGP). Significant LCGP resources were mobilized during the summer, and there have been numerous iterations on the formal paper to put forward to the CERN Council to establish the LCGP. ATLAS, and also the other LHC-experiments, has been very active in this process to maximally influence the outcome. Our main priorities were to ensure that the global aspects are properly taken into account, that the CERN non-member states are also included in the structure, that the experiments are properly involved in the LCGP execution and that the LCGP takes operative responsibility during the data challenges. A Project Launch Board (PLB) was active from the end of July until the 10th of September. It was chaired by Hans Hoffmann and had the IT division leader as secretary. Each experiment had a representative (me for ATLAS), and the large CERN member states were each represented while the smaller were represented as clusters ac...

  14. A practical guide to averaging functions

    CERN Document Server

    Beliakov, Gleb; Calvo Sánchez, Tomasa

    2016-01-01

    This book offers an easy-to-use and practice-oriented reference guide to mathematical averages. It presents different ways of aggregating input values given on a numerical scale, and of choosing and/or constructing aggregating functions for specific applications. Building on a previous monograph by Beliakov et al. published by Springer in 2007, it outlines new aggregation methods developed in the interim, with a special focus on the topic of averaging aggregation functions. It examines recent advances in the field, such as aggregation on lattices, penalty-based aggregation and weakly monotone averaging, and extends many of the already existing methods, such as: ordered weighted averaging (OWA), fuzzy integrals and mixture functions. A substantial mathematical background is not called for, as all the relevant mathematical notions are explained here and reported on together with a wealth of graphical illustrations of distinct families of aggregation functions. The authors mainly focus on practical applications ...

  15. The Colibri Project

    DEFF Research Database (Denmark)

    Lopez, Jose Manuel Guterrez Lopez; Frick, Jan; Kirikova, Marite

    2015-01-01

    Colibri (Collaboration and Innovation for Better, Personalized and IT-Supported Teaching) is a three year project co-funded by the Erasmus + Strategic Partnership, starting in 2014. The project is being carried out by seven academic, two industrial, and one governmental organisation partners from...... eight different countries [1]. The main objective of the project is enhancing the quality and relevance of the learning offer in education by developing and systematically testing new and innovative approaches, and by supporting the dissemination and spreading of best practices. This is in the focus...... course on Future Internet Opportunities is part of the Living Lab, followed by 28 students and given by more than 10 different teachers. Both the teachers and the students represent different fields within computer science, electrical engineering, telecommunications, business informatics, management...

  16. 7 CFR 51.2561 - Average moisture content.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 2 2010-01-01 2010-01-01 false Average moisture content. 51.2561 Section 51.2561... STANDARDS) United States Standards for Grades of Shelled Pistachio Nuts § 51.2561 Average moisture content. (a) Determining average moisture content of the lot is not a requirement of the grades, except when...

  17. Calculation of average molecular parameters, functional groups, and a surrogate molecule for heavy fuel oils using 1H and 13C NMR spectroscopy

    KAUST Repository

    Abdul Jameel, Abdul Gani; Elbaz, Ayman M.; Emwas, Abdul-Hamid M.; Roberts, William L.; Sarathy, Mani

    2016-01-01

    Heavy fuel oil (HFO) is primarily used as fuel in marine engines and in boilers to generate electricity. Nuclear Magnetic Resonance (NMR) is a powerful analytical tool for structure elucidation and in this study, 1H NMR and 13C NMR spectroscopy were used for the structural characterization of 2 HFO samples. The NMR data was combined with elemental analysis and average molecular weight to quantify average molecular parameters (AMPs), such as the number of paraffinic carbons, naphthenic carbons, aromatic hydrogens, olefinic hydrogens, etc. in the HFO samples. Recent formulae published in the literature were used for calculating various derived AMPs like aromaticity factor 〖(f〗_a), C/H ratio, average paraffinic chain length (¯n), naphthenic ring number 〖(R〗_N), aromatic ring number〖 (R〗_A), total ring number〖 (R〗_T), aromatic condensation index (φ) and aromatic condensation degree (Ω). These derived AMPs help in understanding the overall structure of the fuel. A total of 19 functional groups were defined to represent the HFO samples, and their respective concentrations were calculated by formulating balance equations that equate the concentration of the functional groups with the concentration of the AMPs. Heteroatoms like sulfur, nitrogen, and oxygen were also included in the functional groups. Surrogate molecules were finally constructed to represent the average structure of the molecules present in the HFO samples. This surrogate molecule can be used for property estimation of the HFO samples and also serve as a surrogate to represent the molecular structure for use in kinetic studies.

  18. Calculation of average molecular parameters, functional groups, and a surrogate molecule for heavy fuel oils using 1H and 13C NMR spectroscopy

    KAUST Repository

    Abdul Jameel, Abdul Gani

    2016-04-22

    Heavy fuel oil (HFO) is primarily used as fuel in marine engines and in boilers to generate electricity. Nuclear Magnetic Resonance (NMR) is a powerful analytical tool for structure elucidation and in this study, 1H NMR and 13C NMR spectroscopy were used for the structural characterization of 2 HFO samples. The NMR data was combined with elemental analysis and average molecular weight to quantify average molecular parameters (AMPs), such as the number of paraffinic carbons, naphthenic carbons, aromatic hydrogens, olefinic hydrogens, etc. in the HFO samples. Recent formulae published in the literature were used for calculating various derived AMPs like aromaticity factor 〖(f〗_a), C/H ratio, average paraffinic chain length (¯n), naphthenic ring number 〖(R〗_N), aromatic ring number〖 (R〗_A), total ring number〖 (R〗_T), aromatic condensation index (φ) and aromatic condensation degree (Ω). These derived AMPs help in understanding the overall structure of the fuel. A total of 19 functional groups were defined to represent the HFO samples, and their respective concentrations were calculated by formulating balance equations that equate the concentration of the functional groups with the concentration of the AMPs. Heteroatoms like sulfur, nitrogen, and oxygen were also included in the functional groups. Surrogate molecules were finally constructed to represent the average structure of the molecules present in the HFO samples. This surrogate molecule can be used for property estimation of the HFO samples and also serve as a surrogate to represent the molecular structure for use in kinetic studies.

  19. CENELEC Project Report Smart House Roadmap (draft)

    NARCIS (Netherlands)

    Hartog, F.T.H. den; Suters, T.; Parsons, J.; Faller, J.

    2010-01-01

    This CENELEC project report has been drafted by a project team and steering group of representatives of interested parties and is to be endorsed on 2010-11-23. Neither the national members of CENELEC nor the CEN-CENELEC Management Centre can be held accountable for the technical content of this

  20. Pingston hydroelectric project - stage 2 project report : volume 1 of 2

    International Nuclear Information System (INIS)

    1999-01-01

    The British Columbia Environmental Assessment Office has reviewed an application by Canadian Hydro Developers (B.C.) Inc., to develop the Pingston Hydroelectric Project, a run-of-river facility, located 60 km south of Revelstoke, British Columbia. The review includes an evaluation of any social and environmental impacts, including the effects on migratory birds, small mammals, amphibians, fish, wildlife, water quality, noise and visual aspects, that would be associated with the project. One consideration in selecting the project site was its ease of access along existing roadways and close proximity to power lines, another, the absence of any potential effects on fish and fish habitat. The project is also expected to have a negligible effect on water quality- related matters such as sediment transport, nutrient trapping in the headpond, and movement of large organic debris through the system. The project will not impact on archaeological resources, offers a renewable supply of energy and represents an overall positive effect on air quality for British Columbia. BC Hydro has committed to providing electrical interconnection to the main grid. 9 refs., 5 tabs., 14 figs., 9 appendices

  1. Why does project planning fail

    International Nuclear Information System (INIS)

    Foley, M.; Luciano, G.

    1991-01-01

    The technology of project controls have become increasingly sophisticated. Some say that the technological advancements of the Nineteen-Eighties represented a maturation of project control tools. Others say that the advancements were merely bells and whistles that added little or nothing to the project management process. Above it all, as we enter the Nineties, there is a popular outcry to get back to the basics of planning. The genesis of this outcry is the sobering impact of significant cost overruns and schedule extensions, even on projects that have employed the most advanced project control tools and systems. This paper examines the merits of taking a strategic approach to the project planning process. Within that context, there are basic goals of planning which are enduring through the life cycle of the project. Key reasons for failure and inability to achieve the goals of project planning are explained. By examining the goals of project planning and the reasons for failure, insight is provided into the role of project controls specialists and sophisticated project control tools in meeting the challenges of complex project management in the 1990's

  2. Determining average path length and average trapping time on generalized dual dendrimer

    Science.gov (United States)

    Li, Ling; Guan, Jihong

    2015-03-01

    Dendrimer has wide number of important applications in various fields. In some cases during transport or diffusion process, it transforms into its dual structure named Husimi cactus. In this paper, we study the structure properties and trapping problem on a family of generalized dual dendrimer with arbitrary coordination numbers. We first calculate exactly the average path length (APL) of the networks. The APL increases logarithmically with the network size, indicating that the networks exhibit a small-world effect. Then we determine the average trapping time (ATT) of the trapping process in two cases, i.e., the trap placed on a central node and the trap is uniformly distributed in all the nodes of the network. In both case, we obtain explicit solutions of ATT and show how they vary with the networks size. Besides, we also discuss the influence of the coordination number on trapping efficiency.

  3. Alderney 5 complex demonstration project

    Energy Technology Data Exchange (ETDEWEB)

    Stewart, D. [High Performance Energy Systems, Halifax, NS (Canada)

    2008-07-01

    The Halifax Regional Municipality (HRM) is the largest municipality in Atlantic Canada. This presentation described the flagship facility and the energy efficiency retrofit of five HRM-owned buildings called the Alderney 5 complex. The 5 objectives of the demonstration project involved a district-scale cooling project; replacement of chillers with harbour cooling; and replacement of a high exergy system with a low exergy system. Synergies and challenges of the project were also identified. The presentation also referred to borehole thermal energy storage; existing Halifax Harbour cooling; Halifax Harbour temperatures; cold energy geothermal borehole field; and the benefits of advanced concentric boreholes. A project update and progress to date were also provided. The Alderney 5 project represents the first concentric borehole technology for use to store and retrieve cold energy. tabs., figs.

  4. Computation of the bounce-average code

    International Nuclear Information System (INIS)

    Cutler, T.A.; Pearlstein, L.D.; Rensink, M.E.

    1977-01-01

    The bounce-average computer code simulates the two-dimensional velocity transport of ions in a mirror machine. The code evaluates and bounce-averages the collision operator and sources along the field line. A self-consistent equilibrium magnetic field is also computed using the long-thin approximation. Optionally included are terms that maintain μ, J invariance as the magnetic field changes in time. The assumptions and analysis that form the foundation of the bounce-average code are described. When references can be cited, the required results are merely stated and explained briefly. A listing of the code is appended

  5. Project Lifespan-based Nonstationary Hydrologic Design Methods for Changing Environment

    Science.gov (United States)

    Xiong, L.

    2017-12-01

    Under changing environment, we must associate design floods with the design life period of projects to ensure the hydrologic design is really relevant to the operation of the hydrologic projects, because the design value for a given exceedance probability over the project life period would be significantly different from that over other time periods of the same length due to the nonstationarity of probability distributions. Several hydrologic design methods that take the design life period of projects into account have been proposed in recent years, i.e. the expected number of exceedances (ENE), design life level (DLL), equivalent reliability (ER), and average design life level (ADLL). Among the four methods to be compared, both the ENE and ER methods are return period-based methods, while DLL and ADLL are risk/reliability- based methods which estimate design values for given probability values of risk or reliability. However, the four methods can be unified together under a general framework through a relationship transforming the so-called representative reliability (RRE) into the return period, i.e. m=1/1(1-RRE), in which we compute the return period m using the representative reliability RRE.The results of nonstationary design quantiles and associated confidence intervals calculated by ENE, ER and ADLL were very similar, since ENE or ER was a special case or had a similar expression form with respect to ADLL. In particular, the design quantiles calculated by ENE and ADLL were the same when return period was equal to the length of the design life. In addition, DLL can yield similar design values if the relationship between DLL and ER/ADLL return periods is considered. Furthermore, ENE, ER and ADLL had good adaptability to either an increasing or decreasing situation, yielding not too large or too small design quantiles. This is important for applications of nonstationary hydrologic design methods in actual practice because of the concern of choosing the emerging

  6. STUDENTS’ COMMUNICATION SKILLS ASSESSMENT BY EXTERNAL LECTURERS AND INDUSTRY REPRESENTATIVES

    Directory of Open Access Journals (Sweden)

    NOORFAZILA KAMAL

    2016-11-01

    Full Text Available Soft skills, especially communication skills, are important skills which each graduate must possess. Accordingly, several courses and approaches have been carried out in universities to train students in communication skills. The skills are normally evaluated by course lecturers. However, assessments by a third party from outside the university are valuable so that the students’ ahievements may be weighed against external evaluators’ point of views. In the Department of Electrical, Electronic and Systems Engineering (DEESE, Universiti Kebangsaan Malaysia (UKM, communication skills assessment by external lecturers and industry representatives are performed on Hari Poster JKEES, where students present their final year project poster. There are two categories of evaluation, namely project and communication skills. The project evaluation covers content, result and impact, while communication skills evaluation covers poster layout and design, and delivery. This study only analyse the students’ communication skills achievement. The participants of this study consists of 109 final year students from two batches, of which 51 students are from year 2014 and the other 58 students from year 2015. The results show that for the year 2014 students, the mean mark given by external lecturers in layout and design category is 6.7, while the mean mark from industry evaluators is 6.5. For the 2015 batch, the mean mark in the layout and design category is 6.3 from external lecturers and 5.9 from industry evaluators. In the delivery category, the mean marks for the 2014 batch are 7.1 and 6.6 from external lecturers and industry evaluators, espectively. Meanwhile, for the 2015 batch, the mean marks by external lecturers and industry evaluators are 6.3 and 5.8, respectively. The results show that both external lecturers and industry representatives judged DEESE students’ communication skills to be good.

  7. The Importance of Investment Decisions in Project Management

    Directory of Open Access Journals (Sweden)

    Cosmina Mădălina Pop

    2008-10-01

    Full Text Available The financing decision alongside with investment decision represents the core of financial management. The paper presents the main steps in realizing investment projects, the option criteria used in pre-investment analysis, the techniques of financing the investments project, but also the methods of evaluation applied in selecting the most suitable project, accordingly with company’s developlment strategy and policy.

  8. THE SPALLATION NEUTRON SOURCE PROJECT - PHYSICAL CHALLENGES.

    Energy Technology Data Exchange (ETDEWEB)

    WEI,J.

    2002-06-03

    The Spallation Neutron Source (SNS) is designed to reach an average proton beam power of 1.4 MW for pulsed neutron production. This paper summarizes design aspects and physical challenges to the project.

  9. Saharan dust contribution to PM levels: The EC LIFE+ DIAPASON project

    Science.gov (United States)

    Gobbi, G. P.; Wille, H.; Sozzi, R.; Angelini, F.; Barnaba, F.; Costabile, F.; Frey, S.; Bolignano, A.; Di Giosa, A.

    2012-04-01

    The contribution of Saharan-dust advections to both daily and annual PM average values can be significant all over Southern Europe. The most important effects of dust on the number of PM exceedances are mostly observed in polluted areas and large cities. While a wide literature exists documenting episodes of Saharan dust transport towards the Euro-Mediterranean region and Europe in general, a limited number of studies are still available providing statistically significant results on the impact of Saharan dust on the particulate matter loads over the continent. A four-year (2001-2004) study performed in Rome (Italy) found these events to contribute to the average ground PM10 with about 15±10 µg/m3 on about 17% of the days in a year. Since the PM10 yearly average of many traffic stations in Rome is close to 40 μg/m3, these events can cause the PM10 concentration to exceed air quality limit values (50 μg/m3 as daily average) set by the EU Air Quality Directive 2008/50/EC. Although the European legislation allows Member States to subtract the contribution of natural sources before counting PM10 exceedances, definition of an optimal methodology to quantitatively assess such contribution is still in progress. On the basis of the current European Guidelines on the assessment of natural contributions to PM, the DIAPASON project ("Desert-dust Impact on Air quality through model-Predictions and Advanced Sensors ObservatioNs", recently funded under the EC LIFE+ program) has been formulated to provide a robust, user-oriented methodology to assess the presence of desert dust and its contribution to PM levels. To this end, in addition to satellite-based data and model forecasts, the DIAPASON methodology will employ innovative and affordable technologies, partly prototyped within the project itself, as an operational Polarization Lidar-Ceilometer (laser radar) capable of detecting and profiling dust clouds from the ground up to 10 km altitude. The DIAPASON Project (2011

  10. Univariate Lp and ɭ p Averaging, 0 < p < 1, in Polynomial Time by Utilization of Statistical Structure

    Directory of Open Access Journals (Sweden)

    John E. Lavery

    2012-10-01

    Full Text Available We present evidence that one can calculate generically combinatorially expensive Lp and lp averages, 0 < p < 1, in polynomial time by restricting the data to come from a wide class of statistical distributions. Our approach differs from the approaches in the previous literature, which are based on a priori sparsity requirements or on accepting a local minimum as a replacement for a global minimum. The functionals by which Lp averages are calculated are not convex but are radially monotonic and the functionals by which lp averages are calculated are nearly so, which are the keys to solvability in polynomial time. Analytical results for symmetric, radially monotonic univariate distributions are presented. An algorithm for univariate lp averaging is presented. Computational results for a Gaussian distribution, a class of symmetric heavy-tailed distributions and a class of asymmetric heavy-tailed distributions are presented. Many phenomena in human-based areas are increasingly known to be represented by data that have large numbers of outliers and belong to very heavy-tailed distributions. When tails of distributions are so heavy that even medians (L1 and l1 averages do not exist, one needs to consider using lp minimization principles with 0 < p < 1.

  11. Artificial neural network optimisation for monthly average daily global solar radiation prediction

    International Nuclear Information System (INIS)

    Alsina, Emanuel Federico; Bortolini, Marco; Gamberi, Mauro; Regattieri, Alberto

    2016-01-01

    Highlights: • Prediction of the monthly average daily global solar radiation over Italy. • Multi-location Artificial Neural Network (ANN) model: 45 locations considered. • Optimal ANN configuration with 7 input climatologic/geographical parameters. • Statistical indicators: MAPE, NRMSE, MPBE. - Abstract: The availability of reliable climatologic data is essential for multiple purposes in a wide set of anthropic activities and operative sectors. Frequently direct measures present spatial and temporal lacks so that predictive approaches become of interest. This paper focuses on the prediction of the Monthly Average Daily Global Solar Radiation (MADGSR) over Italy using Artificial Neural Networks (ANNs). Data from 45 locations compose the multi-location ANN training and testing sets. For each location, 13 input parameters are considered, including the geographical coordinates and the monthly values for the most frequently adopted climatologic parameters. A subset of 17 locations is used for ANN training, while the testing step is against data from the remaining 28 locations. Furthermore, the Automatic Relevance Determination method (ARD) is used to point out the most relevant input for the accurate MADGSR prediction. The ANN best configuration includes 7 parameters, only, i.e. Top of Atmosphere (TOA) radiation, day length, number of rainy days and average rainfall, latitude and altitude. The correlation performances, expressed through statistical indicators as the Mean Absolute Percentage Error (MAPE), range between 1.67% and 4.25%, depending on the number and type of the chosen input, representing a good solution compared to the current standards.

  12. Averaged multivalued solutions and time discretization for conservation laws

    International Nuclear Information System (INIS)

    Brenier, Y.

    1985-01-01

    It is noted that the correct shock solutions can be approximated by averaging in some sense the multivalued solution given by the method of characteristics for the nonlinear scalar conservation law (NSCL). A time discretization for the NSCL equation based on this principle is considered. An equivalent analytical formulation is shown to lead quite easily to a convergence result, and a third formulation is introduced which can be generalized for the systems of conservation laws. Various numerical schemes are constructed from the proposed time discretization. The first family of schemes is obtained by using a spatial grid and projecting the results of the time discretization. Many known schemes are then recognized (mainly schemes by Osher, Roe, and LeVeque). A second way to discretize leads to a particle scheme without space grid, which is very efficient (at least in the scalar case). Finally, a close relationship between the proposed method and the Boltzmann type schemes is established. 14 references

  13. An active learning representative subset selection method using net analyte signal

    Science.gov (United States)

    He, Zhonghai; Ma, Zhenhe; Luan, Jingmin; Cai, Xi

    2018-05-01

    To guarantee accurate predictions, representative samples are needed when building a calibration model for spectroscopic measurements. However, in general, it is not known whether a sample is representative prior to measuring its concentration, which is both time-consuming and expensive. In this paper, a method to determine whether a sample should be selected into a calibration set is presented. The selection is based on the difference of Euclidean norm of net analyte signal (NAS) vector between the candidate and existing samples. First, the concentrations and spectra of a group of samples are used to compute the projection matrix, NAS vector, and scalar values. Next, the NAS vectors of candidate samples are computed by multiplying projection matrix with spectra of samples. Scalar value of NAS is obtained by norm computation. The distance between the candidate set and the selected set is computed, and samples with the largest distance are added to selected set sequentially. Last, the concentration of the analyte is measured such that the sample can be used as a calibration sample. Using a validation test, it is shown that the presented method is more efficient than random selection. As a result, the amount of time and money spent on reference measurements is greatly reduced.

  14. Human-experienced temperature changes exceed global average climate changes for all income groups

    Science.gov (United States)

    Hsiang, S. M.; Parshall, L.

    2009-12-01

    Global climate change alters local climates everywhere. Many climate change impacts, such as those affecting health, agriculture and labor productivity, depend on these local climatic changes, not global mean change. Traditional, spatially averaged climate change estimates are strongly influenced by the response of icecaps and oceans, providing limited information on human-experienced climatic changes. If used improperly by decision-makers, these estimates distort estimated costs of climate change. We overlay the IPCC’s 20 GCM simulations on the global population distribution to estimate local climatic changes experienced by the world population in the 21st century. The A1B scenario leads to a well-known rise in global average surface temperature of +2.0°C between the periods 2011-2030 and 2080-2099. Projected on the global population distribution in 2000, the median human will experience an annual average rise of +2.3°C (4.1°F) and the average human will experience a rise of +2.4°C (4.3°F). Less than 1% of the population will experience changes smaller than +1.0°C (1.8°F), while 25% and 10% of the population will experience changes greater than +2.9°C (5.2°F) and +3.5°C (6.2°F) respectively. 67% of the world population experiences temperature changes greater than the area-weighted average change of +2.0°C (3.6°F). Using two approaches to characterize the spatial distribution of income, we show that the wealthiest, middle and poorest thirds of the global population experience similar changes, with no group dominating the global average. Calculations for precipitation indicate that there is little change in average precipitation, but redistributions of precipitation occur in all income groups. These results suggest that economists and policy-makers using spatially averaged estimates of climate change to approximate local changes will systematically and significantly underestimate the impacts of climate change on the 21st century population. Top: The

  15. Averaging and sampling for magnetic-observatory hourly data

    Directory of Open Access Journals (Sweden)

    J. J. Love

    2010-11-01

    Full Text Available A time and frequency-domain analysis is made of the effects of averaging and sampling methods used for constructing magnetic-observatory hourly data values. Using 1-min data as a proxy for continuous, geomagnetic variation, we construct synthetic hourly values of two standard types: instantaneous "spot" measurements and simple 1-h "boxcar" averages. We compare these average-sample types with others: 2-h average, Gaussian, and "brick-wall" low-frequency-pass. Hourly spot measurements provide a statistically unbiased representation of the amplitude range of geomagnetic-field variation, but as a representation of continuous field variation over time, they are significantly affected by aliasing, especially at high latitudes. The 1-h, 2-h, and Gaussian average-samples are affected by a combination of amplitude distortion and aliasing. Brick-wall values are not affected by either amplitude distortion or aliasing, but constructing them is, in an operational setting, relatively more difficult than it is for other average-sample types. It is noteworthy that 1-h average-samples, the present standard for observatory hourly data, have properties similar to Gaussian average-samples that have been optimized for a minimum residual sum of amplitude distortion and aliasing. For 1-h average-samples from medium and low-latitude observatories, the average of the combination of amplitude distortion and aliasing is less than the 5.0 nT accuracy standard established by Intermagnet for modern 1-min data. For medium and low-latitude observatories, average differences between monthly means constructed from 1-min data and monthly means constructed from any of the hourly average-sample types considered here are less than the 1.0 nT resolution of standard databases. We recommend that observatories and World Data Centers continue the standard practice of reporting simple 1-h-average hourly values.

  16. The average size of ordered binary subgraphs

    NARCIS (Netherlands)

    van Leeuwen, J.; Hartel, Pieter H.

    To analyse the demands made on the garbage collector in a graph reduction system, the change in size of an average graph is studied when an arbitrary edge is removed. In ordered binary trees the average number of deleted nodes as a result of cutting a single edge is equal to the average size of a

  17. The Design of Project Management Structural Organization

    OpenAIRE

    Dumitru Constantinescu; Cristian Etegan

    2007-01-01

    The relationships organization-suppliers-customers have recently known major changes in the structure of services and have made the organization develop its managerial and professional competencies in order to do projects. The qualified organization is the most trust-worthy in the process of doing a project. The participation of an organization in doing projects depends on a multitude of factors. Out of these factors, the structural organization comes forth, as it represents the variable with...

  18. The average cost of measles cases and adverse events following vaccination in industrialised countries

    Directory of Open Access Journals (Sweden)

    Kou Ulla

    2002-09-01

    Full Text Available Abstract Background Even though the annual incidence rate of measles has dramatically decreased in industrialised countries since the implementation of universal immunisation programmes, cases continue to occur in countries where endemic measles transmission has been interrupted and in countries where adequate levels of immunisation coverage have not been maintained. The objective of this study is to develop a model to estimate the average cost per measles case and per adverse event following measles immunisation using the Netherlands (NL, the United Kingdom (UK and Canada as examples. Methods Parameter estimates were based on a review of the published literature. A decision tree was built to represent the complications associated with measles cases and adverse events following imminisation. Monte-Carlo Simulation techniques were used to account for uncertainty. Results From the perspective of society, we estimated the average cost per measles case to be US$276, US$307 and US$254 for the NL, the UK and Canada, respectively, and the average cost of adverse events following immunisation per vaccinee to be US$1.43, US$1.93 and US$1.51 for the NL, UK and Canada, respectively. Conclusions These average cost estimates could be combined with incidence estimates and costs of immunisation programmes to provide estimates of the cost of measles to industrialised countries. Such estimates could be used as a basis to estimate the potential economic gains of global measles eradication.

  19. Project Guardian: Optimizing Electronic Warfare Systems for Ground Combat Vehicles

    National Research Council Canada - National Science Library

    Parks, Jack G; Jackson, William; Revello, James; Soltesz, James

    1995-01-01

    .... The study, Project Guardian, represents a new process for determining the optimum set of sensors and countermeasures for a specific vehicle class under the constraints of threat projection, combat...

  20. The database of the PREDICTS (Projecting Responses of Ecological Diversity In Changing Terrestrial Systems) project

    OpenAIRE

    Hudson, L. N.; Newbold, T.; Contu, S.; Hill, S. L.; Lysenko, I.; De Palma, A.; Phillips, H. R.; Alhusseini, T. I.; Bedford, F. E.; Bennett, D. J.; Booth, H.; Burton, V. J.; Chng, C. W.; Choimes, A.; Correia, D. L.

    2017-01-01

    The PREDICTS project-Projecting Responses of Ecological Diversity In Changing Terrestrial Systems (www.predicts.org.uk)-has collated from published studies a large, reasonably representative database of comparable samples of biodiversity from multiple sites that differ in the nature or intensity of human impacts relating to land use. We have used this evidence base to develop global and regional statistical models of how local biodiversity responds to these measures. We describe and make free...

  1. Transfer metrics analytics project

    CERN Document Server

    Matonis, Zygimantas

    2016-01-01

    This report represents work done towards predicting transfer rates/latencies on Worldwide LHC Computing Grid (WLCG) sites using Machine Learning techniques. Topic covered are technologies used for the project, data preparation for ML suitable format and attribute selection as well as a comparison of different ML algorithms.

  2. Novel MGF-based expressions for the average bit error probability of binary signalling over generalized fading channels

    KAUST Repository

    Yilmaz, Ferkan

    2014-04-01

    The main idea in the moment generating function (MGF) approach is to alternatively express the conditional bit error probability (BEP) in a desired exponential form so that possibly multi-fold performance averaging is readily converted into a computationally efficient single-fold averaging - sometimes into a closed-form - by means of using the MGF of the signal-to-noise ratio. However, as presented in [1] and specifically indicated in [2] and also to the best of our knowledge, there does not exist an MGF-based approach in the literature to represent Wojnar\\'s generic BEP expression in a desired exponential form. This paper presents novel MGF-based expressions for calculating the average BEP of binary signalling over generalized fading channels, specifically by expressing Wojnar\\'s generic BEP expression in a desirable exponential form. We also propose MGF-based expressions to explore the amount of dispersion in the BEP for binary signalling over generalized fading channels.

  3. Half Moon Cove Tidal Project. Feasibility report

    Energy Technology Data Exchange (ETDEWEB)

    1980-11-01

    The proposed Half Moon Cove Tidal Power Project would be located in a small cove in the northern part of Cobscook Bay in the vicinity of Eastport, Maine. The project would be the first tidal electric power generating plant in the United States of America. The basin impounded by the barrier when full will approximate 1.2 square miles. The average tidal range at Eastport is 18.2 feet. The maximum spring tidal range will be 26.2 feet and the neap tidal range 12.8 feet. The project will be of the single pool-type single effect in which generation takes place on the ebb tide only. Utilizing an average mean tidal range of 18.2 feet the mode of operation enables generation for approximately ten and one-half (10-1/2) hours per day or slightly in excess of five (5) hours per tide. The installed capacity will be 12 MW utilizing 2 to 6 MW units. An axial flow, or Bulb type of turbine was selected for this study.

  4. Project valuation when there are two cashflow streams

    International Nuclear Information System (INIS)

    Emhjellen, Magne; Alaouze, Chris M.

    2002-01-01

    Some authors advocate the separate discounting of different cashflows when calculating net present value (NPV). However, some textbooks (Principles of Corporate Finance, Financial Theory and Corporate Policy: 3rd ed.) focus on calculating NPV by discounting the expected net after tax cashflow using the weighted average cost of capital (WACC) as the discount rate. We show that discounting the expected net after tax cashflow of a project using the WACC yields an incorrect project NPV. A new method for calculating project NPV's using a separate cashflow discounting method is proposed and applied to calculating the NPV's of some North Sea oil projects

  5. EFFECTS OF NON-METHANE HYDROCARBONS ON LOWER STRATOSPHERIC AND UPPER TROPOSPHERIC 2-D ZONAL AVERAGE MODEL CLIMATOLOGY. (R826384)

    Science.gov (United States)

    The perspectives, information and conclusions conveyed in research project abstracts, progress reports, final reports, journal abstracts and journal publications convey the viewpoints of the principal investigator and may not represent the views and policies of ORD and EPA. Concl...

  6. RESEARCH ON COMPLEX, LARGE INDUSTRIAL PROJECTS IN TRANSNATIONAL ENVIRONMENT

    Directory of Open Access Journals (Sweden)

    Florin POPESCU

    2016-12-01

    Full Text Available More and more projects from different industrial sectors developed in transnational environment are becoming more characterized as "complex". In recent years, there has been much discussion and controversy about the complexity of the projects, and, despite what has been written and said in various papers, journals and professional conferences, more confusion than clarification was created, complexity of projects being interpreted differently from one author to another. Most of the literature studied is based on linear, analytical and rational approach, focusing on the size of project management planning and control and actually less on projects that are characterized as taking place and grow into a dynamic socio-human environment in a continuous change. This study represents a critical review of existing theoretical models found in literature, highlighting their limitations. The output of this literature study represents an integration of different approaches concerning complexity under one umbrella to provide a common understanding of the evolution of this concept.

  7. Use of CFD modeling for estimating spatial representativeness of urban air pollution monitoring sites and suitability of their locations

    International Nuclear Information System (INIS)

    Santiago, J. L.; Martin, F.

    2015-01-01

    A methodology to estimate the spatial representativeness of air pollution monitoring sites is applied to two urban districts. This methodology is based on high resolution maps of air pollution computed by using Computational Fluid Dynamics (CFD) modelling tools. Traffic-emitted NO 2 dispersion is simulated for several meteorological conditions taking into account the effect of the buildings on air flow and pollutant dispersion and using a steady state CFD-RANS approach. From these results, maps of average pollutant concentrations for January -May 2011 are computed as a combination of the simulated scenarios. Two urban districts of Madrid City were simulated. Spatial representativeness areas for 32 different sites within the same district (including the site of the operative air quality stations) have been estimated by computing the portion of the domains with average NO 2 concentration differing less than a 20% of the concentration at each candidate monitoring site. New parameters such as the ratio AR between the representativeness area and the whole domain area or the representativeness index (IR) has been proposed to discuss and compare the representativeness areas. Significant differences between the spatial representativeness of the candidate sites of both studied districts have been found. The sites of the Escuelas Aguirre district have generally smaller representativeness areas than those of the Plaza de Castilla. More stations are needed to cover the Escuelas Aguirre district than for the Plaza de Castilla one. The operative air quality station of the Escuelas Aguirre district is less representative than the station of the Plaza de Castilla district. The cause of these differences seems to be the differences in urban structure of both districts prompting different ventilation. (Author)

  8. Use of CFD modeling for estimating spatial representativeness of urban air pollution monitoring sites and suitability of their locations

    Energy Technology Data Exchange (ETDEWEB)

    Santiago, J. L.; Martin, F.

    2015-07-01

    A methodology to estimate the spatial representativeness of air pollution monitoring sites is applied to two urban districts. This methodology is based on high resolution maps of air pollution computed by using Computational Fluid Dynamics (CFD) modelling tools. Traffic-emitted NO{sub 2} dispersion is simulated for several meteorological conditions taking into account the effect of the buildings on air flow and pollutant dispersion and using a steady state CFD-RANS approach. From these results, maps of average pollutant concentrations for January -May 2011 are computed as a combination of the simulated scenarios. Two urban districts of Madrid City were simulated. Spatial representativeness areas for 32 different sites within the same district (including the site of the operative air quality stations) have been estimated by computing the portion of the domains with average NO{sub 2} concentration differing less than a 20% of the concentration at each candidate monitoring site. New parameters such as the ratio AR between the representativeness area and the whole domain area or the representativeness index (IR) has been proposed to discuss and compare the representativeness areas. Significant differences between the spatial representativeness of the candidate sites of both studied districts have been found. The sites of the Escuelas Aguirre district have generally smaller representativeness areas than those of the Plaza de Castilla. More stations are needed to cover the Escuelas Aguirre district than for the Plaza de Castilla one. The operative air quality station of the Escuelas Aguirre district is less representative than the station of the Plaza de Castilla district. The cause of these differences seems to be the differences in urban structure of both districts prompting different ventilation. (Author)

  9. Use of CFD modeling for estimating spatial representativeness of urban air pollution monitoring sites and suitability of their locations

    Energy Technology Data Exchange (ETDEWEB)

    Santiago, J.L.; Martin, F.

    2015-07-01

    A methodology to estimate the spatial representativeness of air pollution monitoring sites is applied to two urban districts. This methodology is based on high resolution maps of air pollution computed by using Computational Fluid Dynamics (CFD) modelling tools. Traffic-emitted NO2 dispersion is simulated for several meteorological conditions taking into account the effect of the buildings on air flow and pollutant dispersion and using a steady state CFD-RANS approach. From these results, maps of average pollutant concentrations for January–May 2011 are computed as a combination of the simulated scenarios. Two urban districts of Madrid City were simulated. Spatial representativeness areas for 32 different sites within the same district (including the site of the operative air quality stations) have been estimated by computing the portion of the domains with average NO2 concentration differing less than a 20% of the concentration at each candidate monitoring site. New parameters such as the ratio AR between the representativeness area and the whole domain area or the representativeness index (IR) has been proposed to discuss and compare the representativeness areas. Significant differences between the spatial representativeness of the candidate sites of both studied districts have been found. The sites of the Escuelas Aguirre district have generally smaller representativeness areas than those of the Plaza de Castilla. More stations are needed to cover the Escuelas Aguirre district than for the Plaza de Castilla one. The operative air quality station of the Escuelas Aguirre district is less representative than the station of the Plaza de Castilla district. The cause of these differences seems to be the differences in urban structure of both districts prompting different ventilation. (Author)

  10. Delineation of facial archetypes by 3d averaging.

    Science.gov (United States)

    Shaweesh, Ashraf I; Thomas, C David L; Bankier, Agnes; Clement, John G

    2004-10-01

    The objective of this study was to investigate the feasibility of creating archetypal 3D faces through computerized 3D facial averaging. A 3D surface scanner Fiore and its software were used to acquire the 3D scans of the faces while 3D Rugle3 and locally-developed software generated the holistic facial averages. 3D facial averages were created from two ethnic groups; European and Japanese and from children with three previous genetic disorders; Williams syndrome, achondroplasia and Sotos syndrome as well as the normal control group. The method included averaging the corresponding depth (z) coordinates of the 3D facial scans. Compared with other face averaging techniques there was not any warping or filling in the spaces by interpolation; however, this facial average lacked colour information. The results showed that as few as 14 faces were sufficient to create an archetypal facial average. In turn this would make it practical to use face averaging as an identification tool in cases where it would be difficult to recruit a larger number of participants. In generating the average, correcting for size differences among faces was shown to adjust the average outlines of the facial features. It is assumed that 3D facial averaging would help in the identification of the ethnic status of persons whose identity may not be known with certainty. In clinical medicine, it would have a great potential for the diagnosis of syndromes with distinctive facial features. The system would also assist in the education of clinicians in the recognition and identification of such syndromes.

  11. Averaging Robertson-Walker cosmologies

    International Nuclear Information System (INIS)

    Brown, Iain A.; Robbers, Georg; Behrend, Juliane

    2009-01-01

    The cosmological backreaction arises when one directly averages the Einstein equations to recover an effective Robertson-Walker cosmology, rather than assuming a background a priori. While usually discussed in the context of dark energy, strictly speaking any cosmological model should be recovered from such a procedure. We apply the scalar spatial averaging formalism for the first time to linear Robertson-Walker universes containing matter, radiation and dark energy. The formalism employed is general and incorporates systems of multiple fluids with ease, allowing us to consider quantitatively the universe from deep radiation domination up to the present day in a natural, unified manner. Employing modified Boltzmann codes we evaluate numerically the discrepancies between the assumed and the averaged behaviour arising from the quadratic terms, finding the largest deviations for an Einstein-de Sitter universe, increasing rapidly with Hubble rate to a 0.01% effect for h = 0.701. For the ΛCDM concordance model, the backreaction is of the order of Ω eff 0 ≈ 4 × 10 −6 , with those for dark energy models being within a factor of two or three. The impacts at recombination are of the order of 10 −8 and those in deep radiation domination asymptote to a constant value. While the effective equations of state of the backreactions in Einstein-de Sitter, concordance and quintessence models are generally dust-like, a backreaction with an equation of state w eff < −1/3 can be found for strongly phantom models

  12. The OPTHER Project: Progress toward the THz Amplifier

    DEFF Research Database (Denmark)

    Paoloni, C; Brunetti, F; Di Carlo, A

    2011-01-01

    This paper describes the status of the OPTHER (OPtically driven TeraHertz AmplifiERs) project and progress toward the THz amplifier realization. This project represents a considerable advancement in the field of high frequency amplification. The design and realization of a THz amplifier within...... this project is a consolidation of efforts at the international level from the leading scientific and industrial European organizations working with vacuum electronics....

  13. SPECIFICITY OF THE PROJECTIVE FIELD: REVERIES AND TRANSFORMATIVE POTENTIALITIES

    OpenAIRE

    Tiziana Sola

    2014-01-01

    This article suggests a reading of the projectives Methods in Bionian key, with particular reference to the bionian elaboration by Antonino Ferro, who substantially considers the patient’s associative flows as forms of the oneiric. The projective situation also represents a place of induction to reverie, in resonance with the concept of projective field, the peculiarity of which lie in the introduction of the element “third”, i.e. the test material.Keywords: Projective methods - Projective fi...

  14. The average baboon brain: MRI templates and tissue probability maps from 89 individuals.

    Science.gov (United States)

    Love, Scott A; Marie, Damien; Roth, Muriel; Lacoste, Romain; Nazarian, Bruno; Bertello, Alice; Coulon, Olivier; Anton, Jean-Luc; Meguerditchian, Adrien

    2016-05-15

    The baboon (Papio) brain is a remarkable model for investigating the brain. The current work aimed at creating a population-average baboon (Papio anubis) brain template and its left/right hemisphere symmetric version from a large sample of T1-weighted magnetic resonance images collected from 89 individuals. Averaging the prior probability maps output during the segmentation of each individual also produced the first baboon brain tissue probability maps for gray matter, white matter and cerebrospinal fluid. The templates and the tissue probability maps were created using state-of-the-art, freely available software tools and are being made freely and publicly available: http://www.nitrc.org/projects/haiko89/ or http://lpc.univ-amu.fr/spip.php?article589. It is hoped that these images will aid neuroimaging research of the baboon by, for example, providing a modern, high quality normalization target and accompanying standardized coordinate system as well as probabilistic priors that can be used during tissue segmentation. Copyright © 2016 Elsevier Inc. All rights reserved.

  15. The Secure Information Exchange (SIX) Project at the OPCW

    International Nuclear Information System (INIS)

    Gulay, M.; Milenkovic, G.

    2015-01-01

    The Chemical Weapons Convention (CWC) entered into force in 1997 and the member states of the Organisation for the Prohibition of Chemical Weapons (OPCW) have obligations for making declarations under various articles of the convention. These declarations could contain confidential information and until recently the only mechanism to submit confidential information to the OPCW Technical Secretariat was through physical delivery by the permanent representatives of the member states which introduced delays in the exchange of information in general. In 2012, the Technical Secretariat initiated a strategic project to establish a secure electronic transmission channel that could be used as an alternative option for the exchange of information between the Technical Secretariat and the member states. The Secure Information Exchange (SIX) Project has been given priority by the Director-General and it received support from the member states. A core project team comprising representatives of the main business unit, the office of legal affairs, IT security and implementation teams were established. Following a feasibility study and with continuous communication with the representatives of the member states, the pilot phase of the project was completed successfully in 2013. In the near future, the project will go live and the member states and the Technical Secretariat will benefit from this key initiative. This paper aims to provide an overview of the project: the solution approach, data gathered in order to assess the delays in communication through traditional means, IT security and implementation issues as well as the legal considerations. (author)

  16. PUREX transition project case study

    International Nuclear Information System (INIS)

    Jasen, W.G.

    1996-01-01

    In December 1992, the US Department of Energy (DOE) directed that the Plutonium-Uranium Extraction (PUREX) Plant be shut down and deactivated because it was no longer needed to support the nation's production of weapons-grade plutonium. The PUREX/UO 2 Deactivation Project will establish a safe and environmentally secure configuration for the facility and preserve that configuration for 10 years. The 10-year span is used to predict future maintenance requirements and represents the estimated time needed to define, authorize, and initiate the follow-on decontamination and decommissioning activities. Accomplishing the deactivation project involves many activities. Removing major hazards, such as excess chemicals, spent fuel, and residual plutonium are major goals of the project. The scope of the PUREX Transition Project is described within

  17. Big gas project for Australia

    International Nuclear Information System (INIS)

    Jemain, A.

    2005-01-01

    Australia is re-launching its ambitions in liquefied natural gas (LNG) with the Greater Gorgon project of offshore exploitation of the natural gas reserves of the continental shelf of NW Australia. These reserves would represent 200 million tons of LNG which will be exported towards China and USA. The project will cost 11 billion dollars and will yield 2 billion dollars per year. It is managed by a consortium which groups together Chevron Corp. (50%), Shell (25%) and ExxonMobil (25%). Technip company is partner of the project. The China National Offshore Oil Corp (CNOOC) has announced its intention to become also partner of the project, and maybe Japan, South Korea and Taiwan will wish too. Short paper. (J.S.)

  18. Fluxes by eddy correlation over heterogeneous landscape: How shall we apply the Reynolds average?

    Science.gov (United States)

    Dobosy, R.

    2007-12-01

    Top-down estimates of carbon exchange across the earth's surface are implicitly an integral scheme, deriving bulk exchanges over large areas. Bottom-up estimates explicitly integrate the individual components of exchange to derive a bulk value. If these approaches are to be properly compared, their estimates should represent the same quantity. Over heterogeneous landscape, eddy-covariance flux computations from towers or aircraft intended for comparison with top-down approach face a question of the proper definition of the mean or base state, the departures from which yield the fluxes by Reynolds averaging. 1)≠Use a global base state derived over a representative sample of the surface, insensitive to land use. The departure quantities then fail to sum to zero over any subsample representing an individual surface type, violating Reynolds criteria. Yet fluxes derived from such subsamples can be directly composed into a bulk flux, globally satisfying Reynolds criteria. 2)≠Use a different base state for each surface type. satisfying Reynolds criteria individually. Then some of the flux may get missed if a surface's characteristics significantly bias its base state. Base state≠(2) is natural for tower samples. Base state≠(1) is natural for airborne samples over heterogeneous landscape, especially in patches smaller than an appropriate averaging length. It appears (1) incorporates a more realistic sample of the flux, though desirably there would be no practical difference between the two schemes. The schemes are related by the expression w¯*a*)C - w¯'a¯')C = w¯'ã¯)C+ wtilde ¯a¯')C+ wtilde ¯ã¯)C Here w is vertical motion, and a is some scalar, such as CO2. The star denotes departure from the global base state≠(1), and the prime from the base state≠(2), defined only over surface class≠C. The overbar with round bracket denotes average over samples drawn from class≠C, determined by footprint model. Thus a¯')C = 0 but a¯*)C ≠ 0 in general. The

  19. Obtaining Samples Representative of Contaminant Distribution in an Aquifer

    International Nuclear Information System (INIS)

    Schalla, Ronald; Spane, Frank A.; Narbutovskih, Susan M.; Conley, Scott F.; Webber, William D.

    2002-01-01

    Historically, groundwater samples collected from monitoring wells have been assumed to provide average indications of contaminant concentrations within the aquifer over the well-screen interval. In-well flow circulation, heterogeneity in the surrounding aquifer, and the sampling method utilized, however, can significantly impact the representativeness of samples as contaminant indicators of actual conditions within the surrounding aquifer. This paper identifies the need and approaches essential for providing cost-effective and technically meaningful groundwater-monitoring results. Proper design of the well screen interval is critical. An accurate understanding of ambient (non-pumping) flow conditions within the monitoring well is essential for determining the contaminant distribution within the aquifer. The ambient in-well flow velocity, flow direction and volumetric flux rate are key to this understanding. Not only do the ambient flow conditions need to be identified for preferential flow zones, but also the probable changes that will be imposed under dynamic conditions that occur during groundwater sampling. Once the in-well flow conditions are understood, effective sampling can be conducted to obtain representative samples for specific depth zones or zones of interest. The question of sample representativeness has become an important issue as waste minimization techniques such as low flow purging and sampling are implemented to combat the increasing cost of well purging and sampling at many hazardous waste sites. Several technical approaches (e.g., well tracer techniques and flowmeter surveys) can be used to determine in-well flow conditions, and these are discussed with respect to both their usefulness and limitations. Proper fluid extraction methods using minimal, (low) volume and no purge sampling methods that are used to obtain representative samples of aquifer conditions are presented

  20. Evaluation of Service Station Attendant-Auto Care Project.

    Science.gov (United States)

    Cress, Ronald J.

    The project described offers an approach to providing occupational skills to socially and educationally handicapped youth, specifically the skills necessary for a service station attendant in driveway salesmanship and auto care. The 10-page evaluation report presents project goals and objectives with evaluation data (represented graphically) and…

  1. γ-Ray spectrometry of radon in water and the role of radon to representatively sample aquifers

    International Nuclear Information System (INIS)

    Talha, S.A.; Lindsay, R.; Newman, R.T.; Meijer, R.J. de; Maleka, P.P.; Hlatshwayo, I.N.; Mlwilo, N.A.; Mohanty, A.K.

    2008-01-01

    Measurement of radon in water by γ-ray spectrometry using a HPGe detector has been investigated to determine aquifer characteristics. The radon activity concentration is determined by taking the weighted average of the concentrations derived from γ-ray lines associated with 214 Pb and 214 Bi decay. The role of accurate radon data to representatively sample aquifers was also investigated by studying a semi-cased borehole. A simplified physical model describing the change of radon concentration with the pumping time, reproduces the data and predicts the time for representative sampling of the aquifer

  2. 75 FR 52689 - Multifamily Housing Reform and Affordability Act: Projects Eligible for a Restructuring Plan...

    Science.gov (United States)

    2010-08-27

    ... HUD-held financing. Essentially, these eligible projects are: (1) Those with rents that on average... CFR parts 401 and 402 (71 FR 2120). The regulatory definition of an eligible project was originally... the 1998 interim rule, the definition of ``eligible project'' read: Eligible project means a project...

  3. Autoregressive moving average fitting for real standard deviation in Monte Carlo power distribution calculation

    International Nuclear Information System (INIS)

    Ueki, Taro

    2010-01-01

    The noise propagation of tallies in the Monte Carlo power method can be represented by the autoregressive moving average process of orders p and p-1 (ARMA(p,p-1)], where p is an integer larger than or equal to two. The formula of the autocorrelation of ARMA(p,q), p≥q+1, indicates that ARMA(3,2) fitting is equivalent to lumping the eigenmodes of fluctuation propagation in three modes such as the slow, intermediate and fast attenuation modes. Therefore, ARMA(3,2) fitting was applied to the real standard deviation estimation of fuel assemblies at particular heights. The numerical results show that straightforward ARMA(3,2) fitting is promising but a stability issue must be resolved toward the incorporation in the distributed version of production Monte Carlo codes. The same numerical results reveal that the average performance of ARMA(3,2) fitting is equivalent to that of the batch method in MCNP with a batch size larger than one hundred and smaller than two hundred cycles for a 1100 MWe pressurized water reactor. The bias correction of low lag autocovariances in MVP/GMVP is demonstrated to have the potential of improving the average performance of ARMA(3,2) fitting. (author)

  4. 14 CFR 1260.58 - Designation of new technology representative and patent representative.

    Science.gov (United States)

    2010-01-01

    ... 14 Aeronautics and Space 5 2010-01-01 2010-01-01 false Designation of new technology... of new technology representative and patent representative. Designation of New Technology... of this grant entitled “New Technology,” the following named representatives are hereby designated by...

  5. Key performance indicators for successful simulation projects

    OpenAIRE

    Jahangirian, M; Taylor, SJE; Young, T; Robinson, S

    2016-01-01

    There are many factors that may contribute to the successful delivery of a simulation project. To provide a structured approach to assessing the impact various factors have on project success, we propose a top-down framework whereby 15 Key Performance Indicators (KPI) are developed that represent the level of successfulness of simulation projects from various perspectives. They are linked to a set of Critical Success Factors (CSF) as reported in the simulation literature. A single measure cal...

  6. Self-similarity of higher-order moving averages

    Science.gov (United States)

    Arianos, Sergio; Carbone, Anna; Türk, Christian

    2011-10-01

    In this work, higher-order moving average polynomials are defined by straightforward generalization of the standard moving average. The self-similarity of the polynomials is analyzed for fractional Brownian series and quantified in terms of the Hurst exponent H by using the detrending moving average method. We prove that the exponent H of the fractional Brownian series and of the detrending moving average variance asymptotically agree for the first-order polynomial. Such asymptotic values are compared with the results obtained by the simulations. The higher-order polynomials correspond to trend estimates at shorter time scales as the degree of the polynomial increases. Importantly, the increase of polynomial degree does not require to change the moving average window. Thus trends at different time scales can be obtained on data sets with the same size. These polynomials could be interesting for those applications relying on trend estimates over different time horizons (financial markets) or on filtering at different frequencies (image analysis).

  7. Determinants of College Grade Point Averages

    Science.gov (United States)

    Bailey, Paul Dean

    2012-01-01

    Chapter 2: The Role of Class Difficulty in College Grade Point Averages. Grade Point Averages (GPAs) are widely used as a measure of college students' ability. Low GPAs can remove a students from eligibility for scholarships, and even continued enrollment at a university. However, GPAs are determined not only by student ability but also by the…

  8. Evaluation of the quality consistency of powdered poppy capsule extractive by an averagely linear-quantified fingerprint method in combination with antioxidant activities and two compounds analyses.

    Science.gov (United States)

    Zhang, Yujing; Sun, Guoxiang; Hou, Zhifei; Yan, Bo; Zhang, Jing

    2017-12-01

    A novel averagely linear-quantified fingerprint method was proposed and successfully applied to monitor the quality consistency of alkaloids in powdered poppy capsule extractive. Averagely linear-quantified fingerprint method provided accurate qualitative and quantitative similarities for chromatographic fingerprints of Chinese herbal medicines. The stability and operability of the averagely linear-quantified fingerprint method were verified by the parameter r. The average linear qualitative similarity SL (improved based on conventional qualitative "Similarity") was used as a qualitative criterion in the averagely linear-quantified fingerprint method, and the average linear quantitative similarity PL was introduced as a quantitative one. PL was able to identify the difference in the content of all the chemical components. In addition, PL was found to be highly correlated to the contents of two alkaloid compounds (morphine and codeine). A simple flow injection analysis was developed for the determination of antioxidant capacity in Chinese Herbal Medicines, which was based on the scavenging of 2,2-diphenyl-1-picrylhydrazyl radical by antioxidants. The fingerprint-efficacy relationship linking chromatographic fingerprints and antioxidant activities was investigated utilizing orthogonal projection to latent structures method, which provided important pharmacodynamic information for Chinese herbal medicines quality control. In summary, quantitative fingerprinting based on averagely linear-quantified fingerprint method can be applied for monitoring the quality consistency of Chinese herbal medicines, and the constructed orthogonal projection to latent structures model is particularly suitable for investigating the fingerprint-efficacy relationship. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  9. Amoco-US Environmental Protection Agency, pollution prevention project, Yorktown, Virginia: Project peer review. Report of the Peer Review Committee of the Amoco/EPA Pollution Prevention Project at the Yorktown, Virginia refinery

    International Nuclear Information System (INIS)

    Klee, H.; Podar, M.

    1991-01-01

    The Amoco/EPA Pollution Prevention Project involved a number of representatives from federal and Virginia regulatory agencies, and Amoco's refining business. Participants believed that the Project could benefit from a broader perspective than these organizations along might provide. The Project Work Group selected an independent Peer Review Process which was conducted by Resource for the Future (RFF), a Washington DC think tank. A group of technical, policy and environmental experts from diverse backgrounds served as Peer Review members. The Peer Review Committee met on three occasions to discuss (1) the Project Work Plan (2) sampling data and interpretation and (3) project conclusions and recommendations. The focus of the meeting was on the general scope and content of the project

  10. Experiences of operational costs of HPV vaccine delivery strategies in Gavi-supported demonstration projects

    Science.gov (United States)

    Holroyd, Taylor; Nanda, Shreya; Bloem, Paul; Griffiths, Ulla K.; Sidibe, Anissa; Hutubessy, Raymond C. W.

    2017-01-01

    From 2012 to 2016, Gavi, the Vaccine Alliance, provided support for countries to conduct small-scale demonstration projects for the introduction of the human papillomavirus vaccine, with the aim of determining which human papillomavirus vaccine delivery strategies might be effective and sustainable upon national scale-up. This study reports on the operational costs and cost determinants of different vaccination delivery strategies within these projects across twelve countries using a standardized micro-costing tool. The World Health Organization Cervical Cancer Prevention and Control Costing Tool was used to collect costing data, which were then aggregated and analyzed to assess the costs and cost determinants of vaccination. Across the one-year demonstration projects, the average economic and financial costs per dose amounted to US$19.98 (standard deviation ±12.5) and US$8.74 (standard deviation ±5.8), respectively. The greatest activities representing the greatest share of financial costs were social mobilization at approximately 30% (range, 6–67%) and service delivery at about 25% (range, 3–46%). Districts implemented varying combinations of school-based, facility-based, or outreach delivery strategies and experienced wide variation in vaccine coverage, drop-out rates, and service delivery costs, including transportation costs and per diems. Size of target population, number of students per school, and average length of time to reach an outreach post influenced cost per dose. Although the operational costs from demonstration projects are much higher than those of other routine vaccine immunization programs, findings from our analysis suggest that HPV vaccination operational costs will decrease substantially for national introduction. Vaccination costs may be decreased further by annual vaccination, high initial investment in social mobilization, or introducing/strengthening school health programs. Our analysis shows that drivers of cost are dependent on

  11. Rotational averaging of multiphoton absorption cross sections

    Energy Technology Data Exchange (ETDEWEB)

    Friese, Daniel H., E-mail: daniel.h.friese@uit.no; Beerepoot, Maarten T. P.; Ruud, Kenneth [Centre for Theoretical and Computational Chemistry, University of Tromsø — The Arctic University of Norway, N-9037 Tromsø (Norway)

    2014-11-28

    Rotational averaging of tensors is a crucial step in the calculation of molecular properties in isotropic media. We present a scheme for the rotational averaging of multiphoton absorption cross sections. We extend existing literature on rotational averaging to even-rank tensors of arbitrary order and derive equations that require only the number of photons as input. In particular, we derive the first explicit expressions for the rotational average of five-, six-, and seven-photon absorption cross sections. This work is one of the required steps in making the calculation of these higher-order absorption properties possible. The results can be applied to any even-rank tensor provided linearly polarized light is used.

  12. Application of Fourier transform infrared spectroscopy and orthogonal projections to latent structures/partial least squares regression for estimation of procyanidins average degree of polymerisation.

    Science.gov (United States)

    Passos, Cláudia P; Cardoso, Susana M; Barros, António S; Silva, Carlos M; Coimbra, Manuel A

    2010-02-28

    Fourier transform infrared (FTIR) spectroscopy has being emphasised as a widespread technique in the quick assess of food components. In this work, procyanidins were extracted with methanol and acetone/water from the seeds of white and red grape varieties. A fractionation by graded methanol/chloroform precipitations allowed to obtain 26 samples that were characterised using thiolysis as pre-treatment followed by HPLC-UV and MS detection. The average degree of polymerisation (DPn) of the procyanidins in the samples ranged from 2 to 11 flavan-3-ol residues. FTIR spectroscopy within the wavenumbers region of 1800-700 cm(-1) allowed to build a partial least squares (PLS1) regression model with 8 latent variables (LVs) for the estimation of the DPn, giving a RMSECV of 11.7%, with a R(2) of 0.91 and a RMSEP of 2.58. The application of orthogonal projection to latent structures (O-PLS1) clarifies the interpretation of the regression model vectors. Moreover, the O-PLS procedure has removed 88% of non-correlated variations with the DPn, allowing to relate the increase of the absorbance peaks at 1203 and 1099 cm(-1) with the increase of the DPn due to the higher proportion of substitutions in the aromatic ring of the polymerised procyanidin molecules. Copyright 2009 Elsevier B.V. All rights reserved.

  13. Project Hanford management contract quality improvement project management plan; TOPICAL

    International Nuclear Information System (INIS)

    ADAMS, D.E.

    1999-01-01

    On July 13, 1998, the U.S. Department of Energy, Richland Operations Office (DOE-RL) Manager transmitted a letter to Fluor Daniel Hanford, Inc. (FDH) describing several DOE-RL identified failed opportunities for FDH to improve the Quality Assurance (QA) Program and its implementation. In addition, DOE-RL identified specific Quality Program performance deficiencies. FDH was requested to establish a periodic reporting mechanism for the corrective action program. In a July 17, 1998 response to DOE-RL, FDH agreed with the DOE concerns and committed to perform a comprehensive review of the Project Hanford Management Contract (PHMC) QA Program during July and August, 1998. As a result, the Project Hanford Management Contract Quality Improvement Plan (QIP) (FDH-3508) was issued on October 21, 1998. The plan identified corrective actions based upon the results of an in-depth Quality Program Assessment. Immediately following the scheduled October 22, 1998, DOE Office of Enforcement and Investigation (EH-10) Enforcement Conference, FDH initiated efforts to effectively implement the QIP corrective actions. A Quality Improvement Project (QI Project) leadership team was assembled to prepare a Project Management Plan for this project. The management plan was specifically designed to engage a core team and the support of representatives from FDH and the major subcontractors (MSCs) to implement the QIP initiatives; identify, correct, and provide feedback as to the root cause for deficiency; and close out the corrective actions. The QI Project will manage and communicate progress of the process

  14. Critical success factors influencing the performance of development projects: An empirical study of Constituency Development Fund projects in Kenya

    Directory of Open Access Journals (Sweden)

    Debadyuti Das

    2017-12-01

    Full Text Available The present work attempts to identify critical success factors (CSFs influencing the performance of development projects based on their key performance indicators (KPIs. It has considered the case of Constituency Development Fund (CDF projects constructed between 2003 and 2011 in Kenya and secured the perceptions of 175 respondents comprising clients, consultants and contractors involved in the implementation of CDF projects on 30 success variables. Findings reveal that individual items constituting these six factors represent six CSFs namely project-related, client-related, consultant-related, contractor-related, supply chain-related, and external environment-related factor. The findings are also relevant to development projects undertaken in other developing countries.

  15. Ergodic averages via dominating processes

    DEFF Research Database (Denmark)

    Møller, Jesper; Mengersen, Kerrie

    2006-01-01

    We show how the mean of a monotone function (defined on a state space equipped with a partial ordering) can be estimated, using ergodic averages calculated from upper and lower dominating processes of a stationary irreducible Markov chain. In particular, we do not need to simulate the stationary...... Markov chain and we eliminate the problem of whether an appropriate burn-in is determined or not. Moreover, when a central limit theorem applies, we show how confidence intervals for the mean can be estimated by bounding the asymptotic variance of the ergodic average based on the equilibrium chain....

  16. Prevalence of sleep duration on an average school night among 4 nationally representative successive samples of American high school students, 2007-2013.

    Science.gov (United States)

    Basch, Charles E; Basch, Corey H; Ruggles, Kelly V; Rajan, Sonali

    2014-12-11

    Consistency, quality, and duration of sleep are important determinants of health. We describe sleep patterns among demographically defined subgroups from the Youth Risk Behavior Surveillance System reported in 4 successive biennial representative samples of American high school students (2007 to 2013). Across the 4 waves of data collection, 6.2% to 7.7% of females and 8.0% to 9.4% of males reported obtaining 9 or more hours of sleep. Insufficient duration of sleep is pervasive among American high school students. Despite substantive public health implications, intervention research on this topic has received little attention.

  17. Tenth working group meeting of representatives of RCA Member States. Report

    International Nuclear Information System (INIS)

    1989-04-01

    The Tenth RCA (Regional Co-operative Agreement for Research, Development and Training Related to Nuclear Science and Technology) Working Group meeting of representatives of RCA Member States was held at the Chinese Academy of Agricultural Sciences, Beijing, between 11-14 April 1988. The report on the meeting consists of a presentation of the seven technical sessions which dealt with topics such as nuclear techniques in industry, agriculture and medicine, nuclear power and of the project reports under RCA

  18. An assessment of uncertainty in forest carbon budget projections

    Science.gov (United States)

    Linda S. Heath; James E. Smith

    2000-01-01

    Estimates of uncertainty are presented for projections of forest carbon inventory and average annual net carbon flux on private timberland in the US using the model FORCARB. Uncertainty in carbon inventory was approximately ±9% (2000 million metric tons) of the estimated median in the year 2000, rising to 11% (2800 million metric tons) in projection year 2040...

  19. Comparing Life-Cycle Costs of ESPCs and Appropriations-Funded Energy Projects: An Update to the 2002 Report

    International Nuclear Information System (INIS)

    Shonder, John A.; Hughes, Patrick; Atkin, Erica

    2006-01-01

    A study was sponsored by FEMP in 2001 - 2002 to develop methods to compare life-cycle costs of federal energy conservation projects carried out through energy savings performance contracts (ESPCs) and projects that are directly funded by appropriations. The study described in this report follows up on the original work, taking advantage of new pricing data on equipment and on $500 million worth of Super ESPC projects awarded since the end of FY 2001. The methods developed to compare life-cycle costs of ESPCs and directly funded energy projects are based on the following tasks: (1) Verify the parity of equipment prices in ESPC vs. directly funded projects; (2) Develop a representative energy conservation project; (3) Determine representative cycle times for both ESPCs and appropriations-funded projects; (4) Model the representative energy project implemented through an ESPC and through appropriations funding; and (5) Calculate the life-cycle costs for each project.

  20. Phantom and Clinical Study of Differences in Cone Beam Computed Tomographic Registration When Aligned to Maximum and Average Intensity Projection

    Energy Technology Data Exchange (ETDEWEB)

    Shirai, Kiyonori [Department of Radiation Oncology, Osaka Medical Center for Cancer and Cardiovascular Diseases, Osaka (Japan); Nishiyama, Kinji, E-mail: sirai-ki@mc.pref.osaka.jp [Department of Radiation Oncology, Osaka Medical Center for Cancer and Cardiovascular Diseases, Osaka (Japan); Katsuda, Toshizo [Department of Radiology, National Cerebral and Cardiovascular Center, Osaka (Japan); Teshima, Teruki; Ueda, Yoshihiro; Miyazaki, Masayoshi; Tsujii, Katsutomo [Department of Radiation Oncology, Osaka Medical Center for Cancer and Cardiovascular Diseases, Osaka (Japan)

    2014-01-01

    Purpose: To determine whether maximum or average intensity projection (MIP or AIP, respectively) reconstructed from 4-dimensional computed tomography (4DCT) is preferred for alignment to cone beam CT (CBCT) images in lung stereotactic body radiation therapy. Methods and Materials: Stationary CT and 4DCT images were acquired with a target phantom at the center of motion and moving along the superior–inferior (SI) direction, respectively. Motion profiles were asymmetrical waveforms with amplitudes of 10, 15, and 20 mm and a 4-second cycle. Stationary CBCT and dynamic CBCT images were acquired in the same manner as stationary CT and 4DCT images. Stationary CBCT was aligned to stationary CT, and the couch position was used as the baseline. Dynamic CBCT was aligned to the MIP and AIP of corresponding amplitudes. Registration error was defined as the SI deviation of the couch position from the baseline. In 16 patients with isolated lung lesions, free-breathing CBCT (FBCBCT) was registered to AIP and MIP (64 sessions in total), and the difference in couch shifts was calculated. Results: In the phantom study, registration errors were within 0.1 mm for AIP and 1.5 to 1.8 mm toward the inferior direction for MIP. In the patient study, the difference in the couch shifts (mean, range) was insignificant in the right-left (0.0 mm, ≤1.0 mm) and anterior–posterior (0.0 mm, ≤2.1 mm) directions. In the SI direction, however, the couch position significantly shifted in the inferior direction after MIP registration compared with after AIP registration (mean, −0.6 mm; ranging 1.7 mm to the superior side and 3.5 mm to the inferior side, P=.02). Conclusions: AIP is recommended as the reference image for registration to FBCBCT when target alignment is performed in the presence of asymmetrical respiratory motion, whereas MIP causes systematic target positioning error.

  1. Phantom and Clinical Study of Differences in Cone Beam Computed Tomographic Registration When Aligned to Maximum and Average Intensity Projection

    International Nuclear Information System (INIS)

    Shirai, Kiyonori; Nishiyama, Kinji; Katsuda, Toshizo; Teshima, Teruki; Ueda, Yoshihiro; Miyazaki, Masayoshi; Tsujii, Katsutomo

    2014-01-01

    Purpose: To determine whether maximum or average intensity projection (MIP or AIP, respectively) reconstructed from 4-dimensional computed tomography (4DCT) is preferred for alignment to cone beam CT (CBCT) images in lung stereotactic body radiation therapy. Methods and Materials: Stationary CT and 4DCT images were acquired with a target phantom at the center of motion and moving along the superior–inferior (SI) direction, respectively. Motion profiles were asymmetrical waveforms with amplitudes of 10, 15, and 20 mm and a 4-second cycle. Stationary CBCT and dynamic CBCT images were acquired in the same manner as stationary CT and 4DCT images. Stationary CBCT was aligned to stationary CT, and the couch position was used as the baseline. Dynamic CBCT was aligned to the MIP and AIP of corresponding amplitudes. Registration error was defined as the SI deviation of the couch position from the baseline. In 16 patients with isolated lung lesions, free-breathing CBCT (FBCBCT) was registered to AIP and MIP (64 sessions in total), and the difference in couch shifts was calculated. Results: In the phantom study, registration errors were within 0.1 mm for AIP and 1.5 to 1.8 mm toward the inferior direction for MIP. In the patient study, the difference in the couch shifts (mean, range) was insignificant in the right-left (0.0 mm, ≤1.0 mm) and anterior–posterior (0.0 mm, ≤2.1 mm) directions. In the SI direction, however, the couch position significantly shifted in the inferior direction after MIP registration compared with after AIP registration (mean, −0.6 mm; ranging 1.7 mm to the superior side and 3.5 mm to the inferior side, P=.02). Conclusions: AIP is recommended as the reference image for registration to FBCBCT when target alignment is performed in the presence of asymmetrical respiratory motion, whereas MIP causes systematic target positioning error

  2. Predicting Student Grade Point Average at a Community College from Scholastic Aptitude Tests and from Measures Representing Three Constructs in Vroom's Expectancy Theory Model of Motivation.

    Science.gov (United States)

    Malloch, Douglas C.; Michael, William B.

    1981-01-01

    This study was designed to determine whether an unweighted linear combination of community college students' scores on standardized achievement tests and a measure of motivational constructs derived from Vroom's expectance theory model of motivation was predictive of academic success (grade point average earned during one quarter of an academic…

  3. FASSET - An European project for environmental protection

    International Nuclear Information System (INIS)

    Bruchertseifer, F.

    2003-01-01

    The European research project ''FASSET'' (Framework ASSessment of Enviromental ImpacT) will provide a framework of the environmental impact of the effects of ionising radiation and will identify protection aims for the environmental protection. This project represent a collaboration of different organisations from the European community: Sweden, Norway, Finland, Great Britain, Spain, France and Germany. The German participants are the GSF-research centre for environment and health and the German radiation protection office. The project is founded by the 5 th EC research programme. The existing national and international radiation protection regulations are focused to the humans. Other species, like plants and animals, are protected indirectly, if their habitat are close to urban areas or they represent a part of the food chain. The ICRP position ''if the man is protected, nature is protected as well'' is now under reconsideration by a ICRP-Taskgroup. The identification of perilled real reference organism by the FASSET-project is an working tool for the definition of the protection aims. For that purpose the project is divided into three working packages: dosimetry, exposure pathways and effects. Another working package is responsible for the developing of the concept using the results provided by the other working packages. (orig.)

  4. Intercultural Education in Primary School: A Collaborative Project

    Science.gov (United States)

    Santos, Marta; Araújo e Sá, Maria Helena; Simões, Ana Raquel

    2014-01-01

    In this article, we present and discuss a collaborative project on intercultural education developed by a group of educational partners. The group was made up of 12 people representing different institutions in the community, namely primary schools, cultural and social associations and the local council. The project takes an intercultural approach…

  5. Multiphase averaging of periodic soliton equations

    International Nuclear Information System (INIS)

    Forest, M.G.

    1979-01-01

    The multiphase averaging of periodic soliton equations is considered. Particular attention is given to the periodic sine-Gordon and Korteweg-deVries (KdV) equations. The periodic sine-Gordon equation and its associated inverse spectral theory are analyzed, including a discussion of the spectral representations of exact, N-phase sine-Gordon solutions. The emphasis is on physical characteristics of the periodic waves, with a motivation from the well-known whole-line solitons. A canonical Hamiltonian approach for the modulational theory of N-phase waves is prescribed. A concrete illustration of this averaging method is provided with the periodic sine-Gordon equation; explicit averaging results are given only for the N = 1 case, laying a foundation for a more thorough treatment of the general N-phase problem. For the KdV equation, very general results are given for multiphase averaging of the N-phase waves. The single-phase results of Whitham are extended to general N phases, and more importantly, an invariant representation in terms of Abelian differentials on a Riemann surface is provided. Several consequences of this invariant representation are deduced, including strong evidence for the Hamiltonian structure of N-phase modulational equations

  6. Probabilistic 21st and 22nd Century Sea-Level Projections at a Global Network of Tide-Gauge Sites

    Science.gov (United States)

    Kopp, Robert E.; Horton, Radley M.; Little, Christopher M.; Mitrovica, Jerry X.; Oppenheimer, Michael; Rasmussen, D. J.; Strauss, Benjamin H.; Tebaldi, Claudia

    2014-01-01

    Sea-level rise due to both climate change and non-climatic factors threatens coastal settlements, infrastructure, and ecosystems. Projections of mean global sea-level (GSL) rise provide insufficient information to plan adaptive responses; local decisions require local projections that accommodate different risk tolerances and time frames and that can be linked to storm surge projections. Here we present a global set of local sea-level (LSL) projections to inform decisions on timescales ranging from the coming decades through the 22nd century. We provide complete probability distributions, informed by a combination of expert community assessment, expert elicitation, and process modeling. Between the years 2000 and 2100, we project a very likely (90% probability) GSL rise of 0.5–1.2?m under representative concentration pathway (RCP) 8.5, 0.4–0.9?m under RCP 4.5, and 0.3–0.8?m under RCP 2.6. Site-to-site differences in LSL projections are due to varying non-climatic background uplift or subsidence, oceanographic effects, and spatially variable responses of the geoid and the lithosphere to shrinking land ice. The Antarctic ice sheet (AIS) constitutes a growing share of variance in GSL and LSL projections. In the global average and at many locations, it is the dominant source of variance in late 21st century projections, though at some sites oceanographic processes contribute the largest share throughout the century. LSL rise dramatically reshapes flood risk, greatly increasing the expected number of “1-in-10” and “1-in-100” year events.

  7. A Tidally Averaged Sediment-Transport Model for San Francisco Bay, California

    Science.gov (United States)

    Lionberger, Megan A.; Schoellhamer, David H.

    2009-01-01

    A tidally averaged sediment-transport model of San Francisco Bay was incorporated into a tidally averaged salinity box model previously developed and calibrated using salinity, a conservative tracer (Uncles and Peterson, 1995; Knowles, 1996). The Bay is represented in the model by 50 segments composed of two layers: one representing the channel (>5-meter depth) and the other the shallows (0- to 5-meter depth). Calculations are made using a daily time step and simulations can be made on the decadal time scale. The sediment-transport model includes an erosion-deposition algorithm, a bed-sediment algorithm, and sediment boundary conditions. Erosion and deposition of bed sediments are calculated explicitly, and suspended sediment is transported by implicitly solving the advection-dispersion equation. The bed-sediment model simulates the increase in bed strength with depth, owing to consolidation of fine sediments that make up San Francisco Bay mud. The model is calibrated to either net sedimentation calculated from bathymetric-change data or measured suspended-sediment concentration. Specified boundary conditions are the tributary fluxes of suspended sediment and suspended-sediment concentration in the Pacific Ocean. Results of model calibration and validation show that the model simulates the trends in suspended-sediment concentration associated with tidal fluctuations, residual velocity, and wind stress well, although the spring neap tidal suspended-sediment concentration variability was consistently underestimated. Model validation also showed poor simulation of seasonal sediment pulses from the Sacramento-San Joaquin River Delta at Point San Pablo because the pulses enter the Bay over only a few days and the fate of the pulses is determined by intra-tidal deposition and resuspension that are not included in this tidally averaged model. The model was calibrated to net-basin sedimentation to calculate budgets of sediment and sediment-associated contaminants. While

  8. Project valuation when there are two cashflow streams

    Energy Technology Data Exchange (ETDEWEB)

    Emhjellen, Magne [School of Economics, Stavanger University College, P.O. Box 2557, Ullandhaug, N-4091, Stavanger (Norway); Alaouze, Chris M. [School of Economics, The University of New South Wales, 2052 Sydney (Australia)

    2002-09-01

    Some authors advocate the separate discounting of different cashflows when calculating net present value (NPV). However, some textbooks (Principles of Corporate Finance, Financial Theory and Corporate Policy: 3rd ed.) focus on calculating NPV by discounting the expected net after tax cashflow using the weighted average cost of capital (WACC) as the discount rate. We show that discounting the expected net after tax cashflow of a project using the WACC yields an incorrect project NPV. A new method for calculating project NPV's using a separate cashflow discounting method is proposed and applied to calculating the NPV's of some North Sea oil projects.

  9. Tendon surveillance requirements - average tendon force

    International Nuclear Information System (INIS)

    Fulton, J.F.

    1982-01-01

    Proposed Rev. 3 to USNRC Reg. Guide 1.35 discusses the need for comparing, for individual tendons, the measured and predicted lift-off forces. Such a comparison is intended to detect any abnormal tendon force loss which might occur. Recognizing that there are uncertainties in the prediction of tendon losses, proposed Guide 1.35.1 has allowed specific tolerances on the fundamental losses. Thus, the lift-off force acceptance criteria for individual tendons appearing in Reg. Guide 1.35, Proposed Rev. 3, is stated relative to a lower bound predicted tendon force, which is obtained using the 'plus' tolerances on the fundamental losses. There is an additional acceptance criterion for the lift-off forces which is not specifically addressed in these two Reg. Guides; however, it is included in a proposed Subsection IWX to ASME Code Section XI. This criterion is based on the overriding requirement that the magnitude of prestress in the containment structure be sufficeint to meet the minimum prestress design requirements. This design requirement can be expressed as an average tendon force for each group of vertical hoop, or dome tendons. For the purpose of comparing the actual tendon forces with the required average tendon force, the lift-off forces measured for a sample of tendons within each group can be averaged to construct the average force for the entire group. However, the individual lift-off forces must be 'corrected' (normalized) prior to obtaining the sample average. This paper derives the correction factor to be used for this purpose. (orig./RW)

  10. New Production Reactor project-management plan

    International Nuclear Information System (INIS)

    McCrosson, F.J.; Hibbard, L.; Buckner, M.R.

    1982-01-01

    This document provides a project management plan for the first phase of a project to design and build a new production reactor (NPR) at SRP. The design of the NPR is based upon proven SRP heavy water reactor design, with several enhancements such as full containment, moderator detritiation, improved cooling, and modernized control rooms and instrumentation. The first phase of the NPR project includes environmental and safety analyses, preparation of the technical data summary and basic data, site studies, engineering studies, and conceptual design. The project management plan was developed by a 14-member task force comprised of representatives from the Technical Division, the Manufacturing Division, the Departmental Engineer's Office, and the Engineering Department

  11. Average Bandwidth Allocation Model of WFQ

    Directory of Open Access Journals (Sweden)

    Tomáš Balogh

    2012-01-01

    Full Text Available We present a new iterative method for the calculation of average bandwidth assignment to traffic flows using a WFQ scheduler in IP based NGN networks. The bandwidth assignment calculation is based on the link speed, assigned weights, arrival rate, and average packet length or input rate of the traffic flows. We prove the model outcome with examples and simulation results using NS2 simulator.

  12. Climate Benefits of Material Recycling: Inventory of Average Greenhouse Gas Emissions for Denmark, Norway and Sweden

    DEFF Research Database (Denmark)

    Hillman, Karl; Damgaard, Anders; Eriksson, Ola

    The purpose of this project is to compare emissions of greenhouse gases from material recycling with those from virgin material production, both from a material supply perspective and from a recycling system perspective. The method for estimating emissions and climate benefits is based on a review......, followed by a selection, of the most relevant publications on life cycle assessment (LCA) of materials for use in Denmark, Norway and Sweden. The proposed averages show that emissions from material recycling are lower in both perspectives, comparing either material supply or complete recycling systems....... The results can be used by companies and industry associations in Denmark, Norway and Sweden to communicate the current climate benefits of material recycling in general. They may also contribute to discussions on a societal level, as long as their average and historic nature is recognised....

  13. Regional averaging and scaling in relativistic cosmology

    International Nuclear Information System (INIS)

    Buchert, Thomas; Carfora, Mauro

    2002-01-01

    Averaged inhomogeneous cosmologies lie at the forefront of interest, since cosmological parameters such as the rate of expansion or the mass density are to be considered as volume-averaged quantities and only these can be compared with observations. For this reason the relevant parameters are intrinsically scale-dependent and one wishes to control this dependence without restricting the cosmological model by unphysical assumptions. In the latter respect we contrast our way to approach the averaging problem in relativistic cosmology with shortcomings of averaged Newtonian models. Explicitly, we investigate the scale-dependence of Eulerian volume averages of scalar functions on Riemannian three-manifolds. We propose a complementary view of a Lagrangian smoothing of (tensorial) variables as opposed to their Eulerian averaging on spatial domains. This programme is realized with the help of a global Ricci deformation flow for the metric. We explain rigorously the origin of the Ricci flow which, on heuristic grounds, has already been suggested as a possible candidate for smoothing the initial dataset for cosmological spacetimes. The smoothing of geometry implies a renormalization of averaged spatial variables. We discuss the results in terms of effective cosmological parameters that would be assigned to the smoothed cosmological spacetime. In particular, we find that on the smoothed spatial domain B-bar evaluated cosmological parameters obey Ω-bar B-bar m + Ω-bar B-bar R + Ω-bar B-bar A + Ω-bar B-bar Q 1, where Ω-bar B-bar m , Ω-bar B-bar R and Ω-bar B-bar A correspond to the standard Friedmannian parameters, while Ω-bar B-bar Q is a remnant of cosmic variance of expansion and shear fluctuations on the averaging domain. All these parameters are 'dressed' after smoothing out the geometrical fluctuations, and we give the relations of the 'dressed' to the 'bare' parameters. While the former provide the framework of interpreting observations with a 'Friedmannian bias

  14. Pedagogy of stereographic projection

    Science.gov (United States)

    Blenkinsop, Tom

    1999-05-01

    Three sorts of problem are encountered by students learning stereographic projection. Lack of familiarity with compass directions and with the specification of planar and linear orientations is the most fundamental problem. The second problem is one of poor visualisation which prohibits the correct understanding of the principles and practice of projection, although students can sometimes perform remarkably well on certain exercises without this understanding. A third problem is a lack of interest or appreciation of the usefulness of the technique. A solution to the first problem is to provide a thorough grounding in compass directions, and concepts of strike, dip, dip direction, plunge and plunge bearing immediately before teaching stereographic projection. Visual aids, such as a transparent hemisphere to represent half of the projection sphere, and a large stereonet, are helpful for visualisation. Estimation and sketching exercises help to develop an understanding of the geometry of projection. Computer-aided teaching and plotting is recommended after an introduction to manual techniques. Interest can be stimulated through the use of exercises based on real geological maps.

  15. Havsnaes wind farm - The project financing of a Swedish wind farm

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2012-07-01

    In March 2008 the ground breaking project financing of the 95 MW Havsnaes wind farm was completed. Havsnaes represents one of the cornerstones in portfolio of Venus Vind, controlled by HgCapital, where sustainable Scandinavian strength is build through industrial scale wind farms with local presence. At the time, Havsnaes represented the largest energy project financing in the market, it is also the first true project financing of a major Swedish wind farm. The aim of this study is to highlight the process of project financing and additional lessons learnt from the Havsnaes transaction. Investment in renewable energy projects often includes international investors. We welcome the growing Swedish wind market, banks and other financial institutions, politicians and other relevant decision makers to take part of our findings. Sponsorship provided by the Swedish Energy Agency has enabled the completion of this study.

  16. Final Technical Report - Modernization of the Boulder Canyon Hydroelectric Project

    Energy Technology Data Exchange (ETDEWEB)

    Taddeucci, Joe [Dept. of Public Works, Boulder, CO (United States). Utilities Division

    2013-03-29

    The Boulder Canyon Hydroelectric Project (BCH) was purchased by the City of Boulder, CO (the city) in 2001. Project facilities were originally constructed in 1910 and upgraded in the 1930s and 1940s. By 2009, the two 10 MW turbine/generators had reached or were nearing the end of their useful lives. One generator had grounded out and was beyond repair, reducing plant capacity to 10 MW. The remaining 10 MW unit was expected to fail at any time. When the BCH power plant was originally constructed, a sizeable water supply was available for the sole purpose of hydroelectric power generation. Between 1950 and 2001, that water supply had gradually been converted to municipal water supply by the city. By 2001, the water available for hydroelectric power generation at BCH could not support even one 10 MW unit. Boulder lacked the financial resources to modernize the facilities, and Boulder anticipated that when the single, operational historical unit failed, the project would cease operation. In 2009, the City of Boulder applied for and received a U.S. Department of Energy (DOE) grant for $1.18 million toward a total estimated project cost of $5.155 million to modernize BCH. The federal funding allowed Boulder to move forward with plant modifications that would ensure BCH would continue operation. Federal funding was made available through the American Recovery and Reinvestment Act (ARRA) of 2009. Boulder determined that a single 5 MW turbine/generator would be the most appropriate capacity, given the reduced water supply to the plant. Average annual BCH generation with the old 10 MW unit had been about 8,500 MW-hr, whereas annual generation with a new, efficient turbine could average 11,000 to 12,000 MW-hr. The incremental change in annual generation represents a 30% increase in generation over pre-project conditions. The old turbine/generator was a single nozzle Pelton turbine with a 5-to-1 flow turndown and a maximum turbine/generator efficiency of 82%. The new unit is a

  17. Cloud radiative effects and changes simulated by the Coupled Model Intercomparison Project Phase 5 models

    Science.gov (United States)

    Shin, Sun-Hee; Kim, Ok-Yeon; Kim, Dongmin; Lee, Myong-In

    2017-07-01

    Using 32 CMIP5 (Coupled Model Intercomparison Project Phase 5) models, this study examines the veracity in the simulation of cloud amount and their radiative effects (CREs) in the historical run driven by observed external radiative forcing for 1850-2005, and their future changes in the RCP (Representative Concentration Pathway) 4.5 scenario runs for 2006-2100. Validation metrics for the historical run are designed to examine the accuracy in the representation of spatial patterns for climatological mean, and annual and interannual variations of clouds and CREs. The models show large spread in the simulation of cloud amounts, specifically in the low cloud amount. The observed relationship between cloud amount and the controlling large-scale environment are also reproduced diversely by various models. Based on the validation metrics, four models—ACCESS1.0, ACCESS1.3, HadGEM2-CC, and HadGEM2-ES—are selected as best models, and the average of the four models performs more skillfully than the multimodel ensemble average. All models project global-mean SST warming at the increase of the greenhouse gases, but the magnitude varies across the simulations between 1 and 2 K, which is largely attributable to the difference in the change of cloud amount and distribution. The models that simulate more SST warming show a greater increase in the net CRE due to reduced low cloud and increased incoming shortwave radiation, particularly over the regions of marine boundary layer in the subtropics. Selected best-performing models project a significant reduction in global-mean cloud amount of about -0.99% K-1 and net radiative warming of 0.46 W m-2 K-1, suggesting a role of positive feedback to global warming.

  18. Integrated project management type contracts

    International Nuclear Information System (INIS)

    Heisler, S.I.

    1975-01-01

    The concept of integrated project management represents a single source to which the owner can turn for all project management functions excepting for those relating to outside parties such as site purchase, personnel selection etc. Other functions such as design, procurement, construction management, schedule and cost control, quality assurance/quality control are usually handled by the integrated project manager as the agent of the owner. The arrangement is flexible and the responsibilities can be varied to suit the size and experience of the owner. Past experience in the United States indicates an increase in the trend toward IPM work and it appears that overseas this trend is developing also. (orig./RW) [de

  19. Average and local structure of α-CuI by configurational averaging

    International Nuclear Information System (INIS)

    Mohn, Chris E; Stoelen, Svein

    2007-01-01

    Configurational Boltzmann averaging together with density functional theory are used to study in detail the average and local structure of the superionic α-CuI. We find that the coppers are spread out with peaks in the atom-density at the tetrahedral sites of the fcc sublattice of iodines. We calculate Cu-Cu, Cu-I and I-I pair radial distribution functions, the distribution of coordination numbers and the distribution of Cu-I-Cu, I-Cu-I and Cu-Cu-Cu bond-angles. The partial pair distribution functions are in good agreement with experimental neutron diffraction-reverse Monte Carlo, extended x-ray absorption fine structure and ab initio molecular dynamics results. In particular, our results confirm the presence of a prominent peak at around 2.7 A in the Cu-Cu pair distribution function as well as a broader, less intense peak at roughly 4.3 A. We find highly flexible bonds and a range of coordination numbers for both iodines and coppers. This structural flexibility is of key importance in order to understand the exceptional conductivity of coppers in α-CuI; the iodines can easily respond to changes in the local environment as the coppers diffuse, and a myriad of different diffusion-pathways is expected due to the large variation in the local motifs

  20. Experimental demonstration of squeezed-state quantum averaging

    DEFF Research Database (Denmark)

    Lassen, Mikael Østergaard; Madsen, Lars Skovgaard; Sabuncu, Metin

    2010-01-01

    We propose and experimentally demonstrate a universal quantum averaging process implementing the harmonic mean of quadrature variances. The averaged variances are prepared probabilistically by means of linear optical interference and measurement-induced conditioning. We verify that the implemented...

  1. Changes in water availability in the Upper Blue Nile basin under the representative concentration pathways scenario

    NARCIS (Netherlands)

    Haile, Alemseged Tamiru; Akawka, Ashenafi Lekasa; Berhanu, Beza; Rientjes, T.H.M.

    2017-01-01

    Climatic and hydrological changes will likely be intensified in the Upper Blue Nile (UBN) basin by the effects of global warming. The extent of such effects for representative concentration pathways (RCP) climate scenarios is unknown. We evaluated projected changes in rainfall and evapotranspiration

  2. Fisheries Restoration Grant Program Projects [ds168

    Data.gov (United States)

    California Natural Resource Agency — This shapefile (FRGP_All_020209.shp) represents the locations of all ongoing and completed salmonid restoration projects in California with existing records in the...

  3. SPECIFICITY OF THE PROJECTIVE FIELD: REVERIES AND TRANSFORMATIVE POTENTIALITIES

    Directory of Open Access Journals (Sweden)

    Tiziana Sola

    2014-01-01

    Full Text Available This article suggests a reading of the projectives Methods in Bionian key, with particular reference to the bionian elaboration by Antonino Ferro, who substantially considers the patient’s associative flows as forms of the oneiric. The projective situation also represents a place of induction to reverie, in resonance with the concept of projective field, the peculiarity of which lie in the introduction of the element “third”, i.e. the test material.Keywords: Projective methods - Projective field – induction to reverie – activity of symbolization

  4. 16 CFR Appendix K to Part 305 - Representative Average Unit Energy Costs

    Science.gov (United States)

    2010-01-01

    ... Btu 1 Electricity 10.65¢/kWh 2,3 $.1065/kWh $31.21 Natural Gas $1.218/therm 4 $12.53/MCF 5,6 $0... kWh stands for kilo Watt hour. 3 1 kWh = 3,412 Btu. 4 1 therm = 100,000 Btu. Natural gas prices...

  5. Aperture averaging in strong oceanic turbulence

    Science.gov (United States)

    Gökçe, Muhsin Caner; Baykal, Yahya

    2018-04-01

    Receiver aperture averaging technique is employed in underwater wireless optical communication (UWOC) systems to mitigate the effects of oceanic turbulence, thus to improve the system performance. The irradiance flux variance is a measure of the intensity fluctuations on a lens of the receiver aperture. Using the modified Rytov theory which uses the small-scale and large-scale spatial filters, and our previously presented expression that shows the atmospheric structure constant in terms of oceanic turbulence parameters, we evaluate the irradiance flux variance and the aperture averaging factor of a spherical wave in strong oceanic turbulence. Irradiance flux variance variations are examined versus the oceanic turbulence parameters and the receiver aperture diameter are examined in strong oceanic turbulence. Also, the effect of the receiver aperture diameter on the aperture averaging factor is presented in strong oceanic turbulence.

  6. Stochastic Averaging and Stochastic Extremum Seeking

    CERN Document Server

    Liu, Shu-Jun

    2012-01-01

    Stochastic Averaging and Stochastic Extremum Seeking develops methods of mathematical analysis inspired by the interest in reverse engineering  and analysis of bacterial  convergence by chemotaxis and to apply similar stochastic optimization techniques in other environments. The first half of the text presents significant advances in stochastic averaging theory, necessitated by the fact that existing theorems are restricted to systems with linear growth, globally exponentially stable average models, vanishing stochastic perturbations, and prevent analysis over infinite time horizon. The second half of the text introduces stochastic extremum seeking algorithms for model-free optimization of systems in real time using stochastic perturbations for estimation of their gradients. Both gradient- and Newton-based algorithms are presented, offering the user the choice between the simplicity of implementation (gradient) and the ability to achieve a known, arbitrary convergence rate (Newton). The design of algorithms...

  7. 20 CFR 404.220 - Average-monthly-wage method.

    Science.gov (United States)

    2010-04-01

    ... 20 Employees' Benefits 2 2010-04-01 2010-04-01 false Average-monthly-wage method. 404.220 Section... INSURANCE (1950- ) Computing Primary Insurance Amounts Average-Monthly-Wage Method of Computing Primary Insurance Amounts § 404.220 Average-monthly-wage method. (a) Who is eligible for this method. You must...

  8. 2011 Joint Science Education Project: Research Experience in Polar Science

    Science.gov (United States)

    Wilkening, J.; Ader, V.

    2011-12-01

    The Joint Science Education Project (JSEP), sponsored by the National Science Foundation, is a two-part program that brings together students and teachers from the United States, Greenland, and Denmark, for a unique cross-cultural, first-hand experience of the realities of polar science field research in Greenland. During JSEP, students experienced research being conducted on and near the Greenland ice sheet by attending researcher presentations, visiting NSF-funded field sites (including Summit and NEEM field stations, both located on the Greenland ice sheet), and designing and conducting research projects in international teams. The results of two of these projects will be highlighted. The atmospheric project investigated the differences in CO2, UVA, UVB, temperature, and albedo in different Arctic microenvironments, while also examining the interaction between the atmosphere and water present in the given environments. It was found that the carbon dioxide levels varied: glacial environments having the lowest levels, with an average concentration of 272.500 ppm, and non-vegetated, terrestrial environments having the highest, with an average concentration of 395.143 ppm. Following up on these results, it is planned to further investigate the interaction of the water and atmosphere, including water's role in the uptake of carbon dioxide. The ecology project investigated the occurrence of unusual large blooms of Nostoc cyanobacteria in Kangerlussuaq area lakes. The water chemistry of the lakes which contained the cyanobacteria and the lakes that did not were compared. The only noticeable difference was of the lakes' acidity, lakes containing the blooms had an average pH value of 8.58, whereas lakes without the blooms had an average pH value of 6.60. Further investigation of these results is needed to determine whether or not this was a cause or effect of the cyanobacteria blooms. As a next step, it is planned to attempt to grow the blooms to monitor their effects on

  9. Testing a groundwater sampling tool: Are the samples representative?

    International Nuclear Information System (INIS)

    Kaback, D.S.; Bergren, C.L.; Carlson, C.A.; Carlson, C.L.

    1989-01-01

    A ground water sampling tool, the HydroPunch trademark, was tested at the Department of Energy's Savannah River Site in South Carolina to determine if representative ground water samples could be obtained without installing monitoring wells. Chemical analyses of ground water samples collected with the HydroPunch trademark from various depths within a borehole were compared with chemical analyses of ground water from nearby monitoring wells. The site selected for the test was in the vicinity of a large coal storage pile and a coal pile runoff basin that was constructed to collect the runoff from the coal storage pile. Existing monitoring wells in the area indicate the presence of a ground water contaminant plume that: (1) contains elevated concentrations of trace metals; (2) has an extremely low pH; and (3) contains elevated concentrations of major cations and anions. Ground water samples collected with the HydroPunch trademark provide in excellent estimate of ground water quality at discrete depths. Groundwater chemical data collected from various depths using the HydroPunch trademark can be averaged to simulate what a screen zone in a monitoring well would sample. The averaged depth-discrete data compared favorably with the data obtained from the nearby monitoring wells

  10. The average cost of pressure ulcer management in a community dwelling spinal cord injury population.

    Science.gov (United States)

    Chan, Brian C; Nanwa, Natasha; Mittmann, Nicole; Bryant, Dianne; Coyte, Peter C; Houghton, Pamela E

    2013-08-01

    Pressure ulcers (PUs) are a common secondary complication experienced by community dwelling individuals with spinal cord injury (SCI). There is a paucity of literature on the health economic impact of PU in SCI population from a societal perspective. The objective of this study was to determine the resource use and costs in 2010 Canadian dollars of a community dwelling SCI individual experiencing a PU from a societal perspective. A non-comparative cost analysis was conducted on a cohort of community dwelling SCI individuals from Ontario, Canada. Medical resource use was recorded over the study period. Unit costs associated with these resources were collected from publicly available sources and published literature. Average monthly cost was calculated based on 7-month follow-up. Costs were stratified by age, PU history, severity level, location of SCI, duration of current PU and PU surface area. Sensitivity analyses were also carried out. Among the 12 study participants, total average monthly cost per community dwelling SCI individual with a PU was $4745. Hospital admission costs represented the greatest percentage of the total cost (62%). Sensitivity analysis showed that the total average monthly costs were most sensitive to variations in hospitalisation costs. © 2012 The Authors. International Wound Journal © 2012 John Wiley & Sons Ltd and Medicalhelplines.com Inc.

  11. Visualizing the uncertainty in the relationship between seasonal average climate and malaria risk.

    Science.gov (United States)

    MacLeod, D A; Morse, A P

    2014-12-02

    Around $1.6 billion per year is spent financing anti-malaria initiatives, and though malaria morbidity is falling, the impact of annual epidemics remains significant. Whilst malaria risk may increase with climate change, projections are highly uncertain and to sidestep this intractable uncertainty, adaptation efforts should improve societal ability to anticipate and mitigate individual events. Anticipation of climate-related events is made possible by seasonal climate forecasting, from which warnings of anomalous seasonal average temperature and rainfall, months in advance are possible. Seasonal climate hindcasts have been used to drive climate-based models for malaria, showing significant skill for observed malaria incidence. However, the relationship between seasonal average climate and malaria risk remains unquantified. Here we explore this relationship, using a dynamic weather-driven malaria model. We also quantify key uncertainty in the malaria model, by introducing variability in one of the first order uncertainties in model formulation. Results are visualized as location-specific impact surfaces: easily integrated with ensemble seasonal climate forecasts, and intuitively communicating quantified uncertainty. Methods are demonstrated for two epidemic regions, and are not limited to malaria modeling; the visualization method could be applied to any climate impact.

  12. Risk Assessment of Pollution Emergencies in Water Source Areas of the Hanjiang-to-Weihe River Diversion Project

    Science.gov (United States)

    Liu, Luyao; Feng, Minquan

    2018-03-01

    [Objective] This study quantitatively evaluated risk probabilities of sudden water pollution accidents under the influence of risk sources, thus providing an important guarantee for risk source identification during water diversion from the Hanjiang River to the Weihe River. [Methods] The research used Bayesian networks to represent the correlation between accidental risk sources. It also adopted the sequential Monte Carlo algorithm to combine water quality simulation with state simulation of risk sources, thereby determining standard-exceeding probabilities of sudden water pollution accidents. [Results] When the upstream inflow was 138.15 m3/s and the average accident duration was 48 h, the probabilities were 0.0416 and 0.0056 separately. When the upstream inflow was 55.29 m3/s and the average accident duration was 48 h, the probabilities were 0.0225 and 0.0028 separately. [Conclusions] The research conducted a risk assessment on sudden water pollution accidents, thereby providing an important guarantee for the smooth implementation, operation, and water quality of the Hanjiang-to-Weihe River Diversion Project.

  13. Characteristics of nocturnal coastal boundary layer in Ahtopol based on averaged SODAR profiles

    Science.gov (United States)

    Barantiev, Damyan; Batchvarova, Ekaterina; Novitzky, Mikhail

    2014-05-01

    The ground-based remote sensing instruments allow studying the wind regime and the turbulent characteristics of the atmosphere with height, achieving new knowledge and solving practical problems, such as air quality assessments, mesoscale models evaluation with high resolution data, characterization of the exchange processes between the surface and the atmosphere, the climate comfort conditions and the risk for extreme events, etc. Very important parameter in such studies is the height of the atmospheric boundary layer. Acoustic remote sensing data of the coastal atmospheric boundary layer were explored based on over 4-years continuous measurements at the meteorological observatory of Ahtopol (Bulgarian Southern Black Sea Coast) under Bulgarian - Russian scientific agreement. Profiles of 12 parameters from a mid-range acoustic sounding instrument type SCINTEC MFAS are derived and averaged up to about 600 m according filtering based on wind direction (land or sea type of night fowls). From the whole investigated period of 1454 days with 10-minute resolution SODAR data 2296 profiles represented night marine air masses and 1975 profiles represented the night flow from land during the months May to September. Graphics of averaged profiles of 12 SODAR output parameters with different availability of data in height are analyzed for both cases. A marine boundary-layer height of about 300 m is identified in the profiles of standard deviation of vertical wind speed (σw), Turbulent Kinetic Energy (TKE) and eddy dissipation rate (EDR). A nocturnal boundary-layer height of about 420 m was identified from the profiles of the same parameters under flows from land condition. In addition, the Buoyancy Production (BP= σw3/z) profiles were calculated from the standard deviation of the vertical wind speed and the height z above ground.

  14. Light-cone averaging in cosmology: formalism and applications

    International Nuclear Information System (INIS)

    Gasperini, M.; Marozzi, G.; Veneziano, G.; Nugier, F.

    2011-01-01

    We present a general gauge invariant formalism for defining cosmological averages that are relevant for observations based on light-like signals. Such averages involve either null hypersurfaces corresponding to a family of past light-cones or compact surfaces given by their intersection with timelike hypersurfaces. Generalized Buchert-Ehlers commutation rules for derivatives of these light-cone averages are given. After introducing some adapted ''geodesic light-cone'' coordinates, we give explicit expressions for averaging the redshift to luminosity-distance relation and the so-called ''redshift drift'' in a generic inhomogeneous Universe

  15. Time Overrun in Construction Project

    Science.gov (United States)

    Othman, I.; Shafiq, Nasir; Nuruddin, M. F.

    2017-12-01

    Timely completion is the key criteria to achieve success in any project despite the industry. Unfortunately construction industry in Malaysia has been labelled as industry facing poor performance leading to failure in achieving effective time management. As the consequence most of the project face huge amount of time overrun. This study assesses the causes of construction projects time overrun in Malaysia using structured questionnaire survey. Each respondent is asked to assign a one-to-five rating for each of the 18 time factors identified from literature review. Out of the 50 questionnaires sent out, 33 were received back representing 68% of the response rate. Data received from the questionnaires were analysed and processed using the descriptive statistics procedures. Findings from the study revealed that design and documentation issues, project management and contract administration, ineffective project planning and scheduling, contractor’s site management, financial resource management were the major factors that cause the time overrun. This study is hoped to help the practitioners to implement the mitigation measure at planning stage in order to achieve successful construction projects.

  16. Bounds on Average Time Complexity of Decision Trees

    KAUST Repository

    Chikalov, Igor

    2011-01-01

    In this chapter, bounds on the average depth and the average weighted depth of decision trees are considered. Similar problems are studied in search theory [1], coding theory [77], design and analysis of algorithms (e.g., sorting) [38]. For any diagnostic problem, the minimum average depth of decision tree is bounded from below by the entropy of probability distribution (with a multiplier 1/log2 k for a problem over a k-valued information system). Among diagnostic problems, the problems with a complete set of attributes have the lowest minimum average depth of decision trees (e.g, the problem of building optimal prefix code [1] and a blood test study in assumption that exactly one patient is ill [23]). For such problems, the minimum average depth of decision tree exceeds the lower bound by at most one. The minimum average depth reaches the maximum on the problems in which each attribute is "indispensable" [44] (e.g., a diagnostic problem with n attributes and kn pairwise different rows in the decision table and the problem of implementing the modulo 2 summation function). These problems have the minimum average depth of decision tree equal to the number of attributes in the problem description. © Springer-Verlag Berlin Heidelberg 2011.

  17. The Spectrum Prize: A simple algorithm to evaluate the relative sensitivity of γ-ray spectra, representative of detection systems

    Energy Technology Data Exchange (ETDEWEB)

    Spolaore, P.

    2016-03-11

    A simple analysis of gamma spectra selected to represent the performance of different detection systems, or, for one same system, different operation modes or states of progress of the system development, allows to compare the relative average-sensitivities of the represented systems themselves, as operated in the selected cases. The obtained SP figure-of-merit takes into account and correlates the main parameters commonly used to estimate the performance of a system. An example of application is given.

  18. 48 CFR 1852.227-72 - Designation of new technology representative and patent representative.

    Science.gov (United States)

    2010-10-01

    ... CONTRACT CLAUSES Texts of Provisions and Clauses 1852.227-72 Designation of new technology representative... of New Technology Representative and Patent Representative (JUL 1997) (a) For purposes of administration of the clause of this contract entitled “New Technology” or “Patent Rights—Retention by the...

  19. Determination of the diagnostic x-ray tube practical peak voltage (PPV) from average or average peak voltage measurements

    Energy Technology Data Exchange (ETDEWEB)

    Hourdakis, C J, E-mail: khour@gaec.gr [Ionizing Radiation Calibration Laboratory-Greek Atomic Energy Commission, PO Box 60092, 15310 Agia Paraskevi, Athens, Attiki (Greece)

    2011-04-07

    The practical peak voltage (PPV) has been adopted as the reference measuring quantity for the x-ray tube voltage. However, the majority of commercial kV-meter models measure the average peak, U-bar{sub P}, the average, U-bar, the effective, U{sub eff} or the maximum peak, U{sub P} tube voltage. This work proposed a method for determination of the PPV from measurements with a kV-meter that measures the average U-bar or the average peak, U-bar{sub p} voltage. The kV-meter reading can be converted to the PPV by applying appropriate calibration coefficients and conversion factors. The average peak k{sub PPV,kVp} and the average k{sub PPV,Uav} conversion factors were calculated from virtual voltage waveforms for conventional diagnostic radiology (50-150 kV) and mammography (22-35 kV) tube voltages and for voltage ripples from 0% to 100%. Regression equation and coefficients provide the appropriate conversion factors at any given tube voltage and ripple. The influence of voltage waveform irregularities, like 'spikes' and pulse amplitude variations, on the conversion factors was investigated and discussed. The proposed method and the conversion factors were tested using six commercial kV-meters at several x-ray units. The deviations between the reference and the calculated - according to the proposed method - PPV values were less than 2%. Practical aspects on the voltage ripple measurement were addressed and discussed. The proposed method provides a rigorous base to determine the PPV with kV-meters from U-bar{sub p} and U-bar measurement. Users can benefit, since all kV-meters, irrespective of their measuring quantity, can be used to determine the PPV, complying with the IEC standard requirements.

  20. Averaging operations on matrices

    Indian Academy of Sciences (India)

    2014-07-03

    Jul 3, 2014 ... Role of Positive Definite Matrices. • Diffusion Tensor Imaging: 3 × 3 pd matrices model water flow at each voxel of brain scan. • Elasticity: 6 × 6 pd matrices model stress tensors. • Machine Learning: n × n pd matrices occur as kernel matrices. Tanvi Jain. Averaging operations on matrices ...

  1. The Visual Identity Project

    Science.gov (United States)

    Tennant-Gadd, Laurie; Sansone, Kristina Lamour

    2008-01-01

    Identity is the focus of the middle-school visual arts program at Cambridge Friends School (CFS) in Cambridge, Massachusetts. Sixth graders enter the middle school and design a personal logo as their first major project in the art studio. The logo becomes a way for students to introduce themselves to their teachers and to represent who they are…

  2. Climatic Models Ensemble-based Mid-21st Century Runoff Projections: A Bayesian Framework

    Science.gov (United States)

    Achieng, K. O.; Zhu, J.

    2017-12-01

    There are a number of North American Regional Climate Change Assessment Program (NARCCAP) climatic models that have been used to project surface runoff in the mid-21st century. Statistical model selection techniques are often used to select the model that best fits data. However, model selection techniques often lead to different conclusions. In this study, ten models are averaged in Bayesian paradigm to project runoff. Bayesian Model Averaging (BMA) is used to project and identify effect of model uncertainty on future runoff projections. Baseflow separation - a two-digital filter which is also called Eckhardt filter - is used to separate USGS streamflow (total runoff) into two components: baseflow and surface runoff. We use this surface runoff as the a priori runoff when conducting BMA of runoff simulated from the ten RCM models. The primary objective of this study is to evaluate how well RCM multi-model ensembles simulate surface runoff, in a Bayesian framework. Specifically, we investigate and discuss the following questions: How well do ten RCM models ensemble jointly simulate surface runoff by averaging over all the models using BMA, given a priori surface runoff? What are the effects of model uncertainty on surface runoff simulation?

  3. A multiscale filter for noise reduction of low-dose cone beam projections.

    Science.gov (United States)

    Yao, Weiguang; Farr, Jonathan B

    2015-08-21

    The Poisson or compound Poisson process governs the randomness of photon fluence in cone beam computed tomography (CBCT) imaging systems. The probability density function depends on the mean (noiseless) of the fluence at a certain detector. This dependence indicates the natural requirement of multiscale filters to smooth noise while preserving structures of the imaged object on the low-dose cone beam projection. In this work, we used a Gaussian filter, exp(-x2/2σ(2)(f)) as the multiscale filter to de-noise the low-dose cone beam projections. We analytically obtained the expression of σ(f), which represents the scale of the filter, by minimizing local noise-to-signal ratio. We analytically derived the variance of residual noise from the Poisson or compound Poisson processes after Gaussian filtering. From the derived analytical form of the variance of residual noise, optimal σ(2)(f)) is proved to be proportional to the noiseless fluence and modulated by local structure strength expressed as the linear fitting error of the structure. A strategy was used to obtain the reliable linear fitting error: smoothing the projection along the longitudinal direction to calculate the linear fitting error along the lateral direction and vice versa. The performance of our multiscale filter was examined on low-dose cone beam projections of a Catphan phantom and a head-and-neck patient. After performing the filter on the Catphan phantom projections scanned with pulse time 4 ms, the number of visible line pairs was similar to that scanned with 16 ms, and the contrast-to-noise ratio of the inserts was higher than that scanned with 16 ms about 64% in average. For the simulated head-and-neck patient projections with pulse time 4 ms, the visibility of soft tissue structures in the patient was comparable to that scanned with 20 ms. The image processing took less than 0.5 s per projection with 1024   ×   768 pixels.

  4. Enhancing STEM coursework at MSIs through the AMS Climate Studies Diversity Project

    Science.gov (United States)

    Abshire, W. E.; Mills, E. W.; Slough, S. W.; Brey, J. A.; Geer, I. W.; Nugnes, K. A.

    2017-12-01

    The AMS Education Program celebrates a successful completion to its AMS Climate Studies Diversity Project. The project was funded for 6 years (2011-2017) through the National Science Foundation (NSF). It introduced and enhanced geoscience and/or sustainability-focused course components at minority-serving institutions (MSIs) across the U.S., many of which are signatories to the President's Climate Leadership Commitments, administered by Second Nature, and/or members of the Louis Stokes Alliances for Minority Participation. The Project introduced AMS Climate Studies curriculum to approximately 130 faculty representing 113 MSIs. Each year a cohort of, on average, 25 faculty attended a course implementation workshop where they were immersed in the course materials, received presentations from high-level speakers, and trained as change agents for their local institutions. This workshop was held in the Washington, DC area in collaboration with Second Nature, NOAA, NASA Goddard Space Flight Center, Howard University, and other local climate educational and research institutions. Following, faculty introduced and enhanced geoscience curricula on their local campuses with AMS Climate Studies course materials, thereby bringing change from within. Faculty were then invited to the following AMS Annual Meeting to report on their AMS Climate Studies course implementation progress, reconnect with their colleagues, and learn new science presented at the meeting. A longitudinal survey was administered to all Climate Diversity Project faculty participants who attended the course implementation workshops. The survey goals were to assess the effectiveness of the Project in helping faculty implement/enhance their institutional climate science offering, share best practices in offering AMS Climate Studies, and analyze the usefulness of course materials. Results will be presented during this presentation. The AMS Climate Studies Diversity Project builds on highly successful, NSF

  5. 20 CFR 404.221 - Computing your average monthly wage.

    Science.gov (United States)

    2010-04-01

    ... 20 Employees' Benefits 2 2010-04-01 2010-04-01 false Computing your average monthly wage. 404.221... DISABILITY INSURANCE (1950- ) Computing Primary Insurance Amounts Average-Monthly-Wage Method of Computing Primary Insurance Amounts § 404.221 Computing your average monthly wage. (a) General. Under the average...

  6. A whole stand basal area projection model for Appalachian hardwoods

    Science.gov (United States)

    John R. Brooks; Lichun Jiang; Matthew Perkowski; Benktesh Sharma

    2008-01-01

    Two whole-stand basal area projection models were developed for Appalachian hardwood stands. The proposed equations are an algebraic difference projection form based on existing basal area and the change in age, trees per acre, and/or dominant height. Average equation error was less than 10 square feet per acre and residuals exhibited no irregular trends.

  7. Exploiting scale dependence in cosmological averaging

    International Nuclear Information System (INIS)

    Mattsson, Teppo; Ronkainen, Maria

    2008-01-01

    We study the role of scale dependence in the Buchert averaging method, using the flat Lemaitre–Tolman–Bondi model as a testing ground. Within this model, a single averaging scale gives predictions that are too coarse, but by replacing it with the distance of the objects R(z) for each redshift z, we find an O(1%) precision at z<2 in the averaged luminosity and angular diameter distances compared to their exact expressions. At low redshifts, we show the improvement for generic inhomogeneity profiles, and our numerical computations further verify it up to redshifts z∼2. At higher redshifts, the method breaks down due to its inability to capture the time evolution of the inhomogeneities. We also demonstrate that the running smoothing scale R(z) can mimic acceleration, suggesting that it could be at least as important as the backreaction in explaining dark energy as an inhomogeneity induced illusion

  8. The new Wallula CO2 project may revive the old Columbia River Basalt (western USA) nuclear-waste repository project

    Science.gov (United States)

    Schwartz, Michael O.

    2018-02-01

    A novel CO2 sequestration project at Wallula, Washington, USA, makes ample use of the geoscientific data collection of the old nuclear waste repository project at the Hanford Site nearby. Both projects target the Columbia River Basalt (CRB). The new publicity for the old project comes at a time when the approach to high-level nuclear waste disposal has undergone fundamental changes. The emphasis now is on a technical barrier that is chemically compatible with the host rock. In the ideal case, the waste container is in thermodynamic equilibrium with the host-rock groundwater regime. The CRB groundwater has what it takes to represent the ideal case.

  9. Accuracy of past projections of US energy consumption

    International Nuclear Information System (INIS)

    O'Neill, B.C.; Desai, Mausami

    2005-01-01

    Energy forecasts play a key role in development of energy and environmental policy. Evaluations of the accuracy of past projections can provide insight into the uncertainty that may be associated with current forecasts. They can also be used to identify sources of inaccuracies, and potentially lead to improvements in projections over time. Here we assess the accuracy of projections of US energy consumption produced by the Energy Information Administration over the period 1982-2000. We find that energy consumption projections have tended to underestimate future consumption. Projections 10-13 years into the future have had an average error of about 4%, and about half that for shorter time horizons. These errors mask much larger, offsetting errors in the projection of GDP and energy intensity (EI). GDP projections have consistently been too high, and EI projection consistently too low, by more than 15% for projections of 10 years or more. Further work on the source of these sizable inaccuracies should be a high priority. Finally, we find no evidence of improvement in projections of consumption, GDP, or EI since 1982

  10. Averaging in SU(2) open quantum random walk

    International Nuclear Information System (INIS)

    Ampadu Clement

    2014-01-01

    We study the average position and the symmetry of the distribution in the SU(2) open quantum random walk (OQRW). We show that the average position in the central limit theorem (CLT) is non-uniform compared with the average position in the non-CLT. The symmetry of distribution is shown to be even in the CLT

  11. Averaging in SU(2) open quantum random walk

    Science.gov (United States)

    Clement, Ampadu

    2014-03-01

    We study the average position and the symmetry of the distribution in the SU(2) open quantum random walk (OQRW). We show that the average position in the central limit theorem (CLT) is non-uniform compared with the average position in the non-CLT. The symmetry of distribution is shown to be even in the CLT.

  12. European Wind Farm Project Costs History and Projections 2008 Study

    Energy Technology Data Exchange (ETDEWEB)

    2008-10-15

    At the request of Enova (the 'Client'), Garrad Hassan and Partners Limited ('GH') has provided technical advice on capital cost expectations for wind farm developments. In summary, the work provides a survey of the present and future 5-year prognosis for costs and conditions facing developers and suppliers in the European wind power market. The report will be used as a benchmark to support tendering for future Norwegian projects. As such, it will also provide discussion of how project characteristics can influence project cost. Data Used in the Analysis GH has obtained data on the investment costs for 35 projects developed or in development in Europe. The projects represent to the extent possible the characteristics representative of potential Norwegian projects. The data used in this analysis are from actual projects in: France, Germany, Hungary, Ireland, Italy, Northern Ireland, Portugal, Scotland, Spain and Wales. The turbine capacities represented in the data are typically 2 MW or above, except in one case where a mix of turbines sizes was used at the project. GH highlights that because of high demand for turbines, the main manufacturers have recently been offering to meet delivery schedules for new orders from late 2010. For new tenders it is likely that delivery time frames offered will now be for 2011 deliveries. As a result of the current 'Seller's Market', production capacity typically relates directly to the number of turbines sold in the year; therefore for 2007 the annual production capacity was approximately 22 GW. GH is aware that turbine suppliers across the market are working to increase their production capacity in order to ease the pressure on the market, however, there are bottlenecks through the supply chain at the sub component level. As a result, increases in production capacity will likely remain at a relatively steady state in the short term. Energy Assessment The energy assessment of a project is the area

  13. Determining average yarding distance.

    Science.gov (United States)

    Roger H. Twito; Charles N. Mann

    1979-01-01

    Emphasis on environmental and esthetic quality in timber harvesting has brought about increased use of complex boundaries of cutting units and a consequent need for a rapid and accurate method of determining the average yarding distance and area of these units. These values, needed for evaluation of road and landing locations in planning timber harvests, are easily and...

  14. Average Revisited in Context

    Science.gov (United States)

    Watson, Jane; Chick, Helen

    2012-01-01

    This paper analyses the responses of 247 middle school students to items requiring the concept of average in three different contexts: a city's weather reported in maximum daily temperature, the number of children in a family, and the price of houses. The mixed but overall disappointing performance on the six items in the three contexts indicates…

  15. When good = better than average

    Directory of Open Access Journals (Sweden)

    Don A. Moore

    2007-10-01

    Full Text Available People report themselves to be above average on simple tasks and below average on difficult tasks. This paper proposes an explanation for this effect that is simpler than prior explanations. The new explanation is that people conflate relative with absolute evaluation, especially on subjective measures. The paper then presents a series of four studies that test this conflation explanation. These tests distinguish conflation from other explanations, such as differential weighting and selecting the wrong referent. The results suggest that conflation occurs at the response stage during which people attempt to disambiguate subjective response scales in order to choose an answer. This is because conflation has little effect on objective measures, which would be equally affected if the conflation occurred at encoding.

  16. Accounting for downscaling and model uncertainty in fine-resolution seasonal climate projections over the Columbia River Basin

    Science.gov (United States)

    Ahmadalipour, Ali; Moradkhani, Hamid; Rana, Arun

    2018-01-01

    Climate change is expected to have severe impacts on natural systems as well as various socio-economic aspects of human life. This has urged scientific communities to improve the understanding of future climate and reduce the uncertainties associated with projections. In the present study, ten statistically downscaled CMIP5 GCMs at 1/16th deg. spatial resolution from two different downscaling procedures are utilized over the Columbia River Basin (CRB) to assess the changes in climate variables and characterize the associated uncertainties. Three climate variables, i.e. precipitation, maximum temperature, and minimum temperature, are studied for the historical period of 1970-2000 as well as future period of 2010-2099, simulated with representative concentration pathways of RCP4.5 and RCP8.5. Bayesian Model Averaging (BMA) is employed to reduce the model uncertainty and develop a probabilistic projection for each variable in each scenario. Historical comparison of long-term attributes of GCMs and observation suggests a more accurate representation for BMA than individual models. Furthermore, BMA projections are used to investigate future seasonal to annual changes of climate variables. Projections indicate significant increase in annual precipitation and temperature, with varied degree of change across different sub-basins of CRB. We then characterized uncertainty of future projections for each season over CRB. Results reveal that model uncertainty is the main source of uncertainty, among others. However, downscaling uncertainty considerably contributes to the total uncertainty of future projections, especially in summer. On the contrary, downscaling uncertainty appears to be higher than scenario uncertainty for precipitation.

  17. Cost diviation in road construction projects: The case of Palestine

    Directory of Open Access Journals (Sweden)

    Ibrahim Mahamid

    2012-02-01

    Full Text Available This paper investigates the statistical relationship between actual and estimated cost of road construction projects using data from road construction projects awarded in the West Bank in Palestine over the years 2004–2008. The study is based on a sample of 169 road construction projects. Based on this data, regression models are developed. The findings reveal that 100% of projects suffer from cost diverge, it is found that 76% of projects have cost under estimation while 24% have cost over estimation. The discrepancy between estimated and actual cost has an average of 14.6%, ranging from -39% to 98%. The relation between the project size (length and width and the cost diverge is discussed.

  18. The role of Project Finance in the viability of infrastructure projects: case of the petroleum and natural gas sector

    International Nuclear Information System (INIS)

    Faria, Viviana Cardoso de Sa e; Rodrigues, Adriano Pires

    2000-01-01

    Project finance represents neither recent news nor a panacea in the fields of long term financing. It is not able to solve the chronicle scarcity of resources applied in big projects financing in developing countries. In fact, underdeveloped markets as the Brazilian one offer, at the beginning, almost no chances to the project finance solution. In this case a path full of barriers overcome the solutions that project finance may offer. The process to adapt this instrument to the Brazilian reality presents the following hindrances: a different law framework, capital market underdevelopment; economical instability; political and regulating risks; incapability of the national insurance companies to insure big projects; cultural differences and lack of know-how in this area, and lastly, the out of date tax system. (author)

  19. ON THE AVERAGE DENSITY PROFILE OF DARK-MATTER HALOS IN THE INNER REGIONS OF MASSIVE EARLY-TYPE GALAXIES

    International Nuclear Information System (INIS)

    Grillo, C.

    2012-01-01

    We study a sample of 39 massive early-type lens galaxies at redshift z ∼< 0.3 to determine the slope of the average dark-matter density profile in the innermost regions. We keep the strong-lensing and stellar population synthesis modeling as simple as possible to measure the galaxy total and luminous masses. By rescaling the values of the Einstein radius and dark-matter projected mass with the values of the luminous effective radius and mass, we combine all the data of the galaxies in the sample. We find that between 0.3 and 0.9 times the value of the effective radius the average logarithmic slope of the dark-matter projected density profile is –1.0 ± 0.2 (i.e., approximately isothermal) or –0.7 ± 0.5 (i.e., shallower than isothermal), if, respectively, a constant Chabrier or heavier, Salpeter-like stellar initial mass function is adopted. These results provide positive evidence of the influence of the baryonic component on the contraction of the galaxy dark-matter halos, compared to the predictions of dark-matter-only cosmological simulations, and open a new way to test models of structure formation and evolution within the standard ΛCDM cosmological scenario.

  20. The PRINCE Project and its Applications (clone)

    OpenAIRE

    Bescos, Sylvie; Bisière, Christophe; Gailly, Pierre-Joseph; Krautter, Wolfgang

    1992-01-01

    The Esprit project Prince aims at development of an industrial Constraint Logic Programing environment based on the Prolog III language. In parallel, the current technology is being validated within the project on representative real-world examples. This paper describes the current status of the three application domains which were Selected to demonstrate the applicability and usefulness of CLP. These applications deal with industrial systems engineering, medium term banking planning and jobs...

  1. Plastic surgery: quo vadis? Current trends and future projections of aesthetic plastic surgical procedures in the United States.

    Science.gov (United States)

    Broer, P Niclas; Levine, Steven M; Juran, Sabrina

    2014-03-01

    The objective of this article was to evaluate past and current trends regarding aesthetic operations in the United States and to project future changes regarding such procedures. Cosmetic surgery statistics from the American Society for Aesthetic Plastic Surgery from 1997 to 2012 were analyzed by sex, age, and ethnic group. Then, using population projections from the U.S. Census Bureau based on the 2010 census, two projection scenarios of the expected number of aesthetic plastic surgery procedures were generated. The scenarios included the presumed occurrence and nonoccurrence of a recession like that which occurred in 2007. Aesthetic procedures are expected to grow from 1,688,694 in 2012 to 3,847,929 by 2030, representing an average annual growth rate of 7.1 percent. Should another recession of similar degree to the one in 2007 occur, procedures would increase to only 2,086,994, displaying an average annual growth percentage rate of 1.3 percent. Because the age distribution of the patient population will change, preferences for specific procedures according to age influence-and thus are reflected in-future demand for those procedures. Furthermore, the ethnic profile of patients will change significantly, with 32 percent of all procedures being performed on patients other than Caucasians by 2030. Demand for aesthetic plastic surgical procedures is expected to continue to grow, while depending on the economic performance at the macro level and changing demographic dynamics of the U.S. population. Considering all investigated factors and trends among all patients, the most commonly requested procedures by 2030 are likely to be (1) breast augmentations, (2) lipoplasties, and (3) blepharoplasties.

  2. Low-level radioactive waste in the northeast: disposal volume projections

    International Nuclear Information System (INIS)

    1982-10-01

    The northeastern states, with support of the Coalition of Northeastern Governors (CONEG), are developing compact(s) for the disposal and management of low-level radioactive waste (LLRW) generated in the eleven northeastern states (Connecticut, Delaware, Maine, Maryland, Massachusetts, New Hampshire, New Jersey, New York, Pennsylvania, Rhode Island, and Vermont). The Technical Subcommittee has made a projection of future low-level radioactive waste to the year 2000 based on existing waste volume data and anticipated growth in the Northeast states. Aware of the difficulties involved with any long range projection - unforeseen events can drastically change projections based on current assumptions - the Technical Subcommittee believes that waste volume projections should be reviewed annually as updated information becomes available. The Technical Subcommittee made the following findings based upon a conservative projection methodology: volumes of low-level waste produced annually in the eleven states individually and collectively are expected to grow continually through the year 2000 with the rate of increase varying by state; by the year 2000, the Northeast is projected to generate 58,000 m 3 of low-level waste annually, about 1.9 times the current average; and based on current estimates, 47% of the total projected waste volume in the year 2000 will be produced by nuclear power plants, compared to the current average of 54%. Non-reactor wastes will equal 53% of the total in the year 2000 compared to the current 46%

  3. Estimation of catchment averaged sensible heat fluxes using a large aperture scintillometer

    Directory of Open Access Journals (Sweden)

    Samain Bruno

    2012-05-01

    Full Text Available Evapotranspiration rates at the catchment scale are very difficult to quantify. One possible manner to continuously observe this variable could be the estimation of sensible heat fluxes (H across large distances (in the order of kilometers using a large aperture scintillometer (LAS, and inverting these observations into evapotranspiration rates, under the assumption that the LAS observations are representative for the entire catchment. The objective of this paper is to assess whether measured sensible heat fluxes from a LAS over a long distance (9.5 km can be assumed to be valid for a 102.3 km2 heterogeneous catchment. Therefore, a fully process-based water and energy balance model with a spatial resolution of 50 m has been thoroughly calibrated and validated for the Bellebeek catchmentin Belgium. A footprint analysis has been performed. In general, the sensible heat fluxes from the LAS compared well with the modeled sensible heat fluxes within the footprint. Moreover, as the modeled Hwithin the footprint has been found to be almost equal to the modeled catchment averaged H, it can be concluded that the scintillometer measurements over a distance of 9.5 km and an effective heightof 68 m are representative for the entire catchment.

  4. TEP Power Partners Project [Tucson Electric Power

    Energy Technology Data Exchange (ETDEWEB)

    None, None

    2014-02-06

    The Arizona Governor’s Office of Energy Policy, in partnership with Tucson Electric Power (TEP), Tendril, and Next Phase Energy (NPE), formed the TEP Power Partners pilot project to demonstrate how residential customers could access their energy usage data and third party applications using data obtained from an Automatic Meter Reading (AMR) network. The project applied for and was awarded a Smart Grid Data Access grant through the U.S. Department of Energy. The project participants’ goal for Phase I is to actively engage 1,700 residential customers to demonstrate sustained participation, reduction in energy usage (kWh) and cost ($), and measure related aspects of customer satisfaction. This Demonstration report presents a summary of the findings, effectiveness, and customer satisfaction with the 15-month TEP Power Partners pilot project. The objective of the program is to provide residential customers with energy consumption data from AMR metering and empower these participants to better manage their electricity use. The pilot recruitment goals included migrating 700 existing customers from the completed Power Partners Demand Response Load Control Project (DRLC), and enrolling 1,000 new participants. Upon conclusion of the project on November 19, 2013; 1,390 Home Area Networks (HANs) were registered; 797 new participants installed a HAN; Survey respondents’ are satisfied with the program and found value with a variety of specific program components; Survey respondents report feeling greater control over their energy usage and report taking energy savings actions in their homes after participating in the program; On average, 43 % of the participants returned to the web portal monthly and 15% returned weekly; and An impact evaluation was completed by Opinion Dynamics and found average participant savings for the treatment period1 to be 2.3% of their household use during this period.2 In total, the program saved 163 MWh in the treatment period of 2013.

  5. DISPLAY OF PIXEL LOSS AND REPLICATION IN REPROJECTING RASTER DATA FROM THE SINUSOIDAL PROJECTION

    Science.gov (United States)

    Recent studies show the sinusoidal projection to be a superior planar projection for representing global raster datasets. This study uses the sinusoidal projection as a basis for evaluating pixel loss and replication in eight other planar map projections. The percent of pixels ...

  6. Effect of money supply on the Dow Jones Industrial Average stock index

    Directory of Open Access Journals (Sweden)

    Martin Širůček

    2012-01-01

    Full Text Available This focus of this paper are the effects and implications of a change in the money supply for share price indices in the USA during 1959–2011. The money supply will be measured by the M2 and MZM aggregates (money with zero maturity. The US stock market is represented by the Dow Jones Industrial Average index. The objective of this paper is to find, describe and evaluate the effects of changes to the money supply (M2 and MZM on the US stock market. A partial objective of this paper is to determine whether a change in the monetary aggregate shows in the stock index immediately or with a delay of several weeks. Another aim is to determine whether asset prices influence the money supply.

  7. 47 CFR 80.759 - Average terrain elevation.

    Science.gov (United States)

    2010-10-01

    ... 47 Telecommunication 5 2010-10-01 2010-10-01 false Average terrain elevation. 80.759 Section 80.759 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) SAFETY AND SPECIAL RADIO SERVICES STATIONS IN THE MARITIME SERVICES Standards for Computing Public Coast Station VHF Coverage § 80.759 Average terrain elevation. (a)(1) Draw radials...

  8. Exploitation of cloud computing in management of construction projects in Slovakia

    Directory of Open Access Journals (Sweden)

    Mandičák Tomáš

    2016-12-01

    Full Text Available The issue of cloud computing is a highly topical issue. Cloud computing represents a new model for information technology (IT services based on the exploitation of Web (it represents a cloud and other application platforms, as well as software as a service. In general, the exploitation of cloud computing in construction project management has several advantages, as demonstrated by several research reports. Currently, research quantifying the exploitation of cloud computing in the Slovak construction industry has not yet been carried out. The article discusses the issue of exploitation of cloud computing in construction project management in Slovakia. The main objective of the research is to confirm whether factors such as size of construction enterprise, owner of construction enterprise and participant of construction project have any impact on the exploitation level of cloud computing in construction project management. It includes confirmation of differences in use between different participants of the construction project or between construction enterprises broken down by size and shareholders.

  9. Project schedule and cost estimate report

    International Nuclear Information System (INIS)

    1988-03-01

    All cost tables represent obligation dollars, at both a constant FY 1987 level and an estimated escalation level, and are based on the FY 1989 DOE Congressional Budget submittal of December 1987. The cost tables display the total UMTRA Project estimated costs, which include both Federal and state funding. The Total Estimated Cost (TEC) for the UMTRA Project is approximately $992.5 million (in 1987 escalated dollars). Project schedules have been developed that provide for Project completion by September 1994, subject to Congressional approval extending DOE's authorization under Public Law 95-604. The report contains site-specific demographic data, conceptual design assumptions, preliminary cost estimates, and site schedules. A general project overview is also presented, which includes a discussion of the basis for the schedule and cost estimates, contingency assumptions, work breakdown structure, and potential project risks. The schedules and cost estimates will be revised as necessary to reflect appropriate decisions relating to relocation of certain tailings piles, or other special design considerations or circumstances (such as revised EPA groundwater standards), and changes in the Project mission. 27 figs', 97 tabs

  10. A novel Generalized State-Space Averaging (GSSA) model for advanced aircraft electric power systems

    International Nuclear Information System (INIS)

    Ebrahimi, Hadi; El-Kishky, Hassan

    2015-01-01

    Highlights: • A study model is developed for aircraft electric power systems. • A novel GSSA model is developed for the interconnected power grid. • The system’s dynamics are characterized under various conditions. • The averaged results are compared and verified with the actual model. • The obtained measured values are validated with available aircraft standards. - Abstract: The growing complexity of Advanced Aircraft Electric Power Systems (AAEPS) has made conventional state-space averaging models inadequate for systems analysis and characterization. This paper presents a novel Generalized State-Space Averaging (GSSA) model for the system analysis, control and characterization of AAEPS. The primary objective of this paper is to introduce a mathematically elegant and computationally simple model to copy the AAEPS behavior at the critical nodes of the electric grid. Also, to reduce some or all of the drawbacks (complexity, cost, simulation time…, etc) associated with sensor-based monitoring and computer aided design software simulations popularly used for AAEPS characterization. It is shown in this paper that the GSSA approach overcomes the limitations of the conventional state-space averaging method, which fails to predict the behavior of AC signals in a circuit analysis. Unlike conventional averaging method, the GSSA model presented in this paper includes both DC and AC components. This would capture the key dynamic and steady-state characteristics of the aircraft electric systems. The developed model is then examined for the aircraft system’s visualization and accuracy of computation under different loading scenarios. Through several case studies, the applicability and effectiveness of the GSSA method is verified by comparing to the actual real-time simulation model obtained from Powersim 9 (PSIM9) software environment. The simulations results represent voltage, current and load power at the major nodes of the AAEPS. It has been demonstrated that

  11. The Millennium Villages Project: a retrospective, observational, endline evaluation

    Directory of Open Access Journals (Sweden)

    Shira Mitchell, PhD

    2018-05-01

    Full Text Available Summary: Background: The Millennium Villages Project (MVP was a 10 year, multisector, rural development project, initiated in 2005, operating across ten sites in ten sub-Saharan African countries to achieve the Millennium Development Goals (MDGs. In this study, we aimed to estimate the project's impact, target attainment, and on-site spending. Methods: In this endline evaluation of the MVP, we retrospectively selected comparison villages that best matched the project villages on possible confounding variables. Cross-sectional survey data on 40 outcomes of interest were collected from both the project and the comparison villages in 2015. Using these data, as well as on-site spending data collected during the project, we estimated project impacts as differences in outcomes between the project and comparison villages; target attainment as differences between project outcomes and prespecified targets; and on-site spending as expenditures reported by communities, donors, governments, and the project. Spending data were not collected in the comparison villages. Findings: Averaged across the ten project sites, we found that impact estimates for 30 of 40 outcomes were significant (95% uncertainty intervals [UIs] for these outcomes excluded zero and favoured the project villages. In particular, substantial effects were seen in agriculture and health, in which some outcomes were roughly one SD better in the project villages than in the comparison villages. The project was estimated to have no significant impact on the consumption-based measures of poverty, but a significant favourable impact on an index of asset ownership. Impacts on nutrition and education outcomes were often inconclusive (95% UIs included zero. Averaging across outcomes within categories, the project had significant favourable impacts on agriculture, nutrition, education, child health, maternal health, HIV and malaria, and water and sanitation. A third of the targets were met in the

  12. Massachusetts Crystalline Repository Project. Progress report, December 31, 1985

    International Nuclear Information System (INIS)

    1985-01-01

    Project activities which have been undertaken include the following: review and comment on OCRD projects reports; review of pertinent DOE, NRC, EPA, and DOT quidelines and regulations; review of reports and maps released by the federal project group and contractors; attendance at DOE workshops and conferences; implementation of state-specific research activities; interaction with representatives of federal agencies and other participating states; and interface with media, state officials and legislators, and interested citizens

  13. HIGH AVERAGE POWER OPTICAL FEL AMPLIFIERS

    International Nuclear Information System (INIS)

    2005-01-01

    Historically, the first demonstration of the optical FEL was in an amplifier configuration at Stanford University [l]. There were other notable instances of amplifying a seed laser, such as the LLNL PALADIN amplifier [2] and the BNL ATF High-Gain Harmonic Generation FEL [3]. However, for the most part FELs are operated as oscillators or self amplified spontaneous emission devices. Yet, in wavelength regimes where a conventional laser seed can be used, the FEL can be used as an amplifier. One promising application is for very high average power generation, for instance FEL's with average power of 100 kW or more. The high electron beam power, high brightness and high efficiency that can be achieved with photoinjectors and superconducting Energy Recovery Linacs (ERL) combine well with the high-gain FEL amplifier to produce unprecedented average power FELs. This combination has a number of advantages. In particular, we show that for a given FEL power, an FEL amplifier can introduce lower energy spread in the beam as compared to a traditional oscillator. This properly gives the ERL based FEL amplifier a great wall-plug to optical power efficiency advantage. The optics for an amplifier is simple and compact. In addition to the general features of the high average power FEL amplifier, we will look at a 100 kW class FEL amplifier is being designed to operate on the 0.5 ampere Energy Recovery Linac which is under construction at Brookhaven National Laboratory's Collider-Accelerator Department

  14. Reynolds averaged turbulence modelling using deep neural networks with embedded invariance

    International Nuclear Information System (INIS)

    Ling, Julia; Kurzawski, Andrew; Templeton, Jeremy

    2016-01-01

    There exists significant demand for improved Reynolds-averaged Navier–Stokes (RANS) turbulence models that are informed by and can represent a richer set of turbulence physics. This paper presents a method of using deep neural networks to learn a model for the Reynolds stress anisotropy tensor from high-fidelity simulation data. A novel neural network architecture is proposed which uses a multiplicative layer with an invariant tensor basis to embed Galilean invariance into the predicted anisotropy tensor. It is demonstrated that this neural network architecture provides improved prediction accuracy compared with a generic neural network architecture that does not embed this invariance property. Furthermore, the Reynolds stress anisotropy predictions of this invariant neural network are propagated through to the velocity field for two test cases. For both test cases, significant improvement versus baseline RANS linear eddy viscosity and nonlinear eddy viscosity models is demonstrated.

  15. Grassmann Averages for Scalable Robust PCA

    DEFF Research Database (Denmark)

    Hauberg, Søren; Feragen, Aasa; Black, Michael J.

    2014-01-01

    As the collection of large datasets becomes increasingly automated, the occurrence of outliers will increase—“big data” implies “big outliers”. While principal component analysis (PCA) is often used to reduce the size of data, and scalable solutions exist, it is well-known that outliers can...... to vectors (subspaces) or elements of vectors; we focus on the latter and use a trimmed average. The resulting Trimmed Grassmann Average (TGA) is particularly appropriate for computer vision because it is robust to pixel outliers. The algorithm has low computational complexity and minimal memory requirements...

  16. Project Valuation when There are Two Cashflow Streams

    Energy Technology Data Exchange (ETDEWEB)

    Emhjellen, Magne; Alaouze, Chris M

    2002-07-01

    Some authors (Lewellen, 1977, Shall, 1972, Butters et al., 1987, Laughton and Jacoby, 1993, Jacoby and Laughton, 1992, Salahor, 1998) advocate the separate discounting of different cash flows when calculating net present value (NPV). However, some textbooks (Brealy and Myers, 1991, Copeland and Weston, 1992) focus on calculating NPV by discounting the expected net after tax cash flow using the weighted average cost of capital (WACC) as the discount rate. We show that discounting the expected net after tax cash flow of a project using the WACC yields an incorrect project NPV. A new method for calculating project NPV's using a separate cash flow discounting method is proposed and applied to calculating the NPV's of some North Sea oil projects. (author)

  17. Project Valuation when There are Two Cashflow Streams

    Energy Technology Data Exchange (ETDEWEB)

    Emhjellen, Magne; Alaouze, Chris M.

    2002-07-01

    Some authors (Lewellen, 1977, Shall, 1972, Butters et al., 1987, Laughton and Jacoby, 1993, Jacoby and Laughton, 1992, Salahor, 1998) advocate the separate discounting of different cash flows when calculating net present value (NPV). However, some textbooks (Brealy and Myers, 1991, Copeland and Weston, 1992) focus on calculating NPV by discounting the expected net after tax cash flow using the weighted average cost of capital (WACC) as the discount rate. We show that discounting the expected net after tax cash flow of a project using the WACC yields an incorrect project NPV. A new method for calculating project NPV's using a separate cash flow discounting method is proposed and applied to calculating the NPV's of some North Sea oil projects. (author)

  18. Scale-invariant Green-Kubo relation for time-averaged diffusivity

    Science.gov (United States)

    Meyer, Philipp; Barkai, Eli; Kantz, Holger

    2017-12-01

    In recent years it was shown both theoretically and experimentally that in certain systems exhibiting anomalous diffusion the time- and ensemble-averaged mean-squared displacement are remarkably different. The ensemble-averaged diffusivity is obtained from a scaling Green-Kubo relation, which connects the scale-invariant nonstationary velocity correlation function with the transport coefficient. Here we obtain the relation between time-averaged diffusivity, usually recorded in single-particle tracking experiments, and the underlying scale-invariant velocity correlation function. The time-averaged mean-squared displacement is given by 〈δ2¯〉 ˜2 DνtβΔν -β , where t is the total measurement time and Δ is the lag time. Here ν is the anomalous diffusion exponent obtained from ensemble-averaged measurements 〈x2〉 ˜tν , while β ≥-1 marks the growth or decline of the kinetic energy 〈v2〉 ˜tβ . Thus, we establish a connection between exponents that can be read off the asymptotic properties of the velocity correlation function and similarly for the transport constant Dν. We demonstrate our results with nonstationary scale-invariant stochastic and deterministic models, thereby highlighting that systems with equivalent behavior in the ensemble average can differ strongly in their time average. If the averaged kinetic energy is finite, β =0 , the time scaling of 〈δ2¯〉 and 〈x2〉 are identical; however, the time-averaged transport coefficient Dν is not identical to the corresponding ensemble-averaged diffusion constant.

  19. Climate change impact on streamflow in large-scale river basins: projections and their uncertainties sourced from GCMs and RCP scenarios

    Science.gov (United States)

    Nasonova, Olga N.; Gusev, Yeugeniy M.; Kovalev, Evgeny E.; Ayzel, Georgy V.

    2018-06-01

    Climate change impact on river runoff was investigated within the framework of the second phase of the Inter-Sectoral Impact Model Intercomparison Project (ISI-MIP2) using a physically-based land surface model Soil Water - Atmosphere - Plants (SWAP) (developed in the Institute of Water Problems of the Russian Academy of Sciences) and meteorological projections (for 2006-2099) simulated by five General Circulation Models (GCMs) (including GFDL-ESM2M, HadGEM2-ES, IPSL-CM5A-LR, MIROC-ESM-CHEM, and NorESM1-M) for each of four Representative Concentration Pathway (RCP) scenarios (RCP2.6, RCP4.5, RCP6.0, and RCP8.5). Eleven large-scale river basins were used in this study. First of all, SWAP was calibrated and validated against monthly values of measured river runoff with making use of forcing data from the WATCH data set and all GCMs' projections were bias-corrected to the WATCH. Then, for each basin, 20 projections of possible changes in river runoff during the 21st century were simulated by SWAP. Analysis of the obtained hydrological projections allowed us to estimate their uncertainties resulted from application of different GCMs and RCP scenarios. On the average, the contribution of different GCMs to the uncertainty of the projected river runoff is nearly twice larger than the contribution of RCP scenarios. At the same time the contribution of GCMs slightly decreases with time.

  20. Development of a major hydroelectric project in the Pacific Rim

    International Nuclear Information System (INIS)

    Afshar, Y.; Kwiatkowski, R.W.

    1990-01-01

    Developing a major international energy project requires the assembly and close coordination of various factors. Some of the specific factors include identification of the client, specific project identification, establishing the need for the project, identifying the pros and cons of the project, technical and financial feasibility of the project, identifying sources of financing, identifying the competitions, studying possible teaming arrangements, involvement of local firms and representatives, and developing a strategy that will lead to the successful negotiation of the contract. This paper briefly addresses each of the factors

  1. OECD Halden Reactor Project

    International Nuclear Information System (INIS)

    1988-01-01

    The OECD Halden Reactor project is an agreement between OECD member countries. It was first signed in 1958 and since then regularly renewed every third year. The activities at the Project is centred around the Halden heavy water rector, the HBWR. The reseach programme comprizes studies of fuel performance under various operating conditions, and the application of computers for process control. The HBWR is equipped for exposing fuel rods to temperatures and pressures, and at heat ratings met in modern BWR's and PWR's. A range of in-core instruments are available, permitting detailed measurements of the reactions of the fuel, including mechanical deformations, thermal behaviour, fission gas release, and corrosion. In the area of computer application, the studies of the communication between operator and process, and the surveillance and control of the reactor core, are of particular interst for reactor operation. 1988 represents the 30th year since the Project was started, and this publication is produced to mark this event. It gives and account of the activities and achievements of the Project through the years 1958-1988

  2. Combining 3d Volume and Mesh Models for Representing Complicated Heritage Buildings

    Science.gov (United States)

    Tsai, F.; Chang, H.; Lin, Y.-W.

    2017-08-01

    This study developed a simple but effective strategy to combine 3D volume and mesh models for representing complicated heritage buildings and structures. The idea is to seamlessly integrate 3D parametric or polyhedral models and mesh-based digital surfaces to generate a hybrid 3D model that can take advantages of both modeling methods. The proposed hybrid model generation framework is separated into three phases. Firstly, after acquiring or generating 3D point clouds of the target, these 3D points are partitioned into different groups. Secondly, a parametric or polyhedral model of each group is generated based on plane and surface fitting algorithms to represent the basic structure of that region. A "bare-bones" model of the target can subsequently be constructed by connecting all 3D volume element models. In the third phase, the constructed bare-bones model is used as a mask to remove points enclosed by the bare-bones model from the original point clouds. The remaining points are then connected to form 3D surface mesh patches. The boundary points of each surface patch are identified and these boundary points are projected onto the surfaces of the bare-bones model. Finally, new meshes are created to connect the projected points and original mesh boundaries to integrate the mesh surfaces with the 3D volume model. The proposed method was applied to an open-source point cloud data set and point clouds of a local historical structure. Preliminary results indicated that the reconstructed hybrid models using the proposed method can retain both fundamental 3D volume characteristics and accurate geometric appearance with fine details. The reconstructed hybrid models can also be used to represent targets in different levels of detail according to user and system requirements in different applications.

  3. 40 CFR 80.205 - How is the annual refinery or importer average and corporate pool average sulfur level determined?

    Science.gov (United States)

    2010-07-01

    ... volume of gasoline produced or imported in batch i. Si=The sulfur content of batch i determined under § 80.330. n=The number of batches of gasoline produced or imported during the averaging period. i=Individual batch of gasoline produced or imported during the averaging period. (b) All annual refinery or...

  4. Changes in precipitation extremes projected by a 20-km mesh global atmospheric model

    Directory of Open Access Journals (Sweden)

    Akio Kitoh

    2016-03-01

    Full Text Available High-resolution modeling is necessary to project weather and climate extremes and their future changes under global warming. A global high-resolution atmospheric general circulation model with grid size about 20 km is able to reproduce climate fields as well as regional-scale phenomena such as monsoonal rainfall, tropical and extratropical cyclones, and heavy precipitation. This 20-km mesh model is applied to project future changes in weather and climate extremes at the end of the 21st century with four different spatial patterns in sea surface temperature (SST changes: one with the mean SST changes by the 28 models of the Coupled Model Intercomparison Project Phase 5 (CMIP5 under the Representative Concentration Pathways (RCP-8.5 scenario, and the other three obtained from a cluster analysis, in which tropical SST anomalies derived from the 28 CMIP5 models were grouped. Here we focus on future changes in regional precipitation and its extremes. Various precipitation indices averaged over the Twenty-two regional land domains are calculated. Heavy precipitation indices (maximum 5-day precipitation total and maximum 1-day precipitation total increase in all regional domains, even where mean precipitation decrease (Southern Africa, South Europe/Mediterranean, Central America. South Asia is the domain of the largest extreme precipitation increase. In some domains, different SST patterns result in large precipitation changes, possibly related to changes in large-scale circulations in the tropical Pacific.

  5. Graph averaging as a means to compare multichannel EEG coherence networks and its application to the study of mental fatigue and neurodegenerative disease

    NARCIS (Netherlands)

    Crippa, Alessandro; Maurits, Natasha M.; Lorist, Monicque M.; Roerdink, Jos B.T.M.

    A method is proposed for quantifying differences between multichannel EEG coherence networks represented by functional unit (FU) maps. The approach is based on inexact graph matching for attributed relational graphs and graph averaging, adapted to FU-maps. The mean of a set of input FU-maps is

  6. Tools for Closure Project and Contract Management: Development of the Rocky Flats Integrated Closure Project Baseline

    International Nuclear Information System (INIS)

    Gelles, C. M.; Sheppard, F. R.

    2002-01-01

    This paper details the development of the Rocky Flats Integrated Closure Project Baseline - an innovative project management effort undertaken to ensure proactive management of the Rocky Flats Closure Contract in support of the Department's goal for achieving the safe closure of the Rocky Flats Environmental Technology Site (RFETS) in December 2006. The accelerated closure of RFETS is one of the most prominent projects within the Department of Energy (DOE) Environmental Management program. As the first major former weapons plant to be remediated and closed, it is a first-of-kind effort requiring the resolution of multiple complex technical and institutional challenges. Most significantly, the closure of RFETS is dependent upon the shipment of all special nuclear material and wastes to other DOE sites. The Department is actively working to strengthen project management across programs, and there is increasing external interest in this progress. The development of the Rocky Flats Integrated Closure Project Baseline represents a groundbreaking and cooperative effort to formalize the management of such a complex project across multiple sites and organizations. It is original in both scope and process, however it provides a useful precedent for the other ongoing project management efforts within the Environmental Management program

  7. Waste Management Project Contingency Analysis

    International Nuclear Information System (INIS)

    Edward L. Parsons, Jr.

    1999-01-01

    The purpose of this report is to provide the office of Waste Management (WM) with recommended contingency calculation procedures for typical WM projects. Typical projects were defined as conventional construction-type activities that use innovative elements when necessary to meet the project objectives. Projects involve treatment, storage, and disposal of low level, mixed low level, hazardous, transuranic, and high level waste. Cost contingencies are an essential part of Total Cost Management. A contingency is an amount added to a cost estimate to compensate for unexpected expenses resulting from incomplete design, unforeseen and unpredictable conditions, or uncertainties in the project scope (DOE 1994, AACE 1998). Contingency allowances are expressed as percentages of estimated cost and improve cost estimates by accounting for uncertainties. The contingency allowance is large at the beginning of a project because there are more uncertainties, but as a project develops, the allowance shrinks to adjust for costs already incurred. Ideally, the total estimated cost remains the same throughout a project. Project contingency reflects the degree of uncertainty caused by lack of project definition, and process contingency reflects the degree of uncertainty caused by use of new technology. Different cost estimation methods were reviewed and compared with respect to terminology, accuracy, and Cost Guide standards. The Association for the Advancement of Cost Engineering (AACE) methods for cost estimation were selected to represent best industry practice. AACE methodology for contingency analysis can be readily applied to WM Projects, accounts for uncertainties associated with different stages of a project, and considers both project and process contingencies and the stage of technical readiness. As recommended, AACE contingency allowances taper off linearly as a project nears completion

  8. Average sedimentary rock rare Earth element patterns and crustal evolution: Some observations and implications from the 3800 Ma ISUA supracrustal belt, West Greenland

    Science.gov (United States)

    Dymek, R. F.; Boak, J. L.; Gromet, L. P.

    1983-01-01

    Rare earth element (REE) data is given on a set of clastic metasediments from the 3800 Ma Isua Supracrustal belt, West Greenland. Each of two units from the same sedimentary sequence has a distinctive REE pattern, but the average of these rocks bears a very strong resemblance to the REE pattern for the North American Shale Composite (NASC), and departs considerably from previous estimates of REE patterns in Archaean sediments. The possibility that the source area for the Isua sediments resembled that of the NASC is regarded as highly unlikely. However, REE patterns like that in the NASC may be produced by sedimentary recycling of material yielding patterns such as are found at Isua. The results lead to the following tentative conclusions: (1) The REE patterns for Isua Seq. B MBG indicate the existence of crustal materials with fractionated REE and negative Eu anomalies at 3800 Ma, (2) The average Seq. B REE pattern resembles that of the North American Shale Composite (NASC), (3) If the Seq. B average is truly representative of its crustal sources, then this early crust could have been extensively differentiated. In this regard, a proper understanding of the NASC pattern, and its relationship to post-Archaean crustal REE reservoirs, is essential, (4) The Isua results may represent a local effect.

  9. Representative of local committee

    International Nuclear Information System (INIS)

    Barcelo, C.

    2007-01-01

    Carlos Barcelo, spoke about the involvement of local companies in the dismantling project. Local companies realised that they could not undertake the whole project, and informed ENRESA of the tasks in which they would be able to co-operate. On the whole, during the course of the 5-year project the local and regional firms were given more work than they had originally expected. Mr. Barcelo praised the good partnership that evolved between the local firms and the implementer. (author)

  10. The ethanol heavy-duty truck fleet demonstration project

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    1997-06-01

    This project was designed to test and demonstrate the use of a high- percentage ethanol-blended fuel in a fleet of heavy-duty, over-the- road trucks, paying particular attention to emissions, performance, and repair and maintenance costs. This project also represents the first public demonstration of the use of ethanol fuels as a viable alternative to conventional diesel fuel in heavy-duty engines.

  11. Engineering for new-built nuclear power plant projects

    International Nuclear Information System (INIS)

    Gonzalez Lopez, A.

    2012-01-01

    This article reviews the opportunities existing in the market (electrical utilities and reactor vendors) for an engineering company with the profile of Empresarios Agrupados (EA) in new-built nuclear power plant projects. To do this, reference is made to some representative examples of projects in which EA has been participating recently. the article concludes sharing with the reader some lessons learned from this participation. (Author)

  12. Averaging hydraulic head, pressure head, and gravitational head in subsurface hydrology, and implications for averaged fluxes, and hydraulic conductivity

    Directory of Open Access Journals (Sweden)

    G. H. de Rooij

    2009-07-01

    Full Text Available Current theories for water flow in porous media are valid for scales much smaller than those at which problem of public interest manifest themselves. This provides a drive for upscaled flow equations with their associated upscaled parameters. Upscaling is often achieved through volume averaging, but the solution to the resulting closure problem imposes severe restrictions to the flow conditions that limit the practical applicability. Here, the derivation of a closed expression of the effective hydraulic conductivity is forfeited to circumvent the closure problem. Thus, more limited but practical results can be derived. At the Representative Elementary Volume scale and larger scales, the gravitational potential and fluid pressure are treated as additive potentials. The necessary requirement that the superposition be maintained across scales is combined with conservation of energy during volume integration to establish consistent upscaling equations for the various heads. The power of these upscaling equations is demonstrated by the derivation of upscaled water content-matric head relationships and the resolution of an apparent paradox reported in the literature that is shown to have arisen from a violation of the superposition principle. Applying the upscaling procedure to Darcy's Law leads to the general definition of an upscaled hydraulic conductivity. By examining this definition in detail for porous media with different degrees of heterogeneity, a series of criteria is derived that must be satisfied for Darcy's Law to remain valid at a larger scale.

  13. Overview of the Hanford Environmental Dose Reconstruction Project

    International Nuclear Information System (INIS)

    Shipler, D.B.; Napier, B.A.; Ikenberry, T.A.

    1992-04-01

    The objective of the Hanford Environmental Dose Reconstruction (HEDR) Project is to estimate the radiation doses that specific and representative individuals and populations may have received as a result of releases of radioactive materials from historical operations at the Hanford Site. These dose estimates would account for the uncertainties of information regarding facilities operations, environmental monitoring, demography, food consumption and lifestyles, and the variability of natural phenomena. Other objectives of the HEDR Project include: supporting the Hanford Thyroid Disease Study (HTDS), declassifying Hanford-generated information and making it available to the public, performing high-quality, credible science, and conducting the project in an open, public forum. The project is briefly described

  14. GOATS Image Projection Component

    Science.gov (United States)

    Haber, Benjamin M.; Green, Joseph J.

    2011-01-01

    When doing mission analysis and design of an imaging system in orbit around the Earth, answering the fundamental question of imaging performance requires an understanding of the image products that will be produced by the imaging system. GOATS software represents a series of MATLAB functions to provide for geometric image projections. Unique features of the software include function modularity, a standard MATLAB interface, easy-to-understand first-principles-based analysis, and the ability to perform geometric image projections of framing type imaging systems. The software modules are created for maximum analysis utility, and can all be used independently for many varied analysis tasks, or used in conjunction with other orbit analysis tools.

  15. Averaging processes in granular flows driven by gravity

    Science.gov (United States)

    Rossi, Giulia; Armanini, Aronne

    2016-04-01

    One of the more promising theoretical frames to analyse the two-phase granular flows is offered by the similarity of their rheology with the kinetic theory of gases [1]. Granular flows can be considered a macroscopic equivalent of the molecular case: the collisions among molecules are compared to the collisions among grains at a macroscopic scale [2,3]. However there are important statistical differences in dealing with the two applications. In the two-phase fluid mechanics, there are two main types of average: the phasic average and the mass weighed average [4]. The kinetic theories assume that the size of atoms is so small, that the number of molecules in a control volume is infinite. With this assumption, the concentration (number of particles n) doesn't change during the averaging process and the two definitions of average coincide. This hypothesis is no more true in granular flows: contrary to gases, the dimension of a single particle becomes comparable to that of the control volume. For this reason, in a single realization the number of grain is constant and the two averages coincide; on the contrary, for more than one realization, n is no more constant and the two types of average lead to different results. Therefore, the ensamble average used in the standard kinetic theory (which usually is the phasic average) is suitable for the single realization, but not for several realization, as already pointed out in [5,6]. In the literature, three main length scales have been identified [7]: the smallest is the particles size, the intermediate consists in the local averaging (in order to describe some instability phenomena or secondary circulation) and the largest arises from phenomena such as large eddies in turbulence. Our aim is to solve the intermediate scale, by applying the mass weighted average, when dealing with more than one realizations. This statistical approach leads to additional diffusive terms in the continuity equation: starting from experimental

  16. Average-case analysis of numerical problems

    CERN Document Server

    2000-01-01

    The average-case analysis of numerical problems is the counterpart of the more traditional worst-case approach. The analysis of average error and cost leads to new insight on numerical problems as well as to new algorithms. The book provides a survey of results that were mainly obtained during the last 10 years and also contains new results. The problems under consideration include approximation/optimal recovery and numerical integration of univariate and multivariate functions as well as zero-finding and global optimization. Background material, e.g. on reproducing kernel Hilbert spaces and random fields, is provided.

  17. Project as a System and its Management

    Directory of Open Access Journals (Sweden)

    Jiří Skalický

    2017-06-01

    Full Text Available The contribution aims to describe project as a system, to define project control goal and strategy, control variables and their relationships. Three common control variables represented by the project triangle, are extended by two other important variables – project risk and quality. The control system consists of two components: social one – project manager and project team – and technical one – project dynamic simulation model as a decision making support of project manager in project milestones. In the project planning phase, the project baseline with planned controlled variables is created. In milestones after project launch, the actual values of these variables are measured. If the actual values deviate from planned ones, corrective actions are proposed and new baseline for the following control interval is created. Project plan takes into account the actual project progress and optimum corrective actions are determined by simulation, respecting control strategy and availability of resources. The contribution presents list of references to articles dealing with project as a system and its simulation. In most cases, they refer to the project control using the Earned Value Management method and its derivatives. Using of the dynamic simulation model for project monitoring and control, suggested in this contribution, presents a novel approach. The proposed model can serve as departure point to future research of authors and for development of appropriate and applicable tool.

  18. Numerical Tokamak Project code comparison

    International Nuclear Information System (INIS)

    Waltz, R.E.; Cohen, B.I.; Beer, M.A.

    1994-01-01

    The Numerical Tokamak Project undertook a code comparison using a set of TFTR tokamak parameters. Local radial annulus codes of both gyrokinetic and gyrofluid types were compared for both slab and toroidal case limits assuming ion temperature gradient mode turbulence in a pure plasma with adiabatic electrons. The heat diffusivities were found to be in good internal agreement within ± 50% of the group average over five codes

  19. The Difficulty in Following Project Schedule as a Key Project Management Challenge: Family Firm Perspective

    Directory of Open Access Journals (Sweden)

    Joanna SADKOWSKA

    2016-12-01

    Full Text Available The problem of how to manage projects successfully has been gaining growing interest for the last decades. The aforementioned is mainly caused by the fact that project management offers a wide range of methods and tools which, when properly used, can stimulate long-term growth of businesses. As a consequence it offers a particular development opportunity for family enterprises which, due to their specificity, have to overcome many difficulties. The primary objective of this paper is to examine whether family enterprises perceive, and to what extent, the factor of following project schedule as a difficulty in the area of project management. 154 Polish family firms representing different sectors were surveyed. The results of regression analysis show that family firms with global range of business activities, on the contrary to those with local or regional ones, do not find the factor of following project schedule as a difficulty. This results manly from the tools employed and the maturity level they have reached in project management. The findings support the current discussion on the specificity and uniqueness of family businesses in relationship to the knowledge area of project management. It also contributes to filling the gap on understanding the functioning of family firms in the emerging economies of Eastern Europe.

  20. A Comparative Study of the Persistence and Academic Achievement of "Project 60"& Regularly Enrolled Students at Middlesex Community College.

    Science.gov (United States)

    Lee, Glenda E.

    The Project 60 group was an experimental group of 53 disadvantaged students who would not have been admitted to Middlesex Community College through the regular admissions procedures. They had a high school average of 1.7 and were largely below average in reading, math, and composition, both in high school grades and individual testing. Project 60…

  1. Post-model selection inference and model averaging

    Directory of Open Access Journals (Sweden)

    Georges Nguefack-Tsague

    2011-07-01

    Full Text Available Although model selection is routinely used in practice nowadays, little is known about its precise effects on any subsequent inference that is carried out. The same goes for the effects induced by the closely related technique of model averaging. This paper is concerned with the use of the same data first to select a model and then to carry out inference, in particular point estimation and point prediction. The properties of the resulting estimator, called a post-model-selection estimator (PMSE, are hard to derive. Using selection criteria such as hypothesis testing, AIC, BIC, HQ and Cp, we illustrate that, in terms of risk function, no single PMSE dominates the others. The same conclusion holds more generally for any penalised likelihood information criterion. We also compare various model averaging schemes and show that no single one dominates the others in terms of risk function. Since PMSEs can be regarded as a special case of model averaging, with 0-1 random-weights, we propose a connection between the two theories, in the frequentist approach, by taking account of the selection procedure when performing model averaging. We illustrate the point by simulating a simple linear regression model.

  2. Theoretical Study of Vibrationally Averaged Dipole Moments for the Ground and Excited C=O Stretching States of trans-Formic Acid

    Czech Academy of Sciences Publication Activity Database

    Paulson, L. O.; Kaminský, Jakub; Anderson, D. T.; Bouř, Petr; Kubelka, J.

    2010-01-01

    Roč. 6, č. 3 (2010), s. 817-827 ISSN 1549-9618 R&D Projects: GA ČR GA202/07/0732; GA AV ČR IAA400550702 Grant - others:CAREER(US) 0846140; AV ČR(CZ) M200550902 Institutional research plan: CEZ:AV0Z40550506 Keywords : dipole moments * theoretical modelling * vibrational averaging Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 5.138, year: 2010

  3. Average L-shell fluorescence, Auger, and electron yields

    International Nuclear Information System (INIS)

    Krause, M.O.

    1980-01-01

    The dependence of the average L-shell fluorescence and Auger yields on the initial vacancy distribution is shown to be small. By contrast, the average electron yield pertaining to both Auger and Coster-Kronig transitions is shown to display a strong dependence. Numerical examples are given on the basis of Krause's evaluation of subshell radiative and radiationless yields. Average yields are calculated for widely differing vacancy distributions and are intercompared graphically for 40 3 subshell yields in most cases of inner-shell ionization

  4. Projections onto the Pareto surface in multicriteria radiation therapy optimization

    International Nuclear Information System (INIS)

    Bokrantz, Rasmus; Miettinen, Kaisa

    2015-01-01

    Purpose: To eliminate or reduce the error to Pareto optimality that arises in Pareto surface navigation when the Pareto surface is approximated by a small number of plans. Methods: The authors propose to project the navigated plan onto the Pareto surface as a postprocessing step to the navigation. The projection attempts to find a Pareto optimal plan that is at least as good as or better than the initial navigated plan with respect to all objective functions. An augmented form of projection is also suggested where dose–volume histogram constraints are used to prevent that the projection causes a violation of some clinical goal. The projections were evaluated with respect to planning for intensity modulated radiation therapy delivered by step-and-shoot and sliding window and spot-scanned intensity modulated proton therapy. Retrospective plans were generated for a prostate and a head and neck case. Results: The projections led to improved dose conformity and better sparing of organs at risk (OARs) for all three delivery techniques and both patient cases. The mean dose to OARs decreased by 3.1 Gy on average for the unconstrained form of the projection and by 2.0 Gy on average when dose–volume histogram constraints were used. No consistent improvements in target homogeneity were observed. Conclusions: There are situations when Pareto navigation leaves room for improvement in OAR sparing and dose conformity, for example, if the approximation of the Pareto surface is coarse or the problem formulation has too permissive constraints. A projection onto the Pareto surface can identify an inaccurate Pareto surface representation and, if necessary, improve the quality of the navigated plan

  5. Projections onto the Pareto surface in multicriteria radiation therapy optimization.

    Science.gov (United States)

    Bokrantz, Rasmus; Miettinen, Kaisa

    2015-10-01

    To eliminate or reduce the error to Pareto optimality that arises in Pareto surface navigation when the Pareto surface is approximated by a small number of plans. The authors propose to project the navigated plan onto the Pareto surface as a postprocessing step to the navigation. The projection attempts to find a Pareto optimal plan that is at least as good as or better than the initial navigated plan with respect to all objective functions. An augmented form of projection is also suggested where dose-volume histogram constraints are used to prevent that the projection causes a violation of some clinical goal. The projections were evaluated with respect to planning for intensity modulated radiation therapy delivered by step-and-shoot and sliding window and spot-scanned intensity modulated proton therapy. Retrospective plans were generated for a prostate and a head and neck case. The projections led to improved dose conformity and better sparing of organs at risk (OARs) for all three delivery techniques and both patient cases. The mean dose to OARs decreased by 3.1 Gy on average for the unconstrained form of the projection and by 2.0 Gy on average when dose-volume histogram constraints were used. No consistent improvements in target homogeneity were observed. There are situations when Pareto navigation leaves room for improvement in OAR sparing and dose conformity, for example, if the approximation of the Pareto surface is coarse or the problem formulation has too permissive constraints. A projection onto the Pareto surface can identify an inaccurate Pareto surface representation and, if necessary, improve the quality of the navigated plan.

  6. Tomographic reconstruction of the time-averaged density distribution in two-phase flow

    International Nuclear Information System (INIS)

    Fincke, J.R.

    1982-01-01

    The technique of reconstructive tomography has been applied to the measurement of time-average density and density distribution in a two-phase flow field. The technique of reconstructive tomography provides a model-independent method of obtaining flow-field density information. A tomographic densitometer system for the measurement of two-phase flow has two unique problems: a limited number of data values and a correspondingly coarse reconstruction grid. These problems were studied both experimentally through the use of prototype hardware on a 3-in. pipe, and analytically through computer generation of simulated data. The prototype data were taken on phantoms constructed of all Plexiglas and Plexiglas laminated with wood and polyurethane foam. Reconstructions obtained from prototype data are compared with reconstructions from the simulated data. Also presented are some representative results in a horizontal air/water flow

  7. Bayesian Averaging is Well-Temperated

    DEFF Research Database (Denmark)

    Hansen, Lars Kai

    2000-01-01

    Bayesian predictions are stochastic just like predictions of any other inference scheme that generalize from a finite sample. While a simple variational argument shows that Bayes averaging is generalization optimal given that the prior matches the teacher parameter distribution the situation is l...

  8. Planning "and" Sprinting: Use of a Hybrid Project Management Methodology within a CIS Capstone Course

    Science.gov (United States)

    Baird, Aaron; Riggins, Frederick J.

    2012-01-01

    An increasing number of information systems projects in industry are managed using hybrid project management methodologies, but this shift in project management methods is not fully represented in our CIS curriculums. CIS capstone courses often include an applied project that is managed with traditional project management methods (plan first,…

  9. Energy-averaged neutron cross sections of fast-reactor structural materials

    International Nuclear Information System (INIS)

    Smith, A.; McKnight, R.; Smith, D.

    1978-02-01

    The status of energy-averaged cross sections of fast-reactor structural materials is outlined with emphasis on U.S. data programs in the neutron-energy range 1-10 MeV. Areas of outstanding accomplishment and significant uncertainty are noted with recommendations for future efforts. Attention is primarily given to the main constituents of stainless steel (e.g., Fe, Ni, and Cr) and, secondarily, to alternate structural materials (e.g., V, Ti, Nb, Mo, Zr). Generally, the mass regions of interest are A approximately 50 to 60 and A approximately 90 to 100. Neutron total and elastic-scattering cross sections are discussed with the implication on the non-elastic-cross sections. Cross sections governing discrete-inelastic-neutron-energy transfers are examined in detail. Cross sections for the reactions (n;p), (n;n',p), (n;α), (n;n',α) and (n;2n') are reviewed in the context of fast-reactor performance and/or diagnostics. The primary orientation of the discussion is experimental with some additional attention to the applications of theory, the problems of evaluation and the data sensitivity of representative fast-reactor systems

  10. The Ulysses Project. Integrating the Curriculum.

    Science.gov (United States)

    Smith, Susan Swenton

    2001-01-01

    Offers a project developed as an outgrowth of sixth-grade students' study of ancient Greece in history, English, drama, and art classes. Explains that the students created sculptures inspired by Greek sculpture that represented student perceptions of the activities and emotions found in the Ulysses myth. (CMK)

  11. BWR radiation exposure--experience and projection

    International Nuclear Information System (INIS)

    Falk, C.F.; Wilkinson, C.D.; Hollander, W.R.

    1979-01-01

    The BWR/6 Mark III radiation exposures are projected to be about half of those of current average operating experience of 725 man-rem. These projections are said to be realistic and based on current achievements and not on promises of future development. The several BWRs operating with low primary system radiation levels are positive evidence that radiation sources can be reduced. Improvements have been made in reducing the maintenance times for the BWR/6, and further improvements can be made by further attention to cost-effective plant arrangement and layout during detail design to improve accessibility and maintainability of each system and component

  12. Selecting users for participation in IT projects

    DEFF Research Database (Denmark)

    Rasmussen, Rasmus; Christensen, Anders S.; Fjeldsted, Tobias

    2011-01-01

    The selection of users for participation in IT projects involves trade-offs between multiple criteria, one of which is selecting a representative cross-section of users. This criterion is basic because trading it for other criteria means basing designs on information biased toward some user groups...... at the expense of others. Based on interviews in development and customer organizations we find that their criteria for user selection favour persons who can contribute to the progress of the IT project over persons who are representative of the full range of users. A highly valued contribution from...... participating users is the ability to advocate a vision for the system and champion its organizational implementation. A survey in one customer organization shows that respondents’ personal traits explain up to 31% of the variation in their experience of aspects of the usability of a recently introduced system...

  13. The Bitterroot Ecosystem Management Research Project: What we have learned

    Science.gov (United States)

    Helen Y. Smith

    2000-01-01

    The varied topics presented in these symposium proceedings represent the diverse nature of the Bitterroot Ecosystem Management Research Project (BEMRP). Separated into six sections, the papers cover the different themes researched by BEMRP collaborators as well as brief overviews of five other ecosystem management projects. The sections are: Understanding the Ecosystem...

  14. A study on the representative person of critical group for radiological assessment in Korea

    International Nuclear Information System (INIS)

    Lee, Gab-Bock; Lee, Doo-Ho; Yang, Yang-Hee; Jang, Young-Sik; Moon, Jong-Yi

    2011-01-01

    The 'maximum individual' concept based on US NRC Regulatory Guide 1.109 has been used for the public dose calculation in Korea. However, the 'maximum individual' results in largely overestimating the public dose. In Korea, the adoption of the 'representative person' is under consideration. The 'representative person' was recommended by ICRP Publication No. 101. In this study, individual habits (consumption of foodstuffs, use of local resources) of representative person have been analyzed or investigated. The age-specific food consumption rates of the representative person were derived from the Korean national dietary survey by government in 2005. The foodstuffs are classified into 12 classes - grain, fruits, kimchi, vegetables, milk, beef, pork, poultry, fishes, mollusca, crustacea, and seaweeds. The use of local resources such as the self-sufficiency of foodstuffs was investigated by using the questionnaire survey to the public members within 2 km from four nuclear power plant sites in Korea. The 95th percentile values are used for the top-two pathways such as grain and kimchi, but the average values are used for other foods. The results show that the effective dose from new habit data is about 30% lower than that from the existing data. (author)

  15. Reynolds Averaged Navier-Stokes (RANS) equation solutions of wind turbine wakes

    Energy Technology Data Exchange (ETDEWEB)

    Ludwig, Daniel Evandro; Horn, Diego Anderson; Petry, Adriane Prisco [Thermal and Energy Study Group, Mechanical Engeneering Department, Federal University of Rio Grande do Sul, Porto Alegre (Brazil)], E-mail: adrianep@mecanica.ufrgs.br

    2010-07-01

    This paper aims to evaluate the influence of three different turbulence models in the study of a wind turbine wake. Numerical Simulation is used as working tool to characterize the flow through the wind turbines, it is used the numeric simulation. The numerical analysis is based on the finite volume method and the Reynolds Averaged Navier-Stokes (RANS) equations. Three turbulence models are used to represent the total effects of turbulence in the flow: the two equations k-classical and the RNG k- models, based on the turbulent viscosity; and the Shear Stress Transport (SST) model, based on the transport of the Reynolds tensor. The results of the 'u' velocity profiles are compared to experimental data from Vermeer (2003) at distances equivalent to 2, 4, 6, 8, 10 and 16 diameters downstream from the turbine. Results shows that the SST model gives better results until 6 diameters, beyond this distance there is no significant differences between the compared models. (author)

  16. Uranium project. Geochemistry prospection

    International Nuclear Information System (INIS)

    Lambert, J.

    1983-01-01

    Geochemistry studies the distribution of the chemicals elements in the terrestrial crust and its ways to migrate. The terminology used in this report is the following one: 1) Principles of the prospection geochemistry 2) Stages of the prospection geochemistry 3)utility of the prospection geochemistry 4) geochemistry of uranium 5) procedures used within the framework of uranium project 6) Average available 7) Selection of the zones of prospection geochemistry 8) Stages of the prospection, Sample preparation and analisis 9) Presentation of the results

  17. Simultaneous inference for model averaging of derived parameters

    DEFF Research Database (Denmark)

    Jensen, Signe Marie; Ritz, Christian

    2015-01-01

    Model averaging is a useful approach for capturing uncertainty due to model selection. Currently, this uncertainty is often quantified by means of approximations that do not easily extend to simultaneous inference. Moreover, in practice there is a need for both model averaging and simultaneous...... inference for derived parameters calculated in an after-fitting step. We propose a method for obtaining asymptotically correct standard errors for one or several model-averaged estimates of derived parameters and for obtaining simultaneous confidence intervals that asymptotically control the family...

  18. National health expenditure projections, 2013-23: faster growth expected with expanded coverage and improving economy.

    Science.gov (United States)

    Sisko, Andrea M; Keehan, Sean P; Cuckler, Gigi A; Madison, Andrew J; Smith, Sheila D; Wolfe, Christian J; Stone, Devin A; Lizonitz, Joseph M; Poisal, John A

    2014-10-01

    In 2013 health spending growth is expected to have remained slow, at 3.6 percent, as a result of the sluggish economic recovery, the effects of sequestration, and continued increases in private health insurance cost-sharing requirements. The combined effects of the Affordable Care Act's coverage expansions, faster economic growth, and population aging are expected to fuel health spending growth this year and thereafter (5.6 percent in 2014 and 6.0 percent per year for 2015-23). However, the average rate of increase through 2023 is projected to be slower than the 7.2 percent average growth experienced during 1990-2008. Because health spending is projected to grow 1.1 percentage points faster than the average economic growth during 2013-23, the health share of the gross domestic product is expected to rise from 17.2 percent in 2012 to 19.3 percent in 2023. Project HOPE—The People-to-People Health Foundation, Inc.

  19. Key aspects to perform a project on energy management

    International Nuclear Information System (INIS)

    Bachini, R.

    1993-01-01

    A general overview on elements and organisms playing a key role to launch a new industrial project is given, taking as base case an energy management project. Likewise the problematic of training personnel involved in the project is analyzed. Energy management becomes crucial in industries where energy costs represent a big portion of the whole production cost. Main aspects to be analyzed are: - Adequate production procedures to be competitive - Environment protection regarding waste management - Maximization of safety at production installations. (Author)

  20. Analytical expressions for conditional averages: A numerical test

    DEFF Research Database (Denmark)

    Pécseli, H.L.; Trulsen, J.

    1991-01-01

    Conditionally averaged random potential fluctuations are an important quantity for analyzing turbulent electrostatic plasma fluctuations. Experimentally, this averaging can be readily performed by sampling the fluctuations only when a certain condition is fulfilled at a reference position...

  1. Relative importance of first and second derivatives of nuclear magnetic resonance chemical shifts and spin-spin coupling constants for vibrational averaging

    Czech Academy of Sciences Publication Activity Database

    Dračínský, Martin; Kaminský, Jakub; Bouř, Petr

    2009-01-01

    Roč. 130, č. 9 (2009), 094106/1-094106/13 ISSN 0021-9606 R&D Projects: GA ČR GA203/06/0420; GA ČR GA202/07/0732; GA AV ČR IAA400550702 Institutional research plan: CEZ:AV0Z40550506 Keywords : NMR * anharmonic forces * vibrational averaging Subject RIV: CF - Physical ; Theoretical Chemistry Impact factor: 3.093, year: 2009

  2. “Bridge of Generations” - Project of OJSC Atomenergomash

    International Nuclear Information System (INIS)

    2014-01-01

    Aims and Goals of the project: The main aim - create conditions for development and high competitiveness of the AEM Group enterprises. The expected results of the project are: • saving of key knowledge; •increase of efficiency of HR-policy; • creation of organizational conditions for the retention and transfer of knowledge in enterprises of OJSC «Atomenergomash»; • retention of critical knowledge and technologies; • reducing the average age of the personnel; •development of young specialists and raising the level of staff involvement

  3. The PRINCE Project and its Applications- VERSION2

    OpenAIRE

    Bescos, Sylvie; Bisière, Christophe; Gailly, Pierre-Joseph; Krautter, Wolfgang

    1992-01-01

    The Esprit project Prince aims at development of an industrial Constraint Logic Programing environment based on the Prolog III language. In parallel, the current technology is being validated within the project on representative real-world examples. This paper describes the current status of the three application domains which were Selected to demonstrate the applicability and usefulness of CLP. These applications deal with industrial systems engineering, medium term banking planning and jobs...

  4. Procedure for the characterization of radon potential in existing dwellings and to assess the annual average indoor radon concentration

    International Nuclear Information System (INIS)

    Collignan, Bernard; Powaga, Emilie

    2014-01-01

    Risk assessment due to radon exposure indoors is based on annual average indoor radon activity concentration. To assess the radon exposure in a building, measurement is generally performed during at least two months during heating period in order to be representative of the annual average value. This is because radon presence indoors could be very variable during time. This measurement protocol is fairly reliable but may be a limiting in the radon risk management, particularly during a real estate transaction due to the duration of the measurement and the limitation of the measurement period. A previous field study defined a rapid methodology to characterize radon entry in dwellings. The objective of this study was at first, to test this methodology in various dwellings to assess its relevance with a daily test. At second, a ventilation model was used to assess numerically the air renewal of a building, the indoor air quality all along the year and the annual average indoor radon activity concentration, based on local meteorological conditions, some building characteristics and in-situ characterization of indoor pollutant emission laws. Experimental results obtained on thirteen individual dwellings showed that it is generally possible to obtain a representative characterization of radon entry into homes. It was also possible to refine the methodology defined in the previous study. In addition, numerical assessments of annual average indoor radon activity concentration showed generally a good agreement with measured values. These results are encouraging to allow a procedure with a short measurement time to be used to characterize long-term radon potential in dwellings. - Highlights: • Test of a daily procedure to characterize radon potential in dwellings. • Numerical assessment of the annual radon concentration. • Procedure applied on thirteen dwellings, characterization generally satisfactory. • Procedure useful to manage radon risk in dwellings, for real

  5. The Millennium Villages Project: a retrospective, observational, endline evaluation.

    Science.gov (United States)

    Mitchell, Shira; Gelman, Andrew; Ross, Rebecca; Chen, Joyce; Bari, Sehrish; Huynh, Uyen Kim; Harris, Matthew W; Sachs, Sonia Ehrlich; Stuart, Elizabeth A; Feller, Avi; Makela, Susanna; Zaslavsky, Alan M; McClellan, Lucy; Ohemeng-Dapaah, Seth; Namakula, Patricia; Palm, Cheryl A; Sachs, Jeffrey D

    2018-05-01

    The Millennium Villages Project (MVP) was a 10 year, multisector, rural development project, initiated in 2005, operating across ten sites in ten sub-Saharan African countries to achieve the Millennium Development Goals (MDGs). In this study, we aimed to estimate the project's impact, target attainment, and on-site spending. In this endline evaluation of the MVP, we retrospectively selected comparison villages that best matched the project villages on possible confounding variables. Cross-sectional survey data on 40 outcomes of interest were collected from both the project and the comparison villages in 2015. Using these data, as well as on-site spending data collected during the project, we estimated project impacts as differences in outcomes between the project and comparison villages; target attainment as differences between project outcomes and prespecified targets; and on-site spending as expenditures reported by communities, donors, governments, and the project. Spending data were not collected in the comparison villages. Averaged across the ten project sites, we found that impact estimates for 30 of 40 outcomes were significant (95% uncertainty intervals [UIs] for these outcomes excluded zero) and favoured the project villages. In particular, substantial effects were seen in agriculture and health, in which some outcomes were roughly one SD better in the project villages than in the comparison villages. The project was estimated to have no significant impact on the consumption-based measures of poverty, but a significant favourable impact on an index of asset ownership. Impacts on nutrition and education outcomes were often inconclusive (95% UIs included zero). Averaging across outcomes within categories, the project had significant favourable impacts on agriculture, nutrition, education, child health, maternal health, HIV and malaria, and water and sanitation. A third of the targets were met in the project sites. Total on-site spending decreased from US$132

  6. Nonequilibrium statistical averages and thermo field dynamics

    International Nuclear Information System (INIS)

    Marinaro, A.; Scarpetta, Q.

    1984-01-01

    An extension of thermo field dynamics is proposed, which permits the computation of nonequilibrium statistical averages. The Brownian motion of a quantum oscillator is treated as an example. In conclusion it is pointed out that the procedure proposed to computation of time-dependent statistical average gives the correct two-point Green function for the damped oscillator. A simple extension can be used to compute two-point Green functions of free particles

  7. Fiscal planning of private electricity production projects

    International Nuclear Information System (INIS)

    Gauthier, R.

    2002-01-01

    Various fiscal considerations frequently encountered in the context of the planning of private electricity production projects were described. Two major themes were discussed: 1) the different jurisdictional vehicles that can be used during the planning of private electricity production projects and the associated fiscal considerations, and 2) the two main fiscal incentives of the Income Tax Act (Canada) which could impact on the financing and operation costs of such a project, namely the accelerated amortization and the possibility of deducting the costs associated to renewable energies and energy savings in Canada. This was a general presentation that did not go into specific details and did not represent a legal opinion. refs

  8. Time average vibration fringe analysis using Hilbert transformation

    International Nuclear Information System (INIS)

    Kumar, Upputuri Paul; Mohan, Nandigana Krishna; Kothiyal, Mahendra Prasad

    2010-01-01

    Quantitative phase information from a single interferogram can be obtained using the Hilbert transform (HT). We have applied the HT method for quantitative evaluation of Bessel fringes obtained in time average TV holography. The method requires only one fringe pattern for the extraction of vibration amplitude and reduces the complexity in quantifying the data experienced in the time average reference bias modulation method, which uses multiple fringe frames. The technique is demonstrated for the measurement of out-of-plane vibration amplitude on a small scale specimen using a time average microscopic TV holography system.

  9. Heat-related mortality projections for cardiovascular and respiratory disease under the changing climate in Beijing, China

    Science.gov (United States)

    Li, Tiantian; Ban, Jie; Horton, Radley M.; Bader, Daniel A.; Huang, Ganlin; Sun, Qinghua; Kinney, Patrick L.

    2015-08-01

    Because heat-related health effects tend to become more serious at higher temperatures, there is an urgent need to determine the mortality projection of specific heat-sensitive diseases to provide more detailed information regarding the variation of the sensitivity of such diseases. In this study, the specific mortality of cardiovascular and respiratory disease in Beijing was initially projected under five different global-scale General Circulation Models (GCMs) and two Representative Concentration Pathways scenarios (RCPs) in the 2020s, 2050s, and 2080s compared to the 1980s. Multi-model ensembles indicated cardiovascular mortality could increase by an average percentage of 18.4%, 47.8%, and 69.0% in the 2020s, 2050s, and 2080s under RCP 4.5, respectively, and by 16.6%,73.8% and 134% in different decades respectively, under RCP 8.5 compared to the baseline range. The same increasing pattern was also observed in respiratory mortality. The heat-related deaths under the RCP8.5 scenario were found to reach a higher number and to increase more rapidly during the 21st century compared to the RCP4.5 scenario, especially in the 2050s and the 2080s. The projection results show potential trends in cause-specific mortality in the context of climate change, and provide support for public health interventions tailored to specific climate-related future health risks.

  10. Safety Impact of Average Speed Control in the UK

    DEFF Research Database (Denmark)

    Lahrmann, Harry Spaabæk; Brassøe, Bo; Johansen, Jonas Wibert

    2016-01-01

    of automatic speed control was point-based, but in recent years a potentially more effective alternative automatic speed control method has been introduced. This method is based upon records of drivers’ average travel speed over selected sections of the road and is normally called average speed control...... in the UK. The study demonstrates that the introduction of average speed control results in statistically significant and substantial reductions both in speed and in number of accidents. The evaluation indicates that average speed control has a higher safety effect than point-based automatic speed control....

  11. In Second Year Of Marketplaces, New Entrants, ACA 'Co-Ops,' And Medicaid Plans Restrain Average Premium Growth Rates.

    Science.gov (United States)

    Gabel, Jon R; Whitmore, Heidi; Green, Matthew; Stromberg, Sam T; Weinstein, Daniel S; Oran, Rebecca

    2015-12-01

    Premiums for health insurance plans offered through the federally facilitated and state-based Marketplaces remained steady or increased only modestly from 2014 to 2015. We used data from the Marketplaces, state insurance departments, and insurer websites to examine patterns of premium pricing and the factors behind these patterns. Our data came from 2,964 unique plans offered in 2014 and 4,153 unique plans offered in 2015 in forty-nine states and the District of Columbia. Using descriptive and multivariate analysis, we found that the addition of a carrier in a rating area lowered average premiums for the two lowest-cost silver plans and the lowest-cost bronze plan by 2.2 percent. When all plans in a rating area were included, an additional carrier was associated with an average decline in premiums of 1.4 percent. Plans in the Consumer Operated and Oriented Plan Program and Medicaid managed care plans had lower premiums and average premium increases than national commercial and Blue Cross and Blue Shield plans. On average, premiums fell by an appreciably larger amount for catastrophic and bronze plans than for gold plans, and premiums for platinum plans increased. This trend of low premium increases overall is unlikely to continue, however, as insurers are faced with mounting medical claims. Project HOPE—The People-to-People Health Foundation, Inc.

  12. Bounds on Average Time Complexity of Decision Trees

    KAUST Repository

    Chikalov, Igor

    2011-01-01

    In this chapter, bounds on the average depth and the average weighted depth of decision trees are considered. Similar problems are studied in search theory [1], coding theory [77], design and analysis of algorithms (e.g., sorting) [38]. For any

  13. In core fuel management optimization by varying the equilibrium cycle average flux shape for batch refuelled reactors

    International Nuclear Information System (INIS)

    Jong, A.J. de.

    1992-12-01

    We suggest a method to overcome this problem of optimization by varying reloading patterns by characterizing each particular reloading pattern by a set of intermediate parameters that are numbers. Plots of the objective function versus the intermediate parameters can be made. When the intermediate parameters represent the reloading patterns in a unique way, the optimum of the objective function can be found by interpolation within such plots and we can find the optimal reloading pattern in terms of intermediate parameters. These have to be transformed backwards to find an optimal reloading pattern. The intermediate parameters are closely related to the time averaged neutron flux shape in the core during an equilibrium cycle. This flux shape is characterized by a set of ratios of the space averaged fluxes in the fuel zones and the space averaged flux in the zone with the fresh fuel elements. An advantage of this choice of intermediate parameters is that it permits analytical calculation of equilibrium cycle fuel densities in the fuel zones for any applied reloading patten characterized by a set of equilibrium cycle average flux ratios and thus, provides analytical calculations of fuel management objective functions. The method is checked for the burnup of one fissile nuclide in a reactor core with the geometry of the PWR at Borssele. For simplicity, neither the conversion of fuel, nor the buildup of fission products were taken into account in this study. Since these phenomena can also be described by the equilibrium cycle average flux ratios, it is likely that this method can be extended to a more realistic method for global in core fuel management optimization. (orig./GL)

  14. Project Management Approach to Transition of the Miamisburg Closure Project From Environmental Cleanup to Post-Closure Operations

    International Nuclear Information System (INIS)

    Carpenter, C.P.; Marks, M.L.; Smiley, S.L.; Gallaher, D.M.; Williams, K.D.

    2006-01-01

    The U.S. Department of Energy (DOE) used a project management approach to transition the Miamisburg Closure Project from cleanup by the Office of Environmental Management (EM) to post-closure operations by the Office of Legacy Management (LM). Two primary DOE orders were used to guide the site transition: DOE Order 430.1B, Real Property Asset Management, for assessment and disposition of real property assets and DOE Order 413.3, Program and Project Management for Acquisition of Capital Assets, for project closeout of environmental cleanup activities and project transition of post-closure activities. To effectively manage these multiple policy requirements, DOE chose to manage the Miamisburg Closure Project as a project under a cross-member transitional team using representatives from four principal organizations: DOE-LM, the LM contractor S.M. Stoller Corporation, DOE-EM, and the EM contractor CH2M Hill Mound Inc. The mission of LM is to manage the Department's post-transition responsibilities and long-term care of legacy liabilities and to ensure the future protection of human health and the environment for cleanup sites after the EM has completed its cleanup activities. (authors)

  15. An average salary: approaches to the index determination

    Directory of Open Access Journals (Sweden)

    T. M. Pozdnyakova

    2017-01-01

    Full Text Available The article “An average salary: approaches to the index determination” is devoted to studying various methods of calculating this index, both used by official state statistics of the Russian Federation and offered by modern researchers.The purpose of this research is to analyze the existing approaches to calculating the average salary of employees of enterprises and organizations, as well as to make certain additions that would help to clarify this index.The information base of the research is laws and regulations of the Russian Federation Government, statistical and analytical materials of the Federal State Statistics Service of Russia for the section «Socio-economic indexes: living standards of the population», as well as materials of scientific papers, describing different approaches to the average salary calculation. The data on the average salary of employees of educational institutions of the Khabarovsk region served as the experimental base of research. In the process of conducting the research, the following methods were used: analytical, statistical, calculated-mathematical and graphical.The main result of the research is an option of supplementing the method of calculating average salary index within enterprises or organizations, used by Goskomstat of Russia, by means of introducing a correction factor. Its essence consists in the specific formation of material indexes for different categories of employees in enterprises or organizations, mainly engaged in internal secondary jobs. The need for introducing this correction factor comes from the current reality of working conditions of a wide range of organizations, when an employee is forced, in addition to the main position, to fulfill additional job duties. As a result, the situation is frequent when the average salary at the enterprise is difficult to assess objectively because it consists of calculating multiple rates per staff member. In other words, the average salary of

  16. 7 CFR 1437.11 - Average market price and payment factors.

    Science.gov (United States)

    2010-01-01

    ... 7 Agriculture 10 2010-01-01 2010-01-01 false Average market price and payment factors. 1437.11... ASSISTANCE PROGRAM General Provisions § 1437.11 Average market price and payment factors. (a) An average... average market price by the applicable payment factor (i.e., harvested, unharvested, or prevented planting...

  17. Anomalous behavior of q-averages in nonextensive statistical mechanics

    International Nuclear Information System (INIS)

    Abe, Sumiyoshi

    2009-01-01

    A generalized definition of average, termed the q-average, is widely employed in the field of nonextensive statistical mechanics. Recently, it has however been pointed out that such an average value may behave unphysically under specific deformations of probability distributions. Here, the following three issues are discussed and clarified. Firstly, the deformations considered are physical and may be realized experimentally. Secondly, in view of the thermostatistics, the q-average is unstable in both finite and infinite discrete systems. Thirdly, a naive generalization of the discussion to continuous systems misses a point, and a norm better than the L 1 -norm should be employed for measuring the distance between two probability distributions. Consequently, stability of the q-average is shown not to be established in all of the cases

  18. Books average previous decade of economic misery.

    Science.gov (United States)

    Bentley, R Alexander; Acerbi, Alberto; Ormerod, Paul; Lampos, Vasileios

    2014-01-01

    For the 20(th) century since the Depression, we find a strong correlation between a 'literary misery index' derived from English language books and a moving average of the previous decade of the annual U.S. economic misery index, which is the sum of inflation and unemployment rates. We find a peak in the goodness of fit at 11 years for the moving average. The fit between the two misery indices holds when using different techniques to measure the literary misery index, and this fit is significantly better than other possible correlations with different emotion indices. To check the robustness of the results, we also analysed books written in German language and obtained very similar correlations with the German economic misery index. The results suggest that millions of books published every year average the authors' shared economic experiences over the past decade.

  19. Asynchronous Gossip for Averaging and Spectral Ranking

    Science.gov (United States)

    Borkar, Vivek S.; Makhijani, Rahul; Sundaresan, Rajesh

    2014-08-01

    We consider two variants of the classical gossip algorithm. The first variant is a version of asynchronous stochastic approximation. We highlight a fundamental difficulty associated with the classical asynchronous gossip scheme, viz., that it may not converge to a desired average, and suggest an alternative scheme based on reinforcement learning that has guaranteed convergence to the desired average. We then discuss a potential application to a wireless network setting with simultaneous link activation constraints. The second variant is a gossip algorithm for distributed computation of the Perron-Frobenius eigenvector of a nonnegative matrix. While the first variant draws upon a reinforcement learning algorithm for an average cost controlled Markov decision problem, the second variant draws upon a reinforcement learning algorithm for risk-sensitive control. We then discuss potential applications of the second variant to ranking schemes, reputation networks, and principal component analysis.

  20. Benchmarking statistical averaging of spectra with HULLAC

    Science.gov (United States)

    Klapisch, Marcel; Busquet, Michel

    2008-11-01

    Knowledge of radiative properties of hot plasmas is important for ICF, astrophysics, etc When mid-Z or high-Z elements are present, the spectra are so complex that one commonly uses statistically averaged description of atomic systems [1]. In a recent experiment on Fe[2], performed under controlled conditions, high resolution transmission spectra were obtained. The new version of HULLAC [3] allows the use of the same model with different levels of details/averaging. We will take advantage of this feature to check the effect of averaging with comparison with experiment. [1] A Bar-Shalom, J Oreg, and M Klapisch, J. Quant. Spectros. Rad. Transf. 65, 43 (2000). [2] J. E. Bailey, G. A. Rochau, C. A. Iglesias et al., Phys. Rev. Lett. 99, 265002-4 (2007). [3]. M. Klapisch, M. Busquet, and A. Bar-Shalom, AIP Conference Proceedings 926, 206-15 (2007).

  1. Representing Development

    DEFF Research Database (Denmark)

    Representing Development presents the different social representations that have formed the idea of development in Western thinking over the past three centuries. Offering an acute perspective on the current state of developmental science and providing constructive insights into future pathways, ...

  2. Equitable Financial Evaluation Method for Public-Private Partnership Projects

    Institute of Scientific and Technical Information of China (English)

    KE Yongjian; LIU Xinping; WANG Shouqing

    2008-01-01

    The feasibility study of a public-private partnership (PPP) project is regarded as one of the critical factors for successful implementation,but unfortunately the common financial evaluation methods currently used only represent the benefits of the private sector.There is,therefore,an urgent need to develop an equitable financial evaluation method for PPP projects.This paper presents a comprehensive literature review that examines international practices.An equitable financial evaluation method was then developed taking into account the inherent characteristics of PPP projects using six separate indicators and Monte Carlo simulations.The result for a bridge project in Romania shows that the method combines the viewpoints of all the relevant stakeholders to achieve an equitable financial evaluation of PPP projects.

  3. Constructing and Representing: a New Project for 3d Surveying of Yazilikaya - HATTUŠA

    Science.gov (United States)

    Repola, L.; Marazzi, M.; Tilia, S.

    2017-05-01

    Within the cooperation project between the University Suor Orsola Benincasa of Naples and the archaeological mission in Hattuša of the German Archaeological Institute of Istanbul, directed by Andreas Schachner, in agreement with the Turkish Ministry of Culture and Tourism, the workgroup of the University of Naples, has carried out, in September 2015, a first survey campaign of the whole rocky site of Yazılıkaya. The experimentation has been finalized at constructing a global 3D territorial and monumental model of the site, capable that is, through the application of differing scanning procedures, according to the different components (topography, rocky complex, the cultural spaces therein, complex of sculptural reliefs, inscriptions accompanying the divine representations), of virtually reproducing in detail, for safegaurd, exhibition and study purposes (in particular from an epigraphical and historic-artistic point of view) all the aspects characterizing the artefact and not completely visible to the naked eye today.

  4. The Yangtze-Project

    Science.gov (United States)

    Subklew, Günter; Ulrich, Julia; Fürst, Leander; Höltkemeier, Agnes

    2010-05-01

    As an important element in Chinese politics for the development of the Western parts of the country, a large hydraulic engineering project - the Three Gorges Dam - has been set up in order to dam the Yangtze River for a length of over 600 km with an average width of about 1,100 m. It is expected that this results in ecological, technical and social problems of a magnitude hardly dealt with before. With this gigantic project, the national executive is pursuing the aims of - preventing flooding - safeguarding the water supply - encouraging navigation and - generating electric energy. In future, fluctuations of the water level of up to 30 metres will be deliberately applied in the dammed-up section of the river while retaining the flow character of the seasonal variation. The pollution of the Yangtze with a wide range of problem substances is frequently underestimated since in many cases attention is only paid to the low measured concentrations. However, the large volumes of water lead to appreciable loads and thus the danger of an accumulation of pollutants even reaching the human food chain. It should also not be forgotten that the Yangtze represents the major, and in some cases indeed the only, source of drinking and domestic water for the population. A consideration of the water level in the impoundment that will in future arise from management of the reservoir reveals the dramatic change in contrast to the natural inundation regime. In the past, the flood events on the banks of the Yangtze and its tributaries occurred in the summer months. The plants in the riparian zone (water fluctuation zone = WFZ) were previously inundated during the warmer time of year (28 ° July/August) and the terrestrial phase of the WFZ was characterized by cool temperatures (3-5 °C January) that permitted little plant activity. In future, the highest water levels will occur in winter above the dam on the Yangtze and also on the tributaries flowing into it. The plants in the WFZ will

  5. An approximate analytical approach to resampling averages

    DEFF Research Database (Denmark)

    Malzahn, Dorthe; Opper, M.

    2004-01-01

    Using a novel reformulation, we develop a framework to compute approximate resampling data averages analytically. The method avoids multiple retraining of statistical models on the samples. Our approach uses a combination of the replica "trick" of statistical physics and the TAP approach for appr...... for approximate Bayesian inference. We demonstrate our approach on regression with Gaussian processes. A comparison with averages obtained by Monte-Carlo sampling shows that our method achieves good accuracy....

  6. Meeting of the ITER CTA project board

    International Nuclear Information System (INIS)

    Vlasenkov, V.

    2002-01-01

    This is information about meeting of the ITER CTA project board, which took place in Moscow, Russian Federation on 22 April 2002 on the occasion of the Third Negotiators Meeting (N3). Thirteen participants, representing PB members and experts from Canada (CA), the European Union (EU), Japan(JA), the Russian federation (RF) and the International Team (IT) attended the meeting chaired by Acad. E. Velikhov. The Project Board took note of the comments made concerning the status of the Participants Teams(PTs)

  7. High average power supercontinuum sources

    Indian Academy of Sciences (India)

    The physical mechanisms and basic experimental techniques for the creation of high average spectral power supercontinuum sources is briefly reviewed. We focus on the use of high-power ytterbium-doped fibre lasers as pump sources, and the use of highly nonlinear photonic crystal fibres as the nonlinear medium.

  8. Average monthly and annual climate maps for Bolivia

    KAUST Repository

    Vicente-Serrano, Sergio M.

    2015-02-24

    This study presents monthly and annual climate maps for relevant hydroclimatic variables in Bolivia. We used the most complete network of precipitation and temperature stations available in Bolivia, which passed a careful quality control and temporal homogenization procedure. Monthly average maps at the spatial resolution of 1 km were modeled by means of a regression-based approach using topographic and geographic variables as predictors. The monthly average maximum and minimum temperatures, precipitation and potential exoatmospheric solar radiation under clear sky conditions are used to estimate the monthly average atmospheric evaporative demand by means of the Hargreaves model. Finally, the average water balance is estimated on a monthly and annual scale for each 1 km cell by means of the difference between precipitation and atmospheric evaporative demand. The digital layers used to create the maps are available in the digital repository of the Spanish National Research Council.

  9. Systematic impact assessment on inter-basin water transfer projects of the Hanjiang River Basin in China

    Science.gov (United States)

    Zhou, Yanlai; Guo, Shenglian; Hong, Xingjun; Chang, Fi-John

    2017-10-01

    China's inter-basin water transfer projects have gained increasing attention in recent years. This study proposes an intelligent water allocation methodology for establishing optimal inter-basin water allocation schemes and assessing the impacts of water transfer projects on water-demanding sectors in the Hanjiang River Basin of China. We first analyze water demands for water allocation purpose, and then search optimal water allocation strategies for maximizing the water supply to water-demanding sectors and mitigating the negative impacts by using the Standard Genetic Algorithm (SGA) and Adaptive Genetic Algorithm (AGA), respectively. Lastly, the performance indexes of the water supply system are evaluated under different scenarios of inter-basin water transfer projects. The results indicate that: the AGA with adaptive crossover and mutation operators could increase the average annual water transfer from the Hanjiang River by 0.79 billion m3 (8.8%), the average annual water transfer from the Changjiang River by 0.18 billion m3 (6.5%), and the average annual hydropower generation by 0.49 billion kW h (5.4%) as well as reduce the average annual unmet water demand by 0.40 billion m3 (9.7%), as compared with the those of the SGA. We demonstrate that the proposed intelligent water allocation schemes can significantly mitigate the negative impacts of inter-basin water transfer projects on the reliability, vulnerability and resilience of water supply to the demanding sectors in water-supplying basins. This study has a direct bearing on more intelligent and effectual water allocation management under various scenarios of inter-basin water transfer projects.

  10. Advanced neutron source project

    International Nuclear Information System (INIS)

    Gorynina, L.V.; Proskuryakov, S.F.; Tishchenko, V.A.; Uzhanova, V.V.

    1991-01-01

    The project of the ANS improved neutron source intended for fundamental researches in nuclear physics and materials testing is considered. New superhigh-flux heavy-water 350 MW reactor is used for the source creation. The standard fuel is uranium silicide (U 3 Si 2 ). Reactor core volume equals 67.4 l and average power density is 4.9 MW/l. Neutron flux density is 10 16 neutron/(cm 2 xs). The facility construction begin is planned for 1996. The first experiments should be accomplished in 2000

  11. Average stress in a Stokes suspension of disks

    NARCIS (Netherlands)

    Prosperetti, Andrea

    2004-01-01

    The ensemble-average velocity and pressure in an unbounded quasi-random suspension of disks (or aligned cylinders) are calculated in terms of average multipoles allowing for the possibility of spatial nonuniformities in the system. An expression for the stress due to the suspended particles is

  12. Representing dispositions

    Directory of Open Access Journals (Sweden)

    Röhl Johannes

    2011-08-01

    Full Text Available Abstract Dispositions and tendencies feature significantly in the biomedical domain and therefore in representations of knowledge of that domain. They are not only important for specific applications like an infectious disease ontology, but also as part of a general strategy for modelling knowledge about molecular interactions. But the task of representing dispositions in some formal ontological systems is fraught with several problems, which are partly due to the fact that Description Logics can only deal well with binary relations. The paper will discuss some of the results of the philosophical debate about dispositions, in order to see whether the formal relations needed to represent dispositions can be broken down to binary relations. Finally, we will discuss problems arising from the possibility of the absence of realizations, of multi-track or multi-trigger dispositions and offer suggestions on how to deal with them.

  13. Average glandular dose in digital mammography and breast tomosynthesis

    Energy Technology Data Exchange (ETDEWEB)

    Olgar, T. [Ankara Univ. (Turkey). Dept. of Engineering Physics; Universitaetsklinikum Leipzig AoeR (Germany). Klinik und Poliklinik fuer Diagnostische und Interventionelle Radiologie; Kahn, T.; Gosch, D. [Universitaetsklinikum Leipzig AoeR (Germany). Klinik und Poliklinik fuer Diagnostische und Interventionelle Radiologie

    2012-10-15

    Purpose: To determine the average glandular dose (AGD) in digital full-field mammography (2 D imaging mode) and in breast tomosynthesis (3 D imaging mode). Materials and Methods: Using the method described by Boone, the AGD was calculated from the exposure parameters of 2247 conventional 2 D mammograms and 984 mammograms in 3 D imaging mode of 641 patients examined with the digital mammographic system Hologic Selenia Dimensions. The breast glandular tissue content was estimated by the Hologic R2 Quantra automated volumetric breast density measurement tool for each patient from right craniocaudal (RCC) and left craniocaudal (LCC) images in 2 D imaging mode. Results: The mean compressed breast thickness (CBT) was 52.7 mm for craniocaudal (CC) and 56.0 mm for mediolateral oblique (MLO) views. The mean percentage of breast glandular tissue content was 18.0 % and 17.4 % for RCC and LCC projections, respectively. The mean AGD values in 2 D imaging mode per exposure for the standard breast were 1.57 mGy and 1.66 mGy, while the mean AGD values after correction for real breast composition were 1.82 mGy and 1.94 mGy for CC and MLO views, respectively. The mean AGD values in 3 D imaging mode per exposure for the standard breast were 2.19 mGy and 2.29 mGy, while the mean AGD values after correction for the real breast composition were 2.53 mGy and 2.63 mGy for CC and MLO views, respectively. No significant relationship was found between the AGD and CBT in 2 D imaging mode and a good correlation coefficient of 0.98 in 3 D imaging mode. Conclusion: In this study the mean calculated AGD per exposure in 3 D imaging mode was on average 34 % higher than for 2 D imaging mode for patients examined with the same CBT.

  14. Average glandular dose in digital mammography and breast tomosynthesis

    International Nuclear Information System (INIS)

    Olgar, T.; Universitaetsklinikum Leipzig AoeR; Kahn, T.; Gosch, D.

    2012-01-01

    Purpose: To determine the average glandular dose (AGD) in digital full-field mammography (2 D imaging mode) and in breast tomosynthesis (3 D imaging mode). Materials and Methods: Using the method described by Boone, the AGD was calculated from the exposure parameters of 2247 conventional 2 D mammograms and 984 mammograms in 3 D imaging mode of 641 patients examined with the digital mammographic system Hologic Selenia Dimensions. The breast glandular tissue content was estimated by the Hologic R2 Quantra automated volumetric breast density measurement tool for each patient from right craniocaudal (RCC) and left craniocaudal (LCC) images in 2 D imaging mode. Results: The mean compressed breast thickness (CBT) was 52.7 mm for craniocaudal (CC) and 56.0 mm for mediolateral oblique (MLO) views. The mean percentage of breast glandular tissue content was 18.0 % and 17.4 % for RCC and LCC projections, respectively. The mean AGD values in 2 D imaging mode per exposure for the standard breast were 1.57 mGy and 1.66 mGy, while the mean AGD values after correction for real breast composition were 1.82 mGy and 1.94 mGy for CC and MLO views, respectively. The mean AGD values in 3 D imaging mode per exposure for the standard breast were 2.19 mGy and 2.29 mGy, while the mean AGD values after correction for the real breast composition were 2.53 mGy and 2.63 mGy for CC and MLO views, respectively. No significant relationship was found between the AGD and CBT in 2 D imaging mode and a good correlation coefficient of 0.98 in 3 D imaging mode. Conclusion: In this study the mean calculated AGD per exposure in 3 D imaging mode was on average 34 % higher than for 2 D imaging mode for patients examined with the same CBT.

  15. Salecker-Wigner-Peres clock and average tunneling times

    International Nuclear Information System (INIS)

    Lunardi, Jose T.; Manzoni, Luiz A.; Nystrom, Andrew T.

    2011-01-01

    The quantum clock of Salecker-Wigner-Peres is used, by performing a post-selection of the final state, to obtain average transmission and reflection times associated to the scattering of localized wave packets by static potentials in one dimension. The behavior of these average times is studied for a Gaussian wave packet, centered around a tunneling wave number, incident on a rectangular barrier and, in particular, on a double delta barrier potential. The regime of opaque barriers is investigated and the results show that the average transmission time does not saturate, showing no evidence of the Hartman effect (or its generalized version).

  16. Average wind statistics for SRP area meteorological towers

    International Nuclear Information System (INIS)

    Laurinat, J.E.

    1987-01-01

    A quality assured set of average wind Statistics for the seven SRP area meteorological towers has been calculated for the five-year period 1982--1986 at the request of DOE/SR. A Similar set of statistics was previously compiled for the years 1975-- 1979. The updated wind statistics will replace the old statistics as the meteorological input for calculating atmospheric radionuclide doses from stack releases, and will be used in the annual environmental report. This report details the methods used to average the wind statistics and to screen out bad measurements and presents wind roses generated by the averaged statistics

  17. Control of underactuated driftless systems using higher-order averaging theory

    OpenAIRE

    Vela, Patricio A.; Burdick, Joel W.

    2003-01-01

    This paper applies a recently developed "generalized averaging theory" to construct stabilizing feedback control laws for underactuated driftless systems. These controls exponentialy stabilize in the average; the actual system may orbit around the average. Conditions for which the orbit collapses to the averaged trajectory are given. An example validates the theory, demonstrating its utility.

  18. Estimating the average treatment effect on survival based on observational data and using partly conditional modeling.

    Science.gov (United States)

    Gong, Qi; Schaubel, Douglas E

    2017-03-01

    Treatments are frequently evaluated in terms of their effect on patient survival. In settings where randomization of treatment is not feasible, observational data are employed, necessitating correction for covariate imbalances. Treatments are usually compared using a hazard ratio. Most existing methods which quantify the treatment effect through the survival function are applicable to treatments assigned at time 0. In the data structure of our interest, subjects typically begin follow-up untreated; time-until-treatment, and the pretreatment death hazard are both heavily influenced by longitudinal covariates; and subjects may experience periods of treatment ineligibility. We propose semiparametric methods for estimating the average difference in restricted mean survival time attributable to a time-dependent treatment, the average effect of treatment among the treated, under current treatment assignment patterns. The pre- and posttreatment models are partly conditional, in that they use the covariate history up to the time of treatment. The pre-treatment model is estimated through recently developed landmark analysis methods. For each treated patient, fitted pre- and posttreatment survival curves are projected out, then averaged in a manner which accounts for the censoring of treatment times. Asymptotic properties are derived and evaluated through simulation. The proposed methods are applied to liver transplant data in order to estimate the effect of liver transplantation on survival among transplant recipients under current practice patterns. © 2016, The International Biometric Society.

  19. NOAA Average Annual Salinity (3-Zone)

    Data.gov (United States)

    California Natural Resource Agency — The 3-Zone Average Annual Salinity Digital Geography is a digital spatial framework developed using geographic information system (GIS) technology. These salinity...

  20. Cost curves for implantation of small scale hydroelectric power plant project in function of the average annual energy production; Curvas de custo de implantacao de pequenos projetos hidreletricos em funcao da producao media anual de energia

    Energy Technology Data Exchange (ETDEWEB)

    Veja, Fausto Alfredo Canales; Mendes, Carlos Andre Bulhoes; Beluco, Alexandre

    2008-10-15

    Because of its maturity, small hydropower generation is one of the main energy sources to be considered for electrification of areas far from the national grid. Once a site with hydropower potential is identified, technical and economical studies to assess its feasibility shall be done. Cost curves are helpful tools in the appraisal of the economical feasibility of this type of projects. This paper presents a method to determine initial cost curves as a function of the average energy production of the hydropower plant, by using a set of parametric cost curves and the flow duration curve at the analyzed location. The method is illustrated using information related to 18 pre-feasibility studies made in 2002, at the Central-Atlantic rural region of Nicaragua. (author)

  1. An approach to averaging digitized plantagram curves.

    Science.gov (United States)

    Hawes, M R; Heinemeyer, R; Sovak, D; Tory, B

    1994-07-01

    The averaging of outline shapes of the human foot for the purposes of determining information concerning foot shape and dimension within the context of comfort of fit of sport shoes is approached as a mathematical problem. An outline of the human footprint is obtained by standard procedures and the curvature is traced with a Hewlett Packard Digitizer. The paper describes the determination of an alignment axis, the identification of two ray centres and the division of the total curve into two overlapping arcs. Each arc is divided by equiangular rays which intersect chords between digitized points describing the arc. The radial distance of each ray is averaged within groups of foot lengths which vary by +/- 2.25 mm (approximately equal to 1/2 shoe size). The method has been used to determine average plantar curves in a study of 1197 North American males (Hawes and Sovak 1993).

  2. Books Average Previous Decade of Economic Misery

    Science.gov (United States)

    Bentley, R. Alexander; Acerbi, Alberto; Ormerod, Paul; Lampos, Vasileios

    2014-01-01

    For the 20th century since the Depression, we find a strong correlation between a ‘literary misery index’ derived from English language books and a moving average of the previous decade of the annual U.S. economic misery index, which is the sum of inflation and unemployment rates. We find a peak in the goodness of fit at 11 years for the moving average. The fit between the two misery indices holds when using different techniques to measure the literary misery index, and this fit is significantly better than other possible correlations with different emotion indices. To check the robustness of the results, we also analysed books written in German language and obtained very similar correlations with the German economic misery index. The results suggest that millions of books published every year average the authors' shared economic experiences over the past decade. PMID:24416159

  3. 40 CFR 60.4112 - Changing Hg designated representative and alternate Hg designated representative; changes in...

    Science.gov (United States)

    2010-07-01

    ... 40 Protection of Environment 6 2010-07-01 2010-07-01 false Changing Hg designated representative and alternate Hg designated representative; changes in owners and operators. 60.4112 Section 60.4112... Generating Units Hg Designated Representative for Hg Budget Sources § 60.4112 Changing Hg designated...

  4. Using fuzzy logic to improve the project time and cost estimation based on Project Evaluation and Review Technique (PERT

    Directory of Open Access Journals (Sweden)

    Farhad Habibi

    2018-09-01

    Full Text Available Among different factors, correct scheduling is one of the vital elements for project management success. There are several ways to schedule projects including the Critical Path Method (CPM and Program Evaluation and Review Technique (PERT. Due to problems in estimating dura-tions of activities, these methods cannot accurately and completely model actual projects. The use of fuzzy theory is a basic way to improve scheduling and deal with such problems. Fuzzy theory approximates project scheduling models to reality by taking into account uncertainties in decision parameters and expert experience and mental models. This paper provides a step-by-step approach for accurate estimation of time and cost of projects using the Project Evaluation and Review Technique (PERT and expert views as fuzzy numbers. The proposed method included several steps. In the first step, the necessary information for project time and cost is estimated using the Critical Path Method (CPM and the Project Evaluation and Review Technique (PERT. The second step considers the duration and cost of the project activities as the trapezoidal fuzzy numbers, and then, the time and cost of the project are recalculated. The duration and cost of activities are estimated using the questionnaires as well as weighing the expert opinions, averaging and defuzzification based on a step-by-step algorithm. The calculating procedures for evaluating these methods are applied in a real project; and the obtained results are explained.

  5. Feasibility of using the Omaha System to represent public health nurse manager interventions.

    Science.gov (United States)

    Monsen, Karen A; Newsom, Eric T

    2011-01-01

    To test the feasibility of representing public health nurse (PHN) manager interventions using a recognized standardized nursing terminology. A nurse manager in a Midwest local public health agency documented nurse manager interventions using the Omaha System for 5 months. ANALYTIC STRATEGY: The data were analyzed and the results were compared with the results from a parallel analysis of existing PHN intervention data. Interventions for 79 "clients" (projects, teams, or individuals) captured 76% of recorded work hours, and addressed 43% of Omaha System problems. Most problems were addressed at the "community" level (87.1%) versus the "individual" level (12.9%). Nursing practice differed between the 2 knowledge domains of public health family home visiting nursing and public health nursing management. Standardized nursing terminologies have the potential to represent, describe, and quantify nurse manager interventions for future evaluation and research. © 2011 Wiley Periodicals, Inc.

  6. Trends in diabetes incidence from 1992 to 2015 and projections for 2024: A Portuguese General Practitioner's Network study.

    Science.gov (United States)

    de Sousa-Uva, Mafalda; Antunes, L; Nunes, B; Rodrigues, A P; Simões, J A; Ribeiro, R T; Boavida, J M; Matias-Dias, C

    2016-10-01

    Diabetes is known as a major cause of morbidity and mortality worldwide. Portugal is known as the European country with the highest prevalence of this disease. While diabetes prevalence data is updated annually in Portugal, the General Practitioner's (GP) Sentinel Network represents the only data source on diabetes incidence. This study describes the trends in Diabetes incidence, between 1992 and 2015, and estimate projections for the future incidence rates in Portugal until 2024. An ecological time-series study was conducted using data from GP Sentinel Network between 1992 and 2015. Family doctors reported all new cases of Diabetes in their patients' lists. Annual trends were estimated through Poisson regression models as well as the future incidence rates (until 2024), sex and age group stratified. Incidence rate projections were adjusted to the distribution of the resident Portuguese population given Statistics Portugal projections. The average increase in Diabetes incidence rate was in total 4.29% (CI95% 3.80-4.80) per year under study. Until 1998-2000, the annual incidence rate was higher in women, and from 1998-2000 to 2013-2015 turn out to be higher in men. The incidence rate projected for 2022-2024 was 972.77/10(5) inhabitants in total, and 846.74/10(5) and 1114.42/10(5), respectively, in women and men. This is the first study in Portugal to estimate diabetes incidence rate projections. The disturbing reported projections seem realistic if things continue as in the past. Actually, effective public health policies will need to be undertaken to minimize this alarming future scenario. Copyright © 2016 Primary Care Diabetes Europe. Published by Elsevier Ltd. All rights reserved.

  7. Model averaging, optimal inference and habit formation

    Directory of Open Access Journals (Sweden)

    Thomas H B FitzGerald

    2014-06-01

    Full Text Available Postulating that the brain performs approximate Bayesian inference generates principled and empirically testable models of neuronal function – the subject of much current interest in neuroscience and related disciplines. Current formulations address inference and learning under some assumed and particular model. In reality, organisms are often faced with an additional challenge – that of determining which model or models of their environment are the best for guiding behaviour. Bayesian model averaging – which says that an agent should weight the predictions of different models according to their evidence – provides a principled way to solve this problem. Importantly, because model evidence is determined by both the accuracy and complexity of the model, optimal inference requires that these be traded off against one another. This means an agent’s behaviour should show an equivalent balance. We hypothesise that Bayesian model averaging plays an important role in cognition, given that it is both optimal and realisable within a plausible neuronal architecture. We outline model averaging and how it might be implemented, and then explore a number of implications for brain and behaviour. In particular, we propose that model averaging can explain a number of apparently suboptimal phenomena within the framework of approximate (bounded Bayesian inference, focussing particularly upon the relationship between goal-directed and habitual behaviour.

  8. Spherical projections and liftings in geometric tomography

    DEFF Research Database (Denmark)

    Goodey, Paul; Kiderlen, Markus; Weil, Wolfgang

    2011-01-01

    We consider a variety of integral transforms arising in Geometric Tomography. It will be shown that these can be put into a common framework using spherical projection and lifting operators. These operators will be applied to support functions and surface area measures of convex bodies and to rad......We consider a variety of integral transforms arising in Geometric Tomography. It will be shown that these can be put into a common framework using spherical projection and lifting operators. These operators will be applied to support functions and surface area measures of convex bodies...... and to radial functions of star bodies. We then investigate averages of lifted projections and show that they correspond to self-adjoint intertwining operators. We obtain formulas for the eigenvalues of these operators and use them to ascertain circumstances under which tomographic measurements determine...... the original bodies. This approach via mean lifted projections leads us to some unexpected relationships between seemingly disparate geometric constructions....

  9. Bootstrapping pre-averaged realized volatility under market microstructure noise

    DEFF Research Database (Denmark)

    Hounyo, Ulrich; Goncalves, Sílvia; Meddahi, Nour

    The main contribution of this paper is to propose a bootstrap method for inference on integrated volatility based on the pre-averaging approach of Jacod et al. (2009), where the pre-averaging is done over all possible overlapping blocks of consecutive observations. The overlapping nature of the pre......-averaged returns implies that these are kn-dependent with kn growing slowly with the sample size n. This motivates the application of a blockwise bootstrap method. We show that the "blocks of blocks" bootstrap method suggested by Politis and Romano (1992) (and further studied by Bühlmann and Künsch (1995......)) is valid only when volatility is constant. The failure of the blocks of blocks bootstrap is due to the heterogeneity of the squared pre-averaged returns when volatility is stochastic. To preserve both the dependence and the heterogeneity of squared pre-averaged returns, we propose a novel procedure...

  10. Local participation in complex technological projects as bridging between different communities in Belgium

    NARCIS (Netherlands)

    Sips, K.; Craps, M.; Dewulf, A.

    2013-01-01

    Local community participation in complex technological projects, where technological innovations and risks need to be managed, is notoriously challenging. Relations with local inhabitants easily take the form of exclusion, protest, controversy or litigation. While such projects represent

  11. Assessing the representativeness of wind data for wind turbine site evaluation

    Science.gov (United States)

    Renne, D. S.; Corotis, R. B.

    1982-01-01

    Once potential wind turbine sites (either for single installations or clusters) are identified through siting procedures, actual evaluation of the sites must commence. This evaluation is needed to obtain estimates of wind turbine performance and to identify hazards to the machine from the turbulence component of the atmosphere. These estimates allow for more detailed project planning and for preliminary financing arrangements to be secured. The site evaluation process can occur in two stages: (1) utilizing existing nearby data, and (2) establishing and monitoring an onsite measurement program. Since step (2) requires a period of at least 1 yr or more from the time a potential site has been identified, step (1) is often an essential stage in the preliminary evaluation process. Both the methods that have been developed and the unknowns that still exist in assessing the representativeness of available data to a nearby wind turbine site are discussed. How the assessment of the representativeness of available data can be used to develop a more effective onsite meteorological measurement program is also discussed.

  12. MN Temperature Average (1961-1990) - Line

    Data.gov (United States)

    Minnesota Department of Natural Resources — This data set depicts 30-year averages (1961-1990) of monthly and annual temperatures for Minnesota. Isolines and regions were created using kriging and...

  13. MN Temperature Average (1961-1990) - Polygon

    Data.gov (United States)

    Minnesota Department of Natural Resources — This data set depicts 30-year averages (1961-1990) of monthly and annual temperatures for Minnesota. Isolines and regions were created using kriging and...

  14. High average-power induction linacs

    International Nuclear Information System (INIS)

    Prono, D.S.; Barrett, D.; Bowles, E.

    1989-01-01

    Induction linear accelerators (LIAs) are inherently capable of accelerating several thousand amperes of /approximately/ 50-ns duration pulses to > 100 MeV. In this paper we report progress and status in the areas of duty factor and stray power management. These technologies are vital if LIAs are to attain high average power operation. 13 figs

  15. Interpreting Bivariate Regression Coefficients: Going beyond the Average

    Science.gov (United States)

    Halcoussis, Dennis; Phillips, G. Michael

    2010-01-01

    Statistics, econometrics, investment analysis, and data analysis classes often review the calculation of several types of averages, including the arithmetic mean, geometric mean, harmonic mean, and various weighted averages. This note shows how each of these can be computed using a basic regression framework. By recognizing when a regression model…

  16. [Earth Science Technology Office's Computational Technologies Project

    Science.gov (United States)

    Fischer, James (Technical Monitor); Merkey, Phillip

    2005-01-01

    This grant supported the effort to characterize the problem domain of the Earth Science Technology Office's Computational Technologies Project, to engage the Beowulf Cluster Computing Community as well as the High Performance Computing Research Community so that we can predict the applicability of said technologies to the scientific community represented by the CT project and formulate long term strategies to provide the computational resources necessary to attain the anticipated scientific objectives of the CT project. Specifically, the goal of the evaluation effort is to use the information gathered over the course of the Round-3 investigations to quantify the trends in scientific expectations, the algorithmic requirements and capabilities of high-performance computers to satisfy this anticipated need.

  17. Development and significance of a fetal electrocardiogram recorded by signal-averaged high-amplification electrocardiography.

    Science.gov (United States)

    Hayashi, Risa; Nakai, Kenji; Fukushima, Akimune; Itoh, Manabu; Sugiyama, Toru

    2009-03-01

    Although ultrasonic diagnostic imaging and fetal heart monitors have undergone great technological improvements, the development and use of fetal electrocardiograms to evaluate fetal arrhythmias and autonomic nervous activity have not been fully established. We verified the clinical significance of the novel signal-averaged vector-projected high amplification ECG (SAVP-ECG) method in fetuses from 48 gravidas at 32-41 weeks of gestation and in 34 neonates. SAVP-ECGs from fetuses and newborns were recorded using a modified XYZ-leads system. Once noise and maternal QRS waves were removed, the P, QRS, and T wave intervals were measured from the signal-averaged fetal ECGs. We also compared fetal and neonatal heart rates (HRs), coefficients of variation of heart rate variability (CV) as a parasympathetic nervous activity, and the ratio of low to high frequency (LF/HF ratio) as a sympathetic nervous activity. The rate of detection of a fetal ECG by SAVP-ECG was 72.9%, and the fetal and neonatal QRS and QTc intervals were not significantly different. The neonatal CVs and LF/HF ratios were significantly increased compared with those in the fetus. In conclusion, we have developed a fetal ECG recording method using the SAVP-ECG system, which we used to evaluate autonomic nervous system development.

  18. Saturated Zone Flow and Transport Expert Elicitation Project

    Energy Technology Data Exchange (ETDEWEB)

    Coppersmith, Kevin J.; Perman, Roseanne C.

    1998-01-01

    This report presents results of the Saturated Zone Flow and Transport Expert Elicitation (SZEE) project for Yucca Mountain, Nevada. This project was sponsored by the US Department of Energy (DOE) and managed by Geomatrix Consultants, Inc. (Geomatrix), for TRW Environmental Safety Systems, Inc. The DOE's Yucca Mountain Site Characterization Project (referred to as the YMP) is intended to evaluate the suitability of the site for construction of a mined geologic repository for the permanent disposal of spent nuclear fuel and high-level radioactive waste. The SZEE project is one of several that involve the elicitation of experts to characterize the knowledge and uncertainties regarding key inputs to the Yucca Mountain Total System Performance Assessment (TSPA). The objective of the current project was to characterize the uncertainties associated with certain key issues related to the saturated zone system in the Yucca Mountain area and downgradient region. An understanding of saturated zone processes is critical to evaluating the performance of the potential high-level nuclear waste repository at Yucca Mountain. A major goal of the project was to capture the uncertainties involved in assessing the saturated flow processes, including uncertainty in both the models used to represent the physical processes controlling saturated zone flow and transport, and the parameter values used in the models. So that the analysis included a wide range of perspectives, multiple individual judgments were elicited from members of an expert panel. The panel members, who were experts from within and outside the Yucca Mountain project, represented a range of experience and expertise. A deliberate process was followed in facilitating interactions among the experts, in training them to express their uncertainties, and in eliciting their interpretations. The resulting assessments and probability distributions, therefore, provide a reasonable aggregate representation of the knowledge and

  19. GI Joe or Average Joe? The impact of average-size and muscular male fashion models on men's and women's body image and advertisement effectiveness.

    Science.gov (United States)

    Diedrichs, Phillippa C; Lee, Christina

    2010-06-01

    Increasing body size and shape diversity in media imagery may promote positive body image. While research has largely focused on female models and women's body image, men may also be affected by unrealistic images. We examined the impact of average-size and muscular male fashion models on men's and women's body image and perceived advertisement effectiveness. A sample of 330 men and 289 women viewed one of four advertisement conditions: no models, muscular, average-slim or average-large models. Men and women rated average-size models as equally effective in advertisements as muscular models. For men, exposure to average-size models was associated with more positive body image in comparison to viewing no models, but no difference was found in comparison to muscular models. Similar results were found for women. Internalisation of beauty ideals did not moderate these effects. These findings suggest that average-size male models can promote positive body image and appeal to consumers. 2010 Elsevier Ltd. All rights reserved.

  20. Agro-climate Projections for a Warming Alaska

    Science.gov (United States)

    Lader, R.; Walsh, J. E.; Bhatt, U. S.; Bieniek, P.

    2017-12-01

    In the context of greenhouse warming, agro-meteorological indices suggest widespread disruption to current food supply chains during the coming decades. Much of the western United States is projected to have more dry days, and the southern states are likely to experience greater plant heat stress. Considering these difficulties, it could become necessary for more northerly locations, including Alaska, to increase agricultural production to support local communities and offset supply shortages. This study employs multiple dynamically downscaled regional climate model simulations from the CMIP5 to investigate projected changes to agro-climate conditions across Alaska. The metric used here, the start-of-field operations index (SFO), identifies the date during which the sum of daily average temperature, starting from January 1st and excluding negative values, exceeds 200 ˚C. Using the current trajectory of greenhouse radiative forcing, RCP 8.5, this study indicates a doubling to 71,960 km2 of Alaska land area that meets the required thermal accumulation for crop production when comparing a historical period (1981-2010) to the future (2071-2100). The SFO shows a correlation coefficient of 0.91 with the independently produced green-up index for Fairbanks from 1981-2010. Among the land areas that currently reach the necessary thermal accumulation, there is a projected increase in growing season length (63-82 days), earlier date of last spring frost (28-48 days), and later date of first autumn frost (24-47 days) across the five USDA Census of Agriculture areas for Alaska. Both an average statewide decrease of annual frost days (71 fewer), and an increase in days with extreme warmth (28 more) are also projected.