WorldWideScience

Sample records for assessment ioa analysis

  1. Independent Orbiter Assessment (IOA): Analysis of the pyrotechnics subsystem

    Science.gov (United States)

    Robinson, W. W.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA approach features a top-down analysis of the hardware to determine failure modes, criticality, and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. This report documents the independent analysis results corresponding to the Orbiter Pyrotechnics hardware. The IOA analysis process utilized available pyrotechnics hardware drawings and schematics for defining hardware assemblies, components, and hardware items. Each level of hardware was evaluated and analyzed for possible failure modes and effects. Criticality was assigned based upon the severity of the effect for each failure mode.

  2. Independent Orbiter Assessment (IOA): Weibull analysis report

    Science.gov (United States)

    Raffaelli, Gary G.

    1987-01-01

    The Auxiliary Power Unit (APU) and Hydraulic Power Unit (HPU) Space Shuttle Subsystems were reviewed as candidates for demonstrating the Weibull analysis methodology. Three hardware components were identified as analysis candidates: the turbine wheel, the gearbox, and the gas generator. Detailed review of subsystem level wearout and failure history revealed the lack of actual component failure data. In addition, component wearout data were not readily available or would require a separate data accumulation effort by the vendor. Without adequate component history data being available, the Weibull analysis methodology application to the APU and HPU subsystem group was terminated.

  3. Independent Orbiter Assessment (IOA): Analysis of the hydraulics/water spray boiler subsystem

    Science.gov (United States)

    Duval, J. D.; Davidson, W. R.; Parkman, William E.

    1986-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA approach features a top-down analysis of the hardware to determine failure modes, criticality, and potential critical items (PCIs). To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. This report documents the independent analysis results for the Orbiter Hydraulics/Water Spray Boiler Subsystem. The hydraulic system provides hydraulic power to gimbal the main engines, actuate the main engine propellant control valves, move the aerodynamic flight control surfaces, lower the landing gear, apply wheel brakes, steer the nosewheel, and dampen the external tank (ET) separation. Each hydraulic system has an associated water spray boiler which is used to cool the hydraulic fluid and APU lubricating oil. The IOA analysis process utilized available HYD/WSB hardware drawings, schematics and documents for defining hardware assemblies, components, and hardware items. Each level of hardware was evaluated and analyzed for possible failure modes and effects. Criticality was assigned based upon the severity of the effect for each failure mode. Of the 430 failure modes analyzed, 166 were determined to be PCIs.

  4. Independent Orbiter Assessment (IOA): Assessment of the pyrotechnics subsystem

    Science.gov (United States)

    Robinson, W. M.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Pyrotechnics (PYRO) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. This report documents the results of that comparison for the Orbiter Pyrotechnics hardware.

  5. Independent Orbiter Assessment (IOA): Analysis of the landing/deceleration subsystem

    Science.gov (United States)

    Compton, J. M.; Beaird, H. G.; Weissinger, W. D.

    1987-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA approach features a top-down analysis of the hardware to determine failure modes, criticality, and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. This report documents the independent analysis results corresponding to the Orbiter Landing/Deceleration Subsystem hardware. The Landing/Deceleration Subsystem is utilized to allow the Orbiter to perform a safe landing, allowing for landing-gear deploy activities, steering and braking control throughout the landing rollout to wheel-stop, and to allow for ground-handling capability during the ground-processing phase of the flight cycle. Specifically, the Landing/Deceleration hardware consists of the following components: Nose Landing Gear (NLG); Main Landing Gear (MLG); Brake and Antiskid (B and AS) Electrical Power Distribution and Controls (EPD and C); Nose Wheel Steering (NWS); and Hydraulics Actuators. Each level of hardware was evaluated and analyzed for possible failure modes and effects. Criticality was assigned based upon the severity of the effect for each failure mode. Due to the lack of redundancy in the Landing/Deceleration Subsystems there is a high number of critical items.

  6. Independent Orbiter Assessment (IOA): Assessment of the main propulsion subsystem FMEA/CIL, volume 4

    Science.gov (United States)

    Slaughter, B. C.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Main Propulsion System (MPS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were than compared to available data from the Rockwell Downey/NASA JSC FMEA/CIL review. Volume 4 contains the IOA analysis worksheets and the NASA FMEA to IOA worksheet cross reference and recommendations.

  7. Independent Orbiter Assessment (IOA): Assessment of the main propulsion subsystem FMEA/CIL, volume 2

    Science.gov (United States)

    Holden, K. A.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Main Propulsion System (MPS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were than compared to available data from the Rockwell Downey/NASA JSC FMEA/CIL review. Volume 2 continues the presentation of IOA worksheets for MPS hardware items.

  8. Independent Orbiter Assessment (IOA): Assessment of the main propulsion subsystem FMEA/CIL, volume 3

    Science.gov (United States)

    Holden, K. A.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Main Propulsion System (MPS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to available data from the Rockwell Downey/NASA JSC FMEA/CIL review. Volume 3 continues the presentation of IOA worksheets and includes the potential critical items list.

  9. Independent Orbiter Assessment (IOA): Assessment of the backup flight system FMEA/CIL

    Science.gov (United States)

    Prust, E. E.; Ewell, J. J., Jr.; Hinsdale, L. W.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Backup Flight System (BFS) hardware, generating draft failure modes and Potential Critical Items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the proposed NASA Post 51-L FMEA/CIL baseline. A resolution of each discrepancy from the comparison is provided through additional analysis as required. This report documents the results of that comparison for the Orbiter BFS hardware. The IOA product for the BFS analysis consisted of 29 failure mode worksheets that resulted in 21 Potential Critical Items (PCI) being identified. This product was originally compared with the proposed NASA BFS baseline and subsequently compared with the applicable Data Processing System (DPS), Electrical Power Distribution and Control (EPD and C), and Displays and Controls NASA CIL items. The comparisons determined if there were any results which had been found by the IOA but were not in the NASA baseline. The original assessment determined there were numerous failure modes and potential critical items in the IOA analysis that were not contained in the NASA BFS baseline. Conversely, the NASA baseline contained three FMEAs (IMU, ADTA, and Air Data Probe) for CIL items that were not identified in the IOA product.

  10. Independent Orbiter Assessment (IOA): Assessment of the data processing system FMEA/CIL

    Science.gov (United States)

    Lowery, H. J.; Haufler, W. A.

    1986-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Data Processing System (DPS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. The results of that comparison is documented for the Orbiter DPS hardware.

  11. Independent Orbiter Assessment (IOA): FMEA/CIL assessment

    Science.gov (United States)

    Hinsdale, L. W.; Swain, L. J.; Barnes, J. E.

    1988-01-01

    The McDonnell Douglas Astronautics Company (MDAC) was selected to perform an Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL). Direction was given by the Orbiter and GFE Projects Office to perform the hardware analysis and assessment using the instructions and ground rules defined in NSTS 22206. The IOA analysis featured a top-down approach to determine hardware failure modes, criticality, and potential critical items. To preserve independence, the analysis was accomplished without reliance upon the results contained within the NASA and Prime Contractor FMEA/CIL documentation. The assessment process compared the independently derived failure modes and criticality assignments to the proposed NASA post 51-L FMEA/CIL documentation. When possible, assessment issues were discussed and resolved with the NASA subsystem managers. Unresolved issues were elevated to the Orbiter and GFE Projects Office manager, Configuration Control Board (CCB), or Program Requirements Control Board (PRCB) for further resolution. The most important Orbiter assessment finding was the previously unknown stuck autopilot push-button criticality 1/1 failure mode. The worst case effect could cause loss of crew/vehicle when the microwave landing system is not active. It is concluded that NASA and Prime Contractor Post 51-L FMEA/CIL documentation assessed by IOA is believed to be technically accurate and complete. All CIL issues were resolved. No FMEA issues remain that have safety implications. Consideration should be given, however, to upgrading NSTS 22206 with definitive ground rules which more clearly spell out the limits of redundancy.

  12. Independent Orbiter Assessment (IOA): Assessment of the orbital maneuvering system FMEA/CIL, volume 1

    Science.gov (United States)

    Prust, Chet D.; Haufler, W. A.; Marino, A. J.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Orbital Maneuvering System (OMS) hardware and Electrical Power Distribution and Control (EPD and C), generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the proposed Post 51-L NASA FMEA/CIL baseline. This report documents the results of that comparison for the Orbiter OMS hardware. The IOA analysis defined the OMS as being comprised of the following subsystems: helium pressurization, propellant storage and distribution, Orbital Maneuvering Engine, and EPD and C. The IOA product for the OMS analysis consisted of 284 hardware and 667 EPD and C failure mode worksheets that resulted in 160 hardware and 216 EPD and C potential critical items (PCIs) being identified. A comparison was made of the IOA product to the NASA FMEA/CIL baseline which consisted of 101 hardware and 142 EPD and C CIL items.

  13. Independent Orbiter Assessment (IOA): Assessment of the atmospheric revitalization pressure control subsystem FMEA/CIL

    Science.gov (United States)

    Saiidi, M. J.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the atmospheric Revitalization Pressure Control Subsystem (ARPCS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL proposed Post 51-L updates based upon the CCB/PRCB presentations and an informal criticality summary listing. A discussion of each discrepancy from the comparison is provided through additional analysis as required. These discrepancies were flagged as issues, and recommendations were made based on the FMEA data available at the time. This report documents the results of that comparison for the Orbiter ARPCS hardware.

  14. Independent Orbiter Assessment (IOA): Assessment of the orbiter main propulsion system FMEA/CIL, volume 1

    Science.gov (United States)

    Slaughter, B. C.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Main Propulsion System (MPS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to available data from the Rockwell Downey/NASA JSC FMEA/CIL review. The Orbiter MPS is composed of the Propellant Management Subsystem (PMS) consisting of the liquid oxygen (LO2) and liquid hydrogen (LH2) subsystems and the helium subsystem. The PMS is a system of manifolds, distribution lines, and valves by which the liquid propellants pass from the External Tank to the Space Shuttle Main Engine (SSME). The helium subsystem consists of a series of helium supply tanks and their associated regulators, control valves, and distribution lines. Volume 1 contains the MPS description, assessment results, ground rules and assumptions, and some of the IOA worksheets.

  15. Independent Orbiter Assessment (IOA): Assessment of instrumental subsystem FMEA/CIL

    Science.gov (United States)

    Gardner, J. R.; Addis, A. W.

    1988-01-01

    The McDonnell Douglas Astronautics Company (MDAC) was selected in June 1986 to perform an Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL). The IOA effort first completed an analysis of the Instrumentation hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline. A resolution of each discrepancy from the comparison is provided through additional analysis as required. The results of that comparison for the Orbiter Instrumentation hardware are documented. The IOA product for Instrumentation analysis consisted of 107 failure mode worksheets that resulted in 22 critical items being identified. Comparison was made to the Pre 51-L NASA baseline with 14 Post 51-L FMEAs added, which consists of 96 FMEAs and 18 CIL items. This comparison produced agreement on all but 25 FMEAs which caused differences in 5 CIL items.

  16. Independent Orbiter Assessment (IOA): Assessment of the ascent thrust vector control actuator subsystem FMEA/CIL

    Science.gov (United States)

    Wilson, R. E.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Ascent Thrust Vector Control Actuator (ATVD) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. This report documents the results of that comparison for the Orbiter ATVC hardware. The IOA product for the ATVC actuator analysis consisted of 25 failure mode worksheets that resulted in 16 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 21 FMEAs and 13 CIL items. This comparison produced agreement on all CIL items. Based on the Pre 51-L baseline, all non-CIL FMEAs were also in agreement.

  17. Independent Orbiter Assessment (IOA): Assessment of the elevon actuator subsystem FMEA/CIL

    Science.gov (United States)

    Wilson, R. E.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Elevon Subsystem hardware, generating draft failure modes, and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. This report documents the results of that comparison for the Orbiter Elevon hardware. The IOA product for the Elevon analysis consisted of 25 failure mode worksheets that resulted in 17 potential critical items being identified. Comparison was made to the NASA FMEA/CIL, which consisted of 23 FMEAs and 13 CIL items. This comparison produced agreement on all CIL items. Based on the Pre 51-L baseline, all non-CIL FMEAs were also in agreement.

  18. Independent Orbiter Assessment (IOA): Assessment of the remote manipulator system FMEA/CIL

    Science.gov (United States)

    Tangorra, F.; Grasmeder, R. F.; Montgomery, A. D.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Remote Manipulator System (RMS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were than compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. The results of that comparison for the Orbiter RMS hardware are documented. The IOA product for the RMS analysis consisted of 604 failure mode worksheets that resulted in 458 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 45 FMEAs and 321 CIL items. This comparison produced agreement on all but 154 FMEAs which caused differences in 137 CIL items.

  19. Independent Orbiter Assessment (IOA): Assessment of the rudder/speed brake subsystem FMEA/CIL

    Science.gov (United States)

    Wilson, R. E.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Rudder/Speed Brake (RSB) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline along with the proposed Post 51-L CIL updates included. A resolution of each discrepancy from the comparison was provided through additional analysis as required. This report documents the results of that comparison for the Orbiter RSB hardware. The IOA product for the RSB analysis consisted of 38 failure mode worksheets that resulted in 27 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 34 FMEAs and 18 CIL items. This comparison produced agreement on all CIL items. Based on the Pre 51-L baseline, all non-CIL FMEAs were also in agreement.

  20. Independent Orbiter Assessment (IOA): Assessment of the body flap subsystem FMEA/CIL

    Science.gov (United States)

    Wilson, R. E.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Body Flap (BF) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. This report documents the results of that comparison for the Orbiter BF hardware. The IOA product for the BF analysis consisted of 43 failure mode worksheets that resulted in 19 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 34 FMEAs and 15 CIL items. This comparison produced agreement on all CIL items. Based on the Pre 51-L baseline, all non-CIL FMEAs were also in agreement.

  1. Independent Orbiter Assessment (IOA): Assessment of the guidance, navigation, and control subsystem FMEA/CIL

    Science.gov (United States)

    Trahan, W. H.; Odonnell, R. A.; Pietz, K. C.; Drapela, L. J.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Guidance, Navigation, and Control System (GNC) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. The results of that comparison for the Orbiter GNC hardware is documented. The IOA product for the GNC analysis consisted of 141 failure mode worksheets that resulted in 24 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 148 FMEAs and 36 CIL items. This comparison produced agreement on all but 56 FMEAs which caused differences in zero CIL items.

  2. Independent Orbiter Assessment (IOA): Assessment of the electrical power generation/power reactant storage and distribution subsystem FMEA/CIL

    Science.gov (United States)

    Ames, B. E.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) is presented. The IOA effort first completed an analysis of the Electrical Power Generation/Power Reactant Storage and Distribution (EPG/PRSD) subsystem hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baselines with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. The results of that comparison are documented for the Orbiter EPG/PRSD hardware. The comparison produced agreement on all but 27 FMEAs and 9 CIL items. The discrepancy between the number of IOA findings and NASA FMEAs can be partially explained by the different approaches used by IOA and NASA to group failure modes together to form one FMEA. Also, several IOA items represented inner tank components and ground operations failure modes which were not in the NASA baseline.

  3. Independent Orbiter Assessment (IOA): Assessment of the EPD and C/remote manipulator system FMEA/CIL

    Science.gov (United States)

    Robinson, W. W.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Electrical Power Distribution and Control (EPD and C)/Remote Manipulator System (RMS) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA analysis of the EPD and C/RMS hardware initially generated 345 failure mode worksheets and identified 117 Potential Critical Items (PCIs) before starting the assessment process. These analysis results were compared to the proposed NASA Post 51-L baseline of 132 FMEAs and 66 CIL items.

  4. Independent Orbiter Assessment (IOA): Assessment of the landing/deceleration (LDG/DEC) subsystem FMEA/CIL

    Science.gov (United States)

    Odonnell, R. A.; Weissinger, D.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Landing/Deceleration (LDG/DEC) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison is provided through additional analysis as required. This report documents the results of that comparison for the Orbiter LDG/DEC hardware. The IOA product for the LDG/DEC analysis consisted of 259 failure mode worksheets that resulted in 124 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 267 FMEA's and 120 CIL items. This comparison produced agreement on all but 75 FMEA's which caused differences in 51 CIL items.

  5. Independent Orbiter Assessment (IOA): Assessment of the electrical power generation/fuel cell powerplant subsystem FMEA/CIL

    Science.gov (United States)

    Brown, K. L.; Bertsch, P. J.

    1987-01-01

    Results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Electrical Power Generation/Fuel Cell Powerplant (EPG/FCP) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the proposed Post 51-L NASA FMEA/CIL baseline. A resolution of each discrepancy from the comparison was provided through additional analysis as required. This report documents the results of that comparison for the Orbiter EPG/FCP hardware.

  6. Independent Orbiter Assessment (IOA): Assessment of the Electrical Power Distribution and Control/Electrical Power Generation (EPD and C/EPG) FMEA/CIL

    Science.gov (United States)

    Mccants, C. N.; Bearrow, M.

    1988-01-01

    The results of the Independent Orbiter Assessment (IOA) of the Failure Modes and Effects Analysis (FMEA) and Critical Items List (CIL) are presented. The IOA effort first completed an analysis of the Electrical Power Distribution and Control/Electrical Power Generation (EPD and C/EPG) hardware, generating draft failure modes and potential critical items. To preserve independence, this analysis was accomplished without reliance upon the results contained within the NASA FMEA/CIL documentation. The IOA results were then compared to the NASA FMEA/CIL baseline with proposed Post 51-L updates included. A resolution of each discrepancy from the comparison was provided through additional analysis as required. The results of that comparison is documented for the Orbiter EPD and C/EPG hardware. The IOA product for the EPD and C/EPG analysis consisted of 263 failure mode worksheets that resulted in 42 potential critical items being identified. Comparison was made to the NASA baseline which consisted of 211 FMEA and 47 CIL items.

  7. The PCA and IOA approaches for life-cycle analysis of greenhouse gas emissions from Thai commodities and energy consumption

    Directory of Open Access Journals (Sweden)

    Pawinee Suksuntornsiri

    2005-01-01

    Full Text Available The use of more substitutable commodities, selected by their total greenhouse gas (GHG emissions would highly contribute to mitigating the effects of global warming. Life-cycle analysis (LCA is a solution that can evaluate the total emissions from a lifetime production of a commodity. It is widely applied to reveal the actual environmental emissions in many countries, however this data could not be applied in other countries due to different emission and energy consumption structures. LCA emission factors within the same country are also different due to different assumptions on the boundary and lifetime of a considered production process. The process chains analysis (PCA, the conventional LCA approach which is mostly applied in Thailand, is accurate in the direct production process, but its analysis in the higher order production process is usually truncated due to a lack of data. It is laborious, time consuming, and hard to gather all the data from the whole production process. This article presents the pros and cons of PCA and inputoutput analysis (IOA and introduces an appropriate approach in a Thai context. The adaptation of energyrelated GHG emissions findings in the revised 1996 Intergovernmental Panel on Climate Change (IPCC guidelines for national GHG inventories are introduced for a commodity emission factor and traced through the whole production chains by IOA. In conclusion, the IOA gives emissions in average values and the historical economic structure is used to derive the emissions. However, emissions from every single link of production lifecycle can be taken into account. A combined PCA and IOA is recommended for an LCA of GHG emissions in Thailand.

  8. Independent Orbiter Assessment (IOA): FMEA/CIL instructions and ground rules

    Science.gov (United States)

    Traves, S. T.

    1986-01-01

    The McDonnell Douglas Astronautics Company was selected to conduct an independent assessment of the Orbiter Failure Mode and Effects Analysis/Critical Items List (FMEA/CIL). Part of this effort involved an examination of the FMEA/CIL preparation instructions and ground rules. Assessment objectives were to identify omissions and ambiguities in the ground rules that may impede the identification of shuttle orbiter safety and mission critical items, and to ensure that ground rules allow these items to receive proper management visibility for risk assessment. Assessment objectives were followed during the performance of the assessment without being influenced by external considerations such as effects on budget, schedule, and documentation growth. Assessment personnel were employed who had a strong reliability background but no previous space shuttle FMEA/CIL experience to ensure an independent assessment would be achieved. The following observations were made: (1) not all essential items are in the CIL for management visibility; (2) ground rules omit FMEA/CIL coverage of items that perform critical functions; (3) essential items excluded from the CIL do not receive design justification; and (4) FMEAs/CILs are not updated in a timely manner. In addition to the above issues, a number of other issues were identified that correct FMEA/CIL preparation instruction omissions and clarify ambiguities. The assessment was successful in that many of the issues have significant safety implications.

  9. Sex Determination Using Inion-Opistocranium-Asterion (IOA) Triangle in Nigerians' Skulls.

    Science.gov (United States)

    Orish, C N; Didia, B C; Fawehinmi, H B

    2014-01-01

    Background. Determination of sex is an important concern to the forensic anthropologists as it is critical for individual identification. This study has investigated the existence of sexual dimorphism in the dimensions and the area of the IOA triangle. Methods. A total of 100 adult dry skulls, (78 males; 22 females) from departments of anatomy in Nigerian universities were used for this study. Automatic digital calliper was used for the measurement. Coefficient of variation, correlation, linear regression, percentiles, and sexual dimorphism ratio were computed from the IOA triangle measurements. The IOA triangle area was compared between sexes. Results. The male parameters were significantly (P triangle of 1938.88 mm(2) and 1305.68 mm(2) for male and female, respectively, was calculated. The left IOA indices were 46.42% and 37.40% in males and females, respectively, while the right IOA indices for males and females were 47.19% and 38.87%, respectively. Conclusion. The anthropometry of inion-opistocranium-asterion IOA triangle can be a guide in gender determination of unknown individuals.

  10. [Research progress on ecological footprint analysis].

    Science.gov (United States)

    Chen, Dongdong; Gao, Wangsheng; Chen, Yuanquan

    2006-10-01

    Ecological footprint (EF) model, as an indicator of sustainability, has received broad attention and wide use. With the development and refinement of the research work on EF theory and methodology, it appeared various methods which can be applied at different scales. Ecological footprint analysis has been combined with material flow analysis, life cycle assessment or input-output analysis, and especially, the newest progress in EF methods called allocating EF to final consumption categories with input-output analysis helps to develop a "standardized" EF. In this paper, the underlying causes of these methods were interpreted theoretically, and the research methods were classified into progress analysis and input-output analysis (IOA). In addition, the compound and component-based methods as well as IOA were introduced, with their respective features, application, and development progress discussed. A prospect on the development of EF in term of the tendency and application of EF methods in China and abroad was given, i. e. , the common framework should be built at the national and regional scales by using compound analysis, IOA and component-based analysis are expected to develop their application

  11. Regional mapping protocol, descriptive nomenclature and field database coding for iron oxide alkali-alteration (IOAA) ore systems and their IOCG, IOA and affiliated deposit types

    National Research Council Canada - National Science Library

    Corriveau, L; deToni, A; Montreuil, J F; Wahl, J; Guilmette, C; Guilmette, S; Champagne, M; Dourlet, S

    2015-01-01

    ...±apatite (IOA or Kiruna-type) deposits, 2) base-, precious-, specialised- and nuclear-metal ores or by-products within magnetite-, magnetite to hematite- and hematite-group IOCG deposits, and 3...

  12. The mitochondria-targeted imidazole substituted oleic acid 'TPP-IOA' affects mitochondrial bioenergetics and its protective efficacy in cells is influenced by cellular dependence on aerobic metabolism.

    Science.gov (United States)

    Maddalena, Lucas A; Ghelfi, Mikel; Atkinson, Jeffrey; Stuart, Jeffrey A

    2017-01-01

    A variety of mitochondria-targeted small molecules have been invented to manipulate mitochondrial redox activities and improve function in certain disease states. 3-Hydroxypropyl-triphenylphosphonium-conjugated imidazole-substituted oleic acid (TPP-IOA) was developed as a specific inhibitor of cytochrome c peroxidase activity that inhibits apoptosis by preventing cardiolipin oxidation and cytochrome c release to the cytosol. Here we evaluate the effects of TPP-IOA on oxidative phosphorylation in isolated mitochondria and on mitochondrial function in live cells. We demonstrate that, at concentrations similar to those required to achieve inhibition of cytochrome c peroxidase activity, TPP-IOA perturbs oxidative phosphorylation in isolated mitochondria. In live SH-SY5Y cells, TPP-IOA partially collapsed mitochondrial membrane potential, caused extensive fragmentation of the mitochondrial network, and decreased apparent mitochondrial abundance within 3h of exposure. Many cultured cell lines rely primarily on aerobic glycolysis, potentially making them less sensitive to small molecules disrupting oxidative phosphorylation. We therefore determined the anti-apoptotic efficacy of TPP-IOA in SH-SY5Y cells growing in glucose or in galactose, the latter of which increases reliance on oxidative phosphorylation for ATP supply. The anti-apoptotic activity of TPP-IOA that was observed in glucose media was not seen in galactose media. It therefore appears that, at concentrations required to inhibit cytochrome c peroxidase activity, TPP-IOA perturbs oxidative phosphorylation. In light of these data it is predicted that potential future therapeutic applications of TPP-IOA will be restricted to highly glycolytic cell types with limited reliance on oxidative phosphorylation. Copyright © 2016 Elsevier B.V. All rights reserved.

  13. Data Analysis and Assessment Center

    Data.gov (United States)

    Federal Laboratory Consortium — The DoD Supercomputing Resource Center (DSRC) Data Analysis and Assessment Center (DAAC) provides classified facilities to enhance customer interactions with the ARL...

  14. Assessment of cleaner electricity generation technologies for net CO{sub 2} mitigation in Thailand

    Energy Technology Data Exchange (ETDEWEB)

    Limmeechokchai, B.; Suksuntornsiri, P. [Thammasat University, Pathum Thani (Thailand)

    2007-02-15

    The choice of electricity generation technologies not only directly affects the amount of CO{sub 2} emission from the power sector, but also indirectly affects the economy-wide CO{sub 2} emission. It is because electricity is the basic requirement of economic sectors and final consumption within the economy. In Thailand, although the power development plan (PDP) has been planned for the committed capacity to meet the future electricity demand, there are some undecided electricity generation technologies that will be studied for technological options. The economy-wide CO{sub 2} mitigations between selecting cleaner power generation options instead of pulverized coal-thermal technology of the undecided capacity are assessed by energy input-output analysis (IOA). The decomposition of IOA presents the fuel-mix effect, input structural effect, and final demand effect by the change in technology of the undecided capacity. The cleaner technologies include biomass power generation, hydroelectricity and integrated gasification combined cycle (IGCC). Results of the analyses show that if the conventional pulverized coal technology is selected in the undecided capacity, the economy-wide CO{sub 2} emission would be increased from 223 million ton in 2006 to 406 million ton in 2016. Renewable technology presents better mitigation option for replacement of conventional pulverized coal technology than the cleaner coal technology. The major contributor of CO{sub 2} mitigation in cleaner coal technology is the fuel mix effect due to higher conversion efficiency.

  15. Implementation of a Posted Schedule to Increase Class-Wide Interobserver Agreement Assessment

    Science.gov (United States)

    Doucette, Stefanie; DiGennaro Reed, Florence D.; Reed, Derek D.; Maguire, Helena; Marquardt, Heidi

    2012-01-01

    The present study investigated the impact of an antecedent intervention in the form of a daily posted schedule on the interobserver agreement (IOA) assessment of educational goals implemented within a classroom at a private school serving individuals with disabilities. During baseline, the percentage of academic goals with interobserver agreement…

  16. Change point analysis and assessment

    DEFF Research Database (Denmark)

    Müller, Sabine; Neergaard, Helle; Ulhøi, John Parm

    2011-01-01

    The aim of this article is to develop an analytical framework for studying processes such as continuous innovation and business development in high-tech SME clusters that transcends the traditional qualitative-quantitative divide. It integrates four existing and well-recognized approaches to stud...... to studying events, processes and change, mamely change-point analysis, event-history analysis, critical-incident technique and sequence analysis....

  17. Multifractal analysis for nutritional assessment.

    Directory of Open Access Journals (Sweden)

    Youngja Park

    Full Text Available The concept of multifractality is currently used to describe self-similar and complex scaling properties observed in numerous biological signals. Fractals are geometric objects or dynamic variations which exhibit some degree of similarity (irregularity to the original object in a wide range of scales. This approach determines irregularity of biologic signal as an indicator of adaptability, the capability to respond to unpredictable stress, and health. In the present work, we propose the application of multifractal analysis of wavelet-transformed proton nuclear magnetic resonance ((1H NMR spectra of plasma to determine nutritional insufficiency. For validation of this method on (1H NMR signal of human plasma, standard deviation from classical statistical approach and Hurst exponent (H, left slope and partition function from multifractal analysis were extracted from (1H NMR spectra to test whether multifractal indices could discriminate healthy subjects from unhealthy, intensive care unit patients. After validation, the multifractal approach was applied to spectra of plasma from a modified crossover study of sulfur amino acid insufficiency and tested for associations with blood lipids. The results showed that standard deviation and H, but not left slope, were significantly different for sulfur amino acid sufficiency and insufficiency. Quadratic discriminant analysis of H, left slope and the partition function showed 78% overall classification accuracy according to sulfur amino acid status. Triglycerides and apolipoprotein C3 were significantly correlated with a multifractal model containing H, left slope, and standard deviation, and cholesterol and high-sensitivity C-reactive protein were significantly correlated to H. In conclusion, multifractal analysis of (1H NMR spectra provides a new approach to characterize nutritional status.

  18. Assessing Analysis and Reasoning in Bioethics

    Science.gov (United States)

    Pearce, Roger S.

    2008-01-01

    Developing critical thinking is a perceived weakness in current education. Analysis and reasoning are core skills in bioethics making bioethics a useful vehicle to address this weakness. Assessment is widely considered to be the most influential factor on learning (Brown and Glasner, 1999) and this piece describes how analysis and reasoning in…

  19. Prior Learning Assessment with Latent Semantic Analysis

    NARCIS (Netherlands)

    Kalz, Marco; Van Bruggen, Jan; Giesbers, Bas; Koper, Rob

    2007-01-01

    Kalz, M., Van Bruggen, J., Giesbers, B., & Koper, R. (2007). Prior Learning Assessment with Latent Semantic Analysis. In F. Wild, M. Kalz, J. Van Bruggen & R. Koper (Eds.). Proceedings of the First European Workshop on Latent Semantic Analysis in Technology Enhanced Learning (pp. 24-25). Heerlen,

  20. Prior Learning Assessment with Latent Semantic Analysis

    OpenAIRE

    Kalz, Marco; van Bruggen, Jan; Giesbers, Bas; Koper, Rob

    2007-01-01

    Kalz, M., Van Bruggen, J., Giesbers, B., & Koper, R. (2007). Prior Learning Assessment with Latent Semantic Analysis. In F. Wild, M. Kalz, J. Van Bruggen & R. Koper (Eds.). Proceedings of the First European Workshop on Latent Semantic Analysis in Technology Enhanced Learning (pp. 24-25). Heerlen, The Netherlands: Open University of the Netherlands.

  1. Interobserver variability in target volume delineation of hepatocellular carcinoma. An analysis of the working group ''Stereotactic Radiotherapy'' of the German Society for Radiation Oncology (DEGRO)

    Energy Technology Data Exchange (ETDEWEB)

    Gkika, E.; Kirste, S. [Medical Center - University of Freiburg, Department of Radiation Oncology, Freiburg im Breisgau (Germany); Tanadini-Lang, S.; Andratschke, N.; Guckenberger, M. [University Hospital Zuerich, Department of Radiation Oncology, Zurich (Switzerland); Holzner, P.A.; Neeff, H.P. [Medical Center - University of Freiburg, Department of Visceral Surgery, Freiburg (Germany); Rischke, H.C. [Medical Center - University of Freiburg, Department of Radiation Oncology, Freiburg im Breisgau (Germany); Medical Center - University of Freiburg, Department of Nuclear Medicine, Freiburg (Germany); Reese, T. [University Hospital Halle-Wittenberg, Department of Radiation Oncology, Halle-Wittenberg (Germany); Lohaus, F. [Technische Universitaet Dresden, Department of Radiation Oncology, Faculty of Medicine and University Hospital Carl Gustav Carus, Dresden (Germany); German Cancer Research Center (DKFZ), Heidelberg (Germany); Technische Universitaet Dresden, OncoRay - National Center for Radiation Research in Oncology, Faculty of Medicine and University Hospital Carl Gustav Carus, Dresden (Germany); National Center for Tumor Diseases (NCT), Dresden (Germany); German Cancer Consortium (DKTK), Dresden (Germany); Duma, M.N. [Helmholtz Zentrum Munich, Institute of Innovative Radiotherapy, Department of Radiation Sciences, Munich (Germany); TU Munich, Department of Radiation Oncology, Klinikum Rechts der Isar, Munich (Germany); Dieckmann, K. [Medical University Vienna, Department of Radiation Oncology, General Hospital Vienna, Vienna (Austria); Semrau, R. [University Hospital of Cologne, Department of Radiation Oncology, Cologne (Germany); Stockinger, M. [University Hospital Mainz, Department of Radiation Oncology, Mainz (Germany); Imhoff, D. [University Hospital Frankfurt, Department of Radiation Oncology, Frankfurt (Germany); Saphir Radiosurgery Center, Frankfurt (Germany); Kremers, N. [Medical Center - University of Freiburg, Department of Radiology, Freiburg (Germany); Haefner, M.F. [University Hospital Heidelberg, Department of Radiation Oncology, Heidelberg (Germany); Nestle, U.; Grosu, A.L.; Brunner, T.B. [Medical Center - University of Freiburg, Department of Radiation Oncology, Freiburg im Breisgau (Germany); University of Freiburg, Faculty of Medicine, Freiburg (Germany); German Cancer Consortium (DKTK), Freiburg (Germany); German Cancer Research Center (DKFZ), Heidelberg (Germany)

    2017-10-15

    Definition of gross tumor volume (GTV) in hepatocellular carcinoma (HCC) requires dedicated imaging in multiple contrast medium phases. The aim of this study was to evaluate the interobserver agreement (IOA) in gross tumor delineation of HCC in a multicenter panel. The analysis was performed within the ''Stereotactic Radiotherapy'' working group of the German Society for Radiation Oncology (DEGRO). The GTVs of three anonymized HCC cases were delineated by 16 physicians from nine centers using multiphasic CT scans. In the first case the tumor was well defined. The second patient had multifocal HCC (one conglomerate and one peripheral tumor) and was previously treated with transarterial chemoembolization (TACE). The peripheral lesion was adjacent to the previous TACE site. The last patient had an extensive HCC with a portal vein thrombosis (PVT) and an inhomogeneous liver parenchyma due to cirrhosis. The IOA was evaluated according to Landis and Koch. The IOA for the first case was excellent (kappa: 0.85); for the second case moderate (kappa: 0.48) for the peripheral tumor and substantial (kappa: 0.73) for the conglomerate. In the case of the peripheral tumor the inconsistency is most likely explained by the necrotic tumor cavity after TACE caudal to the viable tumor. In the last case the IOA was fair, with a kappa of 0.34, with significant heterogeneity concerning the borders of the tumor and the PVT. The IOA was very good among the cases were the tumor was well defined. In complex cases, where the tumor did not show the typical characteristics, or in cases with Lipiodol (Guerbet, Paris, France) deposits, IOA agreement was compromised. (orig.) [German] Die Definition des makroskopischen Tumorvolumens (GTV) bei hepatozellulaeren Karzinomen (HCC) erfordert eine dezidierte Bildgebung in mehreren Kontrastmittelphasen. Ziel dieser Studie war es, die Interobservervariabilitaet (IOA) bei der Konturierung von HCC-Laesionen durch ein multizentrisches

  2. Asbestos Workshop: Sampling, Analysis, and Risk Assessment

    Science.gov (United States)

    2012-03-01

    1EMDQ March 2012 ASBESTOS WORKSHOP: SAMPLING, ANALYSIS, AND RISK ASSESSMENT Paul Black, PhD, Neptune and Company Ralph Perona, DABT, Neptune and...NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Neptune and Company, Inc,1435 Garrison Street, Suite 110,Denver,CO...8-98) Prescribed by ANSI Std Z39-18 2 Presentation Objective Provide an overview of asbestos-related risk assessment: • Focus on risk from

  3. Sensitivity analysis in life cycle assessment

    NARCIS (Netherlands)

    Groen, E.A.; Heijungs, R.; Bokkers, E.A.M.; Boer, de I.J.M.

    2014-01-01

    Life cycle assessments require many input parameters and many of these parameters are uncertain; therefore, a sensitivity analysis is an essential part of the final interpretation. The aim of this study is to compare seven sensitivity methods applied to three types of case stud-ies. Two

  4. Phytochemical analysis and assessment of bioethanol production ...

    African Journals Online (AJOL)

    In this research, the phytochemical and bioethanol content of Cymbopogon schoenanthus was assessed under three different pH. The phytochemical screening and proximate analysis was carried out according to standard qualitative and quantitative methods. Where alkaloids, balsam, flavonoids, glycosides, saponin, ...

  5. Essay Assessment with Latent Semantic Analysis

    Science.gov (United States)

    Miller, Tristan

    2003-01-01

    Latent semantic analysis (LSA) is an automated, statistical technique for comparing the semantic similarity of words or documents. In this article, I examine the application of LSA to automated essay scoring. I compare LSA methods to earlier statistical methods for assessing essay quality, and critically review contemporary essay-scoring systems…

  6. Surface water quality assessment using factor analysis

    African Journals Online (AJOL)

    2006-01-16

    Jan 16, 2006 ... surface water by rain and stormwater. On the other hand, run- off water increases pollutant concentrations, thereby decreases quality. To assess the water quality of the Buyuk Menderes. River under high-flow conditions, factor analysis was applied to data sets obtained from 21 monitoring stations between ...

  7. Quantitative wear particle analysis for osteoarthritis assessment.

    Science.gov (United States)

    Guo, Meizhai; Lord, Megan S; Peng, Zhongxiao

    2017-12-01

    Osteoarthritis is a degenerative joint disease that affects millions of people worldwide. The aims of this study were (1) to quantitatively characterise the boundary and surface features of wear particles present in the synovial fluid of patients, (2) to select key numerical parameters that describe distinctive particle features and enable osteoarthritis assessment and (3) to develop a model to assess osteoarthritis conditions using comprehensive wear debris information. Discriminant analysis was used to statistically group particles based on differences in their numerical parameters. The analysis methods agreed with the clinical osteoarthritis grades in 63%, 50% and 61% of particles for no osteoarthritis, mild osteoarthritis and severe osteoarthritis, respectively. This study has revealed particle features specific to different osteoarthritis grades and provided further understanding of the cartilage degradation process through wear particle analysis - the technique that has the potential to be developed as an objective and minimally invasive method for osteoarthritis diagnosis.

  8. Supplemental Hazard Analysis and Risk Assessment - Hydrotreater

    Energy Technology Data Exchange (ETDEWEB)

    Lowry, Peter P. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States); Wagner, Katie A. [Pacific Northwest National Lab. (PNNL), Richland, WA (United States)

    2015-04-01

    A supplemental hazard analysis was conducted and quantitative risk assessment performed in response to an independent review comment received by the Pacific Northwest National Laboratory (PNNL) from the U.S. Department of Energy Pacific Northwest Field Office (PNSO) against the Hydrotreater/Distillation Column Hazard Analysis Report issued in April 2013. The supplemental analysis used the hazardous conditions documented by the previous April 2013 report as a basis. The conditions were screened and grouped for the purpose of identifying whether additional prudent, practical hazard controls could be identified, using a quantitative risk evaluation to assess the adequacy of the controls and establish a lower level of concern for the likelihood of potential serious accidents. Calculations were performed to support conclusions where necessary.

  9. Quality Assessment of Urinary Stone Analysis

    DEFF Research Database (Denmark)

    Siener, Roswitha; Buchholz, Noor; Daudon, Michel

    2016-01-01

    , between 2010 and 2014. Each participant received the same blinded test samples for stone analysis. A total of 24 samples, comprising pure substances and mixtures of two or three components, were analysed. The evaluation of the quality of the laboratory in the present study was based on the attainment...... and chemical analysis. The aim of the present study was to assess the quality of urinary stone analysis of laboratories in Europe. Nine laboratories from eight European countries participated in six quality control surveys for urinary calculi analyses of the Reference Institute for Bioanalytics, Bonn, Germany...... of 75% of the maximum total points, i.e. 99 points. The methods of stone analysis used were infrared spectroscopy (n = 7), chemical analysis (n = 1) and X-ray diffraction (n = 1). In the present study only 56% of the laboratories, four using infrared spectroscopy and one using X-ray diffraction...

  10. Sensitivity analysis in quantitative microbial risk assessment.

    Science.gov (United States)

    Zwieterin, M H; van Gerwen, S J

    2000-07-15

    The occurrence of foodborne disease remains a widespread problem in both the developing and the developed world. A systematic and quantitative evaluation of food safety is important to control the risk of foodborne diseases. World-wide, many initiatives are being taken to develop quantitative risk analysis. However, the quantitative evaluation of food safety in all its aspects is very complex, especially since in many cases specific parameter values are not available. Often many variables have large statistical variability while the quantitative effect of various phenomena is unknown. Therefore, sensitivity analysis can be a useful tool to determine the main risk-determining phenomena, as well as the aspects that mainly determine the inaccuracy in the risk estimate. This paper presents three stages of sensitivity analysis. First, deterministic analysis selects the most relevant determinants for risk. Overlooking of exceptional, but relevant cases is prevented by a second, worst-case analysis. This analysis finds relevant process steps in worst-case situations, and shows the relevance of variations of factors for risk. The third, stochastic analysis, studies the effects of variations of factors for the variability of risk estimates. Care must be taken that the assumptions made as well as the results are clearly communicated. Stochastic risk estimates are, like deterministic ones, just as good (or bad) as the available data, and the stochastic analysis must not be used to mask lack of information. Sensitivity analysis is a valuable tool in quantitative risk assessment by determining critical aspects and effects of variations.

  11. Interobserver agreement in the assessment of pulmonary infiltrates on chest radiography in community-acquired pneumonia; Detektion pneumonischer Infiltrate bei ambulant erworbener Pneumonie: Uebereinstimmung in der Befundung der Roentgen-Thoraxaufnahme

    Energy Technology Data Exchange (ETDEWEB)

    Pauls, S.; Billich, C.; Boll, D.; Aschoff, A.J. [Diagnostische und Interventionelle Radiologie, Universitaetskliniken Ulm (Germany); Krueger, S. [Medizinische Klinik I, Universitaetskliniken RWTH Aachen (Germany); Richter, K.; Marre, R.; Gonschior, S. [Mikrobiologie und Hygiene, Universitaetskliniken Ulm (Germany); Muche, R. [Inst. fuer Biometrie, Univ. Ulm (Germany); Welte, T. [Abt. fuer Pneumologie, Medizinische Hochschule Hannover (Germany); Schumann, C. [Medizinische Klinik II, Universitaetskliniken Ulm (Germany); Suttorp, N. [Abt. Innere Medizin, Charite Universitaetsmedizin Berlin (Germany)

    2007-11-15

    Purpose: To assess interobserver agreement (IOA) in the diagnosis of pulmonary infiltrates on chest X-rays for patients with community-acquired pneumonia (CAP). Materials and methods: From 7/2002 to 12/2005, 806 adults with CAP were included in the multicenter study 'CAPNETZ' (7 hospitals). Inclusion criteria were clinical signs of pneumonia and pulmonary opacification on chest X-rays. Each X-ray was reevaluated by two radiologists from the university hospital in consensus reading against the interpreter at the referring hospital in regard to: presence of infiltrate (yes/no/equivocal), transparency ({<=}/> 50%), localization, and pattern of infiltrates (alveolar/interstitial). The following parameters were documented: digital or film radiography, hospitalization, fever, findings of auscultation, microbiological findings. Results: The overall IOA concerning the detection of infiltrates was 77.7% (n = 626; Cl 0.75 - 0.81), the infiltrates were not verified in 16.4% (n = 132) by the referring radiologist with equivocal findings in 5.9% (n = 48). The IOA of the different clinical centers varied between 63.2% (n = 38, Cl 0.48 - 0.78) and 92.3% (n = 65, Cl 0.86 - 0.99). The IOA for the diagnosis of infiltrates was significantly higher for inpatients with 82.6% (n = 546; Cl 0.80-0.85) than for outpatients with 55.2% (n = 80; Cl 0.47 - 0.63), p < 0.0001. The IOA of infiltrates with a transparency > 50% was 95.1% (n = 215; Cl 0.92 - 0.98) versus 80.4% (n = 403; Cl 0.77 - 0.84) for infiltrates with a transparency > 50% (p < 0.0001). In patients with positive auscultation, the IOA was higher (p = 0,034). Chest X-rays of patients with antibiotic therapy or an alveolar infiltrate showed more equivocal findings compared to patients without these features. Conclusion: There is considerable interobserver variability in the diagnosis of pulmonary infiltrates on chest radiographs. The IOA is higher in more opaque infiltrates, positive auscultation and inpatients. (orig.)

  12. Multicriteria analysis in hazards assessment in Libya

    Science.gov (United States)

    Zeleňáková, Martina; Gargar, Ibrahim; Purcz, Pavol

    2012-11-01

    Environmental hazards (natural and man-made) have always constituted problem in many developing and developed countries. Many applications proved that these problems could be solved through planning studies and detailed information about these prone areas. Determining time and location and size of the problem are important for decision makers for planning and management activities. It is important to know the risk represented by those hazards and take actions to protect against them. Multicriteria analysis methods - Analytic hierarchy process, Pairwise comparison, Ranking method are used to analyse which is the most dangerous hazard facing Libya country. The multicriteria analysis ends with a more or less stable ranking of the given alternatives and hence a recommendation as to which alternative(s) problems should be preferred. Regarding our problem of environmental risk assessment, the result will be a ranking or categorisation of hazards with regard to their risk level.

  13. HANFORD SAFETY ANALYSIS & RISK ASSESSMENT HANDBOOK (SARAH)

    Energy Technology Data Exchange (ETDEWEB)

    EVANS, C B

    2004-12-21

    The purpose of the Hanford Safety Analysis and Risk Assessment Handbook (SARAH) is to support the development of safety basis documentation for Hazard Category 2 and 3 (HC-2 and 3) U.S. Department of Energy (DOE) nuclear facilities to meet the requirements of 10 CFR 830, ''Nuclear Safety Management''. Subpart B, ''Safety Basis Requirements.'' Consistent with DOE-STD-3009-94, Change Notice 2, ''Preparation Guide for U.S. Department of Energy Nonreactor Nuclear Facility Documented Safety Analyses'' (STD-3009), and DOE-STD-3011-2002, ''Guidance for Preparation of Basis for Interim Operation (BIO) Documents'' (STD-3011), the Hanford SARAH describes methodology for performing a safety analysis leading to development of a Documented Safety Analysis (DSA) and derivation of Technical Safety Requirements (TSR), and provides the information necessary to ensure a consistently rigorous approach that meets DOE expectations. The DSA and TSR documents, together with the DOE-issued Safety Evaluation Report (SER), are the basic components of facility safety basis documentation. For HC-2 or 3 nuclear facilities in long-term surveillance and maintenance (S&M), for decommissioning activities, where source term has been eliminated to the point that only low-level, residual fixed contamination is present, or for environmental remediation activities outside of a facility structure, DOE-STD-1120-98, ''Integration of Environment, Safety, and Health into Facility Disposition Activities'' (STD-1120), may serve as the basis for the DSA. HC-2 and 3 environmental remediation sites also are subject to the hazard analysis methodologies of this standard.

  14. Topological image texture analysis for quality assessment

    Science.gov (United States)

    Asaad, Aras T.; Rashid, Rasber Dh.; Jassim, Sabah A.

    2017-05-01

    Image quality is a major factor influencing pattern recognition accuracy and help detect image tampering for forensics. We are concerned with investigating topological image texture analysis techniques to assess different type of degradation. We use Local Binary Pattern (LBP) as a texture feature descriptor. For any image construct simplicial complexes for selected groups of uniform LBP bins and calculate persistent homology invariants (e.g. number of connected components). We investigated image quality discriminating characteristics of these simplicial complexes by computing these models for a large dataset of face images that are affected by the presence of shadows as a result of variation in illumination conditions. Our tests demonstrate that for specific uniform LBP patterns, the number of connected component not only distinguish between different levels of shadow effects but also help detect the infected regions as well.

  15. Acoustic analysis assessment in speech pathology detection

    Directory of Open Access Journals (Sweden)

    Panek Daria

    2015-09-01

    Full Text Available Automatic detection of voice pathologies enables non-invasive, low cost and objective assessments of the presence of disorders, as well as accelerating and improving the process of diagnosis and clinical treatment given to patients. In this work, a vector made up of 28 acoustic parameters is evaluated using principal component analysis (PCA, kernel principal component analysis (kPCA and an auto-associative neural network (NLPCA in four kinds of pathology detection (hyperfunctional dysphonia, functional dysphonia, laryngitis, vocal cord paralysis using the a, i and u vowels, spoken at a high, low and normal pitch. The results indicate that the kPCA and NLPCA methods can be considered a step towards pathology detection of the vocal folds. The results show that such an approach provides acceptable results for this purpose, with the best efficiency levels of around 100%. The study brings the most commonly used approaches to speech signal processing together and leads to a comparison of the machine learning methods determining the health status of the patient

  16. Technical Overview of Ecological Risk Assessment - Analysis Phase: Exposure Characterization

    Science.gov (United States)

    Exposure Characterization is the second major component of the analysis phase of a risk assessment. For a pesticide risk assessment, the exposure characterization describes the potential or actual contact of a pesticide with a plant, animal, or media.

  17. An Analysis of Assessment of Peacekeeping Operations

    National Research Council Canada - National Science Library

    Unsal Sigri; ; Ufuk Bas¸ar

    2014-01-01

    This article proposes a model for assessment of peacekeeping operations. Peacekeeping operations began in 1948, under the authority of the United Nations in order to provide and sustain a peaceful environment throughout the world...

  18. Landscape analysisAssessing countries' readiness to scale up ...

    African Journals Online (AJOL)

    The purpose of this paper is to describe the use of the Landscape Analysis to assess strengths and weaknesses in combating malnutrition in Sub-Saharan Africa. The Landscape Analysis is an inter-agency initiative to assess gaps and constraints and to identify opportunities for effective nutrition actions in order to ...

  19. Competency Discourses: An Analysis of Secondary and Tertiary Assessments

    Science.gov (United States)

    McNaughton, Susan M.

    2013-01-01

    Analysis of a sample of assessment tasks used in New Zealand first-year clinical and final-year secondary school Biology and Chemistry National Certificate of Educational Achievement summative assessments was conducted to assess whether similarities and differences existed in secondary and tertiary competency discourses. Findings suggested…

  20. Comparative analysis of seismic risk assessment methodology

    Energy Technology Data Exchange (ETDEWEB)

    Park, Chang Kyoo; Kim, Tae Woon; Hwang, Mi Jung [Korea Atomic Energy Research Institute, Taejon (Korea, Republic of)

    1994-07-01

    SRA methodologies are separated into SPSA and SMM. SPSA methodology that has been widely used for seismic risk analysis has two kinds of methodologies such as Zion method and SSMRP method. SPSA methodology is suitable to interfacing with the analysis of internal event. However, the results of SPSA have uncertainties because of uncertainties in seismic hazard analysis and subjective judgement. Zion method specially developed for commercial use is less expensive and less time consuming but more uncertain than SSMRP method, since the former performs the fragility analysis less in detail than the latter. SMM is impossible to interface with the analysis of internal event but the uncertainties that are occurred during seismic hazard analysis is reduced because of the screening using RLE (review level earthquake). Therefore, if SPSA-based SMM methodology is chosen to be developed, the results of SRA will be more reliable and it requires low costs and time. In addition, the new methodology will require the development of a new evaluating code for SRA. (Author) 26 refs., 25 figs., 16 tabs.

  1. Assessment and uncertainty analysis of groundwater risk.

    Science.gov (United States)

    Li, Fawen; Zhu, Jingzhao; Deng, Xiyuan; Zhao, Yong; Li, Shaofei

    2018-01-01

    Groundwater with relatively stable quantity and quality is commonly used by human being. However, as the over-mining of groundwater, problems such as groundwater funnel, land subsidence and salt water intrusion have emerged. In order to avoid further deterioration of hydrogeological problems in over-mining regions, it is necessary to conduct the assessment of groundwater risk. In this paper, risks of shallow and deep groundwater in the water intake area of the South-to-North Water Transfer Project in Tianjin, China, were evaluated. Firstly, two sets of four-level evaluation index system were constructed based on the different characteristics of shallow and deep groundwater. Secondly, based on the normalized factor values and the synthetic weights, the risk values of shallow and deep groundwater were calculated. Lastly, the uncertainty of groundwater risk assessment was analyzed by indicator kriging method. The results meet the decision maker's demand for risk information, and overcome previous risk assessment results expressed in the form of deterministic point estimations, which ignore the uncertainty of risk assessment. Copyright © 2017 Elsevier Inc. All rights reserved.

  2. phytochemical analysis and assessment of bioethanol production ...

    African Journals Online (AJOL)

    userpc

    In this research, the phytochemical and bioethanol content of Cymbopogon schoenanthus was assessed under three different pH. ... of the bioethanol produced shows the presence of Ethyl alcohol (bioethanol) in all the treatments at different pH. ... food security, water use efficiency of the biomass, land use efficiency etc. To.

  3. Data Analysis and Next Generation Assessments

    Science.gov (United States)

    Pon, Kathy

    2013-01-01

    For the last decade, much of the work of California school administrators has been shaped by the accountability of the No Child Left Behind Act. Now as they stand at the precipice of Common Core Standards and next generation assessments, it is important to reflect on the proficiency educators have attained in using data to improve instruction and…

  4. Surface water quality assessment using factor analysis

    African Journals Online (AJOL)

    2006-01-16

    Jan 16, 2006 ... In this study, the factor analysis technique is applied to surface water quality data sets obtained from the Buyuk Menderes. River Basin, Turkey, during two different hydrological periods. Results show that the indices which changed the quality of water in two seasons and locations differed. During low-flow ...

  5. Assessment and Planning Using Portfolio Analysis

    Science.gov (United States)

    Roberts, Laura B.

    2010-01-01

    Portfolio analysis is a simple yet powerful management tool. Programs and activities are placed on a grid with mission along one axis and financial return on the other. The four boxes of the grid (low mission, low return; high mission, low return; high return, low mission; high return, high mission) help managers identify which programs might be…

  6. SEXUAL MINORITIES: ASSESSMENT AND ANALYSIS PROBLEMS

    Directory of Open Access Journals (Sweden)

    Mikhail V. Karmanov

    2014-01-01

    Full Text Available On a number of the objective and subjective reasons statistical information onprevalence of sexual minorities in societycauses very inconsistent estimates. Thus activity of sexual minorities predeterminesin recent years need of deep judgmentand the analysis of behavior and activityof the persons having nonconventionalsexual orientation.

  7. Material Analysis for a Fire Assessment.

    Energy Technology Data Exchange (ETDEWEB)

    Brown, Alexander; Nemer, Martin B.

    2014-08-01

    This report consolidates technical information on several materials and material classes for a fire assessment. The materials include three polymeric materials, wood, and hydraulic oil. The polymers are polystyrene, polyurethane, and melamine- formaldehyde foams. Samples of two of the specific materials were tested for their behavior in a fire - like environment. Test data and the methods used to test the materials are presented. Much of the remaining data are taken from a literature survey. This report serves as a reference source of properties necessary to predict the behavior of these materials in a fire.

  8. Environmental risk assessment in GMO analysis.

    Science.gov (United States)

    Pirondini, Andrea; Marmiroli, Nelson

    2010-01-01

    Genetically modified or engineered organisms (GMOs, GEOs) are utilised in agriculture, expressing traits of interest, such as insect or herbicide resistance. Soybean, maize, cotton and oilseed rape are the GM crops with the largest acreage in the world. The distribution of GM acreage in the different countries is related with the different positions concerning labelling of GMO products: based on the principle of substantial equivalence, or rather based on the precautionary principle. The paper provides an overview on how the risks associated with release of GMO in the environments can be analysed and predicted, in view of a possible coexistence of GM and non-GM organisms in agriculture.Risk assessment procedures, both qualitative and quantitative, are compared in the context of application to GMOs considering also legislation requirements (Directive 2001/18/EC). Criteria and measurable properties to assess harm for human health and environmental safety are listed, and the possible consequences are evaluated in terms of significance.Finally, a mapping of the possible risks deriving from GMO release is reported, focusing on gene transfer to related species, horizontal gene transfer, direct and indirect effects on non target organisms, development of resistance in target organisms, and effects on biodiversity.

  9. Enhancing e-waste estimates: improving data quality by multivariate Input-Output Analysis.

    Science.gov (United States)

    Wang, Feng; Huisman, Jaco; Stevels, Ab; Baldé, Cornelis Peter

    2013-11-01

    Waste electrical and electronic equipment (or e-waste) is one of the fastest growing waste streams, which encompasses a wide and increasing spectrum of products. Accurate estimation of e-waste generation is difficult, mainly due to lack of high quality data referred to market and socio-economic dynamics. This paper addresses how to enhance e-waste estimates by providing techniques to increase data quality. An advanced, flexible and multivariate Input-Output Analysis (IOA) method is proposed. It links all three pillars in IOA (product sales, stock and lifespan profiles) to construct mathematical relationships between various data points. By applying this method, the data consolidation steps can generate more accurate time-series datasets from available data pool. This can consequently increase the reliability of e-waste estimates compared to the approach without data processing. A case study in the Netherlands is used to apply the advanced IOA model. As a result, for the first time ever, complete datasets of all three variables for estimating all types of e-waste have been obtained. The result of this study also demonstrates significant disparity between various estimation models, arising from the use of data under different conditions. It shows the importance of applying multivariate approach and multiple sources to improve data quality for modelling, specifically using appropriate time-varying lifespan parameters. Following the case study, a roadmap with a procedural guideline is provided to enhance e-waste estimation studies. Copyright © 2013 Elsevier Ltd. All rights reserved.

  10. Cooperative analysis expert situation assessment research

    Science.gov (United States)

    Mccown, Michael G.

    1987-01-01

    For the past few decades, Rome Air Development Center (RADC) has been conducting research in Artificial Intelligence (AI). When the recent advances in hardware technology made many AI techniques practical, the Intelligence and Reconnaissance Directorate of RADC initiated an applications program entitled Knowledge Based Intelligence Systems (KBIS). The goal of the program is the development of a generic Intelligent Analyst System, an open machine with the framework for intelligence analysis, natural language processing, and man-machine interface techniques, needing only the specific problem domain knowledge to be operationally useful. The development of KBIS is described.

  11. Analysis and Assessment of Computer-Supported Collaborative Learning Conversations

    NARCIS (Netherlands)

    Trausan-Matu, Stefan

    2008-01-01

    Trausan-Matu, S. (2008). Analysis and Assessment of Computer-Supported Collaborative Learning Conversations. Workshop presentation at the symposium Learning networks for professional. November, 14, 2008, Heerlen, Nederland: Open Universiteit Nederland.

  12. PIE Nacelle Flow Analysis and TCA Inlet Flow Quality Assessment

    Science.gov (United States)

    Shieh, C. F.; Arslan, Alan; Sundaran, P.; Kim, Suk; Won, Mark J.

    1999-01-01

    This presentation includes three topics: (1) Analysis of isolated boattail drag; (2) Computation of Technology Concept Airplane (TCA)-installed nacelle effects on aerodynamic performance; and (3) Assessment of TCA inlet flow quality.

  13. Accuracy Assessment and Analysis for GPT2

    Directory of Open Access Journals (Sweden)

    YAO Yibin

    2015-07-01

    Full Text Available GPT(global pressure and temperature is a global empirical model usually used to provide temperature and pressure for the determination of tropospheric delay, there are some weakness to GPT, these have been improved with a new empirical model named GPT2, which not only improves the accuracy of temperature and pressure, but also provides specific humidity, water vapor pressure, mapping function coefficients and other tropospheric parameters, and no accuracy analysis of GPT2 has been made until now. In this paper high-precision meteorological data from ECWMF and NOAA were used to test and analyze the accuracy of temperature, pressure and water vapor pressure expressed by GPT2, testing results show that the mean Bias of temperature is -0.59℃, average RMS is 3.82℃; absolute value of average Bias of pressure and water vapor pressure are less than 1 mb, GPT2 pressure has average RMS of 7 mb, and water vapor pressure no more than 3 mb, accuracy is different in different latitudes, all of them have obvious seasonality. In conclusion, GPT2 model has high accuracy and stability on global scale.

  14. Trace Assessment for BWR ATWS Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Cheng, L.Y.; Diamond, D.; Arantxa Cuadra, Gilad Raitses, Arnold Aronson

    2010-04-22

    A TRACE/PARCS input model has been developed in order to be able to analyze anticipated transients without scram (ATWS) in a boiling water reactor. The model is based on one developed previously for the Browns Ferry reactor for doing loss-of-coolant accident analysis. This model was updated by adding the control systems needed for ATWS and a core model using PARCS. The control systems were based on models previously developed for the TRAC-B code. The PARCS model is based on information (e.g., exposure and moderator density (void) history distributions) obtained from General Electric Hitachi and cross sections for GE14 fuel obtained from an independent source. The model is able to calculate an ATWS, initiated by the closure of main steam isolation valves, with recirculation pump trip, water level control, injection of borated water from the standby liquid control system and actuation of the automatic depres-surization system. The model is not considered complete and recommendations are made on how it should be improved.

  15. Overcoming barriers to integrating economic analysis into risk assessment.

    Science.gov (United States)

    Hoffmann, Sandra

    2011-09-01

    Regulatory risk analysis is designed to provide decisionmakers with a clearer understanding of how policies are likely to affect risk. The systems that produce risk are biological, physical, and social and economic. As a result, risk analysis is an inherently interdisciplinary task. Yet in practice, risk analysis has been interdisciplinary in only limited ways. Risk analysis could provide more accurate assessments of risk if there were better integration of economics and other social sciences into risk assessment itself. This essay examines how discussions about risk analysis policy have influenced the roles of various disciplines in risk analysis. It explores ways in which integrated bio/physical-economic modeling could contribute to more accurate assessments of risk. It reviews examples of the kind of integrated economics-bio/physical modeling that could be used to enhance risk assessment. The essay ends with a discussion of institutional barriers to greater integration of economic modeling into risk assessment and provides suggestions on how these might be overcome. © 2011 Society for Risk Analysis.

  16. Adversarial safety analysis: borrowing the methods of security vulnerability assessments.

    Science.gov (United States)

    Johnston, Roger G

    2004-01-01

    Safety and security share numerous attributes. The author, who heads the (Security) Vulnerability Assessment Team at Los Alamos National Laboratory, therefore argues that techniques used to optimize security might be useful for optimizing safety. There are three main ways to attempt to improve security-security surveys, risk assessment (or "design basis threat"), and vulnerability assessments. The latter is usually the most effective. Vulnerability assessment techniques used to improve security can be applied to safety analysis--even though safety is not ordinarily viewed as having malicious adversaries (other than hazards involving deliberate sabotage). Thinking like a malicious adversary can nevertheless have benefits in identifying safety vulnerabilities. The attributes of an effective safety vulnerability assessment are discussed, and recommendations are offered for how such an adversarial assessment might work. A safety vulnerability assessment can potentially provide new insights, a fresh and vivid perspective on safety hazards, and increased safety awareness.

  17. Uncertainty analysis in integrated assessment: the users' perspective

    NARCIS (Netherlands)

    Gabbert, S.G.M.; Ittersum, van M.K.; Kroeze, C.; Stalpers, S.I.P.; Ewert, F.; Alkan Olsson, J.

    2010-01-01

    Integrated Assessment (IA) models aim at providing information- and decision-support to complex problems. This paper argues that uncertainty analysis in IA models should be user-driven in order to strengthen science–policy interaction. We suggest an approach to uncertainty analysis that starts with

  18. Using Latent Semantic Analysis To Assess Knowledge: Some Technical Considerations.

    Science.gov (United States)

    Rehder, Bob; Schreiner, M. E.; Wolfe, Michael B. W.; Laham, Darrell; Kintsch, Walter; Landauer, Thomas K

    1998-01-01

    Provides a technical analysis of the factors involved in the ability of latent semantic analysis to assess student knowledge (grading essays and matching students with appropriate instructional texts). Addresses the role of technical vocabulary, how long the student essays should be, and how one deals with the directionality of knowledge in the…

  19. Assessment of Available Numerical Tools for Dynamic Mooring Analysis

    DEFF Research Database (Denmark)

    Thomsen, Jonas Bjerg; Eskilsson, Claes; Ferri, Francesco

    cover their capabilities. The result of the assessments should make it possible to choose relevant software that will be used throughout the project and also in general use for mooring design of WECs. The report is a part of Work Package 1 : "Task 1.2: Assessment of Available Numerical Tools for Dynamic......This report covers a preliminary assessment of available numerical tools to be used in upcoming full dynamic analysis of the mooring systems assessed in the project _Mooring Solutions for Large Wave Energy Converters_. The assessments tends to cover potential candidate software and subsequently...... Mooring Analysis" and "Milestone 1: Acquisition of Selected Numerical Tools" of the project and was produced by Aalborg University in cooperation with Chalmers University of Technology....

  20. Analysis of Alternatives for Risk Assessment Methodologies and Tools

    Energy Technology Data Exchange (ETDEWEB)

    Nachtigal, Noel M. [Sandia National Lab. (SNL-CA), Livermore, CA (United States). System Analytics; Fruetel, Julia A. [Sandia National Lab. (SNL-CA), Livermore, CA (United States). Systems Research and Analysis; Gleason, Nathaniel J. [Sandia National Lab. (SNL-CA), Livermore, CA (United States). Systems Research and Analysis; Helms, Jovana [Sandia National Lab. (SNL-CA), Livermore, CA (United States). Systems Research and Analysis; Imbro, Dennis Raymond [Sandia National Lab. (SNL-CA), Livermore, CA (United States). Systems Research and Analysis; Sumner, Matthew C. [Sandia National Lab. (SNL-CA), Livermore, CA (United States). Systems Research and Analysis

    2013-10-01

    The purpose of this document is to provide a basic overview and understanding of risk assessment methodologies and tools from the literature and to assess the suitability of these methodologies and tools for cyber risk assessment. Sandia National Laboratories (SNL) performed this review in support of risk modeling activities performed for the Stakeholder Engagement and Cyber Infrastructure Resilience (SECIR) division of the Department of Homeland Security (DHS) Office of Cybersecurity and Communications (CS&C). The set of methodologies and tools covered in this document is not intended to be exhaustive; instead, it focuses on those that are commonly used in the risk assessment community. The classification of methodologies and tools was performed by a group of analysts with experience in risk analysis and cybersecurity, and the resulting analysis of alternatives has been tailored to address the needs of a cyber risk assessment.

  1. TEXTS SENTIMENT-ANALYSIS APPLICATION FOR PUBLIC OPINION ASSESSMENT

    Directory of Open Access Journals (Sweden)

    I. A. Bessmertny

    2015-01-01

    Full Text Available The paper describes an approach to the emotional tonality assessment of natural language texts based on special dictionaries. A method for an automatic assessment of public opinion by means of sentiment-analysis of reviews and discussions followed by published Web-documents is proposed. The method is based on statistics of words in the documents. A pilot model of the software system implementing the sentiment-analysis of natural language text in Russian based on a linear assessment scale is developed. A syntactic analysis and words lemmatization are used to identify terms more correctly. Tonality dictionaries are presented in editable format and are open for enhancing. The program system implementing a sentiment-analysis of the Russian texts based on open dictionaries of tonality is presented for the first time.

  2. Coastal marine eutrophication assessment: a review on data analysis.

    Science.gov (United States)

    Kitsiou, Dimitra; Karydis, Michael

    2011-05-01

    A wide variety of data analysis techniques have been applied for quantitative assessment of coastal marine eutrophication. Indicators for assessing eutrophication and frequency distributions have been used to develop scales for characterizing oligotrophy and eutrophication. Numerical classification has also contributed to the assessment of eutrophic trends by grouping sampling sites of similar trophic conditions. Applications of eutrophication assessment based on Principal Component Analysis and Multidimensional Scaling have also been carried out. In addition, the rapid development of Geographical Information Systems has provided the framework for applications of spatial methods and mapping techniques on eutrophication studies. Satellite data have also contributed to eutrophication assessment especially at large scale. Multiple criteria analysis methods can integrate eutrophication variables together with socio-economic parameters providing a holistic approach particularly useful to policy makers. As the current concept of eutrophication problems is to be examined as part of a coastal management approach, more complex quantitative procedures are needed to provide a platform useful for implementation of environmental policy. The present work reviews methods of data analysis used for the assessment of coastal marine eutrophication. The difficulties in applying these methods on data collected from the marine environment are discussed as well as the future perspectives of spatial and multiple criteria choice methods. Copyright © 2011 Elsevier Ltd. All rights reserved.

  3. No-Reference Video Quality Assessment using MPEG Analysis

    DEFF Research Database (Denmark)

    Søgaard, Jacob; Forchhammer, Søren; Korhonen, Jari

    2013-01-01

    We present a method for No-Reference (NR) Video Quality Assessment (VQA) for decoded video without access to the bitstream. This is achieved by extracting and pooling features from a NR image quality assessment method used frame by frame. We also present methods to identify the video coding...... and estimate the video coding parameters for MPEG-2 and H.264/AVC which can be used to improve the VQA. The analysis differs from most other video coding analysis methods since it is without access to the bitstream. The results show that our proposed method is competitive with other recent NR VQA methods...

  4. 78 FR 27235 - Technical Guidance for Assessing Environmental Justice in Regulatory Analysis

    Science.gov (United States)

    2013-05-09

    ... AGENCY Technical Guidance for Assessing Environmental Justice in Regulatory Analysis AGENCY..., ``Technical Guidance for Assessing Environmental Justice in Regulatory Analysis.'' The purpose of this... Technical Guidance for Assessing Environmental Justice in Regulatory Analysis Docket, EPA/DC, EPA West, Room...

  5. Strategic Choice Analysis by Expert Panels for Migration Impact Assessment

    NARCIS (Netherlands)

    Kourtit, K.; Nijkamp, P.

    2011-01-01

    International migration is a source of policy and research interest in many countries. This paper provides a review of experiences and findings from migration impact assessment worldwide. Various findings are briefly summarised in the context of a systematic migration SWOT analysis for five distinct

  6. Assessing Visualization: An analysis of Chilean teachers’ guidelines

    DEFF Research Database (Denmark)

    Andrade-Molina, Melissa; Díaz, Leonora

    2017-01-01

    The aim of this paper is to argue on how visualization is recommended, by official curricular guidelines, to be assessed in schools. We contend that spatial abilities have been granted with the status of a key element to improve students’ performances by research and also by policy makers. Howeve...... do not help teachers while assessing visualization in schools; rather its focus is embedded in a tradition of training that leads to a reduction of space.......The aim of this paper is to argue on how visualization is recommended, by official curricular guidelines, to be assessed in schools. We contend that spatial abilities have been granted with the status of a key element to improve students’ performances by research and also by policy makers. However......, this importance seems to fade when it comes to assessing students while learning school mathematics and geometry. We conducted an analysis of the official guidelines for the assessment school mathematics in Chile. The analysis of two of those guides is considered here. The results revealed that these guidelines...

  7. A Preliminary Analysis of a Behavioral Classrooms Needs Assessment

    Directory of Open Access Journals (Sweden)

    Justin B. LEAF

    2016-12-01

    Full Text Available Today many special education classrooms implement procedures based upon the principles of Applied Behavior Analysis (ABA to establish educationally relevant skills and decrease aberrant behaviors. However, it is difficult for school staff and consultants to evaluate the implementation of various components of ABA and general classroom set up. In the present study we developed the Behavioral Classroom Needs Assessment as a tool to measure the quality of implementation of principles derived from ABA, teaching, and classroom set up in special education classrooms. Experiment 1 evaluated the reliability of two observers using the Behavioral Classroom Needs Assessment during 128 different observations across 68 different special education classrooms. An Intraclass Correlation Coefficient and Cronbach Alpha Analysis were utilized to determine reliability, and the results showed a high f of reliability across the 40 questions of the assessment. Experiment 2 compared the quality of intervention using the Behavioral Classroom Needs Assessment in five classrooms who received behavioral consultation and five classrooms that did not receive behavioral consultation. The results showed an improvement in the scores on the Behavioral Classroom Needs Assessment for those classrooms in which consultation occurred.

  8. A preliminary analysis of a behavioral classrooms needs assessment

    Directory of Open Access Journals (Sweden)

    Justin B. Leaf

    2016-12-01

    Full Text Available Today many special education classrooms implement procedures based upon the principles of Applied Behavior Analysis (ABA to establish educationally relevant skills and decrease aberrant behaviors. However, it is difficult for school staff and consultants to evaluate the implementation of various components of ABA and general classroom set up. In the present study we developed the Behavioral Classroom Needs Assessment as a tool to measure the quality of implementation of principles derived from ABA, teaching, and classroom set up in special education classrooms. Experiment 1 evaluated the reliability of two observers using the Behavioral Classroom Needs Assessment during 128 different observations across 68 different special education classrooms. An Intraclass Correlation Coefficient and Cronbach Alpha Analysis were utilized to determine reliability, and the results showed a high f of reliability across the 40 questions of the assessment. Experiment 2 compared the quality of intervention using the Behavioral Classroom Needs Assessment in five classrooms who received behavioral consultation and five classrooms that did not receive behavioral consultation. The results showed an improvement in the scores on the Behavioral Classroom Needs Assessment for those classrooms in which consultation occurred.

  9. Quantitative assessment of gastric atrophy using the syntactic structure analysis.

    Science.gov (United States)

    Zaitoun, A M; al Mardini, H; Record, C O

    1998-12-01

    To assess the topographical relation between gastric glands, using the minimum spanning tree (MST), to derive both a model of neighbourhood and quantitative representation of the tissue's architecture, to assess the characteristic features of gastric atrophy, and to assess the grades of gastric atrophy. Haematoxylin and eosin stained sections from corporal and antral biopsy specimens (n = 139) from normal patients and from patients with nonatrophic gastritis and atrophic gastritis of grades 1, 2, and 3 (Sydney system) were assessed by image analysis system (Prodit 5.2) and 11 syntactic structure features were derived. These included both line and connectivity features. Syntactic structure analysis was correlated with the semiquantitative grading system of gastric atrophy. The study showed significant reductions in the number of points and the length of MST in both body and antrum. The standard deviation of the length of MST was significantly increased in all grades of atrophy. The connectivity to two glands was the highest and most affected by the increased grade of atrophy. The reciprocal values of the Wiener, Randic, and Balaban indices showed significant changes in the volume of gland, abnormality in the shape of glands, and changes in irregularity and branching of the glands in both types of gastric mucosa. There was a complete separation in the MST, connectivity, and index values between low grade and high grade gastric atrophy. (1) Gastric atrophy was characterised by loss of the gland, variation in the volume, reduction in the neighbourhood, irregularity in spacing, and abnormality in the shape of the glands. (2) Syntactic structure analysis significantly differentiated minor changes in gastric gland (low grade atrophy) from high grade atrophy of clinical significance. (3) Syntactic structure analysis is a simple, fast, and highly reproducible technique and appears a promising method for quantitative assessment of atrophy.

  10. System of gait analysis based on ground reaction force assessment

    Directory of Open Access Journals (Sweden)

    František Vaverka

    2015-12-01

    Full Text Available Background: Biomechanical analysis of gait employs various methods used in kinematic and kinetic analysis, EMG, and others. One of the most frequently used methods is kinetic analysis based on the assessment of the ground reaction forces (GRF recorded on two force plates. Objective: The aim of the study was to present a method of gait analysis based on the assessment of the GRF recorded during the stance phase of two steps. Methods: The GRF recorded with a force plate on one leg during stance phase has three components acting in directions: Fx - mediolateral, Fy - anteroposterior, and Fz - vertical. A custom-written MATLAB script was used for gait analysis in this study. This software displays instantaneous force data for both legs as Fx(t, Fy(t and Fz(t curves, automatically determines the extremes of functions and sets the visual markers defining the individual points of interest. Positions of these markers can be easily adjusted by the rater, which may be necessary if the GRF has an atypical pattern. The analysis is fully automated and analyzing one trial takes only 1-2 minutes. Results: The method allows quantification of temporal variables of the extremes of the Fx(t, Fy(t, Fz(t functions, durations of the braking and propulsive phase, duration of the double support phase, the magnitudes of reaction forces in extremes of measured functions, impulses of force, and indices of symmetry. The analysis results in a standardized set of 78 variables (temporal, force, indices of symmetry which can serve as a basis for further research and diagnostics. Conclusions: The resulting set of variable offers a wide choice for selecting a specific group of variables with consideration to a particular research topic. The advantage of this method is the standardization of the GRF analysis, low time requirements allowing rapid analysis of a large number of trials in a short time, and comparability of the variables obtained during different research measurements.

  11. Coupling multi-criteria decision analysis, life-cycle assessment, and risk assessment for emerging threats.

    Science.gov (United States)

    Linkov, Igor; Seager, Thomas P

    2011-06-15

    Emerging environmental threats such as novel chemical compounds, biological agents, and nanomaterials present serious challenges to traditional models of risk analysis and regulatory risk management processes. Even a massive expansion of risk and life-cycle assessment research efforts is unlikely to keep pace with rapid technological change resulting in new and modified materials with changing properties. Therefore, it is essential to have a framework for interpreting available information in the context of high uncertainty and a strategy for prioritizing research efforts to reduce those uncertainties that are most critical. We discuss how integrating the three analytic approaches of risk assessment, life-cycle assessment, and multicriteria decision analysis into a framework permits understanding uncertainty and prioritizes needs for scientific research. Our approach is illustrated with two separate cases: nanomaterials and contaminated sediment remediation.

  12. Discriminant analysis in Polish manufacturing sector performance assessment

    Directory of Open Access Journals (Sweden)

    Józef Dziechciarz

    2004-01-01

    Full Text Available This is a presentation of the preliminary results of a larger project on the determination of the attractiveness of manufacturing branches. Results of the performance assessment of Polish manufacturing branches in 2000 (section D „Manufacturing” – based on NACE – Nomenclatures des Activites de Communite Europeene are shown. In the research, the classical (Fisher’s linear discriminant analysis technique was used for the analysis of the profit generation ability by the firms belonging to a certain production branch. For estimation, the data describing group level was used – for cross-validation, the classes data.

  13. Assessing Spontaneous Combustion Instability with Recurrence Quantification Analysis

    Science.gov (United States)

    Eberhart, Chad J.; Casiano, Matthew J.

    2016-01-01

    Spontaneous instabilities can pose a significant challenge to verification of combustion stability, and characterizing its onset is an important avenue of improvement for stability assessments of liquid propellant rocket engines. Recurrence Quantification Analysis (RQA) is used here to explore nonlinear combustion dynamics that might give insight into instability. Multiple types of patterns representative of different dynamical states are identified within fluctuating chamber pressure data, and markers for impending instability are found. A class of metrics which describe these patterns is also calculated. RQA metrics are compared with and interpreted against another metric from nonlinear time series analysis, the Hurst exponent, to help better distinguish between stable and unstable operation.

  14. Individuals' stress assessment using human-smartphone interaction analysis

    DEFF Research Database (Denmark)

    Ciman, Matteo; Wac, Katarzyna

    2018-01-01

    The increasing presence of stress in people’ lives has motivated much research efforts focusing on continuous stress assessment methods of individuals, leveraging smartphones and wearable devices. These methods have several drawbacks, i.e., they use invasive external devices, thus increasing entry...... costs and reducing user acceptance, or they use some of privacy-related information. This paper presents an approach for stress assessment that leverages data extracted from smartphone sensors, and that is not invasive concerning privacy. Two different approaches are presented. One, based on smartphone...... gestures analysis, e.g., ‘tap’, ‘scroll’, ‘swipe’ and ‘text writing’, and evaluated in laboratory settings with 13 participants (F-measure 79-85% within-subject model, 70-80% global model); the second one based on smartphone usage analysis and tested in-the-wild with 25 participants (F-measure 77...

  15. Hybrid vehicle potential assessment. Volume 2. Mission analysis

    Energy Technology Data Exchange (ETDEWEB)

    Surber, F.T.; Deshpande, G.K.

    1979-09-30

    Seven vehicle missions were selected and defined for use in assessing hybrid vehicle concepts and are summarized. These missions were selected to provide general coverage of current and future vehicle uses and classes. Their travel pattern definitions are based primarily on analysis of the 1969 National Personal Transportation Study travel data. Performance requirements are based on current highway designs and speed limits and the results of studies of performance requirements for safe and non-interfering operation in today's traffic environment. The mission definitions are independent of vehicle technology so even though they were selected and defined for use in assessing hybrid vehicles, they could also be used in assessing the potential of other vehicle technologies such as electric vehicles, turbo-charging, Stirling engines, etc.

  16. No-Reference Video Quality Assessment using Codec Analysis

    DEFF Research Database (Denmark)

    Søgaard, Jacob; Forchhammer, Søren; Korhonen, Jari

    2015-01-01

    A no-reference video quality assessment (VQA) method is presented for videos distorted by H.264/AVC and MPEG-2. The assessment is performed without access to the bit-stream. Instead we analyze and estimate coefficients based on decoded pixels. The approach involves distinguishing between the two...... types of videos, estimating the level of quantization used in the I-frames, and exploiting this information to assess the video quality. In order to do this for H.264/AVC, the distribution of the DCT-coefficients after intra-prediction and deblocking are modeled. To obtain VQA features for H.264/AVC, we...... propose a novel estimation method of the quantization in H.264/AVC videos without bitstream access, which can also be used for Peak Signalto-Noise Ratio (PSNR) estimation. The results from the MPEG-2 and H.264/AVC analysis are mapped to a perceptual measure of video quality by Support Vector Regression...

  17. Brazilian scientific production analysis about psychological assessment in career guidance

    OpenAIRE

    Ambiel,Rodolfo Augusto Matteo; Polli, Mariana Fralleti de

    2011-01-01

    This study performed a survey of Brazilian scientific literature, published in national databases of scientific journals about the psychological assessment in the context of career guidance. To this end, a search was performed on the basis PePSIC and Scielo in August 2011, with selection and exclusion criteria and terms defined a priori, with the scope of analysis a total of 24 articles that were full texts analyzed. The results were organized into three blocks, about authors, journals and me...

  18. Safety analysis, risk assessment, and risk acceptance criteria

    Energy Technology Data Exchange (ETDEWEB)

    Jamali, K. [Dept. of Energy, Germantown, MD (United States). Core Technical Support and Facility Transition; Stack, D.W.; Sullivan, L.H.; Sanzo, D.L. [Los Alamos National Lab., NM (United States)

    1997-08-01

    This paper discusses a number of topics that relate safety analysis as documented in the Department of Energy (DOE) safety analysis reports (SARs), probabilistic risk assessments (PRA) as characterized primarily in the context of the techniques that have assumed some level of formality in commercial nuclear power plant applications, and risk acceptance criteria as an outgrowth of PRA applications. DOE SARs of interest are those that are prepared for DOE facilities under DOE Order 5480.23 and the implementing guidance in DOE STD-3009-94. It must be noted that the primary area of application for DOE STD-3009 is existing DOE facilities and that certain modifications of the STD-3009 approach are necessary in SARs for new facilities. Moreover, it is the hazard analysis (HA) and accident analysis (AA) portions of these SARs that are relevant to the present discussions. Although PRAs can be qualitative in nature, PRA as used in this paper refers more generally to all quantitative risk assessments and their underlying methods. HA as used in this paper refers more generally to all qualitative risk assessments and their underlying methods that have been in use in hazardous facilities other than nuclear power plants. This discussion includes both quantitative and qualitative risk assessment methods. PRA has been used, improved, developed, and refined since the Reactor Safety Study (WASH-1400) was published in 1975 by the Nuclear Regulatory Commission (NRC). Much debate has ensued since WASH-1400 on exactly what the role of PRA should be in plant design, reactor licensing, `ensuring` plant and process safety, and a large number of other decisions that must be made for potentially hazardous activities. Of particular interest in this area is whether the risks quantified using PRA should be compared with numerical risk acceptance criteria (RACs) to determine whether a facility is `safe.` Use of RACs requires quantitative estimates of consequence frequency and magnitude.

  19. Student Assessment of Desirable Technical Skills: A Correspondence Analysis Approach

    OpenAIRE

    András Farkas; Viktor Nagy

    2008-01-01

    The importance rating of the technical skills and their related knowledge ofindustrial engineering and business management graduates are developed. A survey wasconducted at the Faculty of Economics of the Budapest Tech to measure and evaluate 17technical skills required by employers. The study utilized questionnaires to rate and rankthese skills based on student assessment. The research was supported by a multivariatestatistical method referred to as correspondence analysis. Details of the ma...

  20. Public transport risk assessment through fault tree analysis

    Directory of Open Access Journals (Sweden)

    Z. Yaghoubpour

    2016-04-01

    Full Text Available This study focused on the public transport risk assessment in District one of ​​Tehran through Fault Tree Analysis involving the three criteria of human, vehicle and road in Haddon matrix. In fact, it examined the factors contributing to the occurrence of road accidents at several urban black spots within District 1. Relying on road safety checklists and survey of experts, this study made an effort to help urban managers to assess the risks in the public transport and prevent road accidents. Finally, the risk identification and assessment of public transport in District one yielded several results to answer the research questions. The hypotheses analysis suggested that safety issues involved in public transport are concerned by urban managers. The key reactive measures are investigation of accidents, identification of causes and correction of black spots. In addition to high costs, however, the reactive measures give rise to multiple operational problems such as traffic navigation and guaranteeing user safety in every operation. The case study highlighted the same fact. The macro-level management in the metropolis of Tehran is critical. The urban road casualties and losses can be curtailed by preventive measures such as continuous assessment of road safety.

  1. Regional risk assessment for contaminated sites part 1: vulnerability assessment by multicriteria decision analysis.

    Science.gov (United States)

    Zabeo, A; Pizzol, L; Agostini, P; Critto, A; Giove, S; Marcomini, A

    2011-11-01

    As highlighted in the EU Soil Communication, local contamination is one of the main soil threats and it is often related to present and past industrial activities which left a legacy of a high number of contaminated sites in Europe. These contaminated sites can be harmful to many different receptors according to their sensitivity/susceptibility to contamination, and specific vulnerability evaluations are needed in order to manage this widely spread environmental issue. In this paper a novel comprehensive vulnerability assessment framework to assess regional receptor susceptibility to contaminated site is presented. The developed methodology, which combines multi criteria decision analysis (MCDA) techniques and spatial analysis, can be applied to different receptors recognized as relevant for regional assessment. In order to characterize each receptor, picked parameters significant for the estimation of the vulnerability to contaminated sites have been selected, normalized and aggregated by means of multi criteria decision analysis (MCDA) techniques. The developed MCDA methodology, based on the Choquet integral, allows to include expert judgments for the elicitation of synergic and conflicting effects between involved criteria and is applied to all the geographical objects representing the identified receptors. To test the potential of the vulnerability methodology, it has been applied to a specific case study area in the upper Silesia region of Poland where it proved to be reliable and consistent with the environmental experts' expected results. The vulnerability assessment results indicate that groundwater is the most vulnerable receptor characterized by a wide area with vulnerability scores belonging to the highest vulnerability class. As far as the other receptors are concerned, human health and surface water are characterized by quite homogeneous vulnerability scores falling in the medium-high vulnerability classes, while protected areas resulted to be the less

  2. Safety analysis and review system (SARS) assessment report

    Energy Technology Data Exchange (ETDEWEB)

    Browne, E.T.

    1981-03-01

    Under DOE Order 5481.1, Safety Analysis and Review System for DOE Operations, safety analyses are required for DOE projects in order to ensure that: (1) potential hazards are systematically identified; (2) potential impacts are analyzed; (3) reasonable measures have been taken to eliminate, control, or mitigate the hazards; and (4) there is documented management authorization of the DOE operation based on an objective assessment of the adequacy of the safety analysis. This report is intended to provide the DOE Office of Plans and Technology Assessment (OPTA) with an independent evaluation of the adequacy of the ongoing safety analysis effort. As part of this effort, a number of site visits and interviews were conducted, and FE SARS documents were reviewed. The latter included SARS Implementation Plans for a number of FE field offices, as well as safety analysis reports completed for certain FE operations. This report summarizes SARS related efforts at the DOE field offices visited and evaluates the extent to which they fulfill the requirements of DOE 5481.1.

  3. Assessment of table olive fermentation by functional data analysis.

    Science.gov (United States)

    Ruiz-Bellido, M A; Romero-Gil, V; García-García, P; Rodríguez-Gómez, F; Arroyo-López, F N; Garrido-Fernández, A

    2016-12-05

    For the first time, functional data analysis (FDA) was used to assess the effects of different treatments on Protection Denomination of Origin Aloreña de Málaga table olive fermentations, focusing on the evolution of yeast population. The analysis of fermentation by a conventional approach led to scarce information. However, the transformation of microbial (and also physicochemical) data into smooth curves allowed the application of a new battery of statistical tools for the analysis of fermentations (functional pointwise estimation of the averages and standard deviations, maximum, minimum, first and second derivatives, functional regression, and functional F and t-tests). FDA showed that all the treatments assayed led to similar trends in yeast population while changes in pH and titratable acidity profiles led to several significant differences. Therefore, FDA represents a promising and valuable tool for studying table olive fermentations and for food microbiology in general. Copyright © 2016 Elsevier B.V. All rights reserved.

  4. [External quality assessment of semen analysis in Poland].

    Science.gov (United States)

    Jedrzejczak, Piotr; Talarczyk, Joanna; Taszarek-Hauke, Grazyna; Berger, Anna; Hauke, Jan; Pawelczyk, Leszek

    2012-11-01

    Semen analysis is an important part of male infertility diagnosis and should be performed according to international recommendations. The reliability of the results depends mainly on the qualifications of the laboratory analysts. External Quality Assessment Programmes (EQAP) are performed in laboratories worldwide in order to standardize the results. The aim of this study was to perform the first in Poland comparison of the results of sperm analysis from different laboratories. Forty two Polish laboratories were invited to participate in the EQAP and eight laboratories agreed to take part in the analysis. They were sent uniform semen samples, prepared in accordance with the WHO standards: one sample for the assessment of concentration, two for motility (on DVD) and two for the morphology of the sperm (Papanicolau staining). The reference group was comprised of three employees of the Andrologic Laboratory of Poznań University of Medical Sciences, who regularly take part in EQAP organized by ESHRE. The reference group analyzed the samples from the participating laboratories. Statistic features such as mean, median, standard deviation, minimal and maximal value, first and third quartile were assessed for every examined parameter Z-score index was used to compare the differences between assessed laboratories and the reference laboratory. The acceptable Z-score range was +/-1. Average concentration in the reference group was 43 mln/ml, while in the assessed laboratories (L) it was between 31 (L4) and 72.5 mln/ml (L1). Z-score for concentration analysis in three laboratories exceeded +/- 1 (2.51 for L1, 1.42 for L2 and -1.02 for L4). In the analysis of the first sample for progressive motility the reference group received 59%, while the values of participating laboratories varied from 42,5% (L1) to 80% (L4). Most of the centers achieved Z-score within the normal range, except for L1 and L4 (-1.48 and 1.88 respectively). The reference value for the second sample for

  5. Social and ethical analysis in health technology assessment.

    Science.gov (United States)

    Tantivess, Sripen

    2014-05-01

    This paper presents a review of the domestic and international literature on the assessment of the social and ethical implications of health technologies. It gives an overview of the key concepts, principles, and approaches that should be taken into account when conducting a social and ethical analysis within health technology assessment (HTA). Although there is growing consensus among healthcare experts that the social and ethical ramifications of a given technology should be examined before its adoption, the demand for this kind of analysis among policy-makers around the world, including in Thailand, has so far been lacking. Currently decision-makers mainly base technology adoption decisions using evidence on clinical effectiveness, value for money, and budget impact, while social and ethical aspects have been neglected. Despite the recognized importance of considering equity, justice, and social issues when making decisions regarding health resource allocation, the absence of internationally-accepted principles and methodologies, among other factors, hinders research in these areas. Given that developing internationally agreed standards takes time, it has been recommended that priority be given to defining processes that are justifiable, transparent, and contestable. A discussion of the current situation in Thailand concerning social and ethical analysis of health technologies is also presented.

  6. Development and assessment of best estimate integrated safety analysis code

    Energy Technology Data Exchange (ETDEWEB)

    Chung, Bub Dong; Lee, Young Jin; Hwang, Moon Kyu (and others)

    2007-03-15

    Improvement of the integrated safety analysis code MARS3.0 has been carried out and a multi-D safety analysis application system has been established. Iterative matrix solver and parallel processing algorithm have been introduced, and a LINUX version has been generated to enable MARS to run in cluster PCs. MARS variables and sub-routines have been reformed and modularised to simplify code maintenance. Model uncertainty analyses have been performed for THTF, FLECHT, NEPTUN, and LOFT experiments as well as APR1400 plant. Participations in international cooperation research projects such as OECD BEMUSE, SETH, PKL, BFBT, and TMI-2 have been actively pursued as part of code assessment efforts. The assessment, evaluation and experimental data obtained through international cooperation projects have been registered and maintained in the T/H Databank. Multi-D analyses of APR1400 LBLOCA, DVI Break, SLB, and SGTR have been carried out as a part of application efforts in multi-D safety analysis. GUI based 3D input generator has been developed for user convenience. Operation of the MARS Users Group (MUG) was continued and through MUG, the technology has been transferred to 24 organisations. A set of 4 volumes of user manuals has been compiled and the correction reports for the code errors reported during MARS development have been published.

  7. Multicriteria benefit-risk assessment using network meta-analysis.

    Science.gov (United States)

    van Valkenhoef, Gert; Tervonen, Tommi; Zhao, Jing; de Brock, Bert; Hillege, Hans L; Postmus, Douwe

    2012-04-01

    To enable multicriteria benefit-risk (BR) assessment of any number of alternative treatments using all available evidence from a network of clinical trials. We design a general method for multicriteria decision aiding with criteria measurements from Mixed Treatment Comparison (MTC) analyses. To evaluate the method, we apply it to BR assessment of four second-generation antidepressants and placebo in the setting of a published peer-reviewed systematic review. The analysis without preference information shows that placebo is supported by a wide range of possible preferences. Preference information provided by a clinical expert showed that although treatment with antidepressants is warranted for severely depressed patients, for mildly depressed patients placebo is likely to be the best option. It is difficult to choose between the four antidepressants, and the results of the model indicate a high degree of uncertainty. The designed method enables quantitative BR analysis of alternative treatments using all available evidence from a network of clinical trials. The preference-free analysis can be useful in presenting the results of an MTC considering multiple outcomes. Copyright © 2012 Elsevier Inc. All rights reserved.

  8. Layer of protection analysis for reactive chemical risk assessment.

    Science.gov (United States)

    Wei, Chunyang; Rogers, William J; Mannan, M Sam

    2008-11-15

    Reactive chemical hazards have been a significant concern for the chemical process industries (CPI). Without sufficient control and mitigation of chemical reaction hazards, reactive incidents have led to severe consequences, such as release of flammable and toxic materials, fires and explosions, and threats to human lives, properties, and the environment. Consequence of reactive hazards can be well understood through calorimetric testing and computational techniques. However, risks of incidents caused by reactive chemicals have not been well addressed due partly to sparse failure frequency data. In this paper, the semi-quantitative layer of protection analysis (LOPA) approach is used to estimate reactive chemical risk, and the probabilities or frequencies of failure scenarios are addressed. Using LOPA, reactive risks can be evaluated with respect to predefined criteria, and the effectiveness of risk reduction measures can be assessed. The hydroxylamine (HA) production system is employed as a case study to demonstrate the application of LOPA to reactive chemical risk assessment.

  9. Health Monitoring System Technology Assessments: Cost Benefits Analysis

    Science.gov (United States)

    Kent, Renee M.; Murphy, Dennis A.

    2000-01-01

    The subject of sensor-based structural health monitoring is very diverse and encompasses a wide range of activities including initiatives and innovations involving the development of advanced sensor, signal processing, data analysis, and actuation and control technologies. In addition, it embraces the consideration of the availability of low-cost, high-quality contributing technologies, computational utilities, and hardware and software resources that enable the operational realization of robust health monitoring technologies. This report presents a detailed analysis of the cost benefit and other logistics and operational considerations associated with the implementation and utilization of sensor-based technologies for use in aerospace structure health monitoring. The scope of this volume is to assess the economic impact, from an end-user perspective, implementation health monitoring technologies on three structures. It specifically focuses on evaluating the impact on maintaining and supporting these structures with and without health monitoring capability.

  10. Analysis and Comparison of Objective Methods for Image Quality Assessment

    Directory of Open Access Journals (Sweden)

    P. S. Babkin

    2014-01-01

    Full Text Available The purpose of this work is research and modification of the reference objective methods for image quality assessment. The ultimate goal is to obtain a modification of formal assessments that more closely corresponds to the subjective expert estimates (MOS.In considering the formal reference objective methods for image quality assessment we used the results of other authors, which offer results and comparative analyzes of the most effective algorithms. Based on these investigations we have chosen two of the most successful algorithm for which was made a further analysis in the MATLAB 7.8 R 2009 a (PQS and MSSSIM. The publication focuses on the features of the algorithms, which have great importance in practical implementation, but are insufficiently covered in the publications by other authors.In the implemented modification of the algorithm PQS boundary detector Kirsch was replaced by the boundary detector Canny. Further experiments were carried out according to the method of the ITU-R VT.500-13 (01/2012 using monochrome images treated with different types of filters (should be emphasized that an objective assessment of image quality PQS is applicable only to monochrome images. Images were obtained with a thermal imaging surveillance system. The experimental results proved the effectiveness of this modification.In the specialized literature in the field of formal to evaluation methods pictures, this type of modification was not mentioned.The method described in the publication can be applied to various practical implementations of digital image processing.Advisability and effectiveness of using the modified method of PQS to assess the structural differences between the images are shown in the article and this will be used in solving the problems of identification and automatic control.

  11. Supporting analysis and assessments quality metrics: Utility market sector

    Energy Technology Data Exchange (ETDEWEB)

    Ohi, J. [National Renewable Energy Lab., Golden, CO (United States)

    1996-10-01

    In FY96, NREL was asked to coordinate all analysis tasks so that in FY97 these tasks will be part of an integrated analysis agenda that will begin to define a 5-15 year R&D roadmap and portfolio for the DOE Hydrogen Program. The purpose of the Supporting Analysis and Assessments task at NREL is to provide this coordination and conduct specific analysis tasks. One of these tasks is to prepare the Quality Metrics (QM) for the Program as part of the overall QM effort at DOE/EERE. The Hydrogen Program one of 39 program planning units conducting QM, a process begun in FY94 to assess benefits/costs of DOE/EERE programs. The purpose of QM is to inform decisionmaking during budget formulation process by describing the expected outcomes of programs during the budget request process. QM is expected to establish first step toward merit-based budget formulation and allow DOE/EERE to get {open_quotes}most bang for its (R&D) buck.{close_quotes} In FY96. NREL coordinated a QM team that prepared a preliminary QM for the utility market sector. In the electricity supply sector, the QM analysis shows hydrogen fuel cells capturing 5% (or 22 GW) of the total market of 390 GW of new capacity additions through 2020. Hydrogen consumption in the utility sector increases from 0.009 Quads in 2005 to 0.4 Quads in 2020. Hydrogen fuel cells are projected to displace over 0.6 Quads of primary energy in 2020. In future work, NREL will assess the market for decentralized, on-site generation, develop cost credits for distributed generation benefits (such as deferral of transmission and distribution investments, uninterruptible power service), cost credits for by-products such as heat and potable water, cost credits for environmental benefits (reduction of criteria air pollutants and greenhouse gas emissions), compete different fuel cell technologies against each other for market share, and begin to address economic benefits, especially employment.

  12. Assessment of Kidney Function After Allograft Transplantation by Texture Analysis.

    Science.gov (United States)

    Abbasian Ardakani, Ali; Mohammadi, Afshin; Khalili Najafabad, Bahareh; Abolghasemi, Jamileh

    2017-03-01

    Ultrasonography is the preferable imaging technique for monitoring and assessing complications in kidney allograft transplants. Computer-aided diagnostic system based on texture analysis in ultrasonographic imaging is recommended to identify changes in kidney function after allograft transplantation. A total of 61 biopsy-proven kidney allograft recipients (11 rejected and 50 unrejected) were assessed by a computer-aided diagnostic system. Up to 270 statistical texture features were extracted as descriptors for each region of interest in each recipient. Correlations of texture features with serum creatinine level and differences between rejected and unrejected allografts were analyzed. An area under the receiver operating characteristic curve was calculated for each significant texture feature. Linear discriminant analysis was employed to analyze significant features and increase discriminative power. Recipients were classified by the first nearest neighbor classifier. Fourteen texture features had a significant correlation with serum creatinine level and 16 were significantly different between the rejected and unrejected allografts, for which an area under the curve values were in the range of 0.575 for difference entropy S(4,0) to 0.676 for kurtosis. Using all 16 features, linear discriminant analysis indicated higher performance for classification of the two groups with an area under the curve of 0.975, which corresponded to a sensitivity of 90.9%, a specificity of 100%, a positive predictive value of 100%, and a negative predictive value of 98.0%. Texture analysis was a reliable method, with the potential for characterization, and can help physicians to diagnose kidney failure after transplantation on ultrasonographic imaging.

  13. A Big Data Analysis Approach for Rail Failure Risk Assessment.

    Science.gov (United States)

    Jamshidi, Ali; Faghih-Roohi, Shahrzad; Hajizadeh, Siamak; Núñez, Alfredo; Babuska, Robert; Dollevoet, Rolf; Li, Zili; De Schutter, Bart

    2017-08-01

    Railway infrastructure monitoring is a vital task to ensure rail transportation safety. A rail failure could result in not only a considerable impact on train delays and maintenance costs, but also on safety of passengers. In this article, the aim is to assess the risk of a rail failure by analyzing a type of rail surface defect called squats that are detected automatically among the huge number of records from video cameras. We propose an image processing approach for automatic detection of squats, especially severe types that are prone to rail breaks. We measure the visual length of the squats and use them to model the failure risk. For the assessment of the rail failure risk, we estimate the probability of rail failure based on the growth of squats. Moreover, we perform severity and crack growth analyses to consider the impact of rail traffic loads on defects in three different growth scenarios. The failure risk estimations are provided for several samples of squats with different crack growth lengths on a busy rail track of the Dutch railway network. The results illustrate the practicality and efficiency of the proposed approach. © 2017 The Authors Risk Analysis published by Wiley Periodicals, Inc. on behalf of Society for Risk Analysis.

  14. Regression tree analysis of ecological momentary assessment data.

    Science.gov (United States)

    Richardson, Ben; Fuller-Tyszkiewicz, Matthew; O'Donnell, Renee; Ling, Mathew; Staiger, Petra K

    2017-09-01

    An increasingly popular form of data collection in health psychology research is Ecological Momentary Assessment (EMA); that is, using diaries or smartphones to collect intensive longitudinal data. This method is increasingly applied to the study of relationships between state-based aspects of individuals' functioning and health outcomes (e.g., binge eating, alcohol use). Analysis of such data is challenging and regression tree modelling (RTM) may be a useful alternative to multilevel modelling for investigating the association between a set of explanatory variables and a continuous outcome. Furthermore, RTM outputs 'decision trees' that could be used by health practitioners to guide assessment and tailor intervention. In contrast to regression, RTM is able to easily accommodate many complex, higher-order interactions between predictor variables (without the need to create explicit interaction terms). These benefits make the technique useful for those interested in monitoring and intervening upon health and psychological outcomes (e.g., mood, eating behaviour, risky alcohol use, and treatment adherence). Using real data, this paper demonstrates both the benefits and limitations of RTM and how to extend these models to accommodate analysis of nested data; that is, data that arise from EMA where repeated observations are nested within individuals.

  15. Assessing swine thermal comfort by image analysis of postural behaviors.

    Science.gov (United States)

    Xin, H

    1999-01-01

    Postural behavior is an integral response of animals to complex environmental factors. Huddling, nearly contacting one another on the side, and spreading are common postural behaviors of group-housed animals undergoing cold, comfortable, and warm/hot sensations, respectively. These postural patterns have been routinely used by animal caretakers to assess thermal comfort of the animals and to make according adjustment on the environmental settings or management schemes. This manual adjustment approach, however, has the inherent limitations of daily discontinuity and inconsistency between caretakers in interpretation of the animal comfort behavior. The goal of this project was to explore a novel, automated image analysis system that would assess the thermal comfort of swine and make proper environmental adjustments to enhance animal wellbeing and production efficiency. This paper describes the progress and on-going work toward the achievement of our proposed goal. The feasibility of classifying the thermal comfort state of young pigs by neural network (NN) analysis of their postural images was first examined. It included exploration of using certain feature selections of the postural behavioral images as the input to a three-layer NN that was trained to classify the corresponding thermal comfort state as being cold, comfortable, or warm. The image feature selections, a critical step for the classification, examined in this study included Fourier coefficient (FC), moment (M), perimeter and area (P&A), and combination of M and P&A of the processed binary postural images. The result was positive, with the combination of M and P&A as the input feature to the NN yielding the highest correct classification rate. Subsequent work included the development of hardware and computational algorithms that enable automatic image segmentation, motion detection, and the selection of the behavioral images suitable for use in the classification. Work is in progress to quantify the

  16. Improved nuclear medicine uniformity assessment with noise texture analysis.

    Science.gov (United States)

    Nelson, Jeffrey S; Christianson, Olav I; Harkness, Beth A; Madsen, Mark T; Mah, Eugene; Thomas, Stephen R; Zaidi, Habib; Samei, Ehsan

    2014-01-01

    Because γ cameras are generally susceptible to environmental conditions and system vulnerabilities, they require routine evaluation of uniformity performance. The metrics for such evaluations are commonly pixel value-based. Although these metrics are typically successful at identifying regional nonuniformities, they often do not adequately reflect subtle periodic structures; therefore, additional visual inspections are required. The goal of this project was to develop, test, and validate a new uniformity analysis metric capable of accurately identifying structures and patterns present in nuclear medicine flood-field uniformity images. A new uniformity assessment metric, termed the structured noise index (SNI), was based on the 2-dimensional noise power spectrum (NPS). The contribution of quantum noise was subtracted from the NPS of a flood-field uniformity image, resulting in an NPS representing image artifacts. A visual response filter function was then applied to both the original NPS and the artifact NPS. A single quantitative score was calculated on the basis of the magnitude of the artifact. To verify the validity of the SNI, an observer study was performed with 5 expert nuclear medicine physicists. The correlation between the SNI and the visual score was assessed with Spearman rank correlation analysis. The SNI was also compared with pixel value-based assessment metrics modeled on the National Electrical Manufacturers Association standard for integral uniformity in both the useful field of view (UFOV) and the central field of view (CFOV). The SNI outperformed the pixel value-based metrics in terms of its correlation with the visual score (ρ values for the SNI, integral UFOV, and integral CFOV were 0.86, 0.59, and 0.58, respectively). The SNI had 100% sensitivity for identifying both structured and nonstructured nonuniformities; for the integral UFOV and CFOV metrics, the sensitivities were only 62% and 54%, respectively. The overall positive predictive

  17. Enhancing e-waste estimates: Improving data quality by multivariate Input–Output Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Wang, Feng, E-mail: fwang@unu.edu [Institute for Sustainability and Peace, United Nations University, Hermann-Ehler-Str. 10, 53113 Bonn (Germany); Design for Sustainability Lab, Faculty of Industrial Design Engineering, Delft University of Technology, Landbergstraat 15, 2628CE Delft (Netherlands); Huisman, Jaco [Institute for Sustainability and Peace, United Nations University, Hermann-Ehler-Str. 10, 53113 Bonn (Germany); Design for Sustainability Lab, Faculty of Industrial Design Engineering, Delft University of Technology, Landbergstraat 15, 2628CE Delft (Netherlands); Stevels, Ab [Design for Sustainability Lab, Faculty of Industrial Design Engineering, Delft University of Technology, Landbergstraat 15, 2628CE Delft (Netherlands); Baldé, Cornelis Peter [Institute for Sustainability and Peace, United Nations University, Hermann-Ehler-Str. 10, 53113 Bonn (Germany); Statistics Netherlands, Henri Faasdreef 312, 2492 JP Den Haag (Netherlands)

    2013-11-15

    Highlights: • A multivariate Input–Output Analysis method for e-waste estimates is proposed. • Applying multivariate analysis to consolidate data can enhance e-waste estimates. • We examine the influence of model selection and data quality on e-waste estimates. • Datasets of all e-waste related variables in a Dutch case study have been provided. • Accurate modeling of time-variant lifespan distributions is critical for estimate. - Abstract: Waste electrical and electronic equipment (or e-waste) is one of the fastest growing waste streams, which encompasses a wide and increasing spectrum of products. Accurate estimation of e-waste generation is difficult, mainly due to lack of high quality data referred to market and socio-economic dynamics. This paper addresses how to enhance e-waste estimates by providing techniques to increase data quality. An advanced, flexible and multivariate Input–Output Analysis (IOA) method is proposed. It links all three pillars in IOA (product sales, stock and lifespan profiles) to construct mathematical relationships between various data points. By applying this method, the data consolidation steps can generate more accurate time-series datasets from available data pool. This can consequently increase the reliability of e-waste estimates compared to the approach without data processing. A case study in the Netherlands is used to apply the advanced IOA model. As a result, for the first time ever, complete datasets of all three variables for estimating all types of e-waste have been obtained. The result of this study also demonstrates significant disparity between various estimation models, arising from the use of data under different conditions. It shows the importance of applying multivariate approach and multiple sources to improve data quality for modelling, specifically using appropriate time-varying lifespan parameters. Following the case study, a roadmap with a procedural guideline is provided to enhance e

  18. Risk assessment and cost-effectiveness/utility analysis.

    Science.gov (United States)

    Busch, Michael; Walderhaug, Mark; Custer, Brian; Allain, Jean-Pierre; Reddy, Ravi; McDonough, Brian

    2009-04-01

    Decision-makers at all levels of public health and transfusion medicine have always assessed the risks and benefits of their decisions. Decisions are usually guided by immediately available information and a significant amount of experience and judgment. For decisions concerning familiar situations and common problems, judgment and experience may work quite well, but this type of decision process can lack clarity and accountability. Public health challenges are changing as emerging diseases and expensive technologies complicate the decision-makers' task, confronting the decision-maker with new problems that include multiple potential solutions. Decisions regarding policies and adoption of technologies are particularly complex in transfusion medicine due to the scope of the field, implications for public health, and legal, regulatory and public expectations regarding blood safety. To assist decision-makers, quantitative risk assessment and cost-effectiveness analysis are now being more widely applied. This set of articles will introduce risk assessment and cost-effectiveness methodologies and discuss recent applications of these methods in transfusion medicine.

  19. Multi-dimensional flood vulnerability assessment using data envelopment analysis

    Science.gov (United States)

    Zahid, Zalina; Saharizan, Nurul Syuhada; Hamzah, Paezah; Hussin, Siti Aida Sheikh; Khairi, Siti Shaliza Mohd

    2017-11-01

    Malaysia has been greatly impacted by flood during monsoon seasons. Even though flood prone areas are well identified, assessment on the vulnerability of the disaster is lacking. Assessment of flood vulnerability, defined as the potential for loss when a disaster occurs, is addressed in this paper. The focus is on the development of flood vulnerability measurement in 11 states in Peninsular Malaysia using a non-parametric approach of Data Envelopment Analysis. Scores for three dimensions of flood vulnerability (Population Vulnerability, Social Vulnerability and Biophysical) were calculated using secondary data of selected input and output variables across an 11-year period from 2004 to 2014. The results showed that Johor and Pahang were the most vulnerable to flood in terms of Population Vulnerability, followed by Kelantan, the most vulnerable to flood in terms of Social Vulnerability and Kedah, Pahang and Terengganu were the most vulnerable to flood in terms of Biophysical Vulnerability among the eleven states. The results also showed that the state of Johor, Pahang and Kelantan to be most vulnerable across the three dimensions. Flood vulnerability assessment is important as it provides invaluable information that will allow the authority to identify and develop plans for flood mitigation and to reduce the vulnerability of flood at the affected regions.

  20. Multiple criteria decision analysis for health technology assessment.

    Science.gov (United States)

    Thokala, Praveen; Duenas, Alejandra

    2012-12-01

    Multicriteria decision analysis (MCDA) has been suggested by some researchers as a method to capture the benefits beyond quality adjusted life-years in a transparent and consistent manner. The objectives of this article were to analyze the possible application of MCDA approaches in health technology assessment and to describe their relative advantages and disadvantages. This article begins with an introduction to the most common types of MCDA models and a critical review of state-of-the-art methods for incorporating multiple criteria in health technology assessment. An overview of MCDA is provided and is compared against the current UK National Institute for Health and Clinical Excellence health technology appraisal process. A generic MCDA modeling approach is described, and the different MCDA modeling approaches are applied to a hypothetical case study. A comparison of the different MCDA approaches is provided, and the generic issues that need consideration before the application of MCDA in health technology assessment are examined. There are general practical issues that might arise from using an MCDA approach, and it is suggested that appropriate care be taken to ensure the success of MCDA techniques in the appraisal process. Copyright © 2012 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.

  1. Applicability of digital analysis and imaging technology in neuropathology assessment.

    Science.gov (United States)

    Dunn, William D; Gearing, Marla; Park, Yuna; Zhang, Lifan; Hanfelt, John; Glass, Jonathan D; Gutman, David A

    2016-06-01

    Alzheimer's disease (AD) is a progressive neurological disorder that affects more than 30 million people worldwide. While various dementia-related losses in cognitive functioning are its hallmark clinical symptoms, ultimate diagnosis is based on manual neuropathological assessments using various schemas, including Braak staging, CERAD (Consortium to Establish a Registry for Alzheimer's Disease) and Thal phase scoring. Since these scoring systems are based on subjective assessment, there is inevitably some degree of variation between readers, which could affect ultimate neuropathology diagnosis. Here, we report a pilot study investigating the applicability of computer-driven image analysis for characterizing neuropathological features, as well as its potential to supplement or even replace manually derived ratings commonly performed in medical settings. In this work, we quantitatively measured amyloid beta (Aβ) plaque in various brain regions from 34 patients using a robust digital quantification algorithm. We next verified these digitally derived measures to the manually derived pathology ratings using correlation and ordinal logistic regression methods, while also investigating the association with other AD-related neuropathology scoring schema commonly used at autopsy, such as Braak and CERAD. In addition to successfully verifying our digital measurements of Aβ plaques with respective categorical measurements, we found significant correlations with most AD-related scoring schemas. Our results demonstrate the potential for digital analysis to be adapted to more complex staining procedures commonly used in neuropathological diagnosis. As the efficiency of scanning and digital analysis of histology images increases, we believe that the basis of our semi-automatic approach may better standardize quantification of neuropathological changes and AD diagnosis, ultimately leading to a more comprehensive understanding of neurological disorders and more efficient patient

  2. Uncertainty Analysis of the Estimated Risk in Formal Safety Assessment

    Directory of Open Access Journals (Sweden)

    Molin Sun

    2018-01-01

    Full Text Available An uncertainty analysis is required to be carried out in formal safety assessment (FSA by the International Maritime Organization. The purpose of this article is to introduce the uncertainty analysis technique into the FSA process. Based on the uncertainty identification of input parameters, probability and possibility distributions are used to model the aleatory and epistemic uncertainties, respectively. An approach which combines the Monte Carlo random sampling of probability distribution functions with the a-cuts for fuzzy calculus is proposed to propagate the uncertainties. One output of the FSA process is societal risk (SR, which can be evaluated in the two-dimensional frequency–fatality (FN diagram. Thus, the confidence-level-based SR is presented to represent the uncertainty of SR in two dimensions. In addition, a method for time window selection is proposed to estimate the magnitude of uncertainties, which is an important aspect of modeling uncertainties. Finally, a case study is carried out on an FSA study on cruise ships. The results show that the uncertainty analysis of SR generates a two-dimensional area for a certain degree of confidence in the FN diagram rather than a single FN curve, which provides more information to authorities to produce effective risk control measures.

  3. Time-dependent reliability analysis and condition assessment of structures

    Energy Technology Data Exchange (ETDEWEB)

    Ellingwood, B.R. [Johns Hopkins Univ., Baltimore, MD (United States)

    1997-01-01

    Structures generally play a passive role in assurance of safety in nuclear plant operation, but are important if the plant is to withstand the effect of extreme environmental or abnormal events. Relative to mechanical and electrical components, structural systems and components would be difficult and costly to replace. While the performance of steel or reinforced concrete structures in service generally has been very good, their strengths may deteriorate during an extended service life as a result of changes brought on by an aggressive environment, excessive loading, or accidental loading. Quantitative tools for condition assessment of aging structures can be developed using time-dependent structural reliability analysis methods. Such methods provide a framework for addressing the uncertainties attendant to aging in the decision process.

  4. Confirmatory Factor Analysis of the Procrastination Assessment Scale for Students

    Directory of Open Access Journals (Sweden)

    Ronald D. Yockey

    2015-10-01

    Full Text Available The relative fit of one- and two-factor models of the Procrastination Assessment Scale for Students (PASS was investigated using confirmatory factor analysis on an ethnically diverse sample of 345 participants. The results indicated that although the two-factor model provided better fit to the data than the one-factor model, neither model provided optimal fit. However, a two-factor model which accounted for common item theme pairs used by Solomon and Rothblum in the creation of the scale provided good fit to the data. In addition, a significant difference by ethnicity was also found on the fear of failure subscale of the PASS, with Whites having significantly lower scores than Asian Americans or Latino/as. Implications of the results are discussed and recommendations made for future work with the scale.

  5. Hybrid vehicle assessment. Phase 1: Petroleum savings analysis

    Science.gov (United States)

    Levin, R.; Liddle, S.; Deshpande, G.; Trummel, M.; Vivian, H. C.

    1984-01-01

    The results of a comprehensive analysis of near term electric hybrid vehicles are presented, with emphasis on their potential to save significant amounts of petroleum on a national scale in the 1990s. Performance requirements and expected annual usage patterns of these vehicles are first modeled. The projected U.S. fleet composition is estimated, and conceptual hybrid vehicle designs are conceived and analyzed for petroleum use when driven in the expected annual patterns. These petroleum consumption estimates are then compared to similar estimates for projected 1990 conventional vehicles having the same performance and driven in the same patterns. Results are presented in the form of three utility functions and comparisons of sevral conceptual designs are made. The Hybrid Vehicle (HV) design and assessment techniques are discussed and a general method is explained for selecting the optimum energy management strategy for any vehicle mission battery combination. Conclusions and recommendations are presented, and development recommendations are identified.

  6. Assessing farming eco-efficiency: a Data Envelopment Analysis approach.

    Science.gov (United States)

    Picazo-Tadeo, Andrés J; Gómez-Limón, José A; Reig-Martínez, Ernest

    2011-04-01

    This paper assesses farming eco-efficiency using Data Envelopment Analysis (DEA) techniques. Eco-efficiency scores at both farm and environmental pressure-specific levels are computed for a sample of Spanish farmers operating in the rain-fed agricultural system of Campos County. The determinants of eco-efficiency are then studied using truncated regression and bootstrapping techniques. We contribute to previous literature in this field of research by including information on slacks in the assessment of the potential environmental pressure reductions in a DEA framework. Our results reveal that farmers are quite eco-inefficient, with very few differences emerging among specific environmental pressures. Moreover, eco-inefficiency is closely related to technical inefficiencies in the management of inputs. Regarding the determinants of eco-efficiency, farmers benefiting from agri-environmental programs as well as those with university education are found to be more eco-efficient. Concerning the policy implications of these results, public expenditure in agricultural extension and farmer training could be of some help to promote integration between farming and the environment. Furthermore, Common Agricultural Policy agri-environmental programs are an effective policy to improve eco-efficiency, although some doubts arise regarding their cost-benefit balance. Copyright © 2010 Elsevier Ltd. All rights reserved.

  7. Cyber threat impact assessment and analysis for space vehicle architectures

    Science.gov (United States)

    McGraw, Robert M.; Fowler, Mark J.; Umphress, David; MacDonald, Richard A.

    2014-06-01

    This paper covers research into an assessment of potential impacts and techniques to detect and mitigate cyber attacks that affect the networks and control systems of space vehicles. Such systems, if subverted by malicious insiders, external hackers and/or supply chain threats, can be controlled in a manner to cause physical damage to the space platforms. Similar attacks on Earth-borne cyber physical systems include the Shamoon, Duqu, Flame and Stuxnet exploits. These have been used to bring down foreign power generation and refining systems. This paper discusses the potential impacts of similar cyber attacks on space-based platforms through the use of simulation models, including custom models developed in Python using SimPy and commercial SATCOM analysis tools, as an example STK/SOLIS. The paper discusses the architecture and fidelity of the simulation model that has been developed for performing the impact assessment. The paper walks through the application of an attack vector at the subsystem level and how it affects the control and orientation of the space vehicle. SimPy is used to model and extract raw impact data at the bus level, while STK/SOLIS is used to extract raw impact data at the subsystem level and to visually display the effect on the physical plant of the space vehicle.

  8. The use of error analysis to assess resident performance.

    Science.gov (United States)

    D'Angelo, Anne-Lise D; Law, Katherine E; Cohen, Elaine R; Greenberg, Jacob A; Kwan, Calvin; Greenberg, Caprice; Wiegmann, Douglas A; Pugh, Carla M

    2015-11-01

    The aim of this study was to assess validity of a human factors error assessment method for evaluating resident performance during a simulated operative procedure. Seven postgraduate year 4-5 residents had 30 minutes to complete a simulated laparoscopic ventral hernia (LVH) repair on day 1 of a national, advanced laparoscopic course. Faculty provided immediate feedback on operative errors and residents participated in a final product analysis of their repairs. Residents then received didactic and hands-on training regarding several advanced laparoscopic procedures during a lecture session and animate lab. On day 2, residents performed a nonequivalent LVH repair using a simulator. Three investigators reviewed and coded videos of the repairs using previously developed human error classification systems. Residents committed 121 total errors on day 1 compared with 146 on day 2. One of 7 residents successfully completed the LVH repair on day 1 compared with all 7 residents on day 2 (P = .001). The majority of errors (85%) committed on day 2 were technical and occurred during the last 2 steps of the procedure. There were significant differences in error type (P ≤ .001) and level (P = .019) from day 1 to day 2. The proportion of omission errors decreased from day 1 (33%) to day 2 (14%). In addition, there were more technical and commission errors on day 2. The error assessment tool was successful in categorizing performance errors, supporting known-groups validity evidence. Evaluating resident performance through error classification has great potential in facilitating our understanding of operative readiness. Copyright © 2015 Elsevier Inc. All rights reserved.

  9. Bivariate analysis of floods in climate impact assessments.

    Science.gov (United States)

    Brunner, Manuela Irene; Sikorska, Anna E; Seibert, Jan

    2018-03-01

    Climate impact studies regarding floods usually focus on peak discharges and a bivariate assessment of peak discharges and hydrograph volumes is not commonly included. A joint consideration of peak discharges and hydrograph volumes, however, is crucial when assessing flood risks for current and future climate conditions. Here, we present a methodology to develop synthetic design hydrographs for future climate conditions that jointly consider peak discharges and hydrograph volumes. First, change factors are derived based on a regional climate model and are applied to observed precipitation and temperature time series. Second, the modified time series are fed into a calibrated hydrological model to simulate runoff time series for future conditions. Third, these time series are used to construct synthetic design hydrographs. The bivariate flood frequency analysis used in the construction of synthetic design hydrographs takes into account the dependence between peak discharges and hydrograph volumes, and represents the shape of the hydrograph. The latter is modeled using a probability density function while the dependence between the design variables peak discharge and hydrograph volume is modeled using a copula. We applied this approach to a set of eight mountainous catchments in Switzerland to construct catchment-specific and season-specific design hydrographs for a control and three scenario climates. Our work demonstrates that projected climate changes have an impact not only on peak discharges but also on hydrograph volumes and on hydrograph shapes both at an annual and at a seasonal scale. These changes are not necessarily proportional which implies that climate impact assessments on future floods should consider more flood characteristics than just flood peaks. Copyright © 2017. Published by Elsevier B.V.

  10. Environmental analysis of waste foundry sand via life cycle assessment.

    Science.gov (United States)

    Mitterpach, Jozef; Hroncová, Emília; Ladomerský, Juraj; Balco, Karol

    2017-01-01

    The aim of this manuscript is to provide an environmental assessment of the creation and use of waste foundry sand (WFS) via an LCA in a foundry for grey cast iron. A life cycle impact assessment was carried out using SimaPro 8. This environmental analysis assessed the impact of creating waste foundry sand (WFS) in a foundry, Hronec (Slovakia, Central Europe). According to BREF, this foundry is classified as an iron foundry with a production capacity greater than 20 t/day with processes typical for grey cast iron foundries. Molten metal is poured into single-use sand moulds. We identified those factors influencing the creation and use of WFS which significantly affect the quality of the environment. The use of WFS from the production of cores in regenerated moulding mixtures with installed circuits brings marked minimisation of material and energy inputs in the processes of creating WFS and it positively influences the consumption of resources and the quality of the ecosystem. Space for lessening the impact of WFS processes upon the consumption of resources and ecosystem quality is mainly found in recycling WFS in the building sector. In the next step, it is necessary to thoroughly verify the eco-toxicological properties of not only the created WFS and other foundry waste, but mainly the building products for which this waste is used. In terms of transportation, it is important that waste is recycled at local level. The processes of creating WFS have a marked influence upon all the selected waste categories (consumption of resources, ecosystem quality, human health). By minimising material inputs into processes and the effective adjustment of production technology, a foundry can significantly lessen the impacts of processes for creating WFS upon the environment.

  11. 78 FR 39284 - Technical Guidance for Assessing Environmental Justice in Regulatory Analysis

    Science.gov (United States)

    2013-07-01

    ... entitled, ``Technical Guidance for Assessing Environmental Justice in Regulatory Analysis.'' The purpose of... Technical Guidance for Assessing Environmental Justice in Regulatory Analysis Docket, EPA/DC, EPA West, Room... Technical Guidance for Assessing Environmental Justice in Regulatory Analysis is available in the public...

  12. Assessing student teachers' reflective writing through quantitative content analysis

    NARCIS (Netherlands)

    Poldner, Eric; Van der Schaaf, Marieke; Simons, P. Robert Jan; Van Tartwijk, Jan; Wijngaards, Guus

    2014-01-01

    Students' reflective essay writing can be stimulated by the formative assessments provided to them by their teachers. Such assessments contain information about the quality of students' reflective writings and offer suggestions for improvement. Despite the importance of formatively assessing

  13. Assessment of handedness using latent class factor analysis.

    Science.gov (United States)

    Merni, Franco; Di Michele, Rocco; Soffritti, Gabriele

    2014-01-01

    Recently several studies in which handedness was evaluated as a latent construct have been performed. In those studies, handedness was modelled using a qualitative latent variable (latent class models), a continuous latent variable (factor models), or both a qualitative latent variable and a continuous latent trait (mixed Rasch models). The aim of this study was to explore the usefulness and effectiveness of an approach in which handedness is treated as a qualitatively scaled latent variable with ordered categories (latent class factor models). This aim was pursued through an exploratory analysis of a dataset containing information on the hand used by 2236 young Italian sportspeople to perform 10 tasks. For comparison purposes, a latent class analysis was carried out. A cross-validation procedure was implemented. The results of all the analyses revealed that the best fit to the observed handedness patterns was obtained using a latent class factor model. Through this model, individuals were assigned to one of four ordered levels of handedness, and a quantitative index of left-handedness for each individual was computed by taking into account the different effect of the 10 tasks. These results provide support for the use of the latent class factor approach for handedness assessment.

  14. Integrated analysis environment for the Movement Assessment Battery for Children

    Directory of Open Access Journals (Sweden)

    Carlos Norberto Fischer

    2013-12-01

    Full Text Available Developmental Coordination Disorder (DCD, a chronic and usually permanent condition found in children, is characterized by motor impairment that interferes with a child's activities of daily living and with academic achievement. One of the most popular tests for the quantitative diagnosis of DCD is the Movement Assessment Battery for Children (MABC. Based on the Battery's standardized scores, it is possible to identify children with typical development, children at risk of developing DCD, and children with DCD. This article describes a computational system we developed to assist with the analysis of results obtained in the MABC test. The tool was developed for the web environment and its database provides integration of MABC data. Thus, researchers around the world can share data and develop collaborative work in the DCD field. In order to help analysis processes, our system provides services for filtering data to show more specific sets of information and present the results in textual, table, and graphic formats, allowing easier and more comprehensive evaluation of the results.

  15. Lake Michigan Wind Assessment Analysis, 2012 and 2013

    Directory of Open Access Journals (Sweden)

    Charles R Standridge

    2017-03-01

    Full Text Available A study was conducted to address the wind energy potential over Lake Michigan to support a commercial wind farm.  Lake Michigan is an inland sea in the upper mid-western United States.  A laser wind sensor mounted on a floating platform was located at the mid-lake plateau in 2012 and about 10.5 kilometers from the eastern shoreline near Muskegon Michigan in 2013.  Range gate heights for the laser wind sensor were centered at 75, 90, 105, 125, 150, and 175 meters.  Wind speed and direction were measured once each second and aggregated into 10 minute averages.  The two sample t-test and the paired-t method were used to perform the analysis.  Average wind speed stopped increasing between 105 m and 150 m depending on location.  Thus, the collected data is inconsistent with the idea that average wind speed increases with height. This result implies that measuring wind speed at wind turbine hub height is essential as opposed to using the wind energy power law to project the wind speed from lower heights.  Average speed at the mid-lake plateau is no more that 10% greater than at the location near Muskegon.  Thus, it may be possible to harvest much of the available wind energy at a lower height and closer to the shoreline than previously thought.  At both locations, the predominate wind direction is from the south-southwest.  The ability of the laser wind sensor to measure wind speed appears to be affected by a lack of particulate matter at greater heights.   Keywords: wind assessment, Lake Michigan, LIDAR wind sensor, statistical analysis. Article History: Received June 15th 2016; Received in revised form January 16th 2017; Accepted February 2nd 2017 Available online How to Cite This Article: Standridge, C., Zeitler, D., Clark, A., Spoelma, T., Nordman, E., Boezaart, T.A., Edmonson, J.,  Howe, G., Meadows, G., Cotel, A. and Marsik, F. (2017 Lake Michigan Wind Assessment Analysis, 2012 and 2013. Int. Journal of Renewable Energy Development

  16. Royal London space analysis: plaster versus digital model assessment.

    Science.gov (United States)

    Grewal, Balpreet; Lee, Robert T; Zou, Lifong; Johal, Ama

    2017-06-01

    With the advent of digital study models, the importance of being able to evaluate space requirements becomes valuable to treatment planning and the justification for any required extraction pattern. This study was undertaken to compare the validity and reliability of the Royal London space analysis (RLSA) undertaken on plaster as compared with digital models. A pilot study (n = 5) was undertaken on plaster and digital models to evaluate the feasibility of digital space planning. This also helped to determine the sample size calculation and as a result, 30 sets of study models with specified inclusion criteria were selected. All five components of the RLSA, namely: crowding; depth of occlusal curve; arch expansion/contraction; incisor antero-posterior advancement and inclination (assessed from the pre-treatment lateral cephalogram) were accounted for in relation to both model types. The plaster models served as the gold standard. Intra-operator measurement error (reliability) was evaluated along with a direct comparison of the measured digital values (validity) with the plaster models. The measurement error or coefficient of repeatability was comparable for plaster and digital space analyses and ranged from 0.66 to 0.95mm. No difference was found between the space analysis performed in either the upper or lower dental arch. Hence, the null hypothesis was accepted. The digital model measurements were consistently larger, albeit by a relatively small amount, than the plaster models (0.35mm upper arch and 0.32mm lower arch). No difference was detected in the RLSA when performed using either plaster or digital models. Thus, digital space analysis provides a valid and reproducible alternative method in the new era of digital records.

  17. Nurses' critical event risk assessments: a judgement analysis.

    Science.gov (United States)

    Thompson, Carl; Bucknall, Tracey; Estabrookes, Carole A; Hutchinson, Alison; Fraser, Kim; de Vos, Rien; Binnecade, Jan; Barrat, Gez; Saunders, Jane

    2009-02-01

    To explore and explain nurses' use of readily available clinical information when deciding whether a patient is at risk of a critical event. Half of inpatients who suffer a cardiac arrest have documented but unacted upon clinical signs of deterioration in the 24 hours prior to the event. Nurses appear to be both misinterpreting and mismanaging the nursing-knowledge 'basics' such as heart rate, respiratory rate and oxygenation. Whilst many medical interventions originate from nurses, up to 26% of nurses' responses to abnormal signs result in delays of between one and three hours. A double system judgement analysis using Brunswik's lens model of cognition was undertaken with 245 Dutch, UK, Canadian and Australian acute care nurses. Nurses were asked to judge the likelihood of a critical event, 'at-risk' status, and whether they would intervene in response to 50 computer-presented clinical scenarios in which data on heart rate, systolic blood pressure, urine output, oxygen saturation, conscious level and oxygenation support were varied. Nurses were also presented with a protocol recommendation and also placed under time pressure for some of the scenarios. The ecological criterion was the predicted level of risk from the Modified Early Warning Score assessments of 232 UK acute care inpatients. Despite receiving identical information, nurses varied considerably in their risk assessments. The differences can be partly explained by variability in weightings given to information. Time and protocol recommendations were given more weighting than clinical information for key dichotomous choices such as classifying a patient as 'at risk' and deciding to intervene. Nurses' weighting of cues did not mirror the same information's contribution to risk in real patients. Nurses synthesized information in non-linear ways that contributed little to decisional accuracy. The low-moderate achievement (R(a)) statistics suggests that nurses' assessments of risk were largely inaccurate

  18. Imminent cardiac risk assessment via optical intravascular biochemical analysis.

    Science.gov (United States)

    Wetzel, David L; Wetzel, Louis H; Wetzel, Mark D; Lodder, Robert A

    2009-06-01

    Heart disease is by far the biggest killer in the United States, and type II diabetes, which affects 8% of the U.S. population, is on the rise. In many cases, the acute coronary syndrome and/or sudden cardiac death occurs without warning. Atherosclerosis has known behavioral, genetic and dietary risk factors. However, our laboratory studies with animal models and human post-mortem tissue using FT-IR microspectroscopy reveal the chemical microstructure within arteries and in the arterial walls themselves. These include spectra obtained from the aortas of ApoE-/- knockout mice on sucrose and normal diets showing lipid deposition in the former case. Also pre-aneurysm chemical images of knockout mouse aorta walls, and spectra of plaque excised from a living human patient are shown for comparison. In keeping with the theme of the SPEC 2008 conference 'Spectroscopic Diagnosis of Disease...' this paper describes the background and potential value of a new catheter-based system to provide in vivo biochemical analysis of plaque in human coronary arteries. We report the following: (1) results of FT-IR microspectroscopy on animal models of vascular disease to illustrate the localized chemical distinctions between pathological and normal tissue, (2) current diagnostic techniques used for risk assessment of patients with potential unstable coronary syndromes, and (3) the advantages and limitations of each of these techniques illustrated with patent care histories, related in the first person, by the physician coauthors. Note that the physician comments clarify the contribution of each diagnostic technique to imminent cardiac risk assessment in a clinical setting, leading to the appreciation of what localized intravascular chemical analysis can contribute as an add-on diagnostic tool. The quality of medical imaging has improved dramatically since the turn of the century. Among clinical non-invasive diagnostic tools, laboratory tests of body fluids, EKG, and physical examination

  19. Imminent Cardiac Risk Assessment via Optical Intravascular Biochemical Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Wetzel, D.; Wetzel, L; Wetzel, M; Lodder, R

    2009-01-01

    Heart disease is by far the biggest killer in the United States, and type II diabetes, which affects 8% of the U.S. population, is on the rise. In many cases, the acute coronary syndrome and/or sudden cardiac death occurs without warning. Atherosclerosis has known behavioral, genetic and dietary risk factors. However, our laboratory studies with animal models and human post-mortem tissue using FT-IR microspectroscopy reveal the chemical microstructure within arteries and in the arterial walls themselves. These include spectra obtained from the aortas of ApoE-/- knockout mice on sucrose and normal diets showing lipid deposition in the former case. Also pre-aneurysm chemical images of knockout mouse aorta walls, and spectra of plaque excised from a living human patient are shown for comparison. In keeping with the theme of the SPEC 2008 conference Spectroscopic Diagnosis of Disease this paper describes the background and potential value of a new catheter-based system to provide in vivo biochemical analysis of plaque in human coronary arteries. We report the following: (1) results of FT-IR microspectroscopy on animal models of vascular disease to illustrate the localized chemical distinctions between pathological and normal tissue, (2) current diagnostic techniques used for risk assessment of patients with potential unstable coronary syndromes, and (3) the advantages and limitations of each of these techniques illustrated with patent care histories, related in the first person, by the physician coauthors. Note that the physician comments clarify the contribution of each diagnostic technique to imminent cardiac risk assessment in a clinical setting, leading to the appreciation of what localized intravascular chemical analysis can contribute as an add-on diagnostic tool. The quality of medical imaging has improved dramatically since the turn of the century. Among clinical non-invasive diagnostic tools, laboratory tests of body fluids, EKG, and physical examination are

  20. Assessing computer waste generation in Chile using material flow analysis.

    Science.gov (United States)

    Steubing, Bernhard; Böni, Heinz; Schluep, Mathias; Silva, Uca; Ludwig, Christian

    2010-03-01

    The quantities of e-waste are expected to increase sharply in Chile. The purpose of this paper is to provide a quantitative data basis on generated e-waste quantities. A material flow analysis was carried out assessing the generation of e-waste from computer equipment (desktop and laptop PCs as well as CRT and LCD-monitors). Import and sales data were collected from the Chilean Customs database as well as from publications by the International Data Corporation. A survey was conducted to determine consumers' choices with respect to storage, re-use and disposal of computer equipment. The generation of e-waste was assessed in a baseline as well as upper and lower scenarios until 2020. The results for the baseline scenario show that about 10,000 and 20,000 tons of computer waste may be generated in the years 2010 and 2020, respectively. The cumulative e-waste generation will be four to five times higher in the upcoming decade (2010-2019) than during the current decade (2000-2009). By 2020, the shares of LCD-monitors and laptops will increase more rapidly replacing other e-waste including the CRT-monitors. The model also shows the principal flows of computer equipment from production and sale to recycling and disposal. The re-use of computer equipment plays an important role in Chile. An appropriate recycling scheme will have to be introduced to provide adequate solutions for the growing rate of e-waste generation. Copyright 2009 Elsevier Ltd. All rights reserved.

  1. Credibility assessment in child sexual abuse investigations: A descriptive analysis.

    Science.gov (United States)

    Melkman, Eran P; Hershkowitz, Irit; Zur, Ronit

    2017-05-01

    A major challenge in cases of child sexual abuse (CSA) is determining the credibility of children's reports. Consequently cases may be misclassified as false or deemed 'no judgment possible'. Based on a large national sample of reports of CSA made in Israel in 2014, the study examines child and event characteristics contributing to the probability that reports of abuse would be judged credible. National data files of all children aged 3-14, who were referred for investigation following suspected victimization of sexual abuse, and had disclosed sexual abuse, were analyzed. Cases were classified as either 'credible' or 'no judgment possible'. The probability of reaching a 'credible' judgment was examined in relation to characteristics of the child (age, gender, cognitive delay, marital status of the parents,) and of the abusive event (abuse severity, frequency, perpetrator-victim relationship, perpetrator's use of grooming, and perpetrator's use of coercion), controlling for investigator's identity at the cluster level of the analysis. Of 1563 cases analyzed, 57.9% were assessed as credible. The most powerful predictors of a credible judgment were older age and absence of a cognitive delay. Reports of children to married parents, who experienced a single abusive event that involved perpetrator's use of grooming, were also more likely to be judged as credible. Rates of credible judgments found are lower than expected suggesting under-identification of truthful reports of CSA. In particular, those cases of severe and multiple abuse involving younger and cognitively delayed children are the ones with the lowest chances of being assessed as credible. Copyright © 2017 Elsevier Ltd. All rights reserved.

  2. Modern psychometrics for assessing achievement goal orientation: a Rasch analysis.

    Science.gov (United States)

    Muis, Krista R; Winne, Philip H; Edwards, Ordene V

    2009-09-01

    A program of research is needed that assesses the psychometric properties of instruments designed to quantify students' achievement goal orientations to clarify inconsistencies across previous studies and to provide a stronger basis for future research. We conducted traditional psychometric and modern Rasch-model analyses of the Achievement Goals Questionnaire (AGQ, Elliot & McGregor, 2001) and the Patterns of Adaptive Learning Scale (PALS, Midgley et al., 2000) to provide an in-depth analysis of the two most popular instruments in educational psychology. For Study 1, 217 undergraduate students enrolled in educational psychology courses participated. Thirty-four were male and 181 were female (two did not respond). Participants completed the AGQ in the context of their educational psychology class. For Study 2, 126 undergraduate students enrolled in educational psychology courses participated. Thirty were male and 95 were female (one did not respond). Participants completed the PALS in the context of their educational psychology class. Traditional psychometric assessments of the AGQ and PALS replicated previous studies. For both, reliability estimates ranged from good to very good for raw subscale scores and fit for the models of goal orientations were good. Based on traditional psychometrics, the AGQ and PALS are valid and reliable indicators of achievement goals. Rasch analyses revealed that estimates of reliability for items were very good but respondent ability estimates varied from poor to good for both the AGQ and PALS. These findings indicate that items validly and reliably reflect a group's aggregate goal orientation, but using either instrument to characterize an individual's goal orientation is hazardous.

  3. Newborn infant pain assessment using heart rate variability analysis.

    Science.gov (United States)

    Faye, Papa M; De Jonckheere, Julien; Logier, Regis; Kuissi, Eliane; Jeanne, Mathieu; Rakza, Thameur; Storme, Laurent

    2010-01-01

    Systems controlling cardiovascular function are closely coupled with the perception of pain. Heart rate variability (HRV) is a well-established noninvasive measure of cardiac autonomic control. We hypothesized that pain may alter HRV in the newborn infant and that HRV analysis could be used as an indicator of prolonged pain in the newborn infant. To test the hypothesis, we measured the magnitude of the heart rate high-frequency variations using an innovative High Frequency Variability Index (HFVI) in newborn infants at risk of postoperative pain. We investigated newborn infants with a gestational age (GA) more than 34 weeks, and who were admitted after a major surgical procedure. Inclusions ranged from 2 to 72 hours after the surgery. The postoperative pain was scored using EDIN scale (neonatal pain and discomfort scale) at the end of the 2 hours recording period. The infants were separated in: (1) Group "Low EDIN," when EDIN=5. Predictive positive and negative values of a threshold value of HFVI in assessing pain have been studied. Twenty-eight newborn infants were enrolled in the study (mean GA=37.8+/-1.5 wk) at a median delay between the surgery and the recording of 5 hours. Mean EDIN were 2+/-1 and 7+/-2 in respectively the groups "Low EDIN" and "High EDIN." The 2 groups were similar for GA, basal heart and respiratory rates, SpO2, mean arterial blood pressure, and morphine infusion rate. HFVI was significantly lower in the group "High EDIN" than in the group "Low EDIN" (0.7+/-0.2 vs. 1.2+/-0.3, respectively; P=5, with a sensitivity of 90%, and a specificity of 75%. The results of this study indicate that postoperative pain is associated with a decreased high-frequency HRV in full-term newborn infants. Our findings suggest that HRV could be used as an indicator to assess prolonged pain in the newborn infants.

  4. Robotic Mars Sample Return: Risk Assessment and Analysis Report

    Science.gov (United States)

    Lalk, Thomas R.; Spence, Cliff A.

    2003-01-01

    A comparison of the risk associated with two alternative scenarios for a robotic Mars sample return mission was conducted. Two alternative mission scenarios were identified, the Jet Propulsion Lab (JPL) reference Mission and a mission proposed by Johnson Space Center (JSC). The JPL mission was characterized by two landers and an orbiter, and a Mars orbit rendezvous to retrieve the samples. The JSC mission (Direct/SEP) involves a solar electric propulsion (SEP) return to earth followed by a rendezvous with the space shuttle in earth orbit. A qualitative risk assessment to identify and characterize the risks, and a risk analysis to quantify the risks were conducted on these missions. Technical descriptions of the competing scenarios were developed in conjunction with NASA engineers and the sequence of events for each candidate mission was developed. Risk distributions associated with individual and combinations of events were consolidated using event tree analysis in conjunction with Monte Carlo techniques to develop probabilities of mission success for each of the various alternatives. The results were the probability of success of various end states for each candidate scenario. These end states ranged from complete success through various levels of partial success to complete failure. Overall probability of success for the Direct/SEP mission was determined to be 66% for the return of at least one sample and 58% for the JPL mission for the return of at least one sample cache. Values were also determined for intermediate events and end states as well as for the probability of violation of planetary protection. Overall mission planetary protection event probabilities of occurrence were determined to be 0.002% and 1.3% for the Direct/SEP and JPL Reference missions respectively.

  5. Life Assessment of Steam Turbine Components Based on Viscoplastic Analysis

    Science.gov (United States)

    Choi, Woo-Sung; Fleury, Eric; Kim, Bum-Shin; Hyun, Jung-Seob

    Unsteady thermal and mechanical loading in turbine components is caused due to the transient regimes arising during start-ups and shut-downs and due to changes in the operating regime in steam power plants; this results in nonuniform strain and stress distribution. Thus, an accurate knowledge of the stresses caused by various loading conditions is required to ensure the integrity and to ensure an accurate life assessment of the components of a turbine. Although the materials of the components of the steam turbine deform inelastically at a high temperature, currently, only elastic calculations are performed for safety and simplicity. Numerous models have been proposed to describe the viscoplastic (time-dependent) behavior; these models are rather elaborate and it is difficult to incorporate them into a finite element code in order to simulate the loading of complex structures. In this paper, the total lifetime of the components of a steam turbine was calculated by combining the viscoplastic constitutive equation with the ABAQUS finite element code. Viscoplastic analysis was conducted by focusing mainly on simplified constitutive equations with linear kinematic hardening, which is simple enough to be used effectively in computer simulation. The von Mises stress distribution of an HIP turbine rotor was calculated during the cold start-up operation of the rotor, and a reasonable number of cycles were obtained from the equation of Langer.

  6. Assessment of river Po sediment quality by micropollutant analysis.

    Science.gov (United States)

    Camusso, Marina; Galassi, Silvana; Vignati, Davide

    2002-05-01

    Trace metals, PCB congeners and DDT homologues were determined in composite sediment samples collected from 10 representative sites along the river Po in two separate seasons. The aim was to identify the most anthropogenically impacted areas for future monitoring programmes and to aid development of Italian sediment quality criteria. The surface samples were collected during low flow conditions. Trace metal concentrations were assayed by electrothermal (Cd, Co, Cr, Cu, Ni, Pb), flame (Fe, Mn, Zn) or hydride generation (As) atomic absorption spectrometry after microwave assisted acid digestion. Hg was determined on solid samples by automated analyser. Organic microcontaminants were determined by gas-chromatography with 63Ni electron capture detector after Soxhlet extraction. Concentrations of trace metals, total PCB and DDT homologues showed two distinct peaks at the sites immediately downstream of Turin and Milan, respectively, and in each case decreased progressively further downstream. Principal component analysis identified three major factors (from a multi-dimensional space of 35 variables) which explained 85-90% of the total observed variance. The first and second factors corresponded to anthropogenic inputs and geological factors on sediment quality; the third included seasonal processes of minor importance. Sediment quality assessment identified Cd, Cu, Hg, Pb, Zn and organic microcontaminants as posing the most serious threats to river sediment quality. A reference site within the Po basin provided useful background values. Moderate pollution by organochlorine compounds was ascribed both to local sources and to atmospheric deposition.

  7. Assessing state-level active living promotion using network analysis.

    Science.gov (United States)

    Buchthal, Opal Vanessa; Taniguchi, Nicole; Iskandar, Livia; Maddock, Jay

    2013-01-01

    Physical inactivity is a growing problem in the United States, one that is being addressed through the development of active living communities. However, active living promotion requires collaboration among organizations that may not have previously shared goals. A network analysis was conducted to assess Hawaii's active living promotion network. Twenty-six organizations playing a significant role in promoting active living in Hawaii were identified and surveyed about their frequency of contact, level of collaboration, and funding flow with other agencies. A communication network was identified linking all agencies. This network had many long pathways, impeding information flow. The Department of Health (DOH) and the State Nutrition and Physical Activity Coalition (NPAC) were central nodes, but DOH connected state agencies while NPAC linked county and voluntary organizations. Within the network, information sharing was common, but collaboration and formal partnership were low. Linkages between county and state agencies, between counties, and between state agencies with different core agendas were particularly low. Results suggest that in the early stages of development, active living networks may be divided by geography and core missions, requiring work to bridge these divides. Network mapping appears helpful in identifying areas for network development.

  8. Landslides geotechnical analysis. Qualitative assessment by valuation factors

    Science.gov (United States)

    Cuanalo Oscar, Sc D.; Oliva Aldo, Sc D.; Polanco Gabriel, M. E.

    2012-04-01

    In general, a landslide can cause a disaster when it is combined a number of factors such as an extreme event related to a geological phenomenon, vulnerable elements exposed in a specific geographic area, and the probability of loss and damage evaluated in terms of lives and economic assets, in a certain period of time. This paper presents the qualitative evaluation of slope stability through of Valuation Factors, obtained from the characterization of the determinants and triggers factors that influence the instability; for the first the morphology and topography, geology, soil mechanics, hydrogeology and vegetation to the second, the rain, earthquakes, erosion and scour, human activity, and ultimately dependent factors of the stability analysis, and its influence ranges which greatly facilitate the selection of construction processes best suited to improve the behavior of a slope or hillside. The Valuation Factors are a set of parameters for assessing the influence of conditioning and triggering factors that influence the stability of slopes and hillsides. The characteristics of each factor must be properly categorized to involve its effect on behavior; a way to do this is by assigning a weighted value range indicating its effect on the stability of a slope. It is proposed to use Valuation Factors with weighted values between 0 and 1 (arbitrarily selected but common sense and logic), the first corresponds to no or minimal effect on stability (no effect or very little influence) and the second, the greatest impact on it (has a significant influence). The meddle effects are evaluated with intermediate values.

  9. Life Cycle Assessment and Cost Analysis of Water and ...

    Science.gov (United States)

    changes in drinking and wastewater infrastructure need to incorporate a holistic view of the water service sustainability tradeoffs and potential benefits when considering shifts towards new treatment technology, decentralized systems, energy recovery and reuse of treated wastewater. The main goal of this study is to determine the influence of scale on the energy and cost performance of different transitional membrane bioreactors (MBR) in decentralized wastewater treatment (WWT) systems by performing a life cycle assessment (LCA) and cost analysis. LCA is a tool used to quantify sustainability-related metrics from a systems perspective. The study calculates the environmental and cost profiles of both aerobic MBRs (AeMBR) and anaerobic MBRs (AnMBR), which not only recover energy from waste, but also produce recycled water that can displace potable water for uses such as irrigation and toilet flushing. MBRs represent an intriguing technology to provide decentralized WWT services while maximizing resource recovery. A number of scenarios for these WWT technologies are investigated for different scale systems serving various population density and land area combinations to explore the ideal application potentials. MBR systems are examined from 0.05 million gallons per day (MGD) to 10 MGD and serve land use types from high density urban (100,000 people per square mile) to semi-rural single family (2,000 people per square mile). The LCA and cost model was built with ex

  10. Hybrid vehicle assessment. Phase I. Petroleum savings analysis

    Energy Technology Data Exchange (ETDEWEB)

    Levin, R.; Liddle, S.; Deshpande, G.; Trummel, M.; Vivian, H.

    1984-03-01

    This report presents the results of a comprehensive analysis of near-term electric-hybrid vehicles. Its purpose was to estimate their potential to save significant amounts of petroleum on a national scale in the 1990s. Performance requirements and expected annual usage patterns of these vehicles were first modeled. The projected US fleet composition was estimated, and conceptual hybrid vehicle designs were conceived and analyzed for petroleum use when driven in the expected annual patterns. These petroleum consumption estimates were then compared to similar estimates for projected 1990 conventional vehicles having the same performance and driven in the same patterns. Results are presented in the form of three utility functions and comparisons of several conceptual designs are made. The Hybrid Vehicle (HV) design and assessment techniques are discussed and a general method is explained for selecting the optimum energy management strategy for any vehicle-mission-battery combination. A discussion of lessons learned during the construction and test of the General Electric Hybrid Test Vehicle is also presented. Conclusions and recommendations are presented, and development recommendations are identified.

  11. Direct Survival Analysis: a new stock assessment method

    Directory of Open Access Journals (Sweden)

    Eduardo Ferrandis

    2007-03-01

    Full Text Available In this work, a new stock assessment method, Direct Survival Analysis, is proposed and described. The parameter estimation of the Weibull survival model proposed by Ferrandis (2007 is obtained using trawl survey data. This estimation is used to establish a baseline survival function, which is in turn used to estimate the specific survival functions in the different cohorts considered through an adaptation of the separable model of the fishing mortality rates introduced by Pope and Shepherd (1982. It is thus possible to test hypotheses on the evolution of survival during the period studied and to identify trends in recruitment. A link is established between the preceding analysis of trawl survey data and the commercial catch-at-age data that are generally obtained to evaluate the population using analytical models. The estimated baseline survival, with the proposed versions of the stock and catch equations and the adaptation of the Separable Model, may be applied to commercial catch-at-age data. This makes it possible to estimate the survival corresponding to the landing data, the initial size of the cohort and finally, an effective age of first capture, in order to complete the parameter model estimation and consequently the estimation of the whole survival and mortality, along with the reference parameters that are useful for management purposes. Alternatively, this estimation of an effective age of first capture may be obtained by adapting the demographic structure of trawl survey data to that of the commercial fleet through suitable selectivity models of the commercial gears. The complete model provides the evaluation of the stock at any age. The coherence (and hence the mutual “calibration” between the two kinds of information may be analysed and compared with results obtained by other methods, such as virtual population analysis (VPA, in order to improve the diagnosis of the state of exploitation of the population. The model may be

  12. Sustainability in Higher Education : Analysis and Selection of Assessment Systems

    NARCIS (Netherlands)

    Maragakis, A.; van den Dobbelsteen, A.A.J.F.

    2015-01-01

    There is a noticeable increase in interest with regards to sustainability in higher education. As institutions investigate, implement and market sustainability efforts, there is a myriad of sustainability assessment methodologies currently available. Although these assessment systems were created

  13. Analysis of existing risk assessments, and list of suggestions

    CERN Document Server

    Heimsch, Laura

    2016-01-01

    The scope of this project was to analyse risk assessments made at CERN and extracting some crucial information about the different methodologies used, profiles of people who make the risk assessments, and gathering information of whether the risk matrix was used and if the acceptable level of risk was defined. Second step of the project was to trigger discussion inside HSE about risk assessment by suggesting a risk matrix and a risk assessment template.

  14. Risk Assessment and Integration Team (RAIT) Portfolio Risk Analysis Strategy

    Science.gov (United States)

    Edwards, Michelle

    2010-01-01

    Impact at management level: Qualitative assessment of risk criticality in conjunction with risk consequence, likelihood, and severity enable development of an "investment policy" towards managing a portfolio of risks. Impact at research level: Quantitative risk assessments enable researchers to develop risk mitigation strategies with meaningful risk reduction results. Quantitative assessment approach provides useful risk mitigation information.

  15. Image quality assessment based on multiscale geometric analysis.

    Science.gov (United States)

    Gao, Xinbo; Lu, Wen; Tao, Dacheng; Li, Xuelong

    2009-07-01

    Reduced-reference (RR) image quality assessment (IQA) has been recognized as an effective and efficient way to predict the visual quality of distorted images. The current standard is the wavelet-domain natural image statistics model (WNISM), which applies the Kullback-Leibler divergence between the marginal distributions of wavelet coefficients of the reference and distorted images to measure the image distortion. However, WNISM fails to consider the statistical correlations of wavelet coefficients in different subbands and the visual response characteristics of the mammalian cortical simple cells. In addition, wavelet transforms are optimal greedy approximations to extract singularity structures, so they fail to explicitly extract the image geometric information, e.g., lines and curves. Finally, wavelet coefficients are dense for smooth image edge contours. In this paper, to target the aforementioned problems in IQA, we develop a novel framework for IQA to mimic the human visual system (HVS) by incorporating the merits from multiscale geometric analysis (MGA), contrast sensitivity function (CSF), and the Weber's law of just noticeable difference (JND). In the proposed framework, MGA is utilized to decompose images and then extract features to mimic the multichannel structure of HVS. Additionally, MGA offers a series of transforms including wavelet, curvelet, bandelet, contourlet, wavelet-based contourlet transform (WBCT), and hybrid wavelets and directional filter banks (HWD), and different transforms capture different types of image geometric information. CSF is applied to weight coefficients obtained by MGA to simulate the appearance of images to observers by taking into account many of the nonlinearities inherent in HVS. JND is finally introduced to produce a noticeable variation in sensory experience. Thorough empirical studies are carried out upon the LIVE database against subjective mean opinion score (MOS) and demonstrate that 1) the proposed framework has

  16. Analysis Strategy for Fracture Assessment of Defects in Ductile Materials

    Energy Technology Data Exchange (ETDEWEB)

    Dillstroem, Peter; Andersson, Magnus; Sattari-Far, Iradj; Weilin Zang (Inspecta Technology AB, Stockholm (Sweden))

    2009-06-15

    The main purpose of this work is to investigate the significance of the residual stresses for defects (cracks) in ductile materials with nuclear applications, when the applied primary (mechanical) loads are high. The treatment of weld-induced stresses as expressed in the SACC/ProSACC handbook and other fracture assessment procedures such as the ASME XI code and the R6-method is believed to be conservative for ductile materials. This is because of the general approach not to account for the improved fracture resistance caused by ductile tearing. Furthermore, there is experimental evidence that the contribution of residual stresses to fracture diminishes as the degree of yielding increases to a high level. However, neglecting weld-induced stresses in general, though, is doubtful for loads that are mostly secondary (e.g. thermal shocks) and for materials which are not ductile enough to be limit load controlled. Both thin-walled and thick-walled pipes containing surface cracks are studied here. This is done by calculating the relative contribution from the weld residual stresses to CTOD and the J-integral. Both circumferential and axial cracks are analysed. Three different crack geometries are studied here by using the finite element method (FEM). (i) 2D axisymmetric modelling of a V-joint weld in a thin-walled pipe. (ii) 2D axisymmetric modelling of a V-joint weld in a thick-walled pipe. (iii) 3D modelling of a X-joint weld in a thick-walled pipe. t. Each crack configuration is analysed for two load cases; (1) Only primary (mechanical) loading is applied to the model, (2) Both secondary stresses and primary loading are applied to the model. Also presented in this report are some published experimental investigations conducted on cracked components of ductile materials subjected to both primary and secondary stresses. Based on the outcome of this study, an analysis strategy for fracture assessment of defects in ductile materials of nuclear components is proposed. A new

  17. Joint analysis of epistemic and aleatory uncertainty in stability analysis for geo-hazard assessments

    Science.gov (United States)

    Rohmer, Jeremy; Verdel, Thierry

    2017-04-01

    Uncertainty analysis is an unavoidable task of stability analysis of any geotechnical systems. Such analysis usually relies on the safety factor SF (if SF is below some specified threshold), the failure is possible). The objective of the stability analysis is then to estimate the failure probability P for SF to be below the specified threshold. When dealing with uncertainties, two facets should be considered as outlined by several authors in the domain of geotechnics, namely "aleatoric uncertainty" (also named "randomness" or "intrinsic variability") and "epistemic uncertainty" (i.e. when facing "vague, incomplete or imprecise information" such as limited databases and observations or "imperfect" modelling). The benefits of separating both facets of uncertainty can be seen from a risk management perspective because: - Aleatoric uncertainty, being a property of the system under study, cannot be reduced. However, practical actions can be taken to circumvent the potentially dangerous effects of such variability; - Epistemic uncertainty, being due to the incomplete/imprecise nature of available information, can be reduced by e.g., increasing the number of tests (lab or in site survey), improving the measurement methods or evaluating calculation procedure with model tests, confronting more information sources (expert opinions, data from literature, etc.). Uncertainty treatment in stability analysis usually restricts to the probabilistic framework to represent both facets of uncertainty. Yet, in the domain of geo-hazard assessments (like landslides, mine pillar collapse, rockfalls, etc.), the validity of this approach can be debatable. In the present communication, we propose to review the major criticisms available in the literature against the systematic use of probability in situations of high degree of uncertainty. On this basis, the feasibility of using a more flexible uncertainty representation tool is then investigated, namely Possibility distributions (e

  18. Assessment of academic departments efficiency using data envelopment analysis

    Directory of Open Access Journals (Sweden)

    Salah R. Agha

    2011-07-01

    Full Text Available Purpose: In this age of knowledge economy, universities play an important role in the development of a country. As government subsidies to universities have been decreasing, more efficient use of resources becomes important for university administrators. This study evaluates the relative technical efficiencies of academic departments at the Islamic University in Gaza (IUG during the years 2004-2006. Design/methodology/approach: This study applies Data Envelopment Analysis (DEA to assess the relative technical efficiency of the academic departments. The inputs are operating expenses, credit hours and training resources, while the outputs are number of graduates, promotions and public service activities. The potential improvements and super efficiency are computed for inefficient and efficient departments respectively. Further, multiple linear -regression is used to develop a relationship between super efficiency and input and output variables.Findings: Results show that the average efficiency score is 68.5% and that there are 10 efficient departments out of the 30 studied. It is noted that departments in the faculty of science, engineering and information technology have to greatly reduce their laboratory expenses. The department of economics and finance was found to have the highest super efficiency score among the efficient departments. Finally, it was found that promotions have the greatest contribution to the super efficiency scores while public services activities come next.Research limitations/implications: The paper focuses only on academic departments at a single university. Further, DEA is deterministic in nature.Practical implications: The findings offer insights on the inputs and outputs that significantly contribute to efficiencies so that inefficient departments can focus on these factors.Originality/value: Prior studies have used only one type of DEA (BCC and they did not explicitly answer the question posed by the inefficient

  19. Modeling the Assessment of Agricultural Enterprises Headcount Analysis

    OpenAIRE

    Viatkina, Tatyana

    2014-01-01

    The modern procedures of enterprises labour resources have been analyzed. The algorithm for calculation the enterprise performancepotential efficiency ratio and assessment of the performance potential of the enterprise assessment based on quantitativeand qualitative characteristics has been provided. The model for assessment the effectiveness of labour management of enterprise,branch or region subject to such factors as motivation, labour expenses, staff rotation and its qualifications has be...

  20. Assessing the validity of discourse analysis: transdisciplinary convergence

    Science.gov (United States)

    Jaipal-Jamani, Kamini

    2014-12-01

    Research studies using discourse analysis approaches make claims about phenomena or issues based on interpretation of written or spoken text, which includes images and gestures. How are findings/interpretations from discourse analysis validated? This paper proposes transdisciplinary convergence as a way to validate discourse analysis approaches to research. The argument is made that discourse analysis explicitly grounded in semiotics, systemic functional linguistics, and critical theory, offers a credible research methodology. The underlying assumptions, constructs, and techniques of analysis of these three theoretical disciplines can be drawn on to show convergence of data at multiple levels, validating interpretations from text analysis.

  1. Analysis of Management Behavior Assessments and Affect on Productivity

    National Research Council Canada - National Science Library

    Shipley, Jr, Steven E

    2005-01-01

    ... (Virtual Military Health Institute, 2003). The need exists for constructing a reliable behavior assessment instrument that captures data operationalized into correlational relationships between hospital management and employee beliefs...

  2. Concepts of Causality in Psychopathology: Applications in Clinical Assessment, Clinical Case Formulation and Functional Analysis

    NARCIS (Netherlands)

    Haynes, S.H.; O'Brien, W.H.; Kaholokula, J.K.; Witteman, C.L.M.

    2012-01-01

    This paper discusses and integrates concepts of causality in psychopathology, clinical assessment, clinical case formulation and the functional analysis. We propose that identifying causal variables, relations and mechanisms in psychopathology and clinical assessment can lead to more powerful and

  3. Assessment of the Assessment Tool: Analysis of Items in a Non-MCQ Mathematics Exam

    Science.gov (United States)

    Khoshaim, Heba Bakr; Rashid, Saima

    2016-01-01

    Assessment is one of the vital steps in the teaching and learning process. The reported action research examines the effectiveness of an assessment process and inspects the validity of exam questions used for the assessment purpose. The instructors of a college-level mathematics course studied questions used in the final exams during the academic…

  4. Assessing SRI fund performance research : best practices in empirical analysis

    NARCIS (Netherlands)

    Chegut, Andrea; Schenk, H.; Scholtens, B.

    2011-01-01

    We review the socially responsible investment (SRI) mutual fund performance literature to provide best practices in SRI performance attribution analysis. Based on meta-ethnography and content analysis, five themes in this literature require specific attention: data quality, social responsibility

  5. Conceptual risk assessment framework for global change risk analysis SRP

    CSIR Research Space (South Africa)

    Elphinstone, CD

    2007-12-01

    Full Text Available This report is submitted as a deliverable of the SRP project Global Change Risk Analysis which aims at applying risk analysis as a unifying notion for quantifying and communicating threats to ecosystem services originating from global change...

  6. Assessing the Validity of Discourse Analysis: Transdisciplinary Convergence

    Science.gov (United States)

    Jaipal-Jamani, Kamini

    2014-01-01

    Research studies using discourse analysis approaches make claims about phenomena or issues based on interpretation of written or spoken text, which includes images and gestures. How are findings/interpretations from discourse analysis validated? This paper proposes transdisciplinary convergence as a way to validate discourse analysis approaches to…

  7. Using multivariate statistical analysis to assess changes in water ...

    African Journals Online (AJOL)

    Abstract. Multivariate statistical analysis was used to investigate changes in water chemistry at 5 river sites in the Vaal Dam catch- ... analysis (CCA) showed that the environmental variables used in the analysis, discharge and month of sampling, explained ...... DINGENEN R, WILD O and ZENG G (2006) The global atmos-.

  8. Life Cycle Assessment Software for Product and Process Sustainability Analysis

    Science.gov (United States)

    Vervaeke, Marina

    2012-01-01

    In recent years, life cycle assessment (LCA), a methodology for assessment of environmental impacts of products and services, has become increasingly important. This methodology is applied by decision makers in industry and policy, product developers, environmental managers, and other non-LCA specialists working on environmental issues in a wide…

  9. An analysis of ESL teacher educators' current assessment practices ...

    African Journals Online (AJOL)

    Assessment is not an end in itself but a vehicle for educational improvement, and should be a fundamental and integral part of any curriculum based on student learning outcomes. The purpose of this paper is to analyse and identify any shortcomings in the current assessment practices of English Second Language teacher ...

  10. No-Reference Video Quality Assessment by HEVC Codec Analysis

    DEFF Research Database (Denmark)

    Huang, Xin; Søgaard, Jacob; Forchhammer, Søren

    2015-01-01

    the transform coefficients, estimates the distortion, and assesses the video quality. The proposed scheme generates VQA features based on Intra coded frames, and then maps features using an Elastic Net to predict subjective video quality. A set of HEVC coded 4K UHD sequences are tested. Results show......This paper proposes a No-Reference (NR) Video Quality Assessment (VQA) method for videos subject to the distortion given by High Efficiency Video Coding (HEVC). The proposed assessment can be performed either as a BitstreamBased (BB) method or as a Pixel-Based (PB). It extracts or estimates...

  11. The impact of inter-organizational alignment (IOA) on implementation outcomes: evaluating unique and shared organizational influences in education sector mental health.

    Science.gov (United States)

    Lyon, Aaron R; Whitaker, Kelly; Locke, Jill; Cook, Clayton R; King, Kevin M; Duong, Mylien; Davis, Chayna; Weist, Mark D; Ehrhart, Mark G; Aarons, Gregory A

    2018-02-07

    Integrated healthcare delivered by work groups in nontraditional service settings is increasingly common, yet contemporary implementation frameworks typically assume a single organization-or organizational unit-within which system-level processes influence service quality and implementation success. Recent implementation frameworks predict that inter-organizational alignment (i.e., similarity in values, characteristics, activities related to implementation across organizations) may facilitate the implementation of evidence-based practices (EBP), but few studies have evaluated this premise. This study's aims examine the impact of overlapping organizational contexts by evaluating the implementation contexts of externally employed mental health clinicians working in schools-the most common integrated service delivery setting for children and adolescents. Aim 1 is to estimate the effects of unique intra-organizational implementation contexts and combined inter-organizational alignment on implementation outcomes. Aim 2 is to examine the underlying mechanisms through which inter-organizational alignment facilitates or hinders EBP implementation. This study will conduct sequential, exploratory mixed-methods research to evaluate the intra- and inter-organizational implementation contexts of schools and the external community-based organizations that most often employ school-based mental health clinicians, as they relate to mental health EBP implementation. Aim 1 will involve quantitative surveys with school-based, externally-employed mental health clinicians, their supervisors, and proximal school-employed staff (total n = 120 participants) to estimate the effects of each organization's general and implementation-specific organizational factors (e.g., climate, leadership) on implementation outcomes (fidelity, acceptability, appropriateness) and assess the moderating role of the degree of clinician embeddedness in the school setting. Aim 2 will explore the mechanisms

  12. Imprecise probability analysis for integrated assessment of climate change

    OpenAIRE

    Kriegler, Elmar

    2005-01-01

    We present an application of imprecise probability theory to the quantification of uncertainty in the integrated assessment of climate change. Our work is motivated by the fact that uncertainty about climate change is pervasive, and therefore requires a thorough treatment in the integrated assessment process. Classical probability theory faces some severe difficulties in this respect, since it cannot capture very poor states of information in a satisfactory manner. A more general framework is...

  13. Contract Management Process Maturity: Analysis of Recent Organizational Assessments

    OpenAIRE

    Rendon, Rene G.

    2009-01-01

    Proceedings Paper (for Acquisition Research Program) Approved for public release; distribution unlimited. This research builds upon the emerging body of knowledge on organizational assessments of contract management processes. Since the development of the Contract Management Maturity Modelユ_ᄅ in 2003, several DoD, Air Force, Navy, Army, and defense contractor organizations have undergone contract management process assessments as a part of their process-improvement effort. The assessm...

  14. Assessment of Transport Projects: Risk Analysis and Decision Support

    DEFF Research Database (Denmark)

    Salling, Kim Bang

    2008-01-01

    . Even though vast amounts of money are spent upon preliminary models, environmental investigations, public hearings, etc., the resulting outcome is given by point estimates, i.e. in terms of net present values or benefit-cost rates. This thesis highlights the perspective of risks when assessing...... choosing probability distributions and performing real term data fits. The perspective of this Ph.D. study presents newer and better understanding of assigning risks within assessment of transport projects....

  15. Radiological assessment. A textbook on environmental dose analysis

    Energy Technology Data Exchange (ETDEWEB)

    Till, J.E.; Meyer, H.R. (eds.)

    1983-09-01

    Radiological assessment is the quantitative process of estimating the consequences to humans resulting from the release of radionuclides to the biosphere. It is a multidisciplinary subject requiring the expertise of a number of individuals in order to predict source terms, describe environmental transport, calculate internal and external dose, and extrapolate dose to health effects. Up to this time there has been available no comprehensive book describing, on a uniform and comprehensive level, the techniques and models used in radiological assessment. Radiological Assessment is based on material presented at the 1980 Health Physics Society Summer School held in Seattle, Washington. The material has been expanded and edited to make it comprehensive in scope and useful as a text. Topics covered include (1) source terms for nuclear facilities and Medical and Industrial sites; (2) transport of radionuclides in the atmosphere; (3) transport of radionuclides in surface waters; (4) transport of radionuclides in groundwater; (5) terrestrial and aquatic food chain pathways; (6) reference man; a system for internal dose calculations; (7) internal dosimetry; (8) external dosimetry; (9) models for special-case radionuclides; (10) calculation of health effects in irradiated populations; (11) evaluation of uncertainties in environmental radiological assessment models; (12) regulatory standards for environmental releases of radionuclides; (13) development of computer codes for radiological assessment; and (14) assessment of accidental releases of radionuclides.

  16. Quality Assessment and Proximate Analysis of Amaranthus hybridus ...

    African Journals Online (AJOL)

    The aim of this research is to determine the quality and proximate composition of Amaranthus hybridus, Celosia argentea, and Talinum triangulare obtained from open markets in Benin City, Nigeria. Microbiological and proximate analysis were carried out using standard methods. Results of the proximate analysis revealed ...

  17. Using multivariate statistical analysis to assess changes in water ...

    African Journals Online (AJOL)

    Canonical correspondence analysis (CCA) showed that the environmental variables used in the analysis, discharge and month of sampling, explained a small proportion of the total variance in the data set – less than 10% at each site. However, the total data set variance, explained by the 4 hypothetical axes generated by ...

  18. Methods of Assessing Replicability in Canonical Correlation Analysis (CCA).

    Science.gov (United States)

    King, Jason E.

    Theoretical hypotheses generated from data analysis of a single sample should not be advanced until the replicability issue is treated. At least one of three questions usually arises when evaluating the invariance of results obtained from a canonical correlation analysis (CCA): (1) "Will an effect occur in subsequent studies?"; (2)…

  19. Environmental Impact Assessment for Socio-Economic Analysis of Chemicals

    DEFF Research Database (Denmark)

    Calow, Peter; Biddinger, G; Hennes, C

    This report describes the requirements for, and illustrates the application of, a methodology for a socio-economic analysis (SEA) especially as it might be adopted in the framework of REACH.......This report describes the requirements for, and illustrates the application of, a methodology for a socio-economic analysis (SEA) especially as it might be adopted in the framework of REACH....

  20. ASSESSMENT OF REGIONAL EFFICIENCY IN CROATIA USING DATA ENVELOPMENT ANALYSIS

    Directory of Open Access Journals (Sweden)

    Danijela Rabar

    2013-02-01

    Full Text Available In this paper, regional efficiency of Croatian counties is measured in three-year period (2005-2007 using Data Envelopment Analysis (DEA. The set of inputs and outputs consists of seven socioeconomic indicators. Analysis is carried out using models with assumption of variable returns-to-scale. DEA identifies efficient counties as benchmark members and inefficient counties that are analyzed in detail to determine the sources and the amounts of their inefficiency in each source. To enable proper monitoring of development dynamics, window analysis is applied. Based on the results, guidelines for implementing necessary improvements to achieve efficiency are given. Analysis reveals great disparities among counties. In order to alleviate naturally, historically and politically conditioned unequal county positions over which economic policy makers do not have total control, categorical approach is introduced as an extension to the basic DEA models. This approach, combined with window analysis, changes relations among efficiency scores in favor of continental counties.

  1. Windfarm Generation Assessment for ReliabilityAnalysis of Power Systems

    DEFF Research Database (Denmark)

    Negra, Nicola Barberis; Holmstrøm, Ole; Bak-Jensen, Birgitte

    2007-01-01

    Due to the fast development of wind generation in the past ten years, increasing interest has been paid to techniques for assessing different aspects of power systems with a large amount of installed wind generation. One of these aspects concerns power system reliability. Windfarm modelling plays...... in a reliability model and the generation of a windfarm is evaluated by means of sequential Monte Carlo simulation. Results are used to analyse how each mentioned Factor influences the assessment, and why and when they should be included in the model....

  2. Windfarm Generation Assessment for Reliability Analysis of Power Systems

    DEFF Research Database (Denmark)

    Barberis Negra, Nicola; Bak-Jensen, Birgitte; Holmstrøm, O.

    2007-01-01

    Due to the fast development of wind generation in the past ten years, increasing interest has been paid to techniques for assessing different aspects of power systems with a large amount of installed wind generation. One of these aspects concerns power system reliability. Windfarm modelling plays...... in a reliability model and the generation of a windfarm is evaluated by means of sequential Monte Carlo simulation. Results are used to analyse how each mentioned Factor influences the assessment, and why and when they should be included in the model....

  3. Vulnerability and Risk Analysis Program: Overview of Assessment Methodology

    National Research Council Canada - National Science Library

    2001-01-01

    .... Over the last three years, a team of national laboratory experts, working in partnership with the energy industry, has successfully applied the methodology as part of OCIP's Vulnerability and Risk Analysis Program (VRAP...

  4. The predictive value of aptitude assessment in laparoscopic surgery : a meta-analysis

    NARCIS (Netherlands)

    Kramp, Kelvin H.; van Det, Marc J.; Hoff, Christiaan; Veeger, Nic J. G. M.; ten Cate Hoedemaker, Henk O.; Pierie, Jean-Pierre E. N.

    ContextCurrent methods of assessing candidates for medical specialties that involve laparoscopic skills suffer from a lack of instruments to assess the ability to work in a minimally invasive surgery environment. ObjectivesA meta-analysis was conducted to investigate whether aptitude assessment can

  5. 77 FR 48107 - Workshop on Performance Assessments of Near-Surface Disposal Facilities: FEPs Analysis, Scenario...

    Science.gov (United States)

    2012-08-13

    ...; ] NUCLEAR REGULATORY COMMISSION 10 CFR Part 61 RIN-3150-AI92 Workshop on Performance Assessments of Near... Commission (NRC or the Commission) plans to conduct a workshop on performance assessments of near-surface low... on three aspects of a performance assessment: (1) Features, Events, and Processes (FEPs) analysis, (2...

  6. A big data analysis approach for rail failure risk assessment

    NARCIS (Netherlands)

    Jamshidi, A.; Faghih Roohi, S.; Hajizadeh, S.; Nunez Vicencio, Alfredo; Babuska, R.; Dollevoet, R.P.B.J.; Li, Z.; De Schutter, B.H.K.

    2017-01-01

    Railway infrastructure monitoring is a vital task to ensure rail transportation safety. A rail failure could result in not only a considerable impact on train delays and maintenance costs, but also on safety of passengers. In this article, the aim is to assess the risk of a rail failure by

  7. An Analysis of Peer Assessment through Many Facet Rasch Model

    Science.gov (United States)

    Sahin, Melek Gülsah; Teker, Gülsen Tasdelen; Güler, Nese

    2016-01-01

    This study analyses peer assessment through many facet Rasch model (MFRM). The research was performed with 91 undergraduate students and with lecturer teaching the course. The research data were collected with holistic rubric employed by 6 peers and the lecturer in rating the projects prepared by 85 students taking the course. This study analyses…

  8. Sensitivity Analysis of QSAR Models for Assessing Novel Military Compounds

    Science.gov (United States)

    2009-01-01

    pensive computational models and/or read-across approaches for assess- ing aquatic toxicity ( Pavan et al. 2006; Lilienblum et al. 2008). It seems likely...December 4–6, 2007, Washington, DC. Pavan , M., T. I. Netzeva, and A. P. Worth. 2006. Validation of QSAR model for toxicity. SAR Envrion. Res. 17(2

  9. Assessing Higher Level Learning: Developing Rubrics for Case Analysis

    Science.gov (United States)

    Rochford, Linda; Borchert, Patricia S.

    2011-01-01

    Case study analyses allow students to demonstrate proficiency in executing authentic tasks in marketing and management, facilitating faculty evaluation of higher order learning outcomes. Effective and consistent assessment of case analyses depends in large part on the development of sound rubrics. The authors explored the process of rubric…

  10. Design and analysis for thematic map accuracy assessment: Fundamental principles

    Science.gov (United States)

    Stephen V. Stehman; Raymond L. Czaplewski

    1998-01-01

    Land-cover maps are used in numerous natural resource applications to describe the spatial distribution and pattern of land-cover, to estimate areal extent of various cover classes, or as input into habitat suitability models, land-cover change analyses, hydrological models, and risk analyses. Accuracy assessment quantifies data quality so that map users may evaluate...

  11. Cost-Utility Analysis in Educational Needs Assessment

    Science.gov (United States)

    Ross, John A.

    2008-01-01

    Educational needs assessments (hereafter NAs) are inattentive to cost considerations and are frequently dominated by elite stakeholder groups. In this article I make a case for adopting a cost-utility approach, illustrating the argument with data generated in a NA of central library services in a Canadian school district. Using survey data from…

  12. Analysis of Online Quizzes as a Teaching and Assessment Tool

    Science.gov (United States)

    Salas-Morera, Lorenzo; Arauzo-Azofra, Antonio; García-Hernández, Laura

    2012-01-01

    This article deals with the integrated use of online quizzes as a teaching and assessment tool in the general program of the subject Proyectos in the third course of Ingeniero Técnico en Informática de Gestión over five consecutive years. The research undertaken aimed to test quizzes effectiveness on student performance when used, not only as an…

  13. Repeater Analysis for Combining Information from Different Assessments

    Science.gov (United States)

    Haberman, Shelby; Yao, Lili

    2015-01-01

    Admission decisions frequently rely on multiple assessments. As a consequence, it is important to explore rational approaches to combine the information from different educational tests. For example, U.S. graduate schools usually receive both TOEFL iBT® scores and GRE® General scores of foreign applicants for admission; however, little guidance…

  14. Quantitative assessment of human motion using video motion analysis

    Science.gov (United States)

    Probe, John D.

    1993-01-01

    In the study of the dynamics and kinematics of the human body a wide variety of technologies has been developed. Photogrammetric techniques are well documented and are known to provide reliable positional data from recorded images. Often these techniques are used in conjunction with cinematography and videography for analysis of planar motion, and to a lesser degree three-dimensional motion. Cinematography has been the most widely used medium for movement analysis. Excessive operating costs and the lag time required for film development, coupled with recent advances in video technology, have allowed video based motion analysis systems to emerge as a cost effective method of collecting and analyzing human movement. The Anthropometric and Biomechanics Lab at Johnson Space Center utilizes the video based Ariel Performance Analysis System (APAS) to develop data on shirtsleeved and space-suited human performance in order to plan efficient on-orbit intravehicular and extravehicular activities. APAS is a fully integrated system of hardware and software for biomechanics and the analysis of human performance and generalized motion measurement. Major components of the complete system include the video system, the AT compatible computer, and the proprietary software.

  15. Model-Driven Analysis towards Interoperability of Assessments in LMS

    Directory of Open Access Journals (Sweden)

    Zuzana Bizonova

    2010-06-01

    Full Text Available In this article we are focusing on interoperability of two aspects of LMS systems - test question types and assessments as such. The proposed strategy is based on MDA, especially on the Platform Independent Models (PIM. On the higher level of abstraction (PIM it is possible to find commonalities and differences of architectures of various systems and these are the basis for the common generalized model of the assessments. In the three steps methodology we are adding specificities of PIM models of candidate systems with various architecture – Moodle, Olat and Claroline. The correctness of the final common model (General PIM is proved in the implemented system for exchange of tests between the existing systems.

  16. A comparative assessment of statistical methods for extreme weather analysis

    Science.gov (United States)

    Schlögl, Matthias; Laaha, Gregor

    2017-04-01

    Extreme weather exposure assessment is of major importance for scientists and practitioners alike. We compare different extreme value approaches and fitting methods with respect to their value for assessing extreme precipitation and temperature impacts. Based on an Austrian data set from 25 meteorological stations representing diverse meteorological conditions, we assess the added value of partial duration series over the standardly used annual maxima series in order to give recommendations for performing extreme value statistics of meteorological hazards. Results show the merits of the robust L-moment estimation, which yielded better results than maximum likelihood estimation in 62 % of all cases. At the same time, results question the general assumption of the threshold excess approach (employing partial duration series, PDS) being superior to the block maxima approach (employing annual maxima series, AMS) due to information gain. For low return periods (non-extreme events) the PDS approach tends to overestimate return levels as compared to the AMS approach, whereas an opposite behavior was found for high return levels (extreme events). In extreme cases, an inappropriate threshold was shown to lead to considerable biases that may outperform the possible gain of information from including additional extreme events by far. This effect was neither visible from the square-root criterion, nor from standardly used graphical diagnosis (mean residual life plot), but from a direct comparison of AMS and PDS in synoptic quantile plots. We therefore recommend performing AMS and PDS approaches simultaneously in order to select the best suited approach. This will make the analyses more robust, in cases where threshold selection and dependency introduces biases to the PDS approach, but also in cases where the AMS contains non-extreme events that may introduce similar biases. For assessing the performance of extreme events we recommend conditional performance measures that focus

  17. Composite power system adequacy assessment based on postoptimal analysis

    OpenAIRE

    SAFDARIAN, Amir; FIRUZABAD, Mahmud FOTUHI; AMINIFAR, Farrokh

    2014-01-01

    The modeling and evaluation of enormous numbers of contingencies are the most challenging impediments associated with composite power system adequacy assessment, particularly for large-scale power systems. Optimal power flow (OPF) solution, as a widely common approach, is normally employed to model and analyze each individual contingency as an independent problem. However, mathematical representations associated with diverse states are slightly different in one or a few generating un...

  18. A sequential criterion analysis for assessing coping with chronic illness.

    Science.gov (United States)

    Turk, D C; Sobel, H J; Follick, M J; Youkilis, H D

    1980-01-01

    A three-stage sequential criterion analysis based on the Goldfried-D'Zurilla behavior model is presented as a heuristic approach for studying chronic illness. The three interrelated stages consist of: (a) problem identification, (b) response enumeration, and (c) response evaluation. The utility of this systematic approach for understanding the adaptive process, helping to identify patients who might benefit from intervention, providing content for actual intervention, and reducing the myth of homogeneity among chronic patient populations for both clinicians and researches, is discussed. Examples of the utility of the criterion analysis are provided throughout the paper.

  19. Sex Differences in Mathematics and Science Achievement: A Meta-Analysis of National Assessment of Educational Progress Assessments

    Science.gov (United States)

    Reilly, David; Neumann, David L.; Andrews, Glenda

    2015-01-01

    Gender gaps in the development of mathematical and scientific literacy have important implications for the general public's understanding of scientific issues and for the underrepresentation of women in science, technology, engineering, and math. We subjected data from the National Assessment of Educational Progress to a meta-analysis to examine…

  20. Concentration Analysis: A Quantitative Assessment of Student States.

    Science.gov (United States)

    Bao, Lei; Redish, Edward F.

    Multiple-choice tests such as the Force Concept Inventory (FCI) provide useful instruments to probe the distribution of student difficulties on a large scale. However, traditional analysis often relies solely on scores (number of students giving the correct answer). This ignores what can be significant and important information: the distribution…

  1. Visual unit analysis: a descriptive approach to landscape assessment

    Science.gov (United States)

    R. J. Tetlow; S. R. J. Sheppard

    1979-01-01

    Analysis of the visible attributes of landscapes is an important component of the planning process. When landscapes are at regional scale, economical and effective methodologies are critical. The Visual Unit concept appears to offer a logical and useful framework for description and evaluation. The concept subdivides landscape into coherent, spatially-defined units....

  2. Surface water quality assessment using factor analysis | Boyacioglu ...

    African Journals Online (AJOL)

    In this study, the factor analysis technique is applied to surface water quality data sets obtained from the Buyuk Menderes River Basin, Turkey, during two different hydrological periods. Results show that the indices which changed the quality of water in two seasons and locations differed. During low-flow conditions, water ...

  3. Using multivariate statistical analysis to assess changes in water ...

    African Journals Online (AJOL)

    Multivariate statistical analysis was used to investigate changes in water chemistry at 5 river sites in the Vaal Dam catchment, draining the Highveld grasslands. These grasslands receive more than 8 kg sulphur (S) ha-1·year-1 and 6 kg nitrogen (N) ha-1·year-1 via atmospheric deposition. It was hypothesised that between ...

  4. Service reliability assessment using failure mode and effect analysis ...

    African Journals Online (AJOL)

    Despite Failure Mode and Effect Analysis (FMEA) being a strategic technique for creation of error free service operation, detailed survey study and development of opportunity roadmap for FMEA application in service operation is limited in literature. We presented a preliminary literature survey between 1994 and 2010 that ...

  5. Assessing the copula selection for bivariate frequency analysis ...

    Indian Academy of Sciences (India)

    58

    J R. Stat Soc Series B Stat Methodol, 54, 171-183. Karmakar, S. & Simonovic, S. 2009. Bivariate flood frequency analysis. Part 2: a copula‐based approach with mixed marginal distributions. J. Flood Risk Manage., 2, 32-44. Katz, R. W. 2013. Statistical methods for nonstationary extremes. Extremes in a Changing. Climate.

  6. A Preliminary Analysis of a Behavioral Classrooms Needs Assessment

    Science.gov (United States)

    Leaf, Justin B.; Leaf, Ronald; McCray, Cynthia; Lamkins, Carol; Taubman, Mitchell; McEachin, John; Cihon, Joseph H.

    2016-01-01

    Today many special education classrooms implement procedures based upon the principles of Applied Behavior Analysis (ABA) to establish educationally relevant skills and decrease aberrant behaviors. However, it is difficult for school staff and consultants to evaluate the implementation of various components of ABA and general classroom set up. In…

  7. The Analysis of Ease of Doing Business Assessment Methods

    Directory of Open Access Journals (Sweden)

    Mindaugas Samoška

    2011-07-01

    Full Text Available The study deals with the ease of doing business assessment models. Analysed models describe conditions for doing busi­ness in a certain country that is being ranked and evaluated. However obvious need for improvement in methodology accrues while analysing five most known models in a comparative way. Different data aggregation principles differ the results (quantative and qualitive methods of aggregation despite the factors that are evaluated in both models and are quite similar. After analysing all five methods we state critics about them and insights for possible improvement.Article in Lithuanian

  8. Cost-effectiveness analysis of computer-based assessment

    Directory of Open Access Journals (Sweden)

    Pauline Loewenberger

    2003-12-01

    Full Text Available The need for more cost-effective and pedagogically acceptable combinations of teaching and learning methods to sustain increasing student numbers means that the use of innovative methods, using technology, is accelerating. There is an expectation that economies of scale might provide greater cost-effectiveness whilst also enhancing student learning. The difficulties and complexities of these expectations are considered in this paper, which explores the challenges faced by those wishing to evaluate the costeffectiveness of computer-based assessment (CBA. The paper outlines the outcomes of a survey which attempted to gather information about the costs and benefits of CBA.

  9. Use Of Risk Analysis Fremeworks In Urban Flood Assessments

    DEFF Research Database (Denmark)

    Arnbjerg-Nielsen, Karsten; Madsen, Henrik

    , primarily insurance companies and mortgage providers, but also politicians and media are highly interested. Presently two very different approaches are being followed in both research and practice. One is the introduction of risk analysis and risk management tools to provide professionals and politicians...... in extreme precipitation has been observed, corresponding to an increase of design levels of at least 30 %. Analysis of climate change model output has given clear evidence, that further increases in extreme precipitation must be expected in the future due to anthropogenic emissions of greenhouse gasses...... with better decision support tools. Some of the developments are risk frameworks that encompass economic and/or ethic evaluation of climate change adaptation options and improved risk management. This line of development is based on a societal-based evaluation of maximizing the outcome for society...

  10. Assessment of competence in simulated flexible bronchoscopy using motion analysis

    DEFF Research Database (Denmark)

    Collela, Sara; Svendsen, Morten Bo Søndergaard; Konge, Lars

    2015-01-01

    intermediates and 9 experienced bronchoscopy operators performed 3 procedures each on a bronchoscopy simulator. The Microsoft Kinect system was used to automatically measure the total deviation of the scope from a perfectly straight, vertical line. Results: The low-cost motion analysis system could measure...... with the performance on the simulator (virtual-reality simulator score; p simulators might help new bronchoscopists learn how to handle...... the bronchoscope like an expert. © 2015 S. Karger AG, Basel....

  11. Multi-criteria analysis application in the investment projects assessment

    OpenAIRE

    Popović, Žarko; Stanković, Jelena; Veselinović, Ivana

    2013-01-01

    The growth and development of each economic agent is inconceivable without investment or investment activity. In this sense, the investment decision plays a very important role in the realization of the development goals. Economically correct decision is based on the application of various methods of analysis of investment projects and the choice of the most acceptable one. Since a large number of factors affect the decision about the selection, the nature of the problem is multi-criteria. Th...

  12. Assessing Spontaneous Combustion Instability with Nonlinear Time Series Analysis

    Science.gov (United States)

    Eberhart, C. J.; Casiano, M. J.

    2015-01-01

    Considerable interest lies in the ability to characterize the onset of spontaneous instabilities within liquid propellant rocket engine (LPRE) combustion devices. Linear techniques, such as fast Fourier transforms, various correlation parameters, and critical damping parameters, have been used at great length for over fifty years. Recently, nonlinear time series methods have been applied to deduce information pertaining to instability incipiency hidden in seemingly stochastic combustion noise. A technique commonly used in biological sciences known as the Multifractal Detrended Fluctuation Analysis has been extended to the combustion dynamics field, and is introduced here as a data analysis approach complementary to linear ones. Advancing, a modified technique is leveraged to extract artifacts of impending combustion instability that present themselves a priori growth to limit cycle amplitudes. Analysis is demonstrated on data from J-2X gas generator testing during which a distinct spontaneous instability was observed. Comparisons are made to previous work wherein the data were characterized using linear approaches. Verification of the technique is performed by examining idealized signals and comparing two separate, independently developed tools.

  13. Skeletal assessment with finite element analysis: relevance, pitfalls and interpretation.

    Science.gov (United States)

    Campbell, Graeme Michael; Glüer, Claus-C

    2017-07-01

    Finite element models simulate the mechanical response of bone under load, enabling noninvasive assessment of strength. Models generated from quantitative computed tomography (QCT) incorporate the geometry and spatial distribution of bone mineral density (BMD) to simulate physiological and traumatic loads as well as orthopaedic implant behaviour. The present review discusses the current strengths and weakness of finite element models for application to skeletal biomechanics. In cadaver studies, finite element models provide better estimations of strength compared to BMD. Data from clinical studies are encouraging; however, the superiority of finite element models over BMD measures for fracture prediction has not been shown conclusively, and may be sex and site dependent. Therapeutic effects on bone strength are larger than for BMD; however, model validation has only been performed on untreated bone. High-resolution modalities and novel image processing methods may enhance the structural representation and predictive ability. Despite extensive use of finite element models to study orthopaedic implant stability, accurate simulation of the bone-implant interface and fracture progression remains a significant challenge. Skeletal finite element models provide noninvasive assessments of strength and implant stability. Improved structural representation and implant surface interaction may enable more accurate models of fragility in the future.

  14. An Analysis Report of 2014 CALA Self-Assessment Survey

    Directory of Open Access Journals (Sweden)

    Jian Anna Xiong

    2016-06-01

    Full Text Available On the occasion of CALA’s 40th anniversary in 2014, the 2013 Board of Directors appointed a Self-Assessment Task Force to conduct an assessment survey with special focuses on members’ awareness of CALA’s organizational structure and policies, its services to members, the extent of participation in events sponsored by CALA, and the level of satisfaction with CALA leadership. Although only one-fifth of the active members responded to the survey, the answers and feedback have identified areas for organizational improvement and have shown how active members view the current state of CALA. Some essential findings from the survey include: 1 the growth of overseas membership as a demographic trend, 2 a need to recruit student members, 3 a high percentage of CALA members not aware of CALA’s Mission/Vision/Goal, 4 conflicting data on CALA’s leadership, 5 discovery of low ratings (10-30% of respondents on eleven out of twelve rating questions, and 6 strong support for CALA as a representative organization of Chinese American librarians in North America. The findings of the survey will serve as a valuable reference for future strategic planning and for carrying out CALA’s long term goals.

  15. An assessment on the applicability of licensing requirements for safety analysis of CANDU nuclear power plant

    Energy Technology Data Exchange (ETDEWEB)

    Jang, S. H. [Korea Institute of Nuclear Safety, Taejon (Korea, Republic of)

    1992-05-15

    The purpose of this project is to establish safety analysis requirements for CANDU nuclear power plants to be licensed for construction and operation in Korea. Main emphasis is given to assess the applicability of the Canada AECB consultant Document C-6 and to establish the method of application. Examination of the NPP regulation philosophy and safety analysis requirements in Canada. Assessment of the applicability of the C-6 requirements for CANDU plants to be built in Korea. Establishment of the method of application for C-6 requirements. Assessment of the PSA requirements for CANDU. Examination of the validity of computer programs used for CANDU safety analysis.

  16. Modeling and Analysis on Radiological Safety Assessment of Low- and Intermediate Level Radioactive Waste Repository

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Youn Myoung; Jung, Jong Tae; Kang, Chul Hyung (and others)

    2008-04-15

    Modeling study and analysis for technical support for the safety and performance assessment of the low- and intermediate level (LILW) repository partially needed for radiological environmental impact reporting which is essential for the licenses for construction and operation of LILW has been fulfilled. Throughout this study such essential area for technical support for safety and performance assessment of the LILW repository and its licensing as gas generation and migration in and around the repository, risk analysis and environmental impact during transportation of LILW, biosphere modeling and assessment for the flux-to-dose conversion factors for human exposure as well as regional and global groundwater modeling and analysis has been carried out.

  17. Comparative analysis of deterministic and probabilistic fracture mechanical assessment tools

    Energy Technology Data Exchange (ETDEWEB)

    Heckmann, Klaus [Gesellschaft fuer Anlagen- und Reaktorsicherheit (GRS) gGmbH, Koeln (Germany); Saifi, Qais [VTT Technical Research Centre of Finland, Espoo (Finland)

    2016-11-15

    Uncertainties in material properties, manufacturing processes, loading conditions and damage mechanisms complicate the quantification of structural reliability. Probabilistic structure mechanical computing codes serve as tools for assessing leak- and break probabilities of nuclear piping components. Probabilistic fracture mechanical tools were compared in different benchmark activities, usually revealing minor, but systematic discrepancies between results of different codes. In this joint paper, probabilistic fracture mechanical codes are compared. Crack initiation, crack growth and the influence of in-service inspections are analyzed. Example cases for stress corrosion cracking and fatigue in LWR conditions are analyzed. The evolution of annual failure probabilities during simulated operation time is investigated, in order to identify the reasons for differences in the results of different codes. The comparison of the tools is used for further improvements of the codes applied by the partners.

  18. Assessment of frame-averaging algorithms in OCT image analysis.

    Science.gov (United States)

    Wu, Wei; Tan, Ou; Pappuru, Rajeev R; Duan, Huilong; Huang, David

    2013-01-01

    To evaluate frame registration and averaging algorithms for optical coherence tomography. Twenty normal and 20 glaucomatous eyes were imaged. Objective differences were measured by comparing noise variance, spread of edge, and contrast-to-noise ratio. Subjective image quality was also evaluated. Two frame-averaging algorithms (FA400 and FA407) had better noise variance and contrast-to-noise ratio but worse spread of edge than did single frames (P algorithms provided better subjective assessments of structure boundaries than did single images (P algorithms, higher contrast-to-noise ratio and lower noise variance indicated better noise suppression, but spread of edge was most sensitive in comparing edge preservation between algorithms. Copyright 2013, SLACK Incorporated.

  19. Fuel assembly assessment from CVD image analysis: A feasibility study

    Energy Technology Data Exchange (ETDEWEB)

    Lindsay, C.S.; Lindblad, T. [Royal Inst. of Tech., Stockholm (Sweden). Dept. of Physics

    1997-05-01

    The Swedish Nuclear Inspectorate commissioned a feasibility study of automatic assessment of fuel assemblies from images obtained with the digital Cerenkov viewing device currently in development. The goal is to assist the IAEA inspectors in evaluating the fuel since they typically have only a few seconds to inspect an assembly. We report results here in two main areas: Investigation of basic image processing and recognition techniques needed to enhance the images and find the assembly in the image; Study of the properties of the distributions of light from the assemblies to determine whether they provide unique signatures for different burn-up and cooling times for real fuel or indicate presence of non-fuel. 8 refs, 27 figs.

  20. Image analysis for dental bone quality assessment using CBCT imaging

    Science.gov (United States)

    Suprijanto; Epsilawati, L.; Hajarini, M. S.; Juliastuti, E.; Susanti, H.

    2016-03-01

    Cone beam computerized tomography (CBCT) is one of X-ray imaging modalities that are applied in dentistry. Its modality can visualize the oral region in 3D and in a high resolution. CBCT jaw image has potential information for the assessment of bone quality that often used for pre-operative implant planning. We propose comparison method based on normalized histogram (NH) on the region of inter-dental septum and premolar teeth. Furthermore, the NH characteristic from normal and abnormal bone condition are compared and analyzed. Four test parameters are proposed, i.e. the difference between teeth and bone average intensity (s), the ratio between bone and teeth average intensity (n) of NH, the difference between teeth and bone peak value (Δp) of NH, and the ratio between teeth and bone of NH range (r). The results showed that n, s, and Δp have potential to be the classification parameters of dental calcium density.

  1. An Assessment of Image Analysis of Longitudinal Bone Changes

    Energy Technology Data Exchange (ETDEWEB)

    Kim, Young Jin; Kim, Jae Duck [Dept. of Oral and Maxillofacial Radiology, College of Dentistry, Chosun University, Kwangju (Korea, Republic of)

    1997-02-15

    This study was performed to assess the analyzing methods developed to detect clinically and quantitatively longitudinal bone changes. Through preliminary experiment, accuracy of Cu-Eq value conversion to the mass of HA was examined. For main experiment, 15 intraoral radiograms taken at soon, 1st, 2nd, 4th, and 6th week after implantation of mixture in extracted sites of 3 cases were user. We took the radiograms with copper step wedge as test object and HA phantom. X -ray taking was standardized by using Rinn XCP device customized directly to the individual dentition with resin bite block. The images inputted by Quick scanner into computer were digitized and analyzed by NH image program, the stability of the copper equivalent transformation and the usefulness of two analyzing methods by ROI and Reslice were examined. Obtained results as follows : 1. On the Cu equivalent images, the coefficient of variation in the measurement of Cu-Eq. value of ROI ranged from 0.05 to 0.24 and showed high reproducibility. 2. All results obtained by resliced contiguous images were coincident with those obtained from the assessment by ROI an d formation of plot profile. 3. On the stacked and resliced image at the line of interest, we could analyze directly and quantitatively the longitudinal changes at several portions by plot profile and qualitatively by surface plot. 4. Implant area showed marked resorption till 2 weeks after implantation and showed significant increase in Cu-Eq. value at 6th week (P<0.01) and periapical area showed increase in Cu-Eq. value at 6th week compared to after-operation's.

  2. Assessment of Tibetan grassland degeneration via landscape analysis

    Science.gov (United States)

    Sun, Jian; Hou, Ge; Ma, Baibing; Zang, Wenqian

    2017-04-01

    Desertification as one of the most severity social-economic-environmental issues has been extensive researched, and the assessments of desertification can be implemented accurately and efficiently based on the landscape indicators of vegetation coverage. Consequently, we explored the relationships of the degeneration index of the grassland with climate factors (temperature and precipitation), and human disturbance factors (livestock quantity and animal husbandry output value) via a landscape assessment approach across Tibet. The results showed that the vegetation coverage presented an increase tendency in the central region of Tibet, but the adverse phenomenon was observed in the northwest region. Meanwhile, the correlation of vegetation coverage with precipitation presented as positive effect in most region of Tibet except some regions of the alpine steppe, and the positive correlation of vegetation coverage with temperature also was observed in the less northwest region of Tibet. In addition, we found that the livestock quantity play a key roles in regulating vegetation coverage of the central region. Furthermore, the landscape indexes [number of patches (NP), patch density (PD), contagion index (CONTAG), landscape shape index (LSI), aggregation index (AI)] of grasslands were analyzed, the results exposed that vegetation coverage (1%-20%) has the positive influences on CONTAG and AI, but negative affects LSI, PD and NP. Morreover, there are opposite correlations among vegetation coverage and landscape indexes when vegetation coverage is 21%-40%. We concluded that overgrazing is the main reason of grassland degradation in Tibet, especially the number of livestock aggravates the landscape fragmentation. The results highlighted the alpine grassland management in future.

  3. Sensitivity analysis of risk assessment programs to various input parameters.

    Science.gov (United States)

    Vadász, Z; Hassanien, M A; László, E; Dura, G

    2000-07-01

    HESP 2.b risk assessment program was studied in detail concerning the effect of changing different input parameters for the output ADI values calculated by the program. We used the standard Netherlands 1.0 scenario offered by the program. With this we fixed a lot of input parameters which define the area, human and animal recipient parameters etc. The remaining 31 unfixed parameters were fixed at first to "BASE" input values and the BASE output values were calculated by HESP. Later we chose only one parameter at a time and changed it to an another value. The calculated ADI values were then compared to BASE output values. Seven parameters (soil type, soil usage, site length, soil pH, groundwater fraction in drinking water, basement floor type and Qev) were studied. We found, that changing soil pH or Qev have not any influence on the output ADI values in case of any contaminant. Soil type change has not any effect on the output ADI value in case of Pb or Cd but it seems to play important role in all cases of the four organic material we investigated. Changing soil usage have influence on the output ADI value almost in every case. It seems to be linear relation between the maximal concentration of contaminant and calculated ADI. Changing the site length and basement floor type gave in some cases different ADI values compared to BASE values. If we alter the groundwater fraction in drinking water we got usually different ADI values. With Risc Human risk assessment program we got similar results: nor the changes in soil type, site diameter or soil pH gave any changes in output ADI values. Our results hint that using HESP and Risc Human requires enhanced caution.

  4. Jelly pineapple syneresis assessment via univariate and multivariate analysis

    OpenAIRE

    Carlos Alberto da Silva Ledo; Rossana Catie Bueno de Godoy; Arislete Dantas de Aquino; Silvana Licodiedoff

    2010-01-01

    The evaluation of the pineapple jelly is intended to analyze the occurrence of syneresis by univariate and multivariate analysis. The jelly of the pineapple presents low concentration pectin, therefore, it was added high methoxyl pectin in the following concentrations: 0.50%, 0.75% and 1.00% corresponding to slow, medium and fast speed of gel formation process. In this study it was checked the pH, acidity, brix and the syneresis of jelly. The highest concentration of pectin in the jelly showe...

  5. Assessing Irregular Warfare: A Framework for Intelligence Analysis

    Science.gov (United States)

    2008-01-01

    Intelligence Analysis 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6. AUTHOR(S) 5d. PROJECT NUMBER 5e. TASK NUMBER 5f. WORK UNIT NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Rand Corporation,1776 Main Street,PO Box 2138,Santa Monica,CA,90407-2138 8. PERFORMING ORGANIZATION REPORT NUMBER 9. SPONSORING/MONITORING AGENCY NAME(S) AND ADDRESS(ES) 10. SPONSOR/MONITOR’S ACRONYM(S) 11. SPONSOR/MONITOR’S REPORT NUMBER(S) 12. DISTRIBUTION/AVAILABILITY STATEMENT Approved for public release; distribution

  6. Assessing Canadian Bank Branch Operating Efficiency Using Data Envelopment Analysis

    Science.gov (United States)

    Yang, Zijiang

    2009-10-01

    In today's economy and society, performance analyses in the services industries attract more and more attention. This paper presents an evaluation of 240 branches of one big Canadian bank in Greater Toronto Area using Data Envelopment Analysis (DEA). Special emphasis was placed on how to present the DEA results to management so as to provide more guidance to them on what to manage and how to accomplish the changes. Finally the potential management uses of the DEA results were presented. All the findings are discussed in the context of the Canadian banking market.

  7. Orbital data applications for space objects conjunction assessment and situation analysis

    CERN Document Server

    Chen, Lei; Liang, Yan-Gang; Li, Ke-Bo

    2017-01-01

    This book introduces readers to the application of orbital data on space objects in the contexts of conjunction assessment and space situation analysis, including theories and methodologies. It addresses the main topics involved in space object conjunction assessment, such as: orbital error analysis of space objects; close approach analysis; the calculation, analysis and application of collision probability; and the comprehensive assessment of collision risk. In addition, selected topics on space situation analysis are also presented, including orbital anomaly and space event analysis, and so on. The book offers a valuable guide for researchers and engineers in the fields of astrodynamics, space telemetry, tracking and command (TT&C), space surveillance, space situational awareness, and space debris, as well as for graduates majoring in flight vehicle design and related fields.

  8. Assessment of Prospective Physician Characteristics by SWOT Analysis.

    Science.gov (United States)

    Thira, Woratanarat; Patarawan, Woratanarat

    2012-01-01

    Thailand is one of the developing countries encountering medical workforce shortage. From the national registry in 2006, there were 33 166 physicians: 41.5% worked in the government sector, 21.6% worked in the private sector, and the remaining worked in non-medical fields. There is no current data to confirm the effectiveness of the national policy to increase physician production. We demonstrate our findings from the strength, weakness, opportunity, and threat (SWOT) analysis in medical students and the potential impact on national workforce planning. We introduced SWOT analysis to 568 medical students during the 2008-2010 academic years, with the objective of becoming "a good physician in the future". Pertinent issues were grouped into 4 categories: not wanting to be a doctor, having inadequate medical professional skills, not wanting to work in rural or community areas, and planning to pursue training in specialties with high salary/low workload/low risk for lawsuit. The percentages of medical students who described themselves as "do not want to be a doctor" and "do not want to work in rural or community areas" increased from 7.07% and 25.00% in 2008 to 12.56% and 29.65% in 2010, respectively. Further intervention should be considered in order to change the medical students attitudes on the profession and their impact on Thai health system.

  9. Cognitive assessments in multicultural populations using the Rowland Universal Dementia Assessment Scale: a systematic review and meta-analysis.

    Science.gov (United States)

    Naqvi, Raza M; Haider, Sehrish; Tomlinson, George; Alibhai, Shabbir

    2015-03-17

    Canada has a growing multinational immigrant population. Many commonly used cognitive assessment tools have known cultural biases and are difficult to use in ethnically diverse patient populations. The Rowland Universal Dementia Assessment Scale (RUDAS) is a cognitive assessment tool that was created for culturally and linguistically diverse populations. We performed a systematic review and meta-analysis to assess the psychometric characteristics of the RUDAS and to compare it with other available tools. We identified studies that assessed the psychometric properties of the RUDAS compared with a reference standard for diagnosing dementia or compared the RUDAS to other cognitive assessment tools. Two independent reviewers screened the abstracts and full-text articles and reviewed bibliographies for further references. We extracted data using standardized forms and assessed studies for risk of bias. Our search resulted in 148 articles, from which 11 were included. The RUDAS was assessed in 1236 participants and was found to have a pooled sensitivity of 77.2% (95% confidence interval [CI] 67.4-84.5) and a pooled specificity of 85.9% (95% CI 74.8-92.6) yielding a positive likelihood ratio of 5.5 (95% CI 2.9-10.7) and a negative likelihood ratio of 0.27 (95% CI 0.17-0.40). A pooled estimate of the correlation between the RUDAS and the Mini-Mental State Examination (MMSE) was 0.77 (95% CI 0.72-0.81). Results of the RUDAS were less affected by language and education level than the MMSE. The RUDAS is a brief and freely available cognitive assessment tool with reasonable psychometric characteristics that may be particularly useful in patients with diverse backgrounds. © 2015 Canadian Medical Association or its licensors.

  10. A root cause analysis approach to risk assessment of a pipeline network for Kuwait Oil Company

    Energy Technology Data Exchange (ETDEWEB)

    Davies, Ray J.; Alfano, Tony D. [Det Norske Veritas (DNV), Rio de Janeiro, RJ (Brazil); Waheed, Farrukh [Kuwait Oil Company, Ahmadi (Kuwait); Komulainen, Tiina [Kongsberg Oil and Gas Technologies, Sandvika (Norway)

    2009-07-01

    A large scale risk assessment was performed by Det Norske Veritas (DNV) for the entire Kuwait Oil Company (KOC) pipeline network. This risk assessment was unique in that it incorporated the assessment of all major sources of process related risk faced by KOC and included root cause management system related risks in addition to technical risks related to more immediate causes. The assessment was conducted across the entire pipeline network with the scope divided into three major categories:1. Integrity Management 2. Operations 3. Management Systems Aspects of integrity management were ranked and prioritized using a custom algorithm based on critical data sets. A detailed quantitative risk assessment was then used to further evaluate those issues deemed unacceptable, and finally a cost benefit analysis approach was used to compare and select improvement options. The operations assessment involved computer modeling of the entire pipeline network to assess for bottlenecks, surge and erosion analysis, and to identify opportunities within the network that could potentially lead to increased production. The management system assessment was performed by conducting a gap analysis on the existing system and by prioritizing those improvement actions that best aligned with KOC's strategic goals for pipelines. Using a broad and three-pronged approach to their overall risk assessment, KOC achieved a thorough, root cause analysis-based understanding of risks to their system as well as a detailed list of recommended remediation measures that were merged into a 5-year improvement plan. (author)

  11. Multi-criteria decision analysis with probabilistic risk assessment for the management of contaminated ground water

    OpenAIRE

    Khadam, I.; Kaluarachchi, J. J.

    2003-01-01

    Traditionally, environmental decision analysis in subsurface contamination scenarios is performed using cost–benefit analysis. In this paper, we discuss some of the limitations associated with cost–benefit analysis, especially its definition of risk, its definition of cost of risk, and its poor ability to communicate risk-related information. This paper presents an integrated approach for management of contaminated ground water resources using health risk assessment and economic analysis thro...

  12. Seismic Assessment of an RC Building Using Pushover Analysis

    Directory of Open Access Journals (Sweden)

    R. A. Hakim

    2014-08-01

    Full Text Available Current research works and observations indicated that parts of the Kingdom of Saudi Arabia have low to moderate seismic regions. Major parts of buildings were designed only for gravity load and were poorly detailed to accommodate lateral loads. This study aims to investigate building performance on resisting expected seismic loadings. Two 3D frames were investigated using pushover analysis according to ATC-40. One was designed according to a design practice that considers only the gravity load and the other frame was designed according to the Saudi Building Code (SBC-301. Results showed that the building designed considering only the gravity load was found inadequate. On the other hand, the building designed according to SBC-301 satisfies the Immediate Occupancy (IO acceptance criteria according to ATC-40.

  13. Jelly pineapple syneresis assessment via univariate and multivariate analysis

    Directory of Open Access Journals (Sweden)

    Carlos Alberto da Silva Ledo

    2010-09-01

    Full Text Available The evaluation of the pineapple jelly is intended to analyze the occurrence of syneresis by univariate and multivariate analysis. The jelly of the pineapple presents low concentration pectin, therefore, it was added high methoxyl pectin in the following concentrations: 0.50%, 0.75% and 1.00% corresponding to slow, medium and fast speed of gel formation process. In this study it was checked the pH, acidity, brix and the syneresis of jelly. The highest concentration of pectin in the jelly showed a decrease in the release of the water, syneresis. This result showed that the percentage of 1.00% of pectin in jelly is necessary to form the gel and to obtain a suitable texture.

  14. In-field analysis and assessment of nuclear material

    Energy Technology Data Exchange (ETDEWEB)

    Morgado, R.E.; Myers, W.S.; Olivares, J.A.; Phillips, J.R.; York, R.L.

    1996-05-01

    Los Alamos National Laboratory has actively developed and implemented a number of instruments to monitor, detect, and analyze nuclear materials in the field. Many of these technologies, developed under existing US Department of Energy programs, can also be used to effectively interdict nuclear materials smuggled across or within national borders. In particular, two instruments are suitable for immediate implementation: the NAVI-2, a hand-held gamma-ray and neutron system for the detection and rapid identification of radioactive materials, and the portable mass spectrometer for the rapid analysis of minute quantities of radioactive materials. Both instruments provide not only critical information about the characteristics of the nuclear material for law-enforcement agencies and national authorities but also supply health and safety information for personnel handling the suspect materials.

  15. Rorschach assessment of traumatized refugees: an exploratory factor analysis.

    Science.gov (United States)

    Opaas, Marianne; Hartmann, Ellen

    2013-01-01

    Fifty-one multitraumatized mental health patients with refugee backgrounds completed the Rorschach (Meyer & Viglione, 2008), Harvard Trauma Questionnaire, and Hopkins Symptom Checklist-25 (Mollica, McDonald, Massagli, & Silove, 2004), and the World Health Organization Quality of Life-BREF questionnaire (WHOQOL Group, 1998) before the start of treatment. The purpose was to gain more in-depth knowledge of an understudied patient group and to provide a prospective basis for later analyses of treatment outcome. Factor analysis of trauma-related Rorschach variables gave 2 components explaining 60% of the variance; the first was interpreted as trauma-related flooding versus constriction and the second as adequate versus impaired reality testing. Component 1 correlated positively with self-reported reexperiencing symptoms of posttraumatic stress (r = .32, p social relationships domains (r = .34, .32, and .35, p work experience, education, and language skills.

  16. Assessing Computational Steps for CLIP-Seq Data Analysis

    Directory of Open Access Journals (Sweden)

    Qi Liu

    2015-01-01

    Full Text Available RNA-binding protein (RBP is a key player in regulating gene expression at the posttranscriptional level. CLIP-Seq, with the ability to provide a genome-wide map of protein-RNA interactions, has been increasingly used to decipher RBP-mediated posttranscriptional regulation. Generating highly reliable binding sites from CLIP-Seq requires not only stringent library preparation but also considerable computational efforts. Here we presented a first systematic evaluation of major computational steps for identifying RBP binding sites from CLIP-Seq data, including preprocessing, the choice of control samples, peak normalization, and motif discovery. We found that avoiding PCR amplification artifacts, normalizing to input RNA or mRNAseq, and defining the background model from control samples can reduce the bias introduced by RNA abundance and improve the quality of detected binding sites. Our findings can serve as a general guideline for CLIP experiments design and the comprehensive analysis of CLIP-Seq data.

  17. Multidimensional assessment of homework: an analysis of students with ADHD.

    Science.gov (United States)

    Mautone, Jennifer A; Marshall, Stephen A; Costigan, Tracy E; Clarke, Angela T; Power, Thomas J

    2012-10-01

    Homework can have beneficial effects for students; however, it presents challenges, particularly for students with attention problems. Although effective homework interventions exist, intervention development and evaluation has been hampered by the lack of psychometrically sound measures. The primary purpose of this study was to evaluate the construct validity of the Homework Performance Questionnaire (HPQ), Parent and Teacher Versions, in a sample of children with ADHD. A secondary purpose was to examine variations in homework performance as a function of individual characteristics, such as academic achievement, quality of the family-school relationship, and child's diagnostic status. The sample included 91 children (34% female) with ADHD in Grades 2 to 6. Measures included parent and teacher ratings of homework performance and the quality of the parent-teacher relationship as well as direct assessment of child academic achievement and homework performance (i.e., samples of completed assignments). Correlational analyses were used to examine construct validity, and ANOVAs were used to evaluate group differences. Each factor of the HPQ had a significant relationship with other measures of relevant constructs. There were no significant differences in homework performance between groups for ADHD subtype, medication status, or comorbidity, with the exception of learning disability. Children with ADHD and learning disabilities had significantly lower teacher ratings of academic competence. Results of the present study suggest that HPQ scores may be used to make valid inferences about the homework performance of children with attention problems. These rating scales may be helpful in progress monitoring and evaluating intervention effectiveness.

  18. Assessing photoplethysmographic imaging performance beyond facial perfusion analysis

    Science.gov (United States)

    Amelard, Robert; Hughson, Richard L.; Greaves, Danielle K.; Clausi, David A.; Wong, Alexander

    2017-02-01

    Photoplethysmographic imaging (PPGI) systems are relatively new non-contact biophotonic diffuse reflectance systems able to assess arterial pulsations through transient changes in light-tissue interaction. Many PPGI studies have focused on extracting heart rate from the face or hand. Though PPGI systems can be used for widefield imaging of any anatomical area, whole-body investigations are lacking. Here, using a novel PPGI system, coded hemodynamic imaging (CHI), we explored and analyzed the pulsatility at major arterial locations across the whole body, including the neck (carotid artery), arm/wrist (brachial, radial and ulnar arteries), and leg/feet (popliteal and tibial arteries). CHI was positioned 1.5 m from the participant, and diffuse reactance from a broadband tungsten-halogen illumination was filtered using 850{1000 nm bandpass filter for deep tissue penetration. Images were acquired over a highly varying 24-participant sample (11/13 female/male, age 28.7+/-12.4 years, BMI 25.5+/-5.2 kg/m2), and a preliminary case study was performed. B-mode ultrasound images were acquired to validate observations through planar arterial characteristics.

  19. External quality assessment of BRAF molecular analysis in melanoma.

    Science.gov (United States)

    Deans, Zandra C; Wallace, Andrew; O'Sullivan, Brendan; Purvis, Andrew; Camus, Suzanne; Fairley, Jennifer A; Gonzalez, David

    2014-02-01

    The availability of BRAF inhibitors has given metastatic melanoma patients an effective new treatment choice and molecular testing to determine the presence or absence of a BRAF codon 600 mutation is pivotal in the clinical management of these patients. This molecular test must be performed accurately and appropriately to ensure that the patient receives the most suitable treatment in a timely manner. Laboratories have introduced such testing; however, some experience low sample throughput making it critical that an external quality assurance programme is available to help promote a high standard of testing, reporting and provide an educational aspect for BRAF molecular testing. Laboratories took part in three rounds of external quality assessment (EQA) during a 12-month period giving participants a measure of the accuracy of genotyping, clinical interpretation of the result and experience in testing a range of different samples. Formalin fixed paraffin embedded tissue sections from malignant melanoma patients were distributed to participants for BRAF molecular testing. The standard of testing was generally high but distribution of a mutation other than the most common, p.(Val600Glu), highlighted concerns with detection or reporting of the presence of rarer mutations. The main issues raised in the interpretation of the results were the importance of clear unambiguous interpretation of the result tailored to the patient and the understanding that the treatment is different from that given to other stratified medicine programmes. The variability in reporting and wide range of methodologies used indicate a continuing need for EQA in this field.

  20. Assessing the applicability of qualitative comparative analysis for the evaluation of complex projects

    NARCIS (Netherlands)

    S. Verweij (Stefan); L.M. Gerrits (Lasse)

    2012-01-01

    markdownabstract__Abstract__ This paper introduces and assesses Qualitative Comparative Analysis (QCA) as a complexity-informed framework for evaluating complex infrastructure projects coupled with area development. This is done in four steps. First, the properties of infrastructure and area

  1. Cost-effectiveness analysis: adding value to assessment of animal health welfare and production.

    Science.gov (United States)

    Babo Martins, S; Rushton, J

    2014-12-01

    Cost-effectiveness analysis (CEA) has been extensively used in economic assessments in fields related to animal health, namely in human health where it provides a decision-making framework for choices about the allocation of healthcare resources. Conversely, in animal health, cost-benefit analysis has been the preferred tool for economic analysis. In this paper, the use of CEA in related areas and the role of this technique in assessments of animal health, welfare and production are reviewed. Cost-effectiveness analysis can add further value to these assessments, particularly in programmes targeting animal welfare or animal diseases with an impact on human health, where outcomes are best valued in natural effects rather than in monetary units. Importantly, CEA can be performed during programme implementation stages to assess alternative courses of action in real time.

  2. Assessment of current AASHTO LRFD methods for static pile capacity analysis in Rhode Island soils.

    Science.gov (United States)

    2013-07-01

    This report presents an assessment of current AASHTO LRFD methods for static pile capacity analysis in Rhode : Island soils. Current static capacity methods and associated resistance factors are based on pile load test data in sands : and clays. Some...

  3. Cost Analysis of Water Transport for Climate Change Impact Assessment

    Science.gov (United States)

    Szaleniec, V.; Buytaert, W.

    2012-04-01

    It is expected that climate change will have a strong impact on water resources worldwide. Many studies exist that couple the output of global climate models with hydrological models to assess the impact of climate change on physical water availability. However, the water resources topology of many regions and especially that of cities can be very complex. Changes in physical water availability do therefore not translate easily into impacts on water resources for cities. This is especially the case for cities with a complex water supply topology, for instance because of geographical barriers, strong gradients in precipitation patterns, or competing water uses. In this study we explore the use of cost maps to enable the inclusion of water supply topologies in climate change impact studies. We use the city of Lima as a case study. Lima is the second largest desert city in the world. Although Peru as a whole has no water shortage, extreme gradients exist. Most of the economic activities including the city of Lima are located in the coastal desert. This region is geographically disconnected from the wet Amazon basin because of the Andes mountain range. Hence, water supply is precarious, provided by a complex combination of high mountain ecosystems including wetlands and glaciers, as well as groundwater aquifers depending on recharge from the mountains. We investigate the feasibility and costs of different water abstraction scenarios and the impact of climate change using cost functions for different resources. The option of building inter basins tunnels across the Andes is compared to the costs of desalinating seawater from the Pacific Ocean under different climate change scenarios and population growth scenarios. This approach yields recommendations for the most cost-effective options for the future.

  4. Uncertainty analysis for low-level radioactive waste disposal performance assessment at Oak Ridge National Laboratory

    Energy Technology Data Exchange (ETDEWEB)

    Lee, D.W.; Yambert, M.W.; Kocher, D.C.

    1994-12-31

    A performance assessment of the operating Solid Waste Storage Area 6 (SWSA 6) facility for the disposal of low-level radioactive waste at the Oak Ridge National Laboratory has been prepared to provide the technical basis for demonstrating compliance with the performance objectives of DOE Order 5820.2A, Chapter 111.2 An analysis of the uncertainty incorporated into the assessment was performed which addressed the quantitative uncertainty in the data used by the models, the subjective uncertainty associated with the models used for assessing performance of the disposal facility and site, and the uncertainty in the models used for estimating dose and human exposure. The results of the uncertainty analysis were used to interpret results and to formulate conclusions about the performance assessment. This paper discusses the approach taken in analyzing the uncertainty in the performance assessment and the role of uncertainty in performance assessment.

  5. Polygraph Test Results Assessment by Regression Analysis Methods

    Directory of Open Access Journals (Sweden)

    K. A. Leontiev

    2014-01-01

    Full Text Available The paper considers a problem of defining the importance of asked questions for the examinee under judicial and psychophysiological polygraph examination by methods of mathematical statistics. It offers the classification algorithm based on the logistic regression as an optimum Bayesian classifier, considering weight coefficients of information for the polygraph-recorded physiological parameters with no condition for independence of the measured signs.Actually, binary classification is executed by results of polygraph examination with preliminary normalization and standardization of primary results, with check of a hypothesis that distribution of obtained data is normal, as well as with calculation of coefficients of linear regression between input values and responses by method of maximum likelihood. Further, the logistic curve divided signs into two classes of the "significant" and "insignificant" type.Efficiency of model is estimated by means of the ROC analysis (Receiver Operator Characteristics. It is shown that necessary minimum sample has to contain results of 45 measurements at least. This approach ensures a reliable result provided that an expert-polygraphologist possesses sufficient qualification and follows testing techniques.

  6. Uncertainty Assessment of Hydrological Frequency Analysis Using Bootstrap Method

    Directory of Open Access Journals (Sweden)

    Yi-Ming Hu

    2013-01-01

    Full Text Available The hydrological frequency analysis (HFA is the foundation for the hydraulic engineering design and water resources management. Hydrological extreme observations or samples are the basis for HFA; the representativeness of a sample series to the population distribution is extremely important for the estimation reliability of the hydrological design value or quantile. However, for most of hydrological extreme data obtained in practical application, the size of the samples is usually small, for example, in China about 40~50 years. Generally, samples with small size cannot completely display the statistical properties of the population distribution, thus leading to uncertainties in the estimation of hydrological design values. In this paper, a new method based on bootstrap is put forward to analyze the impact of sampling uncertainty on the design value. By bootstrap resampling technique, a large number of bootstrap samples are constructed from the original flood extreme observations; the corresponding design value or quantile is estimated for each bootstrap sample, so that the sampling distribution of design value is constructed; based on the sampling distribution, the uncertainty of quantile estimation can be quantified. Compared with the conventional approach, this method provides not only the point estimation of a design value but also quantitative evaluation on uncertainties of the estimation.

  7. Utility of factor analysis in optimization of resident assessment and faculty evaluation.

    Science.gov (United States)

    Teman, Nicholas R; Minter, Rebecca M; Kasten, Steven J

    2016-06-01

    Increasing focus on more granular assessment in medical education has led to more lengthy instruments, with concern that the increased complexity undermines the utility of these tools. This study evaluated the relative contribution of individual questions in an assessment of resident performance and a faculty performance evaluation by residents. The authors performed factor analysis on the individual items in the resident assessment instrument (3,009 assessments of 71 residents) and faculty evaluations (7,328 evaluations of 61 faculty) collected from 2006 to 2012. Factor analysis of the resident assessment tool revealed that 1 component was responsible for 96.6% of the variance. This component encompassed each question from the assessment form, and could also be termed "overall resident competency." Factor analysis of the attending evaluation form revealed 2 unique components, representing "clinical care" and "interpersonal skills," which accounted for 89.9% of variance. Three components accounted for 90% to 97% of the observed variance in our analysis. Factor analysis represents a useful strategy for analyzing the utility of data obtained from individual items in the assessment and evaluation instruments. Copyright © 2015 Elsevier Inc. All rights reserved.

  8. D.B.H. and Survival Analysis: A New Methodology for Assessing Forest Inventory Mortality

    Science.gov (United States)

    Christopher W. Woodall; Patricia L. Grambsch; William Thomas

    2005-01-01

    Tree mortality has typically been assessed in Forest Inventory and Analysis (FIA) studies through summaries of mortality by location, species, and causal agents. Although these methods have historically been used for most of FIA's tree mortality analyses, they are inadequate for robust assessment of mortality trends and dynamics. To offer a new method of analyzing...

  9. A novel hazard assessment method for biomass gasification stations based on extended set pair analysis.

    Science.gov (United States)

    Yan, Fang; Xu, Kaili; Li, Deshun; Cui, Zhikai

    2017-01-01

    Biomass gasification stations are facing many hazard factors, therefore, it is necessary to make hazard assessment for them. In this study, a novel hazard assessment method called extended set pair analysis (ESPA) is proposed based on set pair analysis (SPA). However, the calculation of the connection degree (CD) requires the classification of hazard grades and their corresponding thresholds using SPA for the hazard assessment. In regard to the hazard assessment using ESPA, a novel calculation algorithm of the CD is worked out when hazard grades and their corresponding thresholds are unknown. Then the CD can be converted into Euclidean distance (ED) by a simple and concise calculation, and the hazard of each sample will be ranked based on the value of ED. In this paper, six biomass gasification stations are introduced to make hazard assessment using ESPA and general set pair analysis (GSPA), respectively. By the comparison of hazard assessment results obtained from ESPA and GSPA, the availability and validity of ESPA can be proved in the hazard assessment for biomass gasification stations. Meanwhile, the reasonability of ESPA is also justified by the sensitivity analysis of hazard assessment results obtained by ESPA and GSPA.

  10. A new tool for risk analysis and assessment in petrochemical plants

    Directory of Open Access Journals (Sweden)

    El-Arkam Mechhoud

    2016-09-01

    Full Text Available The aim of our work was the implementation of a new automated tool dedicated to risk analysis and assessment in petrochemical plants, based on a combination of two analysis methods: HAZOP (HAZard and OPerability and FMEA (Failure Mode and Effect Analysis. Assessment of accident scenarios is also considered. The principal advantage of the two analysis methods is to speed-up hazard identification and risk assessment and forecast the nature and impact of such accidents. Plant parameters are analyzed under a graphical interface to facilitate the exploitation of our developed approach. This automated analysis brings out the different deviations of the operating parameters of any system in the plant. Possible causes of these deviations, their consequences and preventive actions are identified. The result is risk minimization and dependability enhancement of the considered system.

  11. Theoretical assessment of image analysis: statistical vs structural approaches

    Science.gov (United States)

    Lei, Tianhu; Udupa, Jayaram K.

    2003-05-01

    Statistical and structural methods are two major approaches commonly used in image analysis and have demonstrated considerable success. The former is based on statistical properties and stochastic models of the image and the latter utilizes geometric and topological models. In this study, Markov random field (MRF) theory/model based image segmentation and Fuzzy Connectedness (FC) theory/Fuzzy connected objeect delineation are chosen as the representatives for these two approaches, respectively. The comparative study is focused on their theoretical foundations and main operative procedures. The MRF is defined on a lattice and the associated neighborhood system and is based on the Markov property. The FC method is defined on a fuzzy digital space and is based on fuzzy relations. Locally, MRF is characterized by potentials of cliques, and FC is described by fuzzy adjacency and affinity relations. Globally, MRF is characterized by Gibbs distribution, and FC is described by fuzzy connectedness. The task of MRF model based image segmentation is toe seek a realization of the embedded MRF through a two-level operation: partitioning and labeling. The task of FC object delineation is to extract a fuzzy object from a given scene, through a two-step operation: recognition and delineation. Theoretical foundations which underly statistical and structural approaches and the principles of the main operative procedures in image segmentation by these two approaches demonstrate more similarities than differences between them. Two approaches can also complement each other, particularly in seed selection, scale formation, affinity and object membership function design for FC and neighbor set selection and clique potential design for MRF.

  12. Unmanned Aerial Vehicle (UAV) data analysis for fertilization dose assessment

    Science.gov (United States)

    Kavvadias, Antonis; Psomiadis, Emmanouil; Chanioti, Maroulio; Tsitouras, Alexandros; Toulios, Leonidas; Dercas, Nicholas

    2017-10-01

    The growth rate monitoring of crops throughout their biological cycle is very important as it contributes to the achievement of a uniformly optimum production, a proper harvest planning, and reliable yield estimation. Fertilizer application often dramatically increases crop yields, but it is necessary to find out which is the ideal amount that has to be applied in the field. Remote sensing collects spatially dense information that may contribute to, or provide feedback about, fertilization management decisions. There is a potential goal to accurately predict the amount of fertilizer needed so as to attain an ideal crop yield without excessive use of fertilizers cause financial loss and negative environmental impacts. The comparison of the reflectance values at different wavelengths, utilizing suitable vegetation indices, is commonly used to determine plant vigor and growth. Unmanned Aerial Vehicles (UAVs) have several advantages; because they can be deployed quickly and repeatedly, they are flexible regarding flying height and timing of missions, and they can obtain very high-resolution imagery. In an experimental crop field in Eleftherio Larissa, Greece, different dose of pre-plant and in-season fertilization was applied in 27 plots. A total of 102 aerial photos in two flights were taken using an Unmanned Aerial Vehicle based on the scheduled fertilization. Α correlation of experimental fertilization with the change of vegetation indices values and with the increase of the vegetation cover rate during those days was made. The results of the analysis provide useful information regarding the vigor and crop growth rate performance of various doses of fertilization.

  13. Desiccant dehumidification and cooling systems assessment and analysis

    Energy Technology Data Exchange (ETDEWEB)

    Collier, R.K. Jr. [Collier Engineering, Reno, NV (United States)

    1997-09-01

    The objective of this report is to provide a preliminary analysis of the principles, sensitivities, and potential for national energy savings of desiccant cooling and dehumidification systems. The report is divided into four sections. Section I deals with the maximum theoretical performance of ideal desiccant cooling systems. Section II looks at the performance effects of non-ideal behavior of system components. Section III examines the effects of outdoor air properties on desiccant cooling system performance. Section IV analyzes the applicability of desiccant cooling systems to reduce primary energy requirements for providing space conditioning in buildings. A basic desiccation process performs no useful work (cooling). That is, a desiccant material drying air is close to an isenthalpic process. Latent energy is merely converted to sensible energy. Only when heat exchange is applied to the desiccated air is any cooling accomplished. This characteristic is generic to all desiccant cycles and critical to understanding their operation. The analyses of Section I show that desiccant cooling cycles can theoretically achieve extremely high thermal CoP`s (>2). The general conclusion from Section II is that ventilation air processing is the most viable application for the solid desiccant equipment analyzed. The results from the seasonal simulations performed in Section III indicate that, generally, the seasonal performance of the desiccant system does not change significantly from that predicted for outdoor conditions. Results from Section IV show that all of the candidate desiccant systems can save energy relative to standard vapor-compression systems. The largest energy savings are achieved by the enthalpy exchange devise.

  14. A primer on the use of cluster analysis or factor analysis to assess co-occurrence of risk behaviors.

    Science.gov (United States)

    Hofstetter, Hedwig; Dusseldorp, Elise; van Empelen, Pepijn; Paulussen, Theo W G M

    2014-10-01

    The aim of this paper is to provide a guideline to a universal understanding of the analysis of co-occurrence of risk behaviors. The use of cluster analysis and factor analysis was clarified. A theoretical introduction to cluster analysis and factor analysis and examples from literature were provided. A representative sample (N=4395) of the Dutch population, aged 16-40 and participating from fall 2005 to spring 2006, was used to illustrate the use of both techniques in assessing the co-occurrence of risk behaviors. Whereas cluster analysis techniques serve to focus on particular clusters of individuals showing the same behavioral pattern, factor analysis techniques are used to assess possible groups of interrelated health-risk behaviors that can be explained by an unknown common source. Choice between the techniques partly depends on the research question and the aim of the research, and has different implications for inferences and policy. By integrating theory and results from an illustrative example, a guideline has been provided that contributes towards a systematic approach in the assessment of co-occurrence of risk behaviors. Following this guideline, a better comparison between outcomes from various studies is expected, leading to improved effectiveness of multiple behavior change interventions. Copyright © 2014 Elsevier Inc. All rights reserved.

  15. FEBEX II Project Post-mortem analysis EDZ assessment

    Energy Technology Data Exchange (ETDEWEB)

    Bazargan Sabet, B.; Shao, H.; Autio, J.; Elorza, F. J.

    2004-07-01

    variations in the propagation velocities of acoustic waves. A cylindrical block of granite 38.8 cm in diameter and 40 cm high has been analysed along 2D transversal sections in six radial directions. Different inverse tomographic strategies have been used to analyse the measured data, which shown no evidences of the existence of an EDZ in FEBEX gallery. However, a preferential direction in the wave propagation similar to the maximum compression direction of the stress tensor has appeared. As for in situ investigations, the hydraulic connectivity of the drift has been assessed at eleven locations in heated area, including granite matrix and lamprophyre dykes and at six locations in undisturbed zones. In the granite matrix area, pulse test using pressurised air with stepwise pressure increasing was conducted to determine gas entry pressure. In the fractured area, a gas constant flow rate injection test was conducted. Only two locations with higher permeability were detected; one in a natural fracture in the lamprophyre dyke and the other in the interface between lamprophyre and granite. Where numerical investigations are concerned, several analyses of the FEBEX in situ experiment were carried out to determine whether if the generation of a potential EDZ in the surrounding rock was possible or not. Stresses have been calculated by 1D full-coupled thermo-hydromechanical model and by 2D and 3D thermo-mechanical models. Results compared with the available data on compressive strength of the Grimsel granite show that in the worst-case studied, the state of stresses induced by the excavation and the heating phases remains far below the critical curve. (Author)

  16. Assessing Tuberculosis Case Fatality Ratio: A Meta-Analysis

    Science.gov (United States)

    Straetemans, Masja; Glaziou, Philippe; Bierrenbach, Ana L.; Sismanidis, Charalambos; van der Werf, Marieke J.

    2011-01-01

    Background Recently, the tuberculosis (TB) Task Force Impact Measurement acknowledged the need to review the assumptions underlying the TB mortality estimates published annually by the World Health Organization (WHO). TB mortality is indirectly measured by multiplying estimated TB incidence with estimated case fatality ratio (CFR). We conducted a meta-analysis to estimate the TB case fatality ratio in TB patients having initiated TB treatment. Methods We searched for eligible studies in the PubMed and Embase databases through March 4th 2011 and by reference listing of relevant review articles. Main analyses included the estimation of the pooled percentages of: a) TB patients dying due to TB after having initiated TB treatment and b) TB patients dying during TB treatment. Pooled percentages were estimated using random effects regression models on the combined patient population from all studies. Main Results We identified 69 relevant studies of which 22 provided data on mortality due to TB and 59 provided data on mortality during TB treatment. Among HIV infected persons the pooled percentage of TB patients dying due to TB was 9.2% (95% Confidence Interval (CI): 3.7%–14.7%) and among HIV uninfected persons 3.0% (95% CI: −1.2%–7.4%) based on the results of eight and three studies respectively providing data for this analyses. The pooled percentage of TB patients dying during TB treatment was 18.8% (95% CI: 14.8%–22.8%) among HIV infected patients and 3.5% (95% CI: 2.0%–4.92%) among HIV uninfected patients based on the results of 27 and 19 studies respectively. Conclusion The results of the literature review are useful in generating prior distributions of CFR in countries with vital registration systems and have contributed towards revised estimates of TB mortality This literature review did not provide us with all data needed for a valid estimation of TB CFR in TB patients initiating TB treatment. PMID:21738585

  17. Assessing conceptions of cost-benefit analysis among road safety decision-makers : misunderstandings or disputes?

    NARCIS (Netherlands)

    Veisten, K. Elvik, R. & Bax, C.

    2010-01-01

    Statements about economic cost—benefit analysis were assessed in a sample of European road safety decision-makers. These statements related to both principles of cost—benefit analysis and implications for applying the method to road safety projects. A procedure of information reference testing was

  18. A Quality Assessment Tool for Non-Specialist Users of Regression Analysis

    Science.gov (United States)

    Argyrous, George

    2015-01-01

    This paper illustrates the use of a quality assessment tool for regression analysis. It is designed for non-specialist "consumers" of evidence, such as policy makers. The tool provides a series of questions such consumers of evidence can ask to interrogate regression analysis, and is illustrated with reference to a recent study published…

  19. A suite of models to support the quantitative assessment of spread in pest risk analysis

    NARCIS (Netherlands)

    Robinet, C.; Kehlenbeck, H.; Werf, van der W.

    2012-01-01

    In the frame of the EU project PRATIQUE (KBBE-2007-212459 Enhancements of pest risk analysis techniques) a suite of models was developed to support the quantitative assessment of spread in pest risk analysis. This dataset contains the model codes (R language) for the four models in the suite. Three

  20. Different methods for ethical analysis in health technology assessment: an empirical study

    NARCIS (Netherlands)

    Saarni, S.I.; Braunack-Mayer, A.; Hofmann, B.; Wilt, G.J. van der

    2011-01-01

    OBJECTIVES: Ethical analysis can highlight important ethical issues related to implementing a technology, values inherent in the technology itself, and value-decisions underlying the health technology assessment (HTA) process. Ethical analysis is a well-acknowledged part of HTA, yet seldom included

  1. Data Envelopment Analysis to Assess Productivity in the United States Air Force Medical Supply Chain

    Science.gov (United States)

    2011-06-01

    operationalized as application of the appropriate strategy “to get the job done,” was assessed as a function of lead time using average delivery days...37 Data Alibis and Additional Analysis...productivity measurements are given. This research provides AFMOA with an analysis of the current structure as a function of efficiency where sales

  2. Using Citation Analysis Methods to Assess the Influence of Science, Technology, Engineering, and Mathematics Education Evaluations

    Science.gov (United States)

    Greenseid, Lija O.; Lawrenz, Frances

    2011-01-01

    This study explores the use of citation analysis methods to assess the influence of program evaluations conducted within the area of science, technology, engineering, and mathematics (STEM) education. Citation analysis is widely used within scientific research communities to measure the relative influence of scientific research enterprises and/or…

  3. Assessment

    National Research Council Canada - National Science Library

    Gehan Elassal; Mona Elsheikh; Abdel Gawad Abu Zeid

    2014-01-01

    .... Subjects and methods: 80 COPD patients were assessed using SCID for establishing psychiatric diagnosis, Beck depression inventory for assessment of the severity of depressive symptoms, Hamilton anxiety scale...

  4. Seeking missing pieces in science concept assessments: Reevaluating the Brief Electricity and Magnetism Assessment through Rasch analysis

    Directory of Open Access Journals (Sweden)

    Lin Ding

    2014-02-01

    Full Text Available Discipline-based science concept assessments are powerful tools to measure learners’ disciplinary core ideas. Among many such assessments, the Brief Electricity and Magnetism Assessment (BEMA has been broadly used to gauge student conceptions of key electricity and magnetism (E&M topics in college-level introductory physics courses. Differing from typical concept inventories that focus only on one topic of a subject area, BEMA covers a broad range of topics in the electromagnetism domain. In spite of this fact, prior studies exclusively used a single aggregate score to represent individual students’ overall understanding of E&M without explicating the construct of this assessment. Additionally, BEMA has been used to compare traditional physics courses with a reformed course entitled Matter and Interactions (M&I. While prior findings were in favor of M&I, no empirical evidence was sought to rule out possible differential functioning of BEMA that may have inadvertently advantaged M&I students. In this study, we used Rasch analysis to seek two missing pieces regarding the construct and differential functioning of BEMA. Results suggest that although BEMA items generally can function together to measure the same construct of application and analysis of E&M concepts, several items may need further revision. Additionally, items that demonstrate differential functioning for the two courses are detected. Issues such as item contextual features and student familiarity with question settings may underlie these findings. This study highlights often overlooked threats in science concept assessments and provides an exemplar for using evidence-based reasoning to make valid inferences and arguments.

  5. How to assess the Efficiency and "Uncertainty" of Global Sensitivity Analysis?

    Science.gov (United States)

    Haghnegahdar, Amin; Razavi, Saman

    2016-04-01

    Sensitivity analysis (SA) is an important paradigm for understanding model behavior, characterizing uncertainty, improving model calibration, etc. Conventional "global" SA (GSA) approaches are rooted in different philosophies, resulting in different and sometime conflicting and/or counter-intuitive assessment of sensitivity. Moreover, most global sensitivity techniques are highly computationally demanding to be able to generate robust and stable sensitivity metrics over the entire model response surface. Accordingly, a novel sensitivity analysis method called Variogram Analysis of Response Surfaces (VARS) is introduced to overcome the aforementioned issues. VARS uses the Variogram concept to efficiently provide a comprehensive assessment of global sensitivity across a range of scales within the parameter space. Based on the VARS principles, in this study we present innovative ideas to assess (1) the efficiency of GSA algorithms and (2) the level of confidence we can assign to a sensitivity assessment. We use multiple hydrological models with different levels of complexity to explain the new ideas.

  6. Scenario Analysis for the Safety Assessment of Nuclear Waste Repositories: A Critical Review.

    Science.gov (United States)

    Tosoni, Edoardo; Salo, Ahti; Zio, Enrico

    2017-09-05

    A major challenge in scenario analysis for the safety assessment of nuclear waste repositories pertains to the comprehensiveness of the set of scenarios selected for assessing the safety of the repository. Motivated by this challenge, we discuss the aspects of scenario analysis relevant to comprehensiveness. Specifically, we note that (1) it is necessary to make it clear why scenarios usually focus on a restricted set of features, events, and processes; (2) there is not yet consensus on the interpretation of comprehensiveness for guiding the generation of scenarios; and (3) there is a need for sound approaches to the treatment of epistemic uncertainties. © 2017 Society for Risk Analysis.

  7. Underground Test Area Subproject Phase I Data Analysis Task. Volume VIII - Risk Assessment Documentation Package

    Energy Technology Data Exchange (ETDEWEB)

    None

    1996-12-01

    Volume VIII of the documentation for the Phase I Data Analysis Task performed in support of the current Regional Flow Model, Transport Model, and Risk Assessment for the Nevada Test Site Underground Test Area Subproject contains the risk assessment documentation. Because of the size and complexity of the model area, a considerable quantity of data was collected and analyzed in support of the modeling efforts. The data analysis task was consequently broken into eight subtasks, and descriptions of each subtask's activities are contained in one of the eight volumes that comprise the Phase I Data Analysis Documentation.

  8. Alternative model for assessment administration and analysis: An example from the E-CLASS

    CERN Document Server

    Wilcox, Bethany R; Hobbs, Robert D; Aiken, John M; Welch, Nathan M; Lewandowski, H J

    2016-01-01

    The primary model for dissemination of conceptual and attitudinal assessments that has been used within the physics education research (PER) community is to create a high quality, validated assessment, make it available for a wide range of instructors to use, and provide minimal (if any) support to assist with administration or analysis of the results. Here, we present and discuss an alternative model for assessment dissemination, which is characterized by centralized data collection and analysis. This model also provides a greater degree of support for both researchers and instructors. Specifically, we describe our experiences developing a centralized, automated system for an attitudinal assessment we previously created to examine students' epistemologies and expectations about experimental physics. This system provides a proof-of-concept that we use to discuss the advantages associated with centralized administration and data collection for research-based assessments in PER. We also discuss the challenges t...

  9. The role of evidence in humanitarian assessment: the Seed System Security Assessment and the Emergency Market Mapping and Analysis.

    Science.gov (United States)

    Goeldner Byrne, Karri; March, Julie; McGuire, Shawn; Meissner, Laura; Sperling, Louise

    2013-07-01

    This paper reviews advances in the development and use of two evidence-based assessment toolkits: the Seed System Security Assessment (SSSA) and the Emergency Market Mapping and Analysis (EMMA). Both were created in the past five years and have been employed in a range of acute and chronic stress contexts across Africa, Asia, and parts of the Americas, in periods of civil strife, displacement, and drought, as well as following earthquakes, flooding, and political instability. The aims of this paper are threefold: to review advances with regard to each tool; to compare how each toolkit gathers and uses evidence, while considering possibilities for greater complementarity; and to reflect on the nature of 'evidence' used to guide humanitarian response in sudden-onset and chronic crisis situations. A comparison highlights the importance of triangulation and informed analysis for drawing conclusions from imperfect evidence, understanding the limitations of each assessment methodology, and confronting tacit assumptions. © 2013 The Author(s). Journal compilation © Overseas Development Institute, 2013.

  10. Seismic Vulnerability Assessment of a Historical Church: Limit Analysis and Nonlinear Finite Element Analysis

    Directory of Open Access Journals (Sweden)

    G. Castellazzi

    2013-01-01

    Full Text Available The seismic vulnerability of a historical Basilica church located in Italy is studied by means of limit analysis and nonlinear finite element (FE analysis. Attention is posed to the failure mechanisms involving the façade of the church and its interaction with the lateral walls. In particular, the limit analysis and the nonlinear FE analysis provide an estimate of the load collapse multiplier of the failure mechanisms. Results obtained from both approaches are in agreement and can support the selection of possible retrofitting measures to decrease the vulnerability of the church under seismic loads.

  11. Alternative model for administration and analysis of research-based assessments

    Directory of Open Access Journals (Sweden)

    Bethany R. Wilcox

    2016-06-01

    Full Text Available Research-based assessments represent a valuable tool for both instructors and researchers interested in improving undergraduate physics education. However, the historical model for disseminating and propagating conceptual and attitudinal assessments developed by the physics education research (PER community has not resulted in widespread adoption of these assessments within the broader community of physics instructors. Within this historical model, assessment developers create high quality, validated assessments, make them available for a wide range of instructors to use, and provide minimal (if any support to assist with administration or analysis of the results. Here, we present and discuss an alternative model for assessment dissemination, which is characterized by centralized data collection and analysis. This model provides a greater degree of support for both researchers and instructors in order to more explicitly support adoption of research-based assessments. Specifically, we describe our experiences developing a centralized, automated system for an attitudinal assessment we previously created to examine students’ epistemologies and expectations about experimental physics. This system provides a proof of concept that we use to discuss the advantages associated with centralized administration and data collection for research-based assessments in PER. We also discuss the challenges that we encountered while developing, maintaining, and automating this system. Ultimately, we argue that centralized administration and data collection for standardized assessments is a viable and potentially advantageous alternative to the default model characterized by decentralized administration and analysis. Moreover, with the help of online administration and automation, this model can support the long-term sustainability of centralized assessment systems.

  12. Seismic fragility analysis of a nuclear building based on probabilistic seismic hazard assessment and soil-structure interaction analysis

    Energy Technology Data Exchange (ETDEWEB)

    Gonzalez, R.; Ni, S.; Chen, R.; Han, X.M. [CANDU Energy Inc, Mississauga, Ontario (Canada); Mullin, D. [New Brunswick Power, Point Lepreau, New Brunswick (Canada)

    2016-09-15

    Seismic fragility analyses are conducted as part of seismic probabilistic safety assessment (SPSA) for nuclear facilities. Probabilistic seismic hazard assessment (PSHA) has been undertaken for a nuclear power plant in eastern Canada. Uniform Hazard Spectra (UHS), obtained from the PSHA, is characterized by high frequency content which differs from the original plant design basis earthquake spectral shape. Seismic fragility calculations for the service building of a CANDU 6 nuclear power plant suggests that the high frequency effects of the UHS can be mitigated through site response analysis with site specific geological conditions and state-of-the-art soil-structure interaction analysis. In this paper, it is shown that by performing a detailed seismic analysis using the latest technology, the conservatism embedded in the original seismic design can be quantified and the seismic capacity of the building in terms of High Confidence of Low Probability of Failure (HCLPF) can be improved. (author)

  13. Water quality assessment with hierarchical cluster analysis based on Mahalanobis distance.

    Science.gov (United States)

    Du, Xiangjun; Shao, Fengjing; Wu, Shunyao; Zhang, Hanlin; Xu, Si

    2017-07-01

    Water quality assessment is crucial for assessment of marine eutrophication, prediction of harmful algal blooms, and environment protection. Previous studies have developed many numeric modeling methods and data driven approaches for water quality assessment. The cluster analysis, an approach widely used for grouping data, has also been employed. However, there are complex correlations between water quality variables, which play important roles in water quality assessment but have always been overlooked. In this paper, we analyze correlations between water quality variables and propose an alternative method for water quality assessment with hierarchical cluster analysis based on Mahalanobis distance. Further, we cluster water quality data collected form coastal water of Bohai Sea and North Yellow Sea of China, and apply clustering results to evaluate its water quality. To evaluate the validity, we also cluster the water quality data with cluster analysis based on Euclidean distance, which are widely adopted by previous studies. The results show that our method is more suitable for water quality assessment with many correlated water quality variables. To our knowledge, it is the first attempt to apply Mahalanobis distance for coastal water quality assessment.

  14. Assessment of Water Quality Parameters by Using the Multidimensional Scaling Analysis

    Directory of Open Access Journals (Sweden)

    Suheyla Yerel

    2010-01-01

    Full Text Available The surface water quality parameters of the western region of Black Sea in Turkey were assessed by using multidimensional scaling analysis. This technique was applied to the surface water quality parameters obtained from the five monitoring stations. Multidimensional scaling analysis showed that Cl-, SO42-, Na+ and BOD5 are the most important parameters causing difference in the monitoring stations. These analysis results present from the domestic waste and organic pollution affected of surface water quality. Thus, this study represents the usefulness of multidimensional scaling analysis for interpretation in the river pollution problems.

  15. Systems analysis for environmental assessment of urban water and wastewater systems.

    Science.gov (United States)

    Jeppsson, U; Hellström, D

    2002-01-01

    In this paper, two fundamentally different urban wastewater systems are assessed from an environmental impact perspective. One system represents a centralised, high-tech, end-of-pipe structure whereas the second system is primarily based on source-separation strategies. Using material flow analysis in combination with evaluation methods based on life-cycle assessment the systems are evaluated by means of simulation and the results are discussed. A set of priority indicators for environmental assessments of urban water systems is suggested and applied in the analysis. Energy issues are also commented upon. The main intent of the paper is to present the principles of this type of assessment rather than detailed numbers for all possible environmental effects and hazardous substances emitted to air, water and soil. It represents one of several building blocks for a future multi-criteria decision-support system to evaluate urban water management from a sustainability perspective.

  16. A new methodology for assessing health policy and systems research and analysis capacity in African universities.

    Science.gov (United States)

    Lê, Gillian; Mirzoev, Tolib; Orgill, Marsha; Erasmus, Ermin; Lehmann, Uta; Okeyo, Stephen; Goudge, Jane; Maluka, Stephen; Uzochukwu, Benjamin; Aikins, Moses; de Savigny, Don; Tomson, Goran; Gilson, Lucy

    2014-10-08

    The importance of health policy and systems research and analysis (HPSR+A) has been increasingly recognised, but it is still unclear how most effectively to strengthen the capacity of the different organisations involved in this field. Universities are particularly crucial but the expansive literature on capacity development has little to offer the unique needs of HPSR+A activity within universities, and often overlooks the pivotal contribution of capacity assessments to capacity strengthening. The Consortium for Health Policy and Systems Analysis in Africa 2011-2015 designed and implemented a new framework for capacity assessment for HPSR+A within universities. The methodology is reported in detail. Our reflections on developing and conducting the assessment generated four lessons for colleagues in the field. Notably, there are currently no published capacity assessment methodologies for HPSR+A that focus solely on universities - we report a first for the field to initiate the dialogue and exchange of experiences with others. Second, in HPSR+A, the unit of assessment can be a challenge, because HPSR+A groups within universities tend to overlap between academic departments and are embedded in different networks. Third, capacity assessment experience can itself be capacity strengthening, even when taking into account that doing such assessments require capacity. From our experience, we propose that future systematic assessments of HPSR+A capacity need to focus on both capacity assets and needs and assess capacity at individual, organisational, and systems levels, whilst taking into account the networked nature of HPSR+A activity. A genuine partnership process between evaluators and those participating in an assessment can improve the quality of assessment and uptake of results in capacity strengthening.

  17. The Automatic Assessment and Reduction of Noise Using Edge Pattern Analysis in Nonlinear Image Enhancement

    Science.gov (United States)

    Jobson, Daniel J.; Rahman, Zia-ur; Woodells, Glenn A.; Hines, Glenn D.

    2004-01-01

    Noise is the primary visibility limit in the process of non-linear image enhancement, and is no longer a statistically stable additive noise in the post-enhancement image. Therefore novel approaches are needed to both assess and reduce spatially variable noise at this stage in overall image processing. Here we will examine the use of edge pattern analysis both for automatic assessment of spatially variable noise and as a foundation for new noise reduction methods.

  18. Automatic Assessment and Reduction of Noise using Edge Pattern Analysis in Non-Linear Image Enhancement

    Science.gov (United States)

    Jobson, Daniel J.; Rahman, Zia-Ur; Woodell, Glenn A.; Hines, Glenn D.

    2004-01-01

    Noise is the primary visibility limit in the process of non-linear image enhancement, and is no longer a statistically stable additive noise in the post-enhancement image. Therefore novel approaches are needed to both assess and reduce spatially variable noise at this stage in overall image processing. Here we will examine the use of edge pattern analysis both for automatic assessment of spatially variable noise and as a foundation for new noise reduction methods.

  19. Statistical Analysis and Quality Assessment of ChIP-seq Data with DROMPA.

    Science.gov (United States)

    Nakato, Ryuichiro; Shirahige, Katsuhiko

    2018-01-01

    Chromatin immunoprecipitation followed by sequencing (ChIP-seq) analysis can detect protein/DNA-binding and histone-modification sites across an entire genome. As there are various factors during sample preparation that affect the obtained results, multilateral quality assessments are essential. Here, we describe a step-by-step protocol using DROMPA, a program for user-friendly ChIP-seq pipelining. DROMPA can be used for quality assessment, data normalization, visualization, peak calling, and multiple statistical analyses.

  20. Endogenous allergens and compositional analysis in the allergenicity assessment of genetically modified plants.

    Science.gov (United States)

    Fernandez, A; Mills, E N C; Lovik, M; Spoek, A; Germini, A; Mikalsen, A; Wal, J M

    2013-12-01

    Allergenicity assessment of genetically modified (GM) plants is one of the key pillars in the safety assessment process of these products. As part of this evaluation, one of the concerns is to assess that unintended effects (e.g. over-expression of endogenous allergens) relevant for the food safety have not occurred due to the genetic modification. Novel technologies are now available and could be used as complementary and/or alternative methods to those based on human sera for the assessment of endogenous allergenicity. In view of these developments and as a step forward in the allergenicity assessment of GM plants, it is recommended that known endogenous allergens are included in the compositional analysis as additional parameters to be measured. Copyright © 2013 Elsevier Ltd. All rights reserved.

  1. Degradation Assessment and Fault Diagnosis for Roller Bearing Based on AR Model and Fuzzy Cluster Analysis

    Directory of Open Access Journals (Sweden)

    Lingli Jiang

    2011-01-01

    Full Text Available This paper proposes a new approach combining autoregressive (AR model and fuzzy cluster analysis for bearing fault diagnosis and degradation assessment. AR model is an effective approach to extract the fault feature, and is generally applied to stationary signals. However, the fault vibration signals of a roller bearing are non-stationary and non-Gaussian. Aiming at this problem, the set of parameters of the AR model is estimated based on higher-order cumulants. Consequently, the AR parameters are taken as the feature vectors, and fuzzy cluster analysis is applied to perform classification and pattern recognition. Experiments analysis results show that the proposed method can be used to identify various types and severities of fault bearings. This study is significant for non-stationary and non-Gaussian signal analysis, fault diagnosis and degradation assessment.

  2. Elusive Critical Elements of Transformative Risk Assessment Practice and Interpretation: Is Alternatives Analysis the Next Step?

    Science.gov (United States)

    Francis, Royce A

    2015-11-01

    This article argues that "game-changing" approaches to risk analysis must focus on "democratizing" risk analysis in the same way that information technologies have democratized access to, and production of, knowledge. This argument is motivated by the author's reading of Goble and Bier's analysis, "Risk Assessment Can Be a Game-Changing Information Technology-But Too Often It Isn't" (Risk Analysis, 2013; 33: 1942-1951), in which living risk assessments are shown to be "game changing" in probabilistic risk analysis. In this author's opinion, Goble and Bier's article focuses on living risk assessment's potential for transforming risk analysis from the perspective of risk professionals-yet, the game-changing nature of information technologies has typically achieved a much broader reach. Specifically, information technologies change who has access to, and who can produce, information. From this perspective, the author argues that risk assessment is not a game-changing technology in the same way as the printing press or the Internet because transformative information technologies reduce the cost of production of, and access to, privileged knowledge bases. The author argues that risk analysis does not reduce these costs. The author applies Goble and Bier's metaphor to the chemical risk analysis context, and in doing so proposes key features that transformative risk analysis technology should possess. The author also discusses the challenges and opportunities facing risk analysis in this context. These key features include: clarity in information structure and problem representation, economical information dissemination, increased transparency to nonspecialists, democratized manufacture and transmission of knowledge, and democratic ownership, control, and interpretation of knowledge. The chemical safety decision-making context illustrates the impact of changing the way information is produced and accessed in the risk context. Ultimately, the author concludes that although

  3. Application of a Cloud Model-Set Pair Analysis in Hazard Assessment for Biomass Gasification Stations.

    Science.gov (United States)

    Yan, Fang; Xu, Kaili

    2017-01-01

    Because a biomass gasification station includes various hazard factors, hazard assessment is needed and significant. In this article, the cloud model (CM) is employed to improve set pair analysis (SPA), and a novel hazard assessment method for a biomass gasification station is proposed based on the cloud model-set pair analysis (CM-SPA). In this method, cloud weight is proposed to be the weight of index. In contrast to the index weight of other methods, cloud weight is shown by cloud descriptors; hence, the randomness and fuzziness of cloud weight will make it effective to reflect the linguistic variables of experts. Then, the cloud connection degree (CCD) is proposed to replace the connection degree (CD); the calculation algorithm of CCD is also worked out. By utilizing the CCD, the hazard assessment results are shown by some normal clouds, and the normal clouds are reflected by cloud descriptors; meanwhile, the hazard grade is confirmed by analyzing the cloud descriptors. After that, two biomass gasification stations undergo hazard assessment via CM-SPA and AHP based SPA, respectively. The comparison of assessment results illustrates that the CM-SPA is suitable and effective for the hazard assessment of a biomass gasification station and that CM-SPA will make the assessment results more reasonable and scientific.

  4. Assessment of Innovation Competency:A Thematic Analysis of Upper Secondary School Teachers’ Talk

    OpenAIRE

    Nielsen, Jan Alexis

    2015-01-01

    This study employed a three-step qualitative research design with multiple instances of source validation in order to capture expert teachers’ (n=28) reflections on which manifest signs they would look for when they asses students’ innovation competency. This paper reports on the thematic analysis of the recorded talk-in-interaction that occurred in teacher-group discussion sessions at five upper secondary schools. Based on the analysis, it was possible to extrapolate assessment criteria for ...

  5. Instrumented gait analysis for planning and assessment of treatment in cerebral palsy

    OpenAIRE

    Thomas Raji; Ganesh T; George J; Venugopal K; Macaden A; Poonnoose P; Tharion G; Devasahayam S; Bhattacharji S

    2005-01-01

    Background: To improve the walking efficiency of children with cerebral palsy, gait must be documented accurately so that the abnormalities can be assessed and the best treatment option can be selected. Methods: Gait was analysed using Selspot kinematic system with a Kistler force plate and the motion analysis ambulatory EMG system connected to a personal computer. Results: Walking speed and stride length had improved in children. Conclusion: Gait analysis can help to accomplish improv...

  6. Affective sensory tests and internal analysis of preference in acceptability assessment of Brazilian varietal white wines

    OpenAIRE

    BEHRENS, Jorge H.; SILVA, Maria Aparecida A. P. da; WAKELING, Ian N.

    1999-01-01

    The acceptability of nine commercial brazilian varietal white table wines (Riesling, Chardonnay and Gewürztraminer) was evaluated using sensory affective tests. The samples were assessed by 43 consumers of brazilian white wines using he nine-point structured hedonic scale. Judges were recruited based on their responses to a questionnary about consumer?s behavior towards white wines consumption. Subsequently, Analysis of Variance (ANOVA) with means comparision (Tukey test) and Internal Analysi...

  7. Using principal component analysis and annual seasonal trend analysis to assess karst rocky desertification in southwestern China.

    Science.gov (United States)

    Zhang, Zhiming; Ouyang, Zhiyun; Xiao, Yi; Xiao, Yang; Xu, Weihua

    2017-06-01

    Increasing exploitation of karst resources is causing severe environmental degradation because of the fragility and vulnerability of karst areas. By integrating principal component analysis (PCA) with annual seasonal trend analysis (ASTA), this study assessed karst rocky desertification (KRD) within a spatial context. We first produced fractional vegetation cover (FVC) data from a moderate-resolution imaging spectroradiometer normalized difference vegetation index using a dimidiate pixel model. Then, we generated three main components of the annual FVC data using PCA. Subsequently, we generated the slope image of the annual seasonal trends of FVC using median trend analysis. Finally, we combined the three PCA components and annual seasonal trends of FVC with the incidence of KRD for each type of carbonate rock to classify KRD into one of four categories based on K-means cluster analysis: high, moderate, low, and none. The results of accuracy assessments indicated that this combination approach produced greater accuracy and more reasonable KRD mapping than the average FVC based on the vegetation coverage standard. The KRD map for 2010 indicated that the total area of KRD was 78.76 × 103 km2, which constitutes about 4.06% of the eight southwest provinces of China. The largest KRD areas were found in Yunnan province. The combined PCA and ASTA approach was demonstrated to be an easily implemented, robust, and flexible method for the mapping and assessment of KRD, which can be used to enhance regional KRD management schemes or to address assessment of other environmental issues.

  8. Ripening of salami: assessment of colour and aspect evolution using image analysis and multivariate image analysis.

    Science.gov (United States)

    Fongaro, Lorenzo; Alamprese, Cristina; Casiraghi, Ernestina

    2015-03-01

    During ripening of salami, colour changes occur due to oxidation phenomena involving myoglobin. Moreover, shrinkage due to dehydration results in aspect modifications, mainly ascribable to fat aggregation. The aim of this work was the application of image analysis (IA) and multivariate image analysis (MIA) techniques to the study of colour and aspect changes occurring in salami during ripening. IA results showed that red, green, blue, and intensity parameters decreased due to the development of a global darker colour, while Heterogeneity increased due to fat aggregation. By applying MIA, different salami slice areas corresponding to fat and three different degrees of oxidised meat were identified and quantified. It was thus possible to study the trend of these different areas as a function of ripening, making objective an evaluation usually performed by subjective visual inspection. Copyright © 2014 Elsevier Ltd. All rights reserved.

  9. ANALYSIS OF ENVIRONMENTAL FRAGILITY USING MULTI-CRITERIA ANALYSIS (MCE FOR INTEGRATED LANDSCAPE ASSESSMENT

    Directory of Open Access Journals (Sweden)

    Abimael Cereda Junior

    2014-01-01

    Full Text Available The Geographic Information Systems brought greater possibilitie s to the representation and interpretation of the landscap e as well as the integrated a nalysis. However, this approach does not dispense technical and methodological substan tiation for achieving the computational universe. This work is grounded in ecodynamic s and empirical analysis of natural and anthr opogenic environmental Fragility a nd aims to propose and present an integrated paradigm of Multi-criteria Analysis and F uzzy Logic Model of Environmental Fragility, taking as a case study of the Basin of Monjolinho Stream in São Carlos-SP. The use of this methodology allowed for a reduct ion in the subjectivism influences of decision criteria, which factors might have its cartographic expression, respecting the complex integrated landscape.

  10. Predicting child maltreatment: A meta-analysis of the predictive validity of risk assessment instruments.

    Science.gov (United States)

    van der Put, Claudia E; Assink, Mark; Boekhout van Solinge, Noëlle F

    2017-11-01

    Risk assessment is crucial in preventing child maltreatment since it can identify high-risk cases in need of child protection intervention. Despite widespread use of risk assessment instruments in child welfare, it is unknown how well these instruments predict maltreatment and what instrument characteristics are associated with higher levels of predictive validity. Therefore, a multilevel meta-analysis was conducted to examine the predictive accuracy of (characteristics of) risk assessment instruments. A literature search yielded 30 independent studies (N=87,329) examining the predictive validity of 27 different risk assessment instruments. From these studies, 67 effect sizes could be extracted. Overall, a medium significant effect was found (AUC=0.681), indicating a moderate predictive accuracy. Moderator analyses revealed that onset of maltreatment can be better predicted than recurrence of maltreatment, which is a promising finding for early detection and prevention of child maltreatment. In addition, actuarial instruments were found to outperform clinical instruments. To bring risk and needs assessment in child welfare to a higher level, actuarial instruments should be further developed and strengthened by distinguishing risk assessment from needs assessment and by integrating risk assessment with case management. Copyright © 2017 Elsevier Ltd. All rights reserved.

  11. [Study on the Reliability Assessment Method of Heavy Vehicle Gearbox Based on Spectrometric Analysis].

    Science.gov (United States)

    Bao, Ke; Zhang, Zhong; Cao, Yuan-fu; Chen, Yi-jie

    2015-04-01

    Spectrometric oil analysis is of great importance for wear condition monitoring of gearbox. In this context, the contents of main elements compositions in the bench test of heavy vehicle gearbox are obtained by atomic emission spectrometric oil analysis first. Then correlation analysis of the test data and wearing mechanism analysis are carried out to get the metal element which could be used to describe the wearing and failure of the gearbox. The spectrometric data after filling/changing oil are corrected, and the laws of the contents of main elements compositions during tests are expressed as linear functions. After that, the reliability assessment is executed with considering the degradation law and discreteness of test data, in which the mean and standard deviation of normal distribution of spectrometric oil data at each time point are adopted. Finally, the influences of the threshold are discussed. It has been proved that the contents of metal element Cu, which is got by spectrometric oil analysis of different samples, could be used to assess the reliability of heavy vehicle gearbox. The reason is that the metal element Cu is closely related to the general wear state of gearbox, and is easy to be measured. When the threshold of Cu content is treated as a constant, bigger threshold means higher reliability at the same time, and the mean value of threshold has significant impact on the reliability assessment results as R > 0.9. When the threshold is treated as a random variable, bigger dispersion of threshold means smaller slope of reliability against time, and also means lower reliability of gearbox as R > 0.9 at the same time. In this study, the spectrometric oil analysis and probability statistics are used together for the reliability assessment of gear box, which extends the application range of spectrometric analysis.

  12. A Critical Examination of the Assessment Analysis Capabilities of OCLC ACAS

    Science.gov (United States)

    Lyons, Lucy E.

    2005-01-01

    Over 500 libraries have employed OCLC's iCAS and its successor Automated Collection Assessment and Analysis Services (ACAS) as bibliometric tools to evaluate monograph collections. This examination of ACAS reveals both its methodological limitations and its feasibility as an indicator of collecting patterns. The results can be used to maximize the…

  13. Life cycle based dynamic assessment coupled with multiple criteria decision analysis

    DEFF Research Database (Denmark)

    Sohn, Joshua; Kalbar, Pradip; Birkved, Morten

    2017-01-01

    This work looks at coupling Life cycle assessment (LCA) with a dynamic inventory and multiple criteria decision analysis (MCDA) to improve the validity and reliability of single score results for complex systems. This is done using the case study of a representative Danish single family home over...

  14. An assessment of interventions for improving communication and trust in cost benefit analysis processes

    NARCIS (Netherlands)

    Beukers, E.; Bertolini, L.; te Brömmelstroet, M.

    2014-01-01

    Evaluation literature suggests that assessments of integrated transport plans should be an inclusive dialogue, for which it is crucial that participants communicate with and trust each other. However, cost benefit analysis (CBA) of integrated transport plans is often characterized by communication

  15. A Comparison of Functional Behavioral Assessment and Functional Analysis Methodology among Students with Mild Disabilities

    Science.gov (United States)

    Lewis, Timothy J.; Mitchell, Barbara S.; Harvey, Kristin; Green, Ambra; McKenzie, Jennifer

    2015-01-01

    Functional behavioral assessment (FBA) and functional analyses (FA) are grounded in the applied behavior analysis principle that posits problem behavior is functionally related to the environment in which it occurs and is maintained by either providing access to reinforcing outcomes or allowing the individual to avoid or escape that which they…

  16. Assessing Professional Competence by Using Occupational Judgment Tests Derived From Job Analysis Questionnaires

    Science.gov (United States)

    2009-01-01

    these scales may provide indices of procedural knowledge and attitudinal factors that underlie job performance and employee retention (cf. Campbell...knowledge tests, OJTs may be better suited to assess knowledge that underlies procedural performance, career continuance attitudes, and employee ... retention . Meta-analysis has demonstrated that situational judgment tests (SJTs) are correlated with both job performance, ρ = .34, and general cognitive

  17. Morphological analysis in school-age children: dynamic assessment of a word learning strategy.

    Science.gov (United States)

    Larsen, Jennifer A; Nippold, Marilyn A

    2007-07-01

    Morphological analysis is the ability to use knowledge of root words and affixes to determine the meanings of unfamiliar, morphologically complex words. Beginning in the early elementary grades and continuing into the college years, it is a primary strategy that is used to increase one's knowledge of difficult vocabulary. The purpose of this study was to investigate how well school-age children could use morphological analysis to explain word meanings. The study was also designed to examine individual differences in this domain in relation to children's broader literacy skills. The ability of 50 typically developing sixth-grade children to explain the meanings of 15 low-frequency morphologically complex words was measured using a dynamic assessment procedure. Children were individually interviewed and were asked to define each word. As needed, varying degrees of adult scaffolding were provided. Children were also assessed for their word knowledge and reading comprehension. Performance on the dynamic assessment task revealed a wide range of skill levels in these children and was positively related to the children's literacy levels. Although some children readily used morphological analysis to explain the meanings of unfamiliar words, others required greater amounts of adult scaffolding to be successful. Suggestions are offered for ways to employ dynamic assessment to examine children's ability to use morphological analysis as a word learning strategy.

  18. Applying Uncertainty Analysis to a Risk Assessment for the Pesticide Permethrin

    Science.gov (United States)

    We discuss the application of methods of uncertainty analysis from our previous poster to the problem of a risk assessment for exposure to the food-use pesticide permethrin resulting from residential pesticide crack and crevice application. Exposures are simulated by the SHEDS (S...

  19. PHYTOPLANKTON PIGMENT ANALYSIS BY HPLC FOR ASSESSING COMMUNITY COMPOSITION IN THE LAURENTIAN GREAT LAKES

    Science.gov (United States)

    A technique to rapidly assess phytoplankton dynamics is being evaluated for its utility in the Great Lakes. Comparison to traditional microscopic techniques and to more recent in-situ FluoroProbe technology will allow us to determine if HPLC pigment analysis can provide unique a...

  20. Meta-Analysis of Studies Assessing the Efficacy of Projective Techniques in Discriminating Child Sexual Abuse.

    Science.gov (United States)

    West, Maxine M.

    1998-01-01

    This meta-analysis of 12 studies assessed the efficacy of projective techniques to discriminate between sexually abused and nonsexually abused children. Results indicated the techniques were able to discriminate distressed children from nondistressed subjects, but were less able to discriminate the type of distress. (Author/DB)

  1. Cluster Analysis of Assessment in Anatomy and Physiology for Health Science Undergraduates

    Science.gov (United States)

    Brown, Stephen; White, Sue; Power, Nicola

    2016-01-01

    Academic content common to health science programs is often taught to a mixed group of students; however, content assessment may be consistent for each discipline. This study used a retrospective cluster analysis on such a group, first to identify high and low achieving students, and second, to determine the distribution of students within…

  2. Investigating Faculty Familiarity with Assessment Terminology by Applying Cluster Analysis to Interpret Survey Data

    Science.gov (United States)

    Raker, Jeffrey R.; Holme, Thomas A.

    2014-01-01

    A cluster analysis was conducted with a set of survey data on chemistry faculty familiarity with 13 assessment terms. Cluster groupings suggest a high, middle, and low overall familiarity with the terminology and an independent high and low familiarity with terms related to fundamental statistics. The six resultant clusters were found to be…

  3. Dynamic Ecocentric Assessment Combining Emergy and Data Envelopment Analysis: Application to Wind Farms

    Directory of Open Access Journals (Sweden)

    Mario Martín-Gamboa

    2016-01-01

    Full Text Available Most of current life-cycle approaches show an anthropocentric standpoint for the evaluation of human-dominated activities. However, this perspective is insufficient when it comes to assessing the contribution of natural resources to production processes. In this respect, emergy analysis evaluates human-driven systems from a donor-side perspective, accounting for the environmental effort performed to make the resources available. This article presents a novel methodological framework, which combines emergy analysis and dynamic Data Envelopment Analysis (DEA for the ecocentric assessment of multiple resembling entities over an extended period of time. The use of this approach is shown through a case study of wind energy farms. Furthermore, the results obtained are compared with those of previous studies from two different angles. On the one hand, a comparison with results from anthropocentric approaches (combined life cycle assessment and DEA is drawn. On the other hand, results from similar ecocentric approaches, but without a dynamic model, are also subject to comparison. The combined use of emergy analysis and dynamic DEA is found to be a valid methodological framework for the computation of resource efficiency and the valuation of ecosystem services. It complements traditional anthropocentric assessments while appropriately including relevant time effects.

  4. Assessing Model Fit: Caveats and Recommendations for Confirmatory Factor Analysis and Exploratory Structural Equation Modeling

    Science.gov (United States)

    Perry, John L.; Nicholls, Adam R.; Clough, Peter J.; Crust, Lee

    2015-01-01

    Despite the limitations of overgeneralizing cutoff values for confirmatory factor analysis (CFA; e.g., Marsh, Hau, & Wen, 2004), they are still often employed as golden rules for assessing factorial validity in sport and exercise psychology. The purpose of this study was to investigate the appropriateness of using the CFA approach with these…

  5. Designing student peer assessment in higher education: Analysis of written and oral peer feedback

    NARCIS (Netherlands)

    van den Berg, I.|info:eu-repo/dai/nl/288125797; Admiraal, W.|info:eu-repo/dai/nl/120226340; Pilot, A.|info:eu-repo/dai/nl/068350880

    2006-01-01

    Designing student peer assessment in higher education: analysis of written and oral peer feedback Relating it to design features, the present article describes the nature of written and oral peer feedback as it occurred in seven writing courses, each with a different PA design. Results indicate that

  6. Applying HAZOP analysis in assessing remote handling compatibility of ITER port plugs

    NARCIS (Netherlands)

    Duisings, L. P. M.; van Til, S.; Magielsen, A. J.; Ronden, D. M. S.; Elzendoorn, B. S. Q.; Heemskerk, C. J. M.

    2013-01-01

    This paper describes the application of a Hazard and Operability Analysis (HAZOP) methodology in assessing the criticality of remote handling maintenance activities on port plugs in the ITER Hot Cell facility. As part of the ECHUL consortium, the remote handling team at the DIFFER Institute is

  7. Reliability of ^1^H NMR analysis for assessment of lipid oxidation at frying temperatures

    Science.gov (United States)

    The reliability of a method using ^1^H NMR analysis for assessment of oil oxidation at a frying temperature was examined. During heating and frying at 180 °C, changes of soybean oil signals in the ^1^H NMR spectrum including olefinic (5.16-5.30 ppm), bisallylic (2.70-2.88 ppm), and allylic (1.94-2.1...

  8. Self-evaluation of assessment programs: A cross-case analysis

    NARCIS (Netherlands)

    Baartman, Liesbeth; Prins, Frans; Kirschner, Paul A.; Van der Vleuten, Cees

    2011-01-01

    Baartman, L. K. J., Prins, F. J., Kirschner, P. A., & Van der Vleuten, C. P. M. (2011). Self-evaluation of assessment programs: A cross-case analysis. Evaluation and Program Planning, 34, 206-216. doi: 10.1016/j.evalprogplan.2011.03.001

  9. MONTHLY VARIATION IN SPERM MOTILITY IN COMMON CARP ASSESSED USING COMPUTER-ASSISTED SPERM ANALYSIS (CASA)

    Science.gov (United States)

    Sperm motility variables from the milt of the common carp Cyprinus carpio were assessed using a computer-assisted sperm analysis (CASA) system across several months (March-August 1992) known to encompass the natural spawning period. Two-year-old pond-raised males obtained each mo...

  10. Local Assessment: Using Genre Analysis to Validate Directed Self-Placement

    Science.gov (United States)

    Gere, Anne Ruggles; Aull, Laura; Escudero, Moises Damian Perales; Lancaster, Zak; Lei, Elizabeth Vander

    2013-01-01

    Grounded in the principle that writing assessment should be locally developed and controlled, this article describes a study that contextualizes and validates the decisions that students make in the modified Directed Self-Placement (DSP) process used at the University of Michigan. The authors present results of a detailed text analysis of…

  11. Assessing Reading Skill with a Think-Aloud Procedure and Latent Semantic Analysis.

    Science.gov (United States)

    Magliano, Joseph P.; Millis, Keith K.

    2003-01-01

    Two studies examined the viability of assessing reading strategies using a think-aloud protocol combined with latent semantic analysis (LSA). Findings demonstrated that the responses of less-skilled readers semantically overlapped more with focal sentences than with causal antecedent sentences, whereas skilled readers' responses overlapped with…

  12. Assessing Heterogeneity in Meta-Analysis: Q Statistic or I[squared] Index?

    Science.gov (United States)

    Huedo-Medina, Tania B.; Sanchez-Meca, Julio; Marin-Martinez, Fulgencio; Botella, Juan

    2006-01-01

    In meta-analysis, the usual way of assessing whether a set of single studies is homogeneous is by means of the Q test. However, the Q test only informs meta-analysts about the presence versus the absence of heterogeneity, but it does not report on the extent of such heterogeneity. Recently, the I[squared] index has been proposed to quantify the…

  13. Assessment of Smolt Condition for Travel Time Analysis, 1993-1994 Annual Report.

    Energy Technology Data Exchange (ETDEWEB)

    Schrock, Robin M; Beeman, John W; VanderKooi, Scott P [US Geological Survey, Western Fisheries Research Center, Columbia River Research Laboratory, Cook, WA

    1999-02-01

    The assessment of smolt condition for travel time analysis (ASCTTA) project provided information on the level of smoltification in Columbia River hatchery and wild salmonid stocks to the Fish Passage Center (FPC), for the primary purpose of in-river management of flows.

  14. Assessing model fit in latent class analysis when asymptotics do not hold

    NARCIS (Netherlands)

    van Kollenburg, Geert H.; Mulder, Joris; Vermunt, Jeroen K.

    2015-01-01

    The application of latent class (LC) analysis involves evaluating the LC model using goodness-of-fit statistics. To assess the misfit of a specified model, say with the Pearson chi-squared statistic, a p-value can be obtained using an asymptotic reference distribution. However, asymptotic p-values

  15. Identifying Barriers in Implementing Outcomes-Based Assessment Program Review: A Grounded Theory Analysis

    Science.gov (United States)

    Bresciani, Marilee J.

    2011-01-01

    The purpose of this grounded theory study was to identify the typical barriers encountered by faculty and administrators when implementing outcomes-based assessment program review. An analysis of interviews with faculty and administrators at nine institutions revealed a theory that faculty and administrators' promotion, tenure (if applicable),…

  16. Use of USDA forest inventory and analysis data to assess oak tree health in Minnesota

    Science.gov (United States)

    Kathryn W. Kromroy; Jennifer Juzwik; Paul D. Castillo

    2003-01-01

    As a precursor to a regional assessment for the Upper Midwest, three variables were examined as measures of oak health in Minnesota between 1974 and 1990 using USDA Forest Service Inventory and Analysis data. Mortality was 6 percent in the 1986-1990 inventory based on numbers of dead oaks per total oaks on plots with...

  17. Computational Psycholinguistic Analysis and Its Application in Psychological Assessment of College Students

    Science.gov (United States)

    Kucera, Dalibor; Havigerová, Jana M.

    2015-01-01

    The paper deals with the issue of computational psycholinguistic analysis (CPA) and its experimental application in basic psychological and pedagogical assessment. CPA is a new method which may potentially provide interesting, psychologically relevant information about the author of a particular text, regardless of the text's factual (semantic)…

  18. A Bayesian latent group analysis for detecting poor effort in the assessment of malingering

    NARCIS (Netherlands)

    Ortega, A.; Wagenmakers, E.-J.; Lee, M.D.; Markowitsch, H.J.; Piefke, M.

    2012-01-01

    Despite their theoretical appeal, Bayesian methods for the assessment of poor effort and malingering are still rarely used in neuropsychological research and clinical diagnosis. In this article, we outline a novel and easy-to-use Bayesian latent group analysis of malingering whose goal is to

  19. Assessing soil compaction on Forest Inventory & Analysis phase 3 field plots using a pocket penetrometer

    Science.gov (United States)

    Michael C. Amacher; Katherine P. O' Neill

    2004-01-01

    Soil compaction is an important indicator of soil quality, yet few practical methods are available to quantitatively measure this variable. Although an assessment of the areal extent of soil compaction is included as part of the soil indicator portion of the Forest Inventory & Analysis (FIA) program, no quantitative measurement of the degree of soil compaction...

  20. An Interactive Assessment Framework for Visual Engagement: Statistical Analysis of a TEDx Video

    Science.gov (United States)

    Farhan, Muhammad; Aslam, Muhammad

    2017-01-01

    This study aims to assess the visual engagement of the video lectures. This analysis can be useful for the presenter and student to find out the overall visual attention of the videos. For this purpose, a new algorithm and data collection module are developed. Videos can be transformed into a dataset with the help of data collection module. The…

  1. Analysis And Assessment Of The Security Method Against Incidental Contamination In The Collective Water Supply System

    Directory of Open Access Journals (Sweden)

    Szpak Dawid

    2015-09-01

    Full Text Available The paper presents the main types of surface water incidental contaminations and the security method against incidental contamination in water sources. Analysis and assessment the collective water supply system (CWSS protection against incidental contamination was conducted. Failure Mode and Effects Analysis (FMEA was used. The FMEA method allow to use the product or process analysis, identification of weak points, and implementation the corrections and new solutions for eliminating the source of undesirable events. The developed methodology was shown in application case. It was found that the risk of water contamination in water-pipe network of the analyzed CWSS caused by water source incidental contamination is at controlled level.

  2. A multicriteria decision analysis model and risk assessment framework for carbon capture and storage.

    Science.gov (United States)

    Humphries Choptiany, John Michael; Pelot, Ronald

    2014-09-01

    Multicriteria decision analysis (MCDA) has been applied to various energy problems to incorporate a variety of qualitative and quantitative criteria, usually spanning environmental, social, engineering, and economic fields. MCDA and associated methods such as life-cycle assessments and cost-benefit analysis can also include risk analysis to address uncertainties in criteria estimates. One technology now being assessed to help mitigate climate change is carbon capture and storage (CCS). CCS is a new process that captures CO2 emissions from fossil-fueled power plants and injects them into geological reservoirs for storage. It presents a unique challenge to decisionmakers (DMs) due to its technical complexity, range of environmental, social, and economic impacts, variety of stakeholders, and long time spans. The authors have developed a risk assessment model using a MCDA approach for CCS decisions such as selecting between CO2 storage locations and choosing among different mitigation actions for reducing risks. The model includes uncertainty measures for several factors, utility curve representations of all variables, Monte Carlo simulation, and sensitivity analysis. This article uses a CCS scenario example to demonstrate the development and application of the model based on data derived from published articles and publicly available sources. The model allows high-level DMs to better understand project risks and the tradeoffs inherent in modern, complex energy decisions. © 2014 Society for Risk Analysis.

  3. [A guide on the use of factor analysis in the assessment of construct validity].

    Science.gov (United States)

    Kang, Hyuncheol

    2013-10-01

    The purpose of this study is to provide researchers with a simplified approach to undertaking exploratory factor analysis for the assessment of construct validity. All articles published in 2010, 2011, and 2012 in Journal of Korean Academy of Nursing were reviewed and other relevant books and articles were chosen for the review. In this paper, the following were discussed: preliminary analysis process of exploratory factor analysis to examine the sample size, distribution of measured variables, correlation coefficient, and results of KMO measure and Bartlett's test of sphericity. In addition, other areas to be considered in using factor analysis are discussed, including determination of the number of factors, the choice of rotation method or extraction method of the factor structure, and the interpretation of the factor loadings and explained variance. Content validity is the degree to which elements of an assessment instrument are relevant to and representative of the targeted construct for a particular assessment purpose. This measurement is difficult and challenging and takes a lot of time. Factor analysis is considered one of the strongest approaches to establishing construct validity and is the most commonly used method for establishing construct validity measured by an instrument.

  4. Peak Pc Prediction in Conjunction Analysis: Conjunction Assessment Risk Analysis. Pc Behavior Prediction Models

    Science.gov (United States)

    Vallejo, J.J.; Hejduk, M.D.; Stamey, J. D.

    2015-01-01

    Satellite conjunction risk typically evaluated through the probability of collision (Pc). Considers both conjunction geometry and uncertainties in both state estimates. Conjunction events initially discovered through Joint Space Operations Center (JSpOC) screenings, usually seven days before Time of Closest Approach (TCA). However, JSpOC continues to track objects and issue conjunction updates. Changes in state estimate and reduced propagation time cause Pc to change as event develops. These changes a combination of potentially predictable development and unpredictable changes in state estimate covariance. Operationally useful datum: the peak Pc. If it can reasonably be inferred that the peak Pc value has passed, then risk assessment can be conducted against this peak value. If this value is below remediation level, then event intensity can be relaxed. Can the peak Pc location be reasonably predicted?

  5. Empirical assessment of analysis workflows for differential expression analysis of human samples using RNA-Seq.

    Science.gov (United States)

    Williams, Claire R; Baccarella, Alyssa; Parrish, Jay Z; Kim, Charles C

    2017-01-17

    RNA-Seq has supplanted microarrays as the preferred method of transcriptome-wide identification of differentially expressed genes. However, RNA-Seq analysis is still rapidly evolving, with a large number of tools available for each of the three major processing steps: read alignment, expression modeling, and identification of differentially expressed genes. Although some studies have benchmarked these tools against gold standard gene expression sets, few have evaluated their performance in concert with one another. Additionally, there is a general lack of testing of such tools on real-world, physiologically relevant datasets, which often possess qualities not reflected in tightly controlled reference RNA samples or synthetic datasets. Here, we evaluate 219 combinatorial implementations of the most commonly used analysis tools for their impact on differential gene expression analysis by RNA-Seq. A test dataset was generated using highly purified human classical and nonclassical monocyte subsets from a clinical cohort, allowing us to evaluate the performance of 495 unique workflows, when accounting for differences in expression units and gene- versus transcript-level estimation. We find that the choice of methodologies leads to wide variation in the number of genes called significant, as well as in performance as gauged by precision and recall, calculated by comparing our RNA-Seq results to those from four previously published microarray and BeadChip analyses of the same cell populations. The method of differential gene expression identification exhibited the strongest impact on performance, with smaller impacts from the choice of read aligner and expression modeler. Many workflows were found to exhibit similar overall performance, but with differences in their calibration, with some biased toward higher precision and others toward higher recall. There is significant heterogeneity in the performance of RNA-Seq workflows to identify differentially expressed genes

  6. Integrating Life-cycle Assessment into Transport Cost-benefit Analysis

    DEFF Research Database (Denmark)

    Manzo, Stefano; Salling, Kim Bang

    2016-01-01

    Traditional transport Cost-Benefit Analysis (CBA) commonly ignores the indirect environmental impacts of an infrastructure project deriving from the overall life-cycle of the different project components. Such indirect impacts are instead of key importance in order to assess the long......-term sustainability of a transport infrastructure project. In the present study we suggest to overcome this limit by combining a conventional life-cycle assessment approach with standard transport cost-benefit analysis. The suggested methodology is tested upon a case study project related to the construction of a new...... fixed link across the Roskilde fjord in Frederikssund (Denmark). The results are then compared with those from a standard CBA framework. The analysis shows that indirect environmental impacts represent a relevant share of the estimated costs of the project, clearly affecting the final project evaluation...

  7. Syntax and reading comprehension: a meta-analysis of different spoken-syntax assessments.

    Science.gov (United States)

    Brimo, Danielle; Lund, Emily; Sapp, Alysha

    2017-12-18

    Syntax is a language skill purported to support children's reading comprehension. However, researchers who have examined whether children with average and below-average reading comprehension score significantly different on spoken-syntax assessments report inconsistent results. To determine if differences in how syntax is measured affect whether children with average and below-average reading comprehension score significantly different on spoken-syntax assessments. Studies that included a group comparison design, children with average and below-average reading comprehension, and a spoken-syntax assessment were selected for review. Fourteen articles from a total of 1281 reviewed met the inclusionary criteria. The 14 articles were coded for the age of the children, score on the reading comprehension assessment, type of spoken-syntax assessment, type of syntax construct measured and score on the spoken-syntax assessment. A random-effects model was used to analyze the difference between the effect sizes of the types of spoken-syntax assessments and the difference between the effect sizes of the syntax construct measured. There was a significant difference between children with average and below-average reading comprehension on spoken-syntax assessments. Those with average and below-average reading comprehension scored significantly different on spoken-syntax assessments when norm-referenced and researcher-created assessments were compared. However, when the type of construct was compared, children with average and below-average reading comprehension scored significantly different on assessments that measured knowledge of spoken syntax, but not on assessments that measured awareness of spoken syntax. The results of this meta-analysis confirmed that the type of spoken-syntax assessment, whether norm-referenced or researcher-created, did not explain why some researchers reported that there were no significant differences between children with average and below

  8. Evaluation of SOVAT: An OLAP-GIS decision support system for community health assessment data analysis

    Directory of Open Access Journals (Sweden)

    Parmanto Bambang

    2008-06-01

    Full Text Available Abstract Background Data analysis in community health assessment (CHA involves the collection, integration, and analysis of large numerical and spatial data sets in order to identify health priorities. Geographic Information Systems (GIS enable for management and analysis using spatial data, but have limitations in performing analysis of numerical data because of its traditional database architecture. On-Line Analytical Processing (OLAP is a multidimensional datawarehouse designed to facilitate querying of large numerical data. Coupling the spatial capabilities of GIS with the numerical analysis of OLAP, might enhance CHA data analysis. OLAP-GIS systems have been developed by university researchers and corporations, yet their potential for CHA data analysis is not well understood. To evaluate the potential of an OLAP-GIS decision support system for CHA problem solving, we compared OLAP-GIS to the standard information technology (IT currently used by many public health professionals. Methods SOVAT, an OLAP-GIS decision support system developed at the University of Pittsburgh, was compared against current IT for data analysis for CHA. For this study, current IT was considered the combined use of SPSS and GIS ("SPSS-GIS". Graduate students, researchers, and faculty in the health sciences at the University of Pittsburgh were recruited. Each round consisted of: an instructional video of the system being evaluated, two practice tasks, five assessment tasks, and one post-study questionnaire. Objective and subjective measurement included: task completion time, success in answering the tasks, and system satisfaction. Results Thirteen individuals participated. Inferential statistics were analyzed using linear mixed model analysis. SOVAT was statistically significant (α = .01 from SPSS-GIS for satisfaction and time (p Conclusion Using SOVAT, tasks were completed more efficiently, with a higher rate of success, and with greater satisfaction, than the

  9. Analysis of health economics assessment reports for pharmaceuticals in France - understanding the underlying philosophy of CEESP assessment.

    Science.gov (United States)

    Toumi, Mondher; Motrunich, Anastasiia; Millier, Aurélie; Rémuzat, Cécile; Chouaid, Christos; Falissard, Bruno; Aballéa, Samuel

    2017-01-01

    Background : Despite the guidelines for Economic and Public Health Assessment Committee (CEESP) submission having been available for nearly six years, the dossiers submitted continue to deviate from them, potentially impacting product prices. Objective : to review the reports published by CEESP, analyse deviations from the guidelines, and discuss their implications for the pricing and reimbursement process. Study design : CEESP reports published until January 2017 were reviewed, and deviations from the guidelines were extracted. The frequency of deviations was described by type of methodological concern (minor, important or major). Results : In 19 reports, we identified 243 methodological concerns, most often concerning modelling, measurement and valuation of health states and results presentation and sensitivity analyses; nearly 63% were minor, 33% were important and 4.5% were major. All reports included minor methodological concerns, and 17 (89%) included at least one important and/or major methodological concern. Global major methodological concerns completely invalidated the analysis in seven dossiers (37%). Conclusion : The CEESP submission dossiers fail to adhere to the guidelines, potentially invalidating the health economics analysis and resulting in pricing negotiations. As these negotiations tend to be unfavourable for the manufacturer, the industry should strive to improve the quality of the analyses submitted to CEESP.

  10. Using the Ages and Stages Questionnaire to teach medical students developmental assessment: a descriptive analysis

    Directory of Open Access Journals (Sweden)

    Nicol Pam

    2006-05-01

    Full Text Available Abstract Background After a survey of medical graduates' skills found a lack of confidence in developmental assessment, a program was introduced with the broad aims of increasing medical student confidence and respect for the parents' role in childhood developmental assessment. Research has shown that parents' concerns are as accurate as quality screening tests in assessing development, so the program utilised the Ages and Stages Questionnaire, a parent completed, child development assessment tool. Method To evaluate the program, an interpretative analysis was completed on the students' reports written during the program and a questionnaire was administered to the parents to gain their perception of the experience. As well, student confidence levels in assessing growth and development were measured at the end of the paediatric term. Results Although there was an increase in student confidence in developmental assessment at the end of the term, it was not statistically significant. However the findings indicated that students gained increased understanding of the process and enhanced recognition of the parental role, and the study suggested there was increased confidence in some students. Parents indicated that they thought they should be involved in the teaching of students. Conclusion The ASQ was shown to have been useful in an education program at the level of advanced beginners in developmental assessment.

  11. Comparison between visual analysis and microscope assessment of surgical instrument cleanliness from sterile service departments.

    Science.gov (United States)

    Lipscomb, I P; Sihota, A K; Keevil, C W

    2008-01-01

    Modern hospital sterile service departments (SSDs) routinely inspect instruments visually to assess the degree of contamination present after the washer/disinfector cycle. We aimed to test the effectiveness and reliability of this method. Surgical instrument sets were obtained from nine anonymous National Health Service (NHS) Primary Care Trust SSDs to investigate the efficacy of 'in-place' cleaning procedures. The instruments were first inspected visually, followed by a novel technique called episcopic differential interference contrast microscopy. This was combined with a sensitive fluorescent reagent, SYPRO Ruby, to rapidly visualise and assess contamination levels on the medical devices. The application of a Contamination Index (0-4) for both proteinaceous and non-proteinaceous deposits on the surface allowed quantitative assessment. Close correlation was seen for simple instruments between visual assessment and microscopic analysis. For more complex instruments, however, there was a marked difference between the two assessment techniques and the microscopy procedure showed areas of proteinaceous and non-proteinaceous crystalline soiling that was difficult or impossible to see by eye. Visual assessment of cleaning is fraught with possible error. This survey shows how large amounts of contamination could pass undetected using such antiquated methods. The new methodology applied in the assessment of surface contamination is rapid and generally applicable and could be used more widely for routine monitoring of instrument cleanliness.

  12. Viability assessment of Ascaris suum eggs stained with fluorescent dyes using digital colorimetric analysis.

    Science.gov (United States)

    Włodarczyk, Magdalena; Zdybel, Jolanta; Próchniak, Marek; Osiński, Zbigniew; Karamon, Jacek; Kłapeć, Teresa; Cencek, Tomasz

    2017-07-01

    The aim of the study was to develop a method for the colorimetric evaluation of nematode eggs using appropriate instruments. The materials for the study were live and dead (inactivated) eggs of the Ascaris suum. Viability of the eggs was assessed using four different kits for fluorescent staining (for each technique, a series of photos were taken). Images of stained eggs were analysed using graphic software with RGB (red-green-blue) function. The viability of the eggs was assessed according to the relative positions of the distributions of colour intensities of live or dead eggs - distributions area's overlap index (DAOI), and distributions area's separation index (DASI) were calculated. Computer analysis of the intensity of green colour was not satisfactory. However, analysis of images in the spectrum of red colour proved useful for the effective differentiation between live or dead eggs. The best parameters were observed using the Annexin V FITC Apoptosis Detection Kit (DASI = 41 and 67). The investigation confirmed the usefulness of fluorescent dyes used in conjunction with digital analysis for the assessment of the viability of A. suum eggs. The use of computer software allowed a better objectivity of the assessment, especially in the case of doubtful staining. Copyright © 2017 Elsevier Inc. All rights reserved.

  13. Strategic Environmental Assessment Framework for Landscape-Based, Temporal Analysis of Wetland Change in Urban Environments

    Science.gov (United States)

    Sizo, Anton; Noble, Bram F.; Bell, Scott

    2016-03-01

    This paper presents and demonstrates a spatial framework for the application of strategic environmental assessment (SEA) in the context of change analysis for urban wetland environments. The proposed framework is focused on two key stages of the SEA process: scoping and environmental baseline assessment. These stages are arguably the most information-intense phases of SEA and have a significant effect on the quality of the SEA results. The study aims to meet the needs for proactive frameworks to assess and protect wetland habitat and services more efficiently, toward the goal of advancing more intelligent urban planning and development design. The proposed framework, adopting geographic information system and remote sensing tools and applications, supports the temporal evaluation of wetland change and sustainability assessment based on landscape indicator analysis. The framework was applied to a rapidly developing urban environment in the City of Saskatoon, Saskatchewan, Canada, analyzing wetland change and land-use pressures from 1985 to 2011. The SEA spatial scale was rescaled from administrative urban planning units to an ecologically meaningful area. Landscape change assessed was based on a suite of indicators that were subsequently rolled up into a single, multi-dimensional, and easy to understand and communicate index to examine the implications of land-use change for wetland sustainability. The results show that despite the recent extremely wet period in the Canadian prairie region, land-use change contributed to increasing threats to wetland sustainability.

  14. Strategic Environmental Assessment Framework for Landscape-Based, Temporal Analysis of Wetland Change in Urban Environments.

    Science.gov (United States)

    Sizo, Anton; Noble, Bram F; Bell, Scott

    2016-03-01

    This paper presents and demonstrates a spatial framework for the application of strategic environmental assessment (SEA) in the context of change analysis for urban wetland environments. The proposed framework is focused on two key stages of the SEA process: scoping and environmental baseline assessment. These stages are arguably the most information-intense phases of SEA and have a significant effect on the quality of the SEA results. The study aims to meet the needs for proactive frameworks to assess and protect wetland habitat and services more efficiently, toward the goal of advancing more intelligent urban planning and development design. The proposed framework, adopting geographic information system and remote sensing tools and applications, supports the temporal evaluation of wetland change and sustainability assessment based on landscape indicator analysis. The framework was applied to a rapidly developing urban environment in the City of Saskatoon, Saskatchewan, Canada, analyzing wetland change and land-use pressures from 1985 to 2011. The SEA spatial scale was rescaled from administrative urban planning units to an ecologically meaningful area. Landscape change assessed was based on a suite of indicators that were subsequently rolled up into a single, multi-dimensional, and easy to understand and communicate index to examine the implications of land-use change for wetland sustainability. The results show that despite the recent extremely wet period in the Canadian prairie region, land-use change contributed to increasing threats to wetland sustainability.

  15. Bibliometric analysis of global environmental assessment research in a 20-year period

    Energy Technology Data Exchange (ETDEWEB)

    Li, Wei, E-mail: weili@bnu.edu.cn; Zhao, Yang

    2015-01-15

    Based on the samples of 113,468 publications on environmental assessment (EA) from the past 20 years, we used a bibliometric analysis to study the literature in terms of trends of growth, subject categories and journals, international collaboration, geographic distribution of publications, and scientific research issues. By applying thresholds to network centralities, a core group of countries can be distinguished as part of the international collaboration network. A frequently used keywords analysis found that the priority in assessment would gradually change from project environmental impact assessment (EIA) to strategic environmental assessment (SEA). Decision-theoretic approaches (i.e., environmental indicator selection, life cycle assessment, etc.), along with new technologies and methods (i.e., the geographic information system and modeling) have been widely applied in the EA research field over the past 20 years. Hot spots such as “biodiversity” and “climate change” have been emphasized in current EA research, a trend that will likely continue in the future. The h-index has been used to evaluate the research quality among countries all over the world, while the improvement of developing countries' EA systems is becoming a popular research topic. Our study reveals patterns in scientific outputs and academic collaborations and serves as an alternative and innovative way of revealing global research trends in the EA research field.

  16. Using Two Different Approaches to Assess Dietary Patterns: Hypothesis-Driven and Data-Driven Analysis

    Directory of Open Access Journals (Sweden)

    Ágatha Nogueira Previdelli

    2016-09-01

    Full Text Available The use of dietary patterns to assess dietary intake has become increasingly common in nutritional epidemiology studies due to the complexity and multidimensionality of the diet. Currently, two main approaches have been widely used to assess dietary patterns: data-driven and hypothesis-driven analysis. Since the methods explore different angles of dietary intake, using both approaches simultaneously might yield complementary and useful information; thus, we aimed to use both approaches to gain knowledge of adolescents’ dietary patterns. Food intake from a cross-sectional survey with 295 adolescents was assessed by 24 h dietary recall (24HR. In hypothesis-driven analysis, based on the American National Cancer Institute method, the usual intake of Brazilian Healthy Eating Index Revised components were estimated. In the data-driven approach, the usual intake of foods/food groups was estimated by the Multiple Source Method. In the results, hypothesis-driven analysis showed low scores for Whole grains, Total vegetables, Total fruit and Whole fruits, while, in data-driven analysis, fruits and whole grains were not presented in any pattern. High intakes of sodium, fats and sugars were observed in hypothesis-driven analysis with low total scores for Sodium, Saturated fat and SoFAA (calories from solid fat, alcohol and added sugar components in agreement, while the data-driven approach showed the intake of several foods/food groups rich in these nutrients, such as butter/margarine, cookies, chocolate powder, whole milk, cheese, processed meat/cold cuts and candies. In this study, using both approaches at the same time provided consistent and complementary information with regard to assessing the overall dietary habits that will be important in order to drive public health programs, and improve their efficiency to monitor and evaluate the dietary patterns of populations.

  17. Using Two Different Approaches to Assess Dietary Patterns: Hypothesis-Driven and Data-Driven Analysis

    Science.gov (United States)

    Previdelli, Ágatha Nogueira; de Andrade, Samantha Caesar; Fisberg, Regina Mara; Marchioni, Dirce Maria

    2016-01-01

    The use of dietary patterns to assess dietary intake has become increasingly common in nutritional epidemiology studies due to the complexity and multidimensionality of the diet. Currently, two main approaches have been widely used to assess dietary patterns: data-driven and hypothesis-driven analysis. Since the methods explore different angles of dietary intake, using both approaches simultaneously might yield complementary and useful information; thus, we aimed to use both approaches to gain knowledge of adolescents’ dietary patterns. Food intake from a cross-sectional survey with 295 adolescents was assessed by 24 h dietary recall (24HR). In hypothesis-driven analysis, based on the American National Cancer Institute method, the usual intake of Brazilian Healthy Eating Index Revised components were estimated. In the data-driven approach, the usual intake of foods/food groups was estimated by the Multiple Source Method. In the results, hypothesis-driven analysis showed low scores for Whole grains, Total vegetables, Total fruit and Whole fruits), while, in data-driven analysis, fruits and whole grains were not presented in any pattern. High intakes of sodium, fats and sugars were observed in hypothesis-driven analysis with low total scores for Sodium, Saturated fat and SoFAA (calories from solid fat, alcohol and added sugar) components in agreement, while the data-driven approach showed the intake of several foods/food groups rich in these nutrients, such as butter/margarine, cookies, chocolate powder, whole milk, cheese, processed meat/cold cuts and candies. In this study, using both approaches at the same time provided consistent and complementary information with regard to assessing the overall dietary habits that will be important in order to drive public health programs, and improve their efficiency to monitor and evaluate the dietary patterns of populations. PMID:27669289

  18. RELAP5/MOD2 Overview and Developmental. Assessment Results from TMl-1 Plant Transient Analysis

    Energy Technology Data Exchange (ETDEWEB)

    Lin, J. C.; Tsai, C. C.; Ransom, V. H.; Johnsen, G. W. [EGiG Idaho, Inc. Idaho National Engineering Laboratory, P.O. Box 1625, Idaho Falls, ID 83415 (United States)

    2013-02-01

    RELAP5/MOD2 is a new version of the RELAP5 thermal-hydraulic computer code containing improved modeling features that provide a generic capability for pressurized water reactor transient simulation. The objective of this paper is to provide code users with an overview of the code and to report developmental assessment results obtained from a Three Mile Island Unit One plant transient analysis. The assessment shows that the injection of highly sub-cooled water into a high-pressure primary coolant system does not cause unphysical results or pose a problem for RELAP5/MOD2. (author)

  19. An Analysis Of Tensile Test Results to Assess the Innovation Risk for an Additive Manufacturing Technology

    Directory of Open Access Journals (Sweden)

    Adamczak Stanisław

    2015-03-01

    Full Text Available The aim of this study was to assess the innovation risk for an additive manufacturing process. The analysis was based on the results of static tensile tests obtained for specimens made of photocured resin. The assessment involved analyzing the measurement uncertainty by applying the FMEA method. The structure of the causes and effects of the discrepancies was illustrated using the Ishikawa diagram. The risk priority numbers were calculated. The uncertainty of the tensile test measurement was determined for three printing orientations. The results suggest that the material used to fabricate the tensile specimens shows clear anisotropy of the properties in relation to the printing direction.

  20. Dependence assessment in human reliability analysis based on D numbers and AHP

    Energy Technology Data Exchange (ETDEWEB)

    Zhou, Xinyi; Deng, Xinyang [School of Computer and Information Science, Southwest University, Chongqing 400715 (China); Deng, Yong, E-mail: ydeng@swu.edu.cn [School of Computer and Information Science, Southwest University, Chongqing 400715 (China); Institute of Fundamental and Frontier Sciences, University of Electronic Science and Technology of China, Chengdu, Sichuan 610054 (China); Mahadevan, Sankaran [School of Engineering, Vanderbilt University, Nashville, TN 37235 (United States)

    2017-03-15

    Highlights: • D numbers and AHP are combined to implement dependence assessment in HRA. • A new tool, called D numbers, is used to deal with the uncertainty in HRA. • The proposed method can well address the fuzziness and subjectivity in linguistic assessment. • The proposed method is well applicable in dependence assessment which inherently has a linguistic assessment process. - Abstract: Since human errors always cause heavy loss especially in nuclear engineering, human reliability analysis (HRA) has attracted more and more attention. Dependence assessment plays a vital role in HRA, measuring the dependence degree of human errors. Many researches have been done while still have improvement space. In this paper, a dependence assessment model based on D numbers and analytic hierarchy process (AHP) is proposed. Firstly, identify the factors used to measure the dependence level of two human operations. Besides, in terms of the suggested dependence level, determine and quantify the anchor points for each factor. Secondly, D numbers and AHP are adopted in model. Experts evaluate the dependence level of human operations for each factor. Then, the evaluation results are presented as D numbers and fused by D number’s combination rule that can obtain the dependence probability of human operations for each factor. The weights of factors can be determined by AHP. Thirdly, based on the dependence probability for each factor and its corresponding weight, the dependence probability of two human operations and its confidence can be obtained. The proposed method can well address the fuzziness and subjectivity in linguistic assessment. The proposed method is well applicable to assess the dependence degree of human errors in HRA which inherently has a linguistic assessment process.

  1. Software Integration of Life Cycle Assessment and Economic Analysis for Process Evaluation

    DEFF Research Database (Denmark)

    Kalakula, Sawitree; Malakula, Pomthong; Siemanonda, Kitipat

    2013-01-01

    This study is focused on the sustainable process design of bioethanol production from cassava rhizome. The study includes: process simulation, sustainability analysis, economic evaluation and life cycle assessment (LCA). A steady state process simulation if performed to generate a base case design...... of the bioethanol conversion process using cassava rhizome as a feedstock. The sustainability analysis is performed to analyze the relevant indicators in sustainability metrics, to definedesign/retrofit targets for process improvements. Economic analysis is performed to evaluate the profitability of the process........ Also, simultaneously with sustainability analysis, the life cycle impact on environment associated with bioethanol production is performed. Finally, candidate alternative designs are generated and compared with the base case design in terms of LCA, economics, waste, energy usage and enviromental impact...

  2. Assessing the Goodness of Fit of Phylogenetic Comparative Methods: A Meta-Analysis and Simulation Study

    Science.gov (United States)

    Jhwueng, Dwueng-Chwuan

    2013-01-01

    Background Phylogenetic comparative methods (PCMs) have been applied widely in analyzing data from related species but their fit to data is rarely assessed. Question Can one determine whether any particular comparative method is typically more appropriate than others by examining comparative data sets? Data I conducted a meta-analysis of 122 phylogenetic data sets found by searching all papers in JEB, Blackwell Synergy and JSTOR published in 2002–2005 for the purpose of assessing the fit of PCMs. The number of species in these data sets ranged from 9 to 117. Analysis Method I used the Akaike information criterion to compare PCMs, and then fit PCMs to bivariate data sets through REML analysis. Correlation estimates between two traits and bootstrapped confidence intervals of correlations from each model were also compared. Conclusions For phylogenies of less than one hundred taxa, the Independent Contrast method and the independent, non-phylogenetic models provide the best fit.For bivariate analysis, correlations from different PCMs are qualitatively similar so that actual correlations from real data seem to be robust to the PCM chosen for the analysis. Therefore, researchers might apply the PCM they believe best describes the evolutionary mechanisms underlying their data. PMID:23826183

  3. ASSESSMENT OF PLASTIC FLOWS AND STOCKS IN SERBIA USING MATERIAL FLOW ANALYSIS

    Directory of Open Access Journals (Sweden)

    Goran Vujić

    2010-01-01

    Full Text Available Material flow analysis (MFA was used to assess the amounts of plastic materials flows and stocks that are annually produced, consumed, imported, exported, collected, recycled, and disposed in the landfills in Serbia. The analysis revealed that approximatelly 269,000 tons of plastic materials are directly disposed in uncontrolled landfills in Serbia without any preatretment, and that siginificant amounts of these materials have already accumulated in the landfills. The substantial amounts of landfilled plastics represent not only a loss of valuable recourses, but also pose a seriuos treath to the environment and human health, and if the trend of direct plastic landfilling is continued, Serbia will face with grave consecequnces.

  4. Instrumented gait analysis for planning and assessment of treatment in cerebral palsy

    Directory of Open Access Journals (Sweden)

    Thomas Raji

    2005-01-01

    Full Text Available Background: To improve the walking efficiency of children with cerebral palsy, gait must be documented accurately so that the abnormalities can be assessed and the best treatment option can be selected. Methods: Gait was analysed using Selspot kinematic system with a Kistler force plate and the motion analysis ambulatory EMG system connected to a personal computer. Results: Walking speed and stride length had improved in children. Conclusion: Gait analysis can help to accomplish improvement in walking efficiency of children with cerebral palsy.

  5. MeQA: a pipeline for MeDIP-seq data quality assessment and analysis.

    Science.gov (United States)

    Huang, J; Renault, V; Sengenès, J; Touleimat, N; Michel, S; Lathrop, M; Tost, J

    2012-02-15

    We present a pipeline for the pre-processing, quality assessment, read distribution and methylation estimation for methylated DNA immunoprecipitation (MeDIP)-sequence datasets. This is the first MeDIP-seq-specific analytic pipeline that starts at the output of the sequencers. This pipeline will reduce the data analysis load on staff and allows the easy and straightforward analysis of sequencing data for DNA methylation. The pipeline integrates customized scripting and several existing tools, which can deal with both paired and single end data. The package and extensive documentation, and comparison to public data is available at http://life.tongji.edu.cn/meqa/.

  6. Compost maturity assessment using physicochemical, solid-state spectroscopy, and plant bioassay analysis.

    Science.gov (United States)

    Kumar, D Senthil; Kumar, P Satheesh; Rajendran, N M; Anbuganapathi, G

    2013-11-27

    The vermicompost produced from flower waste inoculated with biofertilizers was subjected to compost maturity test: (i) physicochemical method (pH, OC, TN, C:N); (ii) solid state spectroscopic analysis (FTIR and (13)C CPMAS NMR); and (iii) plant bioassay (germination index). The pH of vermicompost was decreased toward neutral, C:N ratio microbes present in the gut of earthworm. Plant bioassays are considered the most conventional assessment of compost maturity analysis, and subsequently, it shows the effect of vermicompost maturity on the germination index of Vigna mungo .

  7. Quantitative assessment of in-solution digestion efficiency identifies optimal protocols for unbiased protein analysis

    DEFF Research Database (Denmark)

    Leon, Ileana R; Schwämmle, Veit; Jensen, Ole N

    2013-01-01

    The majority of mass spectrometry-based protein quantification studies uses peptide-centric analytical methods and thus strongly relies on efficient and unbiased protein digestion protocols for sample preparation. We present a novel objective approach to assess protein digestion efficiency using...... a combination of qualitative and quantitative LC-MS/MS methods and statistical data analysis. In contrast to previous studies we employed both standard qualitative as well as data-independent quantitative workflows to systematically assess trypsin digestion efficiency and bias using mitochondrial protein...... protocols and replicates, with an average 40% protein sequence coverage and an average of 11 peptides identified per protein. Systematic quantitative and statistical analysis of physicochemical parameters demonstrated that deoxycholate-assisted in-solution digestion combined with phase transfer allows...

  8. Combining a building simulation with energy systems analysis to assess the benefits of natural ventilation

    DEFF Research Database (Denmark)

    Oropeza-Perez, Ivan; Østergaard, Poul Alberg; Remmen, Arne

    2013-01-01

    a thermal air flow simulation program - Into the energy systems analysis model. Descriptions of the energy systems in two geographical locations, i.e. Mexico and Denmark, are set up as inputs. Then, the assessment is done by calculating the energy impacts as well as environmental benefits in the energy......This article shows the combination of a thermal air flow simulation program with an energy systems analysis model in order to assess the use of natural ventilation as a method for saving energy within residential buildings in large-scale scenarios. The aim is to show the benefits for utilizing...... natural airflow instead of active systems such as mechanical ventilation or air-conditioning in buildings where the indoor temperature is over the upper limit of the comfort range. The combination is done by introducing the energy saving output - calculated with a model of natural ventilation using...

  9. SU-F-T-246: Evaluation of Healthcare Failure Mode And Effect Analysis For Risk Assessment

    Energy Technology Data Exchange (ETDEWEB)

    Harry, T [Oregon State University, Corvallis, OR (United States); University of California, San Diego, La Jolla, CA (United States); Manger, R; Cervino, L; Pawlicki, T [University of California, San Diego, La Jolla, CA (United States)

    2016-06-15

    Purpose: To evaluate the differences between the Veteran Affairs Healthcare Failure Modes and Effect Analysis (HFMEA) and the AAPM Task Group 100 Failure and Effect Analysis (FMEA) risk assessment techniques in the setting of a stereotactic radiosurgery (SRS) procedure were compared respectively. Understanding the differences in the techniques methodologies and outcomes will provide further insight into the applicability and utility of risk assessments exercises in radiation therapy. Methods: HFMEA risk assessment analysis was performed on a stereotactic radiosurgery procedure. A previous study from our institution completed a FMEA of our SRS procedure and the process map generated from this work was used for the HFMEA. The process of performing the HFMEA scoring was analyzed, and the results from both analyses were compared. Results: The key differences between the two risk assessments are the scoring criteria for failure modes and identifying critical failure modes for potential hazards. The general consensus among the team performing the analyses was that scoring for the HFMEA was simpler and more intuitive then the FMEA. The FMEA identified 25 critical failure modes while the HFMEA identified 39. Seven of the FMEA critical failure modes were not identified by the HFMEA and 21 of the HFMEA critical failure modes were not identified by the FMEA. HFMEA as described by the Veteran Affairs provides guidelines on which failure modes to address first. Conclusion: HFMEA is a more efficient model for identifying gross risks in a process than FMEA. Clinics with minimal staff, time and resources can benefit from this type of risk assessment to eliminate or mitigate high risk hazards with nominal effort. FMEA can provide more in depth details but at the cost of elevated effort.

  10. A framework for techno-economic & environmental sustainability analysis by risk assessment for conceptual process evaluation

    DEFF Research Database (Denmark)

    Loureiro da Costa Lira Gargalo, Carina; Sin, Gürkan; Carvalho, Ana

    2016-01-01

    for techno-economic and environmental sustainability analysis through risk assessment is proposed for the early-stage design and screening of conceptual process alternatives. The alternatives within the design space are analyzed following the framework’s work-flow, which targets the following: (i) quantify...... the economic risk; (ii) perform the monetary valuation of environmental impact categories under uncertainty; (iii) quantify the potential environmental risk; (iv) measure the alternatives’ eco-efficiency identifying possible trade-offs; and, lastly (v) propose a joint risk assessment matrix...... for the quantitative and qualitative assessment of sustainability at the decision-support level. Through the application of appropriate methods in a hierarchical manner, this tool leads to the identification of the potentially best and more sustainable solutions. Furthermore, the application of the framework...

  11. Analysis of the possibilities of using EEG in assessing pilots’ psychophysical condition

    Directory of Open Access Journals (Sweden)

    Marta GALANT

    2017-06-01

    Full Text Available An excessive load on an operator’s cognitive system can cause deterioration in perceptual abilities, decreased reaction time and increased probability of making an incorrect decision, which in turn can lead to a dangerous situation. Researching the cognitive load of an operator can therefore contribute to safer transportation. While there are many methods used in the study of cognitive load, they can be classified as either subjective assessments or objective assessments. This paper presents an analysis of the possibilities of using electroencephalography in assessing the psychophysical condition of the pilot. The investigation was conducted in the Simulation Research Laboratory in the Institute of Combustion Engines and Transport at Poznan University of Technology.

  12. Full-scale performance assessment of aircraft secondary sandwich structure using thermoelastic stress analysis

    OpenAIRE

    Crump, D.A.; Dulieu-Barton, J.M.; Savage, J

    2009-01-01

    The use of resin film infusion (RFI) has been proven to reduce the cost of production of aircraft secondary sandwich structure. In this paper thermoelastic stress analysis (TSA) is used to assess the performance of full scale aircraft sandwich structure panels produced using both the conventional autoclave process and RFI. Finite element (FE) models of both panel types are developed and TSA is used to validate the models.

  13. Assessing a Sport/Cultural Events Network: An Application of Social Network Analysis

    OpenAIRE

    Ziakas, V; Costa, CA

    2009-01-01

    The purpose of this study was to assess the complexity of a sport/cultural events network. To that intent, a social network analysis was conducted in a small community in the US. The study had three main objectives: (1) Examine relationships among organisations involved in planning and implementing sport and cultural events based on their communication, exchange of resources, and assistance; (2) Identify the most important actors within the events network and their relationships; (3) Investig...

  14. Conference Analysis Report of Assessments on Defect and Damage for a High Temperature Structure

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Hyeong Yeon

    2008-11-15

    This report presents the analysis on the state-of-the-art research trends on creep-fatigue damage, defect assessment of high temperature structure, development of heat resistant materials and their behavior at high temperature based on the papers presented in the two international conferences of ASME PVP 2008 which was held in Chicago in July 2008 and CF-5(5th International Conference on Creep, Fatigue and Creep-Fatigue) which was held in Kalpakkam, India in September 2008.

  15. Automatic mechanical fault assessment of small wind energy systems in microgrids using electric signature analysis

    DEFF Research Database (Denmark)

    Skrimpas, Georgios Alexandros; Marhadi, Kun Saptohartyadi; Jensen, Bogi Bech

    2013-01-01

    of islanded operation. In this paper, the fault assessment is achieved efficiently and consistently via electric signature analysis (ESA). In ESA the fault related frequency components are manifested as sidebands of the existing current and voltage time harmonics. The energy content between the fundamental, 5...... element model where dynamic eccentricity and bearing outer race defect are simulated under varying fault severity and electric loading conditions....

  16. USE OF MULTIPARAMETER ANALYSIS OF LABORATORY BIOMARKERS TO ASSESS RHEUMATOID ARTHRITIS ACTIVITY

    Directory of Open Access Journals (Sweden)

    A. A. Novikov

    2015-01-01

    Full Text Available The key component in the management of patients with rheumatoid arthritis (RA is regular control of RA activity. The quantitative assessment of a patient’s status allows the development of standardized indications for anti-rheumatic therapy.Objective: to identify the laboratory biomarkers able to reflect RA activity.Subjects and methods. Fifty-eight patients with RA and 30 age- and sex-matched healthy donors were examined. The patients were divided into high/moderate and mild disease activity groups according to DAS28. The serum concentrations of 30 biomarkers were measured using immunonephelometric assay, enzyme immunoassay, and xMAP technology.Results and discussion. Multivariate analysis could identify the factors mostly related to high/moderate RA activity according to DAS28, such as fibroblast growth factor-2, monocyte chemoattractant protein-1, interleukins (IL 1α, 6, and 15, and tumor necrosis factor-α and could create a prognostic model for RA activity assessment. ROC analysis has shown that this model has excellent diagnostic efficiency in differentiating high/moderate versus low RA activity.Conclusion. To create a subjective assessment-independent immunological multiparameter index of greater diagnostic accuracy than the laboratory parameters routinely used in clinical practice may be a qualitatively new step in assessing and monitoring RA activity.

  17. Analysis of the methods for assessing socio-economic development level of urban areas

    Science.gov (United States)

    Popova, Olga; Bogacheva, Elena

    2017-01-01

    The present paper provides a targeted analysis of current approaches (ratings) in the assessment of socio-economic development of urban areas. The survey focuses on identifying standardized methodologies to area assessment techniques formation that will result in developing the system of intelligent monitoring, dispatching, building management, scheduling and effective management of an administrative-territorial unit. This system is characterized by complex hierarchical structure, including tangible and intangible properties (parameters, attributes). Investigating the abovementioned methods should increase the administrative-territorial unit's attractiveness for investors and residence. The research aims at studying methods for evaluating socio-economic development level of the Russian Federation territories. Experimental and theoretical territory estimating methods were revealed. Complex analysis of the characteristics of the areas was carried out and evaluation parameters were determined. Integral indicators (resulting rating criteria values) as well as the overall rankings (parameters, characteristics) were analyzed. The inventory of the most widely used partial indicators (parameters, characteristics) of urban areas was revealed. The resulting criteria of rating values homogeneity were verified and confirmed by determining the root mean square deviation, i.e. divergence of indices. The principal shortcomings of assessment methodologies were revealed. The assessment methods with enhanced effectiveness and homogeneity were proposed.

  18. Life Cycle Assessment and Life Cycle Cost Analysis of Magnesia Spinel Brick Production

    Directory of Open Access Journals (Sweden)

    Aysun Özkan

    2016-07-01

    Full Text Available Sustainable use of natural resources in the production of construction materials has become a necessity both in Europe and Turkey. Construction products in Europe should have European Conformity (CE and Environmental Product Declaration (EPD, an independently verified and registered document in line with the European standard EN 15804. An EPD certificate can be created by performing a Life Cycle Assessment (LCA study. In this particular work, an LCA study was carried out for a refractory brick production for environmental assessment. In addition to the LCA, the Life Cycle Cost (LCC analysis was also applied for economic assessment. Firstly, a cradle-to-gate LCA was performed for one ton of magnesia spinel refractory brick. The CML IA method included in the licensed SimaPro 8.0.1 software was chosen to calculate impact categories (namely, abiotic depletion, global warming potential, acidification potential, eutrophication potential, human toxicity, ecotoxicity, ozone depletion potential, and photochemical oxidation potential. The LCC analysis was performed by developing a cost model for internal and external cost categories within the software. The results were supported by a sensitivity analysis. According to the results, the production of raw materials and the firing process in the magnesia spinel brick production were found to have several negative effects on the environment and were costly.

  19. Sex Assessment Using the Femur and Tibia in Medieval Skeletal Remains from Ireland: Discriminant Function Analysis.

    Science.gov (United States)

    Novak, Mario

    2016-04-01

    Sex determination based on discriminant function analysis of skeletal measurements is probably the most effective method for assessment of sex in archaeological and contemporary populations due to various reasons, but it also suffers from limitations such as population specificity. In this paper standards for sex assessment from the femur and tibia in the medieval Irish population are presented. Six femoral and six tibial measurements obtained from 56 male and 45 female skeletons were subjected to discriminant function analysis. Average accuracies obtained by this study range between 87.1 and 97%. The highest level of accuracy (97%) was achieved when using combined variables of the femur and tibia (maximum diameter of femoral head and circumference at tibial nutrient foramen), as well as two variables of the tibia (proximal epiphyseal breadth and circumference at nutrient foramen). Discriminant functions using a single variable provided accuracies between 87.1 and 96% with the circumference at the level of the tibial nutrient foramen providing the best separation. High accuracy rates obtained by this research correspond to the data recorded in other studies thus confirming the importance of discriminant function analysis in assessment of sex in both archaeological and forensic contexts.

  20. Assessing the influence of Environmental Impact Assessments on science and policy: an analysis of the Three Gorges Project.

    Science.gov (United States)

    Tullos, Desiree

    2009-07-01

    The need to understand and minimize negative environmental outcomes associated with large dams has both contributed to and benefited from the introduction and subsequent improvements in the Environmental Impact Assessment (EIA) process. However, several limitations in the EIA process remain, including those associated with the uncertainty and significance of impact projections. These limitations are directly related to the feedback between science and policy, with information gaps in scientific understanding discovered through the EIA process contributing valuable recommendations on critical focus areas for prioritizing and funding research within the fields of ecological conservation and river engineering. This paper presents an analysis of the EIA process for the Three Gorges Project (TGP) in China as a case study for evaluating this feedback between the EIA and science and policy. For one of the best-studied public development projects in the world, this paper presents an investigation into whether patterns exist between the scientific interest (via number of publications) in environmental impacts and (a) the identification of impacts as uncertain or priority by the EIA, (b) decisions or political events associated with the dam, and (c) impact type. This analysis includes the compilation of literature on TGP, characterization of ecosystem interactions and responses to TGP through a hierarchy of impacts, coding of EIA impacts as "uncertain" impacts that require additional study and "priority" impacts that have particularly high significance, mapping of an event chronology to relate policies, institutional changes, and decisions about TGP as "events" that could influence the focus and intensity of scientific investigation, and analysis of the number of publications by impact type and order within the impact hierarchy. From these analyses, it appears that the availability and consistency of scientific information limit the accuracy of environmental impact

  1. Rapid quality assessment of Radix Aconiti Preparata using direct analysis in real time mass spectrometry

    Energy Technology Data Exchange (ETDEWEB)

    Zhu Hongbin; Wang Chunyan; Qi Yao [Changchun Center of Mass Spectrometry and Chemical Biology Laboratory, Changchun Institute of Applied Chemistry, Chinese Academy of Sciences, Changchun 130022 (China); University of Chinese Academy of Sciences, Beijing 100039 (China); Song Fengrui, E-mail: songfr@ciac.jl.cn [Changchun Center of Mass Spectrometry and Chemical Biology Laboratory, Changchun Institute of Applied Chemistry, Chinese Academy of Sciences, Changchun 130022 (China); Liu Zhiqiang; Liu Shuying [Changchun Center of Mass Spectrometry and Chemical Biology Laboratory, Changchun Institute of Applied Chemistry, Chinese Academy of Sciences, Changchun 130022 (China)

    2012-11-08

    Highlights: Black-Right-Pointing-Pointer DART MS combined with PCA and HCA was used to rapidly identify markers of Radix Aconiti. Black-Right-Pointing-Pointer The DART MS behavior of six aconitine-type alkaloids was investigated. Black-Right-Pointing-Pointer Chemical markers were recognized between the qualified and unqualified samples. Black-Right-Pointing-Pointer DART MS was shown to be an effective tool for quality control of Radix Aconiti Preparata. - Abstract: This study presents a novel and rapid method to identify chemical markers for the quality control of Radix Aconiti Preparata, a world widely used traditional herbal medicine. In the method, the samples with a fast extraction procedure were analyzed using direct analysis in real time mass spectrometry (DART MS) combined with multivariate data analysis. At present, the quality assessment approach of Radix Aconiti Preparata was based on the two processing methods recorded in Chinese Pharmacopoeia for the purpose of reducing the toxicity of Radix Aconiti and ensuring its clinical therapeutic efficacy. In order to ensure the safety and effectivity in clinical use, the processing degree of Radix Aconiti should be well controlled and assessed. In the paper, hierarchical cluster analysis and principal component analysis were performed to evaluate the DART MS data of Radix Aconiti Preparata samples in different processing times. The results showed that the well processed Radix Aconiti Preparata, unqualified processed and the raw Radix Aconiti could be clustered reasonably corresponding to their constituents. The loading plot shows that the main chemical markers having the most influence on the discrimination amongst the qualified and unqualified samples were mainly some monoester diterpenoid aconitines and diester diterpenoid aconitines, i.e. benzoylmesaconine, hypaconitine, mesaconitine, neoline, benzoylhypaconine, benzoylaconine, fuziline, aconitine and 10-OH-mesaconitine. The established DART MS approach in

  2. A Comparative Analysis on Assessment of Land Carrying Capacity with Ecological Footprint Analysis and Index System Method.

    Science.gov (United States)

    Qian, Yao; Tang, Lina; Qiu, Quanyi; Xu, Tong; Liao, Jiangfu

    2015-01-01

    Land carrying capacity (LCC) explains whether the local land resources are effectively used to support economic activities and/or human population. LCC can be evaluated commonly with two approaches, namely ecological footprint analysis (EFA) and the index system method (ISM). EFA is helpful to investigate the effects of different land categories whereas ISM can be used to evaluate the contributions of social, environmental, and economic factors. Here we compared the two LCC-evaluation approaches with data collected from Xiamen City, a typical region where rapid economic growth and urbanization are found in China. The results show that LCC assessments with EFA and ISM not only complement each other but also are mutually supportive. Both assessments suggest that decreases in arable land and increasingly high energy consumption have major negative effects on LCC and threaten sustainable development for Xiamen City. It is important for the local policy makers, planners and designers to reduce ecological deficits by controlling fossil energy consumption, protecting arable land and forest land from converting into other land types, and slowing down the speed of urbanization, and to promote sustainability by controlling rural-to-urban immigration, increasing hazard-free treatment rate of household garbage, and raising energy consumption per unit industrial added value. Although EFA seems more appropriate for estimating LCC for a resource-output or self-sufficient region and ISM is more suitable for a resource-input region, both approaches should be employed when perform LCC assessment in any places around the world.

  3. A Comparative Analysis on Assessment of Land Carrying Capacity with Ecological Footprint Analysis and Index System Method.

    Directory of Open Access Journals (Sweden)

    Yao Qian

    Full Text Available Land carrying capacity (LCC explains whether the local land resources are effectively used to support economic activities and/or human population. LCC can be evaluated commonly with two approaches, namely ecological footprint analysis (EFA and the index system method (ISM. EFA is helpful to investigate the effects of different land categories whereas ISM can be used to evaluate the contributions of social, environmental, and economic factors. Here we compared the two LCC-evaluation approaches with data collected from Xiamen City, a typical region where rapid economic growth and urbanization are found in China. The results show that LCC assessments with EFA and ISM not only complement each other but also are mutually supportive. Both assessments suggest that decreases in arable land and increasingly high energy consumption have major negative effects on LCC and threaten sustainable development for Xiamen City. It is important for the local policy makers, planners and designers to reduce ecological deficits by controlling fossil energy consumption, protecting arable land and forest land from converting into other land types, and slowing down the speed of urbanization, and to promote sustainability by controlling rural-to-urban immigration, increasing hazard-free treatment rate of household garbage, and raising energy consumption per unit industrial added value. Although EFA seems more appropriate for estimating LCC for a resource-output or self-sufficient region and ISM is more suitable for a resource-input region, both approaches should be employed when perform LCC assessment in any places around the world.

  4. Assessment of blood glucose predictors: the prediction-error grid analysis.

    Science.gov (United States)

    Sivananthan, Sampath; Naumova, Valeriya; Man, Chiara Dalla; Facchinetti, Andrea; Renard, Eric; Cobelli, Claudio; Pereverzyev, Sergei V

    2011-08-01

    Prediction of the future blood glucose (BG) evolution from continuous glucose monitoring (CGM) data is a promising direction in diabetes therapy management, and several glucose predictors have recently been proposed. This raises the problem of their assessment. There were attempts to use for such assessment the continuous glucose-error grid analysis (CG-EGA), originally developed for CGM devices. However, in the CG-EGA the BG rate of change is estimated from past BG readings, whereas predictors provide BG estimation ahead of time. Therefore, the original CG-EGA should be modified to assess predictors. Here we propose a new version of the CG-EGA, the Prediction-Error Grid Analysis (PRED-EGA). The analysis is based both on simulated data and on data from clinical trials, performed in the European FP7-project "DIAdvisor." Simulated data are used to test the ability of the analyzed CG-EGA modifications to capture erroneous predictions in controlled situation. Real data are used to show the impact of the different CG-EGA versions in the evaluation of a predictor. Using the data of 10 virtual and 10 real subjects and analyzing two different predictors, we demonstrate that the straightforward application of the CG-EGA does not adequately classify the prediction performance. For example, we observed that up to 70% of 20 min ahead predictions in the hyperglycemia region that are classified by this application as erroneous are, in fact, accurate. Moreover, for predictions during hypoglycemia the assessments produced by the straightforward application of the CG-EGA are not only too pessimistic (in up to 60% of cases), but this version is not able to detect real erroneous predictions. In contrast, the proposed modification of the CG-EGA, where the rate of change is estimated on the predicted BG profile, is an adequate metric for the assessment of predictions. We propose a new CG-EGA, the PRED-EGA, for the assessment of glucose predictors. The presented analysis shows that

  5. An Analysis of the Most Adopted Rating Systems for Assessing the Environmental Impact of Buildings

    Directory of Open Access Journals (Sweden)

    Elena Bernardi

    2017-07-01

    Full Text Available Rating systems for assessing the environmental impact of buildings are technical instruments that aim to evaluate the environmental impact of buildings and construction projects. In some cases, these rating systems can also cover urban-scale projects, community projects, and infrastructures. These schemes are designed to assist project management in making the projects more sustainable by providing frameworks with precise criteria for assessing the various aspects of a building’s environmental impact. Given the growing interest in sustainable development worldwide, many rating systems for assessing the environmental impact of buildings have been established in recent years, each one with its peculiarities and fields of applicability. The present work is motivated by an interest in emphasizing such differences to better understand these rating systems and extract the main implications to building design. It also attempts to summarize in a user-friendly form the vast and fragmented assortment of information that is available today. The analysis focuses on the six main rating systems: the Building Research Establishment Environmental Assessment Methodology (BREEAM, the Comprehensive Assessment System for Built Environment Efficiency (CASBEE, the Deutsche Gesellschaft für Nachhaltiges Bauen (DGNB, the Haute Qualité Environnementale (HQETM, the Leadership in Energy and Environmental Design (LEED, and the Sustainable Building Tool (SBTool.

  6. Assessment of acute myocarditis by cardiac magnetic resonance imaging: Comparison of qualitative and quantitative analysis methods.

    Science.gov (United States)

    Imbriaco, Massimo; Nappi, Carmela; Puglia, Marta; De Giorgi, Marco; Dell'Aversana, Serena; Cuocolo, Renato; Ponsiglione, Andrea; De Giorgi, Igino; Polito, Maria Vincenza; Klain, Michele; Piscione, Federico; Pace, Leonardo; Cuocolo, Alberto

    2017-10-26

    To compare cardiac magnetic resonance (CMR) qualitative and quantitative analysis methods for the noninvasive assessment of myocardial inflammation in patients with suspected acute myocarditis (AM). A total of 61 patients with suspected AM underwent coronary angiography and CMR. Qualitative analysis was performed applying Lake-Louise Criteria (LLC), followed by quantitative analysis based on the evaluation of edema ratio (ER) and global relative enhancement (RE). Diagnostic performance was assessed for each method by measuring the area under the curves (AUC) of the receiver operating characteristic analyses. The final diagnosis of AM was based on symptoms and signs suggestive of cardiac disease, evidence of myocardial injury as defined by electrocardiogram changes, elevated troponin I, exclusion of coronary artery disease by coronary angiography, and clinical and echocardiographic follow-up at 3 months after admission to the chest pain unit. In all patients, coronary angiography did not show significant coronary artery stenosis. Troponin I levels and creatine kinase were higher in patients with AM compared to those without (both P quantitative (ER 0.89 and global RE 0.80) analyses were also similar. Qualitative and quantitative CMR analysis methods show similar diagnostic accuracy for the diagnosis of AM. These findings suggest that a simplified approach using a shortened CMR protocol including only T2-weighted STIR sequences might be useful to rule out AM in patients with acute coronary syndrome and normal coronary angiography.

  7. Assessing the Goodness of Fit of Phylogenetic Comparative Methods: A Meta-Analysis and Simulation Study.

    Science.gov (United States)

    Jhwueng, Dwueng-Chwuan

    2013-01-01

    Phylogenetic comparative methods (PCMs) have been applied widely in analyzing data from related species but their fit to data is rarely assessed. Can one determine whether any particular comparative method is typically more appropriate than others by examining comparative data sets? I conducted a meta-analysis of 122 phylogenetic data sets found by searching all papers in JEB, Blackwell Synergy and JSTOR published in 2002-2005 for the purpose of assessing the fit of PCMs. The number of species in these data sets ranged from 9 to 117. I used the Akaike information criterion to compare PCMs, and then fit PCMs to bivariate data sets through REML analysis. Correlation estimates between two traits and bootstrapped confidence intervals of correlations from each model were also compared. For phylogenies of less than one hundred taxa, the Independent Contrast method and the independent, non-phylogenetic models provide the best fit.For bivariate analysis, correlations from different PCMs are qualitatively similar so that actual correlations from real data seem to be robust to the PCM chosen for the analysis. Therefore, researchers might apply the PCM they believe best describes the evolutionary mechanisms underlying their data.

  8. Assessing the Goodness of Fit of Phylogenetic Comparative Methods: A Meta-Analysis and Simulation Study.

    Directory of Open Access Journals (Sweden)

    Dwueng-Chwuan Jhwueng

    Full Text Available Phylogenetic comparative methods (PCMs have been applied widely in analyzing data from related species but their fit to data is rarely assessed.Can one determine whether any particular comparative method is typically more appropriate than others by examining comparative data sets?I conducted a meta-analysis of 122 phylogenetic data sets found by searching all papers in JEB, Blackwell Synergy and JSTOR published in 2002-2005 for the purpose of assessing the fit of PCMs. The number of species in these data sets ranged from 9 to 117.I used the Akaike information criterion to compare PCMs, and then fit PCMs to bivariate data sets through REML analysis. Correlation estimates between two traits and bootstrapped confidence intervals of correlations from each model were also compared.For phylogenies of less than one hundred taxa, the Independent Contrast method and the independent, non-phylogenetic models provide the best fit.For bivariate analysis, correlations from different PCMs are qualitatively similar so that actual correlations from real data seem to be robust to the PCM chosen for the analysis. Therefore, researchers might apply the PCM they believe best describes the evolutionary mechanisms underlying their data.

  9. Assessment of ecological risks at former landfill site using TRIAD procedure and multicriteria analysis.

    Science.gov (United States)

    Sorvari, Jaana; Schultz, Eija; Haimi, Jari

    2013-02-01

    Old industrial landfills are important sources of environmental contamination in Europe, including Finland. In this study, we demonstrated the combination of TRIAD procedure, multicriteria decision analysis (MCDA), and statistical Monte Carlo analysis for assessing the risks to terrestrial biota in a former landfill site contaminated by petroleum hydrocarbons (PHCs) and metals. First, we generated hazard quotients by dividing the concentrations of metals and PHCs in soil by the corresponding risk-based ecological benchmarks. Then we conducted ecotoxicity tests using five plant species, earthworms, and potworms, and determined the abundance and diversity of soil invertebrates from additional samples. We aggregated the results in accordance to the methods used in the TRIAD procedure, conducted rating of the assessment methods based on their performance in terms of specific criteria, and weighted the criteria using two alternative weighting techniques to produce performance scores for each method. We faced problems in using the TRIAD procedure, for example, the results from the animal counts had to be excluded from the calculation of integrated risk estimates (IREs) because our reference soil sample showed the lowest biodiversity and abundance of soil animals. In addition, hormesis hampered the use of the results from the ecotoxicity tests. The final probabilistic IREs imply significant risks at all sampling locations. Although linking MCDA with TRIAD provided a useful means to study and consider the performance of the alternative methods in predicting ecological risks, some uncertainties involved still remained outside the quantitative analysis. © 2012 Society for Risk Analysis.

  10. Quantitative assessment of wound healing using high-frequency ultrasound image analysis.

    Science.gov (United States)

    Mohafez, H; Ahmad, S A; Hadizadeh, M; Moghimi, S; Roohi, S A; Marhaban, M H; Saripan, M I; Rampal, S

    2017-05-29

    We aimed to develop a method for quantitative assessment of wound healing in ulcerated diabetic feet. High-frequency ultrasound (HFU) images of 30 wounds were acquired in a controlled environment on post-debridement days 7, 14, 21, and 28. Meaningful features portraying changes in structure and intensity of echoes during healing were extracted from the images, their relevance and discriminatory power being verified by analysis of variance. Relative analysis of tissue healing was conducted by developing a features-based healing function, optimised using the pattern-search method. Its performance was investigated through leave-one-out cross-validation technique and reconfirmed using principal component analysis. The constructed healing function could depict tissue changes during healing with 87.8% accuracy. The first principal component derived from the extracted features demonstrated similar pattern to the constructed healing function, accounting for 86.3% of the data variance. The developed wound analysis technique could be a viable tool in quantitative assessment of diabetic foot ulcers during healing. © 2017 John Wiley & Sons A/S. Published by John Wiley & Sons Ltd.

  11. Using multi-criteria decision analysis to assess the vulnerability of drinking water utilities.

    Science.gov (United States)

    Joerin, Florent; Cool, Geneviève; Rodriguez, Manuel J; Gignac, Marc; Bouchard, Christian

    2010-07-01

    Outbreaks of microbiological waterborne disease have increased governmental concern regarding the importance of drinking water safety. Considering the multi-barrier approach to safe drinking water may improve management decisions to reduce contamination risks. However, the application of this approach must consider numerous and diverse kinds of information simultaneously. This makes it difficult for authorities to apply the approach to decision making. For this reason, multi-criteria decision analysis can be helpful in applying the multi-barrier approach to vulnerability assessment. The goal of this study is to propose an approach based on a multi-criteria analysis method in order to rank drinking water systems (DWUs) based on their vulnerability to microbiological contamination. This approach is illustrated with an application carried out on 28 DWUs supplied by groundwater in the Province of Québec, Canada. The multi-criteria analysis method chosen is measuring attractiveness by a categorical based evaluation technique methodology allowing the assessment of a microbiological vulnerability indicator (MVI) for each DWU. Results are presented on a scale ranking DWUs from less vulnerable to most vulnerable to contamination. MVI results are tested using a sensitivity analysis on barrier weights and they are also compared with historical data on contamination at the utilities. The investigation demonstrates that MVI provides a good representation of the vulnerability of DWUs to microbiological contamination.

  12. Lacunarity analysis: a promising method for the automated assessment of melanocytic naevi and melanoma.

    Directory of Open Access Journals (Sweden)

    Stephen Gilmore

    Full Text Available The early diagnosis of melanoma is critical to achieving reduced mortality and increased survival. Although clinical examination is currently the method of choice for melanocytic lesion assessment, there is a growing interest among clinicians regarding the potential diagnostic utility of computerised image analysis. Recognising that there exist significant shortcomings in currently available algorithms, we are motivated to investigate the utility of lacunarity, a simple statistical measure previously used in geology and other fields for the analysis of fractal and multi-scaled images, in the automated assessment of melanocytic naevi and melanoma. Digitised dermoscopic images of 111 benign melanocytic naevi, 99 dysplastic naevi and 102 melanomas were obtained over the period 2003 to 2008, and subject to lacunarity analysis. We found the lacunarity algorithm could accurately distinguish melanoma from benign melanocytic naevi or non-melanoma without introducing many of the limitations associated with other previously reported diagnostic algorithms. Lacunarity analysis suggests an ordering of irregularity in melanocytic lesions, and we suggest the clinical application of this ordering may have utility in the naked-eye dermoscopic diagnosis of early melanoma.

  13. Evaluation of static analysis tools used to assess software important to nuclear power plant safety

    Directory of Open Access Journals (Sweden)

    Alain Ourghanlian

    2015-03-01

    Full Text Available We describe a comparative analysis of different tools used to assess safety-critical software used in nuclear power plants. To enhance the credibility of safety assessments and to optimize safety justification costs, Electricité de France (EDF investigates the use of methods and tools for source code semantic analysis, to obtain indisputable evidence and help assessors focus on the most critical issues. EDF has been using the PolySpace tool for more than 10 years. Currently, new industrial tools based on the same formal approach, Abstract Interpretation, are available. Practical experimentation with these new tools shows that the precision obtained on one of our shutdown systems software packages is substantially improved. In the first part of this article, we present the analysis principles of the tools used in our experimentation. In the second part, we present the main characteristics of protection-system software, and why these characteristics are well adapted for the new analysis tools. In the last part, we present an overview of the results and the limitations of the tools.

  14. Seeking Missing Pieces in Science Concept Assessments: Reevaluating the Brief Electricity and Magnetism Assessment through Rasch Analysis

    Science.gov (United States)

    Ding, Lin

    2014-01-01

    Discipline-based science concept assessments are powerful tools to measure learners' disciplinary core ideas. Among many such assessments, the Brief Electricity and Magnetism Assessment (BEMA) has been broadly used to gauge student conceptions of key electricity and magnetism (E&M) topics in college-level introductory physics courses.…

  15. Critical analysis of e-health readiness assessment frameworks: suitability for application in developing countries.

    Science.gov (United States)

    Mauco, Kabelo Leonard; Scott, Richard E; Mars, Maurice

    2018-02-01

    Introduction e-Health is an innovative way to make health services more effective and efficient and application is increasing worldwide. e-Health represents a substantial ICT investment and its failure usually results in substantial losses in time, money (including opportunity costs) and effort. Therefore it is important to assess e-health readiness prior to implementation. Several frameworks have been published on e-health readiness assessment, under various circumstances and geographical regions of the world. However, their utility for the developing world is unknown. Methods A literature review and analysis of published e-health readiness assessment frameworks or models was performed to determine if any are appropriate for broad assessment of e-health readiness in the developing world. A total of 13 papers described e-health readiness in different settings. Results and Discussion Eight types of e-health readiness were identified and no paper directly addressed all of these. The frameworks were based upon varying assumptions and perspectives. There was no underlying unifying theory underpinning the frameworks. Few assessed government and societal readiness, and none cultural readiness; all are important in the developing world. While the shortcomings of existing frameworks have been highlighted, most contain aspects that are relevant and can be drawn on when developing a framework and assessment tools for the developing world. What emerged is the need to develop different assessment tools for the various stakeholder sectors. This is an area that needs further research before attempting to develop a more generic framework for the developing world.

  16. Investigation of modern methods of probalistic sensitivity analysis of final repository performance assessment models (MOSEL)

    Energy Technology Data Exchange (ETDEWEB)

    Spiessl, Sabine; Becker, Dirk-Alexander

    2017-06-15

    Sensitivity analysis is a mathematical means for analysing the sensitivities of a computational model to variations of its input parameters. Thus, it is a tool for managing parameter uncertainties. It is often performed probabilistically as global sensitivity analysis, running the model a large number of times with different parameter value combinations. Going along with the increase of computer capabilities, global sensitivity analysis has been a field of mathematical research for some decades. In the field of final repository modelling, probabilistic analysis is regarded a key element of a modern safety case. An appropriate uncertainty and sensitivity analysis can help identify parameters that need further dedicated research to reduce the overall uncertainty, generally leads to better system understanding and can thus contribute to building confidence in the models. The purpose of the project described here was to systematically investigate different numerical and graphical techniques of sensitivity analysis with typical repository models, which produce a distinctly right-skewed and tailed output distribution and can exhibit a highly nonlinear, non-monotonic or even non-continuous behaviour. For the investigations presented here, three test models were defined that describe generic, but typical repository systems. A number of numerical and graphical sensitivity analysis methods were selected for investigation and, in part, modified or adapted. Different sampling methods were applied to produce various parameter samples of different sizes and many individual runs with the test models were performed. The results were evaluated with the different methods of sensitivity analysis. On this basis the methods were compared and assessed. This report gives an overview of the background and the applied methods. The results obtained for three typical test models are presented and explained; conclusions in view of practical applications are drawn. At the end, a recommendation

  17. Comparative SWOT analysis of strategic environmental assessment systems in the Middle East and North Africa region.

    Science.gov (United States)

    Rachid, G; El Fadel, M

    2013-08-15

    This paper presents a SWOT analysis of SEA systems in the Middle East North Africa region through a comparative examination of the status, application and structure of existing systems based on country-specific legal, institutional and procedural frameworks. The analysis is coupled with the multi-attribute decision making method (MADM) within an analytical framework that involves both performance analysis based on predefined evaluation criteria and countries' self-assessment of their SEA system through open-ended surveys. The results show heterogenous status with a general delayed progress characterized by varied levels of weaknesses embedded in the legal and administrative frameworks and poor integration with the decision making process. Capitalizing on available opportunities, the paper highlights measures to enhance the development and enactment of SEA in the region. Copyright © 2013 Elsevier Ltd. All rights reserved.

  18. Building consensus in environmental impact assessment through multicriteria modeling and sensitivity analysis.

    Science.gov (United States)

    Bojórquez-Tapia, Luis A; Sánchez-Colon, Salvadur; Florez, Arturo

    2005-09-01

    Multicriteria decision analysis (MCDA) increasingly is being applied in environmental impact assessment (EIA). In this article, two MCDA techniques, stochastic analytic hierarchy process and compromise programming, are combined to ascertain the environmental impacts of and to rank two alternative sites for Mexico City's new airport. Extensive sensitivity analyses were performed to determine the probability of changes in rank ordering given uncertainty in the hierarchy structure, decision criteria weights, and decision criteria performances. Results demonstrate that sensitivity analysis is fundamental for attaining consensus among members of interdisciplinary teams and for settling debates in controversial projects. It was concluded that sensitivity analysis is critical for achieving a transparent and technically defensible MCDA implementation in controversial EIA.

  19. An assessment of turbulence models for linear hydrodynamic stability analysis of strongly swirling jets

    CERN Document Server

    Rukes, Lothar; Oberleithner, Kilian

    2016-01-01

    Linear stability analysis has proven to be a useful tool in the analysis of dominant coherent structures, such as the von K\\'{a}rm\\'{a}n vortex street and the global spiral mode associated with the vortex breakdown of swirling jets. In recent years, linear stability analysis has been applied successfully to turbulent time-mean flows, instead of laminar base-flows, \\textcolor{black}{which requires turbulent models that account for the interaction of the turbulent field with the coherent structures. To retain the stability equations of laminar flows, the Boussinesq approximation with a spatially nonuniform but isotropic eddy viscosity is typically employed. In this work we assess the applicability of this concept to turbulent strongly swirling jets, a class of flows that is particularly unsuited for isotropic eddy viscosity models. Indeed we find that unsteady RANS simulations only match with experiments with a Reynolds stress model that accounts for an anisotropic eddy viscosity. However, linear stability anal...

  20. The accidental risk assessment methodology for industries (ARAMIS)/layer of protection analysis (LOPA) methodology: a step forward towards convergent practices in risk assessment?

    Science.gov (United States)

    Gowland, Richard

    2006-03-31

    In the last ten years, layer of protection analysis (LOPA) emerged as a simplified form of quantitative risk assessment (QRA). The European Commission funded project Accidental Risk Assessment Methodology for Industries in the context of the Seveso 2 Directive (ARAMIS) has recently been completed. ARAMIS has several modules which give a consistent simplified approach to risk assessment which does not approach the complexity or expense of full QRA. LOPA is potentially a means of carrying out the assessment of barriers required in ARAMIS. This paper attempts to explain the principles of LOPA and the means by which it can be used within ARAMIS.

  1. Validity and consistency assessment of accident analysis methods in the petroleum industry.

    Science.gov (United States)

    Ahmadi, Omran; Mortazavi, Seyed Bagher; Khavanin, Ali; Mokarami, Hamidreza

    2017-11-17

    Accident analysis is the main aspect of accident investigation. It includes the method of connecting different causes in a procedural way. Therefore, it is important to use valid and reliable methods for the investigation of different causal factors of accidents, especially the noteworthy ones. This study aimed to prominently assess the accuracy (sensitivity index [SI]) and consistency of the six most commonly used accident analysis methods in the petroleum industry. In order to evaluate the methods of accident analysis, two real case studies (process safety and personal accident) from the petroleum industry were analyzed by 10 assessors. The accuracy and consistency of these methods were then evaluated. The assessors were trained in the workshop of accident analysis methods. The systematic cause analysis technique and bowtie methods gained the greatest SI scores for both personal and process safety accidents, respectively. The best average results of the consistency in a single method (based on 10 independent assessors) were in the region of 70%. This study confirmed that the application of methods with pre-defined causes and a logic tree could enhance the sensitivity and consistency of accident analysis.

  2. X-chromosome inactivation pattern analysis for the assessment of cell clonality in cats.

    Science.gov (United States)

    Mochizuki, H; Goto-Koshino, Y; Takahashi, M; Fujino, Y; Ohno, K; Tsujimoto, H

    2012-11-01

    X-chromosome inactivation pattern (XCIP) analysis has been widely used to assess cell clonality in various types of neoplasms in humans. In the present study, a polymerase chain reaction-based feline XCIP analysis using the feline androgen receptor gene was developed. To construct the system of the analysis, polymorphism in CAG tandem repeats within the feline androgen receptor gene was explored using somatic DNAs from 50 male and 103 female cats. CAG tandem repeats in exon 1 of the feline androgen receptor gene were found to be polymorphic, containing 15 to 22 CAG repeats. Of the 103 female cats, 70 (68%) were heterozygous for the number of CAG repeats, indicating the possible usefulness of XCIP analysis in cats. Application of the feline XCIP analysis to 3 feline mammary gland adenocarcinoma cell lines revealed distinctly skewed XCIPs in these cell lines, indicating their clonal origins. Twelve (80%) of the 15 primary tissue/cell samples obtained from cats with various neoplastic diseases showed skewed XCIPs. Moreover, bone marrow samples from 3 cats with myelodysplastic syndrome were also found to have skewed XCIPs. The polymerase chain reaction-based XCIP analysis developed in this study can provide information on cell clonality in female cats, potentially facilitating the differential diagnosis of various disorders in cats.

  3. Risky Business: Factor Analysis of Survey Data – Assessing the Probability of Incorrect Dimensionalisation

    Science.gov (United States)

    van der Eijk, Cees; Rose, Jonathan

    2015-01-01

    This paper undertakes a systematic assessment of the extent to which factor analysis the correct number of latent dimensions (factors) when applied to ordered-categorical survey items (so-called Likert items). We simulate 2400 data sets of uni-dimensional Likert items that vary systematically over a range of conditions such as the underlying population distribution, the number of items, the level of random error, and characteristics of items and item-sets. Each of these datasets is factor analysed in a variety of ways that are frequently used in the extant literature, or that are recommended in current methodological texts. These include exploratory factor retention heuristics such as Kaiser’s criterion, Parallel Analysis and a non-graphical scree test, and (for exploratory and confirmatory analyses) evaluations of model fit. These analyses are conducted on the basis of Pearson and polychoric correlations. We find that, irrespective of the particular mode of analysis, factor analysis applied to ordered-categorical survey data very often leads to over-dimensionalisation. The magnitude of this risk depends on the specific way in which factor analysis is conducted, the number of items, the properties of the set of items, and the underlying population distribution. The paper concludes with a discussion of the consequences of over-dimensionalisation, and a brief mention of alternative modes of analysis that are much less prone to such problems. PMID:25789992

  4. Identifying items to assess methodological quality in physical therapy trials: a factor analysis.

    Science.gov (United States)

    Armijo-Olivo, Susan; Cummings, Greta G; Fuentes, Jorge; Saltaji, Humam; Ha, Christine; Chisholm, Annabritt; Pasichnyk, Dion; Rogers, Todd

    2014-09-01

    Numerous tools and individual items have been proposed to assess the methodological quality of randomized controlled trials (RCTs). The frequency of use of these items varies according to health area, which suggests a lack of agreement regarding their relevance to trial quality or risk of bias. The objectives of this study were: (1) to identify the underlying component structure of items and (2) to determine relevant items to evaluate the quality and risk of bias of trials in physical therapy by using an exploratory factor analysis (EFA). A methodological research design was used, and an EFA was performed. Randomized controlled trials used for this study were randomly selected from searches of the Cochrane Database of Systematic Reviews. Two reviewers used 45 items gathered from 7 different quality tools to assess the methodological quality of the RCTs. An exploratory factor analysis was conducted using the principal axis factoring (PAF) method followed by varimax rotation. Principal axis factoring identified 34 items loaded on 9 common factors: (1) selection bias; (2) performance and detection bias; (3) eligibility, intervention details, and description of outcome measures; (4) psychometric properties of the main outcome; (5) contamination and adherence to treatment; (6) attrition bias; (7) data analysis; (8) sample size; and (9) control and placebo adequacy. Because of the exploratory nature of the results, a confirmatory factor analysis is needed to validate this model. To the authors' knowledge, this is the first factor analysis to explore the underlying component items used to evaluate the methodological quality or risk of bias of RCTs in physical therapy. The items and factors represent a starting point for evaluating the methodological quality and risk of bias in physical therapy trials. Empirical evidence of the association among these items with treatment effects and a confirmatory factor analysis of these results are needed to validate these items.

  5. Using generalizability analysis to estimate parameters for anatomy assessments: A multi-institutional study.

    Science.gov (United States)

    Byram, Jessica N; Seifert, Mark F; Brooks, William S; Fraser-Cotlin, Laura; Thorp, Laura E; Williams, James M; Wilson, Adam B

    2017-03-01

    With integrated curricula and multidisciplinary assessments becoming more prevalent in medical education, there is a continued need for educational research to explore the advantages, consequences, and challenges of integration practices. This retrospective analysis investigated the number of items needed to reliably assess anatomical knowledge in the context of gross anatomy and histology. A generalizability analysis was conducted on gross anatomy and histology written and practical examination items that were administered in a discipline-based format at Indiana University School of Medicine and in an integrated fashion at the University of Alabama School of Medicine and Rush University Medical College. Examination items were analyzed using a partially nested design s×(i:o) in which items were nested within occasions (i:o) and crossed with students (s). A reliability standard of 0.80 was used to determine the minimum number of items needed across examinations (occasions) to make reliable and informed decisions about students' competence in anatomical knowledge. Decision study plots are presented to demonstrate how the number of items per examination influences the reliability of each administered assessment. Using the example of a curriculum that assesses gross anatomy knowledge over five summative written and practical examinations, the results of the decision study estimated that 30 and 25 items would be needed on each written and practical examination to reach a reliability of 0.80, respectively. This study is particularly relevant to educators who may question whether the amount of anatomy content assessed in multidisciplinary evaluations is sufficient for making judgments about the anatomical aptitude of students. Anat Sci Educ 10: 109-119. © 2016 American Association of Anatomists. © 2016 American Association of Anatomists.

  6. Natural Hazard Susceptibility Assessment for Road Planning Using Spatial Multi-Criteria Analysis.

    Science.gov (United States)

    Karlsson, Caroline S J; Kalantari, Zahra; Mörtberg, Ulla; Olofsson, Bo; Lyon, Steve W

    2017-08-18

    Inadequate infrastructural networks can be detrimental to society if transport between locations becomes hindered or delayed, especially due to natural hazards which are difficult to control. Thus determining natural hazard susceptible areas and incorporating them in the initial planning process, may reduce infrastructural damages in the long run. The objective of this study was to evaluate the usefulness of expert judgments for assessing natural hazard susceptibility through a spatial multi-criteria analysis approach using hydrological, geological, and land use factors. To utilize spatial multi-criteria analysis for decision support, an analytic hierarchy process was adopted where expert judgments were evaluated individually and in an aggregated manner. The estimates of susceptible areas were then compared with the methods weighted linear combination using equal weights and factor interaction method. Results showed that inundation received the highest susceptibility. Using expert judgment showed to perform almost the same as equal weighting where the difference in susceptibility between the two for inundation was around 4%. The results also showed that downscaling could negatively affect the susceptibility assessment and be highly misleading. Susceptibility assessment through spatial multi-criteria analysis is useful for decision support in early road planning despite its limitation to the selection and use of decision rules and criteria. A natural hazard spatial multi-criteria analysis could be used to indicate areas where more investigations need to be undertaken from a natural hazard point of view, and to identify areas thought to have higher susceptibility along existing roads where mitigation measures could be targeted after in-situ investigations.

  7. [Objective assessment of facial paralysis using infrared thermography and formal concept analysis].

    Science.gov (United States)

    Liu, Xu-Long; Hong, Wen-Xue; Liu, Jie-Min

    2014-04-01

    This paper presented a novel approach to objective assessment of facial nerve paralysis based on infrared thermography and formal concept analysis. Sixty five patients with facial nerve paralysis on one side were included in the study. The facial temperature distribution images of these 65 patients were captured by infrared thermography every five days during one-month period. First, the facial thermal images were pre-processed to identify six potential regions of bilateral symmetry by using image segmentation techniques. Then, the temperature differences on the left and right sides of the facial regions were extracted and analyzed. Finally, the authors explored the relationships between the statistical averages of those temperature differences and the House-Brackmann score for objective assessment degree of nerve damage in a facial nerve paralysis by using formal concept analysis. The results showed that the facial temperature distribution of patients with facial nerve paralysis exhibited a contralateral asymmetry, and the bilateral temperature differences of the facial regions were greater than 0.2 degrees C, whereas in normal healthy individuals these temperature differences were less than 0.2 degrees C. Spearman correlation coefficient between the bilateral temperature differences of the facial regions and the degree of facial nerve damage was an average of 0.508, which was statistically significant (p facial regions was greater than 0.2 degrees C, and all were less than 0.5 degrees C, facial nerve paralysis could be determined as for the mild to moderate; if one of the temperature differences of bilateral symmetry was greater than 0.5 degrees C, facial nerve paralysis could be determined as for serious. In conclusion, this paper presents an automated technique for the computerized analysis of thermal images to objectively assess facial nerve related thermal dysfunction by using formal concept analysis theory, which may benefit the clinical diagnosis and

  8. Using global sensitivity analysis of demographic models for ecological impact assessment.

    Science.gov (United States)

    Aiello-Lammens, Matthew E; Akçakaya, H Resit

    2017-02-01

    Population viability analysis (PVA) is widely used to assess population-level impacts of environmental changes on species. When combined with sensitivity analysis, PVA yields insights into the effects of parameter and model structure uncertainty. This helps researchers prioritize efforts for further data collection so that model improvements are efficient and helps managers prioritize conservation and management actions. Usually, sensitivity is analyzed by varying one input parameter at a time and observing the influence that variation has over model outcomes. This approach does not account for interactions among parameters. Global sensitivity analysis (GSA) overcomes this limitation by varying several model inputs simultaneously. Then, regression techniques allow measuring the importance of input-parameter uncertainties. In many conservation applications, the goal of demographic modeling is to assess how different scenarios of impact or management cause changes in a population. This is challenging because the uncertainty of input-parameter values can be confounded with the effect of impacts and management actions. We developed a GSA method that separates model outcome uncertainty resulting from parameter uncertainty from that resulting from projected ecological impacts or simulated management actions, effectively separating the 2 main questions that sensitivity analysis asks. We applied this method to assess the effects of predicted sea-level rise on Snowy Plover (Charadrius nivosus). A relatively small number of replicate models (approximately 100) resulted in consistent measures of variable importance when not trying to separate the effects of ecological impacts from parameter uncertainty. However, many more replicate models (approximately 500) were required to separate these effects. These differences are important to consider when using demographic models to estimate ecological impacts of management actions. © 2016 Society for Conservation Biology.

  9. Natural Hazard Susceptibility Assessment for Road Planning Using Spatial Multi-Criteria Analysis

    Science.gov (United States)

    Karlsson, Caroline S. J.; Kalantari, Zahra; Mörtberg, Ulla; Olofsson, Bo; Lyon, Steve W.

    2017-11-01

    Inadequate infrastructural networks can be detrimental to society if transport between locations becomes hindered or delayed, especially due to natural hazards which are difficult to control. Thus determining natural hazard susceptible areas and incorporating them in the initial planning process, may reduce infrastructural damages in the long run. The objective of this study was to evaluate the usefulness of expert judgments for assessing natural hazard susceptibility through a spatial multi-criteria analysis approach using hydrological, geological, and land use factors. To utilize spatial multi-criteria analysis for decision support, an analytic hierarchy process was adopted where expert judgments were evaluated individually and in an aggregated manner. The estimates of susceptible areas were then compared with the methods weighted linear combination using equal weights and factor interaction method. Results showed that inundation received the highest susceptibility. Using expert judgment showed to perform almost the same as equal weighting where the difference in susceptibility between the two for inundation was around 4%. The results also showed that downscaling could negatively affect the susceptibility assessment and be highly misleading. Susceptibility assessment through spatial multi-criteria analysis is useful for decision support in early road planning despite its limitation to the selection and use of decision rules and criteria. A natural hazard spatial multi-criteria analysis could be used to indicate areas where more investigations need to be undertaken from a natural hazard point of view, and to identify areas thought to have higher susceptibility along existing roads where mitigation measures could be targeted after in-situ investigations.

  10. Assessment

    Directory of Open Access Journals (Sweden)

    E.S. Mohamed

    2013-12-01

    Full Text Available Soil erosion, salinity and sodicity hazards are serious problems in the northern west coast of Egypt and lead to reducing the soil quality and increasing the degradation of soil resources. Sidi Barrani and Al-Sallum regions are selected as study areas which are located from a longitude of 25°10′00″ to 26°55′00″East and from a latitude of 31°00′0″ to 31°37′30″ North. Erosion hazard was estimated using the ‘Universal Soil Loss Equation’ (USLE, which is a simple empirical model that is widely used for assessing long-term annual soil loss .The salinity and sodicity hazards were estimated based on FAO method as standard reference. The resultant map of annual soil erosion shows a maximum soil loss of 60 t h−1 y−1 with a close relation to foot slopes and wide units on the steep side-slopes (with high LS value and the erodibility value reached to 0.1 t h−1 y−1. Meanwhile sand beaches and sabkha units are characterized by high environmental hazards of both water erosion, salinity and sodicity, while in the overflow basin units are identified as low environmental hazards. The spatial environmental hazards assessment is conducted by using integrated GIS and RS which can serve as effective inputs in deriving strategies for sustainable land use planning and management.

  11. Clinical assessment of infection in nonhealing ulcers analyzed by latent class analysis.

    Science.gov (United States)

    Lorentzen, Henrik F; Gottrup, Finn

    2006-01-01

    The distinction between bacterial colonization and infection relies on clinical judgement. Determining sensitivity and specificity of this judgement are problematic as no gold standard exists. Six specialists in wound management independently assessed 120 nonhealing chronic wounds. Sixty-five (54.2%) patients had venous ulcer, 18 (15%) arterial ulcers, 15 (12.5%) ulcerative pyoderma gangraenosum, 12 (10%) neuropathic or pressure ulcers, six (5%) vasculitis ulcers, and four patients had ulcers caused by a primary or metastatic cancer disease. Unrestricted latent class analysis was used for determining sensitivity and specificity in the observer's assessment of hypergranulation, redness, and overall impression of infection. Interrater agreement among observers was determined by restricted latent class analysis. The observers used the diagnoses (redness, hypergranulation, and overall impression of infection with different frequencies (phypergranulation ranged from 3 to 82%, for redness from 34 to 91% and for overall impression of infection from 37 to 90%. None of the observers were interchangeable. These results indicate that clinical assessment of chronic wounds for the presence of infection are difficult tasks accompanied by great variability and low reliability.

  12. Comparison of local sine wave modeling with harmonic phase analysis for the assessment of myocardial strain.

    Science.gov (United States)

    Miller, Christopher A; Borg, Alex; Clark, David; Steadman, Christopher D; McCann, Gerry P; Clarysse, Patrick; Croisille, Pierre; Schmitt, Matthias

    2013-08-01

    To compare local sine-wave modeling (SinMod) with harmonic phase analysis (HARP), for assessment of left ventricular (LV) circumferential strain (εcc) from tagged cardiovascular magnetic resonance images. Mid-ventricular spatial modulation of magnetization was performed in 60 participants (15 each with hypertrophic, dilated or ischemic cardiomyopathy and 15 healthy controls) at 1.5 Tesla. Global and segmental peak transmural εcc were measured using HARP and SinMod. Repeated measurements were performed on 25% of examinations to assess observer variability. Effect of contrast was assessed in 10 additional patients. SinMod showed a high level of agreement with HARP for global εcc (mean difference -0.02, 95% limits of agreement -6.46 to 6.43%). Agreement was much lower for segmental εcc. Both methods showed excellent observer agreement for global εcc (intraclass correlation coefficient >0.75). Observer agreement for segmental εcc was also excellent with SinMod, but was significantly lower with HARP. Analysis time was significantly shorter using SinMod. Pre- and postcontrast εcc measurements were not significantly different using either technique, although postcontrast measurements showed greater variability with HARP. SinMod and HARP-based measurements of global εcc have a high level of agreement, but segmental agreement is substantially lower. SinMod has generally lower observer variability, is faster and is less affected by contrast, but requires further validation. Copyright © 2012 Wiley Periodicals, Inc.

  13. Is It Working? Distractor Analysis Results from the Test Of Astronomy STandards (TOAST) Assessment Instrument

    Science.gov (United States)

    Slater, Stephanie

    2009-05-01

    The Test Of Astronomy STandards (TOAST) assessment instrument is a multiple-choice survey tightly aligned to the consensus learning goals stated by the American Astronomical Society - Chair's Conference on ASTRO 101, the American Association of the Advancement of Science's Project 2061 Benchmarks, and the National Research Council's National Science Education Standards. Researchers from the Cognition in Astronomy, Physics and Earth sciences Research (CAPER) Team at the University of Wyoming's Science and Math Teaching Center (UWYO SMTC) have been conducting a question-by-question distractor analysis procedure to determine the sensitivity and effectiveness of each item. In brief, the frequency each possible answer choice, known as a foil or distractor on a multiple-choice test, is determined and compared to the existing literature on the teaching and learning of astronomy. In addition to having statistical difficulty and discrimination values, a well functioning assessment item will show students selecting distractors in the relative proportions to how we expect them to respond based on known misconceptions and reasoning difficulties. In all cases, our distractor analysis suggests that all items are functioning as expected. These results add weight to the validity of the Test Of Astronomy STandards (TOAST) assessment instrument, which is designed to help instructors and researchers measure the impact of course-length duration instructional strategies for undergraduate science survey courses with learning goals tightly aligned to the consensus goals of the astronomy education community.

  14. Laser cutting of various materials: Kerf width size analysis and life cycle assessment of cutting process

    Science.gov (United States)

    Yilbas, Bekir Sami; Shaukat, Mian Mobeen; Ashraf, Farhan

    2017-08-01

    Laser cutting of various materials including Ti-6Al-4V alloy, steel 304, Inconel 625, and alumina is carried out to assess the kerf width size variation along the cut section. The life cycle assessment is carried out to determine the environmental impact of the laser cutting in terms of the material waste during the cutting process. The kerf width size is formulated and predicted using the lump parameter analysis and it is measured from the experiments. The influence of laser output power and laser cutting speed on the kerf width size variation is analyzed using the analytical tools including scanning electron and optical microscopes. In the experiments, high pressure nitrogen assisting gas is used to prevent oxidation reactions in the cutting section. It is found that the kerf width size predicted from the lump parameter analysis agrees well with the experimental data. The kerf width size variation increases with increasing laser output power. However, this behavior reverses with increasing laser cutting speed. The life cycle assessment reveals that material selection for laser cutting is critical for the environmental protection point of view. Inconel 625 contributes the most to the environmental damages; however, recycling of the waste of the laser cutting reduces this contribution.

  15. Discrete wavelet transform analysis of surface electromyography for the fatigue assessment of neck and shoulder muscles.

    Science.gov (United States)

    Chowdhury, Suman Kanti; Nimbarte, Ashish D; Jaridi, Majid; Creese, Robert C

    2013-10-01

    Assessment of neuromuscular fatigue is essential for early detection and prevention of risks associated with work-related musculoskeletal disorders. In recent years, discrete wavelet transform (DWT) of surface electromyography (SEMG) has been used to evaluate muscle fatigue, especially during dynamic contractions when the SEMG signal is non-stationary. However, its application to the assessment of work-related neck and shoulder muscle fatigue is not well established. Therefore, the purpose of this study was to establish DWT analysis as a suitable method to conduct quantitative assessment of neck and shoulder muscle fatigue under dynamic repetitive conditions. Ten human participants performed 40min of fatiguing repetitive arm and neck exertions while SEMG data from the upper trapezius and sternocleidomastoid muscles were recorded. The ten of the most commonly used wavelet functions were used to conduct the DWT analysis. Spectral changes estimated using power of wavelet coefficients in the 12-23Hz frequency band showed the highest sensitivity to fatigue induced by the dynamic repetitive exertions. Although most of the wavelet functions tested in this study reasonably demonstrated the expected power trend with fatigue development and recovery, the overall performance of the "Rbio3.1" wavelet in terms of power estimation and statistical significance was better than the remaining nine wavelets. Copyright © 2013 Elsevier Ltd. All rights reserved.

  16. Analysis and assessment of environmental impacts of small hydro power plant in Slovakia

    Science.gov (United States)

    Zeleňáková, M.; Fijko, R.; Remeňáková, I.

    2017-10-01

    Environmental impact assessment (EIA) is an important process that, prior to approval of the investment plan, can provide a detailed examination of the likely and foreseeable impacts of proposed construction activity on the environment. The objective of this paper is to apply a specific methodology for the analysis and evaluation of the environmental impacts of selected constructions, namely, small hydro power plant, using matrix of impacts. This analysis method is intended not only to increase the clarity and precision of the evaluation process, but also to align it with the requirements of the environmental impact assessment system. This modification should improve the reliability of the environmental impact assessment, and could moreover also be applied to other infrastructure projects. Comparison of alternatives and designation of the optimal variant are implemented based on selected criteria that objectively describe the characteristic lines of the planned alternatives of activity and their impact on the environment. The use of proper EIA procedures can help the decision-makers to formulate proper activities based on qualified decisions. The designed project in Spišské Bystré, Slovakia is used as a case study to clarify and exemplify the methodology and techniques.

  17. Qualitative syndrome analysis by neuropsychological assessment in preschoolers with attention deficit disorder with hyperactivity.

    Directory of Open Access Journals (Sweden)

    Quintanar L.

    2015-07-01

    Full Text Available The basis of this article is A. R. Luria’s conception of the qualitative approach in neuropsychology. Attention deficit disorder with hyperactivity is one of the most frequent clinical diagnoses given during preschool age. However, precise qualitative neuropsychological criteria for analysis of this clinical diagnosis do not exist and change from one approach to another. Our objective here is to propose such qualitative criteria for neuropsychological analysis of children with diagnoses of attention deficit disorder with hyperactivity at preschool age. We follow clinical methodology that is traditional for historical and cultural neuropsychology and is an alternative for the psychometrical and cognitive approach. The methodology of the study was qualitative neuropsychological assessment of the syndrome followed by detailed consideration of the types of difficulties in each case. The study analyzes mistakes and typical examples of execution of the tasks of neuropsychological qualitative assessment by regular children and by children with attention deficit disorder with hyperactivity. The results showed differences between these groups of children. The children with attention deficit disorder with hyperactivity made a lot of mistakes during assessments. Their difficulties are related to unfavorable conditions in the three functional brain blocks according to Luria’s conception. We conclude that “attention” cannot be considered the only or the main problem in children who receive this diagnosis by psychiatrists and neurologists.

  18. Assessing the Geographic Expression of Urban Sustainability: A Scenario Based Approach Incorporating Spatial Multicriteria Decision Analysis

    Directory of Open Access Journals (Sweden)

    James K. Lein

    2012-09-01

    Full Text Available Urban sustainability involves a re-examination of urban development including environmental, social and economic policies and practices that acknowledge the role of cities in global environmental change. However, sustainability remains a broadly defined concept that has been applied to mean everything from environmental protection, social cohesion, economic growth, neighborhood design, alternative energy, and green building design. To guide sustainability initiatives and assess progress toward more sustainable development patterns this construct requires a means to place this concept into a decision-centric context where change can be evaluated and the exploitation of resources, the direction of investment, the orientation of technological development, and institutional programs can be made more consistent with future as well as present needs. In this study the problem of sustainability assessment was examined and a method that couples scenario analysis with spatial multicriteria decision analysis was introduced. The integration of a spatial multicriteria decision analysis (MCDA model for sustainable development with scenario planning resulted in an interpretation of sustainability that is more appropriate for local conditions and useful when exploring sustainability’s semantic uncertainties, particularly those alternate perspectives that influence future environments.

  19. Pesticide flow analysis to assess human exposure in greenhouse flower production in Colombia.

    Science.gov (United States)

    Lesmes-Fabian, Camilo; Binder, Claudia R

    2013-03-25

    Human exposure assessment tools represent a means for understanding human exposure to pesticides in agricultural activities and managing possible health risks. This paper presents a pesticide flow analysis modeling approach developed to assess human exposure to pesticide use in greenhouse flower crops in Colombia, focusing on dermal and inhalation exposure. This approach is based on the material flow analysis methodology. The transfer coefficients were obtained using the whole body dosimetry method for dermal exposure and the button personal inhalable aerosol sampler for inhalation exposure, using the tracer uranine as a pesticide surrogate. The case study was a greenhouse rose farm in the Bogota Plateau in Colombia. The approach was applied to estimate the exposure to pesticides such as mancozeb, carbendazim, propamocarb hydrochloride, fosetyl, carboxin, thiram, dimethomorph and mandipropamide. We found dermal absorption estimations close to the AOEL reference values for the pesticides carbendazim, mancozeb, thiram and mandipropamide during the study period. In addition, high values of dermal exposure were found on the forearms, hands, chest and legs of study participants, indicating weaknesses in the overlapping areas of the personal protective equipment parts. These results show how the material flow analysis methodology can be applied in the field of human exposure for early recognition of the dispersion of pesticides and support the development of measures to improve operational safety during pesticide management. Furthermore, the model makes it possible to identify the status quo of the health risk faced by workers in the study area.

  20. Quantitative Assessment of In-solution Digestion Efficiency Identifies Optimal Protocols for Unbiased Protein Analysis*

    Science.gov (United States)

    León, Ileana R.; Schwämmle, Veit; Jensen, Ole N.; Sprenger, Richard R.

    2013-01-01

    The majority of mass spectrometry-based protein quantification studies uses peptide-centric analytical methods and thus strongly relies on efficient and unbiased protein digestion protocols for sample preparation. We present a novel objective approach to assess protein digestion efficiency using a combination of qualitative and quantitative liquid chromatography-tandem MS methods and statistical data analysis. In contrast to previous studies we employed both standard qualitative as well as data-independent quantitative workflows to systematically assess trypsin digestion efficiency and bias using mitochondrial protein fractions. We evaluated nine trypsin-based digestion protocols, based on standard in-solution or on spin filter-aided digestion, including new optimized protocols. We investigated various reagents for protein solubilization and denaturation (dodecyl sulfate, deoxycholate, urea), several trypsin digestion conditions (buffer, RapiGest, deoxycholate, urea), and two methods for removal of detergents before analysis of peptides (acid precipitation or phase separation with ethyl acetate). Our data-independent quantitative liquid chromatography-tandem MS workflow quantified over 3700 distinct peptides with 96% completeness between all protocols and replicates, with an average 40% protein sequence coverage and an average of 11 peptides identified per protein. Systematic quantitative and statistical analysis of physicochemical parameters demonstrated that deoxycholate-assisted in-solution digestion combined with phase transfer allows for efficient, unbiased generation and recovery of peptides from all protein classes, including membrane proteins. This deoxycholate-assisted protocol was also optimal for spin filter-aided digestions as compared with existing methods. PMID:23792921

  1. Rapid assessment of large scale vegetation change based on multi-temporal phenological analysis

    Science.gov (United States)

    Cai, Danlu; Guan, Yanning; Guo, Shan; Yan, Baoping; Xing, Zhi; Zhang, Chunyan; Piao, Yingchao; An, Xudong; Kang, Lihua

    2011-11-01

    Detecting vegetation change is critical for earth system and sustainability science. The existing methods, however, show several limitations, including inevitable selection of imagery acquisition dates, affection from vegetation related noise on temporal trajectory analysis, and assumptions due to vegetation classification model. This paper presents a multitemporal phenological frequency analysis over a relatively short period (MTPFA-SP) methodology to detect vegetation changes. This MTPFA-SP methodology bases on the amplitude components of fast Fourier transforming (FFT) and is implemented with two steps. First, NDVI time series over two periods are transformed with FFT into frequency domain, separately. Second, amplitude components with phenological information from Step 1 are selected for further change comparison. In this methodology, component selection shows physical meanings of natural vegetation process in frequency domain. Comparisons among those selected components help enhance the ability to rapidly detect vegetation changes. To validate this MTPFA-SP methodology, we detect changes between two periods (2001-2005 and 2006-2010) in the eastern Tibet Plateau area and make two kinds of assessments. The first is for a larger scale, including statistic analysis of altitudinal zonality and latitudinal zonality. The second assessment is for rapid detection of vegetation change location. Landsat TM image were employed to validate the result.

  2. Pesticide Flow Analysis to Assess Human Exposure in Greenhouse Flower Production in Colombia

    Directory of Open Access Journals (Sweden)

    Claudia R. Binder

    2013-03-01

    Full Text Available Human exposure assessment tools represent a means for understanding human exposure to pesticides in agricultural activities and managing possible health risks. This paper presents a pesticide flow analysis modeling approach developed to assess human exposure to pesticide use in greenhouse flower crops in Colombia, focusing on dermal and inhalation exposure. This approach is based on the material flow analysis methodology. The transfer coefficients were obtained using the whole body dosimetry method for dermal exposure and the button personal inhalable aerosol sampler for inhalation exposure, using the tracer uranine as a pesticide surrogate. The case study was a greenhouse rose farm in the Bogota Plateau in Colombia. The approach was applied to estimate the exposure to pesticides such as mancozeb, carbendazim, propamocarb hydrochloride, fosetyl, carboxin, thiram, dimethomorph and mandipropamide. We found dermal absorption estimations close to the AOEL reference values for the pesticides carbendazim, mancozeb, thiram and mandipropamide during the study period. In addition, high values of dermal exposure were found on the forearms, hands, chest and legs of study participants, indicating weaknesses in the overlapping areas of the personal protective equipment parts. These results show how the material flow analysis methodology can be applied in the field of human exposure for early recognition of the dispersion of pesticides and support the development of measures to improve operational safety during pesticide management. Furthermore, the model makes it possible to identify the status quo of the health risk faced by workers in the study area.

  3. Robustness of arterial blood gas analysis for assessment of respiratory safety pharmacology in rats.

    Science.gov (United States)

    Whiteside, Garth T; Hummel, Michele; Boulet, Jamie; Beyenhof, Jessica D; Strenkowski, Bryan; John, Janet Dell; Knappenberger, Terri; Maselli, Harry; Koetzner, Lee

    2016-01-01

    Whole body plethysmography using unrestrained animals is a common technique for assessing the respiratory risk of new drugs in safety pharmacology studies in rats. However, wide variations in experimental technique make cross laboratory comparison of data difficult and raise concerns that non-appropriate conditions may mask the deleterious effects of test compounds - in particular with suspected respiratory depressants. Therefore, the objective of this study was to evaluate the robustness of arterial blood gas analysis as an alternative to plethysmography in rats. We sought to do this by assessing the effect of different vehicles and times post-surgical catheterization on blood gas measurements, in addition to determining sensitivity to multiple opioids. Furthermore, we determined intra-lab variability from multiple datasets utilizing morphine and generated within a single lab and lastly, inter-lab variability was measured by comparing datasets generated in two separate labs. Overall, our data show that arterial blood gas analysis is a measure that is both flexible in terms of experimental conditions and highly sensitive to respiratory depressants, two key limitations when using plethysmography. As such, our data strongly advocate the adoption of arterial blood gas analysis as an investigative approach to reliably examine the respiratory depressant effects of opioids. Copyright © 2015 Elsevier Inc. All rights reserved.

  4. Different methods for ethical analysis in health technology assessment: an empirical study.

    Science.gov (United States)

    Saarni, Samuli I; Braunack-Mayer, Annette; Hofmann, Bjørn; van der Wilt, Gert Jan

    2011-10-01

    Ethical analysis can highlight important ethical issues related to implementing a technology, values inherent in the technology itself, and value-decisions underlying the health technology assessment (HTA) process. Ethical analysis is a well-acknowledged part of HTA, yet seldom included in practice. One reason for this is lack of knowledge about the properties and differences between the methods available. This study compares different methods for ethical analysis within HTA. Ethical issues related to bariatric (obesity) surgery were independently evaluated using axiological, casuist, principlist, and EUnetHTA models for ethical analysis within HTA. The methods and results are presented and compared. Despite varying theoretical underpinnings and practical approaches, the four methods identified similar themes: personal responsibility, self-infliction, discrimination, justice, public funding, and stakeholder involvement. The axiological and EUnetHTA models identified a wider range of arguments, whereas casuistry and principlism concentrated more on analyzing a narrower set of arguments deemed more important. Different methods can be successfully used for conducting ethical analysis within HTA. Although our study does not show that different methods in ethics always produce similar results, it supports the view that different methods of ethics can yield relevantly similar results. This suggests that the key conclusions of ethical analyses within HTA can be transferable between methods and countries. The systematic and transparent use of some method of ethics appears more important than the choice of the exact method.

  5. Quality assessment of Fructus Ligustri Lucidi by the simultaneous determination of six compounds and chemometric analysis.

    Science.gov (United States)

    Yao, Weifeng; Dai, Jin; Zheng, Chuanzhu; Bao, Beihua; Cheng, Haibo; Zhang, Li; Ding, Anwei; Li, Wei

    2015-06-01

    A comprehensive strategy was designed for the quality assessment of Fructus Ligustri Lucidi, a well-known and commonly used herbal medicine in clinical practice in China. First, a simple and stable method of high-performance liquid chromatography was developed for the simultaneous quantitative analysis of six compounds, namely, salidroside, nuzhenide, specnuezhenide, oleanic acid, ursolic acid, and acetyl oleanic acid in Fructus Ligustri Lucidi. The separation of analytes was conducted on a C18 column (200 mm × 4.6 mm, 5 μm) at 30°C, and the wavelength of UV detector was set at 210 nm. In quantitative analysis, all of the calibration curves showed good linear regression (R(2) > 0.9994) within the tested ranges, and the mean recoveries of three different concentrations ranged from 95.21-102.34%. The described method was applied to determine 11 batches of samples collected from different stores in China. Then multiple chemometrics analysis including hierarchical cluster analysis and principal component analysis were performed to classify samples and search significant compounds. Three notable compounds, specnuezhenide, oleanic acid, and acetyl oleanic acid, were discovered for better quality control compared with those stated in the China pharmacopeia. The results demonstrated that this strategy could be readily utilized for the comprehensive quality control of Fructus Ligustri Lucidi. © 2015 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  6. Fast eutrophication assessment for stormwater wet detention ponds via fuzzy probit regression analysis under uncertainty.

    Science.gov (United States)

    Tahsin, Subrina; Chang, Ni-Bin

    2016-02-01

    Stormwater wet detention ponds have been a commonly employed best management practice for stormwater management throughout the world for many years. In the past, the trophic state index values have been used to evaluate seasonal changes in water quality and rank lakes within a region or between several regions; yet, to date, there is no similar index for stormwater wet detention ponds. This study aimed to develop a new multivariate trophic state index (MTSI) suitable for conducting a rapid eutrophication assessment of stormwater wet detention ponds under uncertainty with respect to three typical physical and chemical properties. Six stormwater wet detention ponds in Florida were selected for demonstration of the new MTSI with respect to total phosphorus (TP), total nitrogen (TN), and Secchi disk depth (SDD) as cognitive assessment metrics to sense eutrophication potential collectively and inform the environmental impact holistically. Due to the involvement of multiple endogenous variables (i.e., TN, TP, and SDD) for the eutrophication assessment simultaneously under uncertainty, fuzzy synthetic evaluation was applied to first standardize and synchronize the sources of uncertainty in the decision analysis. The ordered probit regression model was then formulated for assessment based on the concept of MTSI with the inputs from the fuzzy synthetic evaluation. It is indicative that the severe eutrophication condition is present during fall, which might be due to frequent heavy summer storm events contributing to high-nutrient inputs in these six ponds.

  7. Sex assessment from the carpals bones: discriminant function analysis in a 20th century Spanish sample.

    Science.gov (United States)

    Mastrangelo, Paola; De Luca, Stefano; Alemán, Inmaculada; Botella, Miguel C

    2011-03-20

    Sex assessment is one of the first essential steps in human identification, in both medico-legal cases and bio-archaeological contexts. Fragmentary human remains compromised by different types of inhumation or physical insults may frustrate the use of the traditional sex estimation methods, such as the analysis of the skull and pelvis. Currently, the application of discriminant functions to sex unidentified skeletal remains is steadily increasing. However, several studies have demonstrated that, due to variation in size and patterns of sexual dimorphism, discriminant function equations are population-specific [1,2,5,12,61]. In this study, in order to improve sex assessment from skeletal remains and to establish population-specific discriminant functions, the diagnostic values of the carpal bones were considered. A sample of 100 individuals (50 males and 50 females) of known sex and age was analyzed. They belong to a 20th century identified collection from the Municipal Cemetery of "San José", Granada (Spain) and housed in the Laboratory of Anthropology, Faculty of Medicine, University of Granada (Spain). The age of the individuals ranged between 22 and 85 years. Between four and nine measurements of each carpal bone were performed [41,59]. Discriminant function statistics showed a sex assessment accuracy as high as 97.8%. The results suggest that carpal bones can be used for assessing sex in both forensic and bio-archaeological identification procedures. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.

  8. Musculoskeletal Disorders Assessment and Posture Analysis by LUBA among Female Hairdressers in Tehran, 2015

    Directory of Open Access Journals (Sweden)

    Mohammad Khandan

    2017-01-01

    Full Text Available Background & Aims of the Study: Musculoskeletal disorders (MSDs are part of the main occupational diseases in the workplace. Occupations such as hairdressers are exposed to multiple risk factors of these problems. The study was conducted to assess MSDs and posture analysis among female hairdressers in Tehran, 2015. Materials and Methods: In this cross-sectional research, 114 participants were studied. To collect data, demographic questionnaire, body map for assessment of MSDs and Postural Loading on the Upper Body Assessment (LUBA method to evaluate postures was used. Also, data were analyzed by Mann-Whitney, Kruskal Wallis and Spearman correlation tests through SPSS-V20. Results: The mean and standard deviation of age and experience of the participants were5.34±8.9 and 10±8 years, respectively. In addition, they worked 9.8 hours per day on average. One hundred and thirteen (99.12% persons have experienced the pain at least in one member of their musculoskeletal system. Most of hairdressers had reported leg, lower back, as well as neck and shoulder pain. According to the posture assessment, 94.2% of people experienced high risk of exposure to risk factors for MSDs. Conclusion: Findings showed MSDs are high among barbers. Also, the work situations require immediate correction. Correction of workstations and tools design, work rest cycle and reduction in repetitive motions can help to improve working conditions.

  9. Resource efficiency of urban sanitation systems. A comparative assessment using material and energy flow analysis

    Energy Technology Data Exchange (ETDEWEB)

    Meinzinger, Franziska

    2010-07-01

    Within the framework of sustainable development it is important to find ways of reducing natural resource consumption and to change towards closed-loop management. As in many other spheres increased resource efficiency has also become an important issue in sanitation. Particularly nutrient recovery for agriculture, increased energy-efficiency and saving of natural water resources, can make a contribution to more resource efficient sanitation systems. To assess the resource efficiency of alternative developments a systems perspective is required. The present study applies a combined cost, energy and material flow analysis (ceMFA) as a system analysis method to assess the resource efficiency of urban sanitation systems. This includes the discussion of relevant criteria and assessment methods. The main focus of this thesis is the comparative assessment of different systems, based on two case studies; Hamburg in Germany and Arba Minch in Ethiopia. A range of possible system developments including source separation (e.g. diversion of urine or blackwater) is defined and compared with the current situation as a reference system. The assessment is carried out using computer simulations based on model equations. The model equations not only integrate mass and nutrient flows, but also the energy and cost balances of the different systems. In order to assess the impact of different assumptions and calculation parameters, sensitivity analyses and parameter variations complete the calculations. Based on the simulations, following general conclusions can be drawn: None of the systems show an overall benefit with regard to all investigated criteria, namely nutrients, energy, water and costs. Yet, the results of the system analysis can be used as basis for decision making if a case-related weighting is introduced. The systems show varying potential for the recovery of nutrients from (source separated) wastewater flows. For the case study of Hamburg up to 29% of the mineral

  10. Nondestructive inspection assessment of eddy current and electrochemical analysis to separate inconel and stainless steel alloys

    Energy Technology Data Exchange (ETDEWEB)

    Moore, D.G.; Sorensen, N.R.

    1998-02-01

    This report presents a nondestructive inspection assessment of eddy current and electrochemical analysis to separate inconel alloys from stainless steel alloys as well as an evaluation of cleaning techniques to remove a thermal oxide layer on aircraft exhaust components. The results of this assessment are presented in terms of how effective each technique classifies a known exhaust material. Results indicate that either inspection technique can separate inconel and stainless steel alloys. Based on the experiments conducted, the electrochemical spot test is the optimum for use by airframe and powerplant mechanics. A spot test procedure is proposed for incorporation into the Federal Aviation Administration Advisory Circular 65-9A Airframe & Powerplant Mechanic - General Handbook. 3 refs., 70 figs., 7 tabs.

  11. Confirmatory factor analysis of the Neuropsychological Assessment Battery of the LADIS study

    DEFF Research Database (Denmark)

    Moleiro, Carla; Madureira, Sofia; Verdelho, Ana

    2013-01-01

    Age-related white matter changes have been associated with cognitive functioning, even though their role is not fully understood. This work aimed to test a 3-factor model of the neuropsychological assessment battery and evaluate how the model fit the data longitudinally. Confirmatory factor...... analysis (CFA) was used to investigate the dimensions of a structured set of neuropsychological tests administered to a multicenter, international sample of independent older adults (LADIS study). Six hundred and thirty-eight older adults completed baseline neuropsychological, clinical, functional...... and motor assessments, which were repeated each year for a 3-year follow-up. CFA provided support for a 3-factor model. These factors involve the dimensions of executive functions, memory functions, and speed and motor control abilities. Performance decreased in most neuropsychological measures. Results...

  12. Neural image analysis in the process of quality assessment: domestic pig oocytes

    Science.gov (United States)

    Boniecki, P.; Przybył, J.; Kuzimska, T.; Mueller, W.; Raba, B.; Lewicki, A.; Przybył, K.; Zaborowicz, M.; Koszela, K.

    2014-04-01

    The questions related to quality classification of animal oocytes are explored by numerous scientific and research centres. This research is important, particularly in the context of improving the breeding value of farm animals. The methods leading to the stimulation of normal development of a larger number of fertilised animal oocytes in extracorporeal conditions are of special importance. Growing interest in the techniques of supported reproduction resulted in searching for new, increasingly effective methods for quality assessment of mammalian gametes and embryos. Progress in the production of in vitro animal embryos in fact depends on proper classification of obtained oocytes. The aim of this paper was the development of an original method for quality assessment of oocytes, performed on the basis of their graphical presentation in the form of microscopic digital images. The classification process was implemented on the basis of the information coded in the form of microphotographic pictures of the oocytes of domestic pig, using the modern methods of neural image analysis.

  13. Using statistical analysis and artificial intelligence tools for automatic assessment of video sequences

    Science.gov (United States)

    Ekobo Akoa, Brice; Simeu, Emmanuel; Lebowsky, Fritz

    2014-01-01

    This paper proposes two novel approaches to Video Quality Assessment (VQA). Both approaches attempt to develop video evaluation techniques capable of replacing human judgment when rating video quality in subjective experiments. The underlying study consists of selecting fundamental quality metrics based on Human Visual System (HVS) models and using artificial intelligence solutions as well as advanced statistical analysis. This new combination enables suitable video quality ratings while taking as input multiple quality metrics. The first method uses a neural network based machine learning process. The second method consists in evaluating the video quality assessment using non-linear regression model. The efficiency of the proposed methods is demonstrated by comparing their results with those of existing work done on synthetic video artifacts. The results obtained by each method are compared with scores from a database resulting from subjective experiments.

  14. Assessment of Static Delamination Propagation Capabilities in Commercial Finite Element Codes Using Benchmark Analysis

    Science.gov (United States)

    Orifici, Adrian C.; Krueger, Ronald

    2010-01-01

    With capabilities for simulating delamination growth in composite materials becoming available, the need for benchmarking and assessing these capabilities is critical. In this study, benchmark analyses were performed to assess the delamination propagation simulation capabilities of the VCCT implementations in Marc TM and MD NastranTM. Benchmark delamination growth results for Double Cantilever Beam, Single Leg Bending and End Notched Flexure specimens were generated using a numerical approach. This numerical approach was developed previously, and involves comparing results from a series of analyses at different delamination lengths to a single analysis with automatic crack propagation. Specimens were analyzed with three-dimensional and two-dimensional models, and compared with previous analyses using Abaqus . The results demonstrated that the VCCT implementation in Marc TM and MD Nastran(TradeMark) was capable of accurately replicating the benchmark delamination growth results and that the use of the numerical benchmarks offers advantages over benchmarking using experimental and analytical results.

  15. A Systematic Approach to Explorative Scenario Analysis in Emergy Assessment with Emphasis on Resilience

    DEFF Research Database (Denmark)

    Kamp, Andreas; Østergård, Hanne

    2016-01-01

    Fossil energy depletion (specifically peak oil) and climate change are imagined to profoundly affect human civilisation. This motivates assessment of resilience, a concept associated with the ability to persist and maintain function. Explorative scenarios may be used to cast light on what...... the future may bring. We develop a systematic approach to explorative scenario analysis and attempt to quantify aspects of resilience specifically for emergy assessment (EmA) of production systems. We group system inputs into five categories: (1) fossil fuels, their derivatives, metals and minerals, (2) on....... A comparison of two specific production systems showed that different expectations of the future lead to contrasting conclusions regarding prioritisation. We use the insight gained in the study to suggest venues for sustainable development under changing societal conditions....

  16. Information Technology Project Portfolio and Strategy Alignment Assessment Based on Data Envelopment Analysis

    Directory of Open Access Journals (Sweden)

    Marisa Analía Sánchez

    2012-11-01

    Full Text Available Recent researches have shown that companies face considerable difficulties in assessing the strategy value contribution of Information Technology (IT investments. One of the major obstacles to achieving strategy alignment is that organizations find extremely difficult to link and quantify the IT investments benefits with strategic goals. The aim of this paper is to define an approach to assess portfolio-strategy alignment. To this end a formal specification of Kaplan and Norton Strategy Map is developed utilizing Unified Modeling Language (UML. The approach uses the Strategy Map as a framework for defining the portfolio value contribution and Data Envelopment Analysis (DEA is used as the methodology for measuring efficiency of project portfolios.DOI:10.5585/gep.v3i2.66

  17. Automatic Speech Signal Analysis for Clinical Diagnosis and Assessment of Speech Disorders

    CERN Document Server

    Baghai-Ravary, Ladan

    2013-01-01

    Automatic Speech Signal Analysis for Clinical Diagnosis and Assessment of Speech Disorders provides a survey of methods designed to aid clinicians in the diagnosis and monitoring of speech disorders such as dysarthria and dyspraxia, with an emphasis on the signal processing techniques, statistical validity of the results presented in the literature, and the appropriateness of methods that do not require specialized equipment, rigorously controlled recording procedures or highly skilled personnel to interpret results. Such techniques offer the promise of a simple and cost-effective, yet objective, assessment of a range of medical conditions, which would be of great value to clinicians. The ideal scenario would begin with the collection of examples of the clients’ speech, either over the phone or using portable recording devices operated by non-specialist nursing staff. The recordings could then be analyzed initially to aid diagnosis of conditions, and subsequently to monitor the clients’ progress and res...

  18. LIFE CYCLE ASSESSMENT AND HAZARD ANALYSIS AND CRITICAL CONTROL POINTS TO THE PASTA PRODUCT

    Directory of Open Access Journals (Sweden)

    Yulexis Meneses Linares

    2016-10-01

    Full Text Available The objective of this work is to combine the Life Cycle Assessment (LCA and Hazard Analysis and Critical Control Points (HACCP methodologies for the determination of risks that the food production represents to the human health and the ecosystem. The environmental performance of the production of pastas in the “Marta Abreu” Pasta Factory of Cienfuegos is assessed, where the critical control points determined by the biological dangers (mushrooms and plagues and the physical dangers (wood, paper, thread and ferromagnetic particles were the raw materials: flour, semolina and its mixtures, and the disposition and extraction of them. Resources are the most affected damage category due to the consumption of fossil fuels.

  19. Automatized Assessment of Protective Group Reactivity: A Step Toward Big Reaction Data Analysis.

    Science.gov (United States)

    Lin, Arkadii I; Madzhidov, Timur I; Klimchuk, Olga; Nugmanov, Ramil I; Antipin, Igor S; Varnek, Alexandre

    2016-11-28

    We report a new method to assess protective groups (PGs) reactivity as a function of reaction conditions (catalyst, solvent) using raw reaction data. It is based on an intuitive similarity principle for chemical reactions: similar reactions proceed under similar conditions. Technically, reaction similarity can be assessed using the Condensed Graph of Reaction (CGR) approach representing an ensemble of reactants and products as a single molecular graph, i.e., as a pseudomolecule for which molecular descriptors or fingerprints can be calculated. CGR-based in-house tools were used to process data for 142,111 catalytic hydrogenation reactions extracted from the Reaxys database. Our results reveal some contradictions with famous Greene's Reactivity Charts based on manual expert analysis. Models developed in this study show high accuracy (ca. 90%) for predicting optimal experimental conditions of protective group deprotection.

  20. Does bioimpedance analysis or measurement of natriuretic peptides aid volume assessment in peritoneal dialysis patients?

    Science.gov (United States)

    Davenport, Andrew

    2013-01-01

    Cardiovascular mortality remains the commonest cause of death for peritoneal dialysis patients. As such, preventing persistent hypervolemia is important. On the other hand, hypovolemia may potentially risk episodes of acute kidney injury and loss of residual renal function, a major determinant of peritoneal dialysis technique survival. Bioimpedance has developed from a single-frequency research tool to a multi-frequency bioelectrical impedance analysis readily available in the clinic and capable of measuring extracellular, intracellular, and total body water. Similarly, natriuretic peptides released from the heart because of myocardial stretch and increased intracardiac volume have also been variously reported to be helpful in assessing volume status in peritoneal dialysis patients. The question then arises whether these newer technologies and biomarkers have supplanted the time-honored clinical assessment of hydration status or whether they are merely adjuncts that aid the experienced clinician.

  1. Addendum to the performance assessment analysis for low-level waste disposal in the 200 west area active burial grounds

    Energy Technology Data Exchange (ETDEWEB)

    Wood, M.I., Westinghouse Hanford

    1996-12-20

    An addendum was completed to the performance assessment (PA) analysis for the active 200 West Area low-level solid waste burial grounds. The addendum includes supplemental information developed during the review of the PA analysis, an ALARA analysis, a comparison of PA results with the Hanford Groundwater Protection Strategy, and a justification for the assumption of 500 year deterrence to the inadvertent intruder.

  2. Mercury analysis in hair: Comparability and quality assessment within the transnational COPHES/DEMOCOPHES project.

    Science.gov (United States)

    Esteban, Marta; Schindler, Birgit Karin; Jiménez, José Antonio; Koch, Holger Martin; Angerer, Jürgen; Rosado, Montserrat; Gómez, Silvia; Casteleyn, Ludwine; Kolossa-Gehring, Marike; Becker, Kerstin; Bloemen, Louis; Schoeters, Greet; Den Hond, Elly; Sepai, Ovnair; Exley, Karen; Horvat, Milena; Knudsen, Lisbeth E; Joas, Anke; Joas, Reinhard; Aerts, Dominique; Biot, Pierre; Borošová, Daniela; Davidson, Fred; Dumitrascu, Irina; Fischer, Marc E; Grander, Margaretha; Janasik, Beata; Jones, Kate; Kašparová, Lucie; Larssen, Thorjørn; Naray, Miklos; Nielsen, Flemming; Hohenblum, Philipp; Pinto, Rui; Pirard, Catherine; Plateel, Gregory; Tratnik, Janja Snoj; Wittsiepe, Jürgen; Castaño, Argelia

    2015-08-01

    Human biomonitoring (HBM) is an effective tool for assessing actual exposure to chemicals that takes into account all routes of intake. Although hair analysis is considered to be an optimal biomarker for assessing mercury exposure, the lack of harmonization as regards sampling and analytical procedures has often limited the comparison of data at national and international level. The European-funded projects COPHES and DEMOCOPHES developed and tested a harmonized European approach to Human Biomonitoring in response to the European Environment and Health Action Plan. Herein we describe the quality assurance program (QAP) for assessing mercury levels in hair samples from more than 1800 mother-child pairs recruited in 17 European countries. To ensure the comparability of the results, standard operating procedures (SOPs) for sampling and for mercury analysis were drafted and distributed to participating laboratories. Training sessions were organized for field workers and four external quality-assessment exercises (ICI/EQUAS), followed by the corresponding web conferences, were organized between March 2011 and February 2012. ICI/EQUAS used native hair samples at two mercury concentration ranges (0.20-0.71 and 0.80-1.63) per exercise. The results revealed relative standard deviations of 7.87-13.55% and 4.04-11.31% for the low and high mercury concentration ranges, respectively. A total of 16 out of 18 participating laboratories the QAP requirements and were allowed to analyze samples from the DEMOCOPHES pilot study. Web conferences after each ICI/EQUAS revealed this to be a new and effective tool for improving analytical performance and increasing capacity building. The procedure developed and tested in COPHES/DEMOCOPHES would be optimal for application on a global scale as regards implementation of the Minamata Convention on Mercury. Copyright © 2014 Elsevier Inc. All rights reserved.

  3. Transient Elastography for Assessment of Liver Fibrosis and Steatosis: An Evidence-Based Analysis

    Science.gov (United States)

    Brener, S

    2015-01-01

    Background Liver fibrosis is a sign of advanced liver disease and is often an indication for treatment. The current standard for diagnosing liver fibrosis and steatosis is biopsy, but noninvasive alternatives are available; one of the most common is transient elastography (FibroScan). Objectives The objective of this analysis was to assess the diagnostic accuracy and clinical utility of transient elastography alone for liver fibrosis and with controlled attenuation parameter (CAP) for steatosis in patients with hepatitis C virus, hepatitis B virus, nonalcoholic fatty liver disease, alcoholic liver disease, or cholestatic diseases. The analysis also aimed to compare the diagnostic accuracy of transient elastography with two alternative noninvasive technologies: FibroTest and acoustic force radiation impulse (ARFI). Data Sources Ovid MEDLINE, Ovid MEDLINE In-Process, Ovid Embase, and all EBM databases were searched for all studies published prior to October 2, 2014. Review Methods An overview of reviews was conducted using a systematic search and assessment approach. The results of the included systematic reviews were summarized, analyzed, and reported for outcomes related to diagnostic accuracy and clinical utility as a measure of impact on diagnoses, therapeutic decisions, and patient outcomes. Results Fourteen systematic reviews were included, summarizing more than 150 studies. The reviews demonstrated that transient elastography (with or without CAP) has good diagnostic accuracy compared to biopsy for the assessment of liver fibrosis and steatosis. Acoustic force radiation impulse and FibroTest were not superior to transient elastography. Limitations None of the included systematic reviews reported on the clinical utility of transient elastography. Conclusions Transient elastography (with or without CAP) offers a noninvasive alternative to biopsy for the assessment of liver fibrosis and steatosis, given its comparable diagnostic accuracy. PMID:26664664

  4. Guidebook in using Cost Benefit Analysis and strategic environmental assessment for environmental planning in China

    Energy Technology Data Exchange (ETDEWEB)

    NONE

    2011-07-01

    Environmental planning in China may benefit from greater use of Cost Benefit Analysis (CBA) and Strategic Environmental Assessment (SEA) methodologies. We provide guidance on using these methodologies. Part I and II show the principles behind the methodologies as well as their theoretical structure. Part III demonstrates the methodologies in action in a range of different good practice examples. The case studies and theoretical expositions are intended to teach by way of example as well as by understanding the principles, and to help planners use the methodologies as correctly as possible.(auth)

  5. A sensitivity analysis of a radiological assessment model for Arctic waters

    DEFF Research Database (Denmark)

    Nielsen, S.P.

    1998-01-01

    A model based on compartment analysis has been developed to simulate the dispersion of radionuclides in Arctic waters for an assessment of doses to man. The model predicts concentrations of radionuclides in the marine environment and doses to man from a range of exposure pathways. A parameter sen...... scavenging, water-sediment interaction, biological uptake, ice transport and fish migration. Two independent evaluations of the release of radioactivity from dumped nuclear waste in the Kara Sea have been used as source terms for the dose calculations....

  6. Risk Assessment of Distribution Network Based on Random set Theory and Sensitivity Analysis

    Science.gov (United States)

    Zhang, Sh; Bai, C. X.; Liang, J.; Jiao, L.; Hou, Z.; Liu, B. Zh

    2017-05-01

    Considering the complexity and uncertainty of operating information in distribution network, this paper introduces the use of random set for risk assessment. The proposed method is based on the operating conditions defined in the random set framework to obtain the upper and lower cumulative probability functions of risk indices. Moreover, the sensitivity of risk indices can effectually reflect information about system reliability and operating conditions, and by use of these information the bottlenecks that suppress system reliability can be found. The analysis about a typical radial distribution network shows that the proposed method is reasonable and effective.

  7. An Analysis of the Air War College’s Executive Health and Fitness Assessment Program.

    Science.gov (United States)

    1985-05-01

    1.6335 60 1.7782 77 1.8865 (SOURC;E: Naval Health Research Center, San Diego, Calitornia) 30 BIBLIOGRAPHY 1. Castelli, W. P. " Epedemiology of Coronary...1% AIR WAR COLLEGE RESEARCH REPORT No. AU-AWC-85-057 0AN ANALYSIS OF THE AIR WAR COLLEGE’S EXECUTIVE HEALTH AND FITNESS ASSESSMENT PROGRAM In I...COVERED 14. DATE OF REPORT (Yr., Mo., Day) 15. PAGE COUNT Research FROM TO May 1985 132 16. SUPPLEMENTARY NOTATION S COSATI CODES 18. SUBJECT TERMS

  8. A Systematic Approach to Explorative Scenario Analysis in Emergy Assessment with Emphasis on Resilience

    DEFF Research Database (Denmark)

    Kamp, Andreas; Østergård, Hanne

    2016-01-01

    the future may bring. We develop a systematic approach to explorative scenario analysis and attempt to quantify aspects of resilience specifically for emergy assessment (EmA) of production systems. We group system inputs into five categories: (1) fossil fuels, their derivatives, metals and minerals, (2) on...... by corresponding narratives. We analyse the aggregated effect on UEVs of these scenarios for production systems that differ with respect to how the emergy flow is distributed among the five input categories. We find that for most production systems, scenario conditions significantly affect the UEV. The production...

  9. Semantic Pattern Analysis for Verbal Fluency Based Assessment of Neurological Disorders

    Energy Technology Data Exchange (ETDEWEB)

    Sukumar, Sreenivas R [ORNL; Ainsworth, Keela C [ORNL; Brown, Tyler C [ORNL

    2014-01-01

    In this paper, we present preliminary results of semantic pattern analysis of verbal fluency tests used for assessing cognitive psychological and neuropsychological disorders. We posit that recent advances in semantic reasoning and artificial intelligence can be combined to create a standardized computer-aided diagnosis tool to automatically evaluate and interpret verbal fluency tests. Towards that goal, we derive novel semantic similarity (phonetic, phonemic and conceptual) metrics and present the predictive capability of these metrics on a de-identified dataset of participants with and without neurological disorders.

  10. Using computerized text analysis to assess communication within an Italian type 1 diabetes Facebook group

    Directory of Open Access Journals (Sweden)

    Alda Troncone

    2015-11-01

    Full Text Available The purpose of this study was to assess messages posted by mothers of children with type 1 diabetes in the Italian Facebook group “Mamme e diabete” using computerized text analysis. The data suggest that these mothers use online discussion boards as a place to seek and provide information to better manage the disease’s daily demands—especially those tasks linked to insulin correction and administration, control of food intake, and bureaucratic duties, as well as to seek and give encouragement and to share experiences regarding diabetes and related impact on their life. The implications of these findings for the management of diabetes are discussed.

  11. Human Error Assessment and Reduction Technique (HEART) and Human Factor Analysis and Classification System (HFACS)

    Science.gov (United States)

    Alexander, Tiffaney Miller

    2017-01-01

    Research results have shown that more than half of aviation, aerospace and aeronautics mishaps incidents are attributed to human error. As a part of Safety within space exploration ground processing operations, the identification and/or classification of underlying contributors and causes of human error must be identified, in order to manage human error. This research provides a framework and methodology using the Human Error Assessment and Reduction Technique (HEART) and Human Factor Analysis and Classification System (HFACS), as an analysis tool to identify contributing factors, their impact on human error events, and predict the Human Error probabilities (HEPs) of future occurrences. This research methodology was applied (retrospectively) to six (6) NASA ground processing operations scenarios and thirty (30) years of Launch Vehicle related mishap data. This modifiable framework can be used and followed by other space and similar complex operations.

  12. Human Error Assessment and Reduction Technique (HEART) and Human Factor Analysis and Classification System (HFACS)

    Science.gov (United States)

    Alexander, Tiffaney Miller

    2017-01-01

    Research results have shown that more than half of aviation, aerospace and aeronautics mishaps incidents are attributed to human error. As a part of Quality within space exploration ground processing operations, the identification and or classification of underlying contributors and causes of human error must be identified, in order to manage human error.This presentation will provide a framework and methodology using the Human Error Assessment and Reduction Technique (HEART) and Human Factor Analysis and Classification System (HFACS), as an analysis tool to identify contributing factors, their impact on human error events, and predict the Human Error probabilities (HEPs) of future occurrences. This research methodology was applied (retrospectively) to six (6) NASA ground processing operations scenarios and thirty (30) years of Launch Vehicle related mishap data. This modifiable framework can be used and followed by other space and similar complex operations.

  13. Statistical analysis for assessing shallow-landslide susceptibility in South Tyrol (south-eastern Alps, Italy)

    Science.gov (United States)

    Piacentini, Daniela; Troiani, Francesco; Soldati, Mauro; Notarnicola, Claudia; Savelli, Daniele; Schneiderbauer, Stefan; Strada, Claudia

    2012-05-01

    This paper conducts a statistical analysis to determine shallow-landslide susceptibility in an approximately 7500-km2 region of the south-eastern Alps (South Tyrol, Italy). The study applies the weight of evidence (WofE) method, which is useful in determining landslide susceptibility in large areas with complex geological and geomorphological settings. The statistical analysis and landslide susceptibility mapping are based on 882 past landslides, three geometric/topographic factors and two anthropogenic factors, which are the most relevant landslide predisposing factors. The quality of the proposed model, particularly the fitting performance, was assessed; the landslide database was divided into a training set to obtain the model and a validation set to estimate the model quality. The results show that the developed susceptibility model predicts an acceptable percentage (75%) of landslides. Therefore, the model can be useful and reliable for land planners and decision makers also due to its cost-effectiveness ratio.

  14. Natural resource damage assessment for the Hebei Spirit oil spill: An application of Habitat Equivalency Analysis.

    Science.gov (United States)

    Kim, Tae-Goun; Opaluch, James; Moon, Daniel Seong-Hyeok; Petrolia, Daniel R

    2017-08-15

    Sixteen oil spills occurred in Korea between 1995 and 2010, including the Hebei Spirit oil spill (HSOS) in 2007, the largest (77,857 barrels) in Korean history. Yet compensation for environmental damages has never been claimed under the International Oil Pollution Compensation Fund for any of these accidents, because there is no adequate natural resource damage assessment (NRDA) procedure and there are no internationally-admissible economic quantification methods established in Korea. The objective of this study is to propose a methodology to overcome these shortcomings. We propose the use of Habitat Equivalency Analysis, which has dominated the US NRDA process for oil spills, and apply it to the HSOS as a case study. A Base Case analysis estimates the compensatory costs of fishery habitat damages (34,703.5ha) with a 10-year recovery rate at $119.4 million. We also conduct sensitivity analyses under several alternative assumptions. Copyright © 2017 Elsevier Ltd. All rights reserved.

  15. Neo-Deterministic and Probabilistic Seismic Hazard Assessments: a Comparative Analysis

    Science.gov (United States)

    Peresan, Antonella; Magrin, Andrea; Nekrasova, Anastasia; Kossobokov, Vladimir; Panza, Giuliano F.

    2016-04-01

    Objective testing is the key issue towards any reliable seismic hazard assessment (SHA). Different earthquake hazard maps must demonstrate their capability in anticipating ground shaking from future strong earthquakes before an appropriate use for different purposes - such as engineering design, insurance, and emergency management. Quantitative assessment of maps performances is an essential step also in scientific process of their revision and possible improvement. Cross-checking of probabilistic models with available observations and independent physics based models is recognized as major validation procedure. The existing maps from the classical probabilistic seismic hazard analysis (PSHA), as well as those from the neo-deterministic analysis (NDSHA), which have been already developed for several regions worldwide (including Italy, India and North Africa), are considered to exemplify the possibilities of the cross-comparative analysis in spotting out limits and advantages of different methods. Where the data permit, a comparative analysis versus the documented seismic activity observed in reality is carried out, showing how available observations about past earthquakes can contribute to assess performances of the different methods. Neo-deterministic refers to a scenario-based approach, which allows for consideration of a wide range of possible earthquake sources as the starting point for scenarios constructed via full waveforms modeling. The method does not make use of empirical attenuation models (i.e. Ground Motion Prediction Equations, GMPE) and naturally supplies realistic time series of ground shaking (i.e. complete synthetic seismograms), readily applicable to complete engineering analysis and other mitigation actions. The standard NDSHA maps provide reliable envelope estimates of maximum seismic ground motion from a wide set of possible scenario earthquakes, including the largest deterministically or historically defined credible earthquake. In addition

  16. Questions to be addressed in the next Yucca Mountain performance assessment analysis

    Energy Technology Data Exchange (ETDEWEB)

    Barnard, R.W.; Barr, G.E.

    1995-12-31

    The next total-system performance-assessment (TSPA) analyses are designed to aid DOE in performing an ``investment analysis`` for Yucca Mountain. This TSPA must try to bound the uncertainties for several issues that will contribute to the decision whether the US should proceed with the development of a nuclear-waste repository at Yucca Mountain. Because site-characterization experiments and data collection will continue for the foreseeable future, the next TSPA (called TSPA-IA) will again only be able to use partially developed models and partial data sets. In contrast to previous analyses however, TSPA-IA must address more specific questions to be of assistance to the investment-analysis deliberations.

  17. Method of assessing a lipid-related health risk based on ion mobility analysis of lipoproteins

    Energy Technology Data Exchange (ETDEWEB)

    Benner, W. Henry (Danville, CA); Krauss, Ronald M. (Berkeley, CA); Blanche, Patricia J. (Berkeley, CA)

    2010-12-14

    A medical diagnostic method and instrumentation system for analyzing noncovalently bonded agglomerated biological particles is described. The method and system comprises: a method of preparation for the biological particles; an electrospray generator; an alpha particle radiation source; a differential mobility analyzer; a particle counter; and data acquisition and analysis means. The medical device is useful for the assessment of human diseases, such as cardiac disease risk and hyperlipidemia, by rapid quantitative analysis of lipoprotein fraction densities. Initially, purification procedures are described to reduce an initial blood sample to an analytical input to the instrument. The measured sizes from the analytical sample are correlated with densities, resulting in a spectrum of lipoprotein densities. The lipoprotein density distribution can then be used to characterize cardiac and other lipid-related health risks.

  18. Human Error Assessment and Reduction Technique (HEART) and Human Factor Analysis and Classification System (HFACS)

    Science.gov (United States)

    Alexander, Tiffaney Miller

    2017-01-01

    Research results have shown that more than half of aviation, aerospace and aeronautics mishaps/incidents are attributed to human error. As a part of Safety within space exploration ground processing operations, the identification and/or classification of underlying contributors and causes of human error must be identified, in order to manage human error. This research provides a framework and methodology using the Human Error Assessment and Reduction Technique (HEART) and Human Factor Analysis and Classification System (HFACS), as an analysis tool to identify contributing factors, their impact on human error events, and predict the Human Error probabilities (HEPs) of future occurrences. This research methodology was applied (retrospectively) to six (6) NASA ground processing operations scenarios and thirty (30) years of Launch Vehicle related mishap data. This modifiable framework can be used and followed by other space and similar complex operations.

  19. Evolution and Implementation of the NASA Robotic Conjunction Assessment Risk Analysis Concept of Operations

    Science.gov (United States)

    Newman, Lauri K.; Frigm, Ryan C.; Duncan, Matthew G.; Hejduk, Matthew D.

    2014-01-01

    Reacting to potential on-orbit collision risk in an operational environment requires timely and accurate communication and exchange of data, information, and analysis to ensure informed decision-making for safety of flight and responsible use of the shared space environment. To accomplish this mission, it is imperative that all stakeholders effectively manage resources: devoting necessary and potentially intensive resource commitment to responding to high-risk conjunction events and preventing unnecessary expenditure of resources on events of low collision risk. After 10 years of operational experience, the NASA Robotic Conjunction Assessment Risk Analysis (CARA) is modifying its Concept of Operations (CONOPS) to ensure this alignment of collision risk and resource management. This evolution manifests itself in the approach to characterizing, reporting, and refining of collision risk. Implementation of this updated CONOPS is expected to have a demonstrated improvement on the efficacy of JSpOC, CARA, and owner/operator resources.

  20. Y-12 site-specific earthquake response analysis and soil liquefaction assessment

    Energy Technology Data Exchange (ETDEWEB)

    Ahmed, S.B.; Hunt, R.J.; Manrod, W.E. III

    1995-09-29

    A site-specific earthquake response analysis and soil liquefaction assessment were performed for the Oak Ridge Y-12 Plant. The main purpose of these studies was to use the results of the analyses for evaluating the safety of the performance category -1, -2, and -3 facilities against the natural phenomena seismic hazards. Earthquake response was determined for seven (7), one dimensional soil columns (Fig. 12) using two horizontal components of the PC-3 design basis 2000-year seismic event. The computer program SHAKE 91 (Ref. 7) was used to calculate the absolute response accelerations on top of ground (soil/weathered shale) and rock outcrop. The SHAKE program has been validated for horizontal response calculations at periods less than 2.0 second at several sites and consequently is widely accepted in the geotechnical earthquake engineering area for site response analysis.

  1. Using venous blood gas analysis in the assessment of COPD exacerbations: a prospective cohort study.

    Science.gov (United States)

    McKeever, Tricia M; Hearson, Glenn; Housley, Gemma; Reynolds, Catherine; Kinnear, William; Harrison, Tim W; Kelly, Anne-Maree; Shaw, Dominick E

    2016-03-01

    Identifying acute hypercapnic respiratory failure is crucial in the initial management of acute exacerbations of COPD. Guidelines recommend obtaining arterial blood samples but these are more difficult to obtain than venous. We assessed whether blood gas values derived from venous blood could replace arterial at initial assessment. Patients requiring hospital treatment for an exacerbation of COPD had paired arterial and venous samples taken. Bland-Altman analyses were performed to assess agreement between arterial and venous pH, CO2 and HCO3-. The relationship between SpO2 and SaO2 was assessed. The number of attempts and pain scores for each sample were measured. 234 patients were studied. There was good agreement between arterial and venous measures of pH and HC)3- (mean difference 0.03 and -0.04, limits of agreement -0.05 to 0.11 and -2.90 to 2.82, respectively), and between SaO2 and SpO2 (in patients with an SpO2 of >80%). Arterial sampling required more attempts and was more painful than venous (mean pain score 4 (IQR 2-5) and 1 (IQR 0-2), respectively, pblood, and between pulse oximetry and arterial blood gas oxygen saturations. These agreements could allow the initial assessment of COPD exacerbations to be based on venous blood gas analysis and pulse oximetry, simplifying the care pathway and improving the patient experience. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/

  2. Trinity amputation and prosthesis experience scales: a psychometric assessment using classical test theory and rasch analysis.

    Science.gov (United States)

    Gallagher, Pamela; Franchignoni, F; Giordano, A; MacLachlan, M

    2010-06-01

    To perform a detailed psychometric analysis using both classical test theory and Rasch analysis of the three main scales of the Trinity Amputation and Prosthesis Experience Scales (TAPES) in people with a lower-limb amputation. A sample of 498 persons who were prosthesis users with a lower-limb amputation was retrospectively studied, pooled from a number of studies undertaken across the United Kingdom and Ireland in the past decade in which the TAPES had been completed as part of a postal survey. Both factor analysis techniques and Rasch analysis were performed on TAPES data. Dimensionality, item fit to the model, response category performance, and internal construct validity were assessed. Category collapsing and item removal were considered to improve the questionnaire. The analyses suggested to restructure the TAPES as follows: (a) three psychosocial adjustment subscales with a four-point rating scale (and a reworded item); (b) an activity restriction scale based on ten items with their original three-point rating scale; and (c) two satisfaction with the prosthesis subscales using a three-point rating scale. All scales and subscales showed acceptable internal consistency and ability to define a distinct hierarchy of persons along the measured construct. This study empirically identified a revised version of the TAPES (TAPES-R) with a simplified general structure and psychometrically suitable for assessing the complex experience of amputation and adjustment to a lower-limb prosthesis. Additional studies are needed to confirm and further explore its measurement properties in other samples, thereby adding clinical validity to the instrument.

  3. Review and analysis of parameters for assessing transport of environmentally released radionuclides through agriculture

    Energy Technology Data Exchange (ETDEWEB)

    Baes, C.F. III; Sharp, R.D.; Sjoreen, A.L.; Shor, R.W.

    1984-09-01

    Most of the default parameters incorporated into the TERRA computer code are documented including a literature review and systematic analysis of element-specific transfer parameters B/sub v/, B/sub r/, F/sub m/, F/sub f/, and K/sub d/. This review and analysis suggests default values which are consistent with the modeling approaches taken in TERRA and may be acceptable for most assessment applications of the computer code. However, particular applications of the code and additional analysis of elemental transport may require alternative default values. Use of the values reported herein in other computer codes simulating terrestrial transport is not advised without careful interpretation of the limitations and scope these analyses. An approach to determination of vegetation-specific interception fractions is also discussed. The limitations of this approach are many, and its use indicates the need for analysis of deposition, interception, and weathering processes. Judgement must be exercised in interpretation of plant surface concentrations generated. Finally, the location-specific agricultural, climatological, and population parameters in the default SITE data base documented. These parameters are intended as alternatives to average values currently used. Indeed, areas in the United States where intensive crop, milk, or beef production occurs will be reflected in the parameter values as will areas where little agricultural activity occurs. However, the original information sources contained some small error and the interpolation and conversion methods used will add more. Parameters used in TERRA not discussed herein are discussed in the companion report to this one - ORNL-5785. In the companion report the models employed in and the coding of TERRA are discussed. These reports together provide documentation of the TERRA code and its use in assessments. 96 references, 78 figures, 21 tables.

  4. GenomeLandscaper: Landscape analysis of genome-fingerprints maps assessing chromosome architecture.

    Science.gov (United States)

    Ai, Hannan; Ai, Yuncan; Meng, Fanmei

    2018-01-18

    Assessing correctness of an assembled chromosome architecture is a central challenge. We create a geometric analysis method (called GenomeLandscaper) to conduct landscape analysis of genome-fingerprints maps (GFM), trace large-scale repetitive regions, and assess their impacts on the global architectures of assembled chromosomes. We develop an alignment-free method for phylogenetics analysis. The human Y chromosomes (GRCh.chrY, HuRef.chrY and YH.chrY) are analysed as a proof-of-concept study. We construct a galaxy of genome-fingerprints maps (GGFM) for them, and a landscape compatibility among relatives is observed. But a long sharp straight line on the GGFM breaks such a landscape compatibility, distinguishing GRCh38p1.chrY (and throughout GRCh38p7.chrY) from GRCh37p13.chrY, HuRef.chrY and YH.chrY. We delete a 1.30-Mbp target segment to rescue the landscape compatibility, matching the antecedent GRCh37p13.chrY. We re-locate it into the modelled centromeric and pericentromeric region of GRCh38p10.chrY, matching a gap placeholder of GRCh37p13.chrY. We decompose it into sub-constituents (such as BACs, interspersed repeats, and tandem repeats) and trace their homologues by phylogenetics analysis. We elucidate that most examined tandem repeats are of reasonable quality, but the BAC-sized repeats, 173U1020C (176.46 Kbp) and 5U41068C (205.34 Kbp), are likely over-repeated. These results offer unique insights into the centromeric and pericentromeric regions of the human Y chromosomes.

  5. Validation of bioelectrical impedance analysis for assessing dry weight of dialysis patients in Pakistan

    Directory of Open Access Journals (Sweden)

    Huma Asmat

    2017-01-01

    Full Text Available Accurate dry weight estimation (DW to achieve euvolemia is one of the key objectives of hemodialysis (HD. While conventionally DW is estimated by clinical examination, bioelectrical impedance analysis (BIA has been proposed as an objective method to determine DW and has been tested extensively in the Western population. We aim to validate BIA for determining DW in a Pakistani population against the conventional clinician’s method. This is a single-center validation study conducted at two outpatient HD units of Aga Khan Hospital, Karachi. One hundred and forty-eight DW readings of patients who were on maintenance HD were taken both by BIA technology and by clinical assessment. The clinician was blinded to readings obtained by BIA. Data analysis was carried out using the Statistical Package for Social Sciences (SPSS version 19. Median age of patients was 63 years (range 12–89. Nearly 54.1% of the samples were female (n = 80. Spearman’s correlation between the clinician’s estimate and BIA-derived DW showed a correlation coefficient of 0.982, which was statistically significant (P <0.001. This association remained significant when stratified analysis was carried out by dividing the sample into subgroups according to age, gender, body mass index, and total body water content. Inter-rater reliability analysis using the kappa statistics showed almost perfect agreement between the two methods, κ = 0.929 (95% confidence interval, 0.878–0.980, P <0.001. BIA has been validated as a tool for DW assessment of HD patients in Pakistan in comparison to clinical method.

  6. Validation of bioelectrical impedance analysis for assessing dry weight of dialysis patients in Pakistan.

    Science.gov (United States)

    Asmat, Huma; Iqbal, Romaina; Sharif, Fatima; Mahmood, Aamer; Abbas, Aamir; Kashif, Waqar

    2017-01-01

    Accurate dry weight estimation (DW) to achieve euvolemia is one of the key objectives of hemodialysis (HD). While conventionally DW is estimated by clinical examination, bioelectrical impedance analysis (BIA) has been proposed as an objective method to determine DW and has been tested extensively in the Western population. We aim to validate BIA for determining DW in a Pakistani population against the conventional clinician's method. This is a single-center validation study conducted at two outpatient HD units of Aga Khan Hospital, Karachi. One hundred and forty-eight DW readings of patients who were on maintenance HD were taken both by BIA technology and by clinical assessment. The clinician was blinded to readings obtained by BIA. Data analysis was carried out using the Statistical Package for Social Sciences (SPSS) version 19. Median age of patients was 63 years (range 12-89). Nearly 54.1% of the samples were female (n = 80). Spearman's correlation between the clinician's estimate and BIA-derived DW showed a correlation coefficient of 0.982, which was statistically significant (P <0.001). This association remained significant when stratified analysis was carried out by dividing the sample into subgroups according to age, gender, body mass index, and total body water content. Inter-rater reliability analysis using the kappa statistics showed almost perfect agreement between the two methods, κ = 0.929 (95% confidence interval, 0.878-0.980, P <0.001). BIA has been validated as a tool for DW assessment of HD patients in Pakistan in comparison to clinical method.

  7. Assessment of the dimensionality of the Wijma delivery expectancy/experience questionnaire using factor analysis and Rasch analysis.

    Science.gov (United States)

    Pallant, J F; Haines, H M; Green, P; Toohill, J; Gamble, J; Creedy, D K; Fenwick, J

    2016-11-21

    Fear of childbirth has negative consequences for a woman's physical and emotional wellbeing. The most commonly used measurement tool for childbirth fear is the Wijma Delivery Expectancy Questionnaire (WDEQ-A). Although originally conceptualized as unidimensional, subsequent investigations have suggested it is multidimensional. This study aimed to undertake a detailed psychometric assessment of the WDEQ-A; exploring the dimensionality and identifying possible subscales that may have clinical and research utility. WDEQ-A was administered to a sample of 1410 Australian women in mid-pregnancy. The dimensionality of WDEQ-A was explored using exploratory (EFA) and confirmatory factor analysis (CFA), and Rasch analysis. EFA identified a four factor solution. CFA failed to support the unidimensional structure of the original WDEQ-A, but confirmed the four factor solution identified by EFA. Rasch analysis was used to refine the four subscales (Negative emotions: five items; Lack of positive emotions: five items; Social isolation: four items; Moment of birth: three items). Each WDEQ-A Revised subscale showed good fit to the Rasch model and adequate internal consistency reliability. The correlation between Negative emotions and Lack of positive emotions was strong, however Moment of birth and Social isolation showed much lower intercorrelations, suggesting they should not be added to create a total score. This study supports the findings of other investigations that suggest the WDEQ-A is multidimensional and should not be used in its original form. The WDEQ-A Revised may provide researchers with a more refined, psychometrically sound tool to explore the differential impact of aspects of childbirth fear.

  8. Application of linear discriminant analysis to the nasometric assessment of resonance disorders: a pilot study.

    Science.gov (United States)

    de Boer, Gillian; Bressmann, Tim

    2015-03-01

    Objective : Nasalance scores have traditionally been used to assess hypernasality. However, resonance disorders are often complex, and hypernasality and nasal obstruction may co-occur in patients with cleft palate. In this study, normal speakers simulated different resonance disorders, and linear discriminant analysis was used to create a tentative diagnostic formula based on nasalance scores for nonnasal and nasal speech stimuli. Materials and Methods : Eleven female participants were recorded with the Nasometer 6450 while reading nonnasal and nasal speech stimuli. Nasalance measurements were taken of their normal resonance and their simulations of hyponasal, hypernasal, and mixed resonance. Results : A repeated-measures analysis of variance revealed a resonance condition-stimuli interaction effect (P < .001). A linear discriminant analysis of the participants' nasalance scores led to formulas correctly classifying 64.4% of the resonance conditions. When the hyponasal and mixed resonance conditions with obstruction of the less patent nostril were removed from the analysis, the resultant formulas correctly classified 88.6% of the resonance conditions. Conclusion : The simulations produced distinctive nasalance scores, enabling the creation of formulas that predicted resonance condition above chance level. The preliminary results demonstrate the potential of this approach for the diagnosis of resonance disorders.

  9. A wearable textile for respiratory monitoring: Feasibility assessment and analysis of sensors position on system response.

    Science.gov (United States)

    Lo Presti, D; Massaroni, C; Saccomandi, P; Caponero, M A; Formica, D; Schena, E

    2017-07-01

    The interest on wearable textiles to monitor vital signs is growing in the research field and clinical scenario related to the increasing demands of long-term monitoring. Despite several smart textile-based solutions have been proposed for assessing the respiratory status, only a limited number of devices allow the respiratory monitoring in a harsh environment or in different positions of the human body. In this paper, we investigated the performances of a smart textile for respiratory rate monitoring characterized by 12 fiber optic sensors (i.e., fiber Bragg grating) placed on specific landmarks for compartmental analysis of the chest wall movements during quiet breathing. We focused on the analysis of the influence of sensor position on both peak-to-peak amplitude of sensors output and accuracy of respiratory rate measurements. This analysis was performed on two participants, who wore the textile in two positions (i.e., standing and supine). Bland-Altman analysis on respiratory rate showed promising results (better than 0.3 breaths per minute). Referring to the peak-to-peak output amplitude, the abdomen compartment showed the highest excursions in both the enrolled participants and positions. Our findings open up new approaches to design and develop smart textile for respiratory rate monitoring.

  10. Assessment of (1)H NMR-based metabolomics analysis for normalization of urinary metals against creatinine.

    Science.gov (United States)

    Cassiède, Marc; Nair, Sindhu; Dueck, Meghan; Mino, James; McKay, Ryan; Mercier, Pascal; Quémerais, Bernadette; Lacy, Paige

    2017-01-01

    Proton nuclear magnetic resonance ((1)H NMR, or NMR) spectroscopy and inductively coupled plasma-mass spectrometry (ICP-MS) are commonly used for metabolomics and metal analysis in urine samples. However, creatinine quantification by NMR for the purpose of normalization of urinary metals has not been validated. We assessed the validity of using NMR analysis for creatinine quantification in human urine samples in order to allow normalization of urinary metal concentrations. NMR and ICP-MS techniques were used to measure metabolite and metal concentrations in urine samples from 10 healthy subjects. For metabolite analysis, two magnetic field strengths (600 and 700MHz) were utilized. In addition, creatinine concentrations were determined by using the Jaffe method. Creatinine levels were strongly correlated (R(2)=0.99) between NMR and Jaffe methods. The NMR spectra were deconvoluted with a target database containing 151 metabolites that are present in urine. A total of 50 metabolites showed good correlation (R(2)=0.7-1.0) at 600 and 700MHz. Metal concentrations determined after NMR-measured creatinine normalization were comparable to previous reports. NMR analysis provided robust urinary creatinine quantification, and was sufficient for normalization of urinary metal concentrations. We found that NMR-measured creatinine-normalized urinary metal concentrations in our control subjects were similar to general population levels in Canada and the United Kingdom. Copyright © 2016 Elsevier B.V. All rights reserved.

  11. Procedure for conducting probabilistic safety assessment: level 1 full power internal event analysis

    Energy Technology Data Exchange (ETDEWEB)

    Jung, Won Dae; Lee, Y. H.; Hwang, M. J. [and others

    2003-07-01

    This report provides guidance on conducting a Level I PSA for internal events in NPPs, which is based on the method and procedure that was used in the PSA for the design of Korea Standard Nuclear Plants (KSNPs). Level I PSA is to delineate the accident sequences leading to core damage and to estimate their frequencies. It has been directly used for assessing and modifying the system safety and reliability as a key and base part of PSA. Also, Level I PSA provides insights into design weakness and into ways of preventing core damage, which in most cases is the precursor to accidents leading to major accidents. So Level I PSA has been used as the essential technical bases for risk-informed application in NPPs. The report consists six major procedural steps for Level I PSA; familiarization of plant, initiating event analysis, event tree analysis, system fault tree analysis, reliability data analysis, and accident sequence quantification. The report is intended to assist technical persons performing Level I PSA for NPPs. A particular aim is to promote a standardized framework, terminology and form of documentation for PSAs. On the other hand, this report would be useful for the managers or regulatory persons related to risk-informed regulation, and also for conducting PSA for other industries.

  12. Assessment of Student Skills for Critiquing Published Primary Scientific Literature Using a Primary Trait Analysis Scale

    Directory of Open Access Journals (Sweden)

    Manuel F. Varela

    2009-12-01

    Full Text Available Instructor evaluation of progressive student skills in the analysis of primary literature is critical for the development of these skills in young scientists. Students in a senior or graduate-level one-semester course in Immunology at a Masters-level comprehensive university were assessed for abilities (primary traits to recognize and evaluate the following elements of a scientific paper: Hypothesis and Rationale, Significance, Methods, Results, Critical Thinking and Analysis, and Conclusions. We tested the hypotheses that average recognition scores vary among elements and that scores change with time differently by trait. Recognition scores (scaled 1 to 5, and differences in scores were analyzed using analysis of variance (ANOVA, regression, and analysis of covariance (ANCOVA (n = 10 papers over 103 days. By multiple comparisons testing, we found that recognition scores statistically fell into two groups: high scores (for Hypothesis and Rationale, Significance, Methods, and Conclusions and low scores (for Results and Critical Thinking and Analysis. Recognition scores only significantly changed with time (increased for Hypothesis and Rationale and Results. ANCOVA showed that changes in recognition scores for these elements were not significantly different in slope (F1,16 = 0.254, P = 0.621 but the Results trait was significantly lower in elevation (F1,17 = 12.456, P = 0.003. Thus, students improved with similar trajectories, but starting and ending with lower Results scores. We conclude that students have greatest difficulty evaluating Results and critically evaluating scientific validity. Our findings show extant student skills, and the significant increase in some traits shows learning. This study demonstrates that students start with variable recognition skills and that student skills may be learned at differential rates. Faculty can use these findings or the primary trait analysis scoring scale to focus on specific paper elements for which

  13. Safety assessment technology on the free drop impact and puncture analysis of the cask for radioactive material transport

    Energy Technology Data Exchange (ETDEWEB)

    Lee, Dew Hey [Korea Institute of Nuclear Safety, Taejon (Korea, Republic of); Lee, Young Shin; Ryu, Chung Hyun; Kim, Hyun Su; Lee, Ho Chul; Hong, Song Jin; Choi, Young Jin; Lee, Jae Hyung; Na, Jae Yun [Chungnam National Univ., Taejon (Korea, Republic of)

    2001-03-15

    In this study, the regulatory condition and analysis condition is analyzed for the free drop and puncture impact analysis to develop the safety assessment technology. Impact analysis is performed with finite element method which is one of the many analysis methods of the shipping cask. LS-DYNA3D and ABAQUS is suitable for the free drop and the puncture impact analysis of the shipping cask. For the analysis model, the KSC-4 that is the shipping cask to transport spent nuclear fuel is investigated. The results of both LS-DYNA3D and ABAQUS is completely corresponded. And The integrity of the shipping cask is verified. Using this study, the reliable safety assessment technology is supplied to the staff. The efficient and reliable regulatory tasks is performed using the standard safety assessment technology.

  14. Correlation Between Resonance Frequency Analysis and Bone Quality Assessments at Dental Implant Recipient Sites.

    Science.gov (United States)

    Fu, Min-Wen; Fu, Earl; Lin, Fu-Gong; Chang, Wei-Jeng; Hsieh, Yao-Dung; Shen, E-Chin

    To evaluate whether primary implant stability could be used to predict bone quality, the association between the implant stability quotient (ISQ) value and the bone type at the implant site was evaluated. Ninety-five implant sites in 50 patients were included. Bone type (categorized by Lekholm and Zarb) at the implant site was initially assessed using presurgical dental radiography. During the preparation of the implant site, a bone core specimen was carefully obtained. The bone type was assessed by tactile sensation during the drilling operation, according to the Misch criteria. The primary stability of the inserted implant was evaluated by resonance frequency analysis (RFA). The ISQ value was recorded. The bone core specimen was then examined by stereomicroscopy or microcomputed tomography (micro-CT), and the bone type was determined by the surface characteristics of the specimen, based on Lekholm and Zarb classification. Agreement between the bone quality assessed by the four methods (ie, presurgical radiography, tactile sensation, stereomicroscopy, and micro-CT) was tested by Cohen's kappa statistics, whereas the association between the ISQ value and the bone type was evaluated by the generalized linear regression model. The mean ISQ score was 72.6, and the score was significantly influenced by the maxillary or mandibular arch (P = .001). The bone type at the implant sites varied according to the assessment method. However, a significant influence of the arch was repeatedly noted when using radiography or tactile sensation. Among the four bone-quality assessment methods, a weak agreement existed only between stereomicroscopy and micro-CT, especially in the maxilla (κ = 0.469). A negative association between the ISQ value and the bone type assessed by stereomicroscopy or by micro-CT was significant in the maxilla, but not in the mandible, after adjustments for sex, age, and right/left side (P = .013 and P = .027 for stereomicroscopy and micro-CT, respectively

  15. On sustainability assessment of technical systems. Experience from systems analysis with the ORWARE and Ecoeffect tools

    Energy Technology Data Exchange (ETDEWEB)

    Assefa, Getachew [Royal Inst. of Technology, Stockholm (Sweden). Dept. of Chemical Engineering

    2006-06-15

    Engineering research and development work is undergoing a reorientation from focusing on specific parts of different systems to a broader perspective of systems level, albeit at a slower pace. This reorientation should be further developed and enhanced with the aim of organizing and structuring our technical systems in meeting sustainability requirements in face of global ecological threats that have far-reaching social and economic implications, which can no longer be captured using conventional approach of research. Until a list of universally acceptable, clear, and measurable indicators of sustainable development is developed, the work with sustainability metrics should continue to evolve as a relative measure of ecological, economic, and social performance of human activities in general, and technical systems in particular. This work can be done by comparing the relative performance of alternative technologies of providing the same well-defined function or service; or by characterizing technologies that enjoy different levels of societal priorities using relevant performance indicators. In both cases, concepts and methods of industrial ecology play a vital role. This thesis is about the development and application of a systematic approach for the assessment of the performance of technical systems from the perspective of systems analysis, sustainability, sustainability assessment, and industrial ecology. The systematic approach developed and characterized in this thesis advocates for a simultaneous assessment of the ecological, economic, and social dimensions of performance of technologies in avoiding sub-optimization and problem shifting between dimensions. It gives a holistic picture by taking a life cycle perspective of all important aspects. The systematic assessment of technical systems provides an even-handed assessment resulting in a cumulative knowledge. A modular structure of the approach makes it flexible enough in terms of comparing a number of

  16. Shaking up the Cost Benefit Analysis process: Issues and directions for improvement when assessing integrated spatial transport plans through a cost benefit analysis

    NARCIS (Netherlands)

    Beukers, E.

    2015-01-01

    This dissertation focusses on the Cost Benefit Analysis (CBA) processes when assessing integrated spatial transport plans, using the Netherlands as a case in point. It answers the following research question: What process issues occur when assessing integrated spatial transport plans through a CBA

  17. Nanoparticle analysis and characterization methodologies in environmental risk assessment of engineered nanoparticles.

    Science.gov (United States)

    Hassellöv, Martin; Readman, James W; Ranville, James F; Tiede, Karen

    2008-07-01

    Environmental risk assessments of engineered nanoparticles require thorough characterization of nanoparticles and their aggregates. Furthermore, quantitative analytical methods are required to determine environmental concentrations and enable both effect and exposure assessments. Many methods still need optimization and development, especially for new types of nanoparticles in water, but extensive experience can be gained from the fields of environmental chemistry of natural nanomaterials and from fundamental colloid chemistry. This review briefly describes most methods that are being exploited in nanoecotoxicology for analysis and characterization of nanomaterials. Methodological aspects are discussed in relation to the fields of nanometrology, particle size analysis and analytical chemistry. Differences in both the type of size measures (length, radius, aspect ratio, etc.), and the type of average or distributions afforded by the specific measures are compared. The strengths of single particle methods, such as electron microscopy and atomic force microscopy, with respect to imaging, shape determinations and application to particle process studies are discussed, together with their limitations in terms of counting statistics and sample preparation. Methods based on the measurement of particle populations are discussed in terms of their quantitative analyses, but the necessity of knowing their limitations in size range and concentration range is also considered. The advantage of combining complementary methods is highlighted.

  18. Foster Care Reentry: A survival analysis assessing differences across permanency type.

    Science.gov (United States)

    Goering, Emily Smith; Shaw, Terry V

    2017-06-01

    Foster care reentry is an important factor for evaluating the overall success of permanency. Rates of reentry are typically only measured for 12-months and are often evaluated only for children who exit foster care to reunification and not across exit types, also known as 'permanency types'. This study examined the odds of reentry across multiple common permanency types for a cohort of 8107 children who achieved permanency between 2009 and 2013. Overall, 14% of children reentered care within 18-months with an average time to reentry of 6.36 months. A Kaplan-Meier survival analysis was used to assess differences in reentry across permanency types (including reunification, relative guardianship and non-relative guardianship). Children who achieved guardianship with kin had the lowest odds of reentry overall, followed by guardianship with non-kin, and reunification with family of origin. Children reunifying against the recommendations of Children and Family Services had the highest odds of reentry. A Cox regression survival analysis was conducted to assess odds of reentry across permanency type while controlling for demographics, services, and other risk factors. In the final model, only permanency type and cumulative risk were found to have a statistically significant impact on odds of reentry. Copyright © 2017 Elsevier Ltd. All rights reserved.

  19. Assessing Urban Wastewater System Upgrades Using Integrated Modeling, Life Cycle Analysis, and Shadow Pricing.

    Science.gov (United States)

    Hadjimichael, Antonia; Morera, Serni; Benedetti, Lorenzo; Flameling, Tony; Corominas, Lluís; Weijers, Stefan; Comas, Joaquim

    2016-12-06

    This study assesses the environmental impacts of four measures proposed for upgrading of the urban wastewater system of Eindhoven and the Dommel River in The Netherlands, against the base case, "do-nothing" option. The measures aim to reduce the overall environmental impact of the Eindhoven urban wastewater system (UWS) by targeting river dissolved oxygen depletion and ammonia peaks, reducing combined sewer overflows, and enhancing nutrient removal. The measures are evaluated using a life cycle analysis with the boundaries including the receiving river section by means of an integrated model of the UWS. An uncertainty analysis of the estimated impacts has been performed to support the outcomes. The study also uses the economic concept of shadow prices to assign relative weights of socio-economic importance to the estimated life cycle impacts. This novel integration of tools complements the assessments of this UWS with the inclusion of long-term global environmental impacts and the investigation of trade-offs between different environmental impacts through a single monetary unit. The results support the selection of deeper clarifiers as the most environmentally beneficial measure for upgrade.

  20. Building Damage Assessment Using Scenario Based Tsunami Numerical Analysis and Fragility Curves

    Directory of Open Access Journals (Sweden)

    Khawar Rehman

    2016-03-01

    Full Text Available A combination of a deterministic approach and fragility analysis is applied to assess tsunami damage caused to buildings. The area selected to validate the model is Imwon Port in Korea. The deterministic approach includes numerical modeling of tsunami propagation in the East Sea following an earthquake on the western coast of Japan. The model is based on the linear shallow-water equations (LSWE augmented with Boussinesq approximation to account for dispersion effects in wave propagation, and coastal wave run-up is modeled by non-linear shallow-water equations (NLSWE. The output from the deterministic model comprises inundation depth. The numerical output is used to perform fragility analysis for buildings vulnerable to flooding by tsunamis in the port area. Recently developed fragility curves—based on the ordinal regression method—are used for damage probability estimates. The extent of structural damage in the areas under a tsunami hazard is identified by the numerical modeling of tsunami features. Our numerical model offers high bathymetric resolution, which enables us to capture flow features at the individual structure level and results in improved estimation of damage probability. This approach can serve as a measure of assessing structure vulnerability for areas with little or no records of tsunami damage and provide planners with a better understanding of structure behavior when a tsunami strikes.

  1. Operational Implementation of a Pc Uncertainty Construct for Conjunction Assessment Risk Analysis

    Science.gov (United States)

    Newman, Lauri K.; Hejduk, Matthew D.; Johnson, Lauren C.

    2016-01-01

    Earlier this year the NASA Conjunction Assessment and Risk Analysis (CARA) project presented the theoretical and algorithmic aspects of a method to include the uncertainties in the calculation inputs when computing the probability of collision (Pc) between two space objects, principally uncertainties in the covariances and the hard-body radius. The output of this calculation approach is to produce rather than a single Pc value an entire probability density function that will represent the range of possible Pc values given the uncertainties in the inputs and bring CA risk analysis methodologies more in line with modern risk management theory. The present study provides results from the exercise of this method against an extended dataset of satellite conjunctions in order to determine the effect of its use on the evaluation of conjunction assessment (CA) event risk posture. The effects are found to be considerable: a good number of events are downgraded from or upgraded to a serious risk designation on the basis of consideration of the Pc uncertainty. The findings counsel the integration of the developed methods into NASA CA operations.

  2. Towards an analysis of leak-before-break assessments in the ductile tearing regime

    Energy Technology Data Exchange (ETDEWEB)

    Parfitt, V.R.

    1991-12-31

    This paper presents the elastic-plastic fracture analysis of a typical semi-elliptical axial surface flaw growing to a thru-thickness flaw and assessing the leak-before break situation in a pressure vessel subject to pressure. The paper first discusses the semi-elliptical flaw J-integral solution and the thru-thickness flaw solution as modified herein. Then a review is presented of the three ductile tearing stability fracture analysis methods based on the fully plastic J-integral solution; (1) the crack driving force diagram, (2) the tearing modulus diagram, and (3) the failure assessment diagram. These methods are then used to determine the factors of safety to initiation of ductile tearing as the crack grows. Factors of safety based on either pressure alone or crack size alone are illustrated. An illustration is given of a leak-before-break solution discussing the semi-elliptical flaw growing to a thru-thickness flaw in the vessel. The paper concludes with a discussion of additional effort needed to better characterize leak-before-break solutions in the ductile tearing regime.

  3. Exergy analysis in the assessment of the sustainability of waste gas treatment systems.

    Science.gov (United States)

    Dewulf, J; Van Langenhove, H; Dirckx, J

    2001-06-12

    This study focuses on the sustainability of different technological options for the treatment of waste gases from a waste water treatment plant loaded with volatile organic compounds. The options considered are biofiltration, active carbon adsorption and catalytic and thermal oxidation. The amount of resources and utilities to construct and operate each system have been investigated from the point of view of the Second Law of thermodynamics. The unit in which all resources are treated is Joules of exergy. It was concluded that biofiltration was the most exergetically efficient system. The cumulative exergy consumption of the resources and utilities for construction and operation have been quantified in exergy terms. Further on, the requirements for the abatement of emissions generated by operating the waste gas treatment systems and the amount of renewables have been taken into account in the assessment of the sustainability of the waste gas treatment technologies. Finally, a comparison between exergy analysis and life cycle analysis in assessing the sustainability of the waste gas treatment options, is presented.

  4. Assessment of hydrocephalus in children based on digital image processing and analysis

    Directory of Open Access Journals (Sweden)

    Fabijańska Anna

    2014-06-01

    Full Text Available Hydrocephalus is a pathological condition of the central nervous system which often affects neonates and young children. It manifests itself as an abnormal accumulation of cerebrospinal fluid within the ventricular system of the brain with its subsequent progression. One of the most important diagnostic methods of identifying hydrocephalus is Computer Tomography (CT. The enlarged ventricular system is clearly visible on CT scans. However, the assessment of the disease progress usually relies on the radiologist’s judgment and manual measurements, which are subjective, cumbersome and have limited accuracy. Therefore, this paper regards the problem of semi-automatic assessment of hydrocephalus using image processing and analysis algorithms. In particular, automated determination of popular indices of the disease progress is considered. Algorithms for the detection, semi-automatic segmentation and numerical description of the lesion are proposed. Specifically, the disease progress is determined using shape analysis algorithms. Numerical results provided by the introduced methods are presented and compared with those calculated manually by a radiologist and a trained operator. The comparison proves the correctness of the introduced approach.

  5. Operational Implementation of a Pc Uncertainty Construct for Conjunction Assessment Risk Analysis

    Science.gov (United States)

    Newman, L.; Hejduk, M.; Johnson, L.

    2016-09-01

    Earlier this year the NASA Conjunction Assessment and Risk Analysis (CARA) project presented the theoretical and algorithmic aspects of a method to include the uncertainties in the calculation inputs when computing the probability of collision (Pc) between two space objects, principally uncertainties in the covariances and the hardbody radius. The output of this calculation approach is to produce rather than a single Pc value an entire probability density function that will represent the range of possible Pc values given the uncertainties in the inputs and bring CA risk analysis methodologies more in line with modern risk management theory. The present study provides results from the exercise of this method against an extended dataset of satellite conjunctions in order to determine the effect of its use on the evaluation of conjunction assessment (CA) event risk posture. The effects are found to be considerable: a good number of events are downgraded from or upgraded to a serious risk designation on the basis of consideration of the Pc uncertainty. The findings counsel the integration of the developed methods into NASA CA operations.

  6. EVA: laparoscopic instrument tracking based on Endoscopic Video Analysis for psychomotor skills assessment.

    Science.gov (United States)

    Oropesa, Ignacio; Sánchez-González, Patricia; Chmarra, Magdalena K; Lamata, Pablo; Fernández, Alvaro; Sánchez-Margallo, Juan A; Jansen, Frank Willem; Dankelman, Jenny; Sánchez-Margallo, Francisco M; Gómez, Enrique J

    2013-03-01

    The EVA (Endoscopic Video Analysis) tracking system is a new system for extracting motions of laparoscopic instruments based on nonobtrusive video tracking. The feasibility of using EVA in laparoscopic settings has been tested in a box trainer setup. EVA makes use of an algorithm that employs information of the laparoscopic instrument's shaft edges in the image, the instrument's insertion point, and the camera's optical center to track the three-dimensional position of the instrument tip. A validation study of EVA comprised a comparison of the measurements achieved with EVA and the TrEndo tracking system. To this end, 42 participants (16 novices, 22 residents, and 4 experts) were asked to perform a peg transfer task in a box trainer. Ten motion-based metrics were used to assess their performance. Construct validation of the EVA has been obtained for seven motion-based metrics. Concurrent validation revealed that there is a strong correlation between the results obtained by EVA and the TrEndo for metrics, such as path length (ρ = 0.97), average speed (ρ = 0.94), or economy of volume (ρ = 0.85), proving the viability of EVA. EVA has been successfully validated in a box trainer setup, showing the potential of endoscopic video analysis to assess laparoscopic psychomotor skills. The results encourage further implementation of video tracking in training setups and image-guided surgery.

  7. Assessment of high resolution melting analysis as a potential SNP genotyping technique in forensic casework.

    Science.gov (United States)

    Venables, Samantha J; Mehta, Bhavik; Daniel, Runa; Walsh, Simon J; van Oorschot, Roland A H; McNevin, Dennis

    2014-11-01

    High resolution melting (HRM) analysis is a simple, cost effective, closed tube SNP genotyping technique with high throughput potential. The effectiveness of HRM for forensic SNP genotyping was assessed with five commercially available HRM kits evaluated on the ViiA™ 7 Real Time PCR instrument. Four kits performed satisfactorily against forensically relevant criteria. One was further assessed to determine the sensitivity, reproducibility, and accuracy of HRM SNP genotyping. The manufacturer's protocol using 0.5 ng input DNA and 45 PCR cycles produced accurate and reproducible results for 17 of the 19 SNPs examined. Problematic SNPs had GC rich flanking regions which introduced additional melting domains into the melting curve (rs1800407) or included homozygotes that were difficult to distinguish reliably (rs16891982; a G to C SNP). A proof of concept multiplexing experiment revealed that multiplexing a small number of SNPs may be possible after further investigation. HRM enables genotyping of a number of SNPs in a large number of samples without extensive optimization. However, it requires more genomic DNA as template in comparison to SNaPshot®. Furthermore, suitably modifying pre-existing forensic intelligence SNP panels for HRM analysis may pose difficulties due to the properties of some SNPs. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  8. Sustainability assessment of tertiary wastewater treatment technologies: a multi-criteria analysis.

    Science.gov (United States)

    Plakas, K V; Georgiadis, A A; Karabelas, A J

    2016-01-01

    The multi-criteria analysis gives the opportunity to researchers, designers and decision-makers to examine decision options in a multi-dimensional fashion. On this basis, four tertiary wastewater treatment (WWT) technologies were assessed regarding their sustainability performance in producing recycled wastewater, considering a 'triple bottom line' approach (i.e. economic, environmental, and social). These are powdered activated carbon adsorption coupled with ultrafiltration membrane separation (PAC-UF), reverse osmosis, ozone/ultraviolet-light oxidation and heterogeneous photo-catalysis coupled with low-pressure membrane separation (photocatalytic membrane reactor, PMR). The participatory method called simple multi-attribute rating technique exploiting ranks was employed for assigning weights to selected sustainability indicators. This sustainability assessment approach resulted in the development of a composite index as a final metric, for each WWT technology evaluated. The PAC-UF technology appears to be the most appropriate technology, attaining the highest composite value regarding the sustainability performance. A scenario analysis confirmed the results of the original scenario in five out of seven cases. In parallel, the PMR was highlighted as the technology with the least variability in its performance. Nevertheless, additional actions and approaches are proposed to strengthen the objectivity of the final results.

  9. Proteomic and bioinformatics analysis of human saliva for the dental-risk assessment

    Directory of Open Access Journals (Sweden)

    Laputková Galina

    2017-10-01

    Full Text Available Background: Dental caries disease is a dynamic process with a multi-factorial etiology. It is manifested by demineralization of enamel followed by damage spreading into the tooth inner structure. Successful early diagnosis could identify caries-risk and improve dental screening, providing a baseline for evaluating personalized dental treatment and prevention strategies. Methodology:\tSalivary proteome of the whole unstimulated saliva (WUS samples was assessed in caries-free and caries-susceptible individuals of older adolescent age with permanent dentition using a nano-HPLC and MALDI-TOF/TOF mass spectrometry. Results: 554 proteins in the caries-free and 695 proteins in the caries-susceptible group were identified. Assessment using bioinformatics tools and Gene Ontology (GO term enrichment analysis revealed qualitative differences between these two proteomes. Members of the caries-susceptible group exhibited a branch of cytokine binding gene products responsible for the regulation of immune and inflammatory responses to infections. Inspection of molecular functions and biological processes of caries-susceptible saliva samples revealed significant categories predominantly related to the activity of proteolytic peptidases, and the regulation of metabolic and catabolic processes of carbohydrates. Conclusions: Proteomic analysis of the whole saliva revealed information about potential risk factors associated with the development of caries-susceptibility and provides a better understanding of tooth protection mechanisms.

  10. Assessing pollution trends in the Guadalquivir River estuary using N-way analysis

    Directory of Open Access Journals (Sweden)

    López-López J. A.

    2013-04-01

    Full Text Available Human activities have contributed to a deployment of environmental quality. In the last decades the problem of water preservation has gained increasing attention, and new policies have been developed for water resources remediation. Statistical techniques for data treatment are based on the organization of data in a bi-dimensional array; thus, some shades on the trend of the distribution tend to be ignored. Multiway techniques, where data are gathered in n directions, allow the analysis of the results through different directions at the same time. In particular, for 3MPCA a principal components analysis is conducted using three modes and a “core” matrix that allows assessing their interactions. The Guadalquivir River estuary has been used as a model system for the application of 3MPCA in the study of long term evolution of pollutants. Nutrients and heavy metals ultra-traces level have been used to characterize the estuary. The 3MPCA was used to assess the relationships within chemical variables, sampling stations and sampling campaigns.

  11. Toward a low-cost gait analysis system for clinical and free-living assessment.

    Science.gov (United States)

    Ladha, Cassim; Del Din, Silvia; Nazarpour, Kianoush; Hickey, Aodhan; Morris, Rosie; Catt, Michael; Rochester, Lynn; Godfrey, Alan

    2016-08-01

    Gait is an important clinical assessment tool since changes in gait may reflect changes in general health. Measurement of gait is a complex process which has been restricted to bespoke clinical facilities until recently. The use of inexpensive wearable technologies is an attractive alternative and offers the potential to assess gait in any environment. In this paper we present the development of a low cost analysis gait system built using entirely open source components. The system is used to capture spatio-temporal gait characteristics derived from an existing conceptual model, sensitive to ageing and neurodegenerative pathology (e.g. Parkinson's disease). We demonstrate the system is suitable for use in a clinical unit and will lead to pragmatic use in a free-living (home) environment. The system consists of a wearable (tri-axial accelerometer and gyroscope) with a Raspberry Pi module for data storage and analysis. This forms ongoing work to develop gait as a low cost diagnostic in modern healthcare.

  12. Bioimpedance Harmonic Analysis as a Diagnostic Tool to Assess Regional Circulation and Neural Activity

    Science.gov (United States)

    Mudraya, I. S.; Revenko, S. V.; Khodyreva, L. A.; Markosyan, T. G.; Dudareva, A. A.; Ibragimov, A. R.; Romich, V. V.; Kirpatovsky, V. I.

    2013-04-01

    The novel technique based on harmonic analysis of bioimpedance microvariations with original hard- and software complex incorporating a high-resolution impedance converter was used to assess the neural activity and circulation in human urinary bladder and penis in patients with pelvic pain, erectile dysfunction, and overactive bladder. The therapeutic effects of shock wave therapy and Botulinum toxin detrusor injections were evaluated quantitatively according to the spectral peaks at low 0.1 Hz frequency (M for Mayer wave), respiratory (R) and cardiac (C) rhythms with their harmonics. Enhanced baseline regional neural activity identified according to M and R peaks was found to be presumably sympathetic in pelvic pain patients, and parasympathetic - in patients with overactive bladder. Total pulsatile activity and pulsatile resonances found in the bladder as well as in the penile spectrum characterised regional circulation and vascular tone. The abnormal spectral parameters characteristic of the patients with genitourinary diseases shifted to the norm in the cases of efficient therapy. Bioimpedance harmonic analysis seems to be a potent tool to assess regional peculiarities of circulatory and autonomic nervous activity in the course of patient treatment.

  13. Microarray analysis reveals the actual specificity of enrichment media used for food safety assessment.

    Science.gov (United States)

    Kostić, Tanja; Stessl, Beatrix; Wagner, Martin; Sessitsch, Angela

    2011-06-01

    Microbial diagnostic microarrays are tools for simultaneous detection and identification of microorganisms in food, clinical, and environmental samples. In comparison to classic methods, microarray-based systems have the potential for high throughput, parallelism, and miniaturization. High specificity and high sensitivity of detection have been demonstrated. A microbial diagnostic microarray for the detection of the most relevant bacterial food- and waterborne pathogens and indicator organisms was developed and thoroughly validated. The microarray platform based on sequence-specific end labeling of oligonucleotides and the phylogenetically robust gyrB marker gene allowed a highly specific (resolution on genus and/or species level) and sensitive (0.1% relative and 10(4) CFU absolute sensitivity) detection of the target pathogens. In initial challenge studies of the applicability of microarray-based food analysis, we obtained results demonstrating the questionable specificity of standardized culture-dependent microbiological detection methods. Taking into consideration the importance of reliable food safety assessment methods, comprehensive performance assessment is essential. Results demonstrate the potential of this new pathogen diagnostic microarray to evaluate culture-based standard methods in microbiological food analysis.

  14. Data Quality Assessment of the Uncertainty Analysis Applied to the Greenhouse Gas Emissions of a Dairy Cow System

    Directory of Open Access Journals (Sweden)

    Chun-Youl Baek

    2017-09-01

    Full Text Available The results of an uncertainty analysis are achieved by the statistical information (standard error, type of probability distributions, and range of minimum and maximum of the selected input parameters. However, there are limitations in identifying sufficient data samples for the selected input parameters for statistical information in the field of life cycle assessment (LCA. Therefore, there is a strong need for a consistent screening procedure to identify the input parameters for use in uncertainty analysis in the area of LCA. The conventional procedure for identifying input parameters for the uncertainty analysis method includes assessing the data quality using the pedigree method and the contribution analysis of the LCA results. This paper proposes a simplified procedure for ameliorating the existing data quality assessment method, which can lead to an efficient uncertainly analysis of LCA results. The proposed method has two salient features: (i a simplified procedure based on contribution analysis followed by a data quality assessment for selecting the input parameters for the uncertainty analysis; and (ii a quantitative data quality assessment method is proposed, based on the pedigree method, that adopts the analytic hierarchy process (AHP method and quality function deployment (QFD. The effects of the uncertainty of the selected input parameters on the LCA results were assessed using the Monte Carlo simulation method. A case study of greenhouse gas (GHG emissions from a dairy cow system was used to demonstrate the applicability of the proposed procedure.

  15. Assessment of occupational safety risks in Floridian solid waste systems using Bayesian analysis.

    Science.gov (United States)

    Bastani, Mehrad; Celik, Nurcin

    2015-10-01

    Safety risks embedded within solid waste management systems continue to be a significant issue and are prevalent at every step in the solid waste management process. To recognise and address these occupational hazards, it is necessary to discover the potential safety concerns that cause them, as well as their direct and/or indirect impacts on the different types of solid waste workers. In this research, our goal is to statistically assess occupational safety risks to solid waste workers in the state of Florida. Here, we first review the related standard industrial codes to major solid waste management methods including recycling, incineration, landfilling, and composting. Then, a quantitative assessment of major risks is conducted based on the data collected using a Bayesian data analysis and predictive methods. The risks estimated in this study for the period of 2005-2012 are then compared with historical statistics (1993-1997) from previous assessment studies. The results have shown that the injury rates among refuse collectors in both musculoskeletal and dermal injuries have decreased from 88 and 15 to 16 and three injuries per 1000 workers, respectively. However, a contrasting trend is observed for the injury rates among recycling workers, for whom musculoskeletal and dermal injuries have increased from 13 and four injuries to 14 and six injuries per 1000 workers, respectively. Lastly, a linear regression model has been proposed to identify major elements of the high number of musculoskeletal and dermal injuries. © The Author(s) 2015.

  16. Discriminant function analysis for sex assessment in pelvic girdle bones: sample from the contemporary Mexican population.

    Science.gov (United States)

    Gómez-Valdés, Jorge Alfredo; Torres Ramírez, Guillermo; Báez Molgado, Socorro; Herrera Sain-Leu, Patricia; Castrejón Caballero, José Luis; Sánchez-Mejorada, Gabriela

    2011-03-01

    Sex assessment of skeletal remains plays an important role in forensic anthropology. The pelvic bones are the most studied part of the postcranial skeleton for the assessment of sex. It is evident that a population-specific approach improves rates of accuracy within the group. The present study proposes a discriminant function method for the sex assessment of skeletal remains from a contemporary Mexican population. A total of 146 adult human pelvic bones (61 females and 85 males) from the skeletal series pertaining to the National Autonomous University of Mexico were evaluated. Twenty-four direct metrical parameters of coxal and sacral bones were measured and subsequently, sides and sex differences were evaluated, applying a stepwise discriminant function analysis. Coxal and sacra functions achieved accuracies of 99% and 87%, respectively. These analyses follow a population-specific approach; nevertheless, we consider that our results are applicable to any other Hispanic samples for purposes of forensic human identification. © 2011 American Academy of Forensic Sciences.

  17. Using Social Network Analysis to Assess Mentorship and Collaboration in a Public Health Network.

    Science.gov (United States)

    Petrescu-Prahova, Miruna; Belza, Basia; Leith, Katherine; Allen, Peg; Coe, Norma B; Anderson, Lynda A

    2015-08-20

    Addressing chronic disease burden requires the creation of collaborative networks to promote systemic changes and engage stakeholders. Although many such networks exist, they are rarely assessed with tools that account for their complexity. This study examined the structure of mentorship and collaboration relationships among members of the Healthy Aging Research Network (HAN) using social network analysis (SNA). We invited 97 HAN members and partners to complete an online social network survey that included closed-ended questions about HAN-specific mentorship and collaboration during the previous 12 months. Collaboration was measured by examining the activity of the network on 6 types of products: published articles, in-progress manuscripts, grant applications, tools, research projects, and presentations. We computed network-level measures such as density, number of components, and centralization to assess the cohesiveness of the network. Sixty-three respondents completed the survey (response rate, 65%). Responses, which included information about collaboration with nonrespondents, suggested that 74% of HAN members were connected through mentorship ties and that all 97 members were connected through at least one form of collaboration. Mentorship and collaboration ties were present both within and across boundaries of HAN member organizations. SNA of public health collaborative networks provides understanding about the structure of relationships that are formed as a result of participation in network activities. This approach may offer members and funders a way to assess the impact of such networks that goes beyond simply measuring products and participation at the individual level.

  18. Integration of Bayesian analysis for eutrophication prediction and assessment in a landscape lake.

    Science.gov (United States)

    Yang, Likun; Zhao, Xinhua; Peng, Sen; Zhou, Guangyu

    2015-01-01

    Eutrophication models have been widely used to assess water quality in landscape lakes. Because flow rate in landscape lakes is relatively low and similar to that of natural lakes, eutrophication is more dominant in landscape lakes. To assess the risk of eutrophication in landscape lakes, a set of dynamic equations was developed to simulate lake water quality for total nitrogen (TN), total phosphorous (TP), dissolve oxygen (DO) and chlorophyll a (Chl a). Firstly, the Bayesian calibration results were described. Moreover, the ability of the model to reproduce adequately the observed mean patterns and major cause-effect relationships for water quality conditions in landscape lakes were presented. Two loading scenarios were used. A Monte Carlo algorithm was applied to calculate the predicated water quality distributions, which were used in the established hierarchical assessment system for lake water quality risk. The important factors affecting the lake water quality risk were defined using linear regression analysis. The results indicated that the variations in the landscape lake receiving recharge water quality caused considerable landscape lake water quality risk in the surrounding area. Moreover, the Chl a concentration in lake water was significantly affected by TP and TN concentrations; the lake TP concentration was the limiting factor for growth of plankton in lake water. The lake water TN concentration provided the basic nutritional requirements. Lastly, lower TN and TP concentrations in the receiving recharge water caused increased lake water quality risk.

  19. Using Habitat Equivalency Analysis to Assess the Cost Effectiveness of Restoration Outcomes in Four Institutional Contexts

    Science.gov (United States)

    Scemama, Pierre; Levrel, Harold

    2016-01-01

    At the national level, with a fixed amount of resources available for public investment in the restoration of biodiversity, it is difficult to prioritize alternative restoration projects. One way to do this is to assess the level of ecosystem services delivered by these projects and to compare them with their costs. The challenge is to derive a common unit of measurement for ecosystem services in order to compare projects which are carried out in different institutional contexts having different goals (application of environmental laws, management of natural reserves, etc.). This paper assesses the use of habitat equivalency analysis (HEA) as a tool to evaluate ecosystem services provided by restoration projects developed in different institutional contexts. This tool was initially developed to quantify the level of ecosystem services required to compensate for non-market impacts coming from accidental pollution in the US. In this paper, HEA is used to assess the cost effectiveness of several restoration projects in relation to different environmental policies, using case studies based in France. Four case studies were used: the creation of a market for wetlands, public acceptance of a port development project, the rehabilitation of marshes to mitigate nitrate loading to the sea, and the restoration of streams in a protected area. Our main conclusion is that HEA can provide a simple tool to clarify the objectives of restoration projects, to compare the cost and effectiveness of these projects, and to carry out trade-offs, without requiring significant amounts of human or technical resources.

  20. Parkinson's disease assessment based on gait analysis using an innovative RGB-D camera system.

    Science.gov (United States)

    Rocha, Ana Patrícia; Choupina, Hugo; Fernandes, José Maria; Rosas, Maria José; Vaz, Rui; Silva Cunha, João Paulo

    2014-01-01

    Movement-related diseases, such as Parkinson's disease (PD), progressively affect the motor function, many times leading to severe motor impairment and dramatic loss of the patients' quality of life. Human motion analysis techniques can be very useful to support clinical assessment of this type of diseases. In this contribution, we present a RGB-D camera (Microsoft Kinect) system and its evaluation for PD assessment. Based on skeleton data extracted from the gait of three PD patients treated with deep brain stimulation and three control subjects, several gait parameters were computed and analyzed, with the aim of discriminating between non-PD and PD subjects, as well as between two PD states (stimulator ON and OFF). We verified that among the several quantitative gait parameters, the variance of the center shoulder velocity presented the highest discriminative power to distinguish between non-PD, PD ON and PD OFF states (p = 0.004). Furthermore, we have shown that our low-cost portable system can be easily mounted in any hospital environment for evaluating patients' gait. These results demonstrate the potential of using a RGB-D camera as a PD assessment tool.

  1. Trainee and Client Experiences of Therapeutic Assessment in a Required Graduate Course: A Qualitative Analysis.

    Science.gov (United States)

    Smith, Justin D; Egan, Kaitlyn N

    2017-01-01

    Surveys indicate that practice and training in psychological assessment, and personality assessment (PA) to a lesser degree, has been stable or increasing over the past quarter-century. However, its future arguably remains threatened due to changes in doctoral training programs and beliefs in the field concerning the utility of PA for treatment success. To increase interest in and use of PA, studies of training methods that include trainees' perspectives are needed. This study evaluated the experiences of 10 graduate trainees and their clients who were trained in and conducted a brief Therapeutic Assessment (TA). Qualitative responses to a self-evaluation administered post-TA were coded using directed content analysis. Results indicated that trainees viewed TA/PA as having clinical utility; they had positive feelings about TA/PA, and they desired or intended to use or continue learning about TA/PA. Clients' responses reflected positive feelings about the TA, having gained new self-awareness or understanding, and having a positive relationship with the assessor. The findings suggest that teaching PA from a TA perspective could produce positive benefits for psychology trainees.

  2. Advancing effects analysis for integrated, large-scale wildfire risk assessment.

    Science.gov (United States)

    Thompson, Matthew P; Calkin, David E; Gilbertson-Day, Julie W; Ager, Alan A

    2011-08-01

    In this article, we describe the design and development of a quantitative, geospatial risk assessment tool intended to facilitate monitoring trends in wildfire risk over time and to provide information useful in prioritizing fuels treatments and mitigation measures. The research effort is designed to develop, from a strategic view, a first approximation of how both fire likelihood and intensity influence risk to social, economic, and ecological values at regional and national scales. Three main components are required to generate wildfire risk outputs: (1) burn probability maps generated from wildfire simulations, (2) spatially identified highly valued resources (HVRs), and (3) response functions that describe the effects of fire (beneficial or detrimental) on the HVR. Analyzing fire effects has to date presented a major challenge to integrated risk assessments, due to a limited understanding of the type and magnitude of changes wrought by wildfire to ecological and other nonmarket values. This work advances wildfire effects analysis, recognizing knowledge uncertainty and appropriately managing it through the use of an expert systems approach. Specifically, this work entailed consultation with 10 fire and fuels program management officials from federal agencies with fire management responsibilities in order to define quantitative resource response relationships as a function of fire intensity. Here, we demonstrate a proof-of-concept application of the wildland fire risk assessment tool, using the state of Oregon as a case study.

  3. a Psycholinguistic Model for Simultaneous Translation, and Proficiency Assessment by Automated Acoustic Analysis of Discourse.

    Science.gov (United States)

    Yaghi, Hussein M.

    Two separate but related issues are addressed: how simultaneous translation (ST) works on a cognitive level and how such translation can be objectively assessed. Both of these issues are discussed in the light of qualitative and quantitative analyses of a large corpus of recordings of ST and shadowing. The proposed ST model utilises knowledge derived from a discourse analysis of the data, many accepted facts in the psychology tradition, and evidence from controlled experiments that are carried out here. This model has three advantages: (i) it is based on analyses of extended spontaneous speech rather than word-, syllable-, or clause -bound stimuli; (ii) it draws equally on linguistic and psychological knowledge; and (iii) it adopts a non-traditional view of language called 'the linguistic construction of reality'. The discourse-based knowledge is also used to develop three computerised systems for the assessment of simultaneous translation: one is a semi-automated system that treats the content of the translation; and two are fully automated, one of which is based on the time structure of the acoustic signals whilst the other is based on their cross-correlation. For each system, several parameters of performance are identified, and they are correlated with assessments rendered by the traditional, subjective, qualitative method. Using signal processing techniques, the acoustic analysis of discourse leads to the conclusion that quality in simultaneous translation can be assessed quantitatively with varying degrees of automation. It identifies as measures of performance (i) three content-based standards; (ii) four time management parameters that reflect the influence of the source on the target language time structure; and (iii) two types of acoustical signal coherence. Proficiency in ST is shown to be directly related to coherence and speech rate but inversely related to omission and delay. High proficiency is associated with a high degree of simultaneity and

  4. A Support Analysis Framework for mass movement damage assessment: applications to case studies in Calabria (Italy)

    Science.gov (United States)

    Petrucci, O.; Gullã, G.

    2009-03-01

    The analysis of data describing damage caused by mass movements in Calabria (Italy) allowed the organisation of the Support Analysis Framework (SAF), a spreadsheet that converts damage descriptions into numerical indices expressing direct, indirect, and intangible damage. The SAF assesses damage indices of past mass movements and the potential outcomes of dormant phenomena re-activations. It is based on the effects on damaged elements and is independent of both physical and geometric phenomenon characteristics. SAF sections that assess direct damage encompass several lines, each describing an element characterised by a value fixed on a relative arbitrary scale. The levels of loss are classified as: L4: complete; L3: high; L2: medium; or L1: low. For a generic line l, the SAF multiplies the value of a damaged element by its level of loss, obtaining dl, the contribution of the line to the damage. Indirect damage is appraised by two sections accounting for: (a) actions aiming to overcome emergency situations and (b) actions aiming to restore pre-movement conditions. The level of loss depends on the number of people involved (a) or the cost of actions (b). For intangible damage, the level of loss depends on the number of people involved. We examined three phenomena, assessing damage using the SAF and SAFL, customised versions of SAF based on the elements actually present in the analysed municipalities that consider the values of elements in the community framework. We show that in less populated, inland, and affluent municipalities, the impact of mass movements is greater than in coastal areas. The SAF can be useful to sort groups of phenomena according to their probable future damage, supplying results significant either for insurance companies or for local authorities involved in both disaster management and planning of defensive measures.

  5. Applications of life cycle assessment and cost analysis in health care waste management

    Energy Technology Data Exchange (ETDEWEB)

    Soares, Sebastiao Roberto, E-mail: soares@ens.ufsc.br [Department of Sanitary Engineering, Federal University of Santa Catarina, UFSC, Campus Universitario, Centro Tecnologico, Trindade, PO Box 476, Florianopolis, SC 88040-970 (Brazil); Finotti, Alexandra Rodrigues, E-mail: finotti@ens.ufsc.br [Department of Sanitary Engineering, Federal University of Santa Catarina, UFSC, Campus Universitario, Centro Tecnologico, Trindade, PO Box 476, Florianopolis, SC 88040-970 (Brazil); Prudencio da Silva, Vamilson, E-mail: vamilson@epagri.sc.gov.br [Department of Sanitary Engineering, Federal University of Santa Catarina, UFSC, Campus Universitario, Centro Tecnologico, Trindade, PO Box 476, Florianopolis, SC 88040-970 (Brazil); EPAGRI, Rod. Admar Gonzaga 1347, Itacorubi, Florianopolis, Santa Catarina 88034-901 (Brazil); Alvarenga, Rodrigo A.F., E-mail: alvarenga.raf@gmail.com [Department of Sanitary Engineering, Federal University of Santa Catarina, UFSC, Campus Universitario, Centro Tecnologico, Trindade, PO Box 476, Florianopolis, SC 88040-970 (Brazil); Ghent University, Department of Sustainable Organic Chemistry and Technology, Coupure Links 653/9000 Gent (Belgium)

    2013-01-15

    Highlights: Black-Right-Pointing-Pointer Three Health Care Waste (HCW) scenarios were assessed through environmental and cost analysis. Black-Right-Pointing-Pointer HCW treatment using microwave oven had the lowest environmental impacts and costs in comparison with autoclave and lime. Black-Right-Pointing-Pointer Lime had the worst environmental and economic results for HCW treatment, in comparison with autoclave and microwave. - Abstract: The establishment of rules to manage Health Care Waste (HCW) is a challenge for the public sector. Regulatory agencies must ensure the safety of waste management alternatives for two very different profiles of generators: (1) hospitals, which concentrate the production of HCW and (2) small establishments, such as clinics, pharmacies and other sources, that generate dispersed quantities of HCW and are scattered throughout the city. To assist in developing sector regulations for the small generators, we evaluated three management scenarios using decision-making tools. They consisted of a disinfection technique (microwave, autoclave and lime) followed by landfilling, where transportation was also included. The microwave, autoclave and lime techniques were tested at the laboratory to establish the operating parameters to ensure their efficiency in disinfection. Using a life cycle assessment (LCA) and cost analysis, the decision-making tools aimed to determine the technique with the best environmental performance. This consisted of evaluating the eco-efficiency of each scenario. Based on the life cycle assessment, microwaving had the lowest environmental impact (12.64 Pt) followed by autoclaving (48.46 Pt). The cost analyses indicated values of US$ 0.12 kg{sup -1} for the waste treated with microwaves, US$ 1.10 kg{sup -1} for the waste treated by the autoclave and US$ 1.53 kg{sup -1} for the waste treated with lime. The microwave disinfection presented the best eco-efficiency performance among those studied and provided a feasible

  6. Life cycle assessment of Italian citrus-based products. Sensitivity analysis and improvement scenarios.

    Science.gov (United States)

    Beccali, Marco; Cellura, Maurizio; Iudicello, Maria; Mistretta, Marina

    2010-07-01

    Though many studies concern the agro-food sector in the EU and Italy, and its environmental impacts, literature is quite lacking in works regarding LCA application on citrus products. This paper represents one of the first studies on the environmental impacts of citrus products in order to suggest feasible strategies and actions to improve their environmental performance. In particular, it is part of a research aimed to estimate environmental burdens associated with the production of the following citrus-based products: essential oil, natural juice and concentrated juice from oranges and lemons. The life cycle assessment of these products, published in a previous paper, had highlighted significant environmental issues in terms of energy consumption, associated CO(2) emissions, and water consumption. Starting from such results the authors carry out an improvement analysis of the assessed production system, whereby sustainable scenarios for saving water and energy are proposed to reduce environmental burdens of the examined production system. In addition, a sensitivity analysis to estimate the effects of the chosen methods will be performed, giving data on the outcome of the study. Uncertainty related to allocation methods, secondary data sources, and initial assumptions on cultivation, transport modes, and waste management is analysed. The results of the performed analyses allow stating that every assessed eco-profile is differently influenced by the uncertainty study. Different assumptions on initial data and methods showed very sensible variations in the energy and environmental performances of the final products. Besides, the results show energy and environmental benefits that clearly state the improvement of the products eco-profile, by reusing purified water use for irrigation, using the railway mode for the delivery of final products, when possible, and adopting efficient technologies, as the mechanical vapour recompression, in the pasteurisation and

  7. Product analysis and initial reliability testing of the total mesorectal excision-quality assessment instrument.

    Science.gov (United States)

    Simunovic, Marko R; DeNardi, Franco G; Coates, Angela J; Szalay, David A; Eva, Kevin W

    2014-07-01

    Product analysis of rectal cancer resection specimens before specimen fixation may provide an immediate and relevant evaluation of surgical performance. We tested the interrater reliability (IRR) of a product analysis tool called the Total Mesorectal Excision-Quality Assessment Instrument (TME-QA). Participants included two gold standard raters, five pathology assistants, and eight pathologists. Domains of the TME-QA reflect total mesorectal excision principles including: (1) completeness of mesorectal margin; (2) completeness of mesorectum; (3) coning of distal mesorectum; (4) physical defects; and (5) overall specimen quality. Specimens were scored independently. We used the generalizability theory to assess the tool's internal consistency and IRR. There were 39 specimens and 120 ratings. Mean overall specimen quality scores for the gold standard raters, pathologists, and assistants were 4.43, 4.43, and 4.50, respectively (p > 0.85). IRR for the first nine items was 0.68 for the full sample, 0.62 for assistants alone, 0.63 for pathologists alone, and 0.74 for gold standard raters alone. IRR for the item overall specimen quality was 0.67 for the full sample, 0.45 for assistants, 0.80 for pathologists, and 0.86 for gold standard raters. IRR increased for all groups when scores were averaged across two raters. Assessment of surgical specimens using the TME-QA may provide rapid and relevant feedback to surgeons about their technical performance. Our results show good internal consistency and IRR when the TME-QA is used by pathologists. However, for pathology assistants, multiple ratings with the averaging of scores may be needed.

  8. Sex assessment from carpals bones: discriminant function analysis in a contemporary Mexican sample.

    Science.gov (United States)

    Mastrangelo, Paola; De Luca, Stefano; Sánchez-Mejorada, Gabriela

    2011-06-15

    Sex assessment is one of the first essential steps in human identification, in both medico-legal cases and bio-archaeological contexts. Fragmentary human remains compromised by different types of burial or physical insults may frustrate the use of the traditional sex estimation methods, such as the analysis of the skull and pelvis. Currently, the application of discriminant functions to sex unidentified skeletal remains is steadily increasing. However, several studies have demonstrated that, due to variation in size and patterns of sexual dimorphism, discriminant functions are population-specific. In this study, in order to improve sex assessment from skeletal remains and to establish population-specific discriminant functions, the diagnostic values of the carpal bones were considered. A sample of 136 individuals (78 males, 58 females) of known sex and age was analyzed. They belong to a contemporary identified collection from the Laboratory of Physical Anthropology, Faculty of Medicine, UNAM (Universidad Nacional Autónoma de México, Mexico City). The age of the individuals ranged between 25 and 85 years. Between four and nine measurements of each carpal bone were taken. Independent t-tests confirm that all carpals are sexually dimorphic. Univariate measurements produce accuracy levels that range from 61.8% to 90.8%. Classification accuracies ranged between 81.3% and 92.3% in the multivariate stepwise discriminant analysis. In addition, intra- and inter-observer error tests were performed. These indicated that replication of measurements was satisfactory for the same observer over time and between observers. These results suggest that carpal bones can be used for assessing sex in both forensic and bio-archaeological identification procedures and that bone dimensions are population specific. Copyright © 2011 Elsevier Ireland Ltd. All rights reserved.

  9. Acceptable variability in external quality assessment programmes for basic semen analysis.

    Science.gov (United States)

    Palacios, E R; Clavero, A; Gonzalvo, M C; Rosales, A; Mozas, J; Martínez, L; Ramírez, J P; Björndahl, L; Morancho-Zaragoza, J; Fernández-Pardo, E; Castilla, J A

    2012-02-01

    External quality assessment is essential in modern andrology laboratories. To assess the proficiency of laboratories participating in an external quality assessment programme (EQAP), limits for acceptable variability must be determined. Limits currently specified largely depend on criteria set by the organizers of individual EQAP schemes. The objective of this study was to evaluate the different criteria described in ISO 13528: 2005 for calculating acceptable variability in EQAP when applied to basic semen analysis parameters. The data used in this study were the means and standard deviations obtained for independent samples from two EQAPs, one national (Spanish) and one international (European). The acceptable variability according to ISO 13528: 2005 was calculated using four types of criteria: (i) ± 3 standard deviations of the results of all participating laboratories; (ii) ± 3 standard deviations of the results of expert laboratories; (iii) quality specifications based on biological variability, state-of-the-art and clinicians' opinions and (iv) the same quality specifications adjusted for the uncertainty of the assigned value. The first two strategies resulted in very wide ranges of acceptable variability. Conversely, the strategy based only on quality specifications resulted in very narrow ranges. For the fourth strategy, acceptable ranges were intermediate between the results produced with the other strategies. The third and fourth strategies did not produce observable differences in acceptable ranges when the model used for calculating the specifications of analytical quality was changed. It is essential that EQAPs for semen parameters should determine the ranges for acceptable variability in results. Moreover, these ranges must be clinically useful, i.e. the variability should have a minimal negative impact on clinical decisions. The exact definition of 'expert laboratory' is more important than the model chosen for estimating analytical quality

  10. Assessment of genetic stability in micropropagules of Jatropha curcas genotypes by RAPD and AFLP analysis

    KAUST Repository

    Sharma, Sweta K.

    2011-07-01

    Jatropha curcas (Euphorbiaceae), a drought resistant non edible oil yielding plant, has acquired significant importance as an alternative renewable energy source. Low and inconsistent yields found in field plantations prompted for identification of high yielding clones and their large scale multiplication by vegetative propagation to obtain true to type plants. In the current investigation plantlets of J. curcas generated by axillary bud proliferation (micropropagation) using nodal segments obtained from selected high yielding genotypes were assessed for their genetic stability using Randomly Amplified Polymorphic DNA (RAPD) and Amplified Fragment Length Polymorphism (AFLP) analyses. For RAPD analysis, 21 out of 52 arbitrary decamer primers screened gave clear reproducible bands. In the micropropagated plantlets obtained from the 2nd sub-culture, 4 out of a total of 177 bands scored were polymorphic, but in the 8th and 16th sub-cultures (culture cycle) no polymorphisms were detected. AFLP analysis revealed 0.63%, 0% and 0% polymorphism in the 2nd, 8th and 16th generations, respectively. When different genotypes, viz. IC 56557 16, IC 56557 34 and IC 56557 13, were assessed by AFLP, 0%, 0.31% and 0.47% polymorphisms were found, respectively, indicating a difference in genetic stability among the different genotypes. To the best of our knowledge this is the first report on assessment of genetic stability of micropropagated plantlets in J. curcas and suggests that axillary shoot proliferation can safely be used as an efficient micropropagation method for mass propagation of J. curcas. © 2011 Elsevier B.V.

  11. A Support Analysis Framework for mass movement damage assessment: applications to case studies in Calabria (Italy

    Directory of Open Access Journals (Sweden)

    O. Petrucci

    2009-03-01

    Full Text Available The analysis of data describing damage caused by mass movements in Calabria (Italy allowed the organisation of the Support Analysis Framework (SAF, a spreadsheet that converts damage descriptions into numerical indices expressing direct, indirect, and intangible damage.

    The SAF assesses damage indices of past mass movements and the potential outcomes of dormant phenomena re-activations. It is based on the effects on damaged elements and is independent of both physical and geometric phenomenon characteristics.

    SAF sections that assess direct damage encompass several lines, each describing an element characterised by a value fixed on a relative arbitrary scale. The levels of loss are classified as: L4: complete; L3: high; L2: medium; or L1: low. For a generic line l, the SAF multiplies the value of a damaged element by its level of loss, obtaining dl, the contribution of the line to the damage.

    Indirect damage is appraised by two sections accounting for: (a actions aiming to overcome emergency situations and (b actions aiming to restore pre-movement conditions. The level of loss depends on the number of people involved (a or the cost of actions (b.

    For intangible damage, the level of loss depends on the number of people involved.

    We examined three phenomena, assessing damage using the SAF and SAFL, customised versions of SAF based on the elements actually present in the analysed municipalities that consider the values of elements in the community framework. We show that in less populated, inland, and affluent municipalities, the impact of mass movements is greater than in coastal areas.

    The SAF can be useful to sort groups of phenomena according to their probable future damage, supplying results significant either for insurance companies or for local authorities involved in both disaster management and planning of defensive measures.

  12. The assessment of neural injury following open heart surgery by physiological tremor analysis.

    Science.gov (United States)

    Németh, Adám; Hejjel, László; Ajtay, Zénó; Kellényi, Lóránd; Solymos, Andor; Bártfai, Imre; Kovács, Norbert; Lenkey, Zsófia; Cziráki, Attila; Szabados, Sándor

    2013-02-21

    The appearance of post-operative cognitive dysfunction as a result of open heart surgery has been proven by several studies. Focal and/or sporadic neuron damage emerging in the central nervous system may not only appear as cognitive dysfunction, but might strongly influence features of physiological tremor. We investigated 110 patients (age: 34-73 years; 76 male, 34 female; 51 coronary artery bypass grafting (CABG), 25 valve replacement, 25 combined open heart surgery, 9 off-pump CABG) before surgery and after open-heart surgery on the 3(rd) to 5(th) post-operative day. The assessment of the physiological tremor analysis was performed with our newly developed equipment based on the Analog Devices ADXL 320 JPC integrated accelerometer chip. Recordings were stored on a PC and spectral analysis was performed by fast Fourier transformation (FFT). We compared power integrals in the 1-4 Hz, 4-8 Hz and 8-12 Hz frequency ranges and these were statistically assessed by the Wilcoxon rank correlation test. We found significant changes in the power spectrum of physiological tremor. The spectrum in the 8-12 Hz range (neuronal oscillation) decreased and a shift was recognised to the lower spectrum (p < 0.01). The magnitude of the shift was not significantly higher for females than for males (p < 0.157). We found no significant difference between the shift and the cross-clamp or perfusion time (p < 0.6450). The assessment of physiological tremor by means of our novel, feasible method may provide a deeper insight into the mechanism of central nervous system damage associated with open heart surgery.

  13. Application of texture analysis to DAT SPECT imaging: Relationship to clinical assessments

    Directory of Open Access Journals (Sweden)

    Arman Rahmim

    2016-01-01

    Full Text Available Dopamine transporter (DAT SPECT imaging is increasingly utilized for diagnostic purposes in suspected Parkinsonian syndromes. We performed a cross-sectional study to investigate whether assessment of texture in DAT SPECT radiotracer uptake enables enhanced correlations with severity of motor and cognitive symptoms in Parkinson's disease (PD, with the long-term goal of enabling clinical utility of DAT SPECT imaging, beyond standard diagnostic tasks, to tracking of progression in PD. Quantitative analysis in routine DAT SPECT imaging, if performed at all, has been restricted to assessment of mean regional uptake. We applied a framework wherein textural features were extracted from the images. Notably, the framework did not require registration to a common template, and worked in the subject-native space. Image analysis included registration of SPECT images onto corresponding MRI images, automatic region-of-interest (ROI extraction on the MRI images, followed by computation of Haralick texture features. We analyzed 141 subjects from the Parkinson's Progressive Marker Initiative (PPMI database, including 85 PD and 56 healthy controls (HC (baseline scans with accompanying 3 T MRI images. We performed univariate and multivariate regression analyses between the quantitative metrics and different clinical measures, namely (i the UPDRS (part III - motor score, disease duration as measured from (ii time of diagnosis (DD-diag. and (iii time of appearance of symptoms (DD-sympt., as well as (iv the Montreal Cognitive Assessment (MoCA score. For conventional mean uptake analysis in the putamen, we showed significant correlations with clinical measures only when both HC and PD were included (Pearson correlation r = −0.74, p-value < 0.001. However, this was not significant when applied to PD subjects only (r = −0.19, p-value = 0.084, and no such correlations were observed in the caudate. By contrast, for the PD subjects, significant correlations

  14. Automated Retroillumination Photography Analysis for Objective Assessment of Fuchs Corneal Dystrophy.

    Science.gov (United States)

    Eghrari, Allen O; Mumtaz, Aisha A; Garrett, Brian; Rezaei, Mahsa; Akhavan, Mina S; Riazuddin, S Amer; Gottsch, John D

    2017-01-01

    Retroillumination photography analysis is an objective tool for the assessment of the number and distribution of guttae in eyes affected with Fuchs corneal dystrophy (FCD). Current protocols include manual processing of images; here, we assess validity and interrater reliability of automated analysis across various levels of FCD severity. Retroillumination photographs of 97 FCD-affected corneas were acquired, and total counts of guttae were previously summated manually. For each cornea, a single image was loaded into ImageJ software. We reduced color variability and subtracted background noise. Reflection of light from each gutta was identified as a local area of maximum intensity and counted automatically. Noise tolerance level was titrated for each cornea by examining a small region of each image with automated overlay to ensure appropriate coverage of individual guttae. We tested interrater reliability of automated counts of guttae across a spectrum of clinical and educational experience. A set of 97 retroillumination photographs was analyzed. Clinical severity as measured by a modified Krachmer scale ranged from a severity level of 1 to 5 in the set of analyzed corneas. Automated counts by an ophthalmologist correlated strongly with Krachmer grading (R = 0.79) and manual counts (R = 0.88). Intraclass correlation coefficients demonstrated strong correlation at 0.924 (95% CI, 0.870-0.958) among cases analyzed by 3 students, and 0.869 (95% CI, 0.797-0.918) among cases for which images were analyzed by an ophthalmologist and 2 students. Automated retroillumination photography analysis allows for grading of FCD severity with high resolution across a spectrum of disease severity.

  15. LAVA (Los Alamos Vulnerability and Risk Assessment Methodology): A conceptual framework for automated risk analysis

    Energy Technology Data Exchange (ETDEWEB)

    Smith, S.T.; Lim, J.J.; Phillips, J.R.; Tisinger, R.M.; Brown, D.C.; FitzGerald, P.D.

    1986-01-01

    At Los Alamos National Laboratory, we have developed an original methodology for performing risk analyses on subject systems characterized by a general set of asset categories, a general spectrum of threats, a definable system-specific set of safeguards protecting the assets from the threats, and a general set of outcomes resulting from threats exploiting weaknesses in the safeguards system. The Los Alamos Vulnerability and Risk Assessment Methodology (LAVA) models complex systems having large amounts of ''soft'' information about both the system itself and occurrences related to the system. Its structure lends itself well to automation on a portable computer, making it possible to analyze numerous similar but geographically separated installations consistently and in as much depth as the subject system warrants. LAVA is based on hierarchical systems theory, event trees, fuzzy sets, natural-language processing, decision theory, and utility theory. LAVA's framework is a hierarchical set of fuzzy event trees that relate the results of several embedded (or sub-) analyses: a vulnerability assessment providing information about the presence and efficacy of system safeguards, a threat analysis providing information about static (background) and dynamic (changing) threat components coupled with an analysis of asset ''attractiveness'' to the dynamic threat, and a consequence analysis providing information about the outcome spectrum's severity measures and impact values. By using LAVA, we have modeled our widely used computer security application as well as LAVA/CS systems for physical protection, transborder data flow, contract awards, and property management. It is presently being applied for modeling risk management in embedded systems, survivability systems, and weapons systems security. LAVA is especially effective in modeling subject systems that include a large human component.

  16. An assessment of turbulence models for linear hydrodynamic stability analysis of strongly swirling jets

    Science.gov (United States)

    Rukes, Lothar; Paschereit, Christian Oliver; Oberleithner, Kilian

    2016-09-01

    Linear stability analysis has proven to be a useful tool in the analysis of dominant coherent structures, such as the von K\\'{a}rm\\'{a}n vortex street and the global spiral mode associated with the vortex breakdown of swirling jets. In recent years, linear stability analysis has been applied successfully to turbulent time-mean flows, instead of laminar base-flows, \\textcolor{black}{which requires turbulent models that account for the interaction of the turbulent field with the coherent structures. To retain the stability equations of laminar flows, the Boussinesq approximation with a spatially nonuniform but isotropic eddy viscosity is typically employed. In this work we assess the applicability of this concept to turbulent strongly swirling jets, a class of flows that is particularly unsuited for isotropic eddy viscosity models. Indeed we find that unsteady RANS simulations only match with experiments with a Reynolds stress model that accounts for an anisotropic eddy viscosity. However, linear stability analysis of the mean flow is shown to accurately predict the global mode growth rate and frequency if the employed isotropic eddy viscosity represents a least-squares approximation of the anisotropic eddy viscosity. Viscosities derived from the $k-\\epsilon$ model did not achieve a good prediction of the mean flow nor did they allow for accurate stability calculations. We conclude from this study that linear stability analysis can be accurate for flows with strongly anisotropic turbulent viscosity and the capability of the Boussinesq approximation in terms of URANS-based mean flow prediction is not a prerequisite.

  17. Assessing social isolation in motor neurone disease: a Rasch analysis of the MND Social Withdrawal Scale.

    Science.gov (United States)

    Gibbons, Chris J; Thornton, Everard W; Ealing, John; Shaw, Pamela J; Talbot, Kevin; Tennant, Alan; Young, Carolyn A

    2013-11-15

    Social withdrawal is described as the condition in which an individual experiences a desire to make social contact, but is unable to satisfy that desire. It is an important issue for patients with motor neurone disease who are likely to experience severe physical impairment. This study aims to reassess the psychometric and scaling properties of the MND Social Withdrawal Scale (MND-SWS) domains and examine the feasibility of a summary scale, by applying scale data to the Rasch model. The MND Social Withdrawal Scale was administered to 298 patients with a diagnosis of MND, alongside the Hospital Anxiety and Depression Scale. The factor structure of the MND Social Withdrawal Scale was assessed using confirmatory factor analysis. Model fit, category threshold analysis, differential item functioning (DIF), dimensionality and local dependency were evaluated. Factor analysis confirmed the suitability of the four-factor solution suggested by the original authors. Mokken scale analysis suggested the removal of item five. Rasch analysis removed a further three items; from the Community (one item) and Emotional (two items) withdrawal subscales. Following item reduction, each scale exhibited excellent fit to the Rasch model. A 14-item Summary scale was shown to fit the Rasch model after subtesting the items into three subtests corresponding to the Community, Family and Emotional subscales, indicating that items from these three subscales could be summed together to create a total measure for social withdrawal. Removal of four items from the Social Withdrawal Scale led to a four factor solution with a 14-item hierarchical Summary scale that were all unidimensional, free for DIF and well fitted to the Rasch model. The scale is reliable and allows clinicians and researchers to measure social withdrawal in MND along a unidimensional construct. © 2013. Published by Elsevier B.V. All rights reserved.

  18. Environmental Condition Assessment of US Military Installations Using GIS Based Spatial Multi-Criteria Decision Analysis

    Science.gov (United States)

    Singer, Steve; Wang, Guangxing; Howard, Heidi; Anderson, Alan

    2012-08-01

    Environment functions in various aspects including soil and water conservation, biodiversity and habitats, and landscape aesthetics. Comprehensive assessment of environmental condition is thus a great challenge. The issues include how to assess individual environmental components such as landscape aesthetics and integrate them into an indicator that can comprehensively quantify environmental condition. In this study, a geographic information systems based spatial multi-criteria decision analysis was used to integrate environmental variables and create the indicator. This approach was applied to Fort Riley Military installation in which land condition and its dynamics due to military training activities were assessed. The indicator was derived by integrating soil erosion, water quality, landscape fragmentation, landscape aesthetics, and noise based on the weights from the experts by assessing and ranking the environmental variables in terms of their importance. The results showed that landscape level indicator well quantified the overall environmental condition and its dynamics, while the indicator at level of patch that is defined as a homogeneous area that is different from its surroundings detailed the spatiotemporal variability of environmental condition. The environmental condition was mostly determined by soil erosion, then landscape fragmentation, water quality, landscape aesthetics, and noise. Overall, environmental condition at both landscape and patch levels greatly varied depending on the degree of ground and canopy disturbance and their spatial patterns due to military training activities and being related to slope. It was also determined the environment itself could be recovered quickly once military training was halt or reduced. Thus, this study provided an effective tool for the army land managers to monitor environmental dynamics and plan military training activities. Its limitation lies at that the obtained values of the indicator vary and are

  19. Environmental condition assessment of US military installations using GIS based spatial multi-criteria decision analysis.

    Science.gov (United States)

    Singer, Steve; Wang, Guangxing; Howard, Heidi; Anderson, Alan

    2012-08-01

    Environment functions in various aspects including soil and water conservation, biodiversity and habitats, and landscape aesthetics. Comprehensive assessment of environmental condition is thus a great challenge. The issues include how to assess individual environmental components such as landscape aesthetics and integrate them into an indicator that can comprehensively quantify environmental condition. In this study, a geographic information systems based spatial multi-criteria decision analysis was used to integrate environmental variables and create the indicator. This approach was applied to Fort Riley Military installation in which land condition and its dynamics due to military training activities were assessed. The indicator was derived by integrating soil erosion, water quality, landscape fragmentation, landscape aesthetics, and noise based on the weights from the experts by assessing and ranking the environmental variables in terms of their importance. The results showed that landscape level indicator well quantified the overall environmental condition and its dynamics, while the indicator at level of patch that is defined as a homogeneous area that is different from its surroundings detailed the spatiotemporal variability of environmental condition. The environmental condition was mostly determined by soil erosion, then landscape fragmentation, water quality, landscape aesthetics, and noise. Overall, environmental condition at both landscape and patch levels greatly varied depending on the degree of ground and canopy disturbance and their spatial patterns due to military training activities and being related to slope. It was also determined the environment itself could be recovered quickly once military training was halt or reduced. Thus, this study provided an effective tool for the army land managers to monitor environmental dynamics and plan military training activities. Its limitation lies at that the obtained values of the indicator vary and are

  20. Confusion assessment method: a systematic review and meta-analysis of diagnostic accuracy

    Directory of Open Access Journals (Sweden)

    Shi Q

    2013-09-01

    Full Text Available Qiyun Shi,1,2 Laura Warren,3 Gustavo Saposnik,2 Joy C MacDermid1 1Health and Rehabilitation Sciences, Western University, London, Ontario, Canada; 2Stroke Outcomes Research Center, Department of Medicine, St Michael's Hospital, University of Toronto, Toronto, Ontario, Canada; 3Dalla Lana School of Public Health, University of Toronto, Toronto, Ontario, Canada Background: Delirium is common in the early stages of hospitalization for a variety of acute and chronic diseases. Objectives: To evaluate the diagnostic accuracy of two delirium screening tools, the Confusion Assessment Method (CAM and the Confusion Assessment Method for the Intensive Care Unit (CAM-ICU. Methods: We searched MEDLINE, EMBASE, and PsychInfo for relevant articles published in English up to March 2013. We compared two screening tools to Diagnostic and Statistical Manual of Mental Disorders IV criteria. Two reviewers independently assessed studies to determine their eligibility, validity, and quality. Sensitivity and specificity were calculated using a bivariate model. Results: Twenty-two studies (n = 2,442 patients met the inclusion criteria. All studies demonstrated that these two scales can be administered within ten minutes, by trained clinical or research staff. The pooled sensitivities and specificity for CAM were 82% (95% confidence interval [CI]: 69%–91% and 99% (95% CI: 87%–100%, and 81% (95% CI: 57%–93% and 98% (95% CI: 86%–100% for CAM-ICU, respectively. Conclusion: Both CAM and CAM-ICU are validated instruments for the diagnosis of delirium in a variety of medical settings. However, CAM and CAM-ICU both present higher specificity than sensitivity. Therefore, the use of these tools should not replace clinical judgment. Keywords: confusion assessment method, diagnostic accuracy, delirium, systematic review, meta-analysis