Sentences with phrase «measurement data for analysis»

The research was conducted as a blind study, which means that the hospital sent us measurement data for analysis without any additional information, says Project Manager Tero Koivisto from the Department of Future Technologies.

Not exact matches

Individual growth curve models were developed for multilevel analysis and specifically designed for exploring longitudinal data on individual changes over time.23 Using this approach, we applied the MIXED procedure in SAS (SAS Institute) to account for the random effects of repeated measurements.24 To specify the correct model for our individual growth curves, we compared a series of MIXED models by evaluating the difference in deviance between nested models.23 Both fixed quadratic and cubic MIXED models fit our data well, but we selected the fixed quadratic MIXED model because the addition of a cubic time term was not statistically significant based on a log - likelihood ratio test.
The analysis provided two outputs: «numbers representing physical measurements of» various chemicals, and a «gender determination» that indicates whether the sample's data fall within the typical range for males or females.
Diving deeper into the complex puzzle of mass strandings, the team decided to expand their analysis and include additional oceanographic and atmospheric data sets from NASA's Earth science missions, including Terra, the Sea - viewing Wide Field - of - view Sensor — or SeaWIFS, for short — and Global Precipitation Measurement, as well as the National Oceanic and Atmospheric Administration's Geostationary Operational Environmental Satellite, or GOES, mission.
But computer analyses of collective animal behavior are limited by the need for constant tracking and measurement data for each individual; hence, the mechanics of social animal interaction are not fully understood.
The Next - Generation Proteomics Measurement Platform is combining new instrumentation and data analysis capabilities for significantly higher throughput proteomics than currently available, as well as informatics capabilities for managing the massive volumes of proteomics data that will be generated.
We aim to push its limits on all fronts to establish a technique which combines nanometre 3D resolution with maximum labelling efficiencies, absolute measurements of protein copy numbers, precise multi-colour measurements, high - throughput for large scale statistics and novel data analysis approaches, to address the vast array of exciting biological questions at the nanoscale, which are becoming accessible only now.
Taking models of galaxies from the Hubble Space Telescope Ultra Deep Field (HUDF) and applying a correction for the HUDF point spread function we generate lensed simulations of deep, opti... ▽ More We present a simulation analysis of weak gravitational lensing flexion and shear measurement using shapelet decomposition, and identify differences between flexion and shear measurement noise in deep survey data.
For each measurement, only cells with 5 of more data points per time window were utilized and the entire first time window was excluded from further analysis.
Included in the PowerPoint: Macroeconomic Objectives (AS Level) a) Aggregate Demand (AD) and Aggregate Supply (AS) analysis - the shape and determinants of AD and AS curves; AD = C+I+G + (X-M)- the distinction between a movement along and a shift in AD and AS - the interaction of AD and AS and the determination of the level of output, prices and employment b) Inflation - the definition of inflation; degrees of inflation and the measurement of inflation; deflation and disinflation - the distinction between money values and real data - the cause of inflation (cost - push and demand - pull inflation)- the consequences of inflation c) Balance of payments - the components of the balance of payments accounts (using the IMF / OECD definition): current account; capital and financial account; balancing item - meaning of balance of payments equilibrium and disequilibrium - causes of balance of payments disequilibrium in each component of the accounts - consequences of balance of payments disequilibrium on domestic and external economy d) Exchange rates - definitions and measurement of exchange rates - nominal, real, trade - weighted exchange rates - the determination of exchange rates - floating, fixed, managed float - the factors underlying changes in exchange rates - the effects of changing exchange rates on the domestic and external economy using AD, Marshall - Lerner and J curve analysis - depreciation / appreciation - devaluation / revaluation e) The Terms of Trade - the measurement of the terms of trade - causes of the changes in the terms of trade - the impact of changes in the terms of trade f) Principles of Absolute and comparative advantage - the distinction between absolute and comparative advantage - free trade area, customs union, monetary union, full economic union - trade creation and trade diversion - the benefits of free trade, including the trading possibility curve g) Protectionism - the meaning of protectionism in the context of international trade - different methods of protection and their impact, for example, tariffs, import duties and quotas, export subsidies, embargoes, voluntary export restraints (VERs) and excessive administrative burdens («red tape»)- the arguments in favor of protectionism This PowerPoint is best used when using worksheets and activities to help reinforce the ideas talked about.
Repeated measures of both teachers and students are planned over a three - year period, with annual analysis making use of latent variable measurement models and accounting for the multilevel and longitudinal structure of the data.
Learning Analytics is the measurement, collection, analysis and reporting of data about learners and their contexts, for purposes of understanding and optimizing learning and the environments in which it occurs.
INCLUDES: 36 Student Activity Books (1 copy of each of the six titles per grade level, 32 - pages each) 4 Answer Cases 1 Teacher Guide FEATURES: Flexibility for task centers, independent or partner work, or one - on - one tutoring / remediation Clearly stated objective for each activity that allows you to differentiate Focus on foundational skills and concepts Engaging puzzle format for a fun challenge Immediate feedback for self - checking Titles: Grade 1: Number and Operations: Counting and Place Value Addition and Subtraction: Properties and Situations Addition and Subtraction: Strategies and Equations Addition and Subtraction: Beyond 20 Measurement and Data: Length, Time, and Analysis Geometry: Shapes and Attributes Grade 2: Addition and Subtraction: To 20 and Beyond Foundations of Multiplication: Equal Groups and Arrays Addition and Subtraction: Properties and Place Value Measurement and Data: Length, Time, and Analysis Measurement and Data: Time, Money, and Analysis Geometry: Shapes and Attributes Grade 3: Number and Operations: Multiply and Divide Multiply and Divide: Problem Solving Fractions: Fractions as Numbers Measurement and Data: Use and Interpret Data Geometric Measurement: Perimeter and Area Geometry: Shapes and Attributes Grade 4: Number and Operations: Whole Numbers Number and Operations Multi-Digit and Fractions Fractions: Equivalence and Ordering Fractions: Operations Measurement and Data: Convert and Solve Problems Geometry: Angles and Plane Figures Grade 5: Operations and Algebraic Thinking: Expressions and Patterns Number and Operations: Whole Numbers and Decimals Fractions: Add and Subtract Measurement and Data: Convert and Interpret Geometric Measurement: Volume Geometry: Graphing and 2 - D Figures Grade 6: Ratio and Proportions: Ratios and Problem Solving The Number System: Rational Numbers The Number System: Factors and Multiples Expressions and Equations: Write, Solve, and Analyze Geometry: Problem Solving Statistics and Probability: Variability and Displays
INCLUDES: 6 Student Activity Books (1 copy of each title, 32 - pages each) 2 Answer Cases 1 Teacher Guide FEATURES: Flexibility for task centers, independent or partner work, or one - on - one tutoring / remediation Clearly stated objective for each activity that allows you to differentiate Focus on foundational skills and concepts Engaging puzzle format for a fun challenge Immediate feedback for self - checking Level 2 Titles: Addition and Subtraction: To 20 and Beyond Foundations of Multiplication: Equal Groups and Arrays Addition and Subtraction: Properties and Place Value Measurement and Data: Length, Time, and Analysis Measurement and Data: Time, Money, and Analysis Geometry: Shapes and Attributes
INCLUDES: 6 Student Activity Books (1 copy of each title, 32 - pages each) 2 Answer Cases 1 Teacher Guide FEATURES: Flexibility for task centers, independent or partner work, or one - on - one tutoring / remediation Clearly stated objective for each activity that allows you to differentiate Focus on foundational skills and concepts Engaging puzzle format for a fun challenge Immediate feedback for self - checking Level 1 Titles: Number and Operations: Counting and Place Value Addition and Subtraction: Properties and Situations Addition and Subtraction: Strategies and Equations Addition and Subtraction: Beyond 20 Measurement and Data: Length, Time, and Analysis Geometry: Shapes and Attributes
An introduction for children to all the major mathematical content domains, including number sense, algebra, measurement, geometry, data analysis and probability beginning in kindergarten
For example, due to the lack of ocean data, secondary data is often used to infer what the ocean is doing — thus, the AMO analysis relies not on ocean temperature measurements, but rather on air pressure measurements as a proxy for ocean behavior — iffy at beFor example, due to the lack of ocean data, secondary data is often used to infer what the ocean is doing — thus, the AMO analysis relies not on ocean temperature measurements, but rather on air pressure measurements as a proxy for ocean behavior — iffy at befor ocean behavior — iffy at best.
Mike's work, like that of previous award winners, is diverse, and includes pioneering and highly cited work in time series analysis (an elegant use of Thomson's multitaper spectral analysis approach to detect spatiotemporal oscillations in the climate record and methods for smoothing temporal data), decadal climate variability (the term «Atlantic Multidecadal Oscillation» or «AMO» was coined by Mike in an interview with Science's Richard Kerr about a paper he had published with Tom Delworth of GFDL showing evidence in both climate model simulations and observational data for a 50 - 70 year oscillation in the climate system; significantly Mike also published work with Kerry Emanuel in 2006 showing that the AMO concept has been overstated as regards its role in 20th century tropical Atlantic SST changes, a finding recently reaffirmed by a study published in Nature), in showing how changes in radiative forcing from volcanoes can affect ENSO, in examining the role of solar variations in explaining the pattern of the Medieval Climate Anomaly and Little Ice Age, the relationship between the climate changes of past centuries and phenomena such as Atlantic tropical cyclones and global sea level, and even a bit of work in atmospheric chemistry (an analysis of beryllium - 7 measurements).
[Response: That «modeller» is me (I don't like that label, as I've done sea - going measurements and published papers on data analysis and theory — my topic is climate, and models are just one tool for its investigation).
There are now improved and expanded data since the TAR, including, for example, measurements at a larger number of sites, improved analysis of borehole temperature data and more extensive analyses of glaciers, corals and sediments.
Even an analysis of the measurements would have to be peer reviewed to be taken seriously, and isn't the presentation of underlying data one of the requirements for a serious peer review?
The use of even more recently computer - reconstructed total solar irradiance data (whatever have large uncertainties) for the period prior to 1976 would not change any of the conclusions in my paper, where quantitative analyses were emphasized on the influences of humans and the Sun on global surface temperature after 1970 when direct measurements became available.
Because the GISS analysis combines available sea surface temperature records with meteorological station measurements, we test alternative choices for the ocean data, showing that global temperature change is sensitive to estimated temperature change in polar regions where observations are limited.
So, the next step will be to revisit the SST bias corrections and refine them, making use of the new information uncovered concerning national measurement practices and new analysis techniques that allow for more accurate corrections in areas and at times where there are few data.
Such an assessment should involve a detailed analysis of the sensitivity of global - mean temperatures derived from these three different measurement systems to the various choices made in the processing of the raw data — e.g., corrections for instrument changes, adjustments for orbital decay effects in the satellite measurements, and procedures for interpolating station data onto grids.
Monthly analyses of rain - gauge data, including from GPCC, are used in addition to satellite data over land, though with adjustments for the biases of the rain - gauge measurements.
The MEDEA program (Measurements of Earth Data for Environmental Analysis), created in 1992 to advise the government on environmental surveillance, was subsequently shut down by the Bush Administration.
At the time, Esper had refused to provide the measurement data or chronologies for Esper et al 2002 and it was pretty much impossible to try to develop the analysis further.
The QCRAD methodology uses climatological analyses of the surface radiation measurements to define reasonable limits for testing the data for unusual data values.
However, early radiosonde sensors suffered from significant measurement biases, particularly for the upper troposphere, and changes in instrumentation with time often lead to artificial discontinuities in the data record... Consequently, most of the analysis of radiosonde humidity has focused on trends for altitudes below 500 hPa and is restricted to those stations and periods for which stable instrumentation and reliable moisture soundings are available.
Yes, there is data analysis involved and there are uncertainties... but if you want to call them «guestimates» for this reason, you'd have to call basically every measurement in the universe a guestimate.
Uncertainties should decrease closer to near - current dates (e.g. from denser and more accurate sampling)-- but note that these products also employ different QC and analysis methods, rely to varying degrees on satellite data, on sea - ice data to constrain polar SST, and on bias adjustments for historical changes in measurement methods.
The breakout session considered requirements for CDRs (particularly in contrast to EDR retrievals) and the adequacy of current (post-Nunn-McCurdy) plans for prelaunch instrument calibration and characterization; on - orbit calibration and validation; measurement overlap and replenishment requirements; and data storage, archiving, distribution, reprocessing, analysis, and interpretation concerns.
While we noted that the WWII records for 1942 - 45 appeared to be dominated by engine - warmed intake data (a point common to all analyses), because 90 % of SST measurements with known provenance in 1970 were bucket rather than engine inlet, we postulated that «business as usual» had resumed after the war, with a resulting preponderance of bucket, rather than inlet data.
In addition, the analysis requires a data set for ocean surface temperature measurements in the presatellite era.
the measurement, collection, analysis and reporting of Internet data for the purposes of understanding and optimizing Web usage.
Site owners that want more detailed (URL level resolution) analysis and insight into their site performance and can use the same APIs to gather detailed real user measurement (RUM) data for their own origins.
Tags for this Online Resume: Atlanta, Alpharetta, Learning, Training, Development, Adult Learning, Employee Performance, Performance Improvement, Training & Development, employee education, Program Management, Project Management, Learning Assessments, Measurement & Reporting, Personal & Team Coaching, Curriculum Development, Learning Management Systems (LMS), Data Analysis, Work Skill Development, Strategic Planning
Tags for this Online Resume: Six Sigma, Process Improvement, Quality Assurance, Total Quality Management, Statistical Process Control, training, Theory of Constraints, Critical Chain Project Management, Design of Experiments, Seven Managment Tools, 5S, Quality Function Deployment, process standardization, data collection and analysis, Failure Modes and Effects Analysis (FMEA), capability studies, lean manufacturing, best practices, ISO - 9001, Microsoft Access, Microsoft Project, Microsoft Excel / VBA, Measurement systems analysis, Supplier Quality Management, Blueprintanalysis, Failure Modes and Effects Analysis (FMEA), capability studies, lean manufacturing, best practices, ISO - 9001, Microsoft Access, Microsoft Project, Microsoft Excel / VBA, Measurement systems analysis, Supplier Quality Management, BlueprintAnalysis (FMEA), capability studies, lean manufacturing, best practices, ISO - 9001, Microsoft Access, Microsoft Project, Microsoft Excel / VBA, Measurement systems analysis, Supplier Quality Management, Blueprintanalysis, Supplier Quality Management, Blueprint reading
Accountable for the analysis of statistical test and measurement data.
Introductory course in research methodology for the social sciences: formulation of a research problem, design, sampling, data collection, measurement, data analysis, interpretation, and writing the research report.
Blood samples were collected for analysis of a lipid profile and data recorded included blood pressure measurements, weight, height and waist and hip circumference.
The choice of multiple t - tests for analysis of the outcome data is considered appropriate, although a simple two - way analysis of variance would have been a better choice and could have examined simultaneously the effects of the treatments on outcome measures and the differences between pre - and post-treatment measurements.
She also has provided technical assistance and conducted training for a variety of different audiences on such topics as evaluation design, data analysis, and using data for program improvement, with a special focus on identification and measurement of child and family outcomes.
This measurement model fit the data well, χ2 (3, N = 119) = 3.41, p =.33, CFI =.99, and RMSEA =.03, and was used in the structural analysis for mediation.
This comparative analysis of interaction effects between two social contexts (parent connectedness and supportive friendships) is an established way to operationalize Bronfenbrenner's conceptualization of the mesosystem (Bronfenbrenner, 1977; Bronfenbrenner & Crouter, 1983), and the inclusion of supportive relationship data from both waves of measurement allows for the assessment of stability versus change within the family — peer mesosystem (Bronfenbrenner & Morris, 1998).
a b c d e f g h i j k l m n o p q r s t u v w x y z