Search results for: measurement errors
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3521

Search results for: measurement errors

1991 Unconventional Dating of Old Peepal Tree of Chandigarh (India) Using Optically Stimulated Luminescence

Authors: Rita Rani, Ramesh Kumar

Abstract:

The intend of the current study is to date an old grand Peepal tree that is still alive. The tree is situated in Kalibard village, Sector 9, Chandigarh (India). Due to its huge structure, it has got the status of ‘Heritage tree.’ Optically Stimulated Luminescence of sediments beneath the roots is used to determine the age of the tree. Optical dating is preferred over conventional dating methods due to more precession. The methodology includes OSL of quartz grain using SAR protocol for accumulated dose measurement. The age determination of an alive tree using sedimentary quartz is in close agreement with the approximated age provided by the related agency. This is the first attempt at using optically stimulated luminescence in the age determination of alive trees in this region. The study concludes that the Luminescence dating of alive trees is the nondestructive and more precise method.

Keywords: luminescence, dose rate, optical dating, sediments

Procedia PDF Downloads 176
1990 Features for Measuring Credibility on Facebook Information

Authors: Kanda Runapongsa Saikaew, Chaluemwut Noyunsan

Abstract:

Nowadays social media information, such as news, links, images, or VDOs, is shared extensively. However, the effectiveness of disseminating information through social media lacks in quality: less fact checking, more biases, and several rumors. Many researchers have investigated about credibility on Twitter, but there is no the research report about credibility information on Facebook. This paper proposes features for measuring credibility on Facebook information. We developed the system for credibility on Facebook. First, we have developed FB credibility evaluator for measuring credibility of each post by manual human’s labelling. We then collected the training data for creating a model using Support Vector Machine (SVM). Secondly, we developed a chrome extension of FB credibility for Facebook users to evaluate the credibility of each post. Based on the usage analysis of our FB credibility chrome extension, about 81% of users’ responses agree with suggested credibility automatically computed by the proposed system.

Keywords: facebook, social media, credibility measurement, internet

Procedia PDF Downloads 356
1989 Numerical Study of Fiber Bragg Grating Sensor: Longitudinal and Transverse Detection of Temperature and Strain

Authors: K. Khelil, H. Ammar, K. Saouchi

Abstract:

Fiber Bragg Grating (FBG) structure is an periodically modulated optical fiber. It acts as a selective filter of wavelength whose reflected peak is called Bragg wavelength and it depends on the period of the fiber and the refractive index. The simulation of FBG is based on solving the Coupled Mode Theory equation by using the Transfer Matrix Method which is carried out using MATLAB. It is found that spectral reflectivity is shifted when the change of temperature and strain is uniform. Under non-uniform temperature or strain perturbation, the spectrum is both shifted and destroyed. In case of transverse loading, reflectivity spectrum is split into two peaks, the first is specific to X axis, and the second belongs to Y axis. FBGs are used in civil engineering to detect perturbations applied to buildings.

Keywords: Bragg wavelength, coupled mode theory, optical fiber, temperature measurement

Procedia PDF Downloads 494
1988 Breakdown Voltage Measurement of High Voltage Transformers Oils Using an Active Microwave Resonator Sensor

Authors: Ahmed A. Al-Mudhafar, Ali A. Abduljabar, Hayder Jawad Albattat

Abstract:

This work suggests a new microwave resonator sensor (MRS) device for measuring the oil’s breakdown voltage of high voltage transformers. A precise high-sensitivity sensor is designed and manufactured based on a microstrip split ring resonator (SRR). To improve the sensor sensitivity, a RF amplifier of 30 dB gain is linked through a transmission line of 50Ω.The sensor operates at a microwave band (L) with a quality factor of 1.35x105 when it is loaded with an empty tube. In this work, the sensor has been tested with three samples of high voltage transformer oil of different ages (new, middle, and damaged) where the quality factor differs with each sample. A mathematical model was built to calculate the breakdown voltage of the transformer oils and the accuracy of the results was higher than 90%.

Keywords: active resonator sensor, oil breakdown voltage, transformers oils, quality factor

Procedia PDF Downloads 269
1987 Hydrothermal Synthesis of Mesoporous Carbon Nanospheres and Their Electrochemical Properties for Glucose Detection

Authors: Ali Akbar Kazemi Asl, Mansour Rahsepar

Abstract:

Mesoporous carbon nanospheres (MCNs) with uniform particle size distribution having an average of 290 nm and large specific surface area (274.4 m²/g) were synthesized by a one-step hydrothermal method followed by the calcination process and then utilized as an enzyme-free glucose biosensor. Morphology, crystal structure, and porous nature of the synthesized nanospheres were characterized by scanning electron microscopy (SEM), X-Ray diffraction (XRD), and Brunauer–Emmett–Teller (BET) analysis, respectively. Also, the electrochemical performance of the MCNs@GCE electrode for the measurement of glucose concentration in alkaline media was investigated by electrochemical impedance spectroscopy (EIS), cyclic voltammetry (CV), and chronoamperometry (CA). MCNs@GCE electrode shows good sensing performance, including a rapid glucose oxidation response within 3.1 s, a wide linear range of 0.026-12 mM, a sensitivity of 212.34 μA.mM⁻¹.cm⁻², and a detection limit of 25.7 μM with excellent selectivity.

Keywords: biosensor, electrochemical, glucose, mesoporous carbon, non-enzymatic

Procedia PDF Downloads 190
1986 Quality of Service of Transportation Networks: A Hybrid Measurement of Travel Time and Reliability

Authors: Chin-Chia Jane

Abstract:

In a transportation network, travel time refers to the transmission time from source node to destination node, whereas reliability refers to the probability of a successful connection from source node to destination node. With an increasing emphasis on quality of service (QoS), both performance indexes are significant in the design and analysis of transportation systems. In this work, we extend the well-known flow network model for transportation networks so that travel time and reliability are integrated into the QoS measurement simultaneously. In the extended model, in addition to the general arc capacities, each intermediate node has a time weight which is the travel time for per unit of commodity going through the node. Meanwhile, arcs and nodes are treated as binary random variables that switch between operation and failure with associated probabilities. For pre-specified travel time limitation and demand requirement, the QoS of a transportation network is the probability that source can successfully transport the demand requirement to destination while the total transmission time is under the travel time limitation. This work is pioneering, since existing literatures that evaluate travel time reliability via a single optimization path, the proposed QoS focuses the performance of the whole network system. To compute the QoS of transportation networks, we first transfer the extended network model into an equivalent min-cost max-flow network model. In the transferred network, each arc has a new travel time weight which takes value 0. Each intermediate node is replaced by two nodes u and v, and an arc directed from u to v. The newly generated nodes u and v are perfect nodes. The new direct arc has three weights: travel time, capacity, and operation probability. Then the universal set of state vectors is recursively decomposed into disjoint subsets of reliable, unreliable, and stochastic vectors until no stochastic vector is left. The decomposition is made possible by applying existing efficient min-cost max-flow algorithm. Because the reliable subsets are disjoint, QoS can be obtained directly by summing the probabilities of these reliable subsets. Computational experiments are conducted on a benchmark network which has 11 nodes and 21 arcs. Five travel time limitations and five demand requirements are set to compute the QoS value. To make a comparison, we test the exhaustive complete enumeration method. Computational results reveal the proposed algorithm is much more efficient than the complete enumeration method. In this work, a transportation network is analyzed by an extended flow network model where each arc has a fixed capacity, each intermediate node has a time weight, and both arcs and nodes are independent binary random variables. The quality of service of the transportation network is an integration of customer demands, travel time, and the probability of connection. We present a decomposition algorithm to compute the QoS efficiently. Computational experiments conducted on a prototype network show that the proposed algorithm is superior to existing complete enumeration methods.

Keywords: quality of service, reliability, transportation network, travel time

Procedia PDF Downloads 221
1985 Stress Intensity Factor for Dynamic Cracking of Composite Material by X-FEM Method

Authors: S. Lecheb, A. Nour, A. Chellil, H. Mechakra, N. Hamad, H. Kebir

Abstract:

The work involves develops attended by a numerical execution of the eXtend Finite Element Method premises a measurement by the fracture process cracked so many cracked plates an application will be processed for the calculation of the stress intensity factor SIF. In the first we give in statically part the distribution of stress, displacement field and strain of composite plate in two cases uncrack/edge crack, also in dynamical part the first six modes shape. Secondly, we calculate Stress Intensity Factor SIF for different orientation angle θ of central crack with length (2a=0.4mm) in plan strain condition, KI and KII are obtained for mode I and mode II respectively using X-FEM method. Finally from crack inclined involving mixed modes results, the comparison we chose dangerous inclination and the best crack angle when K is minimal.

Keywords: stress intensity factor (SIF), crack orientation, glass/epoxy, natural frequencies, X-FEM

Procedia PDF Downloads 515
1984 Characterization of current–voltage (I–V) and capacitance–voltage–frequency (C–V–f) features of Au/GaN Schottky diodes

Authors: Abdelaziz Rabehi

Abstract:

The current–voltage (I–V) characteristics of Au/GaN Schottky diodes were measured at room temperature. In addition, capacitance–voltage–frequency (C–V–f) characteristics are investigated by considering the interface states (Nss) at frequency range 100 kHz to 1 MHz. From the I–V characteristics of the Schottky diode, ideality factor (n) and barrier height (Φb) values of 1.22 and 0.56 eV, respectively, were obtained from a forward bias I–V plot. In addition, the interface states distribution profile as a function of (Ess − Ev) was extracted from the forward bias I–V measurements by taking into account the bias dependence of the effective barrier height (Φe) for the Schottky diode. The C–V curves gave a barrier height value higher than those obtained from I–V measurements. This discrepancy is due to the different nature of the I–V and C–V measurement techniques.

Keywords: Schottky diodes, frequency dependence, barrier height, interface states

Procedia PDF Downloads 302
1983 Automated User Story Driven Approach for Web-Based Functional Testing

Authors: Mahawish Masud, Muhammad Iqbal, M. U. Khan, Farooque Azam

Abstract:

Manual writing of test cases from functional requirements is a time-consuming task. Such test cases are not only difficult to write but are also challenging to maintain. Test cases can be drawn from the functional requirements that are expressed in natural language. However, manual test case generation is inefficient and subject to errors.  In this paper, we have presented a systematic procedure that could automatically derive test cases from user stories. The user stories are specified in a restricted natural language using a well-defined template.  We have also presented a detailed methodology for writing our test ready user stories. Our tool “Test-o-Matic” automatically generates the test cases by processing the restricted user stories. The generated test cases are executed by using open source Selenium IDE.  We evaluate our approach on a case study, which is an open source web based application. Effectiveness of our approach is evaluated by seeding faults in the open source case study using known mutation operators.  Results show that the test case generation from restricted user stories is a viable approach for automated testing of web applications.

Keywords: automated testing, natural language, restricted user story modeling, software engineering, software testing, test case specification, transformation and automation, user story, web application testing

Procedia PDF Downloads 387
1982 Efficiency Measurement of Indian Sugar Manufacturing Firms - a DEA Approach

Authors: Amit Kumar Dwivedi, Priyanko Ghosh

Abstract:

Data Envelopment analysis (DEA) has been used to calculate the technical and scale efficiency measures of the public and private sugar manufacturing firms of the Indian Sugar Industry (2006 to 2010). Within DEA framework, the input & Output oriented Variable Returns to Scale (VRS) & Constant Return to Scale (CRS) model is employed for the study of Decision making units (DMUs). A representative sample of 43 firms which account for major portion of the total market share is studied. The selection criterion for the inclusion of a firm in the analysis was the total sales of INR 5,000 million or more in the year 2010. After reviewing the literature it is found that no study has been conducted in the context of Indian sugar manufacturing firms in the Post-liberalization era which motivates us to initiate the study.

Keywords: technical efficiency, Indian sugar manufacturing units, DEA, input output oriented

Procedia PDF Downloads 542
1981 Automatic Reporting System for Transcriptome Indel Identification and Annotation Based on Snapshot of Next-Generation Sequencing Reads Alignment

Authors: Shuo Mu, Guangzhi Jiang, Jinsa Chen

Abstract:

The analysis of Indel for RNA sequencing of clinical samples is easily affected by sequencing experiment errors and software selection. In order to improve the efficiency and accuracy of analysis, we developed an automatic reporting system for Indel recognition and annotation based on image snapshot of transcriptome reads alignment. This system includes sequence local-assembly and realignment, target point snapshot, and image-based recognition processes. We integrated high-confidence Indel dataset from several known databases as a training set to improve the accuracy of image processing and added a bioinformatical processing module to annotate and filter Indel artifacts. Subsequently, the system will automatically generate data, including data quality levels and images results report. Sanger sequencing verification of the reference Indel mutation of cell line NA12878 showed that the process can achieve 83% sensitivity and 96% specificity. Analysis of the collected clinical samples showed that the interpretation accuracy of the process was equivalent to that of manual inspection, and the processing efficiency showed a significant improvement. This work shows the feasibility of accurate Indel analysis of clinical next-generation sequencing (NGS) transcriptome. This result may be useful for RNA study for clinical samples with microsatellite instability in immunotherapy in the future.

Keywords: automatic reporting, indel, next-generation sequencing, NGS, transcriptome

Procedia PDF Downloads 191
1980 Easymodel: Web-based Bioinformatics Software for Protein Modeling Based on Modeller

Authors: Alireza Dantism

Abstract:

Presently, describing the function of a protein sequence is one of the most common problems in biology. Usually, this problem can be facilitated by studying the three-dimensional structure of proteins. In the absence of a protein structure, comparative modeling often provides a useful three-dimensional model of the protein that is dependent on at least one known protein structure. Comparative modeling predicts the three-dimensional structure of a given protein sequence (target) mainly based on its alignment with one or more proteins of known structure (templates). Comparative modeling consists of four main steps 1. Similarity between the target sequence and at least one known template structure 2. Alignment of target sequence and template(s) 3. Build a model based on alignment with the selected template(s). 4. Prediction of model errors 5. Optimization of the built model There are many computer programs and web servers that automate the comparative modeling process. One of the most important advantages of these servers is that it makes comparative modeling available to both experts and non-experts, and they can easily do their own modeling without the need for programming knowledge, but some other experts prefer using programming knowledge and do their modeling manually because by doing this they can maximize the accuracy of their modeling. In this study, a web-based tool has been designed to predict the tertiary structure of proteins using PHP and Python programming languages. This tool is called EasyModel. EasyModel can receive, according to the user's inputs, the desired unknown sequence (which we know as the target) in this study, the protein sequence file (template), etc., which also has a percentage of similarity with the primary sequence, and its third structure Predict the unknown sequence and present the results in the form of graphs and constructed protein files.

Keywords: structural bioinformatics, protein tertiary structure prediction, modeling, comparative modeling, modeller

Procedia PDF Downloads 97
1979 Machine Learning Models for the Prediction of Heating and Cooling Loads of a Residential Building

Authors: Aaditya U. Jhamb

Abstract:

Due to the current energy crisis that many countries are battling, energy-efficient buildings are the subject of extensive research in the modern technological era because of growing worries about energy consumption and its effects on the environment. The paper explores 8 factors that help determine energy efficiency for a building: (relative compactness, surface area, wall area, roof area, overall height, orientation, glazing area, and glazing area distribution), with Tsanas and Xifara providing a dataset. The data set employed 768 different residential building models to anticipate heating and cooling loads with a low mean squared error. By optimizing these characteristics, machine learning algorithms may assess and properly forecast a building's heating and cooling loads, lowering energy usage while increasing the quality of people's lives. As a result, the paper studied the magnitude of the correlation between these input factors and the two output variables using various statistical methods of analysis after determining which input variable was most closely associated with the output loads. The most conclusive model was the Decision Tree Regressor, which had a mean squared error of 0.258, whilst the least definitive model was the Isotonic Regressor, which had a mean squared error of 21.68. This paper also investigated the KNN Regressor and the Linear Regression, which had to mean squared errors of 3.349 and 18.141, respectively. In conclusion, the model, given the 8 input variables, was able to predict the heating and cooling loads of a residential building accurately and precisely.

Keywords: energy efficient buildings, heating load, cooling load, machine learning models

Procedia PDF Downloads 96
1978 Identifying Organizational Culture to Implement Knowledge Management: Case Study of BKN, Indonesia

Authors: Maria Margaretha, Elin Cahyaningsih, Dana Indra Sensuse Lukman

Abstract:

One of key success an organization can be seen from its culture. Employee, environment, and so on are factors for organization to achieve goals and build a competitive advantage. Type of organizational culture can be a guide to implementing Knowledge Management (KM) in organization especially in BKN. Culture will determine behavior of employees or environment to support KM. This paper describes the process to decide which culture does organization belong and suggestion and creating strategic moves in the future to implement KM. OCAI (Organizational Culture Assessment Instrument) and its framework (Competing Value Framework) were used to decide the type of organizational culture. To implement KM in organization, clan is an appropriate culture, because clan culture represent cultural values and leader type to implement a successful KM. Result of the measurement will be references for BKN to improve organization culture to achieve its goals and organization effectiveness.

Keywords: organizational culture, government, knowledge management, OCAI

Procedia PDF Downloads 621
1977 Estimation of Natural Convection Heat Transfer from Plate-Fin Heat Sinks in a Closed Enclosure

Authors: Han-Taw Chen, Chung-Hou Lai, Tzu-Hsiang Lin, Ge-Jang He

Abstract:

This study applies the inverse method and three-dimensional CFD commercial software in conjunction with the experimental temperature data to investigate the heat transfer and fluid flow characteristics of the plate-fin heat sink in a closed rectangular enclosure for various values of fin height. The inverse method with the finite difference method and the experimental temperature data is applied to determine the heat transfer coefficient. The k-ε turbulence model is used to obtain the heat transfer and fluid flow characteristics within the fins. To validate the accuracy of the results obtained, the comparison of the average heat transfer coefficient is made. The calculated temperature at selected measurement locations on the plate-fin is also compared with experimental data.

Keywords: inverse method, FLUENT, k-ε model, heat transfer characteristics, plate-fin heat sink

Procedia PDF Downloads 460
1976 Recombination Center Levels in Gold and Platinum Doped N-Type Silicon

Authors: Nam Chol Yu, Kyong Il Chu

Abstract:

Using DLTS measurement techniques, we determined the dominant recombination center levels (defects of both A and B) in gold and platinum doped n-type silicon. Also, the injection and temperature dependence of the Shockley-Read-Hall (SRH) carrier lifetime was studied under low-level injection and high-level injection. Here measurements show that the dominant level under low-level injection located at EC-0.25eV(A) correlated to the Pt+G1 and the dominant level under high-level injection located at EC-0.54eV(B) correlated to the Au+G4. Finally, A and B are the same dominant levels for controlling the lifetime in gold-platinum doped n-silicon.

Keywords: recombination center level, lifetime, carrier lifetime control, gold, platinum, silicon

Procedia PDF Downloads 155
1975 Challenges in Translating Malay Idiomatic Expressions: A Study

Authors: Nor Ruba’Yah Binti Abd Rahim, Norsyahidah Binti Jaafar

Abstract:

Translating Malay idiomatic expressions into other languages presents unique challenges due to the deep cultural nuances and linguistic intricacies embedded within these expressions. This study examined these challenges through a two-pronged methodology: a comparative analysis using survey questionnaires and a quiz administered to 50 semester 6 students who are taking Translation 1 course, and in-depth interviews with their lecturers. The survey aimed to capture students’ experiences and difficulties in translating selected Malay idioms into English, highlighting common errors and misunderstandings. Complementing this, interviews with lecturers provided expert insights into the nuances of these expressions and effective translation strategies. The findings revealed that literal translations often fail to convey the intended meanings, underscoring the importance of cultural competence and contextual awareness. The study also identified key factors that contribute to successful translations, such as the translator’s familiarity with both source and target cultures and their ability to adapt expressions creatively. This research contributed to the field of translation studies by offering practical recommendations for improving the translation of idiomatic expressions, thereby enhancing cross-cultural communication. The insights gained from this study are valuable for translators, educators, and students, emphasizing the need for a nuanced approach that respects the cultural richness of the source language while ensuring clarity in the target language.

Keywords: idiomatic expressions, cultural competence, translation strategies, cross-cultural communication, students’ difficulties

Procedia PDF Downloads 12
1974 Harmonic Distortion Caused by Electric Bus Battery Charger in Alexandria Distribution System

Authors: Mohamed Elhosieny Aly Ismail

Abstract:

The paper illustrates the total voltage and current harmonic distortion impact caused by fast-charging an electric bus and maintaining standard limit compliance. Measuring the current harmonic level in the range of 2 kHz-9 kHz. Also, the impact of the total demand distortions current caused by fast charger electric bus on the utility by measuring at the point of common coupling and comparing the measurement with IEEE519 -2014 standard. The results show that the total harmonic current distortion for the charger is within the limits of IEC 61000-3-12 and the fifth harmonic current was the most dominant frequency then the seventh harmonic current. The harmonic current in the range of 2 kHz- 9 kHz shows the frequency 5.1kHz is the most dominant frequency.

Keywords: electric vehicle, total harmonic distortion, IEEE519-2014, IEC 61000-3-12, super harmonic distortion

Procedia PDF Downloads 101
1973 A Fabrication Method for PEDOT: PSS Based Humidity Sensor

Authors: Nazia Tarannum, M. Ayaz Ahmad

Abstract:

The main goal of this article is to report some interesting features for the fabrication/design of PEDOT:PSS based humidity sensor. Here first we fabricated humidity sensor and then studied its electro-mechanical characteristics. In general the humidity plays an important role in various private and government sectors all over the world. Monitoring and controlling the humidity is a great task for the reliable operation of various systems. The PEDOT:PSS is very much promising humidity sensor and also is fabricated by performing various analyses. The interdigited electrode (IDE) has channel length 200 microns prepared by lithography. Lithography of IDE was done on PPR coated glass substrate using negative mask and exposing it with UV light for 10 secs via DSA. During the above said fabrication, we have taken account for the following steps: •Plasma ashing of IDE •Spincoating of PEDOT:PSS was done @3000 rpm on IDE substrace •Baked the substrace at 130 °C up to time limit 15 mins. •Resistance measurement using Labtracer 2.9 software via Keithley 2400source meter.

Keywords: fabrication method, PEDOT:PSS material, humidity sensor, electro-mechanical

Procedia PDF Downloads 350
1972 Artificial Intelligence in the Design of a Retaining Structure

Authors: Kelvin Lo

Abstract:

Nowadays, numerical modelling in geotechnical engineering is very common but sophisticated. Many advanced input settings and considerable computational efforts are required to optimize the design to reduce the construction cost. To optimize a design, it usually requires huge numerical models. If the optimization is conducted manually, there is a potentially dangerous consequence from human errors, and the time spent on the input and data extraction from output is significant. This paper presents an automation process introduced to numerical modelling (Plaxis 2D) of a trench excavation supported by a secant-pile retaining structure for a top-down tunnel project. Python code is adopted to control the process, and numerical modelling is conducted automatically in every 20m chainage along the 200m tunnel, with maximum retained height occurring in the middle chainage. Python code continuously changes the geological stratum and excavation depth under groundwater flow conditions in each 20m section. It automatically conducts trial and error to determine the required pile length and the use of props to achieve the required factor of safety and target displacement. Once the bending moment of the pile exceeds its capacity, it will increase in size. When the pile embedment reaches the default maximum length, it will turn on the prop system. Results showed that it saves time, increases efficiency, lowers design costs, and replaces human labor to minimize error.

Keywords: automation, numerical modelling, Python, retaining structures

Procedia PDF Downloads 51
1971 Designing Automated Embedded Assessment to Assess Student Learning in a 3D Educational Video Game

Authors: Mehmet Oren, Susan Pedersen, Sevket C. Cetin

Abstract:

Despite the frequently criticized disadvantages of the traditional used paper and pencil assessment, it is the most frequently used method in our schools. Although assessments do an acceptable measurement, they are not capable of measuring all the aspects and the richness of learning and knowledge. Also, many assessments used in schools decontextualize the assessment from the learning, and they focus on learners’ standing on a particular topic but do not concentrate on how student learning changes over time. For these reasons, many scholars advocate that using simulations and games (S&G) as a tool for assessment has significant potentials to overcome the problems in traditionally used methods. S&G can benefit from the change in technology and provide a contextualized medium for assessment and teaching. Furthermore, S&G can serve as an instructional tool rather than a method to test students’ learning at a particular time point. To investigate the potentials of using educational games as an assessment and teaching tool, this study presents the implementation and the validation of an automated embedded assessment (AEA), which can constantly monitor student learning in the game and assess their performance without intervening their learning. The experiment was conducted on an undergraduate level engineering course (Digital Circuit Design) with 99 participant students over a period of five weeks in Spring 2016 school semester. The purpose of this research study is to examine if the proposed method of AEA is valid to assess student learning in a 3D Educational game and present the implementation steps. To address this question, this study inspects three aspects of the AEA for the validation. First, the evidence-centered design model was used to lay out the design and measurement steps of the assessment. Then, a confirmatory factor analysis was conducted to test if the assessment can measure the targeted latent constructs. Finally, the scores of the assessment were compared with an external measure (a validated test measuring student learning on digital circuit design) to evaluate the convergent validity of the assessment. The results of the confirmatory factor analysis showed that the fit of the model with three latent factors with one higher order factor was acceptable (RMSEA < 0.00, CFI =1, TLI=1.013, WRMR=0.390). All of the observed variables significantly loaded to the latent factors in the latent factor model. In the second analysis, a multiple regression analysis was used to test if the external measure significantly predicts students’ performance in the game. The results of the regression indicated the two predictors explained 36.3% of the variance (R2=.36, F(2,96)=27.42.56, p<.00). It was found that students’ posttest scores significantly predicted game performance (β = .60, p < .000). The statistical results of the analyses show that the AEA can distinctly measure three major components of the digital circuit design course. It was aimed that this study can help researchers understand how to design an AEA, and showcase an implementation by providing an example methodology to validate this type of assessment.

Keywords: educational video games, automated embedded assessment, assessment validation, game-based assessment, assessment design

Procedia PDF Downloads 421
1970 Assessing the Structure of Non-Verbal Semantic Knowledge: The Evaluation and First Results of the Hungarian Semantic Association Test

Authors: Alinka Molnár-Tóth, Tímea Tánczos, Regina Barna, Katalin Jakab, Péter Klivényi

Abstract:

Supported by neuroscientific findings, the so-called Hub-and-Spoke model of the human semantic system is based on two subcomponents of semantic cognition, namely the semantic control process and semantic representation. Our semantic knowledge is multimodal in nature, as the knowledge system stored in relation to a conception is extensive and broad, while different aspects of the conception may be relevant depending on the purpose. The motivation of our research is to develop a new diagnostic measurement procedure based on the preservation of semantic representation, which is appropriate to the specificities of the Hungarian language and which can be used to compare the non-verbal semantic knowledge of healthy and aphasic persons. The development of the test will broaden the Hungarian clinical diagnostic toolkit, which will allow for more specific therapy planning. The sample of healthy persons (n=480) was determined by the last census data for the representativeness of the sample. Based on the concept of the Pyramids and Palm Tree Test, and according to the characteristics of the Hungarian language, we have elaborated a test based on different types of semantic information, in which the subjects are presented with three pictures: they have to choose the one that best fits the target word above from the two lower options, based on the semantic relation defined. We have measured 5 types of semantic knowledge representations: associative relations, taxonomy, motional representations, concrete as well as abstract verbs. As the first step in our data analysis, we examined the normal distribution of our results, and since it was not normally distributed (p < 0.05), we used nonparametric statistics further into the analysis. Using descriptive statistics, we could determine the frequency of the correct and incorrect responses, and with this knowledge, we could later adjust and remove the items of questionable reliability. The reliability was tested using Cronbach’s α, and it can be safely said that all the results were in an acceptable range of reliability (α = 0.6-0.8). We then tested for the potential gender differences using the Mann Whitney-U test, however, we found no difference between the two (p < 0.05). Likewise, we didn’t see that the age had any effect on the results using one-way ANOVA (p < 0.05), however, the level of education did influence the results (p > 0.05). The relationships between the subtests were observed by the nonparametric Spearman’s rho correlation matrix, showing statistically significant correlation between the subtests (p > 0.05), signifying a linear relationship between the measured semantic functions. A margin of error of 5% was used in all cases. The research will contribute to the expansion of the clinical diagnostic toolkit and will be relevant for the individualised therapeutic design of treatment procedures. The use of a non-verbal test procedure will allow an early assessment of the most severe language conditions, which is a priority in the differential diagnosis. The measurement of reaction time is expected to advance prodrome research, as the tests can be easily conducted in the subclinical phase.

Keywords: communication disorders, diagnostic toolkit, neurorehabilitation, semantic knowlegde

Procedia PDF Downloads 103
1969 Analytical Development of a Failure Limit and Iso-Uplift Curves for Eccentrically Loaded Shallow Foundations

Authors: N. Abbas, S. Lagomarsino, S. Cattari

Abstract:

Examining existing experimental results for shallow rigid foundations subjected to vertical centric load (N), accompanied or not with a bending moment (M), two main non-linear mechanisms governing the cyclic ‎response of the soil-foundation system can be distinguished: foundation uplift and soil yielding. A soil-foundation failure limit, is defined as a domain of resistance in the two dimensional (2D) load space (N, M) inside of which lie all the admissible combinations of loads; these latter correspond to a pure elastic, non-linear elastic or plastic behavior of the soil-foundation system, while the points lying on the failure limit correspond to a combination of loads leading to a failure of the soil-foundation system. In this study, the proposed resistance domain is constructed analytically based on mechanics. Original elastic limit, uplift initiation ‎limit and iso-uplift limits are constructed inside this domain. These limits give a prediction ‎of the mechanisms activated for each combination of loads applied to the ‎foundation. A comparison of the proposed failure limit with experimental tests existing in the literature shows interesting results. Also, the developed uplift initiation limit and iso-uplift curves are confronted with others already proposed in the literature and widely used due to the absence of other alternatives, and remarkable differences are noted, showing evident errors in the past proposals and relevant accuracy for those given in the present work.

Keywords: foundation uplift, iso-uplift curves, resistance domain, soil yield

Procedia PDF Downloads 383
1968 Separation of Fexofenadine Enantiomers Using Beta Cyclodextrin as Chiral Counter Ion in Mobile Phase

Authors: R. Fegas, S. Zerkout, S. Taberkokt, M. Righezza

Abstract:

The present work demonstrate the potential of Betacyclodextrine (BCD) for the chiral analysis of a drug .Various separation mechanisms were applied and several parameters affecting the separation were studied, including the type and concentration of chiral selector, and pH of buffer. A simple and sensitive high-performance liquid chromatography (HPLC) method was developed as an assay for fexofenadine enantiomers in pharmaceutical preparation. Fexofenadine enantiomers were separated using a mobile phase of 0.25mM NaH2PO4–acetonitrile (65:35, v/v) – Betacyclodextrine on achiral phenyl-urea column at a flow rate of 1ml/min and measurement at 220nm. The chiral mechanism of separation was mainly based on specific interaction between the solute and the stationary phase. The retention was directly controlled by mobile phase composition but not the selectivity which results of the two mechanisms, electrostatic interactions and partition mechanism.

Keywords: fexofenadine enantiomer, HPLC, achiral phenyl-urea column

Procedia PDF Downloads 458
1967 Detection of Lymphedema after Breast Cancer in Yucatecan Women

Authors: Olais A. Ingrid, Peraza G. Leydi, Estrella C. Damaris

Abstract:

Breast cancer is the most common among women worldwide; the different treatments can bring sequels that directly affect the quality of life, such as lymphedema. The objective was to determine if there is presence of lymphedema secondary to breast cancer in Yucatecan women. It was an observational, analytical, cross-sectional study, 92 women were included who met the following criteria: women with surgical treatment for unilateral: breast cancer, aged between 25 and 65 years old, minimum 6 weeks after unilateral breast surgery and have completed any type of chemotherapy or adjuvant radiotherapy treatment for breast cancer. The evaluation was through indirect measurement volume by circometry to determine the presence of lymphedema. 23% of women had lymphedema grade I. It related to the presence of some of the symptoms like stiffness, swelling, decreased range of motion and feeling of heaviness in the arm of the operated side of the breast. It is important to determine the presence of lymphedema to perform physical therapy treatment.

Keywords: breast cancer, lymphedema, physical therapy, Yucatan

Procedia PDF Downloads 349
1966 PIV Measurements of the Instantaneous Velocities for Single and Two-Phase Flows in an Annular Duct

Authors: Marlon M. Hernández Cely, Victor E. C. Baptistella, Oscar M. H. Rodríguez

Abstract:

Particle Image Velocimetry (PIV) is a well-established technique in the field of fluid flow measurement and provides instantaneous velocity fields over global domains. It has been applied to external and internal flows and in single and two-phase flows. Regarding internal flow, works about the application of PIV in annular ducts are scanty. An experimental work is presented, where flow of water is studied in an annular duct of inner diameter of 60 mm and outer diameter of 155 mm and 10.5-m length, with the goal of obtaining detailed velocity measurements. Depending on the flow rates of water, it can be laminar, transitional or turbulent. In this study, the water flow rate was kept at three different values for the annular duct, allowing the analysis of one laminar and two turbulent flows. Velocity fields and statistic quantities of the turbulent flow were calculated.

Keywords: PIV, annular duct, laminar, turbulence, velocity profile

Procedia PDF Downloads 351
1965 Project Objective Structure Model: An Integrated, Systematic and Balanced Approach in Order to Achieve Project Objectives

Authors: Mohammad Reza Oftadeh

Abstract:

The purpose of the article is to describe project objective structure (POS) concept that was developed on research activities and experiences about project management, Balanced Scorecard (BSC) and European Foundation Quality Management Excellence Model (EFQM Excellence Model). Furthermore, this paper tries to define a balanced, systematic, and integrated measurement approach to meet project objectives and project strategic goals based on a process-oriented model. In this paper, POS is suggested in order to measure project performance in the project life cycle. After using the POS model, the project manager can ensure in order to achieve the project objectives on the project charter. This concept can help project managers to implement integrated and balanced monitoring and control project work.

Keywords: project objectives, project performance management, PMBOK, key performance indicators, integration management

Procedia PDF Downloads 378
1964 High-Resolution Spatiotemporal Retrievals of Aerosol Optical Depth from Geostationary Satellite Using Sara Algorithm

Authors: Muhammad Bilal, Zhongfeng Qiu

Abstract:

Aerosols, suspended particles in the atmosphere, play an important role in the earth energy budget, climate change, degradation of atmospheric visibility, urban air quality, and human health. To fully understand aerosol effects, retrieval of aerosol optical properties such as aerosol optical depth (AOD) at high spatiotemporal resolution is required. Therefore, in the present study, hourly AOD observations at 500 m resolution were retrieved from the geostationary ocean color imager (GOCI) using the simplified aerosol retrieval algorithm (SARA) over the urban area of Beijing for the year 2016. The SARA requires top-of-the-atmosphere (TOA) reflectance, solar and sensor geometry information and surface reflectance observations to retrieve an accurate AOD. For validation of the GOCI retrieved AOD, AOD measurements were obtained from the aerosol robotic network (AERONET) version 3 level 2.0 (cloud-screened and quality assured) data. The errors and uncertainties were reported using the root mean square error (RMSE), relative percent mean error (RPME), and the expected error (EE = ± (0.05 + 0.15AOD). Results showed that the high spatiotemporal GOCI AOD observations were well correlated with the AERONET AOD measurements with a correlation coefficient (R) of 0.92, RMSE of 0.07, and RPME of 5%, and 90% of the observations were within the EE. The results suggested that the SARA is robust and has the ability to retrieve high-resolution spatiotemporal AOD observations over the urban area using the geostationary satellite.

Keywords: AEORNET, AOD, SARA, GOCI, Beijing

Procedia PDF Downloads 171
1963 Ambivalence as Ethical Practice: Methodologies to Address Noise, Bias in Care, and Contact Evaluations

Authors: Anthony Townsend, Robyn Fasser

Abstract:

While complete objectivity is a desirable scientific position from which to conduct a care and contact evaluation (CCE), it is precisely the recognition that we are inherently incapable of operating objectively that is the foundation of ethical practice and skilled assessment. Drawing upon recent research from Daniel Kahneman (2021) on the differences between noise and bias, as well as different inherent biases collectively termed “The Elephant in the Brain” by Kevin Simler and Robin Hanson (2019) from Oxford University, this presentation addresses both the various ways in which our judgments, perceptions and even procedures can be distorted and contaminated while conducting a CCE, but also considers the value of second order cybernetics and the psychodynamic concept of ‘ambivalence’ as a conceptual basis to inform our assessment methodologies to limit such errors or at least better identify them. Both a conceptual framework for ambivalence, our higher-order capacity to allow for the convergence and consideration of multiple emotional experiences and cognitive perceptions to inform our reasoning, and a practical methodology for assessment relying on data triangulation, Bayesian inference and hypothesis testing is presented as a means of promoting ethical practice for health care professionals conducting CCEs. An emphasis on widening awareness and perspective, limiting ‘splitting’, is demonstrated both in how this form of emotional processing plays out in alienating dynamics in families as well as the assessment thereof. In addressing this concept, this presentation aims to illuminate the value of ambivalence as foundational to ethical practice for assessors.

Keywords: ambivalence, forensic, psychology, noise, bias, ethics

Procedia PDF Downloads 86
1962 Computer Countenanced Diagnosis of Skin Nodule Detection and Histogram Augmentation: Extracting System for Skin Cancer

Authors: S. Zith Dey Babu, S. Kour, S. Verma, C. Verma, V. Pathania, A. Agrawal, V. Chaudhary, A. Manoj Puthur, R. Goyal, A. Pal, T. Danti Dey, A. Kumar, K. Wadhwa, O. Ved

Abstract:

Background: Skin cancer is now is the buzzing button in the field of medical science. The cyst's pandemic is drastically calibrating the body and well-being of the global village. Methods: The extracted image of the skin tumor cannot be used in one way for diagnosis. The stored image contains anarchies like the center. This approach will locate the forepart of an extracted appearance of skin. Partitioning image models has been presented to sort out the disturbance in the picture. Results: After completing partitioning, feature extraction has been formed by using genetic algorithm and finally, classification can be performed between the trained and test data to evaluate a large scale of an image that helps the doctors for the right prediction. To bring the improvisation of the existing system, we have set our objectives with an analysis. The efficiency of the natural selection process and the enriching histogram is essential in that respect. To reduce the false-positive rate or output, GA is performed with its accuracy. Conclusions: The objective of this task is to bring improvisation of effectiveness. GA is accomplishing its task with perfection to bring down the invalid-positive rate or outcome. The paper's mergeable portion conflicts with the composition of deep learning and medical image processing, which provides superior accuracy. Proportional types of handling create the reusability without any errors.

Keywords: computer-aided system, detection, image segmentation, morphology

Procedia PDF Downloads 150