Search results for: simplified conceptual models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7923

Search results for: simplified conceptual models

3873 Multidimensional Item Response Theory Models for Practical Application in Large Tests Designed to Measure Multiple Constructs

Authors: Maria Fernanda Ordoñez Martinez, Alvaro Mauricio Montenegro

Abstract:

This work presents a statistical methodology for measuring and founding constructs in Latent Semantic Analysis. This approach uses the qualities of Factor Analysis in binary data with interpretations present on Item Response Theory. More precisely, we propose initially reducing dimensionality with specific use of Principal Component Analysis for the linguistic data and then, producing axes of groups made from a clustering analysis of the semantic data. This approach allows the user to give meaning to previous clusters and found the real latent structure presented by data. The methodology is applied in a set of real semantic data presenting impressive results for the coherence, speed and precision.

Keywords: semantic analysis, factorial analysis, dimension reduction, penalized logistic regression

Procedia PDF Downloads 424
3872 Scoring System for the Prognosis of Sepsis Patients in Intensive Care Units

Authors: Javier E. García-Gallo, Nelson J. Fonseca-Ruiz, John F. Duitama-Munoz

Abstract:

Sepsis is a syndrome that occurs with physiological and biochemical abnormalities induced by severe infection and carries a high mortality and morbidity, therefore the severity of its condition must be interpreted quickly. After patient admission in an intensive care unit (ICU), it is necessary to synthesize the large volume of information that is collected from patients in a value that represents the severity of their condition. Traditional severity of illness scores seeks to be applicable to all patient populations, and usually assess in-hospital mortality. However, the use of machine learning techniques and the data of a population that shares a common characteristic could lead to the development of customized mortality prediction scores with better performance. This study presents the development of a score for the one-year mortality prediction of the patients that are admitted to an ICU with a sepsis diagnosis. 5650 ICU admissions extracted from the MIMICIII database were evaluated, divided into two groups: 70% to develop the score and 30% to validate it. Comorbidities, demographics and clinical information of the first 24 hours after the ICU admission were used to develop a mortality prediction score. LASSO (least absolute shrinkage and selection operator) and SGB (Stochastic Gradient Boosting) variable importance methodologies were used to select the set of variables that make up the developed score; each of this variables was dichotomized and a cut-off point that divides the population into two groups with different mean mortalities was found; if the patient is in the group that presents a higher mortality a one is assigned to the particular variable, otherwise a zero is assigned. These binary variables are used in a logistic regression (LR) model, and its coefficients were rounded to the nearest integer. The resulting integers are the point values that make up the score when multiplied with each binary variables and summed. The one-year mortality probability was estimated using the score as the only variable in a LR model. Predictive power of the score, was evaluated using the 1695 admissions of the validation subset obtaining an area under the receiver operating characteristic curve of 0.7528, which outperforms the results obtained with Sequential Organ Failure Assessment (SOFA), Oxford Acute Severity of Illness Score (OASIS) and Simplified Acute Physiology Score II (SAPSII) scores on the same validation subset. Observed and predicted mortality rates within estimated probabilities deciles were compared graphically and found to be similar, indicating that the risk estimate obtained with the score is close to the observed mortality, it is also observed that the number of events (deaths) is indeed increasing as the outcome go from the decile with the lowest probabilities to the decile with the highest probabilities. Sepsis is a syndrome that carries a high mortality, 43.3% for the patients included in this study; therefore, tools that help clinicians to quickly and accurately predict a worse prognosis are needed. This work demonstrates the importance of customization of mortality prediction scores since the developed score provides better performance than traditional scoring systems.

Keywords: intensive care, logistic regression model, mortality prediction, sepsis, severity of illness, stochastic gradient boosting

Procedia PDF Downloads 201
3871 Quantitative Structure Activity Relationship Model for Predicting the Aromatase Inhibition Activity of 1,2,3-Triazole Derivatives

Authors: M. Ouassaf, S. Belaidi

Abstract:

Aromatase is an estrogen biosynthetic enzyme belonging to the cytochrome P450 family, which catalyzes the limiting step in the conversion of androgens to estrogens. As it is relevant for the promotion of tumor cell growth. A set of thirty 1,2,3-triazole derivatives was used in the quantitative structure activity relationship (QSAR) study using regression multiple linear (MLR), We divided the data into two training and testing groups. The results showed a good predictive ability of the MLR model, the models were statistically robust internally (R² = 0.982) and the predictability of the model was tested by several parameters. including external criteria (R²pred = 0.851, CCC = 0.946). The knowledge gained in this study should provide relevant information that contributes to the origins of aromatase inhibitory activity and, therefore, facilitates our ongoing quest for aromatase inhibitors with robust properties.

Keywords: aromatase inhibitors, QSAR, MLR, 1, 2, 3-triazole

Procedia PDF Downloads 100
3870 Assessing Readiness Model for Business Intelligence Implementation in Organization

Authors: Abdul Razak Rahmat, Azizah Ahmad, Azman Ta’aa

Abstract:

The deployment of Business Intelligence (BI) for organization at the beginning phase is very crucial. Results from the previous studies found that more than half of the BI project fails to meet the objective even though a lot money are spent. Based on that problem, the readiness level of BI for the organization is important to identify in order to reduce the risk before the actual BI project is implemented. In this paper, rigorous literature review on the aspect success factors such as Critical Success Factors (CSFs), Readiness Factors (RFs), Success Factors (SFs), are discussed by different authors. The paper also adopted a few models from previous study as a guide for the assessment of BI readiness. The expected finding from this research is the Business Intelligent Readiness Model (BiRM) as a guild before implement the BI system.

Keywords: business intelligence readiness model, business intelligence for higher learning, BI readiness factors, BI critical success factors(CSF)

Procedia PDF Downloads 355
3869 Generating Product Description with Generative Pre-Trained Transformer 2

Authors: Minh-Thuan Nguyen, Phuong-Thai Nguyen, Van-Vinh Nguyen, Quang-Minh Nguyen

Abstract:

Research on automatically generating descriptions for e-commerce products is gaining increasing attention in recent years. However, the generated descriptions of their systems are often less informative and attractive because of lacking training datasets or the limitation of these approaches, which often use templates or statistical methods. In this paper, we explore a method to generate production descriptions by using the GPT-2 model. In addition, we apply text paraphrasing and task-adaptive pretraining techniques to improve the qualify of descriptions generated from the GPT-2 model. Experiment results show that our models outperform the baseline model through automatic evaluation and human evaluation. Especially, our methods achieve a promising result not only on the seen test set but also in the unseen test set.

Keywords: GPT-2, product description, transformer, task-adaptive, language model, pretraining

Procedia PDF Downloads 186
3868 Hybrid Artificial Bee Colony and Least Squares Method for Rule-Based Systems Learning

Authors: Ahcene Habbi, Yassine Boudouaoui

Abstract:

This paper deals with the problem of automatic rule generation for fuzzy systems design. The proposed approach is based on hybrid artificial bee colony (ABC) optimization and weighted least squares (LS) method and aims to find the structure and parameters of fuzzy systems simultaneously. More precisely, two ABC based fuzzy modeling strategies are presented and compared. The first strategy uses global optimization to learn fuzzy models, the second one hybridizes ABC and weighted least squares estimate method. The performances of the proposed ABC and ABC-LS fuzzy modeling strategies are evaluated on complex modeling problems and compared to other advanced modeling methods.

Keywords: automatic design, learning, fuzzy rules, hybrid, swarm optimization

Procedia PDF Downloads 423
3867 Terrain Classification for Ground Robots Based on Acoustic Features

Authors: Bernd Kiefer, Abraham Gebru Tesfay, Dietrich Klakow

Abstract:

The motivation of our work is to detect different terrain types traversed by a robot based on acoustic data from the robot-terrain interaction. Different acoustic features and classifiers were investigated, such as Mel-frequency cepstral coefficient and Gamma-tone frequency cepstral coefficient for the feature extraction, and Gaussian mixture model and Feed forward neural network for the classification. We analyze the system’s performance by comparing our proposed techniques with some other features surveyed from distinct related works. We achieve precision and recall values between 87% and 100% per class, and an average accuracy at 95.2%. We also study the effect of varying audio chunk size in the application phase of the models and find only a mild impact on performance.

Keywords: acoustic features, autonomous robots, feature extraction, terrain classification

Procedia PDF Downloads 348
3866 Perspective Shifting in the Elicited Language Production Can Defy with Aging

Authors: Tuyuan Cheng

Abstract:

As we age, many things become more difficult. Among the abilities are the linguistic and cognitive ones. Competing theories have shown that these two functions could diminish together or that one is selectively affected by the other. In other words, some proposes aging affects sentence production in the same way it affects sentence comprehension and other cognitive functions, while some argues it does not.To address this question, the current investigation is conducted into the critical aspect of sentences as well as cognitive abilities – the syntactic complexity and the number of perspective shifts being contained in the elicited production. Healthy non-pathological aging is often characterized by a cognitive and neural decline in a number of cognitive abilities. Although the language is assumed to be of the more stable domain, a variety of findings in the cognitive aging literature would suggest otherwise. Older adults often show deficits in language production and multiple aspects of comprehension. Nevertheless, while some age differences likely reflect cognitive decline, others might reflect changes in communicative goals, and some even display cognitive advantages. In the domain of language processing, research efforts have been made in tests that probed a variety of communicative abilities. In general, there exists a distinction: Comprehension seems to be selectively unaffected, while production does not. The current study raises a novel question and investigates whether aging affects the production of relative clauses (RCs) under the cognitive factor of perspective shifts. Based on Perspective Hypothesis (MacWhinney, 2000, 2005), our cognitive processes build upon a fundamental system of perspective-taking, and language provides a series of cues to facilitate the construction and shifting of perspectives. These cues include a wide variety of constructions, including RCs structures. In this regard, linguistic complexity can be determined by the number of perspective shifts, and the processing difficulties of RCs can be interpreted within the theory of perspective shifting. Two experiments were conducted to study language production under controlled conditions. In Experiment 1, older healthy participants were tested on standard measures of cognitive aging, including MMSE (Mini-Mental State Examination), ToMI-2 (a simplified Theory of Mind Inventory-2), and a perspective-shifting comprehension task programmed with E-Prime. The results were analyzed to examine if/how they are correlated with aging people’s subsequent production data. In Experiment 2, the production profile of differing RCs, SRC vs. ORC, were collected with healthy aging participants who perform a picture elicitation task. Variable containing 0, 1, or 2 perspective shifts were juxtaposed respectively to the pictures and counterbalanced presented for elicitation. In parallel, a controlled group of young adults were recruited to examine the linguistic and cognitive abilities in question. The results lead us to the discussion whetheraging affects RCs production in a manner determined by its semantic structure or the number of perspective shifts it contains or the status of participants’ mental understanding. The major findingsare: (1) Elders’ production on Chinese RCtypes did not display intrinsic difficulty asymmetry. (2) RC types (the linguistic structural features) and the cognitiveperspective shifts jointly play important roles in the elders’ RCproduction. (3) The production of RC may defy the aging in the case offlexibly preserved cognitive ability.

Keywords: cognition aging, perspective hypothesis, perspective shift, relative clauses, sentence complexity

Procedia PDF Downloads 100
3865 Dynamic Response and Damage Modeling of Glass Fiber Reinforced Epoxy Composite Pipes: Numerical Investigation

Authors: Ammar Maziz, Mostapha Tarfaoui, Said Rechak

Abstract:

The high mechanical performance of composite pipes can be adversely affected by their low resistance to impact loads. Loads in dynamic origin are dangerous and cause consequences on the operation of pipes because the damage is often not detected and can affect the structural integrity of composite pipes. In this work, an advanced 3-D finite element (FE) model, based on the use of intralaminar damage models was developed and used to predict damage under low-velocity impact. The performance of the numerical model is validated with the confrontation with the results of experimental tests. The results show that at low impact energy, the damage happens mainly by matrix cracking and delamination. The model capabilities to simulate the low-velocity impact events on the full-scale composite structures were proved.

Keywords: composite materials, low velocity impact, FEA, dynamic behavior, progressive damage modeling

Procedia PDF Downloads 152
3864 Software Quality Measurement System for Telecommunication Industry in Malaysia

Authors: Nor Fazlina Iryani Abdul Hamid, Mohamad Khatim Hasan

Abstract:

Evolution of software quality measurement has been started since McCall introduced his quality model in year 1977. Starting from there, several software quality models and software quality measurement methods had emerged but none of them focused on telecommunication industry. In this paper, the implementation of software quality measurement system for telecommunication industry was compulsory to accommodate the rapid growth of telecommunication industry. The quality value of the telecommunication related software could be calculated using this system by entering the required parameters. The system would calculate the quality value of the measured system based on predefined quality metrics and aggregated by referring to the quality model. It would classify the quality level of the software based on Net Satisfaction Index (NSI). Thus, software quality measurement system was important to both developers and users in order to produce high quality software product for telecommunication industry.

Keywords: software quality, quality measurement, quality model, quality metric, net satisfaction index

Procedia PDF Downloads 572
3863 Numerical Simulation of Fiber Bragg Grating Spectrum for Mode-І Delamination Detection

Authors: O. Hassoon, M. Tarfoui, A. El Malk

Abstract:

Fiber Bragg optic sensor embedded in composite material to detect and monitor the damage which is occur in composite structure. In this paper we deal with the mode-Ι delamination to determine the resistance of material to crack propagation, and use the coupling mode theory and T-matrix method to simulating the FBGs spectrum for both uniform and non-uniform strain distribution. The double cantilever beam test which is modeling in FEM to determine the Longitudinal strain, there are two models which are used, the first is the global half model, and the second the sub-model to represent the FBGs with refine mesh. This method can simulate the damage in the composite structure and converting the strain to wavelength shifting of the FBG spectrum.

Keywords: fiber bragg grating, delamination detection, DCB, FBG spectrum, structure health monitoring

Procedia PDF Downloads 350
3862 A Corpus-Based Study on the Lexical, Syntactic and Sequential Features across Interpreting Types

Authors: Qianxi Lv, Junying Liang

Abstract:

Among the various modes of interpreting, simultaneous interpreting (SI) is regarded as a ‘complex’ and ‘extreme condition’ of cognitive tasks while consecutive interpreters (CI) do not have to share processing capacity between tasks. Given that SI exerts great cognitive demand, it makes sense to posit that the output of SI may be more compromised than that of CI in the linguistic features. The bulk of the research has stressed the varying cognitive demand and processes involved in different modes of interpreting; however, related empirical research is sparse. In keeping with our interest in investigating the quantitative linguistic factors discriminating between SI and CI, the current study seeks to examine the potential lexical simplification, syntactic complexity and sequential organization mechanism with a self-made inter-model corpus of transcribed simultaneous and consecutive interpretation, translated speech and original speech texts with a total running word of 321960. The lexical features are extracted in terms of the lexical density, list head coverage, hapax legomena, and type-token ratio, as well as core vocabulary percentage. Dependency distance, an index for syntactic complexity and reflective of processing demand is employed. Frequency motif is a non-grammatically-bound sequential unit and is also used to visualize the local function distribution of interpreting the output. While SI is generally regarded as multitasking with high cognitive load, our findings evidently show that CI may impose heavier or taxing cognitive resource differently and hence yields more lexically and syntactically simplified output. In addition, the sequential features manifest that SI and CI organize the sequences from the source text in different ways into the output, to minimize the cognitive load respectively. We reasoned the results in the framework that cognitive demand is exerted both on maintaining and coordinating component of Working Memory. On the one hand, the information maintained in CI is inherently larger in volume compared to SI. On the other hand, time constraints directly influence the sentence reformulation process. The temporal pressure from the input in SI makes the interpreters only keep a small chunk of information in the focus of attention. Thus, SI interpreters usually produce the output by largely retaining the source structure so as to relieve the information from the working memory immediately after formulated in the target language. Conversely, CI interpreters receive at least a few sentences before reformulation, when they are more self-paced. CI interpreters may thus tend to retain and generate the information in a way to lessen the demand. In other words, interpreters cope with the high demand in the reformulation phase of CI by generating output with densely distributed function words, more content words of higher frequency values and fewer variations, simpler structures and more frequently used language sequences. We consequently propose a revised effort model based on the result for a better illustration of cognitive demand during both interpreting types.

Keywords: cognitive demand, corpus-based, dependency distance, frequency motif, interpreting types, lexical simplification, sequential units distribution, syntactic complexity

Procedia PDF Downloads 157
3861 Forecasting Amman Stock Market Data Using a Hybrid Method

Authors: Ahmad Awajan, Sadam Al Wadi

Abstract:

In this study, a hybrid method based on Empirical Mode Decomposition and Holt-Winter (EMD-HW) is used to forecast Amman stock market data. First, the data are decomposed by EMD method into Intrinsic Mode Functions (IMFs) and residual components. Then, all components are forecasted by HW technique. Finally, forecasting values are aggregated together to get the forecasting value of stock market data. Empirical results showed that the EMD- HW outperform individual forecasting models. The strength of this EMD-HW lies in its ability to forecast non-stationary and non- linear time series without a need to use any transformation method. Moreover, EMD-HW has a relatively high accuracy comparing with eight existing forecasting methods based on the five forecast error measures.

Keywords: Holt-Winter method, empirical mode decomposition, forecasting, time series

Procedia PDF Downloads 111
3860 Sensitivity and Uncertainty Analysis of Hydrocarbon-In-Place in Sandstone Reservoir Modeling: A Case Study

Authors: Nejoud Alostad, Anup Bora, Prashant Dhote

Abstract:

Kuwait Oil Company (KOC) has been producing from its major reservoirs that are well defined and highly productive and of superior reservoir quality. These reservoirs are maturing and priority is shifting towards difficult reservoir to meet future production requirements. This paper discusses the results of the detailed integrated study for one of the satellite complex field discovered in the early 1960s. Following acquisition of new 3D seismic data in 1998 and re-processing work in the year 2006, an integrated G&G study was undertaken to review Lower Cretaceous prospectivity of this reservoir. Nine wells have been drilled in the area, till date with only three wells showing hydrocarbons in two formations. The average oil density is around 300API (American Petroleum Institute), and average porosity and water saturation of the reservoir is about 23% and 26%, respectively. The area is dissected by a number of NW-SE trending faults. Structurally, the area consists of horsts and grabens bounded by these faults and hence compartmentalized. The Wara/Burgan formation consists of discrete, dirty sands with clean channel sand complexes. There is a dramatic change in Upper Wara distributary channel facies, and reservoir quality of Wara and Burgan section varies with change of facies over the area. So predicting reservoir facies and its quality out of sparse well data is a major challenge for delineating the prospective area. To characterize the reservoir of Wara/Burgan formation, an integrated workflow involving seismic, well, petro-physical, reservoir and production engineering data has been used. Porosity and water saturation models are prepared and analyzed to predict reservoir quality of Wara and Burgan 3rd sand upper reservoirs. Subsequently, boundary conditions are defined for reservoir and non-reservoir facies by integrating facies, porosity and water saturation. Based on the detailed analyses of volumetric parameters, potential volumes of stock-tank oil initially in place (STOIIP) and gas initially in place (GIIP) were documented after running several probablistic sensitivity analysis using Montecalro simulation method. Sensitivity analysis on probabilistic models of reservoir horizons, petro-physical properties, and oil-water contacts and their effect on reserve clearly shows some alteration in the reservoir geometry. All these parameters have significant effect on the oil in place. This study has helped to identify uncertainty and risks of this prospect particularly and company is planning to develop this area with drilling of new wells.

Keywords: original oil-in-place, sensitivity, uncertainty, sandstone, reservoir modeling, Monte-Carlo simulation

Procedia PDF Downloads 182
3859 A DFT-Based QSARs Study of Kovats Retention Indices of Adamantane Derivatives

Authors: Z. Bayat

Abstract:

A quantitative structure–property relationship (QSPR) study was performed to develop models those relate the structures of 65 Kovats retention index (RI) of adamantane derivatives. Molecular descriptors derived solely from 3D structures of the molecular compounds. The usefulness of the quantum chemical descriptors, calculated at the level of the DFT theories using 6-311+G** basis set for QSAR study of adamantane derivatives was examined. The use of descriptors calculated only from molecular structure eliminates the need to experimental determination of properties for use in the correlation and allows for the estimation of RI for molecules not yet synthesized. The prediction results are in good agreement with the experimental value. A multi-parametric equation containing maximum Four descriptors at B3LYP/6-31+G** method with good statistical qualities (R2train=0.913, Ftrain=97.67, R2test=0.770, Ftest=3.21, Q2LOO=0.895, R2adj=0.904, Q2LGO=0.844) was obtained by Multiple Linear Regression using stepwise method.

Keywords: DFT, adamantane, QSAR, Kovat

Procedia PDF Downloads 354
3858 LaPEA: Language for Preprocessing of Edge Applications in Smart Factory

Authors: Masaki Sakai, Tsuyoshi Nakajima, Kazuya Takahashi

Abstract:

In order to improve the productivity of a factory, it is often the case to create an inference model by collecting and analyzing operational data off-line and then to develop an edge application (EAP) that evaluates the quality of the products or diagnoses machine faults in real-time. To accelerate this development cycle, an edge application framework for the smart factory is proposed, which enables to create and modify EAPs based on prepared inference models. In the framework, the preprocessing component is the key part to make it work. This paper proposes a language for preprocessing of edge applications, called LaPEA, which can flexibly process several sensor data from machines into explanatory variables for an inference model, and proves that it meets the requirements for the preprocessing.

Keywords: edge application framework, edgecross, preprocessing language, smart factory

Procedia PDF Downloads 129
3857 Location3: A Location Scouting Platform for the Support of Film and Multimedia Industries

Authors: Dimitrios Tzilopoulos, Panagiotis Symeonidis, Michael Loufakis, Dimosthenis Ioannidis, Dimitrios Tzovaras

Abstract:

The domestic film industry in Greece has traditionally relied heavily on state support. While film productions are crucial for the country's economy, it has not fully capitalized on attracting and promoting foreign productions. The lack of motivation, organized state support for attraction and licensing, and the absence of location scouting have hindered its potential. Although recent legislative changes have addressed the first two of these issues, the development of a comprehensive location database and a search engine that would effectively support location scouting at the pre-production location scouting is still in its early stages. In addition to the expected benefits of the film, television, marketing, and multimedia industries, a location-scouting service platform has the potential to yield significant financial gains locally and nationally. By promoting featured places like cultural and archaeological sites, natural monuments, and attraction points for visitors, it plays a vital role in both cultural promotion and facilitating tourism development. This study introduces LOCATION3, an internet platform revolutionizing film production location management. It interconnects location providers, film crews, and multimedia stakeholders, offering a comprehensive environment for seamless collaboration. The platform's central geodatabase (PostgreSQL) stores each location’s attributes, while web technologies like HTML, JavaScript, CSS, React.js, and Redux power the user-friendly interface. Advanced functionalities, utilizing deep learning models, developed in Python, are integrated via Node.js. Visual data presentation is achieved using the JS Leaflet library, delivering an interactive map experience. LOCATION3 sets a new standard, offering a range of essential features to enhance the management of film production locations. Firstly, it empowers users to effortlessly upload audiovisual material enriched with geospatial and temporal data, such as location coordinates, photographs, videos, 360-degree panoramas, and 3D location models. With the help of cutting-edge deep learning algorithms, the application automatically tags these materials, while users can also manually tag them. Moreover, the application allows users to record locations directly through its user-friendly mobile application. Users can then embark on seamless location searches, employing spatial or descriptive criteria. This intelligent search functionality considers a combination of relevant tags, dominant colors, architectural characteristics, emotional associations, and unique location traits. One of the application's standout features is the ability to explore locations by their visual similarity to other materials, facilitated by a reverse image search. Also, the interactive map serves as both a dynamic display for locations and a versatile filter, adapting to the user's preferences and effortlessly enhancing location searches. To further streamline the process, the application facilitates the creation of location lightboxes, enabling users to efficiently organize and share their content via email. Going above and beyond location management, the platform also provides invaluable liaison, matchmaking, and online marketplace services. This powerful functionality bridges the gap between visual and three-dimensional geospatial material providers, local agencies, film companies, production companies, etc. so that those interested in a specific location can access additional material beyond what is stored on the platform, as well as access production services supporting the functioning and completion of productions in a location (equipment provision, transportation, catering, accommodation, etc.).

Keywords: deep learning models, film industry, geospatial data management, location scouting

Procedia PDF Downloads 56
3856 Improved Estimation Strategies of Sensitive Characteristics Using Scrambled Response Techniques in Successive Sampling

Authors: S. Suman, G. N. Singh

Abstract:

This research work is an effort to analyse the consequences of scrambled response technique to estimate the current population mean in two-occasion successive sampling when the characteristic of interest is sensitive in nature. The generalized estimation procedures have been proposed using sensitive auxiliary variables under additive and multiplicative scramble models. The properties of resultant estimators have been deeply examined. Simulation, as well as empirical studies, are carried out to evaluate the performances of the proposed estimators with respect to other competent estimators. The results of our studies suggest that the proposed estimation procedures are highly effective under the presence of non-response situation. The result of this study also suggests that additive scrambled response model is a better choice in the perspective of cost of the survey and privacy of the respondents.

Keywords: scrambled response, sensitive characteristic, successive sampling, optimum replacement strategy

Procedia PDF Downloads 168
3855 Voluntary Disclosure Of Sustainability Information In Malaysian Federal-level Statutory Bodies

Authors: Siti Zabedah Saidin, Aidi Ahmi, Azharudin Ali, Wan Norhayati Wan Ahmad

Abstract:

In today's increasingly complex and interconnected world, the concept of sustainability has transcended mere corporate social responsibility, evolving into a fundamental driver of organizational behaviour and disclosure. This content analysis study delves into the Malaysian federal-level statutory bodies’ annual report for the year 2021, aiming to elucidate the extent of sustainability disclosures within the non-financial sections of these reports. The escalating global emphasis on sustainability has prompted organizations to embrace transparency as a means to demonstrate their commitment to environmental, social, and governance (ESG) considerations. Voluntary sustainability disclosure has emerged as a crucial channel through which organizations communicate their efforts, initiatives, and impacts in these areas, thereby fostering trust and accountability with stakeholders. The study aims to identify and examine the types of sustainability information disclosed voluntarily by the federal-level statutory bodies, concentrating on the non-financial sections of the annual reports. To achieve this, the study adopts a simplified disclosure index, a pragmatic tool that quantifies the extent of sustainability reporting in a standardized manner. Using convenience sampling, the study selects a sample of annual reports from the federal-level statutory bodies in Malaysia, as provided on their respective websites. The content analysis is centred on the non-financial sections of these reports, allowing for an in-depth exploration of sustainability disclosures. The findings of the study present the extent to which Malaysian federal-level statutory bodies embrace sustainability reporting. Through thorough content analysis, the study uncovered diverse dimensions of sustainability information, encompassing environmental impact assessments, social engagement endeavours, and governance frameworks. This reveals a deliberate effort by these bodies to encapsulate their holistic organizational contributions and challenges, transcending traditional financial metrics. This research contributes to the existing literature by providing insights into the evolving landscape of sustainability disclosure practices among Malaysian federal-level statutory bodies. The findings underline the proactive nature of these bodies in voluntarily sharing sustainability-related information, reflecting their recognition of the interconnectedness between organizational success and societal well-being. Furthermore, the study underscores the potential influence of regulatory guidelines and societal expectations in shaping the extent and nature of voluntary sustainability disclosures. Organizations are not merely responding to regulatory mandates but are actively aligning with global sustainability goals and stakeholder expectations. As organizations continue to navigate the intricate web of stakeholder expectations and sustainability imperatives, this study enriches the discourse surrounding transparency and sustainability reporting. The analysis emphasizes the important role of non-financial disclosures in portraying a holistic organizational narrative. In an era where stakeholders demand accountability, and the interconnectedness of global challenges necessitates collaborative action, the voluntary disclosure of sustainability information stands as a testament to the commitment of Malaysian federal-level statutory bodies in shaping a more sustainable future.

Keywords: voluntary disclosure, sustainability information, annual report, federal-level statutory body

Procedia PDF Downloads 43
3854 Factors Related to Employee Adherence to Rules in Kuwait Business Organizations

Authors: Ali Muhammad

Abstract:

The purpose of this study is to develop a theoretical framework which demonstrates the effect of four personal factors on employees rule following behavior in Kuwaiti business organizations. The model suggested in this study includes organizational citizenship behavior, affective organizational commitment, organizational trust, and procedural justice as possible predictors of rule following behavior. The study also attempts to compare the effects of the suggested factors on employees rule following behavior. The new model will, hopefully, extend previous research by adding new variables to the models used to explain employees rule following behavior. A discussion of issues related to rule-following behavior is presented, as well as recommendations for future research.

Keywords: employee adherence to rules, organizational justice, organizational commitment, organizational citizenship behavior

Procedia PDF Downloads 441
3853 The Impact of the Global Financial Crises on MILA Stock Markets

Authors: Miriam Sosa, Edgar Ortiz, Alejandra Cabello

Abstract:

This paper examines the volatility changes and leverage effects of the MILA stock markets and their changes since the 2007 global financial crisis. This group integrates the stock markets from Chile, Colombia, Mexico and Peru. Volatility changes and leverage effects are tested with a symmetric GARCH (1,1) and asymmetric TARCH (1,1) models with a dummy variable in the variance equation. Daily closing prices of the stock indexes of Chile (IPSA), Colombia (COLCAP), Mexico (IPC) and Peru (IGBVL) are examined for the period 2003:01 to 2015:02. The evidence confirms the presence of an overall increase in asymmetric market volatility in the Peruvian share market since the 2007 crisis.

Keywords: financial crisis, Latin American Integrated Market, TARCH, GARCH

Procedia PDF Downloads 258
3852 An ICF Framework for Game-Based Experiences in Geriatric Care

Authors: Marlene Rosa, Susana Lopes

Abstract:

Board games have been used for different purposes in geriatric care, demonstrating good results for health in general. However, there is not a conceptual framework that can help professionals and researchers in this area to design intervention programs or to think about future studies in this area. The aim of this study was to provide a pilot collection of board games’ serious purposes in geriatric care, using a WHO framework for health and disability. Study cases were developed in seven geriatric residential institutions from the center region in Portugal that are included in AGILAB program. The AGILAB program is a serious game-based method to train and spread out the implementation of board games in geriatric care. Each institution provides 2-hours/week of experiences using TATI Hand Game for serious purposes and then fulfill questions about a study-case (player characteristics; explain changes in players health according to this game experience). Two independent researchers read the information and classified it according to the International Classification for Functioning and Disability (ICF) categories. Any discrepancy was solved in a consensus meeting. Results indicate an important variability in body functions and structures: specific mental functions (e.g., b140 Attention functions, b144 Memory functions), b156 Perceptual functions, b2 sensory functions and pain (e.g., b230 Hearing functions; b265 Touch function; b280 Sensation of pain), b7 neuromusculoskeletal and movement-related functions (e.g., b730 Muscle power functions; b760 Control of voluntary movement functions; b710 Mobility of joint functions). Less variability was found in activities and participation domains, such as purposeful sensory experiences (d110-d129) (e.g., d115 Listening), communication (d3), d710 basic interpersonal interactions, d920 recreation and leisure (d9200 Play; d9205 Socializing). Concluding, this framework designed from a brief gamed-based experience includes mental, perceptual, sensory, neuromusculoskeletal, and movement-related functions and participation in sensory, communication, and leisure domains. More studies, including different experiences and a high number of users, should be developed to provide a more comprehensive ICF framework for game-based experiences in geriatric care.

Keywords: board game, aging, framework, experience

Procedia PDF Downloads 115
3851 A Hardware-in-the-loop Simulation for the Development of Advanced Control System Design for a Spinal Joint Wear Simulator

Authors: Kaushikk Iyer, Richard M Hall, David Keeling

Abstract:

Hardware-in-the-loop (HIL) simulation is an advanced technique for developing and testing complex real-time control systems. This paper presents the benefits of HIL simulation and how it can be implemented and used effectively to develop, test, and validate advanced control algorithms used in a spinal joint Wear simulator for the Tribological testing of spinal disc prostheses. spinal wear simulator is technologically the most advanced machine currently employed For the in-vitro testing of newly developed spinal Discimplants. However, the existing control techniques, such as a simple position control Does not allow the simulator to test non-sinusoidal waveforms. Thus, there is a need for better and advanced control methods that can be developed and tested Rigorouslybut safely before deploying it into the real simulator. A benchtop HILsetupis was created for experimentation, controller verification, and validation purposes, allowing different control strategies to be tested rapidly in a safe environment. The HIL simulation aspect in this setup attempts to replicate similar spinal motion and loading conditions. The spinal joint wear simulator containsa four-Barlinkpowered by electromechanical actuators. LabVIEW software is used to design a kinematic model of the spinal wear Simulator to Validatehow each link contributes towards the final motion of the implant under test. As a result, the implant articulates with an angular motion specified in the international standards, ISO-18192-1, that define fixed, simplified, and sinusoid motion and load profiles for wear testing of cervical disc implants. Using a PID controller, a velocity-based position control algorithm was developed to interface with the benchtop setup that performs HIL simulation. In addition to PID, a fuzzy logic controller (FLC) was also developed that acts as a supervisory controller. FLC provides intelligence to the PID controller by By automatically tuning the controller for profiles that vary in amplitude, shape, and frequency. This combination of the fuzzy-PID controller is novel to the wear testing application for spinal simulators and demonstrated superior performance against PIDwhen tested for a spectrum of frequency. Kaushikk Iyer is a Ph.D. Student at the University of Leeds and an employee at Key Engineering Solutions, Leeds, United Kingdom, (e-mail: [email protected], phone: +44 740 541 5502). Richard M Hall is with the University of Leeds, the United Kingdom as a professor in the Mechanical Engineering Department (e-mail: [email protected]). David Keeling is the managing director of Key Engineering Solutions, Leeds, United Kingdom (e-mail: [email protected]). Results obtained are successfully validated against the load and motion tolerances specified by the ISO18192-1 standard and fall within limits, that is, ±0.5° at the maxima and minima of the motion and ±2 % of the complete cycle for phasing. The simulation results prove the efficacy of the test setup using HIL simulation to verify and validate the accuracy and robustness of the prospective controller before its deployment into the spinal wear simulator. This method of testing controllers enables a wide range of possibilities to test advanced control algorithms that can potentially test even profiles of patients performing various dailyliving activities.

Keywords: Fuzzy-PID controller, hardware-in-the-loop (HIL), real-time simulation, spinal wear simulator

Procedia PDF Downloads 156
3850 Modern Imputation Technique for Missing Data in Linear Functional Relationship Model

Authors: Adilah Abdul Ghapor, Yong Zulina Zubairi, Rahmatullah Imon

Abstract:

Missing value problem is common in statistics and has been of interest for years. This article considers two modern techniques in handling missing data for linear functional relationship model (LFRM) namely the Expectation-Maximization (EM) algorithm and Expectation-Maximization with Bootstrapping (EMB) algorithm using three performance indicators; namely the mean absolute error (MAE), root mean square error (RMSE) and estimated biased (EB). In this study, we applied the methods of imputing missing values in the LFRM. Results of the simulation study suggest that EMB algorithm performs much better than EM algorithm in both models. We also illustrate the applicability of the approach in a real data set.

Keywords: expectation-maximization, expectation-maximization with bootstrapping, linear functional relationship model, performance indicators

Procedia PDF Downloads 379
3849 Analytical Model for Columns in Existing Reinforced Concrete Buildings

Authors: Chang Seok Lee, Sang Whan Han, Girbo Ko, Debbie Kim

Abstract:

Existing reinforced concrete structures are designed and built without considering seismic loads. The columns in such buildings generally exhibit widely spaced transverse reinforcements without using seismic hooks. Due to the insufficient reinforcement details in columns, brittle shear failure is expected in columns that may cause pre-mature building collapse mechanism during earthquakes. In order to retrofit those columns, the accurate seismic behavior of the columns needs to be predicted with proper analytical models. In this study, an analytical model is proposed for accurately simulating the cyclic behavior of shear critical columns. The parameters for pinching and cyclic deterioration in strength and stiffness are calibrated using test data of column specimens failed by shear.

Keywords: analytical model, cyclic deterioration, existing reinforced concrete columns, shear failure

Procedia PDF Downloads 252
3848 Prediction of Energy Storage Areas for Static Photovoltaic System Using Irradiation and Regression Modelling

Authors: Kisan Sarda, Bhavika Shingote

Abstract:

This paper aims to evaluate regression modelling for prediction of Energy storage of solar photovoltaic (PV) system using Semi parametric regression techniques because there are some parameters which are known while there are some unknown parameters like humidity, dust etc. Here irradiation of solar energy is different for different places on the basis of Latitudes, so by finding out areas which give more storage we can implement PV systems at those places and our need of energy will be fulfilled. This regression modelling is done for daily, monthly and seasonal prediction of solar energy storage. In this, we have used R modules for designing the algorithm. This algorithm will give the best comparative results than other regression models for the solar PV cell energy storage.

Keywords: semi parametric regression, photovoltaic (PV) system, regression modelling, irradiation

Procedia PDF Downloads 364
3847 Real Time Detection, Prediction and Reconstitution of Rain Drops

Authors: R. Burahee, B. Chassinat, T. de Laclos, A. Dépée, A. Sastim

Abstract:

The purpose of this paper is to propose a solution to detect, predict and reconstitute rain drops in real time – during the night – using an embedded material with an infrared camera. To prevent the system from needing too high hardware resources, simple models are considered in a powerful image treatment algorithm reducing considerably calculation time in OpenCV software. Using a smart model – drops will be matched thanks to a process running through two consecutive pictures for implementing a sophisticated tracking system. With this system drops computed trajectory gives information for predicting their future location. Thanks to this technique, treatment part can be reduced. The hardware system composed by a Raspberry Pi is optimized to host efficiently this code for real time execution.

Keywords: reconstitution, prediction, detection, rain drop, real time, raspberry, infrared

Procedia PDF Downloads 399
3846 Performance Analysis of Artificial Neural Network with Decision Tree in Prediction of Diabetes Mellitus

Authors: J. K. Alhassan, B. Attah, S. Misra

Abstract:

Human beings have the ability to make logical decisions. Although human decision - making is often optimal, it is insufficient when huge amount of data is to be classified. medical dataset is a vital ingredient used in predicting patients health condition. In other to have the best prediction, there calls for most suitable machine learning algorithms. This work compared the performance of Artificial Neural Network (ANN) and Decision Tree Algorithms (DTA) as regards to some performance metrics using diabetes data. The evaluations was done using weka software and found out that DTA performed better than ANN. Multilayer Perceptron (MLP) and Radial Basis Function (RBF) were the two algorithms used for ANN, while RegTree and LADTree algorithms were the DTA models used. The Root Mean Squared Error (RMSE) of MLP is 0.3913,that of RBF is 0.3625, that of RepTree is 0.3174 and that of LADTree is 0.3206 respectively.

Keywords: artificial neural network, classification, decision tree algorithms, diabetes mellitus

Procedia PDF Downloads 392
3845 Ikat: Undaunted Journey of a Traditional Textile Practice, a Sublime Connect of Traditionality with Modernity and Calibration for Eco-Sustainable Options

Authors: Purva Khurana

Abstract:

Traditional textile crafts are universally found to have been significantly impeded by the uprise of innovative technologies, but sustained human endeavor, in sync with dynamic market nuances, holds key to these otherwise getting fast-extinct marvels. The metamorphosis of such art-forms into niche markets pre-supposes sharp concentration on adaptability. The author has concentrated on the ancient handicraft of Ikat in Andhra Pradesh (India), a manifestation of their cultural heritage and esoteric cottage industry, so very intrinsic to the development and support of local economy and identity. Like any other traditional practice, ikat weaving has been subjected to the challenges of modernization. However, owing to its unique character, personalize production and adaptability, both of material and process, ikat weaving has stood the test of time by way of judiciously embellishing innovation with contemporary taste. To survive as a living craft as also to justify its role as a universal language of aesthetic sensibility, it is imperative that ikat tradition should lend itself continuous process of experiments, change and growth. Besides, the instant paper aims to examine the contours of ikat production process from its pure form, to more fashion and market oriented production, with upgraded process, material and tools. Over the time, it has adapted well to new style-paradigms, duly matching up with the latest fashion trends, in tandem with the market-sensitivities. Apart, it is an effort to investigate how this craft could respond constructively to the pressure of contemporary technical developments in order to be at cutting edge, while preserving its integrity. In order to approach these issues, the methodology adopted is, conceptual analysis of the craft practices, its unique strength and how they could be used to advance the craft in relation to the emergence of technical developments. The paper summarizes the result of the study carried out by the author on the peculiar advantages of suitably- calibrated vat dyes over natural dyes, in terms of its recycling ability and eco-friendly properties, thus holding definite edge, both in terms of socio-economic as well as environmental concerns.

Keywords: craft, eco-friendly dyes, ikat, metamorphosis

Procedia PDF Downloads 160
3844 Estimation of Pressure Loss Coefficients in Combining Flows Using Artificial Neural Networks

Authors: Shahzad Yousaf, Imran Shafi

Abstract:

This paper presents a new method for calculation of pressure loss coefficients by use of the artificial neural network (ANN) in tee junctions. Geometry and flow parameters are feed into ANN as the inputs for purpose of training the network. Efficacy of the network is demonstrated by comparison of the experimental and ANN based calculated data of pressure loss coefficients for combining flows in a tee junction. Reynolds numbers ranging from 200 to 14000 and discharge ratios varying from minimum to maximum flow for calculation of pressure loss coefficients have been used. Pressure loss coefficients calculated using ANN are compared to the models from literature used in junction flows. The results achieved after the application of ANN agrees reasonably to the experimental values.

Keywords: artificial neural networks, combining flow, pressure loss coefficients, solar collector tee junctions

Procedia PDF Downloads 374