Search results for: working models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 9730

Search results for: working models

9370 Effectiveness of Medication and Non-Medication Therapy on Working Memory of Children with Attention Deficit and Hyperactivity Disorder

Authors: Mohaammad Ahmadpanah, Amineh Akhondi, Mohammad Haghighi, Ali Ghaleiha, Leila Jahangard, Elham Salari

Abstract:

Background: Working memory includes the capability to keep and manipulate information in a short period of time. This capability is the basis of complicated judgments and has been attended to as the specific and constant character of individuals. Children with attention deficit and hyperactivity are among the people suffering from deficiency in the active memory, and this deficiency has been attributed to the problem of frontal lobe. This study utilizes a new approach with suitable tasks and methods for training active memory and assessment of the effects of the trainings. Participants: The children participating in this study were of 7-15 year age, who were diagnosed by the psychiatrist and psychologist as hyperactive and attention deficit based on DSM-IV criteria. The intervention group was consisted of 8 boys and 6 girls with the average age of 11 years and standard deviation of 2, and the control group was consisted of 2 girls and 5 boys with an average age of 11.4 and standard deviation of 3. Three children in the test group and two in the control group were under medicinal therapy. Results: Working memory training meaningfully improved the performance in not-trained areas as visual-spatial working memory as well as the performance in Raven progressive tests which are a perfect example of non-verbal, complicated reasoning tasks. In addition, motional activities – measured based on the number of head movements during computerized measuring program – was meaningfully reduced in the medication group. The results of the second test showed that training similar exercise to teenagers and adults results in the improvement of cognition functions, as in hyperactive people. Discussion: The results of this study showed that the performance of working memory is improved through training, and these trainings are extended and generalized in other areas of cognition functions not receiving any training. Trainings resulted in the improvement of performance in the tasks related to prefrontal. They had also a positive and meaningful impact on the moving activities of hyperactive children.

Keywords: attention deficit hyperactivity disorder, working memory, non-medical treatment, children

Procedia PDF Downloads 367
9369 Review of Theories and Applications of Genetic Programing in Sediment Yield Modeling

Authors: Adesoji Tunbosun Jaiyeola, Josiah Adeyemo

Abstract:

Sediment yield can be considered to be the total sediment load that leaves a drainage basin. The knowledge of the quantity of sediments present in a river at a particular time can lead to better flood capacity in reservoirs and consequently help to control over-bane flooding. Furthermore, as sediment accumulates in the reservoir, it gradually loses its ability to store water for the purposes for which it was built. The development of hydrological models to forecast the quantity of sediment present in a reservoir helps planners and managers of water resources systems, to understand the system better in terms of its problems and alternative ways to address them. The application of artificial intelligence models and technique to such real-life situations have proven to be an effective approach of solving complex problems. This paper makes an extensive review of literature relevant to the theories and applications of evolutionary algorithms, and most especially genetic programming. The successful applications of genetic programming as a soft computing technique were reviewed in sediment modelling and other branches of knowledge. Some fundamental issues such as benchmark, generalization ability, bloat and over-fitting and other open issues relating to the working principles of GP, which needs to be addressed by the GP community were also highlighted. This review aim to give GP theoreticians, researchers and the general community of GP enough research direction, valuable guide and also keep all stakeholders abreast of the issues which need attention during the next decade for the advancement of GP.

Keywords: benchmark, bloat, generalization, genetic programming, over-fitting, sediment yield

Procedia PDF Downloads 446
9368 Study on Flexible Diaphragm In-Plane Model of Irregular Multi-Storey Industrial Plant

Authors: Cheng-Hao Jiang, Mu-Xuan Tao

Abstract:

The rigid diaphragm model may cause errors in the calculation of internal forces due to neglecting the in-plane deformation of the diaphragm. This paper thus studies the effects of different diaphragm in-plane models (including in-plane rigid model and in-plane flexible model) on the seismic performance of structures. Taking an actual industrial plant as an example, the seismic performance of the structure is predicted using different floor diaphragm models, and the analysis errors caused by different diaphragm in-plane models including deformation error and internal force error are calculated. Furthermore, the influence of the aspect ratio on the analysis errors is investigated. Finally, the code rationality is evaluated by assessing the analysis errors of the structure models whose floors were determined as rigid according to the code’s criterion. It is found that different floor models may cause great differences in the distribution of structural internal forces, and the current code may underestimate the influence of the floor in-plane effect.

Keywords: industrial plant, diaphragm, calculating error, code rationality

Procedia PDF Downloads 140
9367 Analysis of Organizational Hybrid Agile Methods Environments: Frameworks, Benefits, and Challenges

Authors: Majid Alsubaie, Hamed Sarbazhosseini

Abstract:

Many working environments have experienced increased uncertainty due to the fast-moving and unpredictable world. IT systems development projects, in particular, face several challenges because of their rapidly changing environments and emerging technologies. Information technology organizations within these contexts adapt systems development methodology and new software approaches to address this issue. One of these methodologies is the Agile method, which has gained huge attention in recent years. However, due to failure rates in IT projects, there is an increasing demand for the use of hybrid Agile methods among organizations. The scarce research in the area means that organizations do not have solid evidence-based knowledge for the use of hybrid Agile. This research was designed to provide further insights into the development of hybrid Agile methods within systems development projects, including how frameworks and processes are used and what benefits and challenges are gained and faced as a result of hybrid Agile methods. This paper presents how three organizations (two government and one private) use hybrid Agile methods in their Agile environments. The data was collected through interviews and a review of relevant documents. The results indicate that these organizations do not predominantly use pure Agile. Instead, they are waterfall organizations by virtue of systems nature and complexity, and Agile is used underneath as the delivery model. Prince2 Agile framework, SAFe, Scrum, and Kanban were the identified models and frameworks followed. This study also found that customer satisfaction and the ability to build quickly are the most frequently perceived benefits of using hybrid Agile methods. In addition, team resistance and scope changes are the common challenges identified by research participants in their working environments. The findings can help to understand Agile environmental conditions and projects that can help get better success rates and customer satisfaction.

Keywords: agile, hybrid, IT systems, management, success rate, technology

Procedia PDF Downloads 108
9366 Responding to the Mental Health Service Needs of Rural-to-Urban Migrant Workers in China: Current Situation and Future Directions

Authors: Yujun Liu, Maosheng Ran

Abstract:

Background: Chinese rural-to-urban migrant workers’ mental health problems raise attentions from different social sectors. However, situation of present mental health services provided to this population has not been discovered. This study attempts to describe the current mental health service situation, identify the gaps and give the future directions based on the quantitative data. Methods: Questionnaire surveys were conducted among 2017 rural-to-urban migrant workers in 13 cities and 100 social work service organizations in 5 cities in 2014. Data was collected by face-to-face structured interview by trained interviewers. Findings: Migrant workers’ mental health status was not good. Compared to the severity of mental distress, mental health service for this population was lacking and insufficient, which accounted for only 14.4% of all services in our sample. And the group work and case work were the most frequently-used methods. By estimating a series of regression models, we revealed that life experiences and working conditions were significantly associated with migrant workers’ mental health status. Therefore, the macro social work practices aimed at this whole group were advocated to promote their mental wellbeing. That is, practitioners should not only focus on the improvement of migrant workers’ emotion management capacity, but also pay attention to raise awareness and improve their living and working condition; not only concentrate on the solving of individuals’ dilemma, but also promote gradual reformation of present labor regime and hukou system in China.

Keywords: Chinese rural-to-urban migrant workers, macro social work practice, mental health service needs, mental health status

Procedia PDF Downloads 281
9365 Assessment of Artists’ Socioeconomic and Working Conditions: The Empirical Case of Lithuania

Authors: Rusne Kregzdaite, Erika Godlevska, Morta Vidunaite

Abstract:

The main aim of this research is to explore existing methodologies for artists’ labour force and create artists’ socio-economic and creative conditions in an assessment model. Artists have dual aims in their creative working process: 1) income and 2) artistic self-expression. The valuation of their conditions takes into consideration both sides: the factors related to income and the satisfaction of the creative process and its result. The problem addressed in the study: tangible and intangible artists' criteria used for assessments creativity conditions. The proposed model includes objective factors (working time, income, etc.) and subjective factors (salary covering essential needs, self-satisfaction). Other intangible indicators are taken into account: the impact on the common culture, social values, and the possibility to receive awards, to represent the country in the international market. The empirical model consists of 59 separate indicators, grouped into eight categories. The deviation of each indicator from the general evaluation allows for identifying the strongest and the weakest components of artists’ conditions.

Keywords: artist conditions, artistic labour force, cultural policy, indicator, assessment model

Procedia PDF Downloads 152
9364 Probing Language Models for Multiple Linguistic Information

Authors: Bowen Ding, Yihao Kuang

Abstract:

In recent years, large-scale pre-trained language models have achieved state-of-the-art performance on a variety of natural language processing tasks. The word vectors produced by these language models can be viewed as dense encoded presentations of natural language that in text form. However, it is unknown how much linguistic information is encoded and how. In this paper, we construct several corresponding probing tasks for multiple linguistic information to clarify the encoding capabilities of different language models and performed a visual display. We firstly obtain word presentations in vector form from different language models, including BERT, ELMo, RoBERTa and GPT. Classifiers with a small scale of parameters and unsupervised tasks are then applied on these word vectors to discriminate their capability to encode corresponding linguistic information. The constructed probe tasks contain both semantic and syntactic aspects. The semantic aspect includes the ability of the model to understand semantic entities such as numbers, time, and characters, and the grammatical aspect includes the ability of the language model to understand grammatical structures such as dependency relationships and reference relationships. We also compare encoding capabilities of different layers in the same language model to infer how linguistic information is encoded in the model.

Keywords: language models, probing task, text presentation, linguistic information

Procedia PDF Downloads 110
9363 Application Difference between Cox and Logistic Regression Models

Authors: Idrissa Kayijuka

Abstract:

The logistic regression and Cox regression models (proportional hazard model) at present are being employed in the analysis of prospective epidemiologic research looking into risk factors in their application on chronic diseases. However, a theoretical relationship between the two models has been studied. By definition, Cox regression model also called Cox proportional hazard model is a procedure that is used in modeling data regarding time leading up to an event where censored cases exist. Whereas the Logistic regression model is mostly applicable in cases where the independent variables consist of numerical as well as nominal values while the resultant variable is binary (dichotomous). Arguments and findings of many researchers focused on the overview of Cox and Logistic regression models and their different applications in different areas. In this work, the analysis is done on secondary data whose source is SPSS exercise data on BREAST CANCER with a sample size of 1121 women where the main objective is to show the application difference between Cox regression model and logistic regression model based on factors that cause women to die due to breast cancer. Thus we did some analysis manually i.e. on lymph nodes status, and SPSS software helped to analyze the mentioned data. This study found out that there is an application difference between Cox and Logistic regression models which is Cox regression model is used if one wishes to analyze data which also include the follow-up time whereas Logistic regression model analyzes data without follow-up-time. Also, they have measurements of association which is different: hazard ratio and odds ratio for Cox and logistic regression models respectively. A similarity between the two models is that they are both applicable in the prediction of the upshot of a categorical variable i.e. a variable that can accommodate only a restricted number of categories. In conclusion, Cox regression model differs from logistic regression by assessing a rate instead of proportion. The two models can be applied in many other researches since they are suitable methods for analyzing data but the more recommended is the Cox, regression model.

Keywords: logistic regression model, Cox regression model, survival analysis, hazard ratio

Procedia PDF Downloads 455
9362 Comparison of Wake Oscillator Models to Predict Vortex-Induced Vibration of Tall Chimneys

Authors: Saba Rahman, Arvind K. Jain, S. D. Bharti, T. K. Datta

Abstract:

The present study compares the semi-empirical wake-oscillator models that are used to predict vortex-induced vibration of structures. These models include those proposed by Facchinetti, Farshidian, and Dolatabadi, and Skop and Griffin. These models combine a wake oscillator model resembling the Van der Pol oscillator model and a single degree of freedom oscillation model. In order to use these models for estimating the top displacement of chimneys, the first mode vibration of the chimneys is only considered. The modal equation of the chimney constitutes the single degree of freedom model (SDOF). The equations of the wake oscillator model and the SDOF are simultaneously solved using an iterative procedure. The empirical parameters used in the wake-oscillator models are estimated using a newly developed approach, and response is compared with experimental data, which appeared comparable. For carrying out the iterative solution, the ode solver of MATLAB is used. To carry out the comparative study, a tall concrete chimney of height 210m has been chosen with the base diameter as 28m, top diameter as 20m, and thickness as 0.3m. The responses of the chimney are also determined using the linear model proposed by E. Simiu and the deterministic model given in Eurocode. It is observed from the comparative study that the responses predicted by the Facchinetti model and the model proposed by Skop and Griffin are nearly the same, while the model proposed by Fashidian and Dolatabadi predicts a higher response. The linear model without considering the aero-elastic phenomenon provides a less response as compared to the non-linear models. Further, for large damping, the prediction of the response by the Euro code is relatively well compared to those of non-linear models.

Keywords: chimney, deterministic model, van der pol, vortex-induced vibration

Procedia PDF Downloads 221
9361 Analysis of Moving Loads on Bridges Using Surrogate Models

Authors: Susmita Panda, Arnab Banerjee, Ajinkya Baxy, Bappaditya Manna

Abstract:

The design of short to medium-span high-speed bridges in critical locations is an essential aspect of vehicle-bridge interaction. Due to dynamic interaction between moving load and bridge, mathematical models or finite element modeling computations become time-consuming. Thus, to reduce the computational effort, a universal approximator using an artificial neural network (ANN) has been used to evaluate the dynamic response of the bridge. The data set generation and training of surrogate models have been conducted over the results obtained from mathematical modeling. Further, the robustness of the surrogate model has been investigated, which showed an error percentage of less than 10% with conventional methods. Additionally, the dependency of the dynamic response of the bridge on various load and bridge parameters has been highlighted through a parametric study.

Keywords: artificial neural network, mode superposition method, moving load analysis, surrogate models

Procedia PDF Downloads 100
9360 An In-Depth Experimental Study of Wax Deposition in Pipelines

Authors: Arias M. L., D’Adamo J., Novosad M. N., Raffo P. A., Burbridge H. P., Artana G.

Abstract:

Shale oils are highly paraffinic and, consequently, can create wax deposits that foul pipelines during transportation. Several factors must be considered when designing pipelines or treatment programs that prevents wax deposition: including chemical species in crude oils, flowrates, pipes diameters and temperature. This paper describes the wax deposition study carried out within the framework of Y-TEC's flow assurance projects, as part of the process to achieve a better understanding on wax deposition issues. Laboratory experiments were performed on a medium size, 1 inch diameter, wax deposition loop of 15 mts long equipped with a solid detector system, online microscope to visualize crystals, temperature and pressure sensors along the loop pipe. A baseline test was performed with diesel with no paraffin or additive content. Tests were undertaken with different temperatures of circulating and cooling fluid at different flow conditions. Then, a solution formed with a paraffin added to the diesel was considered. Tests varying flowrate and cooling rate were again run. Viscosity, density, WAT (Wax Appearance Temperature) with DSC (Differential Scanning Calorimetry), pour point and cold finger measurements were carried out to determine physical properties of the working fluids. The results obtained in the loop were analyzed through momentum balance and heat transfer models. To determine possible paraffin deposition scenarios temperature and pressure loop output signals were studied. They were compared with WAT static laboratory methods. Finally, we scrutinized the effect of adding a chemical inhibitor to the working fluid on the dynamics of the process of wax deposition in the loop.

Keywords: paraffin desposition, flow assurance, chemical inhibitors, flow loop

Procedia PDF Downloads 105
9359 Applying Multiplicative Weight Update to Skin Cancer Classifiers

Authors: Animish Jain

Abstract:

This study deals with using Multiplicative Weight Update within artificial intelligence and machine learning to create models that can diagnose skin cancer using microscopic images of cancer samples. In this study, the multiplicative weight update method is used to take the predictions of multiple models to try and acquire more accurate results. Logistic Regression, Convolutional Neural Network (CNN), and Support Vector Machine Classifier (SVMC) models are employed within the Multiplicative Weight Update system. These models are trained on pictures of skin cancer from the ISIC-Archive, to look for patterns to label unseen scans as either benign or malignant. These models are utilized in a multiplicative weight update algorithm which takes into account the precision and accuracy of each model through each successive guess to apply weights to their guess. These guesses and weights are then analyzed together to try and obtain the correct predictions. The research hypothesis for this study stated that there would be a significant difference in the accuracy of the three models and the Multiplicative Weight Update system. The SVMC model had an accuracy of 77.88%. The CNN model had an accuracy of 85.30%. The Logistic Regression model had an accuracy of 79.09%. Using Multiplicative Weight Update, the algorithm received an accuracy of 72.27%. The final conclusion that was drawn was that there was a significant difference in the accuracy of the three models and the Multiplicative Weight Update system. The conclusion was made that using a CNN model would be the best option for this problem rather than a Multiplicative Weight Update system. This is due to the possibility that Multiplicative Weight Update is not effective in a binary setting where there are only two possible classifications. In a categorical setting with multiple classes and groupings, a Multiplicative Weight Update system might become more proficient as it takes into account the strengths of multiple different models to classify images into multiple categories rather than only two categories, as shown in this study. This experimentation and computer science project can help to create better algorithms and models for the future of artificial intelligence in the medical imaging field.

Keywords: artificial intelligence, machine learning, multiplicative weight update, skin cancer

Procedia PDF Downloads 79
9358 Chemometric Estimation of Inhibitory Activity of Benzimidazole Derivatives by Linear Least Squares and Artificial Neural Networks Modelling

Authors: Sanja O. Podunavac-Kuzmanović, Strahinja Z. Kovačević, Lidija R. Jevrić, Stela Jokić

Abstract:

The subject of this paper is to correlate antibacterial behavior of benzimidazole derivatives with their molecular characteristics using chemometric QSAR (Quantitative Structure–Activity Relationships) approach. QSAR analysis has been carried out on the inhibitory activity of benzimidazole derivatives against Staphylococcus aureus. The data were processed by linear least squares (LLS) and artificial neural network (ANN) procedures. The LLS mathematical models have been developed as a calibration models for prediction of the inhibitory activity. The quality of the models was validated by leave one out (LOO) technique and by using external data set. High agreement between experimental and predicted inhibitory acivities indicated the good quality of the derived models. These results are part of the CMST COST Action No. CM1306 "Understanding Movement and Mechanism in Molecular Machines".

Keywords: Antibacterial, benzimidazoles, chemometric, QSAR.

Procedia PDF Downloads 316
9357 Fusion of MOLA-based DEMs and HiRISE Images for Large-Scale Mars Mapping

Authors: Ahmed F. Elaksher, Islam Omar

Abstract:

In this project, we used MOLA-based DEMs to orthorectify HiRISE optical images. The MOLA data was interpolated using the kriging interpolation technique. Corresponding tie points were then digitized from both datasets. These points were employed in co-registering both datasets using GIS analysis tools. Different transformation models, including the affine and projective transformation models, were used with different sets and distributions of tie points. Additionally, we evaluated the use of the MOLA elevations in co-registering the MOLA and HiRISE datasets. The planimetric RMSEs achieved for each model are reported. Results suggested the use of 3D-2D transformation models.

Keywords: photogrammetry, Mars, MOLA, HiRISE

Procedia PDF Downloads 78
9356 Evaluation of QSRR Models by Sum of Ranking Differences Approach: A Case Study of Prediction of Chromatographic Behavior of Pesticides

Authors: Lidija R. Jevrić, Sanja O. Podunavac-Kuzmanović, Strahinja Z. Kovačević

Abstract:

The present study deals with the selection of the most suitable quantitative structure-retention relationship (QSRR) models which should be used in prediction of the retention behavior of basic, neutral, acidic and phenolic pesticides which belong to different classes: fungicides, herbicides, metabolites, insecticides and plant growth regulators. Sum of ranking differences (SRD) approach can give a different point of view on selection of the most consistent QSRR model. SRD approach can be applied not only for ranking of the QSRR models, but also for detection of similarity or dissimilarity among them. Applying the SRD analysis, the most similar models can be found easily. In this study, selection of the best model was carried out on the basis of the reference ranking (“golden standard”) which was defined as the row average values of logarithm of retention time (logtr) defined by high performance liquid chromatography (HPLC). Also, SRD analysis based on experimental logtr values as reference ranking revealed similar grouping of the established QSRR models already obtained by hierarchical cluster analysis (HCA).

Keywords: chemometrics, chromatography, pesticides, sum of ranking differences

Procedia PDF Downloads 375
9355 On the Development of Medical Additive Manufacturing in Egypt

Authors: Khalid Abdelghany

Abstract:

Additive Manufacturing (AM) is the manufacturing technology that is used to fabricate fast products direct from CAD models in very short time and with minimum operation steps. Jointly with the advancement in medical computer modeling, AM proved to be a very efficient tool to help physicians, orthopedic surgeons and dentists design and fabricate patient-tailored surgical guides, templates and customized implants from the patient’s CT / MRI images. AM jointly with computer-assisted designing/computer-assisted manufacturing (CAD/CAM) technology have enabled medical practitioners to tailor physical models in a patient-and purpose-specific fashion and helped to design and manufacture of templates, appliances and devices with a high range of accuracy using biocompatible materials. In developing countries, there are some technical and financial limitations of implementing such advanced tools as an essential portion of medical applications. CMRDI institute in Egypt has been working in the field of Medical Additive Manufacturing since 2003 and has assisted in the recovery of hundreds of poor patients using these advanced tools. This paper focuses on the surgical and dental use of 3D printing technology in Egypt as a developing country. The presented case studies have been designed and processed using the software tools and additive manufacturing machines in CMRDI through cooperative engineering and medical works. Results showed that the implementation of the additive manufacturing tools in developed countries is successful and could be economical comparing to long treatment plans.

Keywords: additive manufacturing, dental and orthopeadic stents, patient specific surgical tools, titanium implants

Procedia PDF Downloads 315
9354 Analysis of the Advent of Multinational Corporations in Developing Countries: Case Study of Nike Factories Expansion in Vietnam

Authors: Khue Do Phan

Abstract:

Nike has been confronted by the press with their harsh working conditions, underpayment and highly-labor intensive requirement to their manufacturing workers and hiring of underage workers in Vietnam, Nike's largest production center. To analyze this topic critically through an international relations perspective, theory of dependency will be used to criticize the notion of exploitation of resources from developed countries towards developing countries. Theory of economic liberalism will be used to support the notion private property, the free market and generally capitalism as beneficial to both developing and developed countries. Workers are mentally, physically and sexually abused in the factories. In addition to this, their working conditions consist of improper training, lack of safety equipment, exposure of chemicals (glues and pants), their average wage is below the minimum wage in their country; the workers have to work around 60 hours or more a week. Even Nike says that the conditions are regulated often to make sure the workers get a voice to have their work rights and safe working environment. The monitors come to analyze the factories but in the end talk to the employers, whom are the direct abusers to the employees. Health benefits are rarely granted to the employees; they are forced to pay their bills first then the company will reimburse them later. They would also get in trouble for using the bathroom, taking a lunch break or sick days off because this would mean it would decrease their hours of work, leading to an even lower wage and a really angry employer. Of course with the press criticizing Nike’s lack of respect for human rights and working rights, Nike has been working on policy making and implementation to deal with the abuses. Due to its large chains and a great number of outsourcing host countries, the changes that Nike wish or attempt to make have not be in effect as quickly nor spreading to all countries it holds accountable for in its outsourcing factories.

Keywords: dependency theory, economic liberalism, human rights, outsource

Procedia PDF Downloads 334
9353 Dual Language Immersion Models in Theory and Practice

Authors: S. Gordon

Abstract:

Dual language immersion is growing fast in language teaching today. This study provides an overview and evaluation of the different models of Dual language immersion programs in US K-12 schools. First, the paper provides a brief current literature review on the theory of Dual Language Immersion (DLI) in Second Language Acquisition (SLA) studies. Second, examples of several types of DLI language teaching models in US K-12 public schools are presented (including 50/50 models, 90/10 models, etc.). Third, we focus on the unique example of DLI education in the state of Utah, a successful, growing program in K-12 schools that includes: French, Chinese, Spanish, and Portuguese. The project investigates the theory and practice particularly of the case of public elementary and secondary school children that study half their school day in the L1 and the other half in the chosen L2, from kindergarten (age 5-6) through high school (age 17-18). Finally, the project takes the observations of Utah French DLI elementary through secondary programs as a case study. To conclude, we look at the principal challenges, pedagogical objectives and outcomes, and important implications for other US states and other countries (such as France currently) that are in the process of developing similar language learning programs.

Keywords: dual language immersion, second language acquisition, language teaching, pedagogy, teaching, French

Procedia PDF Downloads 175
9352 A Comprehensive Study of a Hybrid System Integrated Solid Oxide Fuel cell, Gas Turbine, Organic Rankine Cycle with Compressed air Energy Storage

Authors: Taiheng Zhang, Hongbin Zhao

Abstract:

Compressed air energy storage become increasingly vital for solving intermittency problem of some renewable energies. In this study, a new hybrid system on a combination of compressed air energy storage (CAES), solid oxide fuel cell (SOFC), gas turbine (GT), and organic Rankine cycle (ORC) is proposed. In the new system, excess electricity during off-peak time is utilized to compress air. Then, the compressed air is stored in compressed air storage tank. During peak time, the compressed air enters the cathode of SOFC directly instead of combustion chamber of traditional CAES. There is no air compressor consumption of SOFC-GT in peak demand, so SOFC- GT can generate power with high-efficiency. In addition, the waste heat of exhaust from GT is recovered by applying an ORC. Three different organic working fluid (R123, R601, R601a) of ORC are chosen to evaluate system performance. Based on Aspen plus and Engineering Equation Solver (EES) software, energy and exergoeconomic analysis are used to access the viability of the combined system. Besides, the effect of two parameters (fuel flow and ORC turbine inlet pressure) on energy efficiency is studied. The effect of low-price electricity at off-peak hours on thermodynamic criteria (total unit exergy cost of products and total cost rate) is also investigated. Furthermore, for three different organic working fluids, the results of round-trip efficiency, exergy efficiency, and exergoeconomic factors are calculated and compared. Based on thermodynamic performance and exergoeconomic performance of different organic working fluids, the best suitable working fluid will be chosen. In conclusion, this study can provide important guidance for system efficiency improvement and viability.

Keywords: CAES, SOFC, ORC, energy and exergoeconomic analysis, organic working fluids

Procedia PDF Downloads 123
9351 Fixed-Bed Column Studies of Green Malachite Removal by Use of Alginate-Encapsulated Aluminium Pillared Clay

Authors: Lazhar mouloud, Chemat Zoubida, Ouhoumna Faiza

Abstract:

The main objective of this study, concerns the modeling of breakthrough curves obtained in the adsorption column of malachite green into alginate-encapsulated aluminium pillared clay in fixed bed according to various operating parameters such as the initial concentration, the feed rate and the height fixed bed, applying mathematical models namely: the model of Bohart and Adams, Wolborska, Bed Depth Service Time, Clark and Yoon-Nelson. These models allow us to express the different parameters controlling the performance of the dynamic adsorption system. The results have shown that all models were found suitable for describing the whole or a definite part of the dynamic behavior of the column with respect to the flow rate, the inlet dye concentration and the height of fixed bed.

Keywords: adsorption column, malachite green, pillared clays, alginate, modeling, mathematic models, encapsulation.

Procedia PDF Downloads 508
9350 An Improvement of a Dynamic Model of the Secondary Sedimentation Tank and Field Validation

Authors: Zahir Bakiri, Saci Nacefa

Abstract:

In this paper a comparison in made between two models, with and without dispersion term, and focused on the characterization of the movement of the sludge blanket in the secondary sedimentation tank using the solid flux theory and the velocity settling. This allowed us develop a one-dimensional models, with and without dispersion based on a thorough experimental study carried out in situ and the application of online data which are the mass load flow, transfer concentration, and influent characteristic. On the other hand, in the proposed model, the new settling velocity law (double-exponential function) used is based on the Vesilind function.

Keywords: wastewater, activated sludge, sedimentation, settling velocity, settling models

Procedia PDF Downloads 388
9349 Mapping Poverty in the Philippines: Insights from Satellite Data and Spatial Econometrics

Authors: Htet Khaing Lin

Abstract:

This study explores the relationship between a diverse set of variables, encompassing both environmental and socio-economic factors, and poverty levels in the Philippines for the years 2012, 2015, and 2018. Employing Ordinary Least Squares (OLS), Spatial Lag Models (SLM), and Spatial Error Models (SEM), this study delves into the dynamics of key indicators, including daytime and nighttime land surface temperature, cropland surface, urban land surface, rainfall, population size, normalized difference water, vegetation, and drought indices. The findings reveal consistent patterns and unexpected correlations, highlighting the need for nuanced policies that address the multifaceted challenges arising from the interplay of environmental and socio-economic factors.

Keywords: poverty analysis, OLS, spatial lag models, spatial error models, Philippines, google earth engine, satellite data, environmental dynamics, socio-economic factors

Procedia PDF Downloads 103
9348 Geopotential Models Evaluation in Algeria Using Stochastic Method, GPS/Leveling and Topographic Data

Authors: M. A. Meslem

Abstract:

For precise geoid determination, we use a reference field to subtract long and medium wavelength of the gravity field from observations data when we use the remove-compute-restore technique. Therefore, a comparison study between considered models should be made in order to select the optimal reference gravity field to be used. In this context, two recent global geopotential models have been selected to perform this comparison study over Northern Algeria. The Earth Gravitational Model (EGM2008) and the Global Gravity Model (GECO) conceived with a combination of the first model with anomalous potential derived from a GOCE satellite-only global model. Free air gravity anomalies in the area under study have been used to compute residual data using both gravity field models and a Digital Terrain Model (DTM) to subtract the residual terrain effect from the gravity observations. Residual data were used to generate local empirical covariance functions and their fitting to the closed form in order to compare their statistical behaviors according to both cases. Finally, height anomalies were computed from both geopotential models and compared to a set of GPS levelled points on benchmarks using least squares adjustment. The result described in details in this paper regarding these two models has pointed out a slight advantage of GECO global model globally through error degree variances comparison and ground-truth evaluation.

Keywords: quasigeoid, gravity aomalies, covariance, GGM

Procedia PDF Downloads 137
9347 Math Anxiety Effects on Complex Addition: An ERP Study

Authors: María Isabel Núñez-Peña, Macarena Suárez Pellicioni

Abstract:

In the present study, we used event-related potentials (ERP) to address the question of whether high (HMA) and low math-anxious (LMA) individuals differ on a complex addition verification task, which involved both carrying and non-carrying additions. ERPs were recorded while seventeen HMA and seventeen LMA individuals performed the verification task. Groups did not differ in trait anxiety or gender distribution. Participants were presented with two-digit additions and were asked to decide whether the proposed solution was correct or incorrect. Behavioral data showed a significant Carrying x Proposed solution x Group interaction for accuracy, showing that carrying additions were more error prone than non-carrying ones for both groups, although the difference non-carrying minus carrying was larger for the HMA group. As for ERPs, a P2 component larger in HMA individuals than in their LMA peers was found both for carrying and non-carrying additions. The P2 was followed by a sustained negative slow wave at parietal positions. Because the negative slow waves are thought to reflect the updating of working memory, these results give support to the relationship among working memory, math performance and math anxiety.

Keywords: math anxiety, carrying, working memory, P2

Procedia PDF Downloads 447
9346 Plant Identification Using Convolution Neural Network and Vision Transformer-Based Models

Authors: Virender Singh, Mathew Rees, Simon Hampton, Sivaram Annadurai

Abstract:

Plant identification is a challenging task that aims to identify the family, genus, and species according to plant morphological features. Automated deep learning-based computer vision algorithms are widely used for identifying plants and can help users narrow down the possibilities. However, numerous morphological similarities between and within species render correct classification difficult. In this paper, we tested custom convolution neural network (CNN) and vision transformer (ViT) based models using the PyTorch framework to classify plants. We used a large dataset of 88,000 provided by the Royal Horticultural Society (RHS) and a smaller dataset of 16,000 images from the PlantClef 2015 dataset for classifying plants at genus and species levels, respectively. Our results show that for classifying plants at the genus level, ViT models perform better compared to CNN-based models ResNet50 and ResNet-RS-420 and other state-of-the-art CNN-based models suggested in previous studies on a similar dataset. ViT model achieved top accuracy of 83.3% for classifying plants at the genus level. For classifying plants at the species level, ViT models perform better compared to CNN-based models ResNet50 and ResNet-RS-420, with a top accuracy of 92.5%. We show that the correct set of augmentation techniques plays an important role in classification success. In conclusion, these results could help end users, professionals and the general public alike in identifying plants quicker and with improved accuracy.

Keywords: plant identification, CNN, image processing, vision transformer, classification

Procedia PDF Downloads 104
9345 Analysis on Heat Transfer in Solar Parabolic Trough Collectors

Authors: Zaid H. Yaseen, Jamel A. Orfi, Zeyad A. Alsuhaibani

Abstract:

Solar power has a huge potential to be employed in the fields of electricity production, water desalination, and multi-generation. There are various types of solar collectors, and parabolic trough collectors (PTCs) are common among these types. In PTCs, a mirror is used to direct the incident radiation on an absorber tube to utilize the heat in power generation. In this work, a PTC covered with a glass tube is presented and analyzed. Results showed that temperatures of 510℃ for steam can be reached for certain parameters. The work also showed the viability of using Benzene as the working fluid in the absorber tube. Also, some analysis regarding changing the absorber’s tube diameter and the efficiency of the solar collector was demonstrated in this work. The effect of changing the heat transfer correlations for the convection phenomena of the working fluid was illustrated. In fact, two heat transfer correlations, the Dittus-Boelter and Gnielinski correlations, were used, and the outcomes showed a resemblance in the results for the maximum attainable temperature in the working fluid.

Keywords: absorber tube, glass tube, incident radiation, parabolic trough collector

Procedia PDF Downloads 11
9344 Sensitivity and Uncertainty Analysis of One Dimensional Shape Memory Alloy Constitutive Models

Authors: A. B. M. Rezaul Islam, Ernur Karadogan

Abstract:

Shape memory alloys (SMAs) are known for their shape memory effect and pseudoelasticity behavior. Their thermomechanical behaviors are modeled by numerous researchers using microscopic thermodynamic and macroscopic phenomenological point of view. Tanaka, Liang-Rogers and Ivshin-Pence models are some of the most popular SMA macroscopic phenomenological constitutive models. They describe SMA behavior in terms of stress, strain and temperature. These models involve material parameters and they have associated uncertainty present in them. At different operating temperatures, the uncertainty propagates to the output when the material is subjected to loading followed by unloading. The propagation of uncertainty while utilizing these models in real-life application can result in performance discrepancies or failure at extreme conditions. To resolve this, we used probabilistic approach to perform the sensitivity and uncertainty analysis of Tanaka, Liang-Rogers, and Ivshin-Pence models. Sobol and extended Fourier Amplitude Sensitivity Testing (eFAST) methods have been used to perform the sensitivity analysis for simulated isothermal loading/unloading at various operating temperatures. As per the results, it is evident that the models vary due to the change in operating temperature and loading condition. The average and stress-dependent sensitivity indices present the most significant parameters at several temperatures. This work highlights the sensitivity and uncertainty analysis results and shows comparison of them at different temperatures and loading conditions for all these models. The analysis presented will aid in designing engineering applications by eliminating the probability of model failure due to the uncertainty in the input parameters. Thus, it is recommended to have a proper understanding of sensitive parameters and the uncertainty propagation at several operating temperatures and loading conditions as per Tanaka, Liang-Rogers, and Ivshin-Pence model.

Keywords: constitutive models, FAST sensitivity analysis, sensitivity analysis, sobol, shape memory alloy, uncertainty analysis

Procedia PDF Downloads 144
9343 A Comprehensive Survey on Machine Learning Techniques and User Authentication Approaches for Credit Card Fraud Detection

Authors: Niloofar Yousefi, Marie Alaghband, Ivan Garibay

Abstract:

With the increase of credit card usage, the volume of credit card misuse also has significantly increased, which may cause appreciable financial losses for both credit card holders and financial organizations issuing credit cards. As a result, financial organizations are working hard on developing and deploying credit card fraud detection methods, in order to adapt to ever-evolving, increasingly sophisticated defrauding strategies and identifying illicit transactions as quickly as possible to protect themselves and their customers. Compounding on the complex nature of such adverse strategies, credit card fraudulent activities are rare events compared to the number of legitimate transactions. Hence, the challenge to develop fraud detection that are accurate and efficient is substantially intensified and, as a consequence, credit card fraud detection has lately become a very active area of research. In this work, we provide a survey of current techniques most relevant to the problem of credit card fraud detection. We carry out our survey in two main parts. In the first part, we focus on studies utilizing classical machine learning models, which mostly employ traditional transnational features to make fraud predictions. These models typically rely on some static physical characteristics, such as what the user knows (knowledge-based method), or what he/she has access to (object-based method). In the second part of our survey, we review more advanced techniques of user authentication, which use behavioral biometrics to identify an individual based on his/her unique behavior while he/she is interacting with his/her electronic devices. These approaches rely on how people behave (instead of what they do), which cannot be easily forged. By providing an overview of current approaches and the results reported in the literature, this survey aims to drive the future research agenda for the community in order to develop more accurate, reliable and scalable models of credit card fraud detection.

Keywords: Credit Card Fraud Detection, User Authentication, Behavioral Biometrics, Machine Learning, Literature Survey

Procedia PDF Downloads 121
9342 Effect of Sedimentation on Torque Transmission in the Larger Radius Magnetorheological Clutch

Authors: Manish Kumar Thakur, Chiranjit Sarkar

Abstract:

Sedimentation of magnetorheological (MR) fluid affects its working. MR fluid is a smart fluid that has unique qualities such as quick responsiveness and easy controllability. It is used in the MR damper, MR brake, and MR clutch. In this work effect of sedimentation on torque transmission in the shear mode operated MR clutch is investigated. A test rig is developed to test the impact of sedimentation on torque transmission in the MR clutch. Torque transmission capability of MR clutch has been measured under two conditions to confirm the result of sedimentation. The first experiment is done just after filling and the other after one week. It has been observed that transmission torque is decreased after sedimentation. Hence sedimentation affects the working of the MR clutch.

Keywords: clutch, magnetorheological fluid, sedimentation, torque

Procedia PDF Downloads 184
9341 Measuring Environmental Efficiency of Energy in OPEC Countries

Authors: Bahram Fathi, Seyedhossein Sajadifar, Naser Khiabani

Abstract:

Data envelopment analysis (DEA) has recently gained popularity in energy efficiency analysis. A common feature of the previously proposed DEA models for measuring energy efficiency performance is that they treat energy consumption as an input within a production framework without considering undesirable outputs. However, energy use results in the generation of undesirable outputs as byproducts of producing desirable outputs. Within a joint production framework of both desirable and undesirable outputs, this paper presents several DEA-type linear programming models for measuring energy efficiency performance. In addition to considering undesirable outputs, our models treat different energy sources as different inputs so that changes in energy mix could be accounted for in evaluating energy efficiency. The proposed models are applied to measure the energy efficiency performances of 12 OPEC countries and the results obtained are presented.

Keywords: energy efficiency, undesirable outputs, data envelopment analysis

Procedia PDF Downloads 736