Search results for: information value method
25264 The Role of Sustainable Development in the Design and Planning of Smart Cities Using GIS Techniques: Models of Arab Cities
Authors: Ahmed M. Jihad
Abstract:
The paper presents the concept of sustainable development, and the role of geographic techniques in the design, planning and presentation of maps of smart cities with geographical vision, and the identification of programs and tools, and models of maps of Arab cities, is the problem of research in how to apply, process and experience these programs? What is the role of geographic techniques in planning and mapping the optimal place for these cities? The paper proposes an addition to the designs of Iraqi cities, as it can be developed in the future to serve as a model for interactive smart cities by developing its services. The importance of this paper stems from the concept of sustainable development dynamic which has become a method of development imposed by the present era in rapid development to achieve social balance and specialized programs in draw paper argues that ensuring sustainable development is achieved through the use of information technology. The paper will follow the theoretical presentation of the importance of the concept of development, design tools and programs. The paper follows the method of analysis of modern systems (System Analysis Approach) through the latest programs will provide results can be said that the new Iraqi cities can be developed with smart technologies, like some of the Arab and European cities that were newly created through the introduction of international investment, and therefore Plans can be made to select the best programs in manufacturing and producing maps and smart cities in the future.Keywords: geographic techniques, planning the cities, smart cities, sustainable development
Procedia PDF Downloads 21025263 Study of Natural Convection Heat Transfer of Plate-Fin Heat Sink
Authors: Han-Taw Chen, Tzu-Hsiang Lin, Chung-Hou Lai
Abstract:
This study applies the inverse method and three-dimensional CFD commercial software in conjunction with the experimental temperature data to investigate the heat transfer and fluid flow characteristics of the plate-fin heat sink in a rectangular closed enclosure. The inverse method with the finite difference method and the experimental temperature data is applied to determine the approximate heat transfer coefficient. Later, based on the obtained results, the zero-equation turbulence model is used to obtain the heat transfer and fluid flow characteristics between two fins. To validate the accuracy of the results obtained, the comparison of the heat transfer coefficient is made. The obtained temperature at selected measurement locations of the fin is also compared with experimental data. The effect of the height of the rectangular enclosure on the obtained results is discussed.Keywords: inverse method, fluent, heat transfer characteristics, plate-fin heat sink
Procedia PDF Downloads 38925262 Telling the Truth to Patients Before Hip Fracture Surgery
Authors: Rawan Masarwa, Merav Ben Natan, Yaron Berkovich
Abstract:
Background: Hip fracture repair surgery carries a certain mortality risk, yet evidence suggests that orthopedic surgeons often refrain from discussing this issue with patients prior to surgery. Aim: This study aims to examine whether orthopedic surgeons address the issue of one-year post-surgery mortality before hip fracture repair surgery and to explore the factors influencing this decision. Method: The study uses a cross-sectional design, administering validated digital questionnaires to 150 orthopedic surgeons. Results: A minority of orthopedic surgeons reported consistently informing patients about the risk of mortality in the year following hip fracture surgery. The primary reasons for not discussing this risk were a desire to avoid frightening patients, time constraints, and concerns about undermining patient hope. Surgeons reported a medium-high level of perceived self-efficacy, with higher self-efficacy linked to a reduced likelihood of discussing one-year mortality risk. In contrast, older age and holding a specialist status in orthopedic surgery were associated with a higher likelihood of discussing this risk with patients. Conclusions: The findings suggest a need for interventions to address communication barriers and ensure consistent provision of essential information to patients undergoing hip fracture surgery. Additionally, they emphasize the importance of considering individual factors such as self-efficacy, age, and expertise in developing strategies to enhance patient-provider communication in orthopedic care settings.Keywords: orthopedic surgeons, hip fracture surgery, mortality risk communication, patient information
Procedia PDF Downloads 2525261 Deep Learning for SAR Images Restoration
Authors: Hossein Aghababaei, Sergio Vitale, Giampaolo Ferraioli
Abstract:
In the context of Synthetic Aperture Radar (SAR) data, polarization is an important source of information for Earth's surface monitoring. SAR Systems are often considered to transmit only one polarization. This constraint leads to either single or dual polarimetric SAR imaging modalities. Single polarimetric systems operate with a fixed single polarization of both transmitted and received electromagnetic (EM) waves, resulting in a single acquisition channel. Dual polarimetric systems, on the other hand, transmit in one fixed polarization and receive in two orthogonal polarizations, resulting in two acquisition channels. Dual polarimetric systems are obviously more informative than single polarimetric systems and are increasingly being used for a variety of remote sensing applications. In dual polarimetric systems, the choice of polarizations for the transmitter and the receiver is open. The choice of circular transmit polarization and coherent dual linear receive polarizations forms a special dual polarimetric system called hybrid polarimetry, which brings the properties of rotational invariance to geometrical orientations of features in the scene and optimizes the design of the radar in terms of reliability, mass, and power constraints. The complete characterization of target scattering, however, requires fully polarimetric data, which can be acquired with systems that transmit two orthogonal polarizations. This adds further complexity to data acquisition and shortens the coverage area or swath of fully polarimetric images compared to the swath of dual or hybrid polarimetric images. The search for solutions to augment dual polarimetric data to full polarimetric data will therefore take advantage of full characterization and exploitation of the backscattered field over a wider coverage with less system complexity. Several methods for reconstructing fully polarimetric images using hybrid polarimetric data can be found in the literature. Although the improvements achieved by the newly investigated and experimented reconstruction techniques are undeniable, the existing methods are, however, mostly based upon model assumptions (especially the assumption of reflectance symmetry), which may limit their reliability and applicability to vegetation and forest scenarios. To overcome the problems of these techniques, this paper proposes a new framework for reconstructing fully polarimetric information from hybrid polarimetric data. The framework uses Deep Learning solutions to augment hybrid polarimetric data without relying on model assumptions. A convolutional neural network (CNN) with a specific architecture and loss function is defined for this augmentation problem by focusing on different scattering properties of the polarimetric data. In particular, the method controls the CNN training process with respect to several characteristic features of polarimetric images defined by the combination of different terms in the cost or loss function. The proposed method is experimentally validated with real data sets and compared with a well-known and standard approach from the literature. From the experiments, the reconstruction performance of the proposed framework is superior to conventional reconstruction methods. The pseudo fully polarimetric data reconstructed by the proposed method also agree well with the actual fully polarimetric images acquired by radar systems, confirming the reliability and efficiency of the proposed method.Keywords: SAR image, polarimetric SAR image, convolutional neural network, deep learnig, deep neural network
Procedia PDF Downloads 6725260 Least Squares Method Identification of Corona Current-Voltage Characteristics and Electromagnetic Field in Electrostatic Precipitator
Authors: H. Nouri, I. E. Achouri, A. Grimes, H. Ait Said, M. Aissou, Y. Zebboudj
Abstract:
This paper aims to analysis the behaviour of DC corona discharge in wire-to-plate electrostatic precipitators (ESP). Current-voltage curves are particularly analysed. Experimental results show that discharge current is strongly affected by the applied voltage. The proposed method of current identification is to use the method of least squares. Least squares problems that of into two categories: linear or ordinary least squares and non-linear least squares, depending on whether or not the residuals are linear in all unknowns. The linear least-squares problem occurs in statistical regression analysis; it has a closed-form solution. A closed-form solution (or closed form expression) is any formula that can be evaluated in a finite number of standard operations. The non-linear problem has no closed-form solution and is usually solved by iterative.Keywords: electrostatic precipitator, current-voltage characteristics, least squares method, electric field, magnetic field
Procedia PDF Downloads 43125259 Bioanalytical Method Development and Validation of Aminophylline in Rat Plasma Using Reverse Phase High Performance Liquid Chromatography: An Application to Preclinical Pharmacokinetics
Authors: S. G. Vasantharaju, Viswanath Guptha, Raghavendra Shetty
Abstract:
Introduction: Aminophylline is a methylxanthine derivative belonging to the class bronchodilator. From the literature survey, reported methods reveals the solid phase extraction and liquid liquid extraction which is highly variable, time consuming, costly and laborious analysis. Present work aims to develop a simple, highly sensitive, precise and accurate high-performance liquid chromatography method for the quantification of Aminophylline in rat plasma samples which can be utilized for preclinical studies. Method: Reverse Phase high-performance liquid chromatography method. Results: Selectivity: Aminophylline and the internal standard were well separated from the co-eluted components and there was no interference from the endogenous material at the retention time of analyte and the internal standard. The LLOQ measurable with acceptable accuracy and precision for the analyte was 0.5 µg/mL. Linearity: The developed and validated method is linear over the range of 0.5-40.0 µg/mL. The coefficient of determination was found to be greater than 0.9967, indicating the linearity of this method. Accuracy and precision: The accuracy and precision values for intra and inter day studies at low, medium and high quality control samples concentrations of aminophylline in the plasma were within the acceptable limits Extraction recovery: The method produced consistent extraction recovery at all 3 QC levels. The mean extraction recovery of aminophylline was 93.57 ± 1.28% while that of internal standard was 90.70 ± 1.30%. Stability: The results show that aminophylline is stable in rat plasma under the studied stability conditions and that it is also stable for about 30 days when stored at -80˚C. Pharmacokinetic studies: The method was successfully applied to the quantitative estimation of aminophylline rat plasma following its oral administration to rats. Discussion: Preclinical studies require a rapid and sensitive method for estimating the drug concentration in the rat plasma. The method described in our article includes a simple protein precipitation extraction technique with ultraviolet detection for quantification. The present method is simple and robust for fast high-throughput sample analysis with less analysis cost for analyzing aminophylline in biological samples. In this proposed method, no interfering peaks were observed at the elution times of aminophylline and the internal standard. The method also had sufficient selectivity, specificity, precision and accuracy over the concentration range of 0.5 - 40.0 µg/mL. An isocratic separation technique was used underlining the simplicity of the presented method.Keywords: Aminophyllin, preclinical pharmacokinetics, rat plasma, RPHPLC
Procedia PDF Downloads 22225258 Technology Valuation of Unconventional Gas R&D Project Using Real Option Approach
Authors: Young Yoon, Jinsoo Kim
Abstract:
The adoption of information and communication technologies (ICT) in all industry is growing under industry 4.0. Many oil companies also are increasingly adopting ICT to improve the efficiency of existing operations, take more accurate and quicker decision making and reduce entire cost by optimization. It is true that ICT is playing an important role in the process of unconventional oil and gas development and companies must take advantage of ICT to gain competitive advantage. In this study, real option approach has been applied to Unconventional gas R&D project to evaluate ICT of them. Many unconventional gas reserves such as shale gas and coal-bed methane(CBM) has developed due to technological improvement and high energy price. There are many uncertainties in unconventional development on the three stage(Exploration, Development, Production). The traditional quantitative benefits-cost method, such as net present value(NPV) is not sufficient for capturing ICT value. We attempted to evaluate the ICT valuation by applying the compound option model; the model is applied to real CBM project case, showing how it consider uncertainties. Variables are treated as uncertain and a Monte Carlo simulation is performed to consider variables effect. Acknowledgement—This work was supported by the Energy Efficiency & Resources Core Technology Program of the Korea Institute of Energy Technology Evaluation and Planning (KETEP) granted financial resource from the Ministry of Trade, Industry & Energy, Republic of Korea (No. 20152510101880) and by the National Research Foundation of Korea Grant funded by the Korean Government (NRF-205S1A3A2046684).Keywords: information and communication technologies, R&D, real option, unconventional gas
Procedia PDF Downloads 22925257 A Performance Analysis Study for Cloud Based ERP Systems
Authors: Burak Erkayman
Abstract:
The manufacturing and service organizations are in the need of using ERP systems to integrate many functions from purchasing to storage, production planning to calculation of costs. Using ERP systems by the integration in the level of information provides companies remarkable advantages in terms of profitability, productivity and efficiency in processes. Cloud computing is one of the most significant changes in information and communication technology. The developments in Cloud Computing attract business world to take advantage of this field. Cloud Computing means much more storage area, more cost saving and faster data transfer rate. In addition to these, it presents new business models, new field of study and practicable solutions for anyone’s use. These developments make inevitable the implementation of ERP systems to cloud environment. In this study, the performance of ERP systems in cloud environment is analyzed through various performance criteria and a comparison between traditional and cloud-ERP systems is presented. At the end of study the transformation and the future of ERP systems is discussed.Keywords: cloud-ERP, ERP system performance, information system transformation
Procedia PDF Downloads 52925256 Online Estimation of Clutch Drag Torque in Wet Dual Clutch Transmission Based on Recursive Least Squares
Authors: Hongkui Li, Tongli Lu , Jianwu Zhang
Abstract:
This paper focuses on developing an estimation method of clutch drag torque in wet DCT. The modelling of clutch drag torque is investigated. As the main factor affecting the clutch drag torque, dynamic viscosity of oil is discussed. The paper proposes an estimation method of clutch drag torque based on recursive least squares by utilizing the dynamic equations of gear shifting synchronization process. The results demonstrate that the estimation method has good accuracy and efficiency.Keywords: clutch drag torque, wet DCT, dynamic viscosity, recursive least squares
Procedia PDF Downloads 31825255 Performance Comparison of Wideband Covariance Matrix Sparse Representation (W-CMSR) with Other Wideband DOA Estimation Methods
Authors: Sandeep Santosh, O. P. Sahu
Abstract:
In this paper, performance comparison of wideband covariance matrix sparse representation (W-CMSR) method with other existing wideband Direction of Arrival (DOA) estimation methods has been made.W-CMSR relies less on a priori information of the incident signal number than the ordinary subspace based methods.Consider the perturbation free covariance matrix of the wideband array output. The diagonal covariance elements are contaminated by unknown noise variance. The covariance matrix of array output is conjugate symmetric i.e its upper right triangular elements can be represented by lower left triangular ones.As the main diagonal elements are contaminated by unknown noise variance,slide over them and align the lower left triangular elements column by column to obtain a measurement vector.Simulation results for W-CMSR are compared with simulation results of other wideband DOA estimation methods like Coherent signal subspace method (CSSM), Capon, l1-SVD, and JLZA-DOA. W-CMSR separate two signals very clearly and CSSM, Capon, L1-SVD and JLZA-DOA fail to separate two signals clearly and an amount of pseudo peaks exist in the spectrum of L1-SVD.Keywords: W-CMSR, wideband direction of arrival (DOA), covariance matrix, electrical and computer engineering
Procedia PDF Downloads 47125254 A Methodological Concept towards a Framework Development for Social Software Adoption in Higher Education System
Authors: Kenneth N. Ohei, Roelien Brink
Abstract:
For decades, teaching and learning processes have centered on the traditional approach (Web 1.0) that promoted teacher-directed pedagogical practices. Currently, there is a realization that the traditional approach is not adequate to effectively address and improve all student-learning outcomes. The subsequent incorporation of social software, Information, and Communication Technology (ICT) tools in universities may serve as complementary to support educational goals, offering students the affordability and opportunity to educational choices and learning platforms. Consequently, educators’ inability to incorporate these instructional ICT tools in their teaching and learning practices remains a challenge. This will signify that educators still lack the ICT skills required to administer lectures and bridging learning gaps. This study probes a methodological concept with the aim of developing a framework towards the adoption of social software in HES to help facilitate business processes and can build social presence among students. A mixed method will be appropriate to develop a comprehensive framework needed in Higher Educational System (HES). After research have been conducted, the adoption of social software will be based on the developed comprehensive framework which is supposed to impact positively on education and approach of delivery, improves learning experience, engagement and finally, increases educational opportunities and easy access to educational contents.Keywords: blended and integrated learning, learning experience and engagement, higher educational system, HES, information and communication technology, ICT, social presence, Web 1.0, Web 2.0, Web 3.0
Procedia PDF Downloads 15725253 Collaborative Planning and Forecasting
Authors: Neha Asthana, Vishal Krishna Prasad
Abstract:
Collaborative planning and forecasting are the innovative and systematic approaches towards productive integration and assimilation of data synergized into information. The changing and variable market dynamics have persuaded global business chains to incorporate collaborative planning and forecasting as an imperative tool. Thus, it is essential for the supply chains to constantly improvise, update its nature, and mould as per changing global environment.Keywords: information transfer, forecasting, optimization, supply chain management
Procedia PDF Downloads 43525252 Design Systems and the Need for a Usability Method: Assessing the Fitness of Components and Interaction Patterns in Design Systems Using Atmosphere Methodology
Authors: Patrik Johansson, Selina Mardh
Abstract:
The present study proposes a usability test method, Atmosphere, to assess the fitness of components and interaction patterns of design systems. The method covers the user’s perception of the components of the system, the efficiency of the logic of the interaction patterns, perceived ease of use as well as the user’s understanding of the intended outcome of interactions. These aspects are assessed by combining measures of first impression, visual affordance and expectancy. The method was applied to a design system developed for the design of an electronic health record system. The study was conducted involving 15 healthcare personnel. It could be concluded that the Atmosphere method provides tangible data that enable human-computer interaction practitioners to analyze and categorize components and patterns based on perceived usability, success rate of identifying interactive components and success rate of understanding components and interaction patterns intended outcome.Keywords: atomic design, atmosphere methodology, design system, expectancy testing, first impression testing, usability testing, visual affordance testing
Procedia PDF Downloads 18025251 A Calibration Method of Portable Coordinate Measuring Arm Using Bar Gauge with Cone Holes
Authors: Rim Chang Hyon, Song Hak Jin, Song Kwang Hyok, Jong Ki Hun
Abstract:
The calibration of the articulated arm coordinate measuring machine (AACMM) is key to improving calibration accuracy and saving calibration time. To reduce the time consumed for calibration, we should choose the proper calibration gauges and develop a reasonable calibration method. In addition, we should get the exact optimal solution by accurately removing the rough errors within the experimental data. In this paper, we present a calibration method of the portable coordinate measuring arm (PCMA) using the 1.2m long bar guage with cone-holes. First, we determine the locations of the bar gauge and establish an optimal objective function for identifying the structural parameter errors. Next, we make a mathematical model of the calibration algorithm and present a new mathematical method to remove the rough errors within calibration data. Finally, we find the optimal solution to identify the kinematic parameter errors by using Levenberg-Marquardt algorithm. The experimental results show that our calibration method is very effective in saving the calibration time and improving the calibration accuracy.Keywords: AACMM, kinematic model, parameter identify, measurement accuracy, calibration
Procedia PDF Downloads 8325250 Towards a Model of Support in the Areas of Services of Educational Assistance and Mentoring in Middle Education in Mexico
Authors: Margarita Zavala, Gabriel Chavira, José González, Jorge Orozco, Julio Rolón, Roberto Pichardo
Abstract:
Adolescence is a neuralgic stage in the formation of every human being, generally this stage is when the middle school level is studied. In 2006, Mexico incorporated 'mentoring' space to assist students in their integration and participation in life. In public middle schools, it is sometimes difficult to be aware of situations that affect students because of the number of them and traditional records management. With this, they lose the opportunity to provide timely support as a preventive way. In order to provide this support, it is required to know the students by detecting the relevant information that has greater impact on their learning process. This research is looking to check if it is possible to identify student’s relevant information to detect when it is at risk, and then to propose a model to manage in a proper way such information.Keywords: adolescence, mentoring, middle school students, mentoring system support
Procedia PDF Downloads 48025249 A Study of Predicting Judgments on Causes of Online Privacy Invasions: Based on U.S Judicial Cases
Authors: Minjung Park, Sangmi Chai, Myoung Jun Lee
Abstract:
Since there are growing concerns on online privacy, enterprises could involve various personal privacy infringements cases resulting legal causations. For companies that are involving online business, it is important for them to pay extra attentions to protect users’ privacy. If firms can aware consequences from possible online privacy invasion cases, they can more actively prevent future online privacy infringements. This study attempts to predict the probability of ruling types caused by various invasion cases under U.S Personal Privacy Act. More specifically, this research explores online privacy invasion cases which was sentenced guilty to identify types of criminal punishments such as penalty, imprisonment, probation as well as compensation in civil cases. Based on the 853 U.S judicial cases ranged from January, 2000 to May, 2016, which related on data privacy, this research examines the relationship between personal information infringements cases and adjudications. Upon analysis results of 41,724 words extracted from 853 regal cases, this study examined online users’ privacy invasion cases to predict the probability of conviction for a firm as an offender in both of criminal and civil law. This research specifically examines that a cause of privacy infringements and a judgment type, whether it leads a civil or criminal liability, from U.S court. This study applies network text analysis (NTA) for data analysis, which is regarded as a useful method to discover embedded social trends within texts. According to our research results, certain online privacy infringement cases caused by online spamming and adware have a high possibility that firms are liable in the case. Our research results provide meaningful insights to academia as well as industry. First, our study is providing a new insight by applying Big Data analytics to legal cases so that it can predict the cause of invasions and legal consequences. Since there are few researches applying big data analytics in the domain of law, specifically in online privacy, this study suggests new area that future studies can explore. Secondly, this study reflects social influences, such as a development of privacy invasion technologies and changes of users’ level of awareness of online privacy on judicial cases analysis by adopting NTA method. Our research results indicate that firms need to improve technical and managerial systems to protect users’ online privacy to avoid negative legal consequences.Keywords: network text analysis, online privacy invasions, personal information infringements, predicting judgements
Procedia PDF Downloads 22925248 A Supervised Learning Data Mining Approach for Object Recognition and Classification in High Resolution Satellite Data
Authors: Mais Nijim, Rama Devi Chennuboyina, Waseem Al Aqqad
Abstract:
Advances in spatial and spectral resolution of satellite images have led to tremendous growth in large image databases. The data we acquire through satellites, radars and sensors consists of important geographical information that can be used for remote sensing applications such as region planning, disaster management. Spatial data classification and object recognition are important tasks for many applications. However, classifying objects and identifying them manually from images is a difficult task. Object recognition is often considered as a classification problem, this task can be performed using machine-learning techniques. Despite of many machine-learning algorithms, the classification is done using supervised classifiers such as Support Vector Machines (SVM) as the area of interest is known. We proposed a classification method, which considers neighboring pixels in a region for feature extraction and it evaluates classifications precisely according to neighboring classes for semantic interpretation of region of interest (ROI). A dataset has been created for training and testing purpose; we generated the attributes by considering pixel intensity values and mean values of reflectance. We demonstrated the benefits of using knowledge discovery and data-mining techniques, which can be on image data for accurate information extraction and classification from high spatial resolution remote sensing imagery.Keywords: remote sensing, object recognition, classification, data mining, waterbody identification, feature extraction
Procedia PDF Downloads 34025247 The Employees' Classification Method in the Space of Their Job Satisfaction, Loyalty and Involvement
Authors: Svetlana Ignatjeva, Jelena Slesareva
Abstract:
The aim of the study is development and adaptation of the method to analyze and quantify the indicators characterizing the relationship between a company and its employees. Diagnostics of such indicators is one of the most complex and actual issues in psychology of labour. The offered method is based on the questionnaire; its indicators reflect cognitive, affective and connotative components of socio-psychological attitude of employees to be as efficient as possible in their professional activities. This approach allows measure not only the selected factors but also such parameters as cognitive and behavioural dissonances. Adaptation of the questionnaire includes factor structure analysis and suitability analysis of phenomena indicators measured in terms of internal consistency of individual factors. Structural validity of the questionnaire was tested by exploratory factor analysis. Extraction Method: Principal Component Analysis. Rotation Method: Varimax with Kaiser Normalization. Factor analysis allows reduce dimension of the phenomena moving from the indicators to aggregative indexes and latent variables. Aggregative indexes are obtained as the sum of relevant indicators followed by standardization. The coefficient Cronbach's Alpha was used to assess the reliability-consistency of the questionnaire items. The two-step cluster analysis in the space of allocated factors allows classify employees according to their attitude to work in the company. The results of psychometric testing indicate possibility of using the developed technique for the analysis of employees’ attitude towards their work in companies and development of recommendations on their optimization.Keywords: involved in the organization, loyalty, organizations, method
Procedia PDF Downloads 35625246 From Electroencephalogram to Epileptic Seizures Detection by Using Artificial Neural Networks
Authors: Gaetano Zazzaro, Angelo Martone, Roberto V. Montaquila, Luigi Pavone
Abstract:
Seizure is the main factor that affects the quality of life of epileptic patients. The diagnosis of epilepsy, and hence the identification of epileptogenic zone, is commonly made by using continuous Electroencephalogram (EEG) signal monitoring. Seizure identification on EEG signals is made manually by epileptologists and this process is usually very long and error prone. The aim of this paper is to describe an automated method able to detect seizures in EEG signals, using knowledge discovery in database process and data mining methods and algorithms, which can support physicians during the seizure detection process. Our detection method is based on Artificial Neural Network classifier, trained by applying the multilayer perceptron algorithm, and by using a software application, called Training Builder that has been developed for the massive extraction of features from EEG signals. This tool is able to cover all the data preparation steps ranging from signal processing to data analysis techniques, including the sliding window paradigm, the dimensionality reduction algorithms, information theory, and feature selection measures. The final model shows excellent performances, reaching an accuracy of over 99% during tests on data of a single patient retrieved from a publicly available EEG dataset.Keywords: artificial neural network, data mining, electroencephalogram, epilepsy, feature extraction, seizure detection, signal processing
Procedia PDF Downloads 18825245 Consumer Knowledge and Behavior in the Aspect of Food Waste
Authors: Katarzyna Neffe-Skocinska, Marzena Tomaszewska, Beata Bilska, Dorota Zielinska, Monika Trzaskowska, Anna Lepecka, Danuta Kolozyn-Krajewska
Abstract:
The aim of the study was to assess Polish consumer behavior towards food waste, including knowledge of information on food labels. The survey was carried out using the CAPI (computer assisted personal interview) method, which involves interviewing the respondent using mobile devices. The research group was a representative sample for Poland due to demographic variables: gender, age, place of residence. A total of 1.115 respondents participated in the study (51.1% were women and 48.9% were men). The questionnaire included questions on five thematic aspects: 1. General knowledge and sources of information on the phenomenon of food waste; 2. Consumption of food after the date of minimum durability; 3. The meanings of the phrase 'best before ...'; 4. Indication of the difference between the meaning of the words 'best before ...' and 'use by'; 5. Indications products marked with the phrase 'best before ...'. It was found that every second surveyed Pole met with the topic of food waste (54.8%). Among the respondents, the most popular source of information related to the research topic was television (89.4%), radio (26%) and the Internet (24%). Over a third of respondents declared that they consume food after the date of minimum durability. Only every tenth (9.8%) respondent does not pay attention to the expiry date and type of consumed products (durable and perishable products). Correctly 39.8% of respondents answered the question: How do you understand the phrase 'best before ...'? In the opinion of 42.8% of respondents, the statements 'best before ...' and 'use by' mean the same thing, while 36% of them think differently. In addition, more than one-fifth of respondents could not respond to the questions. In the case of products of the indication information 'best before ...', more than 40% of the respondents chosen perishable products, e.g., yoghurts and durable, e.g., groats. A slightly lower percentage of indications was recorded for flour (35.1%), sausage (32.8%), canned corn (31.8%), and eggs (25.0%). Based on the assessment of the behavior of Polish consumers towards the phenomenon of food waste, it can be concluded that respondents have elementary knowledge of the study subject. Noteworthy is the good conduct of most respondents in terms of compliance with shelf life and dates of minimum durability of food products. The publication was financed on the basis of an agreement with the National Center for Research and Development No. Gospostrateg 1/385753/1/NCBR/2018 for the implementation and financing of the project under the strategic research and development program social and economic development of Poland in the conditions of globalizing markets – GOSPOSTRATEG - acronym PROM.Keywords: food waste, shelf life, dates of durability, consumer knowledge and behavior
Procedia PDF Downloads 17425244 A Safety Analysis Method for Multi-Agent Systems
Authors: Ching Louis Liu, Edmund Kazmierczak, Tim Miller
Abstract:
Safety analysis for multi-agent systems is complicated by the, potentially nonlinear, interactions between agents. This paper proposes a method for analyzing the safety of multi-agent systems by explicitly focusing on interactions and the accident data of systems that are similar in structure and function to the system being analyzed. The method creates a Bayesian network using the accident data from similar systems. A feature of our method is that the events in accident data are labeled with HAZOP guide words. Our method uses an Ontology to abstract away from the details of a multi-agent implementation. Using the ontology, our methods then constructs an “Interaction Map,” a graphical representation of the patterns of interactions between agents and other artifacts. Interaction maps combined with statistical data from accidents and the HAZOP classifications of events can be converted into a Bayesian Network. Bayesian networks allow designers to explore “what it” scenarios and make design trade-offs that maintain safety. We show how to use the Bayesian networks, and the interaction maps to improve multi-agent system designs.Keywords: multi-agent system, safety analysis, safety model, integration map
Procedia PDF Downloads 41725243 A Holistic Workflow Modeling Method for Business Process Redesign
Authors: Heejung Lee
Abstract:
In a highly competitive environment, it becomes more important to shorten the whole business process while delivering or even enhancing the business value to the customers and suppliers. Although the workflow management systems receive much attention for its capacity to practically support the business process enactment, the effective workflow modeling method remain still challenging and the high degree of process complexity makes it more difficult to gain the short lead time. This paper presents a workflow structuring method in a holistic way that can reduce the process complexity using activity-needs and formal concept analysis, which eventually enhances the key performance such as quality, delivery, and cost in business process.Keywords: workflow management, re-engineering, formal concept analysis, business process
Procedia PDF Downloads 40925242 Fire Smoke Removal over Cu-Mn-Ce Oxide Catalyst with CO₂ Sorbent Addition: Co Oxidation and in-situ CO₂ Sorption
Authors: Jin Lin, Shouxiang Lu, Kim Meow Liew
Abstract:
In a fire accident, fire smoke often poses a serious threat to human safety especially in the enclosed space such as submarine and space-crafts environment. Efficient removal of the hazardous gas products particularly a large amount of CO and CO₂ gases from these confined space is critical for the security of the staff and necessary for the post-fire environment recovery. In this work, Cu-Mn-Ce composite oxide catalysts coupled with CO₂ sorbents were prepared using wet impregnation method, solid-state impregnation method and wet/solid-state impregnation method. The as-prepared samples were tested dynamically and isothermally for CO oxidation and CO₂ sorption and further characterized by the X-ray diffraction (XRD), nitrogen adsorption and desorption, and field emission scanning electron microscopy (FE-SEM). The results showed that all the samples were able to catalyze CO into CO₂ and capture CO₂ in situ by chemisorption. Among all the samples, the sample synthesized by the wet/solid-state impregnation method showed the highest catalytic activity toward CO oxidation and the fine ability of CO₂ sorption. The sample prepared by the solid-state impregnation method showed the second CO oxidation performance, while the coupled sample using the wet impregnation method exhibited much poor CO oxidation activity. The various CO oxidation and CO₂ sorption properties of the samples might arise from the different dispersed states of the CO₂ sorbent in the CO catalyst, owing to the different preparation methods. XRD results confirmed the high-dispersed sorbent phase in the samples prepared by the wet and solid impregnation method, while that of the sample prepared by wet/solid-state impregnation method showed the larger bulk phase as indicated by the high-intensity diffraction peaks. Nitrogen adsorption and desorption results further revealed that the latter sample had a higher surface area and pore volume, which were beneficial for the CO oxidation over the catalyst. Hence, the Cu-Mn-Ce oxide catalyst coupled with CO₂ sorbent using wet/solid-state impregnation method could be a good choice for fire smoke removal in the enclosed space.Keywords: CO oxidation, CO₂ sorption, preparation methods, smoke removal
Procedia PDF Downloads 13925241 The Miller Umwelt Assessment Scale: A Tool for Planning Interventions for Children on the Autism Spectrum
Authors: Sonia Mastrangelo
Abstract:
The Miller Umwelt Assessment Scale is a useful tool for obtaining information about the developmental capacities of children on the autism spectrum. The assessment, made up of 19 tasks in the areas of: body organization, contact with surroundings, expressive and receptive communication, representation, and social-emotional development, has been used with much success over the past 40 years. While many assessments are difficult to administer to children on the autism spectrum, the simplicity of the MUAS reveals key strengths and challenges for both low and high functioning children on the spectrum. The results guide parents and clinicians in providing a curriculum and/or home program that moves children up the developmental ladder.Keywords: autism spectrum disorder, assessment, reading intervention, Miller method
Procedia PDF Downloads 53925240 K-Means Clustering-Based Infinite Feature Selection Method
Authors: Seyyedeh Faezeh Hassani Ziabari, Sadegh Eskandari, Maziar Salahi
Abstract:
Infinite Feature Selection (IFS) algorithm is an efficient feature selection algorithm that selects a subset of features of all sizes (including infinity). In this paper, we present an improved version of it, called clustering IFS (CIFS), by clustering the dataset in advance. To do so, first, we apply the K-means algorithm to cluster the dataset, then we apply IFS. In the CIFS method, the spatial and temporal complexities are reduced compared to the IFS method. Experimental results on 6 datasets show the superiority of CIFS compared to IFS in terms of accuracy, running time, and memory consumption.Keywords: feature selection, infinite feature selection, clustering, graph
Procedia PDF Downloads 12825239 An Optimal Control Model to Determine Body Forces of Stokes Flow
Authors: Yuanhao Gao, Pin Lin, Kees Weijer
Abstract:
In this paper, we will determine the external body force distribution with analysis of stokes fluid motion using mathematical modelling and numerical approaching. The body force distribution is regarded as the unknown variable and could be determined by the idea of optimal control theory. The Stokes flow motion and its velocity are generated by given forces in a unit square domain. A regularized objective functional is built to match the numerical result of flow velocity with the generated velocity data. So that the force distribution could be determined by minimizing the value of objective functional, which is also the difference between the numerical and experimental velocity. Then after utilizing the Lagrange multiplier method, some partial differential equations are formulated consisting the optimal control system to solve. Finite element method and conjugate gradient method are used to discretize equations and deduce the iterative expression of target body force to compute the velocity numerically and body force distribution. Programming environment FreeFEM++ supports the implementation of this model.Keywords: optimal control model, Stokes equation, finite element method, conjugate gradient method
Procedia PDF Downloads 40525238 Data Clustering in Wireless Sensor Network Implemented on Self-Organization Feature Map (SOFM) Neural Network
Authors: Krishan Kumar, Mohit Mittal, Pramod Kumar
Abstract:
Wireless sensor network is one of the most promising communication networks for monitoring remote environmental areas. In this network, all the sensor nodes are communicated with each other via radio signals. The sensor nodes have capability of sensing, data storage and processing. The sensor nodes collect the information through neighboring nodes to particular node. The data collection and processing is done by data aggregation techniques. For the data aggregation in sensor network, clustering technique is implemented in the sensor network by implementing self-organizing feature map (SOFM) neural network. Some of the sensor nodes are selected as cluster head nodes. The information aggregated to cluster head nodes from non-cluster head nodes and then this information is transferred to base station (or sink nodes). The aim of this paper is to manage the huge amount of data with the help of SOM neural network. Clustered data is selected to transfer to base station instead of whole information aggregated at cluster head nodes. This reduces the battery consumption over the huge data management. The network lifetime is enhanced at a greater extent.Keywords: artificial neural network, data clustering, self organization feature map, wireless sensor network
Procedia PDF Downloads 51725237 Secure Data Sharing of Electronic Health Records With Blockchain
Authors: Kenneth Harper
Abstract:
The secure sharing of Electronic Health Records (EHRs) is a critical challenge in modern healthcare, demanding solutions to enhance interoperability, privacy, and data integrity. Traditional standards like Health Information Exchange (HIE) and HL7 have made significant strides in facilitating data exchange between healthcare entities. However, these approaches rely on centralized architectures that are often vulnerable to data breaches, lack sufficient privacy measures, and have scalability issues. This paper proposes a framework for secure, decentralized sharing of EHRs using blockchain technology, cryptographic tokens, and Non-Fungible Tokens (NFTs). The blockchain's immutable ledger, decentralized control, and inherent security mechanisms are leveraged to improve transparency, accountability, and auditability in healthcare data exchanges. Furthermore, we introduce the concept of tokenizing patient data through NFTs, creating unique digital identifiers for each record, which allows for granular data access controls and proof of data ownership. These NFTs can also be employed to grant access to authorized parties, establishing a secure and transparent data sharing model that empowers both healthcare providers and patients. The proposed approach addresses common privacy concerns by employing privacy-preserving techniques such as zero-knowledge proofs (ZKPs) and homomorphic encryption to ensure that sensitive patient information can be shared without exposing the actual content of the data. This ensures compliance with regulations like HIPAA and GDPR. Additionally, the integration of Fast Healthcare Interoperability Resources (FHIR) with blockchain technology allows for enhanced interoperability, enabling healthcare organizations to exchange data seamlessly and securely across various systems while maintaining data governance and regulatory compliance. Through real-world case studies and simulations, this paper demonstrates how blockchain-based EHR sharing can reduce operational costs, improve patient outcomes, and enhance the security and privacy of healthcare data. This decentralized framework holds great potential for revolutionizing healthcare information exchange, providing a transparent, scalable, and secure method for managing patient data in a highly regulated environment.Keywords: blockchain, electronic health records (ehrs), fast healthcare interoperability resources (fhir), health information exchange (hie), hl7, interoperability, non-fungible tokens (nfts), privacy-preserving techniques, tokens, secure data sharing,
Procedia PDF Downloads 2125236 Deep Learning Based Polarimetric SAR Images Restoration
Authors: Hossein Aghababaei, Sergio Vitale, Giampaolo ferraioli
Abstract:
In the context of Synthetic Aperture Radar (SAR) data, polarization is an important source of information for Earth's surface monitoring . SAR Systems are often considered to transmit only one polarization. This constraint leads to either single or dual polarimetric SAR imaging modalities. Single polarimetric systems operate with a fixed single polarization of both transmitted and received electromagnetic (EM) waves, resulting in a single acquisition channel. Dual polarimetric systems, on the other hand, transmit in one fixed polarization and receive in two orthogonal polarizations, resulting in two acquisition channels. Dual polarimetric systems are obviously more informative than single polarimetric systems and are increasingly being used for a variety of remote sensing applications. In dual polarimetric systems, the choice of polarizations for the transmitter and the receiver is open. The choice of circular transmit polarization and coherent dual linear receive polarizations forms a special dual polarimetric system called hybrid polarimetry, which brings the properties of rotational invariance to geometrical orientations of features in the scene and optimizes the design of the radar in terms of reliability, mass, and power constraints. The complete characterization of target scattering, however, requires fully polarimetric data, which can be acquired with systems that transmit two orthogonal polarizations. This adds further complexity to data acquisition and shortens the coverage area or swath of fully polarimetric images compared to the swath of dual or hybrid polarimetric images. The search for solutions to augment dual polarimetric data to full polarimetric data will therefore take advantage of full characterization and exploitation of the backscattered field over a wider coverage with less system complexity. Several methods for reconstructing fully polarimetric images using hybrid polarimetric data can be found in the literature. Although the improvements achieved by the newly investigated and experimented reconstruction techniques are undeniable, the existing methods are, however, mostly based upon model assumptions (especially the assumption of reflectance symmetry), which may limit their reliability and applicability to vegetation and forest scenarios. To overcome the problems of these techniques, this paper proposes a new framework for reconstructing fully polarimetric information from hybrid polarimetric data. The framework uses Deep Learning solutions to augment hybrid polarimetric data without relying on model assumptions. A convolutional neural network (CNN) with a specific architecture and loss function is defined for this augmentation problem by focusing on different scattering properties of the polarimetric data. In particular, the method controls the CNN training process with respect to several characteristic features of polarimetric images defined by the combination of different terms in the cost or loss function. The proposed method is experimentally validated with real data sets and compared with a well-known and standard approach from the literature. From the experiments, the reconstruction performance of the proposed framework is superior to conventional reconstruction methods. The pseudo fully polarimetric data reconstructed by the proposed method also agree well with the actual fully polarimetric images acquired by radar systems, confirming the reliability and efficiency of the proposed method.Keywords: SAR image, deep learning, convolutional neural network, deep neural network, SAR polarimetry
Procedia PDF Downloads 9025235 RP-HPLC Method Development and Its Validation for Simultaneous Estimation of Metoprolol Succinate and Olmesartan Medoxomil Combination in Bulk and Tablet Dosage Form
Authors: S. Jain, R. Savalia, V. Saini
Abstract:
A simple, accurate, precise, sensitive and specific RP-HPLC method was developed and validated for simultaneous estimation of Metoprolol Succinate and Olmesartan Medoxomil in bulk and tablet dosage form. The RP-HPLC method has shown adequate separation for Metoprolol Succinate and Olmesartan Medoxomil from its degradation products. The separation was achieved on a Phenomenex luna ODS C18 (250mm X 4.6mm i.d., 5μm particle size) with an isocratic mixture of acetonitrile: 50mM phosphate buffer pH 4.0 adjusted with glacial acetic acid in the ratio of 55:45 v/v. The mobile phase at a flow rate of 1.0ml/min, Injection volume 20μl and wavelength of detection was kept at 225nm. The retention time for Metoprolol Succinate and Olmesartan Medoxomil was 2.451±0.1min and 6.167±0.1min, respectively. The linearity of the proposed method was investigated in the range of 5-50μg/ml and 2-20μg/ml for Metoprolol Succinate and Olmesartan Medoxomil, respectively. Correlation coefficient was 0.999 and 0.9996 for Metoprolol Succinate and Olmesartan Medoxomil, respectively. The limit of detection was 0.2847μg/ml and 0.1251μg/ml for Metoprolol Succinate and Olmesartan Medoxomil, respectively and the limit of quantification was 0.8630μg/ml and 0.3793μg/ml for Metoprolol and Olmesartan, respectively. Proposed methods were validated as per ICH guidelines for linearity, accuracy, precision, specificity and robustness for estimation of Metoprolol Succinate and Olmesartan Medoxomil in commercially available tablet dosage form and results were found to be satisfactory. Thus the developed and validated stability indicating method can be used successfully for marketed formulations.Keywords: metoprolol succinate, olmesartan medoxomil, RP-HPLC method, validation, ICH
Procedia PDF Downloads 315