Search results for: dynamic capability approach
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 17891

Search results for: dynamic capability approach

15581 Revenue Management of Perishable Products Considering Freshness and Price Sensitive Customers

Authors: Onur Kaya, Halit Bayer

Abstract:

Global grocery and supermarket sales are among the largest markets in the world and perishable products such as fresh produce, dairy and meat constitute the biggest section of these markets. Due to their deterioration over time, the demand for these products depends highly on their freshness. They become totally obsolete after a certain amount of time causing a high amount of wastage and decreases in grocery profits. In addition, customers are asking for higher product variety in perishable product categories, leading to less predictable demand per product and to more out-dating. Effective management of these perishable products is an important issue since it is observed that billions of dollars’ worth of food is expired and wasted every month. We consider coordinated inventory and pricing decisions for perishable products with a time and price dependent random demand function. We use stochastic dynamic programming to model this system for both periodically-reviewed and continuously-reviewed inventory systems and prove certain structural characteristics of the optimal solution. We prove that the optimal ordering decision scenario has a monotone structure and the optimal price value decreases by time. However, the optimal price changes in a non-monotonic structure with respect to inventory size. We also analyze the effect of 1 different parameters on the optimal solution through numerical experiments. In addition, we analyze simple-to-implement heuristics, investigate their effectiveness and extract managerial insights. This study gives valuable insights about the management of perishable products in order to decrease wastage and increase profits.

Keywords: age-dependent demand, dynamic programming, perishable inventory, pricing

Procedia PDF Downloads 247
15580 Applying Knowledge Management and Attitude Based on Holistic Approach in Learning Andragogy, as an Effort to Solve Environmental Problems after Mining Activities

Authors: Aloysius Hardoko, Susilo

Abstract:

The root cause of environmental damage post coal mining activities as determined by the province of East Kalimantan as a corridor of economic activity masterplan acceleration of economic development expansion (MP3EI) is the behavior of adults. Adult behavior can be changed through knowledge management and attitude. Based on the root of the problem, the objective of the research is to apply knowledge management and attitude based on holistic approach in learning andragogy as an effort to solve environmental problems after coal mining activities. Research methods to achieve the objective of using quantitative research with pretest posttest group design. Knowledge management and attitudes based on a holistic approach in adult learning are applied through initial learning activities, core and case-based cover of environmental damage. The research instrument is a description of the case of environmental damage. The data analysis uses t-test to see the effect of knowledge management attitude based on holistic approach before and after adult learning. Location and sample of representative research of adults as many as 20 people in Kutai Kertanegara District, one of the districts in East Kalimantan province, which suffered the worst environmental damage. The conclusion of the research result is the application of knowledge management and attitude in adult learning influence to adult knowledge and attitude to overcome environmental problem post coal mining activity.

Keywords: knowledge management and attitude, holistic approach, andragogy learning, environmental damage

Procedia PDF Downloads 242
15579 A Weighted Approach to Unconstrained Iris Recognition

Authors: Yao-Hong Tsai

Abstract:

This paper presents a weighted approach to unconstrained iris recognition. Nowadays, commercial systems are usually characterized by strong acquisition constraints based on the subject’s cooperation. However, it is not always achievable for real scenarios in our daily life. Researchers have been focused on reducing these constraints and maintaining the performance of the system by new techniques at the same time. With large variation in the environment, there are two main improvements to develop the proposed iris recognition system. For solving extremely uneven lighting condition, statistic based illumination normalization is first used on eye region to increase the accuracy of iris feature. The detection of the iris image is based on Adaboost algorithm. Secondly, the weighted approach is designed by Gaussian functions according to the distance to the center of the iris. Furthermore, local binary pattern (LBP) histogram is then applied to texture classification with the weight. Experiment showed that the proposed system provided users a more flexible and feasible way to interact with the verification system through iris recognition.

Keywords: authentication, iris recognition, adaboost, local binary pattern

Procedia PDF Downloads 225
15578 Wasting Human and Computer Resources

Authors: Mária Csernoch, Piroska Biró

Abstract:

The legends about “user-friendly” and “easy-to-use” birotical tools (computer-related office tools) have been spreading and misleading end-users. This approach has led us to the extremely high number of incorrect documents, causing serious financial losses in the creating, modifying, and retrieving processes. Our research proved that there are at least two sources of this underachievement: (1) The lack of the definition of the correctly edited, formatted documents. Consequently, end-users do not know whether their methods and results are correct or not. They are not aware of their ignorance. They are so ignorant that their ignorance does not allow them to realize their lack of knowledge. (2) The end-users’ problem-solving methods. We have found that in non-traditional programming environments end-users apply, almost exclusively, surface approach metacognitive methods to carry out their computer related activities, which are proved less effective than deep approach methods. Based on these findings we have developed deep approach methods which are based on and adapted from traditional programming languages. In this study, we focus on the most popular type of birotical documents, the text-based documents. We have provided the definition of the correctly edited text, and based on this definition, adapted the debugging method known in programming. According to the method, before the realization of text editing, a thorough debugging of already existing texts and the categorization of errors are carried out. With this method in advance to real text editing users learn the requirements of text-based documents and also of the correctly formatted text. The method has been proved much more effective than the previously applied surface approach methods. The advantages of the method are that the real text handling requires much less human and computer sources than clicking aimlessly in the GUI (Graphical User Interface), and the data retrieval is much more effective than from error-prone documents.

Keywords: deep approach metacognitive methods, error-prone birotical documents, financial losses, human and computer resources

Procedia PDF Downloads 382
15577 Deploying a Platform as a Service Cloud Solution to Support Student Learning

Authors: Jiangping Wang

Abstract:

This presentation describes the design and implementation of PaaS (platform as a service) cloud-based labs that are used in database-related courses to teach students practical skills. Traditionally, all labs are implemented in a desktop-based environment where students have to install heavy client software to access database servers. In order to release students from that burden, we have successfully deployed the cloud-based solution to support database-related courses, from which students and teachers can practice and learn database topics in various database courses via cloud access. With its development environment, execution runtime, web server, database server, and collaboration capability, it offers a shared pool of configurable computing resources and comprehensive environment that supports students’ needs without the complexity of maintaining the infrastructure.

Keywords: PaaS, database environment, e-learning, web server

Procedia PDF Downloads 268
15576 Perforation Analysis of the Aluminum Alloy Sheets Subjected to High Rate of Loading and Heated Using Thermal Chamber: Experimental and Numerical Approach

Authors: A. Bendarma, T. Jankowiak, A. Rusinek, T. Lodygowski, M. Klósak, S. Bouslikhane

Abstract:

The analysis of the mechanical characteristics and dynamic behavior of aluminum alloy sheet due to perforation tests based on the experimental tests coupled with the numerical simulation is presented. The impact problems (penetration and perforation) of the metallic plates have been of interest for a long time. Experimental, analytical as well as numerical studies have been carried out to analyze in details the perforation process. Based on these approaches, the ballistic properties of the material have been studied. The initial and residual velocities laser sensor is used during experiments to obtain the ballistic curve and the ballistic limit. The energy balance is also reported together with the energy absorbed by the aluminum including the ballistic curve and ballistic limit. The high speed camera helps to estimate the failure time and to calculate the impact force. A wide range of initial impact velocities from 40 up to 180 m/s has been covered during the tests. The mass of the conical nose shaped projectile is 28 g, its diameter is 12 mm, and the thickness of the aluminum sheet is equal to 1.0 mm. The ABAQUS/Explicit finite element code has been used to simulate the perforation processes. The comparison of the ballistic curve was obtained numerically and was verified experimentally, and the failure patterns are presented using the optimal mesh densities which provide the stability of the results. A good agreement of the numerical and experimental results is observed.

Keywords: aluminum alloy, ballistic behavior, failure criterion, numerical simulation

Procedia PDF Downloads 312
15575 Estimation of Harmonics in Three-Phase and Six-Phase-Phase (Multi-Phase) Load Circuits

Authors: Zakir Husain, Deepak Kumar

Abstract:

The harmonics are very harmful within an electrical system and can have serious consequences such as reducing the life of apparatus, stress on cable and equipment etc. This paper cites extensive analytical study of harmonic characteristics of multiphase (six-phase) and three-phase system equipped with two and three level inverters for non-linear loads. Multilevel inverter has elevated voltage capability with voltage limited devices, low harmonic distortion, abridged switching losses. Multiphase technology also pays a promising role in harmonic reduction. Matlab simulation is carried out to compare the advantage of multi-phase over three phase systems equipped with two or three level inverters for non-linear load harmonic reduction. The extensive simulation results are presented based on case studies.

Keywords: fast Fourier transform (FFT), harmonics, inverter, ripples, total harmonic distortion (THD)

Procedia PDF Downloads 553
15574 Printed Electronics for Enhanced Monitoring of Organ-on-Chip Culture Media Parameters

Authors: Alejandra Ben-Aissa, Martina Moreno, Luciano Sappia, Paul Lacharmoise, Ana Moya

Abstract:

Organ-on-Chip (OoC) stands out as a highly promising approach for drug testing, presenting a cost-effective and ethically superior alternative to conventional in vivo experiments. These cutting-edge devices emerge from the integration of tissue engineering and microfluidic technology, faithfully replicating the physiological conditions of targeted organs. Consequently, they offer a more precise understanding of drug responses without the ethical concerns associated with animal testing. When addressing the limitations of OoC due to conventional and time-consuming techniques, Lab-On-Chip (LoC) emerge as a disruptive technology capable of providing real-time monitoring without compromising sample integrity. This work develops LoC platforms that can be integrated within OoC platforms to monitor essential culture media parameters, including glucose, oxygen, and pH, facilitating the straightforward exchange of sensing units within a dynamic and controlled environment without disrupting cultures. This approach preserves the experimental setup, minimizes the impact on cells, and enables efficient, prolonged measurement. The LoC system is fabricated following the patented methodology protected by EU patent EP4317957A1. One of the key challenges of integrating sensors in a biocompatible, feasible, robust, and scalable manner is addressed through fully printed sensors, ensuring a customized, cost-effective, and scalable solution. With this technique, sensor reliability is enhanced, providing high sensitivity and selectivity for accurate parameter monitoring. In the present study, LoC is validated measuring a complete culture media. The oxygen sensor provided a measurement range from 0 mgO2/L to 6.3 mgO2/L. The pH sensor demonstrated a measurement range spanning 2 pH units to 9.5 pH units. Additionally, the glucose sensor achieved a measurement range from 0 mM to 11 mM. All the measures were performed with the sensors integrated in the LoC. In conclusion, this study showcases the impactful synergy of OoC technology with LoC systems using fully printed sensors, marking a significant step forward in ethical and effective biomedical research, particularly in drug development. This innovation not only meets current demands but also lays the groundwork for future advancements in precision and customization within scientific exploration.

Keywords: organ on chip, lab on chip, real time monitoring, biosensors

Procedia PDF Downloads 20
15573 River Stage-Discharge Forecasting Based on Multiple-Gauge Strategy Using EEMD-DWT-LSSVM Approach

Authors: Farhad Alizadeh, Alireza Faregh Gharamaleki, Mojtaba Jalilzadeh, Houshang Gholami, Ali Akhoundzadeh

Abstract:

This study presented hybrid pre-processing approach along with a conceptual model to enhance the accuracy of river discharge prediction. In order to achieve this goal, Ensemble Empirical Mode Decomposition algorithm (EEMD), Discrete Wavelet Transform (DWT) and Mutual Information (MI) were employed as a hybrid pre-processing approach conjugated to Least Square Support Vector Machine (LSSVM). A conceptual strategy namely multi-station model was developed to forecast the Souris River discharge more accurately. The strategy used herein was capable of covering uncertainties and complexities of river discharge modeling. DWT and EEMD was coupled, and the feature selection was performed for decomposed sub-series using MI to be employed in multi-station model. In the proposed feature selection method, some useless sub-series were omitted to achieve better performance. Results approved efficiency of the proposed DWT-EEMD-MI approach to improve accuracy of multi-station modeling strategies.

Keywords: river stage-discharge process, LSSVM, discrete wavelet transform, Ensemble Empirical Decomposition Mode, multi-station modeling

Procedia PDF Downloads 175
15572 Relative Effectiveness of Inquiry: Approach and Expository Instructional Methods in Fostering Students’ Retention in Chemistry

Authors: Joy Johnbest Egbo

Abstract:

The study was designed to investigate the relative effectiveness of inquiry role approach and expository instructional methods in fostering students’ retention in chemistry. Two research questions were answered and three null hypotheses were formulated and tested at 0.05 level of significance. A quasi-experimental (the non-equivalent pretest, posttest control group) design was adopted for the study. The population for the study comprised all senior secondary school class two (SS II) students who were offering Chemistry in single sex schools in Enugu Education Zone. The instrument for data collection was a self-developed Chemistry Retention Test (CRT). Relevant data were collected from a sample of one hundred and forty–one (141) students drawn from two secondary schools (1 male and 1 female schools) using simple random sampling technique. A reliability co-efficient of 0.82 was obtained for the instrument using Kuder Richardson formular20 (K-R20). Mean and Standard deviation scores were used to answer the research questions while two–way analysis of covariance (ANCOVA) was used to test the hypotheses. The findings showed that the students taught with Inquiry role approach retained the chemistry concept significantly higher than their counterparts taught with expository method. Female students retained slightly higher than their male counterparts. There is significant interaction between instructional packages and gender on Chemistry students’ retention. It was recommended, among others, that teachers should be encouraged to employ the use of Inquiry-role approach more in the teaching of chemistry and other subjects in general. By so doing, students’ retention of the subject could be increased.

Keywords: inquiry role approach, retention, exposition method, chemistry

Procedia PDF Downloads 513
15571 Copula-Based Estimation of Direct and Indirect Effects in Path Analysis Model

Authors: Alam Ali, Ashok Kumar Pathak

Abstract:

Path analysis is a statistical technique used to evaluate the strength of the direct and indirect effects of variables. One or more structural regression equations are used to estimate a series of parameters in order to find the better fit of data. Sometimes, exogenous variables do not show a significant strength of their direct and indirect effect when the assumption of classical regression (ordinary least squares (OLS)) are violated by the nature of the data. The main motive of this article is to investigate the efficacy of the copula-based regression approach over the classical regression approach and calculate the direct and indirect effects of variables when data violates the OLS assumption and variables are linked through an elliptical copula. We perform this study using a well-organized numerical scheme. Finally, a real data application is also presented to demonstrate the performance of the superiority of the copula approach.

Keywords: path analysis, copula-based regression models, direct and indirect effects, k-fold cross validation technique

Procedia PDF Downloads 72
15570 A Parametric Study of the Effect of Size, Position, and Number of Flexible Membranes Attached to a Circular Cylinder on the Fluid Flow Behavior

Authors: Nabaouia.Maktouf, Ali Ben Moussa, Saïd Turki

Abstract:

This paper discusses the effect of an attached flexible membrane on the control of fluid around a circular cylinder. A parametric study has been investigated for different positions, sizes, modes as well as frequencies of oscillation of the flexible membrane. The numerical investigation was conducted for a Reynolds number equal to 150 using the commercial code Fluent 16.0 and parallel calculation into 4 processors. The motion of the flexible membrane was managed by the dynamic mesh and compiled into Fluent as a user-defined function. The first part of this paper discusses the effect of changing the position of a flexible membrane sized 8° as an angle of aperture on the aerodynamic coefficients. Results show that the flexible membrane placed at 110° from the stagnation point presents more non-linearity on the behavior of the drag coefficient compared to the drag behavior when placed at 180°, relative to the stagnation point. The effect of the size of the flexible surface was studied for the corresponding angles of aperture: 32° and 42°, respectively. The effect of modes (modes 1, 2, and 3) of vibrations has been investigated at a constant frequency of vibration f=2Hz for angles 32° and 42°. All the calculations have been done with a constant amplitude A =0.001m. A non-linearity of the drag coefficient was clearly observed for all the sizes, modes as well as frequencies of excitation. The Fast Fourier transformation shows the appearance of the natural shedding frequency and the multiples of the frequency of excitation. An increase in the modes of oscillation leads to a more linear behavior of the drag coefficient.

Keywords: fluid flow control, numerical simulation, dynamic mesh, aerodynamic forces, flexible membrane

Procedia PDF Downloads 75
15569 Classification Earthquake Distribution in the Banda Sea Collision Zone with Point Process Approach

Authors: H. J. Wattimanela, U. S. Passaribu, N. T. Puspito, S. W. Indratno

Abstract:

Banda Sea collision zone (BSCZ) of is the result of the interaction and convergence of Indo-Australian plate, Eurasian plate and Pacific plate. This location in the eastern part of Indonesia. This zone has a very high seismic activity. In this research, we will be calculated rate (λ) and Mean Square Eror (MSE). By this result, we will identification of Poisson distribution of earthquakes in the BSCZ with the point process approach. Chi-square test approach and test Anscombe made in the process of identifying a Poisson distribution in the partition area. The data used are earthquakes with Magnitude ≥ 6 SR and its period 1964-2013 and sourced from BMKG Jakarta. This research is expected to contribute to the Moluccas Province and surrounding local governments in performing spatial plan document related to disaster management.

Keywords: molluca banda sea collision zone, earthquakes, mean square error, poisson distribution, chi-square test, anscombe test

Procedia PDF Downloads 300
15568 Laser Registration and Supervisory Control of neuroArm Robotic Surgical System

Authors: Hamidreza Hoshyarmanesh, Hosein Madieh, Sanju Lama, Yaser Maddahi, Garnette R. Sutherland, Kourosh Zareinia

Abstract:

This paper illustrates the concept of an algorithm to register specified markers on the neuroArm surgical manipulators, an image-guided MR-compatible tele-operated robot for microsurgery and stereotaxy. Two range-finding algorithms, namely time-of-flight and phase-shift, are evaluated for registration and supervisory control. The time-of-flight approach is implemented in a semi-field experiment to determine the precise position of a tiny retro-reflective moving object. The moving object simulates a surgical tool tip. The tool is a target that would be connected to the neuroArm end-effector during surgery inside the magnet bore of the MR imaging system. In order to apply flight approach, a 905-nm pulsed laser diode and an avalanche photodiode are utilized as the transmitter and receiver, respectively. For the experiment, a high frequency time to digital converter was designed using a field-programmable gate arrays. In the phase-shift approach, a continuous green laser beam with a wavelength of 530 nm was used as the transmitter. Results showed that a positioning error of 0.1 mm occurred when the scanner-target point distance was set in the range of 2.5 to 3 meters. The effectiveness of this non-contact approach exhibited that the method could be employed as an alternative for conventional mechanical registration arm. Furthermore, the approach is not limited by physical contact and extension of joint angles.

Keywords: 3D laser scanner, intraoperative MR imaging, neuroArm, real time registration, robot-assisted surgery, supervisory control

Procedia PDF Downloads 286
15567 Anatomically-Based Oropharyngeal Rehabilitation for the Patients with Obstructive Sleep Apnea Using a Multilevel Approach

Authors: Hsin-Yu Lin, Ching-Hsia Hung

Abstract:

Obstructive sleep apnea (OSA) is characterized by a complete or partial obstruction of the upper airway during sleep. The vulnerable sites of upper airway collapses are consequences of sleep state-dependent reductions in tone in specific pharyngeal dilators. Clinical examinations reveal multilevel collapses of the upper airway among the patients with OSA. Therefore, an anatomically-based oropharyngeal rehabilitation should comprise a multilevel approach, including retropalatal, retroglossal, hypopharyngeal, temporomandibular, and facial levels, all of which involve different muscle groups and contribute to multifunctional interaction and coordination, such as swallowing, breathing, and phonation. The purpose of the study was to exam the effects of this rehabilitation program with a multilevel approach. In this study, fifteen subjects with newly diagnosed moderate or severe OSA (Apnea-Hypopnea-Index≥15) were randomized into an intervention group and control group. The intervention group (N=8) underwent a 12-week-intervention of a hospital-based rehabilitation program, while the control group (N=7) was kept on the waiting list. The 12-week-intervention comprised an anatomically based multilevel approach. The primary outcome was Polysomnography (PSG) data, and the secondary outcome was oropharyngeal and respiratory muscle function. In the intervention group, Apnea-Hypopnea-Index significantly improved (46.96±19.45 versus 32.78±10.78 events/h, p=0.017) compared with control group (35.77±17.49 versus 42.96±17.32 events/h, p=0.043). While the control group remained no change, the intervention group demonstrated other PSG outcomes significantly improvement, including arousal index (46.04±18.9 versus 32.98±8.35/h, p=0.035), mean SpO2 (92.88±2.1 versus 94.13±1.46%, p=0.039). Besides, the intervention group demonstrated significant improvement in oropharyngeal and respiratory muscle function compared to the control group. This anatomically-based oropharyngeal rehabilitation with a multilevel approach can be proven as a non-invasive therapy for patients with OSA.

Keywords: obstructive sleep apnea, upper airway, oropharyngeal rehabilitation, multilevel approach

Procedia PDF Downloads 102
15566 Electromagnetic Simulation Based on Drift and Diffusion Currents for Real-Time Systems

Authors: Alexander Norbach

Abstract:

The script in this paper describes the use of advanced simulation environment using electronic systems (Microcontroller, Operational Amplifiers, and FPGA). The simulation may be used for all dynamic systems with the diffusion and the ionisation behaviour also. By additionally required observer structure, the system works with parallel real-time simulation based on diffusion model and the state-space representation for other dynamics. The proposed deposited model may be used for electrodynamic effects, including ionising effects and eddy current distribution also. With the script and proposed method, it is possible to calculate the spatial distribution of the electromagnetic fields in real-time. For further purpose, the spatial temperature distribution may be used also. With upon system, the uncertainties, unknown initial states and disturbances may be determined. This provides the estimation of the more precise system states for the required system, and additionally, the estimation of the ionising disturbances that occur due to radiation effects. The results have shown that a system can be also developed and adopted specifically for space systems with the real-time calculation of the radiation effects only. Electronic systems can take damage caused by impacts with charged particle flux in space or radiation environment. In order to be able to react to these processes, it must be calculated within a shorter time that ionising radiation and dose is present. All available sensors shall be used to observe the spatial distributions. By measured value of size and known location of the sensors, the entire distribution can be calculated retroactively or more accurately. With the formation, the type of ionisation and the direct effect to the systems and thus possible prevent processes can be activated up to the shutdown. The results show possibilities to perform more qualitative and faster simulations independent of kind of systems space-systems and radiation environment also. The paper gives additionally an overview of the diffusion effects and their mechanisms. For the modelling and derivation of equations, the extended current equation is used. The size K represents the proposed charge density drifting vector. The extended diffusion equation was derived and shows the quantising character and has similar law like the Klein-Gordon equation. These kinds of PDE's (Partial Differential Equations) are analytically solvable by giving initial distribution conditions (Cauchy problem) and boundary conditions (Dirichlet boundary condition). For a simpler structure, a transfer function for B- and E- fields was analytically calculated. With known discretised responses g₁(k·Ts) and g₂(k·Ts), the electric current or voltage may be calculated using a convolution; g₁ is the direct function and g₂ is a recursive function. The analytical results are good enough for calculation of fields with diffusion effects. Within the scope of this work, a proposed model of the consideration of the electromagnetic diffusion effects of arbitrary current 'waveforms' has been developed. The advantage of the proposed calculation of diffusion is the real-time capability, which is not really possible with the FEM programs available today. It makes sense in the further course of research to use these methods and to investigate them thoroughly.

Keywords: advanced observer, electrodynamics, systems, diffusion, partial differential equations, solver

Procedia PDF Downloads 131
15565 New Approach for Constructing a Secure Biometric Database

Authors: A. Kebbeb, M. Mostefai, F. Benmerzoug, Y. Chahir

Abstract:

The multimodal biometric identification is the combination of several biometric systems. The challenge of this combination is to reduce some limitations of systems based on a single modality while significantly improving performance. In this paper, we propose a new approach to the construction and the protection of a multimodal biometric database dedicated to an identification system. We use a topological watermarking to hide the relation between face image and the registered descriptors extracted from other modalities of the same person for more secure user identification.

Keywords: biometric databases, multimodal biometrics, security authentication, digital watermarking

Procedia PDF Downloads 391
15564 Route Planning for Optimization Approach PSO_GA Sharing System (Scooter Sharing-Public Transportation) with Hybrid Optimization Approach PSO_GA

Authors: Mohammad Ali Farrokhpour

Abstract:

In the current decade and sustainable transportation systems, scooter sharing has attracted widespread attention as an environmentally-friendly means of public transportation which can help develop public transportation. The combination of scooters and subway in the area of sustainable transportation systems can provide a great many opportunities for developing access to public transportation. Of the challenges which have arisen and initiated discussions of interest about the implementation of a scooter-subway system to replace personal vehicles is the issue of routing in the aforementioned system. This has been chosen as the main subject of the present paper. Thus, the present paper provides an account for routing in this system. Because the issue of routing includes multiple factors such as time, costs, traffic, green spaces, etc., the above-mentioned problem is considered to be a multi-objective NP-hard optimization problem. For this purpose, the hybrid optimization approach of PSO-GA has been put forward in the present paper for the provided answers to be of higher accuracy and validity than those of normal optimization methods. The results obtained from modeling and problem solving for the case study in the MATLAB software are indicative of the efficiency and desirability of the model and the proposed approach for solving the model

Keywords: route planning, scooter sharing, public transportation, sharing system

Procedia PDF Downloads 84
15563 Finding a Set of Long Common Substrings with Repeats from m Input Strings

Authors: Tiantian Li, Lusheng Wang, Zhaohui Zhan, Daming Zhu

Abstract:

In this paper, we propose two string problems, and study algorithms and complexity of various versions for those problems. Let S = {s₁, s₂, . . . , sₘ} be a set of m strings. A common substring of S is a substring appearing in every string in S. Given a set of m strings S = {s₁, s₂, . . . , sₘ} and a positive integer k, we want to find a set C of k common substrings of S such that the k common substrings in C appear in the same order and have no overlap among the m input strings in S, and the total length of the k common substring in C is maximized. This problem is referred to as the longest total length of k common substrings from m input strings (LCSS(k, m) for short). The other problem we study here is called the longest total length of a set of common substrings with length more than l from m input string (LSCSS(l, m) for short). Given a set of m strings S = {s₁, s₂, . . . , sₘ} and a positive integer l, for LSCSS(l, m), we want to find a set of common substrings of S, each is of length more than l, such that the total length of all the common substrings is maximized. We show that both problems are NP-hard when k and m are variables. We propose dynamic programming algorithms with time complexity O(k n₁n₂) and O(n₁n₂) to solve LCSS(k, 2) and LSCSS(l, 2), respectively, where n1 and n₂ are the lengths of the two input strings. We then design an algorithm for LSCSS(l, m) when every length > l common substring appears once in each of the m − 1 input strings. The running time is O(n₁²m), where n1 is the length of the input string with no restriction on length > l common substrings. Finally, we propose a fixed parameter algorithm for LSCSS(l, m), where each length > l common substring appears m − 1 + c times among the m − 1 input strings (other than s1). In other words, each length > l common substring may repeatedly appear at most c times among the m − 1 input strings {s₂, s₃, . . . , sₘ}. The running time of the proposed algorithm is O((n12ᶜ)²m), where n₁ is the input string with no restriction on repeats. The LSCSS(l, m) is proposed to handle whole chromosome sequence alignment for different strains of the same species, where more than 98% of letters in core regions are identical.

Keywords: dynamic programming, algorithm, common substrings, string

Procedia PDF Downloads 16
15562 Synthesis of Polystyrene Grafted Filler Nanoparticles: Effect of Grafting on Mechanical Reinforcement

Authors: M. Khlifa, A. Youssef, A. F. Zaed, A. Kraft, V. Arrighi

Abstract:

A series of PS-nanoparticles were prepared by grafting PS from both aggregated silica and colloidally silica using atom-transfer radical polymerisation (ATRP). The mechanical behaviour of the nanocomposites have been examined by differential scanning calorimetry (DSC)and dynamic mechanical thermal analysis (DMTA).

Keywords: ATRP, nanocomposites, polystyrene, reinforcement

Procedia PDF Downloads 626
15561 An Adaptive Controller Method Based on Full-State Linear Model of Variable Cycle Engine

Authors: Jia Li, Huacong Li, Xiaobao Han

Abstract:

Due to the more variable geometry parameters of VCE (variable cycle aircraft engine), presents an adaptive controller method based on the full-state linear model of VCE and has simulated to solve the multivariate controller design problem of the whole flight envelops. First, analyzes the static and dynamic performances of bypass ratio and other state parameters caused by variable geometric components, and develops nonlinear component model of VCE. Then based on the component model, through small deviation linearization of main fuel (Wf), the area of tail nozzle throat (A8) and the angle of rear bypass ejector (A163), setting up multiple linear model which variable geometric parameters can be inputs. Second, designs the adaptive controllers for VCE linear models of different nominal points. Among them, considering of modeling uncertainties and external disturbances, derives the adaptive law by lyapunov function. The simulation results showed that, the adaptive controller method based on full-state linear model used the angle of rear bypass ejector as input and effectively solved the multivariate control problems of VCE. The performance of all nominal points could track the desired closed-loop reference instructions. The adjust time was less than 1.2s, and the system overshoot was less than 1%, at the same time, the errors of steady states were less than 0.5% and the dynamic tracking errors were less than 1%. In addition, the designed controller could effectively suppress interference and reached the desired commands with different external random noise signals.

Keywords: variable cycle engine (VCE), full-state linear model, adaptive control, by-pass ratio

Procedia PDF Downloads 318
15560 Cryptocurrency Realities: Insights from Social and Economic Psychology

Authors: Sarah Marie

Abstract:

In today's dynamic financial landscape, cryptocurrencies represent a paradigm shift characterized by innovation and intense debate. This study probes into their transformative potential and the challenges they present, offering a balanced perspective that recognizes both their promise and pitfalls. Emulating the engaging style of a TED Talk, this research goes beyond academic analysis, serving as a critical bridge to reconcile the perspectives of cryptocurrency skeptics and enthusiasts, fostering a well-informed dialogue. The study employs a mixed-method approach, analyzing current trends, regulatory landscapes, and public perceptions in the cryptocurrency domain. It distinguishes genuine innovators in this field from ostentatious opportunists, echoing the sentiment that real innovation should be separated from mere showmanship. If one is unfamiliar with who is being referenced, they can likely spot them leaning against their Lamborghinis outside "Crypto" conventions, looking greasy. Major findings reveal a complex scenario dominated by regulatory uncertainties, market volatility, and security issues, emphasizing the need for a coherent regulatory framework that balances innovation with risk management and sustainable practices. The study underscores the importance of transparency and consumer protection in fostering responsible growth within the cryptocurrency ecosystem. In conclusion, the research advocates for education, innovation, and ethical governance in the realm of cryptocurrencies. It calls for collaborative efforts to navigate the intricacies of this evolving landscape and to realize its full potential in a responsible, inclusive, and forward-thinking manner.

Keywords: financial landscape, innovation, public perception, transparency

Procedia PDF Downloads 52
15559 A Review of Research on Pre-training Technology for Natural Language Processing

Authors: Moquan Gong

Abstract:

In recent years, with the rapid development of deep learning, pre-training technology for natural language processing has made great progress. The early field of natural language processing has long used word vector methods such as Word2Vec to encode text. These word vector methods can also be regarded as static pre-training techniques. However, this context-free text representation brings very limited improvement to subsequent natural language processing tasks and cannot solve the problem of word polysemy. ELMo proposes a context-sensitive text representation method that can effectively handle polysemy problems. Since then, pre-training language models such as GPT and BERT have been proposed one after another. Among them, the BERT model has significantly improved its performance on many typical downstream tasks, greatly promoting the technological development in the field of natural language processing, and has since entered the field of natural language processing. The era of dynamic pre-training technology. Since then, a large number of pre-trained language models based on BERT and XLNet have continued to emerge, and pre-training technology has become an indispensable mainstream technology in the field of natural language processing. This article first gives an overview of pre-training technology and its development history, and introduces in detail the classic pre-training technology in the field of natural language processing, including early static pre-training technology and classic dynamic pre-training technology; and then briefly sorts out a series of enlightening technologies. Pre-training technology, including improved models based on BERT and XLNet; on this basis, analyze the problems faced by current pre-training technology research; finally, look forward to the future development trend of pre-training technology.

Keywords: natural language processing, pre-training, language model, word vectors

Procedia PDF Downloads 57
15558 Stacking Ensemble Approach for Combining Different Methods in Real Estate Prediction

Authors: Sol Girouard, Zona Kostic

Abstract:

A home is often the largest and most expensive purchase a person makes. Whether the decision leads to a successful outcome will be determined by a combination of critical factors. In this paper, we propose a method that efficiently handles all the factors in residential real estate and performs predictions given a feature space with high dimensionality while controlling for overfitting. The proposed method was built on gradient descent and boosting algorithms and uses a mixed optimizing technique to improve the prediction power. Usually, a single model cannot handle all the cases thus our approach builds multiple models based on different subsets of the predictors. The algorithm was tested on 3 million homes across the U.S., and the experimental results demonstrate the efficiency of this approach by outperforming techniques currently used in forecasting prices. With everyday changes on the real estate market, our proposed algorithm capitalizes from new events allowing more efficient predictions.

Keywords: real estate prediction, gradient descent, boosting, ensemble methods, active learning, training

Procedia PDF Downloads 277
15557 A Comparative Analysis of ARIMA and Threshold Autoregressive Models on Exchange Rate

Authors: Diteboho Xaba, Kolentino Mpeta, Tlotliso Qejoe

Abstract:

This paper assesses the in-sample forecasting of the South African exchange rates comparing a linear ARIMA model and a SETAR model. The study uses a monthly adjusted data of South African exchange rates with 420 observations. Akaike information criterion (AIC) and the Schwarz information criteria (SIC) are used for model selection. Mean absolute error (MAE), root mean squared error (RMSE) and mean absolute percentage error (MAPE) are error metrics used to evaluate forecast capability of the models. The Diebold –Mariano (DM) test is employed in the study to check forecast accuracy in order to distinguish the forecasting performance between the two models (ARIMA and SETAR). The results indicate that both models perform well when modelling and forecasting the exchange rates, but SETAR seemed to outperform ARIMA.

Keywords: ARIMA, error metrices, model selection, SETAR

Procedia PDF Downloads 244
15556 A Multi-Agent Intelligent System for Monitoring Health Conditions of Elderly People

Authors: Ayman M. Mansour

Abstract:

In this paper, we propose a multi-agent intelligent system that is used for monitoring the health conditions of elderly people. Monitoring the health condition of elderly people is a complex problem that involves different medical units and requires continuous monitoring. Such expert system is highly needed in rural areas because of inadequate number of available specialized physicians or nurses. Such monitoring must have autonomous interactions between these medical units in order to be effective. A multi-agent system is formed by a community of agents that exchange information and proactively help one another to achieve the goal of elderly monitoring. The agents in the developed system are equipped with intelligent decision maker that arms them with the rule-based reasoning capability that can assist the physicians in making decisions regarding the medical condition of elderly people.

Keywords: fuzzy logic, inference system, monitoring system, multi-agent system

Procedia PDF Downloads 608
15555 Performing the Landscape: Temporary and Performative Practices in Landscape Production

Authors: Miguel Costa

Abstract:

Despite the "time" element being an intrinsic characteristic of the work with the landscape, its execution and completion are also often dependent on external factors, i.e., the slow bureaucratic procedures required for the implementation of a project. In the urban areas of the city, these conditions are even more present — some landscape projects are articulated with the architectural/urban design, transporting itself long, expensive and inflexible processes related with the constant transformations of contemporary urban culture, where the needs and expectations could change before the project is finished. However, despite the renewed interest and growing concern for issues related to the landscapes (particularly since the European Landscape Convention, its scope and fields of action, extended to all the landscapes and not just the selected ones), still lacks the need for a greater inclusion of citizens in its protection and construction processes as well as a greater transparency and clarity of the consequences and results of their active participation. This article aims to reflect on the production processes of urban landscapes, on its completion runtime and its relationship with the citizens by introducing temporary projects as a fieldwork methodology, as well as using the contribution of different professional practices and knowledge for its monitoring, execution, and implementation. These strategies address a more interdisciplinary, transdisciplinary and performative approach, not only from the ephemeral experience of objects and actions but also from the processes and the dynamic events that are organized from these objects and actions over the landscape. The goal is to discuss the results of these approaches on its different dimensions: critical dimension; experimental and strategic dimension; pedagogical dimension; political dimension; cultural.

Keywords: landscape fieldwork, interdisciplinarity, public inclusion, public participation, temporary projects, transdisciplinarity

Procedia PDF Downloads 323
15554 A Lightweight Blockchain: Enhancing Internet of Things Driven Smart Buildings Scalability and Access Control Using Intelligent Direct Acyclic Graph Architecture and Smart Contracts

Authors: Syed Irfan Raza Naqvi, Zheng Jiangbin, Ahmad Moshin, Pervez Akhter

Abstract:

Currently, the IoT system depends on a centralized client-servant architecture that causes various scalability and privacy vulnerabilities. Distributed ledger technology (DLT) introduces a set of opportunities for the IoT, which leads to practical ideas for existing components at all levels of existing architectures. Blockchain Technology (BCT) appears to be one approach to solving several IoT problems, like Bitcoin (BTC) and Ethereum, which offer multiple possibilities. Besides, IoTs are resource-constrained devices with insufficient capacity and computational overhead to process blockchain consensus mechanisms; the traditional BCT existing challenge for IoTs is poor scalability, energy efficiency, and transaction fees. IOTA is a distributed ledger based on Direct Acyclic Graph (DAG) that ensures M2M micro-transactions are free of charge. IOTA has the potential to address existing IoT-related difficulties such as infrastructure scalability, privacy and access control mechanisms. We proposed an architecture, SLDBI: A Scalable, lightweight DAG-based Blockchain Design for Intelligent IoT Systems, which adapts the DAG base Tangle and implements a lightweight message data model to address the IoT limitations. It enables the smooth integration of new IoT devices into a variety of apps. SLDBI enables comprehensive access control, energy efficiency, and scalability in IoT ecosystems by utilizing the Masked Authentication Message (MAM) protocol and the IOTA Smart Contract Protocol (ISCP). Furthermore, we suggest proof-of-work (PoW) computation on the full node in an energy-efficient way. Experiments have been carried out to show the capability of a tangle to achieve better scalability while maintaining energy efficiency. The findings show user access control management at granularity levels and ensure scale up to massive networks with thousands of IoT nodes, such as Smart Connected Buildings (SCBDs).

Keywords: blockchain, IOT, direct acyclic graphy, scalability, access control, architecture, smart contract, smart connected buildings

Procedia PDF Downloads 122
15553 Isolation and Molecular Identification of Two Fungal Strains Capable of Degrading hydrocarbon Contaminants on Saudi Arabian Environment

Authors: Amr A. EL Hanafy, Yasir Anwar, Saleh A. Mohamed, Saleh Mohamed Saleh Al-Garni, Jamal S. M. Sabir , Osama A. H. Abu Zinadah, Mohamed Morsi Ahmed

Abstract:

In the vicinity of the red sea about 15 fungi species were isolated from oil contaminated sites. On the basis of aptitude to degrade the crude oil and DCPIP assay, two fungal isolates were selected amongst 15 oil degrading strains. Analysis of ITS-1, ITS-2 and amplicon pyrosequencing studies of fungal diversity revealed that these strains belong to Penicillium and Aspergillus species. Two strains that proved to be the most efficient in degrading crude oil was Aspergillus niger (54 %) and Penicillium commune (48 %) Subsequent to two weeks of cultivation in BHS medium the degradation rate were recorded by using spectrophotometer and GC-MS. Hence, it is cleared that these fungal strains has the capability of degradation and can be utilized for cleaning the Saudi Arabian environment.

Keywords: fungal strains, hydrocarbon contaminants, molecular identification, biodegradation, GC-MS

Procedia PDF Downloads 523
15552 A Study of Lurking Behavior: The Desire Perspective

Authors: Hsiu-Hua Cheng, Chi-Wei Chen

Abstract:

Lurking behaviour is common in information-seeking oriented communities. Transferring users with lurking behaviour to be contributors can assist virtual communities to obtain competitive advantages. Based on the ecological cognition framework, this study proposes a model to examine the antecedents of lurking behaviour in information-seeking oriented virtual communities. This study argues desire for emotional support, desire for information support, desire for performance-approach, desire for performance -avoidance, desire for mastery-approach, desire for mastery-avoidance, desire for ability trust, desire for benevolence trust, and desire for integrity trust effect on lurking behaviour. This study offers an approach to understanding the determinants of lurking behaviour in online contexts.

Keywords: lurking behaviour, the ecological cognition framework, Information-seeking oriented virtual communities, desire

Procedia PDF Downloads 278