Search results for: low order model
23385 Hull Detection from Handwritten Digit Image
Authors: Sriraman Kothuri, Komal Teja Mattupalli
Abstract:
In this paper we proposed a novel algorithm for recognizing hulls in a hand written digits. This is an extension to the work on “Digit Recognition Using Freeman Chain code”. In order to find out the hulls in a user given digit it is necessary to follow three steps. Those are pre-processing, Boundary Extraction and at last apply the Hull Detection system in a way to attain the better results. The detection of Hull Regions is mainly intended to increase the machine learning capability in detection of characters or digits. This can also extend this in order to get the hull regions and their intensities in Black Holes in Space Exploration.Keywords: chain code, machine learning, hull regions, hull recognition system, SASK algorithm
Procedia PDF Downloads 40023384 GPS Devices to Increase Efficiency of Indian Auto-Rickshaw Segment
Authors: Sanchay Vaidya, Sourabh Gupta, Gouresh Singhal
Abstract:
There are various modes of transport in metro cities in India, auto-rickshaws being one of them. Auto-rickshaws provide connectivity to all the places in the city offering last mile connectivity. Among all the modes of transport, the auto-rickshaw industry is the most unorganized and inefficient. Although unions exist in different cities they aren’t good enough to cope up with the upcoming advancements in the field of technology. An introduction of simple technology in this field may do wonder and help increase the revenues. This paper aims to organize this segment under a single umbrella using GPS devices and mobile phones. The paper includes surveys of about 300 auto-rickshaw drivers and 1000 plus commuters across 6 metro cities in India. Carrying out research and analysis provides a base for the development of this model and implementation of this innovative technique, which is discussed in this paper in detail with ample emphasis given on the implementation of this model.Keywords: auto-rickshaws, business model, GPS device, mobile application
Procedia PDF Downloads 22723383 A Novel Approach of Power Transformer Diagnostic Using 3D FEM Parametrical Model
Authors: M. Brandt, A. Peniak, J. Makarovič, P. Rafajdus
Abstract:
This paper deals with a novel approach of power transformers diagnostics. This approach identifies the exact location and the range of a fault in the transformer and helps to reduce operation costs related to handling of the faulty transformer, its disassembly and repair. The advantage of the approach is a possibility to simulate healthy transformer and also all faults, which can occur in transformer during its operation without its disassembling, which is very expensive in practice. The approach is based on creating frequency dependent impedance of the transformer by sweep frequency response analysis measurements and by 3D FE parametrical modeling of the fault in the transformer. The parameters of the 3D FE model are the position and the range of the axial short circuit. Then, by comparing the frequency dependent impedances of the parametrical models with the measured ones, the location and the range of the fault is identified. The approach was tested on a real transformer and showed high coincidence between the real fault and the simulated one.Keywords: transformer, parametrical model of transformer, fault, sweep frequency response analysis, finite element method
Procedia PDF Downloads 48323382 Excitation Modeling for Hidden Markov Model-Based Speech Synthesis Based on Wavelet Analysis
Authors: M. Kiran Reddy, K. Sreenivasa Rao
Abstract:
The conventional Hidden Markov Model (HMM)-based speech synthesis system (HTS) uses only a pulse excitation model, which significantly differs from natural excitation signal. Hence, buzziness can be perceived in the speech generated using HTS. This paper proposes an efficient excitation modeling method that can significantly reduce the buzziness, and improve the quality of HMM-based speech synthesis. The proposed approach models the pitch-synchronous residual frames extracted from the residual excitation signal. Each pitch synchronous residual frame is parameterized using 30 wavelet coefficients. These 30 wavelet coefficients are found to accurately capture the perceptually important information present in the residual waveform. In synthesis phase, the residual frames are reconstructed from the generated wavelet coefficients and are pitch-synchronously overlap-added to generate the excitation signal. The proposed excitation modeling method is integrated into HMM-based speech synthesis system. Evaluation results indicate that the speech synthesized by the proposed excitation model is significantly better than the speech generated using state-of-the-art excitation modeling methods.Keywords: excitation modeling, hidden Markov models, pitch-synchronous frames, speech synthesis, wavelet coefficients
Procedia PDF Downloads 24823381 CoP-Networks: Virtual Spaces for New Faculty’s Professional Development in the 21st Higher Education
Authors: Eman AbuKhousa, Marwan Z. Bataineh
Abstract:
The 21st century higher education and globalization challenge new faculty members to build effective professional networks and partnership with industry in order to accelerate their growth and success. This creates the need for community of practice (CoP)-oriented development approaches that focus on cognitive apprenticeship while considering individual predisposition and future career needs. This work adopts data mining, clustering analysis, and social networking technologies to present the CoP-Network as a virtual space that connects together similar career-aspiration individuals who are socially influenced to join and engage in a process for domain-related knowledge and practice acquisitions. The CoP-Network model can be integrated into higher education to extend traditional graduate and professional development programs.Keywords: clustering analysis, community of practice, data mining, higher education, new faculty challenges, social network, social influence, professional development
Procedia PDF Downloads 18323380 Adopting the Transition Management Model as a Tool for Sustainable Groundwater Management in Nigeria
Authors: Ali Bakari Mohammed
Abstract:
Transitioning is a continuous process of radical change in a society which involves co-evolution of institutional, technological, socio-cultural, and ecological developments at different scales and levels. Transition management model is a methodology that influences structural change of complex systems over a period (0-30 years) by experimenting and implementing new techniques. A transition management in the context of groundwater is a radical change from the current operate and control system to a next generation integrated and sustainable system that takes into account quality protection and sustained supply into the future. This study evaluates the transition management model in adopting it as a viable tool for the attainment of sustainable groundwater management. The outcome of the evaluation shows that there are three levels (strategic, tactical and operational) of operating the transition management model. At the strategic level, long-term goals for sustainable groundwater management are formulated, at the tactical level activities such as inter institutional networking, negotiation, planning and financing are carried out, and at the operational level, transition experiments and strategic niche management are carried out at the societal level. Overall, different actors and set of activities are required to partake at each management level. The outcome of this paper will provide basis for the implementation of the Sustainable Development Goal (SDG) 6 in Nigeria.Keywords: transition management, groundwater, sustainable management, tool, Nigeria
Procedia PDF Downloads 27123379 The Toxic Effects of Kynurenine Metabolites on SH-SY5Y Neuroblastoma Cells
Authors: Susan Hall, Gary D. Grant, Catherine McDermott, Devinder Arora
Abstract:
Introduction /Aim: The kynurenine pathway is thought to play an important role in the pathophysiology of numerous neurodegenerative diseases including depression, Alzheimer’s disease, and Parkinson’s disease. Numerous neuroactive compounds, including the neurotoxic 3-hydroxyanthranilic acid, 3-hydroxykynurenine and quinolinic acid and the neuroprotective kynurenic acid and picolinic acid, are produced through the metabolism of kynurenine and are thought to be the causative agents responsible for neurodegeneration. The toxicity of 3-hydroxykynurenine, 3-hydroxyanthranilic acid and quinolinic acid has been widely evaluated and demonstrated in primary cell cultures but to date only 3-hydroxykynurenine and 3-hydroxyanthranilic acid have been shown to cause toxicity in immortal tumour cells. The aim of this study was to evaluate the toxicity of kynurenine metabolites, both individually and in combination, on SH-SY5Y neuroblastoma cells after 24 and 72 h exposure in order to explore a cost-effective model to study their neurotoxic effects and potential protective agents. Methods: SH-SY5Y neuroblastoma cells were exposed to various concentrations of the neuroactive kynurenine metabolites, both individually and in combination, for 24 and 72 h, and viability was subsequently evaluated using the Resazurin (Alamar blue) proliferation assay. Furthermore, the effects of these compounds, alone and in combination, on specific death pathways including apoptosis, necrosis and free radical production was evaluated using various assays. Results: Consistent with literature, toxicity was shown with short-term 24-hour treatments at 1000 μM concentrations for both 3-hydroxykynurenine and 3-hydroxyanthranilic acid. Combinations of kynurenine metabolites showed modest toxicity towards SH-SY5Y neuroblastoma cells in a concentration-dependent manner. Specific cell death pathways, including apoptosis, necrosis and free radical production were shown to be increased after both 24 and 72 h exposure of SH-SY5Y neuroblastoma cells to 3-hydroxykynurenine and 3-hydroxyanthranilic acid and various combinations of neurotoxic kynurenine metabolites. Conclusion: It is well documented that neurotoxic kynurenine metabolites show toxicity towards primary human neurons in the nanomolar to low micromolar concentration range. Results show that the concentrations required to show significant cell death are in the range of 1000 µM for 3-hydroxykynurenine and 3-hydroxyanthranilic acid and toxicity of quinolinic acid towards SH-SY5Y was unable to be shown. This differs significantly from toxicities observed in primary human neurons. Combinations of the neurotoxic metabolites were shown to have modest toxicity towards these cells with increased toxicity and activation of cell death pathways observed after 72 h exposure. This study suggests that the 24 h model is unsuitable for use in neurotoxicity studies, however, the 72 h model better represents the observations of the studies using primary human neurons and may provide some benefit in providing a cost-effective model to assess possible protective agents against kynurenine metabolite toxicities.Keywords: kynurenine metabolites, neurotoxicity, quinolinic acid, SH-SY5Y neuroblastoma
Procedia PDF Downloads 41723378 Uplift Segmentation Approach for Targeting Customers in a Churn Prediction Model
Authors: Shivahari Revathi Venkateswaran
Abstract:
Segmenting customers plays a significant role in churn prediction. It helps the marketing team with proactive and reactive customer retention. For the reactive retention, the retention team reaches out to customers who already showed intent to disconnect by giving some special offers. When coming to proactive retention, the marketing team uses churn prediction model, which ranks each customer from rank 1 to 100, where 1 being more risk to churn/disconnect (high ranks have high propensity to churn). The churn prediction model is built by using XGBoost model. However, with the churn rank, the marketing team can only reach out to the customers based on their individual ranks. To profile different groups of customers and to frame different marketing strategies for targeted groups of customers are not possible with the churn ranks. For this, the customers must be grouped in different segments based on their profiles, like demographics and other non-controllable attributes. This helps the marketing team to frame different offer groups for the targeted audience and prevent them from disconnecting (proactive retention). For segmentation, machine learning approaches like k-mean clustering will not form unique customer segments that have customers with same attributes. This paper finds an alternate approach to find all the combination of unique segments that can be formed from the user attributes and then finds the segments who have uplift (churn rate higher than the baseline churn rate). For this, search algorithms like fast search and recursive search are used. Further, for each segment, all customers can be targeted using individual churn ranks from the churn prediction model. Finally, a UI (User Interface) is developed for the marketing team to interactively search for the meaningful segments that are formed and target the right set of audience for future marketing campaigns and prevent them from disconnecting.Keywords: churn prediction modeling, XGBoost model, uplift segments, proactive marketing, search algorithms, retention, k-mean clustering
Procedia PDF Downloads 7123377 Determination of Activation Energy for Thermal Decomposition of Selected Soft Tissues Components
Authors: M. Ekiert, T. Uhl, A. Mlyniec
Abstract:
Tendons are the biological soft tissue structures composed of collagen, proteoglycan, glycoproteins, water and cells of extracellular matrix (ECM). Tendons, which primary function is to transfer force generated by the muscles to the bones causing joints movement, are exposed to many micro and macro damages. In fact, tendons and ligaments trauma are one of the most numerous injuries of human musculoskeletal system, causing for many people (particularly for athletes and physically active people), recurring disorders, chronic pain or even inability of movement. The number of tendons reconstruction and transplantation procedures is increasing every year. Therefore, studies on soft tissues storage conditions (influencing i.e. tissue aging) seem to be an extremely important issue. In this study, an atomic-scale investigation on the kinetics of decomposition of two selected tendon components – collagen type I (which forms a 60-85% of a tendon dry mass) and elastin protein (which combine with ECM creates elastic fibers of connective tissues) is presented. A molecular model of collagen and elastin was developed based on crystal structure of triple-helical collagen-like 1QSU peptide and P15502 human elastin protein, respectively. Each model employed 4 linear strands collagen/elastin strands per unit cell, distributed in 2x2 matrix arrangement, placed in simulation box filled with water molecules. A decomposition phenomena was simulated with molecular dynamics (MD) method using ReaxFF force field and periodic boundary conditions. A set of NVT-MD runs was performed for 1000K temperature range in order to obtained temperature-depended rate of production of decomposition by-products. Based on calculated reaction rates activation energies and pre-exponential factors, required to formulate Arrhenius equations describing kinetics of decomposition of tested soft tissue components, were calculated. Moreover, by adjusting a model developed for collagen, system scalability and correct implementation of the periodic boundary conditions were evaluated. An obtained results provide a deeper insight into decomposition of selected tendon components. A developed methodology may also be easily transferred to other connective tissue elements and therefore might be used for further studies on soft tissues aging.Keywords: decomposition, molecular dynamics, soft tissue, tendons
Procedia PDF Downloads 21023376 Agriculture and Global Economy vis-à-vis the Climate Change
Authors: Assaad Ghazouani, Ati Abdessatar
Abstract:
In the world, agriculture maintains a social and economic importance in the national economy. Its importance is distinguished by its ripple effects not only downstream but also upstream vis-à-vis the non-agricultural sector. However, the situation is relatively fragile because of weather conditions. In this work, we propose a model to highlight the impacts of climate change (CC) on economic growth in the world where agriculture is considered as a strategic sector. The CC is supposed to directly and indirectly affect economic growth by reducing the performance of the agricultural sector. The model is tested for Tunisia. The results validate the hypothesis that the potential economic damage of the CC is important. Indeed, an increase in CO2 concentration (temperatures and disruption of rainfall patterns) will have an impact on global economic growth particularly by reducing the performance of the agricultural sector. Analysis from a vector error correction model also highlights the magnitude of climate impact on the performance of the agricultural sector and its repercussions on economic growthKeywords: Climate Change, Agriculture, Economic Growth, World, VECM, Cointegration.
Procedia PDF Downloads 61923375 A Selection Approach: Discriminative Model for Nominal Attributes-Based Distance Measures
Authors: Fang Gong
Abstract:
Distance measures are an indispensable part of many instance-based learning (IBL) and machine learning (ML) algorithms. The value difference metrics (VDM) and inverted specific-class distance measure (ISCDM) are among the top-performing distance measures that address nominal attributes. VDM performs well in some domains owing to its simplicity and poorly in others that exist missing value and non-class attribute noise. ISCDM, however, typically works better than VDM on such domains. To maximize their advantages and avoid disadvantages, in this paper, a selection approach: a discriminative model for nominal attributes-based distance measures is proposed. More concretely, VDM and ISCDM are built independently on a training dataset at the training stage, and the most credible one is recorded for each training instance. At the test stage, its nearest neighbor for each test instance is primarily found by any of VDM and ISCDM and then chooses the most reliable model of its nearest neighbor to predict its class label. It is simply denoted as a discriminative distance measure (DDM). Experiments are conducted on the 34 University of California at Irvine (UCI) machine learning repository datasets, and it shows DDM retains the interpretability and simplicity of VDM and ISCDM but significantly outperforms the original VDM and ISCDM and other state-of-the-art competitors in terms of accuracy.Keywords: distance measure, discriminative model, nominal attributes, nearest neighbor
Procedia PDF Downloads 11423374 Sea-Spray Calculations Using the MESO-NH Model
Authors: Alix Limoges, William Bruch, Christophe Yohia, Jacques Piazzola
Abstract:
A number of questions arise concerning the long-term impact of the contribution of marine aerosol fluxes generated at the air-sea interface on the occurrence of intense events (storms, floods, etc.) in the coastal environment. To this end, knowledge is needed on sea-spray emission rates and the atmospheric dynamics of the corresponding particles. Our aim is to implement the mesoscale model MESO-NH on the study area using an accurate sea-spray source function to estimate heat fluxes and impact on the precipitations. Based on an original and complete sea-spray source function, which covers a large size spectrum since taking into consideration the sea-spray produced by both bubble bursting and surface tearing process, we propose a comparison between model simulations and experimental data obtained during an oceanic scientific cruise on board the navy ship Atalante. The results show the relevance of the sea-spray flux calculations as well as their impact on the heat fluxes and AOD.Keywords: atmospheric models, sea-spray source, sea-spray dynamics, aerosols
Procedia PDF Downloads 14923373 The Intention to Use Telecare in People of Fall Experience: Application of Fuzzy Neural Network
Authors: Jui-Chen Huang, Shou-Hsiung Cheng
Abstract:
This study examined their willingness to use telecare for people who have had experience falling in the last three months in Taiwan. This study adopted convenience sampling and a structural questionnaire to collect data. It was based on the definition and the constructs related to the Health Belief Model (HBM). HBM is comprised of seven constructs: perceived benefits (PBs), perceived disease threat (PDT), perceived barriers of taking action (PBTA), external cues to action (ECUE), internal cues to action (ICUE), attitude toward using (ATT), and behavioral intention to use (BI). This study adopted Fuzzy Neural Network (FNN) to put forward an effective method. It shows the dependence of ATT on PB, PDT, PBTA, ECUE, and ICUE. The training and testing data RMSE (root mean square error) are 0.028 and 0.166 in the FNN, respectively. The training and testing data RMSE are 0.828 and 0.578 in the regression model, respectively. On the other hand, as to the dependence of ATT on BI, as presented in the FNN, the training and testing data RMSE are 0.050 and 0.109, respectively. The training and testing data RMSE are 0.529 and 0.571 in the regression model, respectively. The results show that the FNN method is better than the regression analysis. It is an effective and viable good way.Keywords: fall, fuzzy neural network, health belief model, telecare, willingness
Procedia PDF Downloads 20123372 The Case for Strategic Participation: How Facilitated Engagement Can Be Shown to Reduce Resistance and Improve Outcomes Through the Use of Strategic Models
Authors: Tony Mann
Abstract:
This paper sets out the case for involving and engaging employees/workers/stakeholders/staff in any significant change that is being considered by the senior executives of the organization. It establishes the rationale, the approach, the methodology of engagement and the benefits of a participative approach. It challenges the new norm of imposing change for fear of resistance and instead suggests that involving people has better outcomes and a longer-lasting impact. Various strategic models are introduced and illustrated to explain how the process can be most effective. The paper highlights one model in particular (the Process Iceberg® Organizational Change model) that has proven to be instrumental in developing effective change. Its use is demonstrated in its various forms and explains why so much change fails to address the key elements and how we can be more productive in managing change. ‘Participation’ in change is too often seen as negative, expensive and unwieldy. The paper aims to show that another model: UIA=O+E, can offset the difficulties and, in fact, produce much more positive and effective change.Keywords: facilitation, stakeholders, buy-in, digital workshops
Procedia PDF Downloads 10923371 Investigation of a Technology Enabled Model of Home Care: the eShift Model of Palliative Care
Authors: L. Donelle, S. Regan, R. Booth, M. Kerr, J. McMurray, D. Fitzsimmons
Abstract:
Palliative home health care provision within the Canadian context is challenged by: (i) a shortage of registered nurses (RN) and RNs with palliative care expertise, (ii) an aging population, (iii) reliance on unpaid family caregivers to sustain home care services with limited support to conduct this ‘care work’, (iv) a model of healthcare that assumes client self-care, and (v) competing economic priorities. In response, an interprofessional team of service provider organizations, a software/technology provider, and health care providers developed and implemented a technology-enabled model of home care, the eShift model of palliative home care (eShift). The eShift model combines communication and documentation technology with non-traditional utilization of health human resources to meet patient needs for palliative care in the home. The purpose of this study was to investigate the structure, processes, and outcomes of the eShift model of care. Methodology: Guided by Donebedian’s evaluation framework for health care, this qualitative-descriptive study investigated the structure, processes, and outcomes care of the eShift model of palliative home care. Interviews and focus groups were conducted with health care providers (n= 45), decision-makers (n=13), technology providers (n=3) and family care givers (n=8). Interviews were recorded, transcribed, and a deductive analysis of transcripts was conducted. Study Findings (1) Structure: The eShift model consists of a remotely-situated RN using technology to direct care provision virtually to patients in their home. The remote RN is connected virtually to a health technician (an unregulated care provider) in the patient’s home using real-time communication. The health technician uses a smartphone modified with the eShift application and communicates with the RN who uses a computer with the eShift application/dashboard. Documentation and communication about patient observations and care activities occur in the eShift portal. The RN is typically accountable for four to six health technicians and patients over an 8-hour shift. The technology provider was identified as an important member of the healthcare team. Other members of the team include family members, care coordinators, nurse practitioners, physicians, and allied health. (2) Processes: Conventionally, patient needs are the focus of care; however within eShift, the patient and the family caregiver were the focus of care. Enhanced medication administration was seen as one of the most important processes, and family caregivers reported high satisfaction with the care provided. There was perceived enhanced teamwork among health care providers. (3) Outcomes: Patients were able to die at home. The eShift model enabled consistency and continuity of care, and effective management of patient symptoms and caregiver respite. Conclusion: More than a technology solution, the eShift model of care was viewed as transforming home care practice and an innovative way to resolve the shortage of palliative care nurses within home care.Keywords: palliative home care, health information technology, patient-centred care, interprofessional health care team
Procedia PDF Downloads 41823370 Multi-Objective Simulated Annealing Algorithms for Scheduling Just-In-Time Assembly Lines
Authors: Ghorbanali Mohammadi
Abstract:
New approaches to sequencing mixed-model manufacturing systems are present. These approaches have attracted considerable attention due to their potential to deal with difficult optimization problems. This paper presents Multi-Objective Simulated Annealing Algorithms (MOSAA) approaches to the Just-In-Time (JIT) sequencing problem where workload-smoothing (WL) and the number of set-ups (St) are to be optimized simultaneously. Mixed-model assembly lines are types of production lines where varieties of product models similar in product characteristics are assembled. Moreover, this type of problem is NP-hard. Two annealing methods are proposed to solve the multi-objective problem and find an efficient frontier of all design configurations. The performances of the two methods are tested on several problems from the literature. Experimentation demonstrates the relative desirable performance of the presented methodology.Keywords: scheduling, just-in-time, mixed-model assembly line, sequencing, simulated annealing
Procedia PDF Downloads 12823369 A Sharp Interface Model for Simulating Seawater Intrusion in the Coastal Aquifer of Wadi Nador (Algeria)
Authors: Abdelkader Hachemi, Boualem Remini
Abstract:
Seawater intrusion is a significant challenge faced by coastal aquifers in the Mediterranean basin. This study aims to determine the position of the sharp interface between seawater and freshwater in the aquifer of Wadi Nador, located in the Wilaya of Tipaza, Algeria. A numerical areal sharp interface model using the finite element method is developed to investigate the spatial and temporal behavior of seawater intrusion. The aquifer is assumed to be homogeneous and isotropic. The simulation results are compared with geophysical prospection data obtained through electrical methods in 2011 to validate the model. The simulation results demonstrate a good agreement with the geophysical prospection data, confirming the accuracy of the sharp interface model. The position of the sharp interface in the aquifer is found to be approximately 1617 meters from the sea. Two scenarios are proposed to predict the interface position for the year 2024: one without pumping and the other with pumping. The results indicate a noticeable retreat of the sharp interface position in the first scenario, while a slight decline is observed in the second scenario. The findings of this study provide valuable insights into the dynamics of seawater intrusion in the Wadi Nador aquifer. The predicted changes in the sharp interface position highlight the potential impact of pumping activities on the aquifer's vulnerability to seawater intrusion. This study emphasizes the importance of implementing measures to manage and mitigate seawater intrusion in coastal aquifers. The sharp interface model developed in this research can serve as a valuable tool for assessing and monitoring the vulnerability of aquifers to seawater intrusion.Keywords: seawater intrusion, sharp interface, coastal aquifer, algeria
Procedia PDF Downloads 12023368 Secure Image Encryption via Enhanced Fractional Order Chaotic Map
Authors: Ismail Haddad, Djamel Herbadji, Aissa Belmeguenai, Selma Boumerdassi
Abstract:
in this paper, we provide a novel approach for image encryption that employs the Fibonacci matrix and an enhanced fractional order chaotic map. The enhanced map overcomes the drawbacks of the classical map, especially the limited chaotic range and non-uniform distribution of chaotic sequences, resulting in a larger encryption key space. As a result, this strategy improves the encryption system's security. Our experimental results demonstrate that our proposed algorithm effectively encrypts grayscale images with exceptional efficiency. Furthermore, our technique is resistant to a wide range of potential attacks, including statistical and entropy attacks.Keywords: image encryption, logistic map, fibonacci matrix, grayscale images
Procedia PDF Downloads 31823367 Molecular Communication Noise Effect Analysis of Diffusion-Based Channel for Considering Minimum-Shift Keying and Molecular Shift Keying Modulations
Authors: A. Azari, S. S. K. Seyyedi
Abstract:
One of the unaddressed and open challenges in the nano-networking is the characteristics of noise. The previous analysis, however, has concentrated on end-to-end communication model with no separate modelings for propagation channel and noise. By considering a separate signal propagation and noise model, the design and implementation of an optimum receiver will be much easier. In this paper, we justify consideration of a separate additive Gaussian noise model of a nano-communication system based on the molecular communication channel for which are applicable for MSK and MOSK modulation schemes. The presented noise analysis is based on the Brownian motion process, and advection molecular statistics, where the received random signal has a probability density function whose mean is equal to the mean number of the received molecules. Finally, the justification of received signal magnitude being uncorrelated with additive non-stationary white noise is provided.Keywords: molecular, noise, diffusion, channel
Procedia PDF Downloads 27923366 A Mutually Exclusive Task Generation Method Based on Data Augmentation
Authors: Haojie Wang, Xun Li, Rui Yin
Abstract:
In order to solve the memorization overfitting in the model-agnostic meta-learning MAML algorithm, a method of generating mutually exclusive tasks based on data augmentation is proposed. This method generates a mutex task by corresponding one feature of the data to multiple labels so that the generated mutex task is inconsistent with the data distribution in the initial dataset. Because generating mutex tasks for all data will produce a large number of invalid data and, in the worst case, lead to an exponential growth of computation, this paper also proposes a key data extraction method that only extract part of the data to generate the mutex task. The experiments show that the method of generating mutually exclusive tasks can effectively solve the memorization overfitting in the meta-learning MAML algorithm.Keywords: mutex task generation, data augmentation, meta-learning, text classification.
Procedia PDF Downloads 14323365 Decoding Socio-Cultural Trends in Indian Urban Youth Using Ogilvy 3E Model
Authors: Falguni Vasavada, Pradyumna Malladi
Abstract:
The research focuses on studying the ecosystem of the youth using Ogilvy's 3E model, Ethnography and Thematic Analysis. It has been found that urban Indian youth today is an honest generation, hungry for success, living life by the moment, fiercely independent, are open about sex, sexuality and embrace individual differences. Technology and social media dominate their life. However, they are also phobic about commitments, often drifting along life and engage in unsubstantiated brave-talk.Keywords: ethnography, youth, culture, track, buyer behavior
Procedia PDF Downloads 36023364 Estimation of Optimum Parameters of Non-Linear Muskingum Model of Routing Using Imperialist Competition Algorithm (ICA)
Authors: Davood Rajabi, Mojgan Yazdani
Abstract:
Non-linear Muskingum model is an efficient method for flood routing, however, the efficiency of this method is influenced by three applied parameters. Therefore, efficiency assessment of Imperialist Competition Algorithm (ICA) to evaluate optimum parameters of non-linear Muskingum model was addressed through this study. In addition to ICA, Genetic Algorithm (GA) and Particle Swarm Optimization (PSO) were also used aiming at an available criterion to verdict ICA. In this regard, ICA was applied for Wilson flood routing; then, routing of two flood events of DoAab Samsami River was investigated. In case of Wilson flood that the target function was considered as the sum of squared deviation (SSQ) of observed and calculated discharges. Routing two other floods, in addition to SSQ, another target function was also considered as the sum of absolute deviations of observed and calculated discharge. For the first floodwater based on SSQ, GA indicated the best performance, however, ICA was on first place, based on SAD. For the second floodwater, based on both target functions, ICA indicated a better operation. According to the obtained results, it can be said that ICA could be used as an appropriate method to evaluate the parameters of Muskingum non-linear model.Keywords: Doab Samsami river, genetic algorithm, imperialist competition algorithm, meta-exploratory algorithms, particle swarm optimization, Wilson flood
Procedia PDF Downloads 50423363 Time Series Analysis the Case of China and USA Trade Examining during Covid-19 Trade Enormity of Abnormal Pricing with the Exchange rate
Authors: Md. Mahadi Hasan Sany, Mumenunnessa Keya, Sharun Khushbu, Sheikh Abujar
Abstract:
Since the beginning of China's economic reform, trade between the U.S. and China has grown rapidly, and has increased since China's accession to the World Trade Organization in 2001. The US imports more than it exports from China, reducing the trade war between China and the U.S. for the 2019 trade deficit, but in 2020, the opposite happens. In international and U.S. trade, Washington launched a full-scale trade war against China in March 2016, which occurred a catastrophic epidemic. The main goal of our study is to measure and predict trade relations between China and the U.S., before and after the arrival of the COVID epidemic. The ML model uses different data as input but has no time dimension that is present in the time series models and is only able to predict the future from previously observed data. The LSTM (a well-known Recurrent Neural Network) model is applied as the best time series model for trading forecasting. We have been able to create a sustainable forecasting system in trade between China and the US by closely monitoring a dataset published by the State Website NZ Tatauranga Aotearoa from January 1, 2015, to April 30, 2021. Throughout the survey, we provided a 180-day forecast that outlined what would happen to trade between China and the US during COVID-19. In addition, we have illustrated that the LSTM model provides outstanding outcome in time series data analysis rather than RFR and SVR (e.g., both ML models). The study looks at how the current Covid outbreak affects China-US trade. As a comparative study, RMSE transmission rate is calculated for LSTM, RFR and SVR. From our time series analysis, it can be said that the LSTM model has given very favorable thoughts in terms of China-US trade on the future export situation.Keywords: RFR, China-U.S. trade war, SVR, LSTM, deep learning, Covid-19, export value, forecasting, time series analysis
Procedia PDF Downloads 19823362 Modeling of Electrokinetic Mixing in Lab on Chip Microfluidic Devices
Authors: Virendra J. Majarikar, Harikrishnan N. Unni
Abstract:
This paper sets to demonstrate a modeling of electrokinetic mixing employing electroosmotic stationary and time-dependent microchannel using alternate zeta patches on the lower surface of the micromixer in a lab on chip microfluidic device. Electroosmotic flow is amplified using different 2D and 3D model designs with alternate and geometric zeta potential values such as 25, 50, and 100 mV, respectively, to achieve high concentration mixing in the electrokinetically-driven microfluidic system. The enhancement of electrokinetic mixing is studied using Finite Element Modeling, and simulation workflow is accomplished with defined integral steps. It can be observed that the presence of alternate zeta patches can help inducing microvortex flows inside the channel, which in turn can improve mixing efficiency. Fluid flow and concentration fields are simulated by solving Navier-Stokes equation (implying Helmholtz-Smoluchowski slip velocity boundary condition) and Convection-Diffusion equation. The effect of the magnitude of zeta potential, the number of alternate zeta patches, etc. are analysed thoroughly. 2D simulation reveals that there is a cumulative increase in concentration mixing, whereas 3D simulation differs slightly with low zeta potential as that of the 2D model within the T-shaped micromixer for concentration 1 mol/m3 and 0 mol/m3, respectively. Moreover, 2D model results were compared with those of 3D to indicate the importance of the 3D model in a microfluidic design process.Keywords: COMSOL Multiphysics®, electrokinetic, electroosmotic, microfluidics, zeta potential
Procedia PDF Downloads 24223361 The Asymptotic Hole Shape in Long Pulse Laser Drilling: The Influence of Multiple Reflections
Authors: Torsten Hermanns, You Wang, Stefan Janssen, Markus Niessen, Christoph Schoeler, Ulrich Thombansen, Wolfgang Schulz
Abstract:
In long pulse laser drilling of metals, it can be demonstrated that the ablation shape approaches a so-called asymptotic shape such that it changes only slightly or not at all with further irradiation. These findings are already known from ultra short pulse (USP) ablation of dielectric and semiconducting materials. The explanation for the occurrence of an asymptotic shape in long pulse drilling of metals is identified, a model for the description of the asymptotic hole shape numerically implemented, tested and clearly confirmed by comparison with experimental data. The model assumes a robust process in that way that the characteristics of the melt flow inside the arising melt film does not change qualitatively by changing the laser or processing parameters. Only robust processes are technically controllable and thus of industrial interest. The condition for a robust process is identified by a threshold for the mass flow density of the assist gas at the hole entrance which has to be exceeded. Within a robust process regime the melt flow characteristics can be captured by only one model parameter, namely the intensity threshold. In analogy to USP ablation (where it is already known for a long time that the resulting hole shape results from a threshold for the absorbed laser fluency) it is demonstrated that in the case of robust long pulse ablation the asymptotic shape forms in that way that along the whole contour the absorbed heat flux density is equal to the intensity threshold. The intensity threshold depends on the special material and radiation properties and has to be calibrated be one reference experiment. The model is implemented in a numerical simulation which is called AsymptoticDrill and requires such a few amount of resources that it can run on common desktop PCs, laptops or even smart devices. Resulting hole shapes can be calculated within seconds what depicts a clear advantage over other simulations presented in literature in the context of industrial every day usage. Against this background the software additionally is equipped with a user-friendly GUI which allows an intuitive usage. Individual parameters can be adjusted using sliders while the simulation result appears immediately in an adjacent window. A platform independent development allow a flexible usage: the operator can use the tool to adjust the process in a very convenient manner on a tablet during the developer can execute the tool in his office in order to design new processes. Furthermore, at the best knowledge of the authors AsymptoticDrill is the first simulation which allows the import of measured real beam distributions and thus calculates the asymptotic hole shape on the basis of the real state of the specific manufacturing system. In this paper the emphasis is placed on the investigation of the effect of multiple reflections on the asymptotic hole shape which gain in importance when drilling holes with large aspect ratios.Keywords: asymptotic hole shape, intensity threshold, long pulse laser drilling, robust process
Procedia PDF Downloads 21323360 Competency Based Talent Acquisition: Concept, Practice, and Model, with Reference to Indian Industries
Authors: Manasi V. Shah
Abstract:
Organizations, in the competitive era, are participating in the competency act. They have discerned that, strategically researched and defined competencies when put up on the shelf, can help in achieving business goals. The research focuses on critical elements of competency-based talent acquisition process from practical vantage, with significant experience in a variety of business settings. The research is exploratory and descriptive in nature. The research conduct and outcome is the hinge on with reference to Indian Industries. It elaborates about the concept, practice and a brief model that human resource practitioner can use for effective talent acquisition process, which in turn would be in alignment with business performance. The research helps to present a prudent understanding of recruiting and selecting apt human capital, that can fit in a given job role and has action oriented competency based assessment approach for measuring the probable success of a job incumbent in a given job role.Keywords: competency based talent acquisition, competency model, talent acquisition concept, talent acquisition practice
Procedia PDF Downloads 31223359 Using Social Network Analysis for Cyber Threat Intelligence
Authors: Vasileios Anastopoulos
Abstract:
Cyber threat intelligence assists organizations in understanding the threats they face and helps them make educated decisions on preparing their defenses. Sharing of threat intelligence and threat information is increasingly leveraged by organizations and enterprises, and various software solutions are already available, with the open-source malware information sharing platform (MISP) being a popular one. In this work, a methodology for the production of cyber threat intelligence using the threat information stored in MISP is proposed. The methodology leverages the discipline of social network analysis and the diamond model, a model used for intrusion analysis, to produce cyber threat intelligence. The workings are demonstrated with a case study on a production MISP instance of a real organization. The paper concluded with a discussion on the proposed methodology and possible directions for further research.Keywords: cyber threat intelligence, diamond model, malware information sharing platform, social network analysis
Procedia PDF Downloads 17823358 Attention-based Adaptive Convolution with Progressive Learning in Speech Enhancement
Authors: Tian Lan, Yixiang Wang, Wenxin Tai, Yilan Lyu, Zufeng Wu
Abstract:
The monaural speech enhancement task in the time-frequencydomain has a myriad of approaches, with the stacked con-volutional neural network (CNN) demonstrating superiorability in feature extraction and selection. However, usingstacked single convolutions method limits feature represen-tation capability and generalization ability. In order to solvethe aforementioned problem, we propose an attention-basedadaptive convolutional network that integrates the multi-scale convolutional operations into a operation-specific blockvia input dependent attention to adapt to complex auditoryscenes. In addition, we introduce a two-stage progressivelearning method to enlarge the receptive field without a dra-matic increase in computation burden. We conduct a series ofexperiments based on the TIMIT corpus, and the experimen-tal results prove that our proposed model is better than thestate-of-art models on all metrics.Keywords: speech enhancement, adaptive convolu-tion, progressive learning, time-frequency domain
Procedia PDF Downloads 12223357 Optimization of Operational Parameters and Design of an Electrochlorination System to Produce Naclo
Authors: Pablo Ignacio Hernández Arango, Niels Lindemeyer
Abstract:
Chlorine, as Sodium Hypochlorite (NaClO) solution in water, is an effective, worldwide spread, and economical substance to eliminate germs in the water. The disinfection potential of chlorine lies in its ability to degrade the outer surfaces of bacterial cells and viruses. This contribution reports the main parameters of the brine electrolysis for the production of NaClO, which is afterward used for the disinfection of water either for drinking or recreative uses. Herein, the system design was simulated, optimized, build, and tested based on titanium electrodes. The process optimization considers the whole process, from the salt (NaCl) dilution tank in order to maximize its operation time util the electrolysis itself in order to maximize the chlorine production reducing the energy and raw material (salt and water) consumption. One novel idea behind this optimization process is the modification of the flow pattern inside the electrochemical reactors. The increasing turbulence and residence time impact positively the operations figures. The operational parameters, which are defined in this study were compared and benchmarked with the parameters of actual commercial systems in order to validate the pertinency of those results.Keywords: electrolysis, water disinfection, sodium hypochlorite, process optimization
Procedia PDF Downloads 12823356 Digital Employment of Disabled People: Empirical Study from Shanghai
Abstract:
Across the globe, ICTs are influencing employment both as an industry that creates jobs and as a tool that empowers disabled people to access new forms of work, in innovative and more flexible ways. The advancements in ICT and the number of apps and solutions that support persons with physical, cognitive and intellectual disabilities challenge traditional biased notions and offer a pathway out of traditional sheltered workshops. As the global leader in digital technology innovation, China is arguably a leader in the use of digital technology as a 'lever' in ending the economic and social marginalization of the disabled. This study investigates factors that influence adoption and use of employment-oriented ICT applications among disabled people in China and seeks to integrate three theoretical approaches: the technology acceptance model (TAM), the uses and gratifications (U&G) approach, and the social model of disability. To that end, the study used data from self-reported survey of 214 disabled adults who have been involved in two top-down 'Internet + employment' programs promoted by local disabled persons’ federation in Shanghai. A structural equation model employed in the study demonstrates that the use of employment-oriented ICT applications is affected by demographic factors of gender, categories of disability, education and marital status. The organizational support of local social organizations demonstrates significate effects on the motivations of disabled people. Results from the focus group interviews particularly suggested that to maximize the positive impact of ICTs on employment, there is significant need to build stakeholder capacity on how ICTs could benefits persons with disabilities.Keywords: disabled people, ICTs, technology acceptance model, uses and gratifications, the social model of disability
Procedia PDF Downloads 108