Search results for: computational intelligence
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3431

Search results for: computational intelligence

911 Reed: An Approach Towards Quickly Bootstrapping Multilingual Acoustic Models

Authors: Bipasha Sen, Aditya Agarwal

Abstract:

Multilingual automatic speech recognition (ASR) system is a single entity capable of transcribing multiple languages sharing a common phone space. Performance of such a system is highly dependent on the compatibility of the languages. State of the art speech recognition systems are built using sequential architectures based on recurrent neural networks (RNN) limiting the computational parallelization in training. This poses a significant challenge in terms of time taken to bootstrap and validate the compatibility of multiple languages for building a robust multilingual system. Complex architectural choices based on self-attention networks are made to improve the parallelization thereby reducing the training time. In this work, we propose Reed, a simple system based on 1D convolutions which uses very short context to improve the training time. To improve the performance of our system, we use raw time-domain speech signals directly as input. This enables the convolutional layers to learn feature representations rather than relying on handcrafted features such as MFCC. We report improvement on training and inference times by atleast a factor of 4x and 7.4x respectively with comparable WERs against standard RNN based baseline systems on SpeechOcean's multilingual low resource dataset.

Keywords: convolutional neural networks, language compatibility, low resource languages, multilingual automatic speech recognition

Procedia PDF Downloads 116
910 Synthesis, Characterization and Biological Activites of Azomethine Derivatives

Authors: Lynda Golea, Rachid Chebaki

Abstract:

Schiff bases contain heterocyclic structural units with N and O donor atoms which plays an important role in coordination chemistry. Azomethine groups are a broad class of widely used compounds with applications in many fields, including analytical, inorganic chemistry and biological. Schiff's base is of promising research interest due to the widespread antibacterial resistance in medical science. In addition, the research is essential to generate Schiff base metal complexes with various applications. Schiff complexes have been used as drugs and have antibacterial, antifungal, antiviral, and anti-inflammatory properties. The various donor atoms they contain offer a special ability for metal binding. In this research on the physicochemical properties of azomethine groups, we synthesized and studied the Schiff base compounds by a condensation reaction of tryptamines and acetophenone in ethanol. The structure of the prepared compound was interpreted using 1H NMR, 13C NMR, UV-vis and FT-IR. A computational analysis at the level of DFT with functional B3LYP in conjunction with the base 6-311+G (d, p) was conducted to study its electronic and molecular structure. The biological study was performed on three bacterial strains usually causing infection, including Gram-positive and Gram-negative, for antibacterial activity. Results showed moderate biological activity and proportional activity with increasing concentration.

Keywords: azomethine, HOMO, LUMO, RMN, molecular docking

Procedia PDF Downloads 58
909 Turbulent Channel Flow Synthesis using Generative Adversarial Networks

Authors: John M. Lyne, K. Andrea Scott

Abstract:

In fluid dynamics, direct numerical simulations (DNS) of turbulent flows require large amounts of nodes to appropriately resolve all scales of energy transfer. Due to the size of these databases, sharing these datasets amongst the academic community is a challenge. Recent work has been done to investigate the use of super-resolution to enable database sharing, where a low-resolution flow field is super-resolved to high resolutions using a neural network. Recently, Generative Adversarial Networks (GAN) have grown in popularity with impressive results in the generation of faces, landscapes, and more. This work investigates the generation of unique high-resolution channel flow velocity fields from a low-dimensional latent space using a GAN. The training objective of the GAN is to generate samples in which the distribution of the generated samplesis ideally indistinguishable from the distribution of the training data. In this study, the network is trained using samples drawn from a statistically stationary channel flow at a Reynolds number of 560. Results show that the turbulent statistics and energy spectra of the generated flow fields are within reasonable agreement with those of the DNS data, demonstrating that GANscan produce the intricate multi-scale phenomena of turbulence.

Keywords: computational fluid dynamics, channel flow, turbulence, generative adversarial network

Procedia PDF Downloads 201
908 De-Novo Structural Elucidation from Mass/NMR Spectra

Authors: Ismael Zamora, Elisabeth Ortega, Tatiana Radchenko, Guillem Plasencia

Abstract:

The structure elucidation based on Mass Spectra (MS) data of unknown substances is an unresolved problem that affects many different fields of application. The recent overview of software available for structure elucidation of small molecules has shown the demand for efficient computational tool that will be able to perform structure elucidation of unknown small molecules and peptides. We developed an algorithm for De-Novo fragment analysis based on MS data that proposes a set of scored and ranked structures that are compatible with the MS and MSMS spectra. Several different algorithms were developed depending on the initial set of fragments and the structure building processes. Also, in all cases, several scores for the final molecule ranking were computed. They were validated with small and middle databases (DB) with the eleven test set compounds. Similar results were obtained from any of the databases that contained the fragments of the expected compound. We presented an algorithm. Or De-Novo fragment analysis based on only mass spectrometry (MS) data only that proposed a set of scored/ranked structures that was validated on different types of databases and showed good results as proof of concept. Moreover, the solutions proposed by Mass Spectrometry were submitted to the prediction of NMR spectra in order to elucidate which of the proposed structures was compatible with the NMR spectra collected.

Keywords: De Novo, structure elucidation, mass spectrometry, NMR

Procedia PDF Downloads 287
907 Brain-Computer Interfaces That Use Electroencephalography

Authors: Arda Ozkurt, Ozlem Bozkurt

Abstract:

Brain-computer interfaces (BCIs) are devices that output commands by interpreting the data collected from the brain. Electroencephalography (EEG) is a non-invasive method to measure the brain's electrical activity. Since it was invented by Hans Berger in 1929, it has led to many neurological discoveries and has become one of the essential components of non-invasive measuring methods. Despite the fact that it has a low spatial resolution -meaning it is able to detect when a group of neurons fires at the same time-, it is a non-invasive method, making it easy to use without possessing any risks. In EEG, electrodes are placed on the scalp, and the voltage difference between a minimum of two electrodes is recorded, which is then used to accomplish the intended task. The recordings of EEGs include, but are not limited to, the currents along dendrites from synapses to the soma, the action potentials along the axons connecting neurons, and the currents through the synaptic clefts connecting axons with dendrites. However, there are some sources of noise that may affect the reliability of the EEG signals as it is a non-invasive method. For instance, the noise from the EEG equipment, the leads, and the signals coming from the subject -such as the activity of the heart or muscle movements- affect the signals detected by the electrodes of the EEG. However, new techniques have been developed to differentiate between those signals and the intended ones. Furthermore, an EEG device is not enough to analyze the data from the brain to be used by the BCI implication. Because the EEG signal is very complex, to analyze it, artificial intelligence algorithms are required. These algorithms convert complex data into meaningful and useful information for neuroscientists to use the data to design BCI devices. Even though for neurological diseases which require highly precise data, invasive BCIs are needed; non-invasive BCIs - such as EEGs - are used in many cases to help disabled people's lives or even to ease people's lives by helping them with basic tasks. For example, EEG is used to detect before a seizure occurs in epilepsy patients, which can then prevent the seizure with the help of a BCI device. Overall, EEG is a commonly used non-invasive BCI technique that has helped develop BCIs and will continue to be used to detect data to ease people's lives as more BCI techniques will be developed in the future.

Keywords: BCI, EEG, non-invasive, spatial resolution

Procedia PDF Downloads 69
906 AI for Efficient Geothermal Exploration and Utilization

Authors: Velimir "monty" Vesselinov, Trais Kliplhuis, Hope Jasperson

Abstract:

Artificial intelligence (AI) is a powerful tool in the geothermal energy sector, aiding in both exploration and utilization. Identifying promising geothermal sites can be challenging due to limited surface indicators and the need for expensive drilling to confirm subsurface resources. Geothermal reservoirs can be located deep underground and exhibit complex geological structures, making traditional exploration methods time-consuming and imprecise. AI algorithms can analyze vast datasets of geological, geophysical, and remote sensing data, including satellite imagery, seismic surveys, geochemistry, geology, etc. Machine learning algorithms can identify subtle patterns and relationships within this data, potentially revealing hidden geothermal potential in areas previously overlooked. To address these challenges, a SIML (Science-Informed Machine Learning) technology has been developed. SIML methods are different from traditional ML techniques. In both cases, the ML models are trained to predict the spatial distribution of an output (e.g., pressure, temperature, heat flux) based on a series of inputs (e.g., permeability, porosity, etc.). The traditional ML (a) relies on deep and wide neural networks (NNs) based on simple algebraic mappings to represent complex processes. In contrast, the SIML neurons incorporate complex mappings (including constitutive relationships and physics/chemistry models). This results in ML models that have a physical meaning and satisfy physics laws and constraints. The prototype of the developed software, called GeoTGO, is accessible through the cloud. Our software prototype demonstrates how different data sources can be made available for processing, executed demonstrative SIML analyses, and presents the results in a table and graphic form.

Keywords: science-informed machine learning, artificial inteligence, exploration, utilization, hidden geothermal

Procedia PDF Downloads 43
905 Human Factors Interventions for Risk and Reliability Management of Defence Systems

Authors: Chitra Rajagopal, Indra Deo Kumar, Ila Chauhan, Ruchi Joshi, Binoy Bhargavan

Abstract:

Reliability and safety are essential for the success of mission-critical and safety-critical defense systems. Humans are part of the entire life cycle of defense systems development and deployment. The majority of industrial accidents or disasters are attributed to human errors. Therefore, considerations of human performance and human reliability are critical in all complex systems, including defense systems. Defense systems are operating from the ground, naval and aerial platforms in diverse conditions impose unique physical and psychological challenges to the human operators. Some of the safety and mission-critical defense systems with human-machine interactions are fighter planes, submarines, warships, combat vehicles, aerial and naval platforms based missiles, etc. Human roles and responsibilities are also going through a transition due to the infusion of artificial intelligence and cyber technologies. Human operators, not accustomed to such challenges, are more likely to commit errors, which may lead to accidents or loss events. In such a scenario, it is imperative to understand the human factors in defense systems for better systems performance, safety, and cost-effectiveness. A case study using Task Analysis (TA) based methodology for assessment and reduction of human errors in the Air and Missile Defense System in the context of emerging technologies were presented. Action-oriented task analysis techniques such as Hierarchical Task Analysis (HTA) and Operator Action Event Tree (OAET) along with Critical Action and Decision Event Tree (CADET) for cognitive task analysis was used. Human factors assessment based on the task analysis helps in realizing safe and reliable defense systems. These techniques helped in the identification of human errors during different phases of Air and Missile Defence operations, leading to meet the requirement of a safe, reliable and cost-effective mission.

Keywords: defence systems, reliability, risk, safety

Procedia PDF Downloads 133
904 Analysis of Hard Turning Process of AISI D3-Thermal Aspects

Authors: B. Varaprasad, C. Srinivasa Rao

Abstract:

In the manufacturing sector, hard turning has emerged as vital machining process for cutting hardened steels. Besides many advantages of hard turning operation, one has to implement to achieve close tolerances in terms of surface finish, high product quality, reduced machining time, low operating cost and environmentally friendly characteristics. In the present study, three-dimensional CAE (Computer Aided Engineering) based simulation of  hard turning by using commercial software DEFORM 3D has been compared to experimental results of  stresses, temperatures and tool forces in machining of AISI D3 steel using mixed Ceramic inserts (CC6050). In the present analysis, orthogonal cutting models are proposed, considering several processing parameters such as cutting speed, feed, and depth of cut. An exhaustive friction modeling at the tool-work interfaces is carried out. Work material flow around the cutting edge is carefully modeled with adaptive re-meshing simulation capability. In process simulations, feed rate and cutting speed are constant (i.e.,. 0.075 mm/rev and 155 m/min), and analysis is focused on stresses, forces, and temperatures during machining. Close agreement is observed between CAE simulation and experimental values.

Keywords: hard turning, computer aided engineering, computational machining, finite element method

Procedia PDF Downloads 450
903 Multi-Criteria Test Case Selection Using Ant Colony Optimization

Authors: Niranjana Devi N.

Abstract:

Test case selection is to select the subset of only the fit test cases and remove the unfit, ambiguous, redundant, unnecessary test cases which in turn improve the quality and reduce the cost of software testing. Test cases optimization is the problem of finding the best subset of test cases from a pool of the test cases to be audited. It will meet all the objectives of testing concurrently. But most of the research have evaluated the fitness of test cases only on single parameter fault detecting capability and optimize the test cases using a single objective. In the proposed approach, nine parameters are considered for test case selection and the best subset of parameters for test case selection is obtained using Interval Type-2 Fuzzy Rough Set. Test case selection is done in two stages. The first stage is the fuzzy entropy-based filtration technique, used for estimating and reducing the ambiguity in test case fitness evaluation and selection. The second stage is the ant colony optimization-based wrapper technique with a forward search strategy, employed to select test cases from the reduced test suite of the first stage. The results are evaluated using the Coverage parameters, Precision, Recall, F-Measure, APSC, APDC, and SSR. The experimental evaluation demonstrates that by this approach considerable computational effort can be avoided.

Keywords: ant colony optimization, fuzzy entropy, interval type-2 fuzzy rough set, test case selection

Procedia PDF Downloads 663
902 Solving Transient Conduction and Radiation using Finite Volume Method

Authors: Ashok K. Satapathy, Prerana Nashine

Abstract:

Radiative heat transfer in participating medium was anticipated using the finite volume method. The radiative transfer equations are formulated for absorbing and anisotropically scattering and emitting medium. The solution strategy is discussed and the conditions for computational stability are conferred. The equations have been solved for transient radiative medium and transient radiation incorporated with transient conduction. Results have been obtained for irradiation and corresponding heat fluxes for both the cases. The solutions can be used to conclude incident energy and surface heat flux. Transient solutions were obtained for a slab of heat conducting in slab by thermal radiation. The effect of heat conduction during the transient phase is to partially equalize the internal temperature distribution. The solution procedure provides accurate temperature distributions in these regions. A finite volume procedure with variable space and time increments is used to solve the transient energy equation. The medium in the enclosure absorbs, emits, and anisotropically scatters radiative energy. The incident radiations and the radiative heat fluxes are presented in graphical forms. The phase function anisotropy plays a significant role in the radiation heat transfer when the boundary condition is non-symmetric.

Keywords: participating media, finite volume method, radiation coupled with conduction, heat transfer

Procedia PDF Downloads 375
901 Artificial Intelligence Protecting Birds against Collisions with Wind Turbines

Authors: Aleksandra Szurlej-Kielanska, Lucyna Pilacka, Dariusz Górecki

Abstract:

The dynamic development of wind energy requires the simultaneous implementation of effective systems minimizing the risk of collisions between birds and wind turbines. Wind turbines are installed in more and more challenging locations, often close to the natural environment of birds. More and more countries and organizations are defining guidelines for the necessary functionality of such systems. The minimum bird detection distance, trajectory tracking, and shutdown time are key factors in eliminating collisions. Since 2020, we have continued the survey on the validation of the subsequent version of the BPS detection and reaction system. Bird protection system (BPS) is a fully automatic camera system which allows one to estimate the distance of the bird to the turbine, classify its size and autonomously undertake various actions depending on the bird's distance and flight path. The BPS was installed and tested in a real environment at a wind turbine in northern Poland and Central Spain. The performed validation showed that at a distance of up to 300 m, the BPS performs at least as well as a skilled ornithologist, and large bird species are successfully detected from over 600 m. In addition, data collected by BPS systems installed in Spain showed that 60% of the detections of all birds of prey were from individuals approaching the turbine, and these detections meet the turbine shutdown criteria. Less than 40% of the detections of birds of prey took place at wind speeds below 2 m/s while the turbines were not working. As shown by the analysis of the data collected by the system over 12 months, the system classified the improved size of birds with a wingspan of more than 1.1 m in 90% and the size of birds with a wingspan of 0.7 - 1 m in 80% of cases. The collected data also allow the conclusion that some species keep a certain distance from the turbines at a wind speed of over 8 m/s (Aquila sp., Buteo sp., Gyps sp.), but Gyps sp. and Milvus sp. remained active at this wind speed on the tested area. The data collected so far indicate that BPS is effective in detecting and stopping wind turbines in response to the presence of birds of prey with a wingspan of more than 1 m.

Keywords: protecting birds, birds monitoring, wind farms, green energy, sustainable development

Procedia PDF Downloads 67
900 Chaos Fuzzy Genetic Algorithm

Authors: Mohammad Jalali Varnamkhasti

Abstract:

The genetic algorithms have been very successful in handling difficult optimization problems. The fundamental problem in genetic algorithms is premature convergence. This paper, present a new fuzzy genetic algorithm based on chaotic values instead of the random values in genetic algorithm processes. In this algorithm, for initial population is used chaotic sequences and then a new sexual selection proposed for selection mechanism. In this technique, the population is divided such that the male and female would be selected in an alternate way. The layout of the male and female chromosomes in each generation is different. A female chromosome is selected by tournament selection size from the female group. Then, the male chromosome is selected, in order of preference based on the maximum Hamming distance between the male chromosome and the female chromosome or The highest fitness value of male chromosome (if more than one male chromosome is having the maximum Hamming distance existed), or Random selection. The selections of crossover and mutation operators are achieved by running the fuzzy logic controllers, the crossover and mutation probabilities are varied on the basis of the phenotype and genotype characteristics of the chromosome population. Computational experiments are conducted on the proposed techniques and the results are compared with some other operators, heuristic and local search algorithms commonly used for solving p-median problems published in the literature.

Keywords: genetic algorithm, fuzzy system, chaos, sexual selection

Procedia PDF Downloads 380
899 Prediction of Temperature Distribution during Drilling Process Using Artificial Neural Network

Authors: Ali Reza Tahavvor, Saeed Hosseini, Nazli Jowkar, Afshin Karimzadeh Fard

Abstract:

Experimental & numeral study of temperature distribution during milling process, is important in milling quality and tools life aspects. In the present study the milling cross-section temperature is determined by using Artificial Neural Networks (ANN) according to the temperature of certain points of the work piece and the points specifications and the milling rotational speed of the blade. In the present work, at first three-dimensional model of the work piece is provided and then by using the Computational Heat Transfer (CHT) simulations, temperature in different nods of the work piece are specified in steady-state conditions. Results obtained from CHT are used for training and testing the ANN approach. Using reverse engineering and setting the desired x, y, z and the milling rotational speed of the blade as input data to the network, the milling surface temperature determined by neural network is presented as output data. The desired points temperature for different milling blade rotational speed are obtained experimentally and by extrapolation method for the milling surface temperature is obtained and a comparison is performed among the soft programming ANN, CHT results and experimental data and it is observed that ANN soft programming code can be used more efficiently to determine the temperature in a milling process.

Keywords: artificial neural networks, milling process, rotational speed, temperature

Procedia PDF Downloads 396
898 Risk Assessment of Radiation Hazard for a Typical WWER1000: Cancer Risk Analysis during a Hypothetical Accident

Authors: R. Gharari, N. Kojouri, R. Hosseini Aghdam, E. Alibeigi, B. Salmasian

Abstract:

In this research, the WWER1000/V446 (a PWR Russian type reactor) is chosen as the case study. It is assumed that radioactive materials that release into the environment are more than allowable limit due to a complete failure of the ventilation system (reactor stack). In the following, the HOTSPOT and the RASCAL computational codes have been used and coupled with a developed program using MATLAB software to evaluate Total effective dose equivalent (TEDE) and cancer risk according to the BEIR equations for various human organs. In addition, effects of the containment spray system and climate conditions on the TEDE have been investigated. According to the obtained results, there is an inverse correlation between the received dose and the wind speed; the amount of the TEDE for wind speed 2 m/s and is more than wind speed for 14 m/s during the class A of the climate (2.168 and 0.444 mSv, respectively). Also, containment spray system can effect and reduce the amount of the fission products and TEDE. Furthermore, the probability of the cancer risk for women is more than men, and for children is more than adults. In addition, a specific emergency zonal planning is proposed. Results are promising in which the site selection of the WWER1000/V446 were considered safe for the public in this situation.

Keywords: TEDE, total effective dose equivalent, RASCAL and HOTSPOT codes, BEIR equations, cancer risk

Procedia PDF Downloads 162
897 Long Waves Inundating through and around an Array of Circular Cylinders

Authors: Christian Klettner, Ian Eames, Tristan Robinson

Abstract:

Tsunami is characterised by their very long time periods and can have devastating consequences when these inundate through built-up coastal regions as in the 2004 Indian Ocean and 2011 Tohoku Tsunami. This work aims to investigate the effect of these long waves on the flow through and around a group of buildings, which are abstracted to circular cylinders. The research approach used in this study was using experiments and numerical simulations. Large-scale experiments were carried out at HR Wallingford. The novelty of these experiments is (I) the number of bodies present (up to 64), (II) the long wavelength of the input waves (80 seconds) and (III) the width of the tank (4m) which gives the unique opportunity to investigate three length scales, namely the diameter of the building, the diameter of the array and the width of the tank. To complement the experiments, dam break flow past the same arrays is investigated using three-dimensional numerical simulations in OpenFOAM. Dam break flow was chosen as it is often used as a surrogate for the tsunami in previous research and is used here as there are well defined initial conditions and high quality previous experimental data for the case of a single cylinder is available. The focus of this work is to better understand the effect of the solid void fraction on the force and flow through and around the array. New qualitative and quantitative diagnostics are developed and tested to analyse the complex coupled interaction between the cylinders.

Keywords: computational fluid dynamics, tsunami, forces, complex geometry

Procedia PDF Downloads 191
896 Towards a Reinvented Cash Management Function: Mobilising Innovative Advances for Enhanced Performance and Optimised Cost Management: Insights from Large Moroccan Companies in the Casablanca-Settat Region

Authors: Badrane Nohayla, Bamousse Zineb

Abstract:

Financial crises, exchange rate volatility, fluctuations in commodity prices, increased competitive pressures, and environmental issues are all threats that businesses face. In light of these diverse challenges, proactive, agile, and innovative cash management becomes an indispensable financial shield, allowing companies to thrive despite the adverse conditions of the global environment. In the same spirit, uncertainty, turbulence, volatility, and competitiveness continue to disrupt economic environments, compelling companies to swiftly master innovative breakthroughs that provide added value. In such a context, innovation emerges as a catalytic vector for performance, aiming to reduce costs, strengthen growth, and ultimately ensure the sustainability of Moroccan companies in the national arena. Moreover, innovation in treasury management promises to be one of the key pillars of financial stability, enabling companies to navigate the tumultuous waters of a globalized environment. Therefore, the objective of this study is to better understand the impact of innovative treasury management on cost optimization and, by extension, performance improvement. To elucidate this relationship, we conducted an exploratory qualitative study with 20 large Moroccan companies operating in the Casablanca-Settat region. The results highlight that innovation at the heart of treasury management is a guarantee of sustainability against the risks of failure and stands as a true pivot of the performance of Moroccan companies, an important parameter of their financial balance and a catalytic vector of their growth in the national economic landscape. In this regard, the present study aims to explore the extent to which innovation at the core of the treasury function serves as an indispensable tool for boosting performance while optimising costs in large Moroccan companies.

Keywords: innovative cash management, artificial intelligence, financial performance, risk management, cost savings

Procedia PDF Downloads 22
895 A Simple Computational Method for the Gravitational and Seismic Soil-Structure-Interaction between New and Existent Buildings Sites

Authors: Nicolae Daniel Stoica, Ion Mierlus Mazilu

Abstract:

This work is one of numerical research and aims to address the issue of the design of new buildings in a 3D location of existing buildings. In today's continuous development and congestion of urban centers is a big question about the influence of the new buildings on an already existent vicinity site. Thus, in this study, we tried to focus on how existent buildings may be affected by any newly constructed buildings and in how far this influence is really decreased. The problem of modeling the influence of interaction between buildings is not simple in any area in the world, and neither in Romania. Unfortunately, most often the designers not done calculations that can determine how close to reality these 3D influences nor the simplified method and the more superior methods. In the most literature making a "shield" (the pilots or molded walls) is absolutely sufficient to stop the influence between the buildings, and so often the soil under the structure is ignored in the calculation models. The main causes for which the soil is neglected in the analysis are related to the complexity modeling of interaction between soil and structure. In this paper, based on a new simple but efficient methodology we tried to determine for a lot of study cases the influence, in terms of assessing the interaction land structure on the behavior of structures that influence a new building on an existing one. The study covers additional subsidence that may occur during the execution of new works and after its completion. It also highlighted the efforts diagrams and deflections in the soil for both the original case and the final stage. This is necessary to see to what extent the expected impact of the new building on existing areas.

Keywords: soil, structure, interaction, piles, earthquakes

Procedia PDF Downloads 288
894 Computational Studies of the Reactivity Descriptors and the Optoelectronic Properties on the Efficiency Free-Base- and Zn-Porphyrin-Sensitized Solar Cells

Authors: Soraya Abtouche, Zeyneb Ghoualem, Syrine Daoudi, Lina Ouldmohamed, Xavier Assfeld

Abstract:

This work reports density functional theory calculations of the optimized geometries, molecular reactivity, energy gap,and thermodynamic properties of the free base (H2P) and their Zn (II) metallated (ZnP), bearing one, two, or three carboxylic acid groups using the hybrid functional B3LYP, Cam-B3lYP, wb97xd with 6-31G(d,p) basis sets. When donating groups are attached to the molecular dye, the bond lengths are slightly decreased, which is important for the easy transfer of an electron from donating to the accepting group. For all dyes, the highest occupied molecular orbital/lowest occupied molecular orbital analysis results in positive outcomes upon electron injection to the semiconductor and subsequent dye regeneration by the electrolyte. The ionization potential increases with increasing conjugation; therefore, the compound dye attached to one carboxylic acid group has the highest ionization potential. The results show higher efficiencies of those sensitized with ZnP. These results have been explained, taking into account the electronic character of the metal ion, which acts as a mediator in the injection step, and, on the other hand, considering the number of anchoring groups to which it binds to the surface of TiO2.

Keywords: DSSC, porphyrin, TD-DFT, electronic properties, donor-acceptor groups

Procedia PDF Downloads 72
893 Scientific Linux Cluster for BIG-DATA Analysis (SLBD): A Case of Fayoum University

Authors: Hassan S. Hussein, Rania A. Abul Seoud, Amr M. Refaat

Abstract:

Scientific researchers face in the analysis of very large data sets that is increasing noticeable rate in today’s and tomorrow’s technologies. Hadoop and Spark are types of software that developed frameworks. Hadoop framework is suitable for many Different hardware platforms. In this research, a scientific Linux cluster for Big Data analysis (SLBD) is presented. SLBD runs open source software with large computational capacity and high performance cluster infrastructure. SLBD composed of one cluster contains identical, commodity-grade computers interconnected via a small LAN. SLBD consists of a fast switch and Gigabit-Ethernet card which connect four (nodes). Cloudera Manager is used to configure and manage an Apache Hadoop stack. Hadoop is a framework allows storing and processing big data across the cluster by using MapReduce algorithm. MapReduce algorithm divides the task into smaller tasks which to be assigned to the network nodes. Algorithm then collects the results and form the final result dataset. SLBD clustering system allows fast and efficient processing of large amount of data resulting from different applications. SLBD also provides high performance, high throughput, high availability, expandability and cluster scalability.

Keywords: big data platforms, cloudera manager, Hadoop, MapReduce

Procedia PDF Downloads 354
892 Big Data Analytics and Public Policy: A Study in Rural India

Authors: Vasantha Gouri Prathapagiri

Abstract:

Innovations in ICT sector facilitate qualitative life style for citizens across the globe. Countries that facilitate usage of new techniques in ICT, i.e., big data analytics find it easier to fulfil the needs of their citizens. Big data is characterised by its volume, variety, and speed. Analytics involves its processing in a cost effective way in order to draw conclusion for their useful application. Big data also involves into the field of machine learning, artificial intelligence all leading to accuracy in data presentation useful for public policy making. Hence using data analytics in public policy making is a proper way to march towards all round development of any country. The data driven insights can help the government to take important strategic decisions with regard to socio-economic development of her country. Developed nations like UK and USA are already far ahead on the path of digitization with the support of Big Data analytics. India is a huge country and is currently on the path of massive digitization being realised through Digital India Mission. Internet connection per household is on the rise every year. This transforms into a massive data set that has the potential to improvise the public services delivery system into an effective service mechanism for Indian citizens. In fact, when compared to developed nations, this capacity is being underutilized in India. This is particularly true for administrative system in rural areas. The present paper focuses on the need for big data analytics adaptation in Indian rural administration and its contribution towards development of the country on a faster pace. Results of the research focussed on the need for increasing awareness and serious capacity building of the government personnel working for rural development with regard to big data analytics and its utility for development of the country. Multiple public policies are framed and implemented for rural development yet the results are not as effective as they should be. Big data has a major role to play in this context as can assist in improving both policy making and implementation aiming at all round development of the country.

Keywords: Digital India Mission, public service delivery system, public policy, Indian administration

Procedia PDF Downloads 155
891 Maximizing Bidirectional Green Waves for Major Road Axes

Authors: Christian Liebchen

Abstract:

Both from an environmental perspective and with respect to road traffic flow quality, planning so-called green waves along major road axes is a well-established target for traffic engineers. For one-way road axes (e.g. the Avenues in Manhattan), this is a trivial downstream task. For bidirectional arterials, the well-known necessary condition for establishing a green wave in both directions is that the driving times between two subsequent crossings must be an integer multiple of half of the cycle time of the signal programs at the nodes. In this paper, we propose an integer linear optimization model to establish fixed-time green waves in both directions that are as long and as wide as possible, even in the situation where the driving time condition is not fulfilled. In particular, we are considering an arterial along whose nodes separate left-turn signal groups are realized. In our computational results, we show that scheduling left-turn phases before or after the straight phases can reduce waiting times along the arterial. Moreover, we show that there is always a solution with green waves in both directions that are as long and as wide as possible, where absolute priority is put on just one direction. Compared to optimizing both directions together, establishing an ideal green wave into one direction can only provide suboptimal quality when considering prioritized parts of a green band (e.g., first few seconds).

Keywords: traffic light coordination, synchronization, phase sequencing, green waves, integer programming

Procedia PDF Downloads 110
890 On the Design of a Secure Two-Party Authentication Scheme for Internet of Things Using Cancelable Biometrics and Physically Unclonable Functions

Authors: Behnam Zahednejad, Saeed Kosari

Abstract:

Widespread deployment of Internet of Things (IoT) has raised security and privacy issues in this environment. Designing a secure two-factor authentication scheme between the user and server is still a challenging task. In this paper, we focus on Cancelable Biometric (CB) as an authentication factor in IoT. We show that previous CB-based scheme fail to provide real two-factor security, Perfect Forward Secrecy (PFS) and suffer database attacks and traceability of the user. Then we propose our improved scheme based on CB and Physically Unclonable Functions (PUF), which can provide real two-factor security, PFS, user’s unlinkability, and resistance to database attack. In addition, Key Compromise Impersonation (KCI) resilience is achieved in our scheme. We also prove the security of our proposed scheme formally using both Real-Or-Random (RoR) model and the ProVerif analysis tool. For the usability of our scheme, we conducted a performance analysis and showed that our scheme has the least communication cost compared to the previous CB-based scheme. The computational cost of our scheme is also acceptable for the IoT environment.

Keywords: IoT, two-factor security, cancelable biometric, key compromise impersonation resilience, perfect forward secrecy, database attack, real-or-random model, ProVerif

Procedia PDF Downloads 92
889 Smart Signature - Medical Communication without Barrier

Authors: Chia-Ying Lin

Abstract:

This paper explains how to enhance doctor-patient communication and nurse-patient communication through multiple intelligence signing methods and user-centered. It is hoped that through the implementation of the "electronic consent", the problems faced by the paper consent can be solved: storage methods, resource utilization, convenience, correctness of information, integrated management, statistical analysis and other related issues. Make better use and allocation of resources to provide better medical quality. First, invite the medical records department to assist in the inventory of paper consent in the hospital: organising, classifying, merging, coding, and setting. Second, plan the electronic consent configuration file: set the form number, consent form group, fields and templates, and the corresponding doctor's order code. Next, Summarize four types of rapid methods of electronic consent: according to the doctor's order, according to the medical behavior, according to the schedule, and manually generate the consent form. Finally, system promotion and adjustment: form an "electronic consent promotion team" to improve, follow five major processes: planning, development, testing, release, and feedback, and invite clinical units to raise the difficulties faced in the promotion, and make improvements to the problems. The electronic signature rate of the whole hospital will increase from 4% in January 2022 to 79% in November 2022. Use the saved resources more effectively, including: reduce paper usage (reduce carbon footprint), reduce the cost of ink cartridges, re-plan and use the space for paper medical records, and save human resources to provide better services. Through the introduction of information technology and technology, the main spirit of "lean management" is implemented. Transforming and reengineering the process to eliminate unnecessary waste is also the highest purpose of this project.

Keywords: smart signature, electronic consent, electronic medical records, user-centered, doctor-patient communication, nurse-patient communication

Procedia PDF Downloads 119
888 Achieving High Renewable Energy Penetration in Western Australia Using Data Digitisation and Machine Learning

Authors: A. D. Tayal

Abstract:

The energy industry is undergoing significant disruption. This research outlines that, whilst challenging; this disruption is also an emerging opportunity for electricity utilities. One such opportunity is leveraging the developments in data analytics and machine learning. As the uptake of renewable energy technologies and complimentary control systems increases, electricity grids will likely transform towards dense microgrids with high penetration of renewable generation sources, rich in network and customer data, and linked through intelligent, wireless communications. Data digitisation and analytics have already impacted numerous industries, and its influence on the energy sector is growing, as computational capabilities increase to manage big data, and as machines develop algorithms to solve the energy challenges of the future. The objective of this paper is to address how far the uptake of renewable technologies can go given the constraints of existing grid infrastructure and provides a qualitative assessment of how higher levels of renewable energy penetration can be facilitated by incorporating even broader technological advances in the fields of data analytics and machine learning. Western Australia is used as a contextualised case study, given its abundance and diverse renewable resources (solar, wind, biomass, and wave) and isolated networks, making a high penetration of renewables a feasible target for policy makers over coming decades.

Keywords: data, innovation, renewable, solar

Procedia PDF Downloads 359
887 A Lightweight Pretrained Encrypted Traffic Classification Method with Squeeze-and-Excitation Block and Sharpness-Aware Optimization

Authors: Zhiyan Meng, Dan Liu, Jintao Meng

Abstract:

Dependable encrypted traffic classification is crucial for improving cybersecurity and handling the growing amount of data. Large language models have shown that learning from large datasets can be effective, making pre-trained methods for encrypted traffic classification popular. However, attention-based pre-trained methods face two main issues: their large neural parameters are not suitable for low-computation environments like mobile devices and real-time applications, and they often overfit by getting stuck in local minima. To address these issues, we developed a lightweight transformer model, which reduces the computational parameters through lightweight vocabulary construction and Squeeze-and-Excitation Block. We use sharpness-aware optimization to avoid local minima during pre-training and capture temporal features with relative positional embeddings. Our approach keeps the model's classification accuracy high for downstream tasks. We conducted experiments on four datasets -USTC-TFC2016, VPN 2016, Tor 2016, and CICIOT 2022. Even with fewer than 18 million parameters, our method achieves classification results similar to methods with ten times as many parameters.

Keywords: sharpness-aware optimization, encrypted traffic classification, squeeze-and-excitation block, pretrained model

Procedia PDF Downloads 14
886 Development of a French to Yorùbá Machine Translation System

Authors: Benjamen Nathaniel, Eludiora Safiriyu Ijiyemi, Egume Oneme Lucky

Abstract:

A review on machine translation systems shows that a lot of computational artefacts has been carried out to translate written or spoken texts from a source language to Yorùbá language through Machine Translation systems. However, there are no work on French to Yorùbá language machine translation system; hence, the study investigated the process involved in the translation of French-to-Yorùbá language equivalent with the view to adopting a rule- based MT approach to build a Machine Translation framework from simple sentences administered through questionnaire. Articles and relevant textbooks were reviewed with key speakers of both languages interviewed to find out the processes involved in the translation of French language and their equivalent in Yorùbálanguage simple sentences using home domain terminologies. Achieving this, a model was formulated using phrase grammar structure, re-write rule, parse tree, automata theory- based techniques, designed and implemented respectively with unified modeling language (UML) and python programming language. Analysing the result, it was observed when carrying out the result that, the Machine Translation system performed 18.45% above Experimental Subject Respondent and 2.7% below Linguistics Expert when analysed with word orthography, sentence syntax and semantic correctness of the sentences. And, when compared with Google Machine Translation system, it was noticed that the developed system performed better on lexicons of the target language.

Keywords: machine translation (MT), rule-based, French language, Yoru`ba´ language

Procedia PDF Downloads 66
885 Insecurity and Insurgency on Economic Development of Nigeria

Authors: Uche Lucy Onyekwelu, Uche B. Ugwuanyi

Abstract:

Suffice to say that socio-economic disruptions of any form is likely to affect the wellbeing of the citizenry. The upsurge of social disequilibrium caused by the incessant disruptive tendencies exhibited by youths and some others in Nigeria are not helping matters. In Nigeria the social unrest has caused different forms of draw backs in Socio Economic Development. This study has empirically evaluated the impact of insecurity and insurgency on the Economic Development of Nigeria. The paper noted that the different forms of insecurity in Nigeria are namely: Insurgency and Banditry as witnessed in Northern Nigeria; Militancy: Niger Delta area and self-determination groups pursuing various forms of agenda such as Sit –at- Home Syndrome in the South Eastern Nigeria and other secessionist movements. All these have in one way or the other hampered Economic development in Nigeria. Data for this study were collected through primary and secondary sources using questionnaire and some existing documentations. Cost of investment in different aspects of security outfits in Nigeria represents the independent variable while the differentials in the Gross Domestic Product(GDP) and Human Development Index(HDI) are the measures of the dependent variable. Descriptive statistics and Simple Linear Regression analytical tool were employed in the data analysis. The result revealed that Insurgency/Insecurity negatively affect the economic development of the different parts of Nigeria. Following the findings, a model to analyse the effect of insecurity and insurgency was developed, named INSECUREDEVNIG. It implies that the economic development of Nigeria will continue to deteriorate if insurgency and insecurity continue. The study therefore recommends that the government should do all it could to nurture its human capital, adequately fund the state security apparatus and employ individuals of high integrity to manage the various security outfits in Nigeria. The government should also as a matter of urgency train the security personnel in intelligence cum Information and Communications Technology to enable them ensure the effectiveness of implementation of security policies needed to sustain Gross Domestic Product and Human Capital Index of Nigeria.

Keywords: insecurity, insurgency, gross domestic product, human development index, Nigeria

Procedia PDF Downloads 95
884 A Mechanical Diagnosis Method Based on Vibration Fault Signal down-Sampling and the Improved One-Dimensional Convolutional Neural Network

Authors: Bowei Yuan, Shi Li, Liuyang Song, Huaqing Wang, Lingli Cui

Abstract:

Convolutional neural networks (CNN) have received extensive attention in the field of fault diagnosis. Many fault diagnosis methods use CNN for fault type identification. However, when the amount of raw data collected by sensors is massive, the neural network needs to perform a time-consuming classification task. In this paper, a mechanical fault diagnosis method based on vibration signal down-sampling and the improved one-dimensional convolutional neural network is proposed. Through the robust principal component analysis, the low-rank feature matrix of a large amount of raw data can be separated, and then down-sampling is realized to reduce the subsequent calculation amount. In the improved one-dimensional CNN, a smaller convolution kernel is used to reduce the number of parameters and computational complexity, and regularization is introduced before the fully connected layer to prevent overfitting. In addition, the multi-connected layers can better generalize classification results without cumbersome parameter adjustments. The effectiveness of the method is verified by monitoring the signal of the centrifugal pump test bench, and the average test accuracy is above 98%. When compared with the traditional deep belief network (DBN) and support vector machine (SVM) methods, this method has better performance.

Keywords: fault diagnosis, vibration signal down-sampling, 1D-CNN

Procedia PDF Downloads 126
883 Healthcare-SignNet: Advanced Video Classification for Medical Sign Language Recognition Using CNN and RNN Models

Authors: Chithra A. V., Somoshree Datta, Sandeep Nithyanandan

Abstract:

Sign Language Recognition (SLR) is the process of interpreting and translating sign language into spoken or written language using technological systems. It involves recognizing hand gestures, facial expressions, and body movements that makeup sign language communication. The primary goal of SLR is to facilitate communication between hearing- and speech-impaired communities and those who do not understand sign language. Due to the increased awareness and greater recognition of the rights and needs of the hearing- and speech-impaired community, sign language recognition has gained significant importance over the past 10 years. Technological advancements in the fields of Artificial Intelligence and Machine Learning have made it more practical and feasible to create accurate SLR systems. This paper presents a distinct approach to SLR by framing it as a video classification problem using Deep Learning (DL), whereby a combination of Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) has been used. This research targets the integration of sign language recognition into healthcare settings, aiming to improve communication between medical professionals and patients with hearing impairments. The spatial features from each video frame are extracted using a CNN, which captures essential elements such as hand shapes, movements, and facial expressions. These features are then fed into an RNN network that learns the temporal dependencies and patterns inherent in sign language sequences. The INCLUDE dataset has been enhanced with more videos from the healthcare domain and the model is evaluated on the same. Our model achieves 91% accuracy, representing state-of-the-art performance in this domain. The results highlight the effectiveness of treating SLR as a video classification task with the CNN-RNN architecture. This approach not only improves recognition accuracy but also offers a scalable solution for real-time SLR applications, significantly advancing the field of accessible communication technologies.

Keywords: sign language recognition, deep learning, convolution neural network, recurrent neural network

Procedia PDF Downloads 18
882 The Effect of Sorafenibe on Soat1 Protein by Using Molecular Docking Method

Authors: Mahdiyeh Gholaminezhad

Abstract:

Context: The study focuses on the potential impact of Sorafenib on SOAT1 protein in liver cancer treatment, addressing the need for more effective therapeutic options. Research aim: To explore the effects of Sorafenib on the activity of SOAT1 protein in liver cancer cells. Methodology: Molecular docking was employed to analyze the interaction between Sorafenib and SOAT1 protein. Findings: The study revealed a significant effect of Sorafenib on the stability and activity of SOAT1 protein, suggesting its potential as a treatment for liver cancer. Theoretical importance: This research highlights the molecular mechanism underlying Sorafenib's anti-cancer properties, contributing to the understanding of its therapeutic effects. Data collection: Data on the molecular structure of Sorafenib and SOAT1 protein were obtained from computational simulations and databases. Analysis procedures: Molecular docking simulations were performed to predict the binding interactions between Sorafenib and SOAT1 protein. Question addressed: How does Sorafenib influence the activity of SOAT1 protein and what are the implications for liver cancer treatment? Conclusion: The study demonstrates the potential of Sorafenib as a targeted therapy for liver cancer by affecting the activity of SOAT1 protein. Reviewers' Comments: The study provides valuable insights into the molecular basis of Sorafenib's action on SOAT1 protein, suggesting its therapeutic potential. To enhance the methodology, the authors could consider validating the docking results with experimental data for further validation.

Keywords: liver cancer, sorafenib, SOAT1, molecular docking

Procedia PDF Downloads 17