Search results for: absolutely continuous function
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 7124

Search results for: absolutely continuous function

6674 Literature as a Tool for Sustenance of Human Dignity in the 21st Century

Authors: Arubi Thompson Abari

Abstract:

Globally, a writer is absolutely necessary to the society, for he mirrors and projects the society, grumbles and protects against the ills that hinders its development. A writer is committed to the language, social-cultural, political and economic factors that determine the sustenance of human dignity in the society. In this 21st century. The literary artist holds literature as a tool for the restoration and sustenance of human dignity. In Nigeria, literature is politically committed because colonialism gives birth to the modern Nigerian literature. Literature thus was regarded as one of the greatest weapons against colonialism in Nigeria. Nigerian literature is aimed at the restoration and sustenance of the dignity of Nigerians in the 21st century. A literary writer is a member of the society and his sensibility is conditioned by the socio-political situations around him. A writer cannot be excused from the task of regeneration and restoration of his past lost glorious days that must be done. This academic paper therefore showcases the efficacy of literature in bringing about the sustenance of human dignity in the 21st century. Consequently, the paper in its introduction clarifies some vital concepts. It discusses the forms of literature, portrays the ability and capability of literature as a tool for the sustenance of human dignity globally, and makes useful recommendations for the growth of knowledge in the 21st century and beyond.

Keywords: literature, sustenance, human dignity, 21st century

Procedia PDF Downloads 91
6673 Automated Ultrasound Carotid Artery Image Segmentation Using Curvelet Threshold Decomposition

Authors: Latha Subbiah, Dhanalakshmi Samiappan

Abstract:

In this paper, we propose denoising Common Carotid Artery (CCA) B mode ultrasound images by a decomposition approach to curvelet thresholding and automatic segmentation of the intima media thickness and adventitia boundary. By decomposition, the local geometry of the image, its direction of gradients are well preserved. The components are combined into a single vector valued function, thus removes noise patches. Double threshold is applied to inherently remove speckle noise in the image. The denoised image is segmented by active contour without specifying seed points. Combined with level set theory, they provide sub regions with continuous boundaries. The deformable contours match to the shapes and motion of objects in the images. A curve or a surface under constraints is developed from the image with the goal that it is pulled into the necessary features of the image. Region based and boundary based information are integrated to achieve the contour. The method treats the multiplicative speckle noise in objective and subjective quality measurements and thus leads to better-segmented results. The proposed denoising method gives better performance metrics compared with other state of art denoising algorithms.

Keywords: curvelet, decomposition, levelset, ultrasound

Procedia PDF Downloads 343
6672 Termite Brick Temperature and Relative Humidity by Continuous Monitoring Technique

Authors: Khalid Abdullah Alshuhail, Syrif Junidi, Ideisan Abu-Abdoum, Abdulsalam Aldawoud

Abstract:

For the intention of reducing energy consumption, a proposed construction brick was made of imitation termite mound soil referred here as termite brick (TB). To calculate the thermal performance, a real case model was constructed by using this biomimetic brick for testing purposes. This paper aims at investigating the thermal performance of this brick during different climatic months. Its thermal behaviour was thoroughly studied over the course of four months by using continuous method (CMm). The main parameters were focused on temperature and relative humidity. It was found that the TB does not perform similarly in all four months and/or in all orientations. Each four-month model study was deeply analyzed. By using the CMm method, the model was also examined. The measuring period shows generally that internal temperature and internal humidity are higher in the roof within 2 degrees and lowest at north wall orientation. The relative humidity was also investigated systematically. The paper reveals more interesting findings.

Keywords: building material, continious monitoring, orientation, wall, temprature

Procedia PDF Downloads 125
6671 Evaluation and Association of Thyroid Function Tests with Liver Function Parameters LDL and LDH Level Before and after I131 Therapy

Authors: Sabika Rafiq, Rubaida Mehmood, Sajid Hussain, Atia Iqbal

Abstract:

Background and objectives: The pathogenesis of liver function abnormalities and cardiac dysfunction in hyperthyroid patients after I131 treatment is still unclear. This study aimed to determine the effects of radioiodine I131 on liver function parameters, lactate dehydrogenase (LDH) and low-density lipoproteins (LDL) before and after I131 therapy hyperthyroidism patients. Material & Methods: A total of 52 patients of hyperthyroidism recommended for I131were involved in this study with ages ranging from 12–65 years (mean age=38.6±14.8 & BMI=11.5±3.7). The significance of the differences between the results of 1st, 2nd and 3rd-time serum analysis was assessed by unpaired student’s t-test. Associations between the parameters were assessed by Spearman correlation analysis. Results: Significant variations were observed for thyroid profile free FT3 (p=0.04), FT4 (p=0.01), TSH (p=0.005) during the follow-up treatment. Before taking I131 (serum analyzed at 1st time), negative correlation of FT3 with AST (r=-0.458, p=0.032) and LDL (r=-0.454, p=0.039) were observed. During 2nd time (after stopping carbimazole), no correlation was assessed. Two months after the administration of I131 drops, a significant negative association of FT3 (r=-0.62, p=0.04) and FT4(r=-0.61, p=0.02) with ALB were observed. FT3(r=-0.82, p=0.00) & FT4 (r=-0.71, p=0.00) also showed negative correlation with LDL after I131 therapy. Whereas TSH showed significant positive association with ALB (r=0.61, p=0.01) and LDL (r=0.70, p=0.00) respectively. Conclusion: Current findings suggested that the association of TFTs with biochemical parameters in patients with goiter recommended for iodine therapy is an important diagnostic and therapeutic tool. The significant changes increased in transaminases and low-density lipoprotein levels after taking I131drops are alarming signs for heart and liver function abnormalities and warrant physicians' attention on an urgent basis.

Keywords: hyperthyroidism, carbimazole, radioiodine I131, liver functions, low-density lipoprotein

Procedia PDF Downloads 155
6670 Modeling Engagement with Multimodal Multisensor Data: The Continuous Performance Test as an Objective Tool to Track Flow

Authors: Mohammad H. Taheri, David J. Brown, Nasser Sherkat

Abstract:

Engagement is one of the most important factors in determining successful outcomes and deep learning in students. Existing approaches to detect student engagement involve periodic human observations that are subject to inter-rater reliability. Our solution uses real-time multimodal multisensor data labeled by objective performance outcomes to infer the engagement of students. The study involves four students with a combined diagnosis of cerebral palsy and a learning disability who took part in a 3-month trial over 59 sessions. Multimodal multisensor data were collected while they participated in a continuous performance test. Eye gaze, electroencephalogram, body pose, and interaction data were used to create a model of student engagement through objective labeling from the continuous performance test outcomes. In order to achieve this, a type of continuous performance test is introduced, the Seek-X type. Nine features were extracted including high-level handpicked compound features. Using leave-one-out cross-validation, a series of different machine learning approaches were evaluated. Overall, the random forest classification approach achieved the best classification results. Using random forest, 93.3% classification for engagement and 42.9% accuracy for disengagement were achieved. We compared these results to outcomes from different models: AdaBoost, decision tree, k-Nearest Neighbor, naïve Bayes, neural network, and support vector machine. We showed that using a multisensor approach achieved higher accuracy than using features from any reduced set of sensors. We found that using high-level handpicked features can improve the classification accuracy in every sensor mode. Our approach is robust to both sensor fallout and occlusions. The single most important sensor feature to the classification of engagement and distraction was shown to be eye gaze. It has been shown that we can accurately predict the level of engagement of students with learning disabilities in a real-time approach that is not subject to inter-rater reliability, human observation or reliant on a single mode of sensor input. This will help teachers design interventions for a heterogeneous group of students, where teachers cannot possibly attend to each of their individual needs. Our approach can be used to identify those with the greatest learning challenges so that all students are supported to reach their full potential.

Keywords: affective computing in education, affect detection, continuous performance test, engagement, flow, HCI, interaction, learning disabilities, machine learning, multimodal, multisensor, physiological sensors, student engagement

Procedia PDF Downloads 95
6669 Group Decision Making through Interval-Valued Intuitionistic Fuzzy Soft Set TOPSIS Method Using New Hybrid Score Function

Authors: Syed Talib Abbas Raza, Tahseen Ahmed Jilani, Saleem Abdullah

Abstract:

This paper presents interval-valued intuitionistic fuzzy soft sets based TOPSIS method for group decision making. The interval-valued intuitionistic fuzzy soft set is a mutation of an interval-valued intuitionistic fuzzy set and soft set. In group decision making problems IVIFSS makes the process much more algebraically elegant. We have used weighted arithmetic averaging operator for aggregating the information and define a new Hybrid Score Function as metric tool for comparison between interval-valued intuitionistic fuzzy values. In an illustrative example we have applied the developed method to a criminological problem. We have developed a group decision making model for integrating the imprecise and hesitant evaluations of multiple law enforcement agencies working on target killing cases in the country.

Keywords: group decision making, interval-valued intuitionistic fuzzy soft set, TOPSIS, score function, criminology

Procedia PDF Downloads 605
6668 Spatial Interpolation of Intermediate Soil Properties to Enhance Geotechnical Surveying for Foundation Design

Authors: Yelbek B. Utepov, Assel T. Mukhamejanova, Aliya K. Aldungarova, Aida G. Nazarova, Sabit A. Karaulov, Nurgul T. Alibekova, Aigul K. Kozhas, Dias Kazhimkanuly, Akmaral K. Tleubayeva

Abstract:

This research focuses on enhancing geotechnical surveying for foundation design through the spatial interpolation of intermediate soil properties. Traditional geotechnical practices rely on discrete data from borehole drilling, soil sampling, and laboratory analyses, often neglecting the continuous nature of soil properties and disregarding values in intermediate locations. This study challenges these omissions by emphasizing interpolation techniques such as Kriging, Inverse Distance Weighting, and Spline interpolation to capture the nuanced spatial variations in soil properties. The methodology is applied to geotechnical survey data from two construction sites in Astana, Kazakhstan, revealing continuous representations of Young's Modulus, Cohesion, and Friction Angle. The spatial heatmaps generated through interpolation offered valuable insights into the subsurface environment, highlighting heterogeneity and aiding in more informed foundation design decisions for considered cites. Moreover, intriguing patterns of heterogeneity, as well as visual clusters and transitions between soil classes, were explored within seemingly uniform layers. The study bridges the gap between discrete borehole samples and the continuous subsurface, contributing to the evolution of geotechnical engineering practices. The proposed approach, utilizing open-source software geographic information systems, provides a practical tool for visualizing soil characteristics and may pave the way for future advancements in geotechnical surveying and foundation design.

Keywords: soil mechanical properties, spatial interpolation, inverse distance weighting, heatmaps

Procedia PDF Downloads 86
6667 Application of Causal Inference and Discovery in Curriculum Evaluation and Continuous Improvement

Authors: Lunliang Zhong, Bin Duan

Abstract:

The undergraduate graduation project is a vital part of the higher education curriculum, crucial for engineering accreditation. Current evaluations often summarize data without identifying underlying issues. This study applies the Peter-Clark algorithm to analyze causal relationships within the graduation project data of an Electronics and Information Engineering program, creating a causal model. Structural equation modeling confirmed the model's validity. The analysis reveals key teaching stages affecting project success, uncovering problems in the process. Introducing causal discovery and inference into project evaluation helps identify issues and propose targeted improvement measures. The effectiveness of these measures is validated by comparing the learning outcomes of two student cohorts, stratified by confounding factors, leading to improved teaching quality.

Keywords: causal discovery, causal inference, continuous improvement, Peter-Clark algorithm, structural equation modeling

Procedia PDF Downloads 20
6666 Study the effect of bulk traps on Solar Blind Photodetector Based on an IZTO/β Ga2O3/ITO Schottky Diode

Authors: Laboratory of Semiconducting, Metallic Materials (LMSM) Biskra Algeria

Abstract:

InZnSnO2 (IZTO)/β-Ga2O3 Schottky solar barrier photodetector (PhD) exposed to 255 nm was simulated and compared to the measurement. Numerical simulations successfully reproduced the photocurrent at reverse bias and response by taking into account several factors, such as conduction mechanisms and material parameters. By adopting reducing the density of the trap as an improvement. The effect of reducing the bulk trap densities on the photocurrent, response, and time-dependent (continuous conductivity) was studied. As the trap density decreased, the photocurrent increased. The response was 0.04 A/W for the low Ga2O3 trap density. The estimated decay time for the lowest intensity ET (0.74, 1.04 eV) is 0.05 s and is shorter at ∼0.015 s for ET (0.55 eV). This indicates that the shallow traps had the dominant effect (ET = 0.55 eV) on the continuous photoconductivity phenomenon. Furthermore, with decreasing trap densities, this PhD can be considered as a self-powered solar-blind photodiode (SBPhD).

Keywords: IZTO/β-Ga2O3, self-powered solar-blind photodetector, numerical simulation, bulk traps

Procedia PDF Downloads 87
6665 A Distribution Free Test for Censored Matched Pairs

Authors: Ayman Baklizi

Abstract:

This paper discusses the problem of testing hypotheses about the lifetime distributions of a matched pair based on censored data. A distribution free test based on a runs statistic is proposed. Its null distribution and power function are found in a simple convenient form. Some properties of the test statistic and its power function are studied.

Keywords: censored data, distribution free, matched pair, runs statistics

Procedia PDF Downloads 288
6664 Cellular Automata Using Fractional Integral Model

Authors: Yasser F. Hassan

Abstract:

In this paper, a proposed model of cellular automata is studied by means of fractional integral function. A cellular automaton is a decentralized computing model providing an excellent platform for performing complex computation with the help of only local information. The paper discusses how using fractional integral function for representing cellular automata memory or state. The architecture of computing and learning model will be given and the results of calibrating of approach are also given.

Keywords: fractional integral, cellular automata, memory, learning

Procedia PDF Downloads 415
6663 Characterization of Surface Suction Grippers for Continuous-Discontinuous Fiber Reinforced Semi-Finished Parts of an Automated Handling and Preforming Operation

Authors: Jürgen Fleischer, Woramon Pangboonyanon, Dominic Lesage

Abstract:

Non-metallic lightweight materials such as fiber reinforced plastics (FRP) become very significant at present. Prepregs e.g. SMC and unidirectional tape (UD-tape) are one of raw materials used to produce FRP. This study concerns with the manufacturing steps of handling and preforming of this UD-SMC and focuses on the investigation of gripper characteristics regarding gripping forces in normal and lateral direction, in order to identify suitable operating pressures for a secure gripping operation. A reliable handling and preforming operation results in a higher adding value of the overall process chain. As a result, the suitable operating pressures depending on travelling direction for each material type could be shown. Moreover, system boundary conditions regarding allowable pulling force in normal and lateral directions during preforming could be measured.

Keywords: continuous-discontinuous fiber reinforced plastics, UD-SMC-prepreg, handling, preforming, prepregs, sheet moulding compounds, surface suction gripper

Procedia PDF Downloads 223
6662 Establishing a Computational Screening Framework to Identify Environmental Exposures Using Untargeted Gas-Chromatography High-Resolution Mass Spectrometry

Authors: Juni C. Kim, Anna R. Robuck, Douglas I. Walker

Abstract:

The human exposome, which includes chemical exposures over the lifetime and their effects, is now recognized as an important measure for understanding human health; however, the complexity of the data makes the identification of environmental chemicals challenging. The goal of our project was to establish a computational workflow for the improved identification of environmental pollutants containing chlorine or bromine. Using the “pattern. search” function available in the R package NonTarget, we wrote a multifunctional script that searches mass spectral clusters from untargeted gas-chromatography high-resolution mass spectrometry (GC-HRMS) for the presence of spectra consistent with chlorine and bromine-containing organic compounds. The “pattern. search” function was incorporated into a different function that allows the evaluation of clusters containing multiple analyte fragments, has multi-core support, and provides a simplified output identifying listing compounds containing chlorine and/or bromine. The new function was able to process 46,000 spectral clusters in under 8 seconds and identified over 150 potential halogenated spectra. We next applied our function to a deidentified dataset from patients diagnosed with primary biliary cholangitis (PBC), primary sclerosing cholangitis (PSC), and healthy controls. Twenty-two spectra corresponded to potential halogenated compounds in the PSC and PBC dataset, including six significantly different in PBC patients, while four differed in PSC patients. We have developed an improved algorithm for detecting halogenated compounds in GC-HRMS data, providing a strategy for prioritizing exposures in the study of human disease.

Keywords: exposome, metabolome, computational metabolomics, high-resolution mass spectrometry, exposure, pollutants

Procedia PDF Downloads 138
6661 Analysis of Advancements in Process Modeling and Reengineering at Fars Regional Electric Company, Iran

Authors: Mohammad Arabi

Abstract:

Business Process Reengineering (BPR) is a systematic approach to fundamentally redesign organizational processes to achieve significant improvements in organizational performance. At Fars Regional Electric Company, implementing BPR is deemed essential to increase productivity, reduce costs, and improve service quality. This article examines how BPR can help enhance the performance of Fars Regional Electric Company. The objective of this research is to evaluate and analyze the advancements in process modeling and reengineering at Fars Regional Electric Company and to provide solutions for improving the productivity and efficiency of organizational processes. This study aims to demonstrate how BPR can be used to improve organizational processes and enhance the overall performance of the company. This research employs both qualitative and quantitative research methods and includes interviews with senior managers and experts at Fars Regional Electric Company. The analytical tools include process modeling software such as Bizagi and ARIS, and statistical analysis software such as SPSS and Minitab. Data analysis was conducted using advanced statistical methods. The results indicate that the use of BPR techniques can lead to a significant reduction in process execution time and overall improvement in quality. Implementing BPR at Fars Regional Electric Company has led to increased productivity, reduced costs, and improved overall performance of the company. This study shows that with proper implementation of BPR and the use of modeling tools, the company can achieve significant improvements in its processes. Recommendations: (1) Continuous Training for Staff: Invest in continuous training of staff to enhance their skills and knowledge in BPR. (2) Use of Advanced Technologies: Utilize modeling and analysis software to improve processes. (3) Implementation of Effective Management Systems: Employ knowledge and information management systems to enhance organizational performance. (4) Continuous Monitoring and Review of Processes: Regularly review and revise processes to ensure ongoing improvements. This article highlights the importance of improving organizational processes at Fars Regional Electric Company and recommends that managers and decision-makers at the company seriously consider reengineering processes and utilizing modeling technologies to achieve developmental goals and continuous improvement.

Keywords: business process reengineering, electric company, Fars province, process modeling advancements

Procedia PDF Downloads 51
6660 Low-Voltage and Low-Power Bulk-Driven Continuous-Time Current-Mode Differentiator Filters

Authors: Ravi Kiran Jaladi, Ezz I. El-Masry

Abstract:

Emerging technologies such as ultra-wide band wireless access technology that operate at ultra-low power present several challenges due to their inherent design that limits the use of voltage-mode filters. Therefore, Continuous-time current-mode (CTCM) filters have become very popular in recent times due to the fact they have a wider dynamic range, improved linearity, and extended bandwidth compared to their voltage-mode counterparts. The goal of this research is to develop analog filters which are suitable for the current scaling CMOS technologies. Bulk-driven MOSFET is one of the most popular low power design technique for the existing challenges, while other techniques have obvious shortcomings. In this work, a CTCM Gate-driven (GD) differentiator has been presented with a frequency range from dc to 100MHz which operates at very low supply voltage of 0.7 volts. A novel CTCM Bulk-driven (BD) differentiator has been designed for the first time which reduces the power consumption multiple times that of GD differentiator. These GD and BD differentiator has been simulated using CADENCE TSMC 65nm technology for all the bilinear and biquadratic band-pass frequency responses. These basic building blocks can be used to implement the higher order filters. A 6th order cascade CTCM Chebyshev band-pass filter has been designed using the GD and BD techniques. As a conclusion, a low power GD and BD 6th order chebyshev stagger-tuned band-pass filter was simulated and all the parameters obtained from all the resulting realizations are analyzed and compared. Monte Carlo analysis is performed for both the 6th order filters and the results of sensitivity analysis are presented.

Keywords: bulk-driven (BD), continuous-time current-mode filters (CTCM), gate-driven (GD)

Procedia PDF Downloads 261
6659 Continuous Functions Modeling with Artificial Neural Network: An Improvement Technique to Feed the Input-Output Mapping

Authors: A. Belayadi, A. Mougari, L. Ait-Gougam, F. Mekideche-Chafa

Abstract:

The artificial neural network is one of the interesting techniques that have been advantageously used to deal with modeling problems. In this study, the computing with artificial neural network (CANN) is proposed. The model is applied to modulate the information processing of one-dimensional task. We aim to integrate a new method which is based on a new coding approach of generating the input-output mapping. The latter is based on increasing the neuron unit in the last layer. Accordingly, to show the efficiency of the approach under study, a comparison is made between the proposed method of generating the input-output set and the conventional method. The results illustrated that the increasing of the neuron units, in the last layer, allows to find the optimal network’s parameters that fit with the mapping data. Moreover, it permits to decrease the training time, during the computation process, which avoids the use of computers with high memory usage.

Keywords: neural network computing, continuous functions generating the input-output mapping, decreasing the training time, machines with big memories

Procedia PDF Downloads 283
6658 The High Precision of Magnetic Detection with Microwave Modulation in Solid Spin Assembly of NV Centres in Diamond

Authors: Zongmin Ma, Shaowen Zhang, Yueping Fu, Jun Tang, Yunbo Shi, Jun Liu

Abstract:

Solid-state quantum sensors are attracting wide interest because of their high sensitivity at room temperature. In particular, spin properties of nitrogen–vacancy (NV) color centres in diamond make them outstanding sensors of magnetic fields, electric fields and temperature under ambient conditions. Much of the work on NV magnetic sensing has been done so as to achieve the smallest volume, high sensitivity of NV ensemble-based magnetometry using micro-cavity, light-trapping diamond waveguide (LTDW), nano-cantilevers combined with MEMS (Micro-Electronic-Mechanical System) techniques. Recently, frequency-modulated microwaves with continuous optical excitation method have been proposed to achieve high sensitivity of 6 μT/√Hz using individual NV centres at nanoscale. In this research, we built-up an experiment to measure static magnetic field through continuous wave optical excitation with frequency-modulated microwaves method under continuous illumination with green pump light at 532 nm, and bulk diamond sample with a high density of NV centers (1 ppm). The output of the confocal microscopy was collected by an objective (NA = 0.7) and detected by a high sensitivity photodetector. We design uniform and efficient excitation of the micro strip antenna, which is coupled well with the spin ensembles at 2.87 GHz for zero-field splitting of the NV centers. Output of the PD signal was sent to an LIA (Lock-In Amplifier) modulated signal, generated by the microwave source by IQ mixer. The detected signal is received by the photodetector, and the reference signal enters the lock-in amplifier to realize the open-loop detection of the NV atomic magnetometer. We can plot ODMR spectra under continuous-wave (CW) microwave. Due to the high sensitivity of the lock-in amplifier, the minimum detectable value of the voltage can be measured, and the minimum detectable frequency can be made by the minimum and slope of the voltage. The magnetic field sensitivity can be derived from η = δB√T corresponds to a 10 nT minimum detectable shift in the magnetic field. Further, frequency analysis of the noise in the system indicates that at 10Hz the sensitivity less than 10 nT/√Hz.

Keywords: nitrogen-vacancy (NV) centers, frequency-modulated microwaves, magnetic field sensitivity, noise density

Procedia PDF Downloads 440
6657 An Algorithm to Find Fractional Edge Domination Number and Upper Fractional Edge Domination Number of an Intuitionistic Fuzzy Graph

Authors: Karunambigai Mevani Govindasamy, Sathishkumar Ayyappan

Abstract:

In this paper, we formulate the algorithm to find out the dominating function parameters of Intuitionistic Fuzzy Graphs(IFG). The methodology we adopted here is converting any physical problem into an IFG, and that has been transformed into Intuitionistic Fuzzy Matrix. Using Linear Program Solver software (LiPS), we found the defined parameters for the given IFG. We obtained these parameters for a path and cycle IFG. This study can be extended to other varieties of IFG. In particular, we obtain the definition of edge dominating function, minimal edge dominating function, fractional edge domination number (γ_if^') and upper fractional edge domination number (Γ_if^') of an intuitionistic fuzzy graph. Also, we formulated an algorithm which is appropriate to work on LiPS to find fractional edge domination number and upper fractional edge domination number of an IFG.

Keywords: fractional edge domination number, intuitionistic fuzzy cycle, intuitionistic fuzzy graph, intuitionistic fuzzy path

Procedia PDF Downloads 177
6656 Application of Genetic Algorithm with Multiobjective Function to Improve the Efficiency of Photovoltaic Thermal System

Authors: Sonveer Singh, Sanjay Agrawal, D. V. Avasthi, Jayant Shekhar

Abstract:

The aim of this paper is to improve the efficiency of photovoltaic thermal (PVT) system with the help of Genetic Algorithms with multi-objective function. There are some parameters that affect the efficiency of PVT system like depth and length of the channel, velocity of flowing fluid through the channel, thickness of the tedlar and glass, temperature of inlet fluid i.e. all above parameters are considered for optimization. An attempt has been made to the model and optimizes the parameters of glazed hybrid single channel PVT module when two objective functions have been considered separately. The two objective function for optimization of PVT module is overall electrical and thermal efficiency. All equations for PVT module have been derived. Using genetic algorithms (GAs), above two objective functions of the system has been optimized separately and analysis has been carried out for two cases. Two cases are: Case-I; Improvement in electrical and thermal efficiency when overall electrical efficiency is optimized, Case-II; Improvement in electrical and thermal efficiency when overall thermal efficiency is optimized. All the parameters that are used in genetic algorithms are the parameters that could be changed, and the non-changeable parameters, like solar radiation, ambient temperature cannot be used in the algorithm. It has been observed that electrical efficiency (14.08%) and thermal efficiency (19.48%) are obtained when overall thermal efficiency was an objective function for optimization. It is observed that GA is a very efficient technique to estimate the design parameters of hybrid single channel PVT module.

Keywords: genetic algorithm, energy, exergy, PVT module, optimization

Procedia PDF Downloads 606
6655 A Versatile Data Processing Package for Ground-Based Synthetic Aperture Radar Deformation Monitoring

Authors: Zheng Wang, Zhenhong Li, Jon Mills

Abstract:

Ground-based synthetic aperture radar (GBSAR) represents a powerful remote sensing tool for deformation monitoring towards various geohazards, e.g. landslides, mudflows, avalanches, infrastructure failures, and the subsidence of residential areas. Unlike spaceborne SAR with a fixed revisit period, GBSAR data can be acquired with an adjustable temporal resolution through either continuous or discontinuous operation. However, challenges arise from processing high temporal-resolution continuous GBSAR data, including the extreme cost of computational random-access-memory (RAM), the delay of displacement maps, and the loss of temporal evolution. Moreover, repositioning errors between discontinuous campaigns impede the accurate measurement of surface displacements. Therefore, a versatile package with two complete chains is developed in this study in order to process both continuous and discontinuous GBSAR data and address the aforementioned issues. The first chain is based on a small-baseline subset concept and it processes continuous GBSAR images unit by unit. Images within a window form a basic unit. By taking this strategy, the RAM requirement is reduced to only one unit of images and the chain can theoretically process an infinite number of images. The evolution of surface displacements can be detected as it keeps temporarily-coherent pixels which are present only in some certain units but not in the whole observation period. The chain supports real-time processing of the continuous data and the delay of creating displacement maps can be shortened without waiting for the entire dataset. The other chain aims to measure deformation between discontinuous campaigns. Temporal averaging is carried out on a stack of images in a single campaign in order to improve the signal-to-noise ratio of discontinuous data and minimise the loss of coherence. The temporal-averaged images are then processed by a particular interferometry procedure integrated with advanced interferometric SAR algorithms such as robust coherence estimation, non-local filtering, and selection of partially-coherent pixels. Experiments are conducted using both synthetic and real-world GBSAR data. Displacement time series at the level of a few sub-millimetres are achieved in several applications (e.g. a coastal cliff, a sand dune, a bridge, and a residential area), indicating the feasibility of the developed GBSAR data processing package for deformation monitoring of a wide range of scientific and practical applications.

Keywords: ground-based synthetic aperture radar, interferometry, small baseline subset algorithm, deformation monitoring

Procedia PDF Downloads 163
6654 Fractal Analysis of Some Bifurcations of Discrete Dynamical Systems in Higher Dimensions

Authors: Lana Horvat Dmitrović

Abstract:

The main purpose of this paper is to study the box dimension as fractal property of bifurcations of discrete dynamical systems in higher dimensions. The paper contains the fractal analysis of the orbits near the hyperbolic and non-hyperbolic fixed points in discrete dynamical systems. It is already known that in one-dimensional case the orbit near the hyperbolic fixed point has the box dimension equal to zero. On the other hand, the orbit near the non-hyperbolic fixed point has strictly positive box dimension which is connected to the non-degeneracy condition of certain bifurcation. One of the main results in this paper is the generalisation of results about box dimension near the hyperbolic and non-hyperbolic fixed points to higher dimensions. In the process of determining box dimension, the restriction of systems to stable, unstable and center manifolds, Lipschitz property of box dimension and the notion of projective box dimension are used. The analysis of the bifurcations in higher dimensions with one multiplier on the unit circle is done by using the normal forms on one-dimensional center manifolds. This specific change in box dimension of an orbit at the moment of bifurcation has already been explored for some bifurcations in one and two dimensions. It was shown that specific values of box dimension are connected to appropriate bifurcations such as fold, flip, cusp or Neimark-Sacker bifurcation. This paper further explores this connection of box dimension as fractal property to some specific bifurcations in higher dimensions, such as fold-flip and flip-Neimark-Sacker. Furthermore, the application of the results to the unit time map of continuous dynamical system near hyperbolic and non-hyperbolic singularities is presented. In that way, box dimensions which are specific for certain bifurcations of continuous systems can be obtained. The approach to bifurcation analysis by using the box dimension as specific fractal property of orbits can lead to better understanding of bifurcation phenomenon. It could also be useful in detecting the existence or nonexistence of bifurcations of discrete and continuous dynamical systems.

Keywords: bifurcation, box dimension, invariant manifold, orbit near fixed point

Procedia PDF Downloads 255
6653 Energy Consumption Modeling for Strawberry Greenhouse Crop by Adaptive Nero Fuzzy Inference System Technique: A Case Study in Iran

Authors: Azar Khodabakhshi, Elham Bolandnazar

Abstract:

Agriculture as the most important food manufacturing sector is not only the energy consumer, but also is known as energy supplier. Using energy is considered as a helpful parameter for analyzing and evaluating the agricultural sustainability. In this study, the pattern of energy consumption of strawberry greenhouses of Jiroft in Kerman province of Iran was surveyed. The total input energy required in the strawberries production was calculated as 113314.71 MJ /ha. Electricity with 38.34% contribution of the total energy was considered as the most energy consumer in strawberry production. In this study, Neuro Fuzzy networks was used for function modeling in the production of strawberries. Results showed that the best model for predicting the strawberries function had a correlation coefficient, root mean square error (RMSE) and mean absolute percentage error (MAPE) equal to 0.9849, 0.0154 kg/ha and 0.11% respectively. Regards to these results, it can be said that Neuro Fuzzy method can be well predicted and modeled the strawberry crop function.

Keywords: crop yield, energy, neuro-fuzzy method, strawberry

Procedia PDF Downloads 383
6652 Distances over Incomplete Diabetes and Breast Cancer Data Based on Bhattacharyya Distance

Authors: Loai AbdAllah, Mahmoud Kaiyal

Abstract:

Missing values in real-world datasets are a common problem. Many algorithms were developed to deal with this problem, most of them replace the missing values with a fixed value that was computed based on the observed values. In our work, we used a distance function based on Bhattacharyya distance to measure the distance between objects with missing values. Bhattacharyya distance, which measures the similarity of two probability distributions. The proposed distance distinguishes between known and unknown values. Where the distance between two known values is the Mahalanobis distance. When, on the other hand, one of them is missing the distance is computed based on the distribution of the known values, for the coordinate that contains the missing value. This method was integrated with Wikaya, a digital health company developing a platform that helps to improve prevention of chronic diseases such as diabetes and cancer. In order for Wikaya’s recommendation system to work distance between users need to be measured. Since there are missing values in the collected data, there is a need to develop a distance function distances between incomplete users profiles. To evaluate the accuracy of the proposed distance function in reflecting the actual similarity between different objects, when some of them contain missing values, we integrated it within the framework of k nearest neighbors (kNN) classifier, since its computation is based only on the similarity between objects. To validate this, we ran the algorithm over diabetes and breast cancer datasets, standard benchmark datasets from the UCI repository. Our experiments show that kNN classifier using our proposed distance function outperforms the kNN using other existing methods.

Keywords: missing values, incomplete data, distance, incomplete diabetes data

Procedia PDF Downloads 225
6651 A Lifetime-Enhancing Monitoring Node Distribution Using Minimum Spanning Tree in Mobile Ad Hoc Networks

Authors: Sungchul Ha, Hyunwoo Kim

Abstract:

In mobile ad hoc networks, all nodes in a network only have limited resources and calculation ability. Therefore communication topology which have long lifetime is good for all nodes in mobile ad hoc networks. There are a variety of researches on security problems in wireless ad hoc networks. The existing many researches try to make efficient security schemes to reduce network power consumption and enhance network lifetime. Because a new node can join the network at any time, the wireless ad hoc networks are exposed to various threats and can be destroyed by attacks. Resource consumption is absolutely necessary to secure networks, but more resource consumption can be a critical problem to network lifetime. This paper focuses on efficient monitoring node distribution to enhance network lifetime in wireless ad hoc networks. Since the wireless ad hoc networks cannot use centralized infrastructure and security systems of wired networks, a new special IDS scheme is necessary. The scheme should not only cover all nodes in a network but also enhance the network lifetime. In this paper, we propose an efficient IDS node distribution scheme using minimum spanning tree (MST) method. The simulation results show that the proposed algorithm has superior performance in comparison with existing algorithms.

Keywords: MANETs, IDS, power control, minimum spanning tree

Procedia PDF Downloads 373
6650 Stabilization of Rotational Motion of Spacecrafts Using Quantized Two Torque Inputs Based on Random Dither

Authors: Yusuke Kuramitsu, Tomoaki Hashimoto, Hirokazu Tahara

Abstract:

The control problem of underactuated spacecrafts has attracted a considerable amount of interest. The control method for a spacecraft equipped with less than three control torques is useful when one of the three control torques had failed. On the other hand, the quantized control of systems is one of the important research topics in recent years. The random dither quantization method that transforms a given continuous signal to a discrete signal by adding artificial random noise to the continuous signal before quantization has also attracted a considerable amount of interest. The objective of this study is to develop the control method based on random dither quantization method for stabilizing the rotational motion of a rigid spacecraft with two control inputs. In this paper, the effectiveness of random dither quantization control method for the stabilization of rotational motion of spacecrafts with two torque inputs is verified by numerical simulations.

Keywords: spacecraft control, quantized control, nonlinear control, random dither method

Procedia PDF Downloads 180
6649 An Optimal Algorithm for Finding (R, Q) Policy in a Price-Dependent Order Quantity Inventory System with Soft Budget Constraint

Authors: S. Hamid Mirmohammadi, Shahrazad Tamjidzad

Abstract:

This paper is concerned with the single-item continuous review inventory system in which demand is stochastic and discrete. The budget consumed for purchasing the ordered items is not restricted but it incurs extra cost when exceeding specific value. The unit purchasing price depends on the quantity ordered under the all-units discounts cost structure. In many actual systems, the budget as a resource which is occupied by the purchased items is limited and the system is able to confront the resource shortage by charging more costs. Thus, considering the resource shortage costs as a part of system costs, especially when the amount of resource occupied by the purchased item is influenced by quantity discounts, is well motivated by practical concerns. In this paper, an optimization problem is formulated for finding the optimal (R, Q) policy, when the system is influenced by the budget limitation and a discount pricing simultaneously. Properties of the cost function are investigated and then an algorithm based on a one-dimensional search procedure is proposed for finding an optimal (R, Q) policy which minimizes the expected system costs .

Keywords: (R, Q) policy, stochastic demand, backorders, limited resource, quantity discounts

Procedia PDF Downloads 641
6648 Developing a Structured Example Space for Finding the Collision Points of Functions and Their Inverse

Authors: M. Saeed, A. Shahidzadeh

Abstract:

Interaction between teachers and learners requires applying a set of samples (examples) which helps to create coordination between the goals and methods. The main result and achievement and application of samples (examples) are that they can bring the teacher and learner to a shared understanding of the concept. mathematical concepts, and also one of the challenging issues in the discussion of the function is to find the collision points of functions of and, regarding that the example space of teachers is different in this issue, this paper aims to present an example space including several problems of the secondary school with the help of intuition and drawing various graphs of functions of and for more familiarity of teachers.

Keywords: inverse function, educational example, Mathematic example, example space

Procedia PDF Downloads 179
6647 Study on Optimal Control Strategy of PM2.5 in Wuhan, China

Authors: Qiuling Xie, Shanliang Zhu, Zongdi Sun

Abstract:

In this paper, we analyzed the correlation relationship among PM2.5 from other five Air Quality Indices (AQIs) based on the grey relational degree, and built a multivariate nonlinear regression equation model of PM2.5 and the five monitoring indexes. For the optimal control problem of PM2.5, we took the partial large Cauchy distribution of membership equation as satisfaction function. We established a nonlinear programming model with the goal of maximum performance to price ratio. And the optimal control scheme is given.

Keywords: grey relational degree, multiple linear regression, membership function, nonlinear programming

Procedia PDF Downloads 301
6646 Pyrroloquinoline Quinone Enhances the Mitochondrial Function by Increasing Beta-Oxidation and a Balanced Mitochondrial Recycling in Mice Granulosa Cells

Authors: Moustafa Elhamouly, Masayuki Shimada

Abstract:

The production of competent oocytes is essential for reproductivity in mammals. Maintenance of mitochondrial efficiency is required to supply the ATP necessary for granulosa cell proliferation during the follicular development process. Treatment with Pyrroloquinoline quinone (PQQ) has been reported to increase the number of ovulated oocytes and pups per delivery in mice by maintaining healthy mitochondrial function. This study aimed to elucidate how PQQ maintains mitochondrial function during ovarian follicle growth. To do this, both in vitro and in vivo experiments were performed with granulosa cells from superovulated immature (3-week-old) mice that were pretreated with or without PQQ. The effects of PQQ on beta-oxidation, mitochondrial function, mitophagy, and mitochondrial biogenesis were examined. PQQ increased beta-oxidation-related genes and CPT1 protein content in granulosa cells and this was associated with a decreased phosphorylation of P38 signaling protein. Using the fatty acid oxidation assay on the flux analyzer, PQQ increased the reliance of beta-oxidation on the endogenous fatty acids and was associated with a mild UCP-dependant mitochondrial uncoupling, ATP production, mitophagy, and mitochondrial biogenesis. PQQ also increased the expression of endogenous antioxidant enzymes. Thus, PQQ induced beta-oxidation in growing granulosa cells relying on endogenous fatty acids. And reduced the Reactive oxygen species (ROS) production by inducing a mild mitochondrial uncoupling with keeping high mitochondrial function. Damaged mitochondria were recycled by the induced mitophagy and replaced by the increased mitochondrial biogenesis. Collectively, PQQ may enhance reproductivity by maintaining the efficiency of mitochondria to produce enough ATP required for normal folliculogenesis.

Keywords: granulosa cells, mitochondrial uncoupling, mitophagy, pyrroloquinoline quinone (PQQ), reactive oxygen species (ROS).

Procedia PDF Downloads 83
6645 Data Driven Infrastructure Planning for Offshore Wind farms

Authors: Isha Saxena, Behzad Kazemtabrizi, Matthias C. M. Troffaes, Christopher Crabtree

Abstract:

The calculations done at the beginning of the life of a wind farm are rarely reliable, which makes it important to conduct research and study the failure and repair rates of the wind turbines under various conditions. This miscalculation happens because the current models make a simplifying assumption that the failure/repair rate remains constant over time. This means that the reliability function is exponential in nature. This research aims to create a more accurate model using sensory data and a data-driven approach. The data cleaning and data processing is done by comparing the Power Curve data of the wind turbines with SCADA data. This is then converted to times to repair and times to failure timeseries data. Several different mathematical functions are fitted to the times to failure and times to repair data of the wind turbine components using Maximum Likelihood Estimation and the Posterior expectation method for Bayesian Parameter Estimation. Initial results indicate that two parameter Weibull function and exponential function produce almost identical results. Further analysis is being done using the complex system analysis considering the failures of each electrical and mechanical component of the wind turbine. The aim of this project is to perform a more accurate reliability analysis that can be helpful for the engineers to schedule maintenance and repairs to decrease the downtime of the turbine.

Keywords: reliability, bayesian parameter inference, maximum likelihood estimation, weibull function, SCADA data

Procedia PDF Downloads 87