Search results for: parallel output
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 3146

Search results for: parallel output

2246 Predicting Daily Patient Hospital Visits Using Machine Learning

Authors: Shreya Goyal

Abstract:

The study aims to build user-friendly software to understand patient arrival patterns and compute the number of potential patients who will visit a particular health facility for a given period by using a machine learning algorithm. The underlying machine learning algorithm used in this study is the Support Vector Machine (SVM). Accurate prediction of patient arrival allows hospitals to operate more effectively, providing timely and efficient care while optimizing resources and improving patient experience. It allows for better allocation of staff, equipment, and other resources. If there's a projected surge in patients, additional staff or resources can be allocated to handle the influx, preventing bottlenecks or delays in care. Understanding patient arrival patterns can also help streamline processes to minimize waiting times for patients and ensure timely access to care for patients in need. Another big advantage of using this software is adhering to strict data protection regulations such as the Health Insurance Portability and Accountability Act (HIPAA) in the United States as the hospital will not have to share the data with any third party or upload it to the cloud because the software can read data locally from the machine. The data needs to be arranged in. a particular format and the software will be able to read the data and provide meaningful output. Using software that operates locally can facilitate compliance with these regulations by minimizing data exposure. Keeping patient data within the hospital's local systems reduces the risk of unauthorized access or breaches associated with transmitting data over networks or storing it in external servers. This can help maintain the confidentiality and integrity of sensitive patient information. Historical patient data is used in this study. The input variables used to train the model include patient age, time of day, day of the week, seasonal variations, and local events. The algorithm uses a Supervised learning method to optimize the objective function and find the global minima. The algorithm stores the values of the local minima after each iteration and at the end compares all the local minima to find the global minima. The strength of this study is the transfer function used to calculate the number of patients. The model has an output accuracy of >95%. The method proposed in this study could be used for better management planning of personnel and medical resources.

Keywords: machine learning, SVM, HIPAA, data

Procedia PDF Downloads 57
2245 Classification Based on Deep Neural Cellular Automata Model

Authors: Yasser F. Hassan

Abstract:

Deep learning structure is a branch of machine learning science and greet achievement in research and applications. Cellular neural networks are regarded as array of nonlinear analog processors called cells connected in a way allowing parallel computations. The paper discusses how to use deep learning structure for representing neural cellular automata model. The proposed learning technique in cellular automata model will be examined from structure of deep learning. A deep automata neural cellular system modifies each neuron based on the behavior of the individual and its decision as a result of multi-level deep structure learning. The paper will present the architecture of the model and the results of simulation of approach are given. Results from the implementation enrich deep neural cellular automata system and shed a light on concept formulation of the model and the learning in it.

Keywords: cellular automata, neural cellular automata, deep learning, classification

Procedia PDF Downloads 183
2244 A Different Approach to Optimize Fuzzy Membership Functions with Extended FIR Filter

Authors: Jun-Ho Chung, Sung-Hyun Yoo, In-Hwan Choi, Hyun-Kook Lee, Moon-Kyu Song, Choon-Ki Ahn

Abstract:

The extended finite impulse response (EFIR) filter is addressed to optimize membership functions (MFs) of the fuzzy model that has strong nonlinearity. MFs are important parts of the fuzzy logic system (FLS) and, thus optimizing MFs of FLS is one of approaches to improve the performance of output. We employ the EFIR as an alternative optimization option to nonlinear fuzzy model. The performance of EFIR is demonstrated on a fuzzy cruise control via a numerical example.

Keywords: fuzzy logic system, optimization, membership function, extended FIR filter

Procedia PDF Downloads 713
2243 Quantum Cum Synaptic-Neuronal Paradigm and Schema for Human Speech Output and Autism

Authors: Gobinathan Devathasan, Kezia Devathasan

Abstract:

Objective: To improve the current modified Broca-Wernicke-Lichtheim-Kussmaul speech schema and provide insight into autism. Methods: We reviewed the pertinent literature. Current findings, involving Brodmann areas 22, 46, 9,44,45,6,4 are based on neuropathology and functional MRI studies. However, in primary autism, there is no lucid explanation and changes described, whether neuropathology or functional MRI, appear consequential. Findings: We forward an enhanced model which may explain the enigma related to autism. Vowel output is subcortical and does need cortical representation whereas consonant speech is cortical in origin. Left lateralization is needed to commence the circuitry spin as our life have evolved with L-amino acids and left spin of electrons. A fundamental species difference is we are capable of three syllable-consonants and bi-syllable expression whereas cetaceans and songbirds are confined to single or dual consonants. The 4 key sites for speech are superior auditory cortex, Broca’s two areas, and the supplementary motor cortex. Using the Argand’s diagram and Reimann’s projection, we theorize that the Euclidean three dimensional synaptic neuronal circuits of speech are quantized to coherent waves, and then decoherence takes place at area 6 (spherical representation). In this quantum state complex, 3-consonant languages are instantaneously integrated and multiple languages can be learned, verbalized and differentiated. Conclusion: We postulate that evolutionary human speech is elevated to quantum interaction unlike cetaceans and birds to achieve the three consonants/bi-syllable speech. In classical primary autism, the sudden speech switches off and on noted in several cases could now be explained not by any anatomical lesion but failure of coherence. Area 6 projects directly into prefrontal saccadic area (8); and this further explains the second primary feature in autism: lack of eye contact. The third feature which is repetitive finger gestures, located adjacent to the speech/motor areas, are actual attempts to communicate with the autistic child akin to sign language for the deaf.

Keywords: quantum neuronal paradigm, cetaceans and human speech, autism and rapid magnetic stimulation, coherence and decoherence of speech

Procedia PDF Downloads 180
2242 Comparative Performance Analysis of Nonlinearity Cancellation Techniques for MOS-C Realization in Integrator Circuits

Authors: Hasan Çiçekli, Ahmet Gökçen, Uğur Çam

Abstract:

In this paper, a comparative performance analysis of mostly used four nonlinearity cancellation techniques used to realize the passive resistor by MOS transistors is presented. The comparison is done by using an integrator circuit which is employing sequentially Op-amp, OTRA and ICCII as active element. All of the circuits are implemented by MOS-C realization and simulated by PSPICE program using 0.35 µm process TSMC MOSIS model parameters. With MOS-C realization, the circuits became electronically tunable and fully integrable which is very important in IC design. The output waveforms, frequency responses, THD analysis results and features of the nonlinearity cancellation techniques are also given.

Keywords: integrator circuits, MOS-C realization, nonlinearity cancellation, tuneable resistors

Procedia PDF Downloads 523
2241 Analytical Study Of Holographic Polymer Dispersed Liquid Crystals Using Finite Difference Time Domain Method

Authors: N. R. Mohamad, H. Ono, H. Haroon, A. Salleh, N. M. Z. Hashim

Abstract:

In this research, we have studied and analyzed the modulation of light and liquid crystal in HPDLCs using Finite Domain Time Difference (FDTD) method. HPDLCs are modeled as a mixture of polymer and liquid crystals (LCs) that categorized as an anisotropic medium. FDTD method is directly solves Maxwell’s equation with less approximation, so this method can analyze more flexible and general approach for the arbitrary anisotropic media. As the results from FDTD simulation, the highest diffraction efficiency occurred at ±19 degrees (Bragg angle) using p polarization incident beam to Bragg grating, Q > 10 when the pitch is 1µm. Therefore, the liquid crystal is assumed to be aligned parallel to the grating constant vector during these parameters.

Keywords: birefringence, diffraction efficiency, finite domain time difference, nematic liquid crystals

Procedia PDF Downloads 451
2240 An Improved Photovolatic System Balancer Architecture

Authors: Chih-Chiang Hua, Yi-Hsiung Fang, Cyuan-Jyun Wong

Abstract:

An improved PV balancer for photovoltaic applications is proposed in this paper. The proposed PV balancer senses the voltage and current of PV module and adjusts the output voltage of converter. Thus, the PV system can implement maximum power point tracking (MPPT) independently for each module whether it is under shading, different irradiation or degradation of PV cell. In addition, the cost of PV balancer can be reduced due to the low power rating of converter. To assess the effectiveness of the proposed system, two PV balancers are designed and verified through simulation under different shading conditions. The proposed PV balancers can provide more energy than the traditional PV balancer.

Keywords: MPPT, partial shading, PV System, converter

Procedia PDF Downloads 283
2239 Effect of Fault Depth on Near-Fault Peak Ground Velocity

Authors: Yanyan Yu, Haiping Ding, Pengjun Chen, Yiou Sun

Abstract:

Fault depth is an important parameter to be determined in ground motion simulation, and peak ground velocity (PGV) demonstrates good application prospect. Using numerical simulation method, the variations of distribution and peak value of near-fault PGV with different fault depth were studied in detail, and the reason of some phenomena were discussed. The simulation results show that the distribution characteristics of PGV of fault-parallel (FP) component and fault-normal (FN) component are distinctly different; the value of PGV FN component is much larger than that of FP component. With the increase of fault depth, the distribution region of the FN component strong PGV moves forward along the rupture direction, while the strong PGV zone of FP component becomes gradually far away from the fault trace along the direction perpendicular to the strike. However, no matter FN component or FP component, the strong PGV distribution area and its value are both quickly reduced with increased fault depth. The results above suggest that the fault depth have significant effect on both FN component and FP component of near-fault PGV.

Keywords: fault depth, near-fault, PGV, numerical simulation

Procedia PDF Downloads 338
2238 Determination of Optimum Fin Wave Angle and Its Effect on the Performance of an Intercooler

Authors: Mahdi Hamzehei, Seyyed Amin Hakim, Nahid Taherian

Abstract:

Fins play an important role in increasing the efficiency of compact shell and tube heat exchangers by increasing heat transfer. The objective of this paper is to determine the optimum fin wave angle, as one of the geometric parameters affecting the efficiency of the heat exchangers. To this end, finite volume method is used to model and simulate the flow in heat exchanger. In this study, computational fluid dynamics simulations of wave channel are done. The results show that the wave angle affects the temperature output of the heat exchanger.

Keywords: fin wave angle, tube, intercooler, optimum, performance

Procedia PDF Downloads 370
2237 Three-Dimensional Fluid-Structure-Thermal Coupling Dynamics Simulation Model of a Gas-Filled Fluid-Resistance Damper and Experimental Verification

Authors: Wenxue Xu

Abstract:

Fluid resistance damper is an important damping element to attenuate vehicle vibration. It converts vibration energy into thermal energy dissipation through oil throttling. It is a typical fluid-solid-heat coupling problem. A complete three-dimensional flow-structure-thermal coupling dynamics simulation model of a gas-filled fluid-resistance damper was established. The flow-condition-based interpolation (FCBI) method and direct coupling calculation method, the unit's FCBI-C fluid numerical analysis method and iterative coupling calculation method are used to achieve the damper dynamic response of the piston rod under sinusoidal excitation; the air chamber inflation pressure, spring compression characteristics, constant flow passage cross-sectional area and oil parameters, etc. The system parameters, excitation frequency, and amplitude and other excitation parameters are analyzed and compared in detail for the effects of differential pressure characteristics, velocity characteristics, flow characteristics and dynamic response of valve opening, floating piston response and piston rod output force characteristics. Experiments were carried out on some simulation analysis conditions. The results show that the node-based FCBI (flow-condition-based interpolation) fluid numerical analysis method and direct coupling calculation method can better guarantee the conservation of flow field calculation, and the calculation step is larger, but the memory is also larger; if the chamber inflation pressure is too low, the damper will become cavitation. The inflation pressure will cause the speed characteristic hysteresis to increase, and the sealing requirements are too strict. The spring compression characteristics have a great influence on the damping characteristics of the damper, and reasonable damping characteristic needs to properly design the spring compression characteristics; the larger the cross-sectional area of the constant flow channel, the smaller the maximum output force, but the more stable when the valve plate is opening.

Keywords: damper, fluid-structure-thermal coupling, heat generation, heat transfer

Procedia PDF Downloads 135
2236 MIMO PID Controller of a Power Plant Boiler–Turbine Unit

Authors: N. Ben-Mahmoud, M. Elfandi, A. Shallof

Abstract:

This paper presents a methodology to design multivariable PID controllers for multi-input and multi-output systems. The proposed control strategy, which is centralized, combines of PID controllers. The proportional gains in the P controllers act as tuning parameters of (SISO) in order to modify the behavior of the loops almost independently. The design procedure consists of three steps: first, an ideal decoupler including integral action is determined. Second, the decoupler is approximated with PID controllers. Third, the proportional gains are tuned to achieve the specified performance. The proposed method is applied to representative processes.

Keywords: boiler turbine, MIMO, PID controller, control by decoupling, anti wind-up techniques

Procedia PDF Downloads 322
2235 Physical Parameters Influencing the Yield of Nigella Sativa Oil Extracted by Hydraulic Pressing

Authors: Hadjadj Naima, K. Mahdi, D. Belhachat, F. S. Ait Chaouche, A. Ferradji

Abstract:

The Nigella Sativa oil yield extracted by hydraulic pressing is influenced by the pressure temperature and size particles. The optimization of oil extraction is investigated. The rate of extraction of the whole seeds is very weak, a crushing of seeds is necessary to facilitate the extraction. This rate augments with the rise of the temperature and the pressure, and decrease of size particles. The best output (66%) is obtained for a granulometry lower than 1mm, a temperature of 50°C and a pressure of 120 bars.

Keywords: oil, Nigella sativa, extraction, optimization, temperature, pressure

Procedia PDF Downloads 469
2234 Overcoming 4-to-1 Decryption Failure of the Rabin Cryptosystem

Authors: Muhammad Rezal Kamel Ariffin, Muhammad Asyraf Asbullah

Abstract:

The square root modulo problem is a known primitive in designing an asymmetric cryptosystem. It was first attempted by Rabin. Decryption failure of the Rabin cryptosystem caused by the 4-to-1 decryption output is overcome efficiently in this work. The proposed scheme to overcome the decryption failure issue (known as the AAβ-cryptosystem) is constructed using a simple mathematical structure, it has low computational requirements and would enable communication devices with low computing power to deploy secure communication procedures efficiently.

Keywords: Rabin cryptosystem, 4-to-1 decryption failure, square root modulo problem, integer factorization problem

Procedia PDF Downloads 456
2233 Regularizing Software for Aerosol Particles

Authors: Christine Böckmann, Julia Rosemann

Abstract:

We present an inversion algorithm that is used in the European Aerosol Lidar Network for the inversion of data collected with multi-wavelength Raman lidar. These instruments measure backscatter coefficients at 355, 532, and 1064 nm, and extinction coefficients at 355 and 532 nm. The algorithm is based on manually controlled inversion of optical data which allows for detailed sensitivity studies and thus provides us with comparably high quality of the derived data products. The algorithm allows us to derive particle effective radius, volume, surface-area concentration with comparably high confidence. The retrieval of the real and imaginary parts of the complex refractive index still is a challenge in view of the accuracy required for these parameters in climate change studies in which light-absorption needs to be known with high accuracy. Single-scattering albedo (SSA) can be computed from the retrieve microphysical parameters and allows us to categorize aerosols into high and low absorbing aerosols. From mathematical point of view the algorithm is based on the concept of using truncated singular value decomposition as regularization method. This method was adapted to work for the retrieval of the particle size distribution function (PSD) and is called hybrid regularization technique since it is using a triple of regularization parameters. The inversion of an ill-posed problem, such as the retrieval of the PSD, is always a challenging task because very small measurement errors will be amplified most often hugely during the solution process unless an appropriate regularization method is used. Even using a regularization method is difficult since appropriate regularization parameters have to be determined. Therefore, in a next stage of our work we decided to use two regularization techniques in parallel for comparison purpose. The second method is an iterative regularization method based on Pade iteration. Here, the number of iteration steps serves as the regularization parameter. We successfully developed a semi-automated software for spherical particles which is able to run even on a parallel processor machine. From a mathematical point of view, it is also very important (as selection criteria for an appropriate regularization method) to investigate the degree of ill-posedness of the problem which we found is a moderate ill-posedness. We computed the optical data from mono-modal logarithmic PSD and investigated particles of spherical shape in our simulations. We considered particle radii as large as 6 nm which does not only cover the size range of particles in the fine-mode fraction of naturally occurring PSD but also covers a part of the coarse-mode fraction of PSD. We considered errors of 15% in the simulation studies. For the SSA, 100% of all cases achieve relative errors below 12%. In more detail, 87% of all cases for 355 nm and 88% of all cases for 532 nm are well below 6%. With respect to the absolute error for non- and weak-absorbing particles with real parts 1.5 and 1.6 in all modes the accuracy limit +/- 0.03 is achieved. In sum, 70% of all cases stay below +/-0.03 which is sufficient for climate change studies.

Keywords: aerosol particles, inverse problem, microphysical particle properties, regularization

Procedia PDF Downloads 334
2232 Dual Active Bridge Converter with Photovoltaic Arrays for DC Microgrids: Design and Analysis

Authors: Ahmed Atef, Mohamed Alhasheem, Eman Beshr

Abstract:

In this paper, an enhanced DC microgrid design is proposed using the DAB converter as a conversion unit in order to harvest the maximum power from the PV array. Each connected DAB converter is controlled with an enhanced control strategy. The controller is based on the artificial intelligence (AI) technique to regulate the terminal PV voltage through the phase shift angle of each DAB converter. In this manner, no need for a Maximum Power Point Tracking (MPPT) unit to set the reference of the PV terminal voltage. This strategy overcomes the stability issues of the DC microgrid as the response of converters is superior compared to the conventional strategies. The proposed PV interface system is modelled and simulated using MATLAB/SIMULINK. The simulation results reveal an accurate and fast response of the proposed design in case of irradiance changes.

Keywords: DC microgrid, DAB converter, parallel operation, artificial intelligence, fast response

Procedia PDF Downloads 773
2231 The Application of Video Segmentation Methods for the Purpose of Action Detection in Videos

Authors: Nassima Noufail, Sara Bouhali

Abstract:

In this work, we develop a semi-supervised solution for the purpose of action detection in videos and propose an efficient algorithm for video segmentation. The approach is divided into video segmentation, feature extraction, and classification. In the first part, a video is segmented into clips, and we used the K-means algorithm for this segmentation; our goal is to find groups based on similarity in the video. The application of k-means clustering into all the frames is time-consuming; therefore, we started by the identification of transition frames where the scene in the video changes significantly, and then we applied K-means clustering into these transition frames. We used two image filters, the gaussian filter and the Laplacian of Gaussian. Each filter extracts a set of features from the frames. The Gaussian filter blurs the image and omits the higher frequencies, and the Laplacian of gaussian detects regions of rapid intensity changes; we then used this vector of filter responses as an input to our k-means algorithm. The output is a set of cluster centers. Each video frame pixel is then mapped to the nearest cluster center and painted with a corresponding color to form a visual map. The resulting visual map had similar pixels grouped. We then computed a cluster score indicating how clusters are near each other and plotted a signal representing frame number vs. clustering score. Our hypothesis was that the evolution of the signal would not change if semantically related events were happening in the scene. We marked the breakpoints at which the root mean square level of the signal changes significantly, and each breakpoint is an indication of the beginning of a new video segment. In the second part, for each segment from part 1, we randomly selected a 16-frame clip, then we extracted spatiotemporal features using convolutional 3D network C3D for every 16 frames using a pre-trained model. The C3D final output is a 512-feature vector dimension; hence we used principal component analysis (PCA) for dimensionality reduction. The final part is the classification. The C3D feature vectors are used as input to a multi-class linear support vector machine (SVM) for the training model, and we used a multi-classifier to detect the action. We evaluated our experiment on the UCF101 dataset, which consists of 101 human action categories, and we achieved an accuracy that outperforms the state of art by 1.2%.

Keywords: video segmentation, action detection, classification, Kmeans, C3D

Procedia PDF Downloads 67
2230 Efficient Filtering of Graph Based Data Using Graph Partitioning

Authors: Nileshkumar Vaishnav, Aditya Tatu

Abstract:

An algebraic framework for processing graph signals axiomatically designates the graph adjacency matrix as the shift operator. In this setup, we often encounter a problem wherein we know the filtered output and the filter coefficients, and need to find out the input graph signal. Solution to this problem using direct approach requires O(N3) operations, where N is the number of vertices in graph. In this paper, we adapt the spectral graph partitioning method for partitioning of graphs and use it to reduce the computational cost of the filtering problem. We use the example of denoising of the temperature data to illustrate the efficacy of the approach.

Keywords: graph signal processing, graph partitioning, inverse filtering on graphs, algebraic signal processing

Procedia PDF Downloads 301
2229 The High Precision of Magnetic Detection with Microwave Modulation in Solid Spin Assembly of NV Centres in Diamond

Authors: Zongmin Ma, Shaowen Zhang, Yueping Fu, Jun Tang, Yunbo Shi, Jun Liu

Abstract:

Solid-state quantum sensors are attracting wide interest because of their high sensitivity at room temperature. In particular, spin properties of nitrogen–vacancy (NV) color centres in diamond make them outstanding sensors of magnetic fields, electric fields and temperature under ambient conditions. Much of the work on NV magnetic sensing has been done so as to achieve the smallest volume, high sensitivity of NV ensemble-based magnetometry using micro-cavity, light-trapping diamond waveguide (LTDW), nano-cantilevers combined with MEMS (Micro-Electronic-Mechanical System) techniques. Recently, frequency-modulated microwaves with continuous optical excitation method have been proposed to achieve high sensitivity of 6 μT/√Hz using individual NV centres at nanoscale. In this research, we built-up an experiment to measure static magnetic field through continuous wave optical excitation with frequency-modulated microwaves method under continuous illumination with green pump light at 532 nm, and bulk diamond sample with a high density of NV centers (1 ppm). The output of the confocal microscopy was collected by an objective (NA = 0.7) and detected by a high sensitivity photodetector. We design uniform and efficient excitation of the micro strip antenna, which is coupled well with the spin ensembles at 2.87 GHz for zero-field splitting of the NV centers. Output of the PD signal was sent to an LIA (Lock-In Amplifier) modulated signal, generated by the microwave source by IQ mixer. The detected signal is received by the photodetector, and the reference signal enters the lock-in amplifier to realize the open-loop detection of the NV atomic magnetometer. We can plot ODMR spectra under continuous-wave (CW) microwave. Due to the high sensitivity of the lock-in amplifier, the minimum detectable value of the voltage can be measured, and the minimum detectable frequency can be made by the minimum and slope of the voltage. The magnetic field sensitivity can be derived from η = δB√T corresponds to a 10 nT minimum detectable shift in the magnetic field. Further, frequency analysis of the noise in the system indicates that at 10Hz the sensitivity less than 10 nT/√Hz.

Keywords: nitrogen-vacancy (NV) centers, frequency-modulated microwaves, magnetic field sensitivity, noise density

Procedia PDF Downloads 429
2228 Assessments of Internal Erosion in a Landfill Due to Changes in the Groundwater Level

Authors: Siamak Feizi, Gunvor Baardvik

Abstract:

Soil erosion has special consequences for landfills that are more serious than those found at conventional construction sites. Different potential heads between two sides of a landfill and the subsequent movement of water through pores within the soil body could trigger the soil erosion and construction instability. Such a condition was encountered in a landfill project in the southern part of Norway. To check the risk of internal erosion due to changes in the groundwater level (because of seasonal flooding in the river), a series of numerical simulations by means of Geo-Seep software was conducted. Output of this study provides a total picture of the landfill stability, possibilities of erosions, and necessary measures to prevent or reduce the risk for the landfill operator.

Keywords: erosion, seepage, landfill, stability

Procedia PDF Downloads 128
2227 A Risk Management Approach for Nigeria Manufacturing Industries

Authors: Olaniyi O. Omoyajowo

Abstract:

To be successful in today’s competitive global environment, manufacturing industry must be able to respond quickly to changes in technology. These changes in technology introduce new risks and hazards. The management of risk/hazard in a manufacturing process recommends method through which the success rate of an organization can be increased. Thus, there is a continual need for manufacturing industries to invest significant amount of resources in risk management, which in turn optimizes the production output and profitability of any manufacturing industry (if implemented properly). To help improve the existing risk prevention and mitigation practices in Small and Medium Enterprise (SME) in Nigeria Manufacturing Industries (NMI), the researcher embarks on this research to develop a systematic Risk Management process.

Keywords: manufacturing management, risk, risk management, SMEs

Procedia PDF Downloads 389
2226 The Soviet Union-Style of Urban Planning in China: Historical Review and Enlightenment from the Output Mode of Contemporary Cooperative Parks

Authors: Yifeng Shi, Xingping Wang

Abstract:

The Soviet Union-style of urban planning has produced a broad and profound influence on China’s urban planning system. The study on extendibility and development experience of Soviet planning in China helps to change the current embarrassing situation 'one-hand planning practice, second-hand planning theory', and also beneficial to facilitate the establishment of China's domestic urban planning theory from the planning source, especially the overseas cooperation parks rich in 'Chinese characteristics'. In practice, as the world’s major infrastructure country, China is exporting to the world especially countries along 'the Belt and Road' a development model featuring cooperation parks as Chinese characteristics. This is of great significance to evaluate and summarize the experiences of Soviet Union-style of planning for China's development objectively and rationally, from removing ideological factors and extracting positive factors to carry them forward in overseas cooperation parks. This article briefly reviews the Soviet influence on urban planning after the founding of China and divided the influences stages into 'guidance, internalization and absorption, selective learning, decline' four periods. The impact includes production-oriented planning and planning concepts continue to be implemented, the establishment of the regional planning, master planning, detailed planning of the basic framework of urban planning, and homogenized cellular structure of the space, as well as planning techniques, professional training, planning techniques and so on. China and even most socialist countries now still carry such planning genes. At present, in the process of implementing 'the Belt and Road' strategy, the planning and construction of China’s overseas cooperation parks generally encounter many problems as lack of strategic planning and systematic planning, lack of top-level design, uncoordinated planning and layout in parks, and redundant construction in some areas. After sublating the planning genes of the Soviet Union-style of urban planning for the development of the socialist countries, especially the industrial planning system, this paper puts forward some views as follows to realize the overseas output and development of China's planning model and technology. Firstly the future development of overseas cooperation park should be from a rational planning point of view. Secondly the government should not only rigidly and equitably allocate the resources of the parks but also closely integrate the national economic plans or economic development strategies. Lastly management department should frame the threshold of development rationally, give full play to the pragmatic planning style in accordance with the local land system and planning system. It has an important guiding and reference role for the development of China's overseas cooperation park under the 'go global' strategy, after objectively evaluating the impact of the Soviet Union-style urban planning and absorbing the beneficial components on China. However, we should also recognize that the cooperation parks and the urban industrial system behind it are only part of urban development. More attention should be payed on the design of the local and the general rules of urban development to take the lead effect of cooperation parks suitable. Foundation item: Under the auspices of the Specific Plan for Strategic International Cooperation in Scientific and Technological Innovation, the National Key Research and Development Plan 'Research Cooperation and Exemplary Application in Planning of Development of Overseas Industrial Parks' (No 2016YFE0201000).

Keywords: China cooperative parks, history of urban planning, output mode, The Soviet Union

Procedia PDF Downloads 241
2225 A Comparative Study of Sampling-Based Uncertainty Propagation with First Order Error Analysis and Percentile-Based Optimization

Authors: M. Gulam Kibria, Shourav Ahmed, Kais Zaman

Abstract:

In system analysis, the information on the uncertain input variables cause uncertainty in the system responses. Different probabilistic approaches for uncertainty representation and propagation in such cases exist in the literature. Different uncertainty representation approaches result in different outputs. Some of the approaches might result in a better estimation of system response than the other approaches. The NASA Langley Multidisciplinary Uncertainty Quantification Challenge (MUQC) has posed challenges about uncertainty quantification. Subproblem A, the uncertainty characterization subproblem, of the challenge posed is addressed in this study. In this subproblem, the challenge is to gather knowledge about unknown model inputs which have inherent aleatory and epistemic uncertainties in them with responses (output) of the given computational model. We use two different methodologies to approach the problem. In the first methodology we use sampling-based uncertainty propagation with first order error analysis. In the other approach we place emphasis on the use of Percentile-Based Optimization (PBO). The NASA Langley MUQC’s subproblem A is developed in such a way that both aleatory and epistemic uncertainties need to be managed. The challenge problem classifies each uncertain parameter as belonging to one the following three types: (i) An aleatory uncertainty modeled as a random variable. It has a fixed functional form and known coefficients. This uncertainty cannot be reduced. (ii) An epistemic uncertainty modeled as a fixed but poorly known physical quantity that lies within a given interval. This uncertainty is reducible. (iii) A parameter might be aleatory but sufficient data might not be available to adequately model it as a single random variable. For example, the parameters of a normal variable, e.g., the mean and standard deviation, might not be precisely known but could be assumed to lie within some intervals. It results in a distributional p-box having the physical parameter with an aleatory uncertainty, but the parameters prescribing its mathematical model are subjected to epistemic uncertainties. Each of the parameters of the random variable is an unknown element of a known interval. This uncertainty is reducible. From the study, it is observed that due to practical limitations or computational expense, the sampling is not exhaustive in sampling-based methodology. That is why the sampling-based methodology has high probability of underestimating the output bounds. Therefore, an optimization-based strategy to convert uncertainty described by interval data into a probabilistic framework is necessary. This is achieved in this study by using PBO.

Keywords: aleatory uncertainty, epistemic uncertainty, first order error analysis, uncertainty quantification, percentile-based optimization

Procedia PDF Downloads 230
2224 Magnetic Fluctuations in the Terrestrial Magnetosheath

Authors: Alexandre Gurchumelia, Luca Sorriso-Valvo, David Burgess, Khatuna Elbakidze, Oleg Kharshiladze, Diana Kvaratskhelia

Abstract:

The terrestrial magnetosheath is a highly turbulent medium, with a high level of magnetic1field fluctuations throughout a broad range of scales. These often include an inertial range where a2magnetohydrodynamic turbulent cascade is observed. The multifractal properties of the turbulent3cascade, strictly related to intermittency, are observed here during the transition from quasi-parallel to4quasi-perpendicular magnetic field with respect to the bow-shock normal. The different multifractal5behavior in the two regions is analyzed. A standard coarse-graining technique has been used6to evaluate the generalized dimensions and the corresponding multifractal spectrumf(α). A7p-model fit provided a quantitative measure of multifractality and intermittency, to be compared with8standard indicators: the width of the multifractal spectrum, the peak of the kurtosis, and its scaling9exponent. Results show a clear transition and sharp differences in the intermittency properties for the two regions.

Keywords: magnetos heath, turbulence, multifractal, instabilities

Procedia PDF Downloads 170
2223 Parallel among Urinary Tract Infection in Diabetic and Non-Diabetic Patients: A Case Study

Authors: Khaled Khleifat

Abstract:

This study detects the bacterial species that responsible for UTI in both diabetic patients and non-diabetic patients, Jordan. 116 urine samples were investigated in order to determine UTI-causing bacteria. These samples distributed unequally between diabetic male (12) and diabetic female (25) and also non-diabetic male (13) and non-diabetic female (66). The results represent that E.coli is responsible for UTI in both diabetic and non-diabetic patients (15.5% and 29.3% respectively) with large proportion (44.8%). This study showed that not all bacterial species that isolated from the non-diabetic sample could be isolated from diabetic samples. E. coli (15.5%), P. aeruginosa (4.3%), K. pneumonia (1.7%), P. mirabilis (2.6%), S. marcescens (0.9%), S. aureus (1.7%), S. pyogenes (1.7%), E. faecalis (0.9%), S. epidermidis (1.7%) and S. saprophyticus (0.9%). But E. aerogenes, E. cloacae, C. freundii, A. baumannii and B. subtilis are five bacterial species that can’t isolate from all diabetic samples. This study shows that for the treatment of UTI in both diabetic and non-diabetic patients, Chloramphenicol (30 μg), Ciprofloxacin (5 μg) and Vancomycin (30 μg) are more favorable than other antibiotics. In the same time, Cephalothin (30μg) is not recommended.

Keywords: urinary tract infections, diabetes mellitus, bacterial species, infections

Procedia PDF Downloads 318
2222 Motion Estimator Architecture with Optimized Number of Processing Elements for High Efficiency Video Coding

Authors: Seongsoo Lee

Abstract:

Motion estimation occupies the heaviest computation in HEVC (high efficiency video coding). Many fast algorithms such as TZS (test zone search) have been proposed to reduce the computation. Still the huge computation of the motion estimation is a critical issue in the implementation of HEVC video codec. In this paper, motion estimator architecture with optimized number of PEs (processing element) is presented by exploiting early termination. It also reduces hardware size by exploiting parallel processing. The presented motion estimator architecture has 8 PEs, and it can efficiently perform TZS with very high utilization of PEs.

Keywords: motion estimation, test zone search, high efficiency video coding, processing element, optimization

Procedia PDF Downloads 353
2221 FLC with 3DSVM for 4LEG 4WIRE Shunt Active Power Filter

Authors: Abdelhalim Kessal, Ali Chebabhi

Abstract:

In this paper, a controller based on fuzzy logic control (FLC) associated to Three Dimensional Space Vector Modulation (3DSVM) is applied for shunt active filter in αβo axes domain. The main goals are to improve power quality under disturbed loads, minimize source currents harmonics and reduce neutral current magnitude in the four-wire structure. FLC is used to obtain the reference current and control the DC-bus voltage at the inverter output. The switching signals of the four-leg inverter are generating through a Three Dimensional Space Vector Modulation (3DSVM). Selected simulation results have been shown to validate the proposed system.

Keywords: flc, 3dsvm, sapf, harmonic, inverter

Procedia PDF Downloads 488
2220 Spectroscopic Autoradiography of Alpha Particles on Geologic Samples at the Thin Section Scale Using a Parallel Ionization Multiplier Gaseous Detector

Authors: Hugo Lefeuvre, Jerôme Donnard, Michael Descostes, Sophie Billon, Samuel Duval, Tugdual Oger, Herve Toubon, Paul Sardini

Abstract:

Spectroscopic autoradiography is a method of interest for geological sample analysis. Indeed, researchers may face different issues such as radioelement identification and quantification in the field of environmental studies. Imaging gaseous ionization detectors find their place in geosciences for conducting specific measurements of radioactivity to improve the monitoring of natural processes using naturally-occurring radioactive tracers, but also for the nuclear industry linked to the mining sector. In geological samples, the location and identification of the radioactive-bearing minerals at the thin-section scale remains a major challenge as the detection limit of the usual elementary microprobe techniques is far higher than the concentration of most of the natural radioactive decay products. The spatial distribution of each decay product in the case of uranium in a geomaterial is interesting for relating radionuclides concentration to the mineralogy. The present study aims to provide spectroscopic autoradiography analysis method for measuring the initial energy of alpha particles with a parallel ionization multiplier gaseous detector. The analysis method has been developed thanks to Geant4 modelling of the detector. The track of alpha particles recorded in the gas detector allow the simultaneous measurement of the initial point of emission and the reconstruction of the initial particle energy by a selection based on the linear energy distribution. This spectroscopic autoradiography method was successfully used to reproduce the alpha spectra from a 238U decay chain on a geological sample at the thin-section scale. The characteristics of this measurement are an energy spectrum resolution of 17.2% (FWHM) at 4647 keV and a spatial resolution of at least 50 µm. Even if the efficiency of energy spectrum reconstruction is low (4.4%) compared to the efficiency of a simple autoradiograph (50%), this novel measurement approach offers the opportunity to select areas on an autoradiograph to perform an energy spectrum analysis within that area. This opens up possibilities for the detailed analysis of heterogeneous geological samples containing natural alpha emitters such as uranium-238 and radium-226. This measurement will allow the study of the spatial distribution of uranium and its descendants in geo-materials by coupling scanning electron microscope characterizations. The direct application of this dual modality (energy-position) of analysis will be the subject of future developments. The measurement of the radioactive equilibrium state of heterogeneous geological structures, and the quantitative mapping of 226Ra radioactivity are now being actively studied.

Keywords: alpha spectroscopy, digital autoradiography, mining activities, natural decay products

Procedia PDF Downloads 141
2219 Measuring the Effect of the Privatization of the Kuwait Stock Exchange on Its Performance

Authors: Mohamad H. Atyeh, Wael Alrashed, Steven Telford

Abstract:

The main objective of this research is to measure if there have been any notable changes in the trading actives of the Kuwait stock Exchange (KSE) after the privatization process that took place on the 25th of April 2016. The data that are used to test if there is any change in the KSE market performance are the daily indices for the period from the 25th of April 2016 till the 24th of October 2016 (after privatization) and a similar six months period before the date of the privatization from the 24th of October 2015 till the 24th of April 2016. In addition, as a control, the study included a period that is a period parallel to the six months period after the privatization. The results indicate that privatization is associated with lower variability for the majority of variables, but that the observed switch in slope direction is not actually a product of privatization, but rather one of serial correlation.

Keywords: privatization, Kuwait stock exchange (KSE), market capitalization (MCAP), capital markets authority (CMA), Boursa Kuwait securities company (BKSC)

Procedia PDF Downloads 289
2218 A Method of Improving Out Put Using a Feedback Supply Chain System: Case Study Bramlima

Authors: Samuel Atongaba Danji, Veseke Moleke

Abstract:

The increase of globalization is a very important part of today’s changing environment and due to this, manufacturing industries have to always come up with methods of continuous improvement of their manufacturing methods in order to be competitive, without which may lead them to be left out of the market due to constant changing customers requirement. Due to this, the need is an advance supply chain system which prevents a number of issues that can prevent a company from being competitive. In this work, we developed a feedback control supply chain system which streamline the entire process in order to improve competitiveness and the result shows that when applied in a different geographical area, the output varies.

Keywords: globalization, supply chain, improvement, manufacturing

Procedia PDF Downloads 321
2217 Current Starved Ring Oscillator Image Sensor

Authors: Devin Atkin, Orly Yadid-Pecht

Abstract:

The continual demands for increasing resolution and dynamic range in CMOS image sensors have resulted in exponential increases in the amount of data that needs to be read out of an image sensor, and existing readouts cannot keep up with this demand. Interesting approaches such as sparse and burst readouts have been proposed and show promise, but at considerable trade-offs in other specifications. To this end, we have begun designing and evaluating various new readout topologies centered around an attempt to parallelize the sensor readout. In this paper, we have designed, simulated, and started testing a new light-controlled oscillator topology with dual column and row readouts. We expect the parallel readout structure to offer greater speed and alleviate the trade-off typical in this topology, where slow pixels present a major framerate bottleneck.

Keywords: CMOS image sensors, high-speed capture, wide dynamic range, light controlled oscillator

Procedia PDF Downloads 76