Search results for: Data structures
4831 Nonlinear Analysis of Postural Sway in Multiple Sclerosis
Authors: Hua Cao, Laurent Peyrodie, Olivier Agnani, Cécile Donzé
Abstract:
Multiple Sclerosis (MS) is a disease which affects the central nervous system and causes balance problem. In clinical, this disorder is usually evaluated using static posturography. Some linear or nonlinear measures, extracted from the posturographic data (i.e. center of pressure, COP) recorded during a balance test, has been used to analyze postural control of MS patients. In this study, the trend (TREND) and the sample entropy (SampEn), two nonlinear parameters were chosen to investigate their relationships with the expanded disability status scale (EDSS) score. 40 volunteers with different EDSS scores participated in our experiments with eyes open (EO) and closed (EC). TREND and 2 types of SampEn (SampEn1 and SampEn2) were calculated for each combined COP’s position signal. The results have shown that TREND had a weak negative correlation to EDSS while SampEn2 had a strong positive correlation to EDSS. Compared to TREND and SampEn1, SampEn2 showed a better significant correlation to EDSS and an ability to discriminate the MS patients in the EC case. In addition, the outcome of the study suggests that the multi-dimensional nonlinear analysis could provide some information about the impact of disability progression in MS on dynamics of the COP data.Keywords: Balance, multiple sclerosis, nonlinear analysis, postural sway.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19714830 Applying Multiple Kinect on the Development of a Rapid 3D Mannequin Scan Platform
Authors: Shih-Wen Hsiao, Yi-Cheng Tsao
Abstract:
In the field of reverse engineering and creative industries, applying 3D scanning process to obtain geometric forms of the objects is a mature and common technique. For instance, organic objects such as faces and nonorganic objects such as products could be scanned to acquire the geometric information for further application. However, although the data resolution of 3D scanning device is increasing and there are more and more abundant complementary applications, the penetration rate of 3D scanning for the public is still limited by the relative high price of the devices. On the other hand, Kinect, released by Microsoft, is known for its powerful functions, considerably low price, and complete technology and database support. Therefore, related studies can be done with the applying of Kinect under acceptable cost and data precision. Due to the fact that Kinect utilizes optical mechanism to extracting depth information, limitations are found due to the reason of the straight path of the light. Thus, various angles are required sequentially to obtain the complete 3D information of the object when applying a single Kinect for 3D scanning. The integration process which combines the 3D data from different angles by certain algorithms is also required. This sequential scanning process costs much time and the complex integration process often encounter some technical problems. Therefore, this paper aimed to apply multiple Kinects simultaneously on the field of developing a rapid 3D mannequin scan platform and proposed suggestions on the number and angles of Kinects. In the content, a method of establishing the coordination based on the relation between mannequin and the specifications of Kinect is proposed, and a suggestion of angles and number of Kinects is also described. An experiment of applying multiple Kinect on the scanning of 3D mannequin is constructed by Microsoft API, and the results show that the time required for scanning and technical threshold can be reduced in the industries of fashion and garment design.
Keywords: 3D scan, depth sensor, fashion and garment design, mannequin, multiple kinect sensor.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22764829 A Multigrid Approach for Three-Dimensional Inverse Heat Conduction Problems
Authors: Jianhua Zhou, Yuwen Zhang
Abstract:
A two-step multigrid approach is proposed to solve the inverse heat conduction problem in a 3-D object under laser irradiation. In the first step, the location of the laser center is estimated using a coarse and uniform grid system. In the second step, the front-surface temperature is recovered in good accuracy using a multiple grid system in which fine mesh is used at laser spot center to capture the drastic temperature rise in this region but coarse mesh is employed in the peripheral region to reduce the total number of sensors required. The effectiveness of the two-step approach and the multiple grid system are demonstrated by the illustrative inverse solutions. If the measurement data for the temperature and heat flux on the back surface do not contain random error, the proposed multigrid approach can yield more accurate inverse solutions. When the back-surface measurement data contain random noise, accurate inverse solutions cannot be obtained if both temperature and heat flux are measured on the back surface.
Keywords: Conduction, inverse problems, conjugated gradient method, laser.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8444828 Variability of Metal Composition and Concentrations in Road Dust in the Urban Environment
Authors: Sandya Mummullage, Prasanna Egodawatta, Ashantha Goonetilleke, Godwin A. Ayoko
Abstract:
Urban road dust comprises of a range of potentially toxic metal elements and plays a critical role in degrading urban receiving water quality. Hence, assessing the metal composition and concentration in urban road dust is a high priority. This study investigated the variability of metal composition and concentrations in road dust in 4 different urban land uses in Gold Coast, Australia. Samples from 16 road sites were collected and tested for selected 12 metal species. The data set was analyzed using both univariate and multivariate techniques. Outcomes of the data analysis revealed that the metal concentrations inroad dust differs considerably within and between different land uses. Iron, aluminum, magnesium and zinc are the most abundant in urban land uses. It was also noted that metal species such as titanium, nickel, copper and zinc have the highest concentrations in industrial land use. The study outcomes revealed that soil and traffic related sources as key sources of metals deposited on road surfaces.
Keywords: Metals build-up, Pollutant accumulation, Stormwater quality, Urban road dust.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 23504827 Cooperative Energy Efficient Routing for Wireless Sensor Networks in Smart Grid Communications
Authors: Ghazi AL-Sukkar, Iyad Jafar, Khalid Darabkh, Raed Al-Zubi, Mohammed Hawa
Abstract:
Smart Grids employ wireless sensor networks for their control and monitoring. Sensors are characterized by limitations in the processing power, energy supply and memory spaces, which require a particular attention on the design of routing and data management algorithms. Since most routing algorithms for sensor networks, focus on finding energy efficient paths to prolong the lifetime of sensor networks, the power of sensors on efficient paths depletes quickly, and consequently sensor networks become incapable of monitoring events from some parts of their target areas. In consequence, the design of routing protocols should consider not only energy efficiency paths, but also energy efficient algorithms in general. In this paper we propose an energy efficient routing protocol for wireless sensor networks without the support of any location information system. The reliability and the efficiency of this protocol have been demonstrated by simulation studies where we compare them to the legacy protocols. Our simulation results show that these algorithms scale well with network size and density.Keywords: Data-centric storage, Dynamic Address Allocation, Sensor networks, Smart Grid Communications.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18524826 Investigation of Hydraulic and Thermal Performances of Fin Array at Different Shield Positions without By-Pass
Authors: Ramy H. Mohammed
Abstract:
In heat sinks, the flow within the core exhibits separation and hence does not lend itself to simple analytical boundary layer or duct flow analysis of the wall friction. In this paper, we present some findings from an experimental and numerical study aimed to obtain physical insight into the influence of the presence of the shield and its position on the hydraulic and thermal performance of square pin fin heat sink without top by-pass. The variations of the Nusselt number and friction factor are obtained under varied parameters, such as the Reynolds number and the shield position. The numerical code is validated by comparing the numerical results with the available experimental data. It is shown that, there is a good agreement between the temperature predictions based on the model and the experimental data. Results show that, as the presence of the shield, the heat transfer of fin array is enhanced and the flow resistance increased. The surface temperature distribution of the heat sink base is more uniform when the dimensionless shield position equals to 1/3 or 2/3. The comprehensive performance evaluation approach based on identical pumping power criteria is adopted and shows that the optimum shield position is at x/l=0.43.
Keywords: Shield, Fin array, Performance evaluation, Heat transfer, Validation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18384825 Fast Adjustable Threshold for Uniform Neural Network Quantization
Authors: Alexander Goncharenko, Andrey Denisov, Sergey Alyamkin, Evgeny Terentev
Abstract:
The neural network quantization is highly desired procedure to perform before running neural networks on mobile devices. Quantization without fine-tuning leads to accuracy drop of the model, whereas commonly used training with quantization is done on the full set of the labeled data and therefore is both time- and resource-consuming. Real life applications require simplification and acceleration of quantization procedure that will maintain accuracy of full-precision neural network, especially for modern mobile neural network architectures like Mobilenet-v1, MobileNet-v2 and MNAS. Here we present a method to significantly optimize training with quantization procedure by introducing the trained scale factors for discretization thresholds that are separate for each filter. Using the proposed technique, we quantize the modern mobile architectures of neural networks with the set of train data of only ∼ 10% of the total ImageNet 2012 sample. Such reduction of train dataset size and small number of trainable parameters allow to fine-tune the network for several hours while maintaining the high accuracy of quantized model (accuracy drop was less than 0.5%). Ready-for-use models and code are available in the GitHub repository.Keywords: Distillation, machine learning, neural networks, quantization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7324824 Person-Environment Fit (PE Fit): Evidence from Brazil
Authors: Jucelia Appio, Danielle Deimling De Carli, Bruno Henrique Rocha Fernandes, Nelson Natalino Frizon
Abstract:
The purpose of this paper is to investigate if there are positive and significant correlations between the dimensions of Person-Environment Fit (Person-Job, Person-Organization, Person-Group and Person-Supervisor) at the “Best Companies to Work for” in Brazil in 2017. For that, a quantitative approach was used with a descriptive method being defined as a research sample the "150 Best Companies to Work for", according to data base collected in 2017 and provided by Fundação Instituto of Administração (FIA) of the University of São Paulo (USP). About the data analysis procedures, asymmetry and kurtosis, factorial analysis, Kaiser-Meyer-Olkin (KMO) tests, Bartlett sphericity and Cronbach's alpha were used for the 69 research variables, and as a statistical technique for the purpose of analyzing the hypothesis, Pearson's correlation analysis was performed. As a main result, we highlight that there was a positive and significant correlation between the dimensions of Person-Environment Fit, corroborating the H1 hypothesis that there is a positive and significant correlation between Person-Job Fit, Person-Organization Fit, Person-Group Fit and Person-Supervisor Fit.
Keywords: Human resource management, person-environment fit, strategic people management, best companies to work for.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10004823 Investigating the Vehicle-Bicyclists Conflicts Using LIDAR Sensor Technology at Signalized Intersections
Authors: Alireza Ansariyar, Mansoureh Jeihani
Abstract:
Light Detection and Ranging (LiDAR) sensors is capable of recording traffic data including the number of passing vehicles and bicyclists, the speed of vehicles and bicyclists, and the number of conflicts among both road users. In order to collect real-time traffic data and investigate the safety of different road users, a LiDAR sensor was installed at Cold Spring Ln – Hillen Rd intersection in Baltimore city. The frequency and severity of collected real-time conflicts were analyzed and the results highlighted that 122 conflicts were recorded over a 10-month time interval from May 2022 to February 2023. By employing an image-processing algorithm, a safety Measure of Effectiveness (MOE) aims to identify critical zones for bicyclists upon entering each respective zone at the signalized intersection. Considering the trajectory of conflicts, the results of analysis demonstrated that conflicts in the northern approach (zone N) are more frequent and severe. Additionally, sunny weather is more likely to cause severe vehicle-bike conflicts.
Keywords: LiDAR sensor, Post Encroachment Time threshold, vehicle-bike conflicts, measure of effectiveness, weather condition.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 1434822 Application of Geographic Information Systems(GIS) in the History of Cartography
Authors: Bangbo Hu
Abstract:
This paper discusses applications of a revolutionary information technology, Geographic Information Systems (GIS), in the field of the history of cartography by examples, including assessing accuracy of early maps, establishing a database of places and historical administrative units in history, integrating early maps in GIS or digital images, and analyzing social, political, and economic information related to production of early maps. GIS provides a new mean to evaluate the accuracy of early maps. Four basic steps using GIS for this type of study are discussed. In addition, several historical geographical information systems are introduced. These include China Historical Geographic Information Systems (CHGIS), the United States National Historical Geographic Information System (NHGIS), and the Great Britain Historical Geographical Information System. GIS also provides digital means to display and analyze the spatial information on the early maps or to layer them with modern spatial data. How GIS relational data structure may be used to analyze social, political, and economic information related to production of early maps is also discussed in this paper. Through discussion on these examples, this paper reveals value of GIS applications in this field.Keywords: Cartography, GIS, history, maps.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 31564821 New Features for Specific JPEG Steganalysis
Authors: Johann Barbier, Eric Filiol, Kichenakoumar Mayoura
Abstract:
We present in this paper a new approach for specific JPEG steganalysis and propose studying statistics of the compressed DCT coefficients. Traditionally, steganographic algorithms try to preserve statistics of the DCT and of the spatial domain, but they cannot preserve both and also control the alteration of the compressed data. We have noticed a deviation of the entropy of the compressed data after a first embedding. This deviation is greater when the image is a cover medium than when the image is a stego image. To observe this deviation, we pointed out new statistic features and combined them with the Multiple Embedding Method. This approach is motivated by the Avalanche Criterion of the JPEG lossless compression step. This criterion makes possible the design of detectors whose detection rates are independent of the payload. Finally, we designed a Fisher discriminant based classifier for well known steganographic algorithms, Outguess, F5 and Hide and Seek. The experiemental results we obtained show the efficiency of our classifier for these algorithms. Moreover, it is also designed to work with low embedding rates (< 10-5) and according to the avalanche criterion of RLE and Huffman compression step, its efficiency is independent of the quantity of hidden information.
Keywords: Compressed frequency domain, Fisher discriminant, specific JPEG steganalysis.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21624820 Comparison of Stochastic Point Process Models of Rainfall in Singapore
Abstract:
Extensive rainfall disaggregation approaches have been developed and applied in climate change impact studies such as flood risk assessment and urban storm water management.In this study, five rainfall models that were capable ofdisaggregating daily rainfall data into hourly one were investigated for the rainfall record in theChangi Airport, Singapore. The objectives of this study were (i) to study the temporal characteristics of hourly rainfall in Singapore, and (ii) to evaluate the performance of variousdisaggregation models. The used models included: (i) Rectangular pulse Poisson model (RPPM), (ii) Bartlett-Lewis Rectangular pulse model (BLRPM), (iii) Bartlett-Lewis model with 2 cell types (BL2C), (iv) Bartlett-Lewis Rectangular with cell depth distribution dependent on duration (BLRD), and (v) Neyman-Scott Rectangular pulse model (NSRPM). All of these models werefitted using hourly rainfall data ranging from 1980 to 2005 (which was obtained from Changimeteorological station).The study results indicated that the weight scheme of inversely proportional variance could deliver more accurateoutputs for fitting rainfall patterns in tropical areas, and BLRPM performedrelatively better than other disaggregation models.
Keywords: Rainfall disaggregation, statistical properties, poisson processed, Bartlett-Lewis model, Neyman-Scott model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22814819 Surveying Earthquake Vulnerabilities of District 13 of Kabul City, Afghanistan
Authors: Mohsen Mohammadi, Toshio Fujimi
Abstract:
High population and irregular urban development in Kabul city, Afghanistan's capital, are among factors that increase its vulnerability to earthquake disasters (on top of its location in a high seismic region); this can lead to widespread economic loss and casualties. This study aims to evaluate earthquake risks in Kabul's 13th district based on scientific data. The research data, which include hazard curves of Kabul, vulnerability curves, and a questionnaire survey through sampling in district 13, have been incorporated to develop risk curves. To estimate potential casualties, we used a set of M parameters in a model developed by Coburn and Spence. The results indicate that in the worst case scenario, more than 90% of district 13, which comprises mostly residential buildings, is exposed to high risk; this may lead to nearly 1000 million USD economic loss and 120 thousand casualties (equal to 25.88% of the 13th district's population) for a nighttime earthquake. To reduce risks, we present the reconstruction of the most vulnerable buildings, which are primarily adobe and masonry buildings. A comparison of risk reduction between reconstructing adobe and masonry buildings indicates that rebuilding adobe buildings would be more effective.
Keywords: Earthquake risk evaluation, Kabul, mitigation, vulnerability.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16344818 Biosynthesis and Metabolism of Anthraquinone Derivatives
Authors: Dmitry Yu. Korulkin, Raissa A. Muzychkina
Abstract:
In review the generalized data about biosynthetic routs formation anthraquinone molecules in natural cells. The basic possibilities of various ways of biosynthesis of different quinoid substances are shown.
Keywords: Anthraquinones, biochemical evolution, biosynthesis, metabolism.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 39664817 In vitro Cytotoxic and Genotoxic Effects of Arsenic Trioxide on Human Keratinocytes
Authors: H. Bouaziz, M. Sefi, J. de Lapuente, M. Borras, N. Zeghal
Abstract:
Although, arsenic trioxide has been the subject of toxicological research, in vitro cytotoxicity and genotoxicity studies using relevant cell models and uniform methodology are not well elucidated. Hence, the aim of the present study was to evaluate the cytotoxicity and genotoxicity induced by arsenic trioxide in human keratinocytes (HaCaT) using the MTT [3-(4, 5-dimethylthiazol-2-yl)- 2,5-diphenyltetrazolium bromide] and alkaline single cell gel electrophoresis (Comet) assays, respectively. Human keratinocytes were treated with different doses of arsenic trioxide for 4 h prior to cytogenetic assessment. Data obtained from the MTT assay indicated that arsenic trioxide significantly reduced the viability of HaCaT cells in a dose-dependent manner, showing an IC50 value of 34.18 ± 0.6 μM. Data generated from the comet assay also indicated a significant dose-dependent increase in DNA damage in HaCaT cells associated with arsenic trioxide exposure. We observed a significant increase in comet tail length and tail moment, showing an evidence of arsenic trioxide -induced genotoxic damage in HaCaT cells. This study confirms that the comet assay is a sensitive and effective method to detect DNA damage caused by arsenic.
Keywords: Arsenic trioxide, cytotoxixity, genotoxicity, HaCaT.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22284816 Characterization and Optimization of Culture Conditions for Sulphur Oxidizing Bacteria after Isolation from Rhizospheric Mustard Soil, Decomposing Sites and Pit House
Authors: Suman Chaudhary, Rinku Dhanker, Tanvi, Sneh Goyal
Abstract:
Sulphur oxidizing bacteria (SOB) have marked their significant role in perspectives of maintaining healthy environment as researchers from all over the world tested and apply these in waste water treatment plants, bioleaching of heavy metals, deterioration of bridge structures, concrete and for bioremediation purposes, etc. Also, these SOB are well adapted in all kinds of environment ranging from normal soil, water habitats to extreme natural sources like geothermal areas, volcanic eruptions, black shale and acid rock drainage (ARD). SOB have been isolated from low pH environment of anthropogenic origin like acid mine drainage (AMD) and bioleaching heaps, hence these can work efficiently in different environmental conditions. Besides having many applications in field of environment science, they may be proven to be very beneficial in area of agriculture as sulphur is the fourth major macronutrients required for the growth of plants. More amount of sulphur is needed by pulses and oilseed crops with respect to the cereal grains. Due to continuous use of land for overproduction of more demanding sulphur utilizing crops and without application of sulphur fertilizers, its concentration is decreasing day by day, and thus, sulphur deficiency is becoming a great problem as it affects the crop productivity and quality. Sulphur is generally found in soils in many forms which are unavailable for plants (cannot be use by plants) like elemental sulphur, thiosulphate which can be taken up by bacteria and converted into simpler forms usable by plants by undergoing a series of transformations. So, keeping the importance of sulphur in view for various soil types, oilseed crops and role of microorganisms in making them available to plants, we made an effort to isolate, optimize, and characterize SOB. Three potential strains of bacteria were isolated, namely SSF7, SSA21, and SSS6, showing sulphate production of concentration, i.e. 2.268, 3.102, and 2.785 mM, respectively. Also, these were optimized for various culture conditions like carbon, nitrogen source, pH, temperature, and incubation time, and characterization was also done.
Keywords: Sulphur oxidizing bacteria, isolation, optimization, characterization, sulphate production.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11834815 Designing Information Systems in Education as Prerequisite for Successful Management Results
Authors: Vladimir Simovic, Matija Varga, Tonco Marusic
Abstract:
This research paper shows matrix technology models and examples of information systems in education (in the Republic of Croatia and in the Germany) in support of business, education (when learning and teaching) and e-learning. Here we researched and described the aims and objectives of the main process in education and technology, with main matrix classes of data. In this paper, we have example of matrix technology with detailed description of processes related to specific data classes in the processes of education and an example module that is support for the process: ‘Filling in the directory and the diary of work’ and ‘evaluation’. Also, on the lower level of the processes, we researched and described all activities which take place within the lower process in education. We researched and described the characteristics and functioning of modules: ‘Fill the directory and the diary of work’ and ‘evaluation’. For the analysis of the affinity between the aforementioned processes and/or sub-process we used our application model created in Visual Basic, which was based on the algorithm for analyzing the affinity between the observed processes and/or sub-processes.Keywords: Designing, education management, information systems, matrix technology, process affinity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 10964814 STEP Implementation on Turn-mill Manufacturing Environment
Authors: Ahmad Majdi Bin Abdul-Rani, Mesfin Gizaw, Yusri Yusof
Abstract:
Researches related to standard product model and development of neutral manufacturing interfaces for numerical control machines becomes a significant topic since the last 25 years. In this paper, a detail description of STEP implementation on turnmill manufacturing has been discussed. It shows requirements of information contents from ISO14649 data model. It covers to describe the design of STEP-NC framework applicable to turn-mill manufacturing. In the framework, EXPRESS-G and UML modeling tools are used to depict the information contents of the system and established the bases of information model requirement. A product and manufacturing data model applicable for STEP compliant manufacturing. The next generation turn-mill operations requirements have been represented by a UML diagram. An object oriented classes of ISO1449 has been developed on Visual Basic dot NET platform for binding the static information model represented by the UML diagram. An architect of the proposed system implementation has been given on the bases of the design and manufacturing module of STEP-NC interface established. Finally, a part 21 file process plan generated for an illustration of turn-mill components.Keywords: CAPP, ISO14649, Product modeling, STEP-NC
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 16624813 Experimental and Numerical Study of Ultra-High-Performance Fiber-Reinforced Concrete Column Subjected to Axial and Eccentric Loads
Authors: Chengfeng Fang, Mohamed Ali Sadakkathulla, Abdul Sheikh
Abstract:
Ultra-high-performance fiber reinforced concrete (UHPFRC) is a specially formulated cement-based composite characterized with an ultra-high compressive strength (fc’ = 240 MPa) and a low water-cement ratio (W/B= 0.2). With such material characteristics, UHPFRC is favored for the design and constructions of structures required high structural performance and slender geometries. Unlike conventional concrete, the structural performance of members manufactured with UHPFRC has not yet been fully studied, particularly, for UHPFRC columns with high slenderness. In this study, the behaviors of slender UHPFRC columns under concentric or eccentric load will be investigated both experimentally and numerically. Four slender UHPFRC columns were tested under eccentric loads with eccentricities, of 0 mm, 35 mm, 50 mm, and 85 mm, respectively, and one UHPFRC beam was tested under four-point bending. Finite element (FE) analysis was conducted with concrete damage plasticity (CDP) modulus to simulating the load-middle height or middle span deflection relationships and damage patterns of all UHPFRC members. Simulated results were compared against the experimental results and observation to gain the confidence of FE model, and this model was further extended to conduct parametric studies, which aim to investigate the effects of slenderness regarding failure modes and load-moment interaction relationships. Experimental results showed that the load bearing capacities of the slender columns reduced with an increase in eccentricity. Comparisons between load-middle height and middle span deflection relationships as well as damage patterns of all UHPFRC members obtained both experimentally and numerically demonstrated high accuracy of the FE simulations. Based on the available FE model, the following parametric study indicated that a further increase in the slenderness of column resulted in significant decreases in the load-bearing capacities, ductility index, and flexural bending capacities.
Keywords: Eccentric loads, ductility index, RC column, slenderness, UHPFRC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9744812 Retarding Potential Analyzer Design and Result Analysis for Ion Energy Distribution Measurement of the Thruster Plume in the Laboratory
Authors: Ma Ya-li, Tang Fu-jun, Xue Yu-xiong, Chen Yi-feng, Gao Xin, Wang Yi, Tian Kai, Yan Ze-dong
Abstract:
Plasma plume will be produced and arrive at spacecraft when the electric thruster operates on orbit. It-s important to characterize the thruster plasma parameters because the plume has significant effects or hazards on spacecraft sub-systems and parts. Through the ground test data of the desired parameters, the major characteristics of the thruster plume will be achieved. Also it is very important for optimizing design of Ion thruster. Retarding Potential Analyzer (RPA) is an effective instrument for plasma ion energy per unit charge distribution measurement. Special RPA should be designed according to certain plume plasma parameters range and feature. In this paper, major principles usable for good RPA design are discussed carefully. Conform to these principles, a four-grid planar electrostatic energy analyzer RPA was designed to avoid false data, and details were discussed including construction, materials, aperture diameter and so on. At the same time, it was designed more suitable for credible and long-duration measurements in the laboratory. In the end, RPA measurement results in the laboratory were given and discussed.
Keywords: Thruster plume ion energy distributions, retarding potential analyzer, ground test.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 40104811 The Relationship between Competency-Based Learning and Learning Efficiency of Media Communication Students at Suan Sunandha Rajabhat University
Authors: Somtop Keawchuer
Abstract:
This research aims to study (1) the relationship between competency-based learning and learning efficiency of new media communication students at Suan Sunandha University (2) the demographic factor effect on learning efficiency of students at Suan Sunandha University. This research method will use quantitative research; data was collected by questionnaires distributed to students from new media communication in management science faculty of Suan Sunandha Rajabhat University for 1340 sample by purposive sampling method. Data was analyzed by descriptive statistic including percentage, mean, standard deviation and inferential statistic including T-test, ANOVA and Pearson correlation for hypothesis testing. The results showed that the competency-based learning in term of ability to communicate, ability to think and solve the problem, life skills and ability to use technology has a significant relationship with learning efficiency in term of the cognitive domain, psychomotor domain and affective domain at the 0.05 level and which is in harmony with the research hypotheses.
Keywords: Competency-based learning, learning efficiency, new media communication students, Suan Sunandha Rajabhat University.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11794810 Multi-Agent System for Irrigation Using Fuzzy Logic Algorithm and Open Platform Communication Data Access
Authors: T. Wanyama, B. Far
Abstract:
Automatic irrigation systems usually conveniently protect landscape investment. While conventional irrigation systems are known to be inefficient, automated ones have the potential to optimize water usage. In fact, there is a new generation of irrigation systems that are smart in the sense that they monitor the weather, soil conditions, evaporation and plant water use, and automatically adjust the irrigation schedule. In this paper, we present an agent based smart irrigation system. The agents are built using a mix of commercial off the shelf software, including MATLAB, Microsoft Excel and KEPServer Ex5 OPC server, and custom written code. The Irrigation Scheduler Agent uses fuzzy logic to integrate the information that affect the irrigation schedule. In addition, the Multi-Agent system uses Open Platform Connectivity (OPC) technology to share data. OPC technology enables the Irrigation Scheduler Agent to communicate over the Internet, making the system scalable to a municipal or regional agent based water monitoring, management, and optimization system. Finally, this paper presents simulation and pilot installation test result that show the operational effectiveness of our system.
Keywords: Community water usage, fuzzy logic, irrigation, multi-agent system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13384809 On Adaptive Optimization of Filter Performance Based on Markov Representation for Output Prediction Error
Authors: Hong Son Hoang, Remy Baraille
Abstract:
This paper addresses the problem of how one can improve the performance of a non-optimal filter. First the theoretical question on dynamical representation for a given time correlated random process is studied. It will be demonstrated that for a wide class of random processes, having a canonical form, there exists a dynamical system equivalent in the sense that its output has the same covariance function. It is shown that the dynamical approach is more effective for simulating and estimating a Markov and non- Markovian random processes, computationally is less demanding, especially with increasing of the dimension of simulated processes. Numerical examples and estimation problems in low dimensional systems are given to illustrate the advantages of the approach. A very useful application of the proposed approach is shown for the problem of state estimation in very high dimensional systems. Here a modified filter for data assimilation in an oceanic numerical model is presented which is proved to be very efficient due to introducing a simple Markovian structure for the output prediction error process and adaptive tuning some parameters of the Markov equation.Keywords: Statistical simulation, canonical form, dynamical system, Markov and non-Markovian processes, data assimilation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 12984808 Validating Condition-Based Maintenance Algorithms Through Simulation
Authors: Marcel Chevalier, Léo Dupont, Sylvain Marié, Frédérique Roffet, Elena Stolyarova, William Templier, Costin Vasile
Abstract:
Industrial end users are currently facing an increasing need to reduce the risk of unexpected failures and optimize their maintenance. This calls for both short-term analysis and long-term ageing anticipation. At Schneider Electric, we tackle those two issues using both Machine Learning and First Principles models. Machine learning models are incrementally trained from normal data to predict expected values and detect statistically significant short-term deviations. Ageing models are constructed from breaking down physical systems into sub-assemblies, then determining relevant degradation modes and associating each one to the right kinetic law. Validating such anomaly detection and maintenance models is challenging, both because actual incident and ageing data are rare and distorted by human interventions, and incremental learning depends on human feedback. To overcome these difficulties, we propose to simulate physics, systems and humans – including asset maintenance operations – in order to validate the overall approaches in accelerated time and possibly choose between algorithmic alternatives.
Keywords: Degradation models, ageing, anomaly detection, soft sensor, incremental learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 3294807 The Bright Side of Organizational Politics as a Driver of Firm Competitiveness: The Mediating Role of Corporate Entrepreneurship
Authors: Monika Kulikowska-Pawlak, Katarzyna Bratnicka-Myśliwiec, Tomasz Ingram
Abstract:
This study seeks to contribute to the literature on firm competitiveness by advancing the perspective of organizational politics that views this process as a driver which creates identifiable differences in firm performance. The hypothesized relationships were tested on the basis of data from 355 Polish medium and large-sized enterprises. Data were analyzed using correlation analysis, EFA and robustness tests. The main result of the conducted analyses proved the coexistence, previously examined in the literature, of corporate entrepreneurship and firm performance. The obtained research findings made it possible to add organizational politics to a wide range of elements determining corporate entrepreneurship, followed by competitive advantage, in addition to antecedents such as strategic leadership, corporate culture, opportunity-oriented resource-based management, etc. Also, the empirical results suggest that four dimensions of organizational politics (dominant coalition, influence exertion, making organizational changes, and information openness) are positively related to firm competitiveness. In addition, these findings seem to underline a supposition that corporate entrepreneurship is an important mediator which strengthens the competitive effects of organizational politics.
Keywords: Corporate entrepreneurship, firm competitiveness organizational politics, sensemaking.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 9344806 Ideal Disinfectant Characteristics According Data in Published Literature
Authors: Saimir Heta, Ilma Robo, Rialda Xhizdari, Kers Kapaj
Abstract:
The stability of an ideal disinfectant should be constant regardless of the change in the atmospheric conditions of the environment where it is kept. If the conditions such as temperature or humidity change, it is understood that it will also be necessary to approach possible changes in the holding materials such as plastic or glass bottles with the aim of protecting the disinfectant, for example, from the excessive lighting of the environment, which can also be translated as an increase in the temperature of disinfectant as a fluid. In this study, an attempt was made to find the most recent published data about the best possible combination of disinfectants indicated for use after dental procedures. This purpose of the study was realized by comparing the basic literature that is studied in the field of dentistry by students with the most published data in the literature of recent years about this topic. Each disinfectant is represented by a number called the disinfectant count, in which different factors can influence the increase or reduction of variables whose production remains a specific statistic for a specific disinfectant. The changes in the atmospheric conditions where the disinfectant is deposited and stored in the environment are known to affect the stability of the disinfectant as a fluid; this fact is known and even cited in the leaflets accompanying the manufactured boxes of disinfectants. It is these cares, in the form of advice, which are based not only on the preservation of the disinfectant but also on the application in order to have the desired clinical result. Aldehydes have the highest constant among the types of disinfectants, followed by acids. The lowest value of the constant belongs to the class of glycols, the predecessors of which were the halogens, in which class there are some representatives with disinfection applications. The class of phenols and acids have almost the same intervals of constants. If the goal were to find the ideal disinfectant among the large variety of disinfectants produced, a good starting point would be to find something unchanging or a fixed, unchanging element on the basis of which the comparison can be made properties of different disinfectants. Precisely based on the results of this study, the role of the specific constant according to the specific disinfectant is highlighted. Finding an ideal disinfectant, like finding a medication or the ideal antibiotic, is an ongoing but unattainable goal.
Keywords: Different disinfectants, phenols, aldehydes, specific constant, dental procedures.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 434805 Analyzing the Factors that Cause Parallel Performance Degradation in Parallel Graph-Based Computations Using Graph500
Authors: Mustafa Elfituri, Jonathan Cook
Abstract:
Recently, graph-based computations have become more important in large-scale scientific computing as they can provide a methodology to model many types of relations between independent objects. They are being actively used in fields as varied as biology, social networks, cybersecurity, and computer networks. At the same time, graph problems have some properties such as irregularity and poor locality that make their performance different than regular applications performance. Therefore, parallelizing graph algorithms is a hard and challenging task. Initial evidence is that standard computer architectures do not perform very well on graph algorithms. Little is known exactly what causes this. The Graph500 benchmark is a representative application for parallel graph-based computations, which have highly irregular data access and are driven more by traversing connected data than by computation. In this paper, we present results from analyzing the performance of various example implementations of Graph500, including a shared memory (OpenMP) version, a distributed (MPI) version, and a hybrid version. We measured and analyzed all the factors that affect its performance in order to identify possible changes that would improve its performance. Results are discussed in relation to what factors contribute to performance degradation.
Keywords: Graph computation, Graph500 benchmark, parallel architectures, parallel programming, workload characterization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5484804 Developing Manufacturing Process for the Graphene Sensors
Authors: Abdullah Faqihi, John Hedley
Abstract:
Biosensors play a significant role in the healthcare sectors, scientific and technological progress. Developing electrodes that are easy to manufacture and deliver better electrochemical performance is advantageous for diagnostics and biosensing. They can be implemented extensively in various analytical tasks such as drug discovery, food safety, medical diagnostics, process controls, security and defence, in addition to environmental monitoring. Development of biosensors aims to create high-performance electrochemical electrodes for diagnostics and biosensing. A biosensor is a device that inspects the biological and chemical reactions generated by the biological sample. A biosensor carries out biological detection via a linked transducer and transmits the biological response into an electrical signal; stability, selectivity, and sensitivity are the dynamic and static characteristics that affect and dictate the quality and performance of biosensors. In this research, a developed experimental study for laser scribing technique for graphene oxide inside a vacuum chamber for processing of graphene oxide is presented. The processing of graphene oxide (GO) was achieved using the laser scribing technique. The effect of the laser scribing on the reduction of GO was investigated under two conditions: atmosphere and vacuum. GO solvent was coated onto a LightScribe DVD. The laser scribing technique was applied to reduce GO layers to generate rGO. The micro-details for the morphological structures of rGO and GO were visualised using scanning electron microscopy (SEM) and Raman spectroscopy so that they could be examined. The first electrode was a traditional graphene-based electrode model, made under normal atmospheric conditions, whereas the second model was a developed graphene electrode fabricated under a vacuum state using a vacuum chamber. The purpose was to control the vacuum conditions, such as the air pressure and the temperature during the fabrication process. The parameters to be assessed include the layer thickness and the continuous environment. Results presented show high accuracy and repeatability achieving low cost productivity.Keywords: Laser scribing, LightScribe DVD, graphene oxide, scanning electron microscopy.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6654803 A New Approach for Predicting and Optimizing Weld Bead Geometry in GMAW
Authors: Farhad Kolahan, Mehdi Heidari
Abstract:
Gas Metal Arc Welding (GMAW) processes is an important joining process widely used in metal fabrication industries. This paper addresses modeling and optimization of this technique using a set of experimental data and regression analysis. The set of experimental data has been used to assess the influence of GMAW process parameters in weld bead geometry. The process variables considered here include voltage (V); wire feed rate (F); torch Angle (A); welding speed (S) and nozzle-to-plate distance (D). The process output characteristics include weld bead height, width and penetration. The Taguchi method and regression modeling are used in order to establish the relationships between input and output parameters. The adequacy of the model is evaluated using analysis of variance (ANOVA) technique. In the next stage, the proposed model is embedded into a Simulated Annealing (SA) algorithm to optimize the GMAW process parameters. The objective is to determine a suitable set of process parameters that can produce desired bead geometry, considering the ranges of the process parameters. Computational results prove the effectiveness of the proposed model and optimization procedure.Keywords: Weld Bead Geometry, GMAW welding, Processparameters Optimization, Modeling, SA algorithm
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21874802 Consumers’ Perceptions of Noncommunicable Diseases and Perceived Product Value Impacts on Healthy Food Purchasing Decisions
Authors: Khatesiree Sripoothon, Usanee Sengpanich, Rattana Sittioum
Abstract:
The objective of this study is to examine the factors influencing consumer purchasing decisions about healthy food. This model consists of two latent variables: Consumer Perception relating to NCDs and Consumer Perceived Product Value. The study was conducted in the northern provinces of Thailand, which are popular with tourists and have received support from the government for health and wellness tourism. A survey was used as the data collection method, and the questionnaire was applied to 385 consumers. An accidental sampling method was used to identify the sample. The statistics of frequency, percentage, mean, and structural equation model were used to analyze the data obtained. Additionally, all factors had a significant positive influence on healthy food purchasing decisions (p<0.001) and were predictive of healthy food purchasing decisions at 46.20% (R2=0.462). Also, these findings seem to underline the supposition that consumer perceptions of NCDs and perceived product value are key variables that strengthen the competitive effects of healthy-friendly business entrepreneurs. Moreover, it reduces the countries' public health costs for treating patients with the disease of NCDs in Thailand.
Keywords: healthy food, perceived product value, perception of noncommunicable diseases, purchasing decisions
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 561