Search results for: patch based similarity metric.
1354 Research of Strong-Column-Weak-Beam Criteria of Reinforced Concrete Frames Subjected to Biaxial Seismic Excitation
Authors: Chong Zhang, Mu-Xuan Tao
Abstract:
In several earthquakes, numerous reinforced concrete (RC) frames subjected to seismic excitation demonstrated a collapse pattern characterized by column hinges, though designed according to the Strong-Column-Weak-Beam (S-C-W-B) criteria. The effect of biaxial seismic excitation on the disparity between design and actual performance is carefully investigated in this article. First, a modified load contour method is proposed to derive a closed-form equation of biaxial bending moment strength, which is verified by numerical and experimental tests. Afterwards, a group of time history analyses of a simple frame modeled by fiber beam-column elements subjected to biaxial seismic excitation are conducted to verify that the current S-C-W-B criteria are not adequate to prevent the occurrence of column hinges. A biaxial over-strength factor is developed based on the proposed equation, and the reinforcement of columns is appropriately amplified with this factor to prevent the occurrence of column hinges under biaxial excitation, which is proved to be effective by another group of time history analyses.
Keywords: Biaxial bending moment strength, biaxial seismic excitation, fiber beam-column model, load contour method, strong-column-weak-beam.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6241353 Isolation and Classification of Red Blood Cells in Anemic Microscopic Images
Authors: Jameela Ali Alkrimi, Loay E. George, Azizah Suliman, Abdul Rahim Ahmad, Karim Al-Jashamy
Abstract:
Red blood cells (RBCs) are among the most commonly and intensively studied type of blood cells in cell biology. Anemia is a lack of RBCs is characterized by its level compared to the normal hemoglobin level. In this study, a system based image processing methodology was developed to localize and extract RBCs from microscopic images. Also, the machine learning approach is adopted to classify the localized anemic RBCs images. Several textural and geometrical features are calculated for each extracted RBCs. The training set of features was analyzed using principal component analysis (PCA). With the proposed method, RBCs were isolated in 4.3secondsfrom an image containing 18 to 27 cells. The reasons behind using PCA are its low computation complexity and suitability to find the most discriminating features which can lead to accurate classification decisions. Our classifier algorithm yielded accuracy rates of 100%, 99.99%, and 96.50% for K-nearest neighbor (K-NN) algorithm, support vector machine (SVM), and neural network RBFNN, respectively. Classification was evaluated in highly sensitivity, specificity, and kappa statistical parameters. In conclusion, the classification results were obtained within short time period, and the results became better when PCA was used.
Keywords: Red blood cells, pre-processing image algorithms, classification algorithms, principal component analysis PCA, confusion matrix, kappa statistical parameters, ROC.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 32021352 Combined Feature Based Hyperspectral Image Classification Technique Using Support Vector Machines
Authors: Mrs.K.Kavitha, S.Arivazhagan
Abstract:
A spatial classification technique incorporating a State of Art Feature Extraction algorithm is proposed in this paper for classifying a heterogeneous classes present in hyper spectral images. The classification accuracy can be improved if and only if both the feature extraction and classifier selection are proper. As the classes in the hyper spectral images are assumed to have different textures, textural classification is entertained. Run Length feature extraction is entailed along with the Principal Components and Independent Components. A Hyperspectral Image of Indiana Site taken by AVIRIS is inducted for the experiment. Among the original 220 bands, a subset of 120 bands is selected. Gray Level Run Length Matrix (GLRLM) is calculated for the selected forty bands. From GLRLMs the Run Length features for individual pixels are calculated. The Principle Components are calculated for other forty bands. Independent Components are calculated for next forty bands. As Principal & Independent Components have the ability to represent the textural content of pixels, they are treated as features. The summation of Run Length features, Principal Components, and Independent Components forms the Combined Features which are used for classification. SVM with Binary Hierarchical Tree is used to classify the hyper spectral image. Results are validated with ground truth and accuracies are calculated.
Keywords: Multi-class, Run Length features, PCA, ICA, classification and Support Vector Machines.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15241351 Oscillatory Electroosmotic Flow of Power-Law Fluids in a Microchannel
Authors: Rubén Bãnos, José Arcos, Oscar Bautista, Federico Méndez
Abstract:
The Oscillatory electroosmotic flow (OEOF) in power law fluids through a microchannel is studied numerically. A time-dependent external electric field (AC) is suddenly imposed at the ends of the microchannel which induces the fluid motion. The continuity and momentum equations in the x and y direction for the flow field were simplified in the limit of the lubrication approximation theory (LAT), and then solved using a numerical scheme. The solution of the electric potential is based on the Debye-H¨uckel approximation which suggest that the surface potential is small,say, smaller than 0.025V and for a symmetric (z : z) electrolyte. Our results suggest that the velocity profiles across the channel-width are controlled by the following dimensionless parameters: the angular Reynolds number, Reω, the electrokinetic parameter, ¯κ, defined as the ratio of the characteristic length scale to the Debye length, the parameter λ which represents the ratio of the Helmholtz-Smoluchowski velocity to the characteristic length scale and the flow behavior index, n. Also, the results reveal that the velocity profiles become more and more non-uniform across the channel-width as the Reω and ¯κ are increased, so oscillatory OEOF can be really useful in micro-fluidic devices such as micro-mixers.Keywords: Oscillatory electroosmotic flow, Non-Newtonian fluids, power-law model, low zeta potentials.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 8871350 An Efficient Biometric Cryptosystem using Autocorrelators
Authors: R. Bremananth, A. Chitra
Abstract:
Cryptography provides the secure manner of information transmission over the insecure channel. It authenticates messages based on the key but not on the user. It requires a lengthy key to encrypt and decrypt the sending and receiving the messages, respectively. But these keys can be guessed or cracked. Moreover, Maintaining and sharing lengthy, random keys in enciphering and deciphering process is the critical problem in the cryptography system. A new approach is described for generating a crypto key, which is acquired from a person-s iris pattern. In the biometric field, template created by the biometric algorithm can only be authenticated with the same person. Among the biometric templates, iris features can efficiently be distinguished with individuals and produces less false positives in the larger population. This type of iris code distribution provides merely less intra-class variability that aids the cryptosystem to confidently decrypt messages with an exact matching of iris pattern. In this proposed approach, the iris features are extracted using multi resolution wavelets. It produces 135-bit iris codes from each subject and is used for encrypting/decrypting the messages. The autocorrelators are used to recall original messages from the partially corrupted data produced by the decryption process. It intends to resolve the repudiation and key management problems. Results were analyzed in both conventional iris cryptography system (CIC) and non-repudiation iris cryptography system (NRIC). It shows that this new approach provides considerably high authentication in enciphering and deciphering processes.Keywords: Autocorrelators, biometrics cryptography, irispatterns, wavelets.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 15291349 A Domain Specific Modeling Language Semantic Model for Artefact Orientation
Authors: Bunakiye R. Japheth, Ogude U. Cyril
Abstract:
Since the process of transforming user requirements to modeling constructs are not very well supported by domain-specific frameworks, it became necessary to integrate domain requirements with the specific architectures to achieve an integrated customizable solutions space via artifact orientation. Domain-specific modeling language specifications of model-driven engineering technologies focus more on requirements within a particular domain, which can be tailored to aid the domain expert in expressing domain concepts effectively. Modeling processes through domain-specific language formalisms are highly volatile due to dependencies on domain concepts or used process models. A capable solution is given by artifact orientation that stresses on the results rather than expressing a strict dependence on complicated platforms for model creation and development. Based on this premise, domain-specific methods for producing artifacts without having to take into account the complexity and variability of platforms for model definitions can be integrated to support customizable development. In this paper, we discuss methods for the integration capabilities and necessities within a common structure and semantics that contribute a metamodel for artifact-orientation, which leads to a reusable software layer with concrete syntax capable of determining design intents from domain expert. These concepts forming the language formalism are established from models explained within the oil and gas pipelines industry.
Keywords: Control process, metrics of engineering, structured abstraction, semantic model.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7431348 Embodiment Design of an Azimuth-Altitude Solar Tracker
Authors: M. Culman, O. Lengerke
Abstract:
To provide an efficient solar generation system, the embodiment design of a two axis solar tracker for an array of photovoltaic (PV) panels destiny to supply the power demand on off-the-grid areas was developed. Photovoltaic cells have high costs in relation to t low efficiency; and while a lot of research and investment has been made to increases its efficiency a few points, there is a profitable solution that increases by 30-40% the annual power production: two axis solar trackers. A solar tracker is a device that supports a load in a perpendicular position toward the sun during daylight. Mounted on solar trackers, the solar panels remain perpendicular to the incoming sunlight at day and seasons so the maximum amount of energy is outputted. Through a preview research done it was justified why the generation of solar energy through photovoltaic panels mounted on dual axis structures is an attractive solution to bring electricity to remote off-the-grid areas. The work results are the embodiment design of an azimuth-altitude solar tracker to guide an array of photovoltaic panels based on a specific design methodology. The designed solar tracker is mounted on a pedestal that uses two slewing drives‚ with a nominal torque of 1950 Nm‚ to move a solar array that provides 3720 W from 12 PV panels.
Keywords: Azimuth-altitude sun tracker, dual-axis solar tracker, photovoltaic system, solar energy, stand-alone power system.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 17871347 Problems and Needs of Frozen Shrimp Industry Small and Medium Enterprises in the Central Region of the Lower Three Provinces
Authors: P. Thepnarintra
Abstract:
Frozen shrimp industry plays an important role in the development of production industry of the country. There has been a continuing development to response the increasing demand; however, there have been some problems in running the enterprises. The purposes of this study are to: 1) investigate problems related to basic factors in operating frozen shrimp industry based on the entrepreneurs’ points of view. The enterprises involved in this study were small and medium industry receiving Thai Frozen Foods Association. 2) Compare the problems of the frozen shrimp industry according to their sizes of operation in 3 provinces of the central region Thailand. Population in this study consisted of 148 managers from 148 frozen shrimp enterprises Thai Frozen Foods Association which 77 were small size and 71 were medium size. The data were analyzed to find percentage, arithmetic mean, standard deviation, and independent sample T-test with the significant hypothesis at .05. The results revealed that the problems of the frozen shrimp industries of both size were in high level. The needs for government supporting were in high level. The comparison of the problems and the basic factors between the small and medium size enterprises showed no statistically significant level. The problems that they mentioned included raw materials, labors, production, marketing, and the need for academic supporting from the government sector.Keywords: Frozen shrimp industry, problems, related to the enterprise, operation.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11061346 Investigating the Potential for Introduction of Warm Mix Asphalt in Kuwait Using the Volcanic Ash
Authors: H. Al-Baghli, F. Al-Asfour
Abstract:
The current applied asphalt technology for Kuwait roads pavement infrastructure is the hot mix asphalt (HMA) pavement, including both pen grade and polymer modified bitumen (PMBs), that is produced and compacted at high temperature levels ranging from 150 to 180 °C. There are no current specifications for warm and cold mix asphalts in Kuwait’s Ministry of Public Works (MPW) asphalt standard and specifications. The process of the conventional HMA is energy intensive and directly responsible for the emission of greenhouse gases and other environmental hazards into the atmosphere leading to significant environmental impacts and raising health risk to labors at site. Warm mix asphalt (WMA) technology, a sustainable alternative preferred in multiple countries, has many environmental advantages because it requires lower production temperatures than HMA by 20 to 40 °C. The reduction of temperatures achieved by WMA originates from multiple technologies including foaming and chemical or organic additives that aim to reduce bitumen and improve mix workability. This paper presents a literature review of WMA technologies and techniques followed by an experimental study aiming to compare the results of produced WMA samples, using a water containing additive (foaming process), at different compaction temperatures with the HMA control volumetric properties mix designed in accordance to the new MPW’s specifications and guidelines.
Keywords: Warm-mix asphalt, water-bearing additives, foaming-based process, chemical additives, organic additives.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 5031345 Detecting Geographically Dispersed Overlay Communities Using Community Networks
Authors: Madhushi Bandara, Dharshana Kasthurirathna, Danaja Maldeniya, Mahendra Piraveenan
Abstract:
Community detection is an extremely useful technique in understanding the structure and function of a social network. Louvain algorithm, which is based on Newman-Girman modularity optimization technique, is extensively used as a computationally efficient method extract the communities in social networks. It has been suggested that the nodes that are in close geographical proximity have a higher tendency of forming communities. Variants of the Newman-Girman modularity measure such as dist-modularity try to normalize the effect of geographical proximity to extract geographically dispersed communities, at the expense of losing the information about the geographically proximate communities. In this work, we propose a method to extract geographically dispersed communities while preserving the information about the geographically proximate communities, by analyzing the ‘community network’, where the centroids of communities would be considered as network nodes. We suggest that the inter-community link strengths, which are normalized over the community sizes, may be used to identify and extract the ‘overlay communities’. The overlay communities would have relatively higher link strengths, despite being relatively apart in their spatial distribution. We apply this method to the Gowalla online social network, which contains the geographical signatures of its users, and identify the overlay communities within it.Keywords: Social networks, community detection, modularity optimization, geographically dispersed communities.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 13001344 Experimental Testing of Statistical Size Effect in Civil Engineering Structures
Authors: Jana Kaděrová, Miroslav Vořechovský
Abstract:
The presented paper copes with an experimental evaluation of a model based on modified Weibull size effect theory. Classical statistical Weibull theory was modified by introducing a new parameter (correlation length lp) representing the spatial autocorrelation of a random mechanical properties of material. This size effect modification was observed on two different materials used in civil engineering: unreinforced (plain) concrete and multi-filament yarns made of alkaliresistant (AR) glass which are used for textile-reinforced concrete. The behavior under flexural, resp. tensile loading was investigated by laboratory experiments. A high number of specimens of different sizes was tested to obtain statistically significant data which were subsequently corrected and statistically processed. Due to a distortion of the measured displacements caused by the unstiff experiment device, only the maximal load values were statistically evaluated. Results of the experiments showed a decreasing strength with an increasing sample length. Size effect curves were obtained and the correlation length was fitted according to measured data. Results did not exclude the existence of the proposed new parameter lp.
Keywords: Statistical size effect, concrete, multi filaments yarns, experiment, autocorrelation length.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19861343 The Significance of Awareness about Gender Diversity for the Future of Work: A Multi-Method Study of Organizational Structures and Policies Considering Trans and Gender Diversity
Authors: Robin C. Ladwig
Abstract:
The future of work becomes less predictable which requires increasing adaptability of organizations to social and work changes. Society is transforming regarding gender identity in the sense that more people come forward to identify as trans and gender diverse (TGD). Organizations are ill-equipped to provide a safe and encouraging work environment by lacking inclusive organizational structures. The qualitative multi-method research about TGD inclusivity in the workplace explores the enablers and barriers for TGD individuals to satisfactorily engage in the work environment and organizational culture. Furthermore, these TGD insights are analyzed based on organizational implications and awareness from a leadership and management perspective. The semi-structured online interviews with TGD individuals and the photo-elicit open-ended questionnaire addressed to leadership and management in diversity, career development, and human resources have been analyzed with a critical grounded theory approach. Findings demonstrated the significance of TGD voices, the support of leadership and management, as well as the synergy between voices and leadership. Hence, it indicates practical implications such as the revision of exclusive language used in policies, data collection, or communication and reconsideration of organizational decision-making by leaders to include TGD voices.
Keywords: Future of work, occupational identity, organizational decision-making, trans and gender diverse identity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4711342 Learning Classifier Systems Approach for Automated Discovery of Crisp and Fuzzy Hierarchical Production Rules
Authors: Suraiya Jabin, Kamal K. Bharadwaj
Abstract:
This research presents a system for post processing of data that takes mined flat rules as input and discovers crisp as well as fuzzy hierarchical structures using Learning Classifier System approach. Learning Classifier System (LCS) is basically a machine learning technique that combines evolutionary computing, reinforcement learning, supervised or unsupervised learning and heuristics to produce adaptive systems. A LCS learns by interacting with an environment from which it receives feedback in the form of numerical reward. Learning is achieved by trying to maximize the amount of reward received. Crisp description for a concept usually cannot represent human knowledge completely and practically. In the proposed Learning Classifier System initial population is constructed as a random collection of HPR–trees (related production rules) and crisp / fuzzy hierarchies are evolved. A fuzzy subsumption relation is suggested for the proposed system and based on Subsumption Matrix (SM), a suitable fitness function is proposed. Suitable genetic operators are proposed for the chosen chromosome representation method. For implementing reinforcement a suitable reward and punishment scheme is also proposed. Experimental results are presented to demonstrate the performance of the proposed system.Keywords: Hierarchical Production Rule, Data Mining, Learning Classifier System, Fuzzy Subsumption Relation, Subsumption matrix, Reinforcement Learning.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 14581341 Expectation-Confirmation Model of Information System Continuance: A Meta-Analysis
Authors: Hui-Min Lai, Chin-Pin Chen, Yung-Fu Chang
Abstract:
The expectation-confirmation model (ECM) is one of the most widely used models for evaluating information system continuance, and this model has been extended to other study backgrounds, or expanded with other theoretical perspectives. However, combining ECM with other theories or investigating the background problem may produce some disparities, thus generating inaccurate conclusions. Habit is considered to be an important factor that influences the user’s continuance behavior. This paper thus critically examines seven pairs of relationships from the original ECM and the habit variable. A meta-analysis was used to tackle the development of ECM research over the last 10 years from a range of journals and conference papers published in 2005–2014. Forty-six journal articles and 19 conference papers were selected for analysis. The results confirm our prediction that a high effect size for the seven pairs of relationships was obtained (ranging from r=0.386 to r=0.588). Furthermore, a meta-analytic structural equation modeling was performed to simultaneously test all relationships. The results show that habit had a significant positive effect on continuance intention at p<=0.05 and that the six other pairs of relationships were significant at p<0.10. Based on the findings, we refined our original research model and an alternative model was proposed for understanding and predicting information system continuance. Some theoretical implications are also discussed.Keywords: Expectation-confirmation theory, expectation- confirmation model, meta-analysis, meta-analytic structural equation modeling.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27451340 Radial Basis Surrogate Model Integrated to Evolutionary Algorithm for Solving Computation Intensive Black-Box Problems
Authors: Abdulbaset Saad, Adel Younis, Zuomin Dong
Abstract:
For design optimization with high-dimensional expensive problems, an effective and efficient optimization methodology is desired. This work proposes a series of modification to the Differential Evolution (DE) algorithm for solving computation Intensive Black-Box Problems. The proposed methodology is called Radial Basis Meta-Model Algorithm Assisted Differential Evolutionary (RBF-DE), which is a global optimization algorithm based on the meta-modeling techniques. A meta-modeling assisted DE is proposed to solve computationally expensive optimization problems. The Radial Basis Function (RBF) model is used as a surrogate model to approximate the expensive objective function, while DE employs a mechanism to dynamically select the best performing combination of parameters such as differential rate, cross over probability, and population size. The proposed algorithm is tested on benchmark functions and real life practical applications and problems. The test results demonstrate that the proposed algorithm is promising and performs well compared to other optimization algorithms. The proposed algorithm is capable of converging to acceptable and good solutions in terms of accuracy, number of evaluations, and time needed to converge.
Keywords: Differential evolution, engineering design, expensive computations, meta-modeling, radial basis function, optimization.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11771339 AI-Driven Cloud Security: Proactive Defense Against Evolving Cyber Threats
Authors: Ashly Joseph
Abstract:
Cloud computing has become an essential component of enterprises and organizations globally in the current era of digital technology. The cloud has a multitude of advantages, including scalability, flexibility, and cost-effectiveness, rendering it an appealing choice for data storage and processing. The increasing storage of sensitive information in cloud environments has raised significant concerns over the security of such systems. The frequency of cyber threats and attacks specifically aimed at cloud infrastructure has been increasing, presenting substantial dangers to the data, reputation, and financial stability of enterprises. Conventional security methods can become inadequate when confronted with ever intricate and dynamic threats. Artificial Intelligence (AI) technologies possess the capacity to significantly transform cloud security through their ability to promptly identify and thwart assaults, adjust to emerging risks, and offer intelligent perspectives for proactive security actions. The objective of this research study is to investigate the utilization of AI technologies in augmenting the security measures within cloud computing systems. This paper aims to offer significant insights and recommendations for businesses seeking to protect their cloud-based assets by analyzing the present state of cloud security, the capabilities of AI, and the possible advantages and obstacles associated with using AI into cloud security policies.
Keywords: Machine Learning, Natural Learning Processing, Denial-of-Service attacks, Sentiment Analysis, Cloud computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2231338 Effects of Initial Moisture Content on the Physical and Mechanical Properties of Norway Spruce Briquettes
Authors: Miloš Matúš, Peter Križan, Ľubomír Šooš, Juraj Beniak
Abstract:
The moisture content of densified biomass is a limiting parameter influencing the quality of this solid biofuel. It influences its calorific value, density, mechanical strength and dimensional stability as well as affecting its production process. This paper deals with experimental research into the effect of moisture content of the densified material on the final quality of biofuel in the form of logs (briquettes or pellets). Experiments based on the singleaxis densification of the spruce sawdust were carried out with a hydraulic piston press (piston and die), where the densified logs were produced at room temperature. The effect of moisture content on the qualitative properties of the logs, including density, change of moisture, expansion and physical changes, and compressive and impact resistance were studied. The results show the moisture ranges required for producing good-quality logs. The experiments were evaluated and the moisture content of the tested material was optimized to achieve the optimum value for the best quality of the solid biofuel. The dense logs also have high-energy content per unit volume. The research results could be used to develop and optimize industrial technologies and machinery for biomass densification to achieve high quality solid biofuel.Keywords: Biomass, briquettes, densification, fuel quality, moisture content, density.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 27701337 Characteristics of Regional Issues in Local Municipalities of Japan in Consideration of Socio-Economic Condition
Authors: Akiko Kondo, Akio Kondo
Abstract:
We are facing serious problems related to long-term depopulation and an aging society with a falling birth rate in Japan. In this situation, we are suffering from a shortfall in human resources as well as a shortage of workforce in rural regions. In addition, we are struggling with a protracted economic slump and excess concentration of population in the Tokyo Metropolitan area. It is an urgent national issue to consider how to live in this country and what kind of structure of society and administration policy is needed. It is necessary to clarify people’s desire for their way of living and social assistance to be provided. The aim of this study is to clarify the characteristics of regional issues and the degree of their seriousness in local municipalities of Japan. We conducted a questionnaire survey about regional agenda in all local municipalities in Japan. We obtained responses concerning the degree of seriousness of regional issues and degree of importance of policies. Based on the data gathered from the survey, it is apparent that many local municipalities are facing an aging population and declining population. We constructed a model to analyze factors for declining population. Using the model, it was clarified that a population’s age structure, job opportunities and income level affect the decline of population. In addition, we showed the way of the evaluation of state of local municipality.
Keywords: Evaluation, Local municipality, Regional analysis, Regional issue.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18641336 A Study on the Attractiveness of Heavy Duty Motorcycle
Authors: Kaishuan Shen, Pan Changyu, Yuhsiang Lu, Zongshao Liu, Chishxsin Chuang, Minyuan Ma
Abstract:
The culture of riding heavy motorcycles originates from advanced countries and mainly comes from Europe, North America, and Japan. Heavy duty motorcycle riders are different from people who view motorcycles as a convenient mean of transportation. They regard riding them as a kind of enjoyment and high-level taste. The activities of riding heavy duty motorcycles have formes a distinctive landscape in domestic land in Taiwan. Previous studies which explored motorcycle culture in Taiwan still focused on the objects of motorcycle engine displacement under 50 cc.. The study aims to study the heavy duty motorcycles of engine displacement over 550 cc. and explores where their attractiveness is. For finding the attractiveness of heavy duty motorcycle, the study chooses Miryoku Engineering (Preference-Based Design) approach. Two steps are adopted to proceed the research. First, through arranging the letters obtained from interviewing experts, EGM (The Evaluation Grid Method) was applied to find out the structure of attractiveness. The attractive styles are eye-dazzling, leisure, classic, and racing competitive styles. Secondarily, Quantification Theory Type I analysis was adopted as a tool for analyzing the importance of attractiveness. The relationship between style and attractive parts was also discussed. The results could contribute to the design and research development of heavy duty motorcycle industry in Taiwan.Keywords: attractiveness, evaluation, heavy dutymotorcycle, miryoku engineering
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19201335 An Investigation to Study the Moisture Dependency of Ground Enhancement Compound
Authors: Arunima Shukla, Vikas Almadi, Devesh Jaiswal, Sunil Saini, Bhusan S. Patil
Abstract:
Lightning protection consists of three main parts; mainly air termination system, down conductor, and earth termination system. Earth termination system is the most important part as earth is the sink and source of charges. Therefore, even when the charges are captured and delivered to the ground, and an easy path is not provided to the charges, earth termination system would lead to problems. Soil has significantly different resistivities ranging from 10 Ωm for wet organic soil to 10000 Ωm for bedrock. Different methods have been discussed and used conventionally such as deep-ground-well method and altering the length of the rod. Those methods are not considered economical. Therefore, it was a general practice to use charcoal along with salt to reduce the soil resistivity. Bentonite is worldwide acceptable material, that had led our interest towards study of bentonite at first. It was concluded that bentonite is a clay which is non-corrosive, environment friendly. Whereas bentonite is suitable only when there is moisture present in the soil, as in the absence of moisture, cracks will appear on the surface which will provide an open passage to the air, resulting into increase in the resistivity. Furthermore, bentonite without moisture does not have enough bonding property, moisture retention, conductivity, and non-leachability. Therefore, bentonite was used along with the other backfill material to overcome the dependency of bentonite on moisture. Different experiments were performed to get the best ratio of bentonite and carbon backfill. It was concluded that properties will highly depend on the quantity of bentonite and carbon-based backfill material.
Keywords: Backfill material, bentonite, conducting soil, grounding material, low resistivity.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4531334 Analysis and Evaluation of the Public Responses to Traffic Congestion Pricing Schemes in Urban Streets
Authors: Saeed Sayyad Hagh Shomar
Abstract:
Traffic congestion pricing in urban streets is one of the most suitable options for solving the traffic problems and environment pollutions in the cities of the country. Unlike its acceptable outcomes, there are problems concerning the necessity to pay by the mass. Regarding the fact that public response in order to succeed in this strategy is so influential, studying their response and behavior to get the feedback and improve the strategies is of great importance. In this study, a questionnaire was used to examine the public reactions to the traffic congestion pricing schemes at the center of Tehran metropolis and the factors involved in people’s decision making in accepting or rejecting the congestion pricing schemes were assessed based on the data obtained from the questionnaire as well as the international experiences. Then, by analyzing and comparing the schemes, guidelines to reduce public objections to them are discussed. The results of reviewing and evaluating the public reactions show that all the pros and cons must be considered to guarantee the success of these projects. Consequently, with targeted public education and consciousness-raising advertisements, prior to initiating a scheme and ensuring the mechanism of the implementation after the start of the project, the initial opposition is reduced and, with the gradual emergence of the real and tangible benefits of its implementation, users’ satisfaction will increase.
Keywords: Demand management, international experiences, traffic congestion pricing, public acceptance, public objection.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 6561333 The Analysis of Secondary Case Studies as a Starting Point for Grounded Theory Studies: An Example from the Enterprise Software Industry
Authors: Abilio Avila, Orestis Terzidis
Abstract:
A fundamental principle of Grounded Theory (GT) is to prevent the formation of preconceived theories. This implies the need to start a research study with an open mind and to avoid being absorbed by the existing literature. However, to start a new study without an understanding of the research domain and its context can be extremely challenging. This paper presents a research approach that simultaneously supports a researcher to identify and to focus on critical areas of a research project and prevent the formation of prejudiced concepts by the current body of literature. This approach comprises of four stages: Selection of secondary case studies, analysis of secondary case studies, development of an initial conceptual framework, development of an initial interview guide. The analysis of secondary case studies as a starting point for a research project allows a researcher to create a first understanding of a research area based on real-world cases without being influenced by the existing body of theory. It enables a researcher to develop through a structured course of actions a firm guide that establishes a solid starting point for further investigations. Thus, the described approach may have significant implications for GT researchers who aim to start a study within a given research area.Keywords: Grounded theory, qualitative research, secondary case studies, secondary data analysis, interview guide.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18471332 A Combined Conventional and Differential Evolution Method for Model Order Reduction
Authors: J. S. Yadav, N. P. Patidar, J. Singhai, S. Panda, C. Ardil
Abstract:
In this paper a mixed method by combining an evolutionary and a conventional technique is proposed for reduction of Single Input Single Output (SISO) continuous systems into Reduced Order Model (ROM). In the conventional technique, the mixed advantages of Mihailov stability criterion and continued Fraction Expansions (CFE) technique is employed where the reduced denominator polynomial is derived using Mihailov stability criterion and the numerator is obtained by matching the quotients of the Cauer second form of Continued fraction expansions. Then, retaining the numerator polynomial, the denominator polynomial is recalculated by an evolutionary technique. In the evolutionary method, the recently proposed Differential Evolution (DE) optimization technique is employed. DE method is based on the minimization of the Integral Squared Error (ISE) between the transient responses of original higher order model and the reduced order model pertaining to a unit step input. The proposed method is illustrated through a numerical example and compared with ROM where both numerator and denominator polynomials are obtained by conventional method to show its superiority.
Keywords: Reduced Order Modeling, Stability, Mihailov Stability Criterion, Continued Fraction Expansions, Differential Evolution, Integral Squared Error.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 21661331 Uniform Overlapped Multi-Carrier PWM for a Six-Level Diode Clamped Inverter
Authors: S.Srinivas
Abstract:
Multi-level voltage source inverters offer several advantages such as; derivation of a refined output voltage with reduced total harmonic distortion (THD), reduction of voltage ratings of the power semiconductor switching devices and also the reduced electro-magnetic-interference problems etc. In this paper, new carrier-overlapped phase-disposition or sub-harmonic sinusoidal pulse width modulation (CO-PD-SPWM) and also the carrieroverlapped phase-disposition space vector modulation (CO-PDSVPWM) schemes for a six-level diode-clamped inverter topology are proposed. The principle of the proposed PWM schemes is similar to the conventional PD-PWM with a little deviation from it in the sense that the triangular carriers are all overlapped. The overlapping of the triangular carriers on one hand results in an increased number of switchings, on the other hand this facilitates an improved spectral performance of the output voltage. It is demonstrated through simulation studies that the six-level diode-clamped inverter with the use of CO-PD-SPWM and CO-PD-SVPWM proposed in this paper is capable of generating multiple levels in its output voltage. The advantages of the proposed PWM schemes can be derived to benefit, especially at lower modulation indices of the inverter and hence this aspect of the proposed PWM schemes can be well exploited in high power applications requiring low speeds of operation of the drive.Keywords: Diode clamped inverter, Pulse width modulation, Six level inverter, carrier based PWM.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 19221330 Design of an Ensemble Learning Behavior Anomaly Detection Framework
Authors: Abdoulaye Diop, Nahid Emad, Thierry Winter, Mohamed Hilia
Abstract:
Data assets protection is a crucial issue in the cybersecurity field. Companies use logical access control tools to vault their information assets and protect them against external threats, but they lack solutions to counter insider threats. Nowadays, insider threats are the most significant concern of security analysts. They are mainly individuals with legitimate access to companies information systems, which use their rights with malicious intents. In several fields, behavior anomaly detection is the method used by cyber specialists to counter the threats of user malicious activities effectively. In this paper, we present the step toward the construction of a user and entity behavior analysis framework by proposing a behavior anomaly detection model. This model combines machine learning classification techniques and graph-based methods, relying on linear algebra and parallel computing techniques. We show the utility of an ensemble learning approach in this context. We present some detection methods tests results on an representative access control dataset. The use of some explored classifiers gives results up to 99% of accuracy.Keywords: Cybersecurity, data protection, access control, insider threat, user behavior analysis, ensemble learning, high performance computing.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 11681329 Assessment of Analytical Equations for the Derivation of Young’s Modulus of Bonded Rubber Materials
Authors: Z. N. Haji, S. O. Oyadiji, H. Samami, O. Farrell
Abstract:
The prediction of the vibration response of rubber products by analytical or numerical method depends mainly on the predefined intrinsic material properties such as Young’s modulus, damping factor and Poisson’s ratio. Such intrinsic properties are determined experimentally by subjecting a bonded rubber sample to compression tests. The compression tests on such a sample yield an apparent Young’s modulus which is greater in magnitude than the intrinsic Young’s modulus of the rubber. As a result, many analytical equations have been developed to determine Young’s modulus from an apparent Young’s modulus of bonded rubber materials. In this work, the applicability of some of these analytical equations is assessed via experimental testing. The assessment is based on testing of vulcanized nitrile butadiene rubber (NBR70) samples using tensile test and compression test methods. The analytical equations are used to determine the intrinsic Young’s modulus from the apparent modulus that is derived from the compression test data of the bonded rubber samples. Then, these Young’s moduli are compared with the actual Young’s modulus that is derived from the tensile test data. The results show significant discrepancy between the Young’s modulus derived using the analytical equations and the actual Young’s modulus.
Keywords: Bonded rubber, quasi-static test, shape factor, apparent Young’s modulus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 7561328 Experimental Study on the Variation of Young's Modulus of Hollow Clay Brick Obtained from Static and Dynamic Tests
Authors: M. Aboudalle, Le Btth, M. Sari, F. Meftah
Abstract:
In parallel with the appearance of new materials, brick masonry had and still has an essential part of the construction market today, with new technical challenges in designing bricks to meet additional requirements. Being used in structural applications, predicting the performance of clay brick masonry allows a significant cost reduction, in terms of practical experimentation. The behavior of masonry walls depends on the behavior of their elementary components, such as bricks, joints, and coatings. Therefore, it is necessary to consider it at different scales (from the scale of the intrinsic material to the real scale of the wall) and then to develop appropriate models, using numerical simulations. The work presented in this paper focuses on the mechanical characterization of the terracotta material at ambient temperature. As a result, the static Young’s modulus obtained from the flexural test shows different values in comparison with the compression test, as well as with the dynamic Young’s modulus obtained from the Impulse excitation of vibration test. Moreover, the Young's modulus varies according to the direction in which samples are extracted, where the values in the extrusion direction diverge from the ones in the orthogonal directions. Based on these results, hollow bricks can be considered as transversely isotropic bimodulus material.
Keywords: Bimodulus material, hollow clay brick, impulse excitation of vibration, transversely isotropic material, Young’s modulus.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 4621327 A Novel Approach for Coin Identification using Eigenvalues of Covariance Matrix, Hough Transform and Raster Scan Algorithms
Authors: J. Prakash, K. Rajesh
Abstract:
In this paper we present a new method for coin identification. The proposed method adopts a hybrid scheme using Eigenvalues of covariance matrix, Circular Hough Transform (CHT) and Bresenham-s circle algorithm. The statistical and geometrical properties of the small and large Eigenvalues of the covariance matrix of a set of edge pixels over a connected region of support are explored for the purpose of circular object detection. Sparse matrix technique is used to perform CHT. Since sparse matrices squeeze zero elements and contain only a small number of non-zero elements, they provide an advantage of matrix storage space and computational time. Neighborhood suppression scheme is used to find the valid Hough peaks. The accurate position of the circumference pixels is identified using Raster scan algorithm which uses geometrical symmetry property. After finding circular objects, the proposed method uses the texture on the surface of the coins called texton, which are unique properties of coins, refers to the fundamental micro structure in generic natural images. This method has been tested on several real world images including coin and non-coin images. The performance is also evaluated based on the noise withstanding capability.Keywords: Circular Hough Transform, Coin detection, Covariance matrix, Eigenvalues, Raster scan Algorithm, Texton.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 18821326 Treatment of Oily Wastewater by Fibrous Coalescer Process: Stage Coalescer and Model Prediction
Authors: Pisut Painmanakul, Kotchakorn Kongkangwarn, Nattawin Chawaloesphonsiya
Abstract:
The coalescer process is one of the methods for oily water treatment by increasing the oil droplet size in order to enhance the separating velocity and thus effective separation. However, the presence of surfactants in an oily emulsion can limit the obtained mechanisms due to the small oil size related with stabilized emulsion. In this regard, the purpose of this research is to improve the efficiency of the coalescer process for treating the stabilized emulsion. The effects of bed types, bed height, liquid flow rate and stage coalescer (step-bed) on the treatment efficiencies in term of COD values were studied. Note that the treatment efficiency obtained experimentally was estimated by using the COD values and oil droplet size distribution. The study has shown that the plastic media has more effective to attach with oil particles than the stainless one due to their hydrophobic properties. Furthermore, the suitable bed height (3.5 cm) and step bed (3.5 cm with 2 steps) were necessary in order to well obtain the coalescer performance. The application of step bed coalescer process in reactor has provided the higher treatment efficiencies in term of COD removal than those obtained with classical process. The proposed model for predicting the area under curve and thus treatment efficiency, based on the single collector efficiency (ηT) and the attachment efficiency (α), provides relatively a good coincidence between the experimental and predicted values of treatment efficiencies in this study.
Keywords: Stage coalescer, stabilized emulsions, treatment efficiency, model prediction.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 22001325 An Approach of Quantum Steganography through Special SSCE Code
Authors: Indradip Banerjee, Souvik Bhattacharyya, Gautam Sanyal
Abstract:
Encrypted messages sending frequently draws the attention of third parties, perhaps causing attempts to break and reveal the original messages. Steganography is introduced to hide the existence of the communication by concealing a secret message in an appropriate carrier like text, image, audio or video. Quantum steganography where the sender (Alice) embeds her steganographic information into the cover and sends it to the receiver (Bob) over a communication channel. Alice and Bob share an algorithm and hide quantum information in the cover. An eavesdropper (Eve) without access to the algorithm can-t find out the existence of the quantum message. In this paper, a text quantum steganography technique based on the use of indefinite articles (a) or (an) in conjunction with the nonspecific or non-particular nouns in English language and quantum gate truth table have been proposed. The authors also introduced a new code representation technique (SSCE - Secret Steganography Code for Embedding) at both ends in order to achieve high level of security. Before the embedding operation each character of the secret message has been converted to SSCE Value and then embeds to cover text. Finally stego text is formed and transmits to the receiver side. At the receiver side different reverse operation has been carried out to get back the original information.Keywords: Quantum Steganography, SSCE (Secret SteganographyCode for Embedding), Security, Cover Text, Stego Text.
Procedia APA BibTeX Chicago EndNote Harvard JSON MLA RIS XML ISO 690 PDF Downloads 2110