Search results for: calibration data requirements
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 26407

Search results for: calibration data requirements

25837 Refining Employee's Customer Service Performance through an Inter-Organizational Climate Study: A Way Forward

Authors: Zainal Abu Zatim, Hafizah Omar Zaki

Abstract:

Substantial research had been done on refining employee’s customer service performance. Thus, there were very limited empirical studies that are engage in an inter-organizational climate study in assessing employee’s customer service performance. With the current economic situation as well as emerging needs and requirements, all businesses either from public or private sector serving customers put greater attention on fulfilling those needs and requirements. In this state of affairs, the act of polishing its employees’ skills, knowledge, teamwork and passion is very important in ensuring better performance deliverance. A study conducted in one of the telecommunication service provider company in Malaysia had been done to test its inter-organizational climate study. The Internal Climate Study was done to benchmark opinions and perceptions of its employees. The study had provided baseline information about perceptions that exist in the internal environment and ways forward to improve customer service performance. The approach used is through the use of focus group and qualitative interview.

Keywords: employees, Customer Service Performance, inter-organizational climate study, public and private sector

Procedia PDF Downloads 385
25836 Transaction Cost Analysis, Execution Quality, and Best Execution under MiFID II

Authors: Rodrigo Zepeda

Abstract:

Transaction cost analysis (TCA) is a way of analyzing the relative performance of different intermediaries and different trading strategies for trades undertaken in financial instruments. It is a way for an investor to determine the overall quality of execution of a particular trade, and there are many different approaches to undertaking TCA. Under the updated Markets in Financial Instruments Directive (2014/65/EU) (MiFID II), investment firms are required when executing orders, to take all sufficient steps to obtain the best possible result for their clients. This requirement for 'Best Execution' must take into account price, costs, speed, likelihood of execution and settlement, size, nature or any other consideration relevant to the execution of the order. The new regulatory compliance framework under MiFID II will now also apply across a very broad range of financial instruments. This article will provide a comprehensive technical analysis of how TCA and Best Execution will significantly change under MiFID II. It will also explain why harmonization of post-trade reporting requirements under MiFID II could potentially support the development of peer group analysis, which in turn could provide a new and highly advanced framework for TCA that could more effectively support Best Execution requirements under MiFID II. The study is significant because there are no studies that have dealt with TCA and Best Execution under MiFID II in the literature.

Keywords: transaction cost analysis, execution quality, best execution, MiFID II, financial instruments

Procedia PDF Downloads 278
25835 Workflow Based Inspection of Geometrical Adaptability from 3D CAD Models Considering Production Requirements

Authors: Tobias Huwer, Thomas Bobek, Gunter Spöcker

Abstract:

Driving forces for enhancements in production are trends like digitalization and individualized production. Currently, such developments are restricted to assembly parts. Thus, complex freeform surfaces are not addressed in this context. The need for efficient use of resources and near-net-shape production will require individualized production of complex shaped workpieces. Due to variations between nominal model and actual geometry, this can lead to changes in operations in Computer-aided process planning (CAPP) to make CAPP manageable for an adaptive serial production. In this context, 3D CAD data can be a key to realizing that objective. Along with developments in the geometrical adaptation, a preceding inspection method based on CAD data is required to support the process planner by finding objective criteria to make decisions about the adaptive manufacturability of workpieces. Nowadays, this kind of decisions is depending on the experience-based knowledge of humans (e.g. process planners) and results in subjective decisions – leading to a variability of workpiece quality and potential failure in production. In this paper, we present an automatic part inspection method, based on design and measurement data, which evaluates actual geometries of single workpiece preforms. The aim is to automatically determine the suitability of the current shape for further machining, and to provide a basis for an objective decision about subsequent adaptive manufacturability. The proposed method is realized by a workflow-based approach, keeping in mind the requirements of industrial applications. Workflows are a well-known design method of standardized processes. Especially in applications like aerospace industry standardization and certification of processes are an important aspect. Function blocks, providing a standardized, event-driven abstraction to algorithms and data exchange, will be used for modeling and execution of inspection workflows. Each analysis step of the inspection, such as positioning of measurement data or checking of geometrical criteria, will be carried out by function blocks. One advantage of this approach is its flexibility to design workflows and to adapt algorithms specific to the application domain. In general, within the specified tolerance range it will be checked if a geometrical adaption is possible. The development of particular function blocks is predicated on workpiece specific information e.g. design data. Furthermore, for different product lifecycle phases, appropriate logics and decision criteria have to be considered. For example, tolerances for geometric deviations are different in type and size for new-part production compared to repair processes. In addition to function blocks, appropriate referencing systems are important. They need to support exact determination of position and orientation of the actual geometries to provide a basis for precise analysis. The presented approach provides an inspection methodology for adaptive and part-individual process chains. The analysis of each workpiece results in an inspection protocol and an objective decision about further manufacturability. A representative application domain is the product lifecycle of turbine blades containing a new-part production and a maintenance process. In both cases, a geometrical adaptation is required to calculate individual production data. In contrast to existing approaches, the proposed initial inspection method provides information to decide between different potential adaptive machining processes.

Keywords: adaptive, CAx, function blocks, turbomachinery

Procedia PDF Downloads 286
25834 Destination Decision Model for Cruising Taxis Based on Embedding Model

Authors: Kazuki Kamada, Haruka Yamashita

Abstract:

In Japan, taxi is one of the popular transportations and taxi industry is one of the big businesses. However, in recent years, there has been a difficult problem of reducing the number of taxi drivers. In the taxi business, mainly three passenger catching methods are applied. One style is "cruising" that drivers catches passengers while driving on a road. Second is "waiting" that waits passengers near by the places with many requirements for taxies such as entrances of hospitals, train stations. The third one is "dispatching" that is allocated based on the contact from the taxi company. Above all, the cruising taxi drivers need the experience and intuition for finding passengers, and it is difficult to decide "the destination for cruising". The strong recommendation system for the cruising taxies supports the new drivers to find passengers, and it can be the solution for the decreasing the number of drivers in the taxi industry. In this research, we propose a method of recommending a destination for cruising taxi drivers. On the other hand, as a machine learning technique, the embedding models that embed the high dimensional data to a low dimensional space is widely used for the data analysis, in order to represent the relationship of the meaning between the data clearly. Taxi drivers have their favorite courses based on their experiences, and the courses are different for each driver. We assume that the course of cruising taxies has meaning such as the course for finding business man passengers (go around the business area of the city of go to main stations) and course for finding traveler passengers (go around the sightseeing places or big hotels), and extract the meaning of their destinations. We analyze the cruising history data of taxis based on the embedding model and propose the recommendation system for passengers. Finally, we demonstrate the recommendation of destinations for cruising taxi drivers based on the real-world data analysis using proposing method.

Keywords: taxi industry, decision making, recommendation system, embedding model

Procedia PDF Downloads 126
25833 The Impact of Misogyny on Women's Leadership in the Local Sphere of Government: The Case of Dr. Kenneth Kaunda District Municipality

Authors: Josephine Eghonghon Ahiante, Barry Hanyane

Abstract:

To give effect to the constitutional rights of gender equality, the South African government instituted various legislative policy frameworks and legislations to equalise the public service. Nonetheless, gender inequality in senior management positions remains a rift in government institutions, particularly the local sphere of government. The methodology for gathering and analysing data for this study was based on both primary and secondary data sources, namely literature review, qualitative and quantitative data collection and analysis, triangulation, and inductive and deductive thematic analysis. The study found that misogynist tendencies which are manifest in organisational culture suffocate the good intentions of government in ensuring social justices, leadership diversity, and women equality. It also demonstrates that traditional gender role expectation still informs the ground in which senior management positions are allocated, men perceive women as non-leadership fit and discriminate against them during recruitment, selection, and promotion into high positions. The analyses from the study portray that, while government legislation and framework has been instrumental in the leadership acceleration of women, much more has to be done to deconstruct internalised leadership stereotypes on women's gender roles and leadership requirements. The study recommends that gender bias training intervention is needed to teach public employees on management excellence.

Keywords: gender, leadership, misogyny, orgnisational cultural, patriachy

Procedia PDF Downloads 135
25832 Disclosure Extension of Oil and Gas Reserve Quantum

Authors: Ali Alsawayeh, Ibrahim Eldanfour

Abstract:

This paper examines the extent of disclosure of oil and gas reserve quantum in annual reports of international oil and gas exploration and production companies, particularly companies in untested international markets, such as Canada, the UK and the US, and seeks to determine the underlying factors that affect the level of disclosure on oil reserve quantum. The study is concerned with the usefulness of disclosure of oil and gas reserves quantum to investors and other users. Given the primacy of the annual report (10-k) as a source of supplemental reserves data about the company and as the channel through which companies disseminate information about their performance, the annual reports for one year (2009) were the central focus of the study. This comparative study seeks to establish whether differences exist between the sample companies, based on new disclosure requirements by the Securities and Exchange Commission (SEC) in respect of reserves classification and definition. The extent of disclosure of reserve is provided and compared among the selected companies. Statistical analysis is performed to determine whether any differences exist in the extent of disclosure of reserve under the determinant variables. This study shows that some factors would affect the extent of disclosure of reserve quantum in the above-mentioned countries, namely: company’s size, leverage and quality of auditor. Companies that provide reserves quantum in detail appear to display higher size. The findings also show that the level of leverage has affected companies’ reserves quantum disclosure. Indeed, companies that provide detailed reserves quantum disclosure tend to employ a ‘high-quality auditor’. In addition, the study found significant independent variable such as Profit Sharing Contracts (PSC). This factor could explain variations in the level of disclosure of oil reserve quantum between the contractor and host governments. The implementation of SEC oil and gas reporting requirements do not enhance companies’ valuation because the new rules are based only on past and present reserves information (proven reserves); hence, future valuation of oil and gas companies is missing for the market.

Keywords: comparison, company characteristics, disclosure, reserve quantum, regulation

Procedia PDF Downloads 391
25831 Laser-Dicing Modeling: Implementation of a High Accuracy Tool for Laser-Grooving and Cutting Application

Authors: Jeff Moussodji, Dominique Drouin

Abstract:

The highly complex technology requirements of today’s integrated circuits (ICs), lead to the increased use of several materials types such as metal structures, brittle and porous low-k materials which are used in both front end of line (FEOL) and back end of line (BEOL) process for wafer manufacturing. In order to singulate chip from wafer, a critical laser-grooving process, prior to blade dicing, is used to remove these layers of materials out of the dicing street. The combination of laser-grooving and blade dicing allows to reduce the potential risk of induced mechanical defects such micro-cracks, chipping, on the wafer top surface where circuitry is located. It seems, therefore, essential to have a fundamental understanding of the physics involving laser-dicing in order to maximize control of these critical process and reduce their undesirable effects on process efficiency, quality, and reliability. In this paper, the study was based on the convergence of two approaches, numerical and experimental studies which allowed us to investigate the interaction of a nanosecond pulsed laser and BEOL wafer materials. To evaluate this interaction, several laser grooved samples were compared with finite element modeling, in which three different aspects; phase change, thermo-mechanical and optic sensitive parameters were considered. The mathematical model makes it possible to highlight a groove profile (depth, width, etc.) of a single pulse or multi-pulses on BEOL wafer material. Moreover, the heat affected zone, and thermo-mechanical stress can be also predicted as a function of laser operating parameters (power, frequency, spot size, defocus, speed, etc.). After modeling validation and calibration, a satisfying correlation between experiment and modeling, results have been observed in terms of groove depth, width and heat affected zone. The study proposed in this work is a first step toward implementing a quick assessment tool for design and debug of multiple laser grooving conditions with limited experiments on hardware in industrial application. More correlations and validation tests are in progress and will be included in the full paper.

Keywords: laser-dicing, nano-second pulsed laser, wafer multi-stack, multiphysics modeling

Procedia PDF Downloads 191
25830 Structuring Highly Iterative Product Development Projects by Using Agile-Indicators

Authors: Guenther Schuh, Michael Riesener, Frederic Diels

Abstract:

Nowadays, manufacturing companies are faced with the challenge of meeting heterogeneous customer requirements in short product life cycles with a variety of product functions. So far, some of the functional requirements remain unknown until late stages of the product development. A way to handle these uncertainties is the highly iterative product development (HIP) approach. By structuring the development project as a highly iterative process, this method provides customer oriented and marketable products. There are first approaches for combined, hybrid models comprising deterministic-normative methods like the Stage-Gate process and empirical-adaptive development methods like SCRUM on a project management level. However, almost unconsidered is the question, which development scopes can preferably be realized with either empirical-adaptive or deterministic-normative approaches. In this context, a development scope constitutes a self-contained section of the overall development objective. Therefore, this paper focuses on a methodology that deals with the uncertainty of requirements within the early development stages and the corresponding selection of the most appropriate development approach. For this purpose, internal influencing factors like a company’s technology ability, the prototype manufacturability and the potential solution space as well as external factors like the market accuracy, relevance and volatility will be analyzed and combined into an Agile-Indicator. The Agile-Indicator is derived in three steps. First of all, it is necessary to rate each internal and external factor in terms of the importance for the overall development task. Secondly, each requirement has to be evaluated for every single internal and external factor appropriate to their suitability for empirical-adaptive development. Finally, the total sums of internal and external side are composed in the Agile-Indicator. Thus, the Agile-Indicator constitutes a company-specific and application-related criterion, on which the allocation of empirical-adaptive and deterministic-normative development scopes can be made. In a last step, this indicator will be used for a specific clustering of development scopes by application of the fuzzy c-means (FCM) clustering algorithm. The FCM-method determines sub-clusters within functional clusters based on the empirical-adaptive environmental impact of the Agile-Indicator. By means of the methodology presented in this paper, it is possible to classify requirements, which are uncertainly carried out by the market, into empirical-adaptive or deterministic-normative development scopes.

Keywords: agile, highly iterative development, agile-indicator, product development

Procedia PDF Downloads 229
25829 Transparency Obligations under the AI Act Proposal: A Critical Legal Analysis

Authors: Michael Lognoul

Abstract:

In April 2021, the European Commission released its AI Act Proposal, which is the first policy proposal at the European Union level to target AI systems comprehensively, in a horizontal manner. This Proposal notably aims to achieve an ecosystem of trust in the European Union, based on the respect of fundamental rights, regarding AI. Among many other requirements, the AI Act Proposal aims to impose several generic transparency obligationson all AI systems to the benefit of natural persons facing those systems (e.g. information on the AI nature of systems, in case of an interaction with a human). The Proposal also provides for more stringent transparency obligations, specific to AI systems that qualify as high-risk, to the benefit of their users, notably on the characteristics, capabilities, and limitations of the AI systems they use. Against that background, this research firstly presents all such transparency requirements in turn, as well as related obligations, such asthe proposed obligations on record keeping. Secondly, it focuses on a legal analysis of their scope of application, of the content of the obligations, and on their practical implications. On the scope of transparency obligations tailored for high-risk AI systems, the research notably notes that it seems relatively narrow, given the proposed legal definition of the notion of users of AI systems. Hence, where end-users do not qualify as users, they may only receive very limited information. This element might potentially raise concern regarding the objective of the Proposal. On the content of the transparency obligations, the research highlights that the information that should benefit users of high-risk AI systems is both very broad and specific, from a technical perspective. Therefore, the information required under those obligations seems to create, prima facie, an adequate framework to ensure trust for users of high-risk AI systems. However, on the practical implications of these transparency obligations, the research notes that concern arises due to potential illiteracy of high-risk AI systems users. They might not benefit from sufficient technical expertise to fully understand the information provided to them, despite the wording of the Proposal, which requires that information should be comprehensible to its recipients (i.e. users).On this matter, the research points that there could be, more broadly, an important divergence between the level of detail of the information required by the Proposal and the level of expertise of users of high-risk AI systems. As a conclusion, the research provides policy recommendations to tackle (part of) the issues highlighted. It notably recommends to broaden the scope of transparency requirements for high-risk AI systems to encompass end-users. It also suggests that principles of explanation, as they were put forward in the Guidelines for Trustworthy AI of the High Level Expert Group, should be included in the Proposal in addition to transparency obligations.

Keywords: aI act proposal, explainability of aI, high-risk aI systems, transparency requirements

Procedia PDF Downloads 277
25828 A Hierarchical Bayesian Calibration of Data-Driven Models for Composite Laminate Consolidation

Authors: Nikolaos Papadimas, Joanna Bennett, Amir Sakhaei, Timothy Dodwell

Abstract:

Composite modeling of consolidation processes is playing an important role in the process and part design by indicating the formation of possible unwanted prior to expensive experimental iterative trial and development programs. Composite materials in their uncured state display complex constitutive behavior, which has received much academic interest, and this with different models proposed. Errors from modeling and statistical which arise from this fitting will propagate through any simulation in which the material model is used. A general hyperelastic polynomial representation was proposed, which can be readily implemented in various nonlinear finite element packages. In our case, FEniCS was chosen. The coefficients are assumed uncertain, and therefore the distribution of parameters learned using Markov Chain Monte Carlo (MCMC) methods. In engineering, the approach often followed is to select a single set of model parameters, which on average, best fits a set of experiments. There are good statistical reasons why this is not a rigorous approach to take. To overcome these challenges, A hierarchical Bayesian framework was proposed in which population distribution of model parameters is inferred from an ensemble of experiments tests. The resulting sampled distribution of hyperparameters is approximated using Maximum Entropy methods so that the distribution of samples can be readily sampled when embedded within a stochastic finite element simulation. The methodology is validated and demonstrated on a set of consolidation experiments of AS4/8852 with various stacking sequences. The resulting distributions are then applied to stochastic finite element simulations of the consolidation of curved parts, leading to a distribution of possible model outputs. With this, the paper, as far as the authors are aware, represents the first stochastic finite element implementation in composite process modelling.

Keywords: data-driven , material consolidation, stochastic finite elements, surrogate models

Procedia PDF Downloads 132
25827 Design for Error-Proofing Assembly: A Systematic Approach to Prevent Assembly Issues since Early Design Stages, an Industrial Case Study

Authors: Gabriela Estrada, Joaquim Lloveras

Abstract:

Design for error-proofing assembly is a new DFX approach to prevent assembly issues since early design stages. Assembly issues that can happen during the life phases of a system such as: production, installation, operation, and replacement phases. This prevention is possible by designing the product with poka-yoke or error-proofing characteristics. This approach guide designers to make decisions based on poka-yoke assembly design requirements. As a result of applying these requirements designers are able to create solutions to prevent assembly issues for the product in development stage. This paper integrates the needs to design products in an error proofing way into the systematic approach of design process by Pahl and Beitz. A case study is presented applying this approach.

Keywords: poka-yoke, error-proofing, assembly issues, design process, life phases of a system

Procedia PDF Downloads 359
25826 Design for Error-Proofing Assembly: A Systematic Approach to Prevent Assembly Issues since Early Design Stages. An Industry Case Study

Authors: Gabriela Estrada, Joaquim Lloveras

Abstract:

Design for error-proofing assembly is a new DFX approach to prevent assembly issues since early design stages. Assembly issues that can happen during the life phases of a system such as: production, installation, operation and replacement phases. This prevention is possible by designing the product with poka-yoke or error-proofing characteristics. This approach guide designers to make decisions based on poka-yoke assembly design requirements. As a result of applying these requirements designers are able to create solutions to prevent assembly issues for the product in development stage. This paper integrates the needs to design products in an error proofing way into the systematic approach of design process by Pahl and Beitz. A case study is presented applying this approach.

Keywords: poka-yoke, error-proofing, assembly issues, design process, life phases of a system

Procedia PDF Downloads 301
25825 Realistic Testing Procedure of Power Swing Blocking Function in Distance Relay

Authors: Farzad Razavi, Behrooz Taheri, Mohammad Parpaei, Mehdi Mohammadi Ghalesefidi, Siamak Zarei

Abstract:

As one of the major problems in protecting large-dimension power systems, power swing and its effect on distance have caused a lot of damages to energy transfer systems in many parts of the world. Therefore, power swing has gained attentions of many researchers, which has led to invention of different methods for power swing detection. Power swing detection algorithm is highly important in distance relay, but protection relays should have general requirements such as correct fault detection, response rate, and minimization of disturbances in a power system. To ensure meeting the requirements, protection relays need different tests during development, setup, maintenance, configuration, and troubleshooting steps. This paper covers power swing scheme of the modern numerical relay protection, 7sa522 to address the effect of the different fault types on the function of the power swing blocking. In this study, it was shown that the different fault types during power swing cause different time for unblocking distance relay.

Keywords: power swing, distance relay, power system protection, relay test, transient in power system

Procedia PDF Downloads 363
25824 Regional Adjustment to the Analytical Attenuation Coefficient in the GMPM BSSA 14 for the Region of Spain

Authors: Gonzalez Carlos, Martinez Fransisco

Abstract:

There are various types of analysis that allow us to involve seismic phenomena that cause strong requirements for structures that are designed by society; one of them is a probabilistic analysis which works from prediction equations that have been created based on metadata seismic compiled in different regions. These equations form models that are used to describe the 5% damped pseudo spectra response for the various zones considering some easily known input parameters. The biggest problem for the creation of these models requires data with great robust statistics that support the results, and there are several places where this type of information is not available, for which the use of alternative methodologies helps to achieve adjustments to different models of seismic prediction.

Keywords: GMPM, 5% damped pseudo-response spectra, models of seismic prediction, PSHA

Procedia PDF Downloads 59
25823 Extension of Moral Agency to Artificial Agents

Authors: Sofia Quaglia, Carmine Di Martino, Brendan Tierney

Abstract:

Artificial Intelligence (A.I.) constitutes various aspects of modern life, from the Machine Learning algorithms predicting the stocks on Wall streets to the killing of belligerents and innocents alike on the battlefield. Moreover, the end goal is to create autonomous A.I.; this means that the presence of humans in the decision-making process will be absent. The question comes naturally: when an A.I. does something wrong when its behavior is harmful to the community and its actions go against the law, which is to be held responsible? This research’s subject matter in A.I. and Robot Ethics focuses mainly on Robot Rights and its ultimate objective is to answer the questions: (i) What is the function of rights? (ii) Who is a right holder, what is personhood and the requirements needed to be a moral agent (therefore, accountable for responsibility)? (iii) Can an A.I. be a moral agent? (ontological requirements) and finally (iv) if it ought to be one (ethical implications). With the direction to answer this question, this research project was done via a collaboration between the School of Computer Science in the Technical University of Dublin that oversaw the technical aspects of this work, as well as the Department of Philosophy in the University of Milan, who supervised the philosophical framework and argumentation of the project. Firstly, it was found that all rights are positive and based on consensus; they change with time based on circumstances. Their function is to protect the social fabric and avoid dangerous situations. The same goes for the requirements considered necessary to be a moral agent: those are not absolute; in fact, they are constantly redesigned. Hence, the next logical step was to identify what requirements are regarded as fundamental in real-world judicial systems, comparing them to that of ones used in philosophy. Autonomy, free will, intentionality, consciousness and responsibility were identified as the requirements to be considered a moral agent. The work went on to build a symmetrical system between personhood and A.I. to enable the emergence of the ontological differences between the two. Each requirement is introduced, explained in the most relevant theories of contemporary philosophy, and observed in its manifestation in A.I. Finally, after completing the philosophical and technical analysis, conclusions were drawn. As underlined in the research questions, there are two issues regarding the assignment of moral agency to artificial agent: the first being that all the ontological requirements must be present and secondly being present or not, whether an A.I. ought to be considered as an artificial moral agent. From an ontological point of view, it is very hard to prove that an A.I. could be autonomous, free, intentional, conscious, and responsible. The philosophical accounts are often very theoretical and inconclusive, making it difficult to fully detect these requirements on an experimental level of demonstration. However, from an ethical point of view it makes sense to consider some A.I. as artificial moral agents, hence responsible for their own actions. When considering artificial agents as responsible, there can be applied already existing norms in our judicial system such as removing them from society, and re-educating them, in order to re-introduced them to society. This is in line with how the highest profile correctional facilities ought to work. Noticeably, this is a provisional conclusion and research must continue further. Nevertheless, the strength of the presented argument lies in its immediate applicability to real world scenarios. To refer to the aforementioned incidents, involving the murderer of innocents, when this thesis is applied it is possible to hold an A.I. accountable and responsible for its actions. This infers removing it from society by virtue of its un-usability, re-programming it and, only when properly functioning, re-introducing it successfully

Keywords: artificial agency, correctional system, ethics, natural agency, responsibility

Procedia PDF Downloads 170
25822 Research on the Strategy of Whole-Life-Cycle Campus Design from the Perspective of Sustainable Concept: A Case Study on Hangzhou Senior High School in Zhejiang

Authors: Fan Yang

Abstract:

With the development of social economy and the popularization of quality education, the Chinese government invests more and more funding in education. Campus constructions are experiencing a great development phase. Under the trend of sustainable development, modern green campus design needs to meet new requirements of contemporary, informational and diversified education means and adapt to future education development. Educators, designers and other participants of campus design are facing new challenges. By studying and analyzing the universal unsatisfied current situations and sustainable development requirements of Chinese campuses, this paper summarizes the strategies and intentions of the whole-life-cycle campus design. In addition, a Chinese high school in Zhejiang province is added to illustrate the design cycle in an actual case. It is aimed to make all participants of campus design, especially the designers, to realize the importance of whole-life-cycle campus design and cooperate better. Sustainable campus design is expected to come true in deed instead of becoming a slogan in this way.

Keywords: campus design, green school, sustainable development, whole-life-cycle design

Procedia PDF Downloads 356
25821 Optimization for Autonomous Robotic Construction by Visual Guidance through Machine Learning

Authors: Yangzhi Li

Abstract:

Network transfer of information and performance customization is now a viable method of digital industrial production in the era of Industry 4.0. Robot platforms and network platforms have grown more important in digital design and construction. The pressing need for novel building techniques is driven by the growing labor scarcity problem and increased awareness of construction safety. Robotic approaches in construction research are regarded as an extension of operational and production tools. Several technological theories related to robot autonomous recognition, which include high-performance computing, physical system modeling, extensive sensor coordination, and dataset deep learning, have not been explored using intelligent construction. Relevant transdisciplinary theory and practice research still has specific gaps. Optimizing high-performance computing and autonomous recognition visual guidance technologies improves the robot's grasp of the scene and capacity for autonomous operation. Intelligent vision guidance technology for industrial robots has a serious issue with camera calibration, and the use of intelligent visual guiding and identification technologies for industrial robots in industrial production has strict accuracy requirements. It can be considered that visual recognition systems have challenges with precision issues. In such a situation, it will directly impact the effectiveness and standard of industrial production, necessitating a strengthening of the visual guiding study on positioning precision in recognition technology. To best facilitate the handling of complicated components, an approach for the visual recognition of parts utilizing machine learning algorithms is proposed. This study will identify the position of target components by detecting the information at the boundary and corner of a dense point cloud and determining the aspect ratio in accordance with the guidelines for the modularization of building components. To collect and use components, operational processing systems assign them to the same coordinate system based on their locations and postures. The RGB image's inclination detection and the depth image's verification will be used to determine the component's present posture. Finally, a virtual environment model for the robot's obstacle-avoidance route will be constructed using the point cloud information.

Keywords: robotic construction, robotic assembly, visual guidance, machine learning

Procedia PDF Downloads 69
25820 Factors of Divergence of Shari’Ah Supervisory Opinions and Its Effects on the Harmonization of Islamic Banking Products and Services

Authors: Dlir Abdullah Ahmed

Abstract:

Overall aims of this study are to investigate the effects of differences of opinions among Shari’ah supervisory bodies on standardization and internationalization of Islamic banking products and services. The study has used semi-structured in-depth interview where five respondents from both the Middle East and Malaysia Shari’ah advisors participated in the interview sessions. The data were analyzed by both manual and software techniques. The findings reveal that indeed there are differences of opinions among Shari’ah advisors in different jurisdictions. These differences are due to differences in educational background, schools of thoughts, environment in which they operate, and legal requirements. Moreover, the findings also reveal that these differences in opinions among Shari’ah bodies create confusions among public and bankers, and negatively affect standardization of Islamic banking transactions. In addition, the study has explored the possibility to develop Islamic-based products. However, the finding shows that it is difficult for the industry to have Islamic-based products due to high competition from conventional counterpart, legal constraints and moral hazard. Furthermore, the findings indicate that lack of political will and unity, lack of technology are the main constraints to internationalization of Islamic banking products. Last but not least, the study found that there are possibility of convergence of opinions, standardization of Islamic banking products and services if there are unified international Shari’ah h advisory council, international basic requirements for Islamic Shari’ah h advisors, and increase training and educations of Islamic bankers. This study has several implications to the bankers, policymakers and researchers. The policymakers should be able to resolve their political differences and set up unified international advisory council and international research and development center. The bankers should increase training and educations of the workforce as well improve on their banking infrastructure to facility cross-border transactions.

Keywords: Shari’ah h views, Islamic banking, products & services, standardization.

Procedia PDF Downloads 53
25819 Structure Design of Vacuum Vessel with Large Openings for Spacecraft Thermal Vacuum Test

Authors: Han Xiao, Ruan Qi, Zhang Lei, Qi Yan

Abstract:

Space environment simulator is a facility used to conduct thermal test for spacecraft, and vacuum vessel is the main body of it. According to the requirements for thermal tests of the spacecraft and its solar array panels, the primary vessel and the side vessels are designed to be a combinative structure connected with aperture, which ratio reaches 0.7. Since the vacuum vessel suffers 0.1MPa external pressure during the process of thermal test, in order to ensure the simulator’s reliability and safety, it’s necessary to calculate the vacuum vessel’s intensity and stability. Based on the impact of large openings to vacuum vessel structure, this paper explored the reinforce design and analytical way of vacuum vessel with large openings, using a large space environment simulator’s vacuum vessel design as an example. Tests showed that the reinforce structure is effective to fulfill the requirements of external pressure and the gravity. This ensured the reliability of the space environment simulator, providing a guarantee for developing the spacecraft.

Keywords: vacuum vessel, large opening, space environment simulator, structure design

Procedia PDF Downloads 509
25818 Control the Flow of Big Data

Authors: Shizra Waris, Saleem Akhtar

Abstract:

Big data is a research area receiving attention from academia and IT communities. In the digital world, the amounts of data produced and stored have within a short period of time. Consequently this fast increasing rate of data has created many challenges. In this paper, we use functionalism and structuralism paradigms to analyze the genesis of big data applications and its current trends. This paper presents a complete discussion on state-of-the-art big data technologies based on group and stream data processing. Moreover, strengths and weaknesses of these technologies are analyzed. This study also covers big data analytics techniques, processing methods, some reported case studies from different vendor, several open research challenges and the chances brought about by big data. The similarities and differences of these techniques and technologies based on important limitations are also investigated. Emerging technologies are suggested as a solution for big data problems.

Keywords: computer, it community, industry, big data

Procedia PDF Downloads 177
25817 Estimation of Pressure Profile and Boundary Layer Characteristics over NACA 4412 Airfoil

Authors: Anwar Ul Haque, Waqar Asrar, Erwin Sulaeman, Jaffar S. M. Ali

Abstract:

Pressure distribution data of the standard airfoils is usually used for the calibration purposes in subsonic wind tunnels. Results of such experiments are quite old and obtained by using the model in the spanwise direction. In this manuscript, pressure distribution over NACA 4412 airfoil model was presented by placing the 3D model in the lateral direction. The model is made of metal with pressure ports distributed longitudinally as well as in the lateral direction. The pressure model was attached to the floor of the tunnel with the help of the base plate to give the specified angle of attack to the model. Before the start of the experiments, the pressure tubes of the respective ports of the 128 ports pressure scanner are checked for leakage, and the losses due to the length of the pipes were also incorporated in the results for the specified pressure range. Growth rate maps of the boundary layer thickness were also plotted. It was found that with the increase in the velocity, the dynamic pressure distribution was also increased for the alpha seep. Plots of pressure distribution so obtained were overlapped with those obtained by using XFLR software, a low fidelity tool. It was found that at moderate and high angles of attack, the distribution of the pressure coefficients obtained from the experiments is high when compared with the XFLR ® results obtained along with the span of the wing. This under-prediction by XFLR ® is more obvious on the windward than on the leeward side.

Keywords: subsonic flow, boundary layer, wind tunnel, pressure testing

Procedia PDF Downloads 307
25816 Low-Cost Wireless Power Transfer System for Smart Recycling Containers

Authors: Juan Luis Leal, Rafael Maestre, Ovidio López

Abstract:

As innovation progresses, more possibilities are made available to increase the efficiency and reach of solutions for Smart Cities, most of which require the data provided by the Internet of Things (IoT) devices and may even have higher power requirements such as motors or actuators. A reliable power supply with the lowest maintenance is a requirement for the success of these solutions in the long term. Energy harvesting, mainly solar, becomes the solution of choice in most cases, but only if there is enough power to be harvested, which may depend on the device location (e.g., outdoors vs. indoor). This is the case of Smart Waste Containers with compaction systems, which have moderately high-power requirements, and may be installed in places with little sunlight for solar generation. It should be noted that waste is unloaded from the containers with cranes, so sudden and irregular movements may happen, making wired power unviable. In these cases, a wireless power supply may be a great alternative. This paper proposes a cost-effective two coil resonant wireless power transfer (WPT) system and describes its implementation, which has been carried out within an R&D project and validated in real settings with smart containers. Experimental results prove that the developed system achieves wireless power transmission up to 35W in the range of 5 cm to 1 m with a peak efficiency of 78%. The circuit is operated at relatively low resonant frequencies, which combined with enough wire-to-wire separation between the coil windings, reduce the losses caused by the proximity effect and, therefore, allow the use of common stranded wire instead of Litz wire, this without reducing the efficiency significantly. All these design considerations led to a final system that achieves a high efficiency for the desired charging range, simplifying the energy supply for Smart Containers as well as other devices that may benefit from a cost-effective wireless charging system.

Keywords: electromagnetic coupling, resonant wireless charging, smart recycling containers, wireless power transfer

Procedia PDF Downloads 77
25815 High Performance Computing and Big Data Analytics

Authors: Branci Sarra, Branci Saadia

Abstract:

Because of the multiplied data growth, many computer science tools have been developed to process and analyze these Big Data. High-performance computing architectures have been designed to meet the treatment needs of Big Data (view transaction processing standpoint, strategic, and tactical analytics). The purpose of this article is to provide a historical and global perspective on the recent trend of high-performance computing architectures especially what has a relation with Analytics and Data Mining.

Keywords: high performance computing, HPC, big data, data analysis

Procedia PDF Downloads 501
25814 Operating System Based Virtualization Models in Cloud Computing

Authors: Dev Ras Pandey, Bharat Mishra, S. K. Tripathi

Abstract:

Cloud computing is ready to transform the structure of businesses and learning through supplying the real-time applications and provide an immediate help for small to medium sized businesses. The ability to run a hypervisor inside a virtual machine is important feature of virtualization and it is called nested virtualization. In today’s growing field of information technology, many of the virtualization models are available, that provide a convenient approach to implement, but decision for a single model selection is difficult. This paper explains the applications of operating system based virtualization in cloud computing with an appropriate/suitable model with their different specifications and user’s requirements. In the present paper, most popular models are selected, and the selection was based on container and hypervisor based virtualization. Selected models were compared with a wide range of user’s requirements as number of CPUs, memory size, nested virtualization supports, live migration and commercial supports, etc. and we identified a most suitable model of virtualization.

Keywords: virtualization, OS based virtualization, container based virtualization, hypervisor based virtualization

Procedia PDF Downloads 295
25813 Identification of Location Parameters for Different User Types of the Inner-City Building Stock: An Austrian Example

Authors: Bernhard Bauer, Thomas Meixner, Amir Dini, Detlef Heck

Abstract:

The inner city building stock is characterized by different types of buildings of different decades and centuries and different types of historical constructions. Depending on the natural growth of a city, those types are often located in downtown areas and the surrounding suburbs. Since the population is becoming older and the variation of the different social requirements spread with the so-called 'Silver Society', city quarters have to be seen alternatively. If an area is very attractive for young students to live there because of the busy nightlife, it might not be suitable for the older society. To identify 'Location Types A, B, C' for different user groups, qualitative interviews with 24 citizens of the city of Graz (Austria) have been carried out, in order to identify the most important values for making a location or city quarter 'A', 'B', or 'C'. Furthermore these acknowledgements have been put into a softwaretool for predicting locations that are the most suitable for certain user groups. On the other hands side, investors or owners of buildings can use the tool for determining the most suitable user group for the location of their building or construction project in order to adapt the project or building stock to the requirements of the users.

Keywords: building stock, location parameters, inner city population, built environment

Procedia PDF Downloads 300
25812 Concurrent Engineering Challenges and Resolution Mechanisms from Quality Perspectives

Authors: Grmanesh Gidey Kahsay

Abstract:

In modern technical engineering applications, quality is defined in two ways. The first one is that quality is the parameter that measures a product or service’s characteristics to meet and satisfy the pre-stated or fundamental needs (reliability, durability, serviceability). The second one is the quality of a product or service free of any defect or deficiencies. The American Society for Quality (ASQ) describes quality as a pursuit of optimal solutions to confirm successes and fulfillment to be accountable for the product or service's requirements and expectations. This article focuses on quality engineering tools in modern industrial applications. Quality engineering is a field of engineering that deals with the principles, techniques, models, and applications of the product or service to guarantee quality. Including the entire activities to analyze the product’s design and development, quality engineering emphasizes how to make sure that products and services are designed and developed to meet consumers’ requirements. This episode acquaints with quality tools such as quality systems, auditing, product design, and process control. The finding presents thoughts that aim to improve quality engineering proficiency and effectiveness by introducing essential quality techniques and tools in some selected industries.

Keywords: essential quality tools, quality systems and models, quality management systems, and quality assurance

Procedia PDF Downloads 137
25811 Title: Real World Evidence a Tool to Overcome the Lack of a Comparative Arm in Drug Evaluation in the Context of Rare Diseases

Authors: Mohamed Wahba

Abstract:

Objective: To build a comparative arm for product (X) in specific gene mutated advanced gastrointestinal cancer using real world evidence to fulfill HTA requirements in drug evaluation. Methods: Data for product (X) were collected from phase II clinical trial while real world data for (Y) and (Z) were collected from US database. Real-world (RW) cohorts were matched to clinical trial base line characteristics using weighting by odds method. Outcomes included progression-free survival (PFS) and overall survival (OS) rates. Study location and participants: Internationally (product X, n=80) and from USA (Product Y and Z, n=73) Results: Two comparisons were made: trial cohort 1 (X) versus real-world cohort 1 (Z), trial cohort 2 (X) versus real-world cohort 2 (Y). For first line, the median OS was 9.7 months (95% CI 8.6- 11.5) and the median PFS was 5.2 months (95% CI 4.7- not reached) for real-world cohort 1. For second line, the median OS was 10.6 months (95% CI 4.7- 27.3) for real-world cohort 2 and the median PFS was 5.0 months (95% CI 2.1- 29.3). For OS analysis, results were statistically significant but not for PFS analysis. Conclusion: This study provided the clinical comparative outcomes needed for HTA evaluation.

Keywords: real world evidence, pharmacoeconomics, HTA agencies, oncology

Procedia PDF Downloads 76
25810 Examination of the Satisfaction Levels of Pre-Service Teachers Concerning E-Learning Process in Terms of Different Variables

Authors: Agah Tugrul Korucu

Abstract:

Significant changes have taken place for the better in the bulk of information and in the use of technology available in the field of education induced by technological changes in the 21st century. It is mainly the job of the teachers and pre-service teachers to integrate information and communication technologies into education by means of conveying the use of technology to individuals. While the pre-service teachers are conducting lessons by using technology, the methods they have developed are important factors for the requirements of the lesson and for the satisfaction levels of the students. The study of this study is to examine the satisfaction levels of pre-service teachers as regards e-learning in a technological environment in which there are lesson activities conducted through an online learning environment in terms of various variables. The study group of the research is composed of 156 pre-service teachers that were students in the departments of Computer and Teaching Technologies, Art Teaching and Pre-school Teaching in the academic year of 2014 - 2015. The qualitative research method was adopted for this study; the scanning model was employed in collecting the data. “The Satisfaction Scale regarding the E-learning Process”, developed by Gülbahar, and the personal information form, which was developed by the researcher, were used as means of collecting the data. Cronbach α reliability coefficient, which is the internal consistency coefficient of the scale, is 0.91. SPSS computerized statistical package program and the techniques of medium, standard deviation, percentage, correlation, t-test and variance analysis were used in the analysis of the data.

Keywords: online learning environment, integration of information technologies, e-learning, e-learning satisfaction, pre-service teachers

Procedia PDF Downloads 338
25809 Challenges in Promoting Software Usability and Applying Principles of Usage-Centred Design in Saudi Arabia

Authors: Kholod J. Alotaibi, Andrew M. Gravell

Abstract:

A study was conducted in which 212 software developers in higher education institutions in Saudi Arabia were surveyed to gather an indication of their understanding of the concept of usability, their acceptance of its importance, and to see how well its principles are applied. Interviews were then held with 20 of these developers, and a demonstration of Usage-Centred Design was attempted, a highly usability focused software development methodology, at one select institution for its redesign of an e-learning exam system interface during the requirements gathering phase. The study confirms the need to raise awareness of usability and its importance, and for Usage-Centred Design to be applied in its entirety, also need to encourage greater consultation with potential end-users of software and collaborative practices. The demonstration of Usage-Centred Design confirmed its ability to capture usability requirements more completely and precisely than would otherwise be the case, and hence its usefulness for developers concerned with improving software usability. The concluding discussion delves on the challenges for promoting usability and Usage-Centred Design in light of the research results and findings and recommendations are made for the same.

Keywords: usability, usage-centred, applying principles of usage-centred, Saudi Arabia

Procedia PDF Downloads 373
25808 A Study of Cloud Computing Solution for Transportation Big Data Processing

Authors: Ilgin Gökaşar, Saman Ghaffarian

Abstract:

The need for fast processed big data of transportation ridership (eg., smartcard data) and traffic operation (e.g., traffic detectors data) which requires a lot of computational power is incontrovertible in Intelligent Transportation Systems. Nowadays cloud computing is one of the important subjects and popular information technology solution for data processing. It enables users to process enormous measure of data without having their own particular computing power. Thus, it can also be a good selection for transportation big data processing as well. This paper intends to examine how the cloud computing can enhance transportation big data process with contrasting its advantages and disadvantages, and discussing cloud computing features.

Keywords: big data, cloud computing, Intelligent Transportation Systems, ITS, traffic data processing

Procedia PDF Downloads 443