Search results for: cloud service models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 10575

Search results for: cloud service models

7905 Special Case of Trip Distribution Model and Its Use for Estimation of Detailed Transport Demand in the Czech Republic

Authors: Jiri Dufek

Abstract:

The national model of the Czech Republic has been modified in a detailed way to get detailed travel demand in the municipality level (cities, villages over 300 inhabitants). As a technique for this detailed modelling, three-dimensional procedure for calibrating gravity models, was used. Besides of zone production and attraction, which is usual in gravity models, the next additional parameter for trip distribution was introduced. Usually it is called by “third dimension”. In the model, this parameter is a demand between regions. The distribution procedure involved calculation of appropriate skim matrices and its multiplication by three coefficients obtained by iterative balancing of production, attraction and third dimension. This type of trip distribution was processed in R-project and the results were used in the Czech Republic transport model, created in PTV Vision. This process generated more precise results in local level od the model (towns, villages)

Keywords: trip distribution, three dimension, transport model, municipalities

Procedia PDF Downloads 132
7904 Data-driven Decision-Making in Digital Entrepreneurship

Authors: Abeba Nigussie Turi, Xiangming Samuel Li

Abstract:

Data-driven business models are more typical for established businesses than early-stage startups that strive to penetrate a market. This paper provided an extensive discussion on the principles of data analytics for early-stage digital entrepreneurial businesses. Here, we developed data-driven decision-making (DDDM) framework that applies to startups prone to multifaceted barriers in the form of poor data access, technical and financial constraints, to state some. The startup DDDM framework proposed in this paper is novel in its form encompassing startup data analytics enablers and metrics aligning with startups' business models ranging from customer-centric product development to servitization which is the future of modern digital entrepreneurship.

Keywords: startup data analytics, data-driven decision-making, data acquisition, data generation, digital entrepreneurship

Procedia PDF Downloads 329
7903 The Relationship between Employee Commitment, Job Satisfaction and External Market Orientation in Vietnamese Joint-Stock Commercial Banks

Authors: Nguyen Ngoc Que Tran

Abstract:

Purpose: The purpose of this paper is to investigate the relationship between internal market orientation, external market orientation, employee commitment and job satisfaction. Design/methodology/approach: This study collected data through a survey and utilized simple linear regression and multiple regression analysis to determine if there was any support for the research hypotheses as presented in the previous chapter. Findings: Using data from 256 employees of four leading joint stock banks in Vietnam, the empirical results indicates that employee commitment is positively related with external market orientation, job satisfaction is positively related to employee commitment, and employee commitment and job satisfaction are positively related to external market orientation. However, job satisfaction has no significant positive effect on external market orientation. Theoretical contribution: The primary contribution to marketing theory arising from this study is the integration of job satisfaction, employee commitment, and external market orientation in a single research model. Practical implications: The major contribution to practice is an external market oriented bank has to respond rapidly to the future needs and preferences of its customers. This could result in high levels of commitment to the service process and in doing so provide Vietnamese joint-stock commercial banks with a competitive advantage. The finding is important for the banking service sector in general and the Vietnamese banking industry in particular.

Keywords: employee commitment, job satisfaction and external market orientation, vietnam, bank

Procedia PDF Downloads 417
7902 Native Language Identification with Cross-Corpus Evaluation Using Social Media Data: ’Reddit’

Authors: Yasmeen Bassas, Sandra Kuebler, Allen Riddell

Abstract:

Native language identification is one of the growing subfields in natural language processing (NLP). The task of native language identification (NLI) is mainly concerned with predicting the native language of an author’s writing in a second language. In this paper, we investigate the performance of two types of features; content-based features vs. content independent features, when they are evaluated on a different corpus (using social media data “Reddit”). In this NLI task, the predefined models are trained on one corpus (TOEFL), and then the trained models are evaluated on different data using an external corpus (Reddit). Three classifiers are used in this task; the baseline, linear SVM, and logistic regression. Results show that content-based features are more accurate and robust than content independent ones when tested within the corpus and across corpus.

Keywords: NLI, NLP, content-based features, content independent features, social media corpus, ML

Procedia PDF Downloads 139
7901 Assessment of Food Safety Culture in Select Restaurants and a Produce Market in Doha, Qatar

Authors: Ipek Goktepe, Israa Elnemr, Hammad Asim, Hao Feng, Mosbah Kushad, Hee Park, Sheikha Alzeyara, Mohammad Alhajri

Abstract:

Food safety management in Qatar is under the shared oversight of multiple agencies in two government ministries (Ministry of Public Health and Ministry of Municipality and Environment). Despite the increasing number and diversity of the food service establishments, no systematic food surveillance system is in place in the country, which creates a gap in terms of determining the food safety attitudes and practices applied in the food service operations. Therefore, this study seeks to partially address this gap through determination of food safety knowledge among food handlers, specifically with respect to food preparation and handling practices, and sanitation methods applied in food service providers (FSPs) and a major market in Doha, Qatar. The study covered a sample of 53 FSPs randomly selected out of 200 FSPs. Face-to-face interviews with managers at participating FSPs were conducted using a 40-questions survey. Additionally, 120 produce handlers who are in direct contact with fresh produce at the major produce market in Doha were surveyed using a questionnaire containing 21 questions. A written informed consent was obtained from each survey participant. The survey data were analyzed using the chi-square test and correlation test. The significance was evaluated at p ˂ 0.05. The results from the FSPs surveys indicated that the average age of FSPs was 11 years, with the oldest and newest being established in 1982 and 2015, respectively. Most managers (66%) had college degree and 68% of them were trained on the food safety management system known as HACCP. These surveys revealed that FSP managers’ training and education level were highly correlated with the probability of their employees receiving food safety training while managers with lower education level had no formal training on food safety for themselves nor for their employees. Casual sit-in and fine dine-in restaurants consistently kept records (100%), followed by fast food (36%), and catering establishments (14%). The produce handlers’ survey results showed that none of the workers had any training on safe produce handling practices. The majority of the workers were in the age range of 31-40 years (37%) and only 38% of them had high-school degree. Over 64% of produce handlers claimed to wash their hands 4-5 times per day but field observations pointed limited handwashing as there was soap in the settings. This observation suggests potential food safety risks since a significant correlation (p ˂ 0.01) between the educational level and the hand-washing practices was determined. This assessment on food safety culture through determination of food and produce handlers' level of knowledge and practices, the first of its kind in Qatar, demonstrated that training and education are important factors which directly impact the food safety culture in FSPs and produce markets. These findings should help in identifying the need for on-site training of food handlers for effective food safety practices in food establishments in Qatar.

Keywords: food safety, food safety culture, food service providers, food handlers

Procedia PDF Downloads 342
7900 Temperature Control Improvement of Membrane Reactor

Authors: Pornsiri Kaewpradit, Chalisa Pourneaw

Abstract:

Temperature control improvement of a membrane reactor with exothermic and reversible esterification reaction is studied in this work. It is well known that a batch membrane reactor requires different control strategies from a continuous one due to the fact that it is operated dynamically. Due to the effect of the operating temperature, the suitable control scheme has to be designed based reliable predictive model to achieve a desired objective. In the study, the optimization framework has been preliminary formulated in order to determine an optimal temperature trajectory for maximizing a desired product. In model predictive control scheme, a set of predictive models have been initially developed corresponding to the possible operating points of the system. The multiple predictive control moves have been further calculated on-line using the developed models corresponding to current operating point. It is obviously seen in the simulation results that the temperature control has been improved compared to the performance obtained by the conventional predictive controller. Further robustness tests have also been investigated in this study.

Keywords: model predictive control, batch reactor, temperature control, membrane reactor

Procedia PDF Downloads 470
7899 Hybrid Model: An Integration of Machine Learning with Traditional Scorecards

Authors: Golnush Masghati-Amoli, Paul Chin

Abstract:

Over the past recent years, with the rapid increases in data availability and computing power, Machine Learning (ML) techniques have been called on in a range of different industries for their strong predictive capability. However, the use of Machine Learning in commercial banking has been limited due to a special challenge imposed by numerous regulations that require lenders to be able to explain their analytic models, not only to regulators but often to consumers. In other words, although Machine Leaning techniques enable better prediction with a higher level of accuracy, in comparison with other industries, they are adopted less frequently in commercial banking especially for scoring purposes. This is due to the fact that Machine Learning techniques are often considered as a black box and fail to provide information on why a certain risk score is given to a customer. In order to bridge this gap between the explain-ability and performance of Machine Learning techniques, a Hybrid Model is developed at Dun and Bradstreet that is focused on blending Machine Learning algorithms with traditional approaches such as scorecards. The Hybrid Model maximizes efficiency of traditional scorecards by merging its practical benefits, such as explain-ability and the ability to input domain knowledge, with the deep insights of Machine Learning techniques which can uncover patterns scorecard approaches cannot. First, through development of Machine Learning models, engineered features and latent variables and feature interactions that demonstrate high information value in the prediction of customer risk are identified. Then, these features are employed to introduce observed non-linear relationships between the explanatory and dependent variables into traditional scorecards. Moreover, instead of directly computing the Weight of Evidence (WoE) from good and bad data points, the Hybrid Model tries to match the score distribution generated by a Machine Learning algorithm, which ends up providing an estimate of the WoE for each bin. This capability helps to build powerful scorecards with sparse cases that cannot be achieved with traditional approaches. The proposed Hybrid Model is tested on different portfolios where a significant gap is observed between the performance of traditional scorecards and Machine Learning models. The result of analysis shows that Hybrid Model can improve the performance of traditional scorecards by introducing non-linear relationships between explanatory and target variables from Machine Learning models into traditional scorecards. Also, it is observed that in some scenarios the Hybrid Model can be almost as predictive as the Machine Learning techniques while being as transparent as traditional scorecards. Therefore, it is concluded that, with the use of Hybrid Model, Machine Learning algorithms can be used in the commercial banking industry without being concerned with difficulties in explaining the models for regulatory purposes.

Keywords: machine learning algorithms, scorecard, commercial banking, consumer risk, feature engineering

Procedia PDF Downloads 137
7898 Constructing Evaluation Indicators for the Supply of Urban-Friendly Shelters from the Perspective of the Needs of the Elderly People in Taiwan

Authors: Chuan-Ming Tung, Tzu-Chiao Yuan

Abstract:

This research aims to construct the supply indicators and weights of shelter space from a perspective of the needs of the elderly by virtue of literature review, a systematical compilation of related regulations, and the use of the Analytical Hierarchy Process method, the questionnaires regarding the indicators filled out by 16 experts and scholars. The researcher then used 3 schools and 2 activity centers in Banqiao District, New Taipei City, as study cases to evaluate the ‘friendliness’ degree/level for the supply of shelters meeting the needs of elderly people. The supply evaluation indicators of friendly shelters meeting the needs of the elderly include "Administrative Operations and Service Needs" and "Residence-related and Living Needs"; under the "Administrative Operations and Service Needs" are "Management Operations and Information Provision", "Shelter Space Preparedness and Logistics Support", "Medical Care and Social Support", and "Shelters and Medical Environment", a total of 17 assessment items in four indicators, while under the "Residence-related and Living Needs" are "Dietary Needs", "Sleep Needs", "Hygiene and Sanitation Needs", "Accessibility and Convenience Needs ", etc., a total of 18 assessment items in four indicators. The results show that "Residence-related and Living Needs" is the most important item in the main levels of the supply indicators of the needs for friendly shelters to elderly people (weigh value 0.5504), followed by "Administrative Operations and Service Needs" (0.4496). The order of importance of the supply indicators of friendly shelters for the needs of elderly people is as follows: "Hygiene and Sanitation Needs" (0.1721), "Dietary Needs" (0.1340), "Medical Care and Social Support" (0.1300), "Sleep Needs" (0.1277), "Accessibility and Convenience Needs" (0.1166), "Basic Environment of Shelters" (0.1145), "Shelter Space Preparedness and Logistics Support" (0.1115) and "Management Operations and Information Provision" (0.0936). In addition, it can be noticed from the results of the case evaluation that the provision of refuges and shelters, mainly from schools and activity centers, is extremely inadequate for the needs of the elderly. In a set of comprehensive comparisons and contrasts, the evaluation indicators of refuges and shelters that need to be improved are "Medical Care and Social Support", "Hygiene and Sanitation Needs", "Sleep Needs", "Dietary Needs", and "Shelter Space Preparedness and Logistics Support".

Keywords: needs of the elderly people, urban shelters, evaluation indicators/indices., taiwan

Procedia PDF Downloads 81
7897 A Fast Calculation Approach for Position Identification in a Distance Space

Authors: Dawei Cai, Yuya Tokuda

Abstract:

The market of localization based service (LBS) is expanding. The acquisition of physical location is the fundamental basis for LBS. GPS, the de facto standard for outdoor localization, does not work well in indoor environment due to the blocking of signals by walls and ceiling. To acquire high accurate localization in an indoor environment, many techniques have been developed. Triangulation approach is often used for identifying the location, but a heavy and complex computation is necessary to calculate the location of the distances between the object and several source points. This computation is also time and power consumption, and not favorable to a mobile device that needs a long action life with battery. To provide a low power consumption approach for a mobile device, this paper presents a fast calculation approach to identify the location of the object without online solving solutions to simultaneous quadratic equations. In our approach, we divide the location identification into two parts, one is offline, and other is online. In offline mode, we make a mapping process that maps the location area to distance space and find a simple formula that can be used to identify the location of the object online with very light computation. The characteristic of the approach is a good tradeoff between the accuracy and computational amount. Therefore, this approach can be used in smartphone and other mobile devices that need a long work time. To show the performance, some simulation experimental results are provided also in the paper.

Keywords: indoor localization, location based service, triangulation, fast calculation, mobile device

Procedia PDF Downloads 175
7896 Analysis of a Discrete-time Geo/G/1 Queue Integrated with (s, Q) Inventory Policy at a Service Facility

Authors: Akash Verma, Sujit Kumar Samanta

Abstract:

This study examines a discrete-time Geo/G/1 queueing-inventory system attached with (s, Q) inventory policy. Assume that the customers follow the Bernoulli process on arrival. Each customer demands a single item with arbitrarily distributed service time. The inventory is replenished by an outside supplier, and the lead time for the replenishment is determined by a geometric distribution. There is a single server and infinite waiting space in this facility. Demands must wait in the specified waiting area during a stock-out period. The customers are served on a first-come-first-served basis. With the help of the embedded Markov chain technique, we determine the joint probability distributions of the number of customers in the system and the number of items in stock at the post-departure epoch using the Matrix Analytic approach. We relate the system length distribution at post-departure and outside observer's epochs to determine the joint probability distribution at the outside observer's epoch. We use probability distributions at random epochs to determine the waiting time distribution. We obtain the performance measures to construct the cost function. The optimum values of the order quantity and reordering point are found numerically for the variety of model parameters.

Keywords: discrete-time queueing inventory model, matrix analytic method, waiting-time analysis, cost optimization

Procedia PDF Downloads 46
7895 Water Quality Calculation and Management System

Authors: H. M. B. N Jayasinghe

Abstract:

The water is found almost everywhere on Earth. Water resources contain a lot of pollution. Some diseases can be spread through the water to the living beings. So to be clean water it should undergo a number of treatments necessary to make it drinkable. So it is must to have purification technology for the wastewater. So the waste water treatment plants act a major role in these issues. When considering the procedures taken after the water treatment process was always based on manual calculations and recordings. Water purification plants may interact with lots of manual processes. It means the process taking much time consuming. So the final evaluation and chemical, biological treatment process get delayed. So to prevent those types of drawbacks there are some computerized programmable calculation and analytical techniques going to be introduced to the laboratory staff. To solve this problem automated system will be a solution in which guarantees the rational selection. A decision support system is a way to model data and make quality decisions based upon it. It is widely used in the world for the various kind of process automation. Decision support systems that just collect data and organize it effectively are usually called passive models where they do not suggest a specific decision but only reveal information. This web base system is based on global positioning data adding facility with map location. Most worth feature is SMS and E-mail alert service to inform the appropriate person on a critical issue. The technological influence to the system is HTML, MySQL, PHP, and some other web developing technologies. Current issues in the computerized water chemistry analysis are not much deep in progress. For an example the swimming pool water quality calculator. The validity of the system has been verified by test running and comparison with an existing plant data. Automated system will make the life easier in productively and qualitatively.

Keywords: automated system, wastewater, purification technology, map location

Procedia PDF Downloads 249
7894 Optimizing Bridge Deck Construction: A Deep Neural Network Approach for Limiting Exterior Grider Rotation

Authors: Li Hui, Riyadh Hindi

Abstract:

In the United States, bridge construction often employs overhang brackets to support the deck overhang, the weight of fresh concrete, and loads from construction equipment. This approach, however, can lead to significant torsional moments on the exterior girders, potentially causing excessive girder rotation. Such rotations can result in various safety and maintenance issues, including thinning of the deck, reduced concrete cover, and cracking during service. Traditionally, these issues are addressed by installing temporary lateral bracing systems and conducting comprehensive torsional analysis through detailed finite element analysis for the construction of bridge deck overhang. However, this process is often intricate and time-intensive, with the spacing between temporary lateral bracing systems usually relying on the field engineers’ expertise. In this study, a deep neural network model is introduced to limit exterior girder rotation during bridge deck construction. The model predicts the optimal spacing between temporary bracing systems. To train this model, over 10,000 finite element models were generated in SAP2000, incorporating varying parameters such as girder dimensions, span length, and types and spacing of lateral bracing systems. The findings demonstrate that the deep neural network provides an effective and efficient alternative for limiting the exterior girder rotation for bridge deck construction. By reducing dependence on extensive finite element analyses, this approach stands out as a significant advancement in improving safety and maintenance effectiveness in the construction of bridge decks.

Keywords: bridge deck construction, exterior girder rotation, deep learning, finite element analysis

Procedia PDF Downloads 64
7893 Times Series Analysis of Depositing in Industrial Design in Brazil between 1996 and 2013

Authors: Jonas Pedro Fabris, Alberth Almeida Amorim Souza, Maria Emilia Camargo, Suzana Leitão Russo

Abstract:

With the law Nº. 9279, of May 14, 1996, the Brazilian government regulates rights and obligations relating to industrial property considering the economic development of the country as granting patents, trademark registration, registration of industrial designs and other forms of protection copyright. In this study, we show the application of the methodology of Box and Jenkins in the series of deposits of industrial design at the National Institute of Industrial Property for the period from May 1996 to April 2013. First, a graphical analysis of the data was done by observing the behavior of the data and the autocorrelation function. The best model found, based on the analysis of charts and statistical tests suggested by Box and Jenkins methodology, it was possible to determine the model number for the deposit of industrial design, SARIMA (2,1,0)(2,0,0), with an equal to 9.88% MAPE.

Keywords: ARIMA models, autocorrelation, Box and Jenkins Models, industrial design, MAPE, time series

Procedia PDF Downloads 546
7892 Identification of Nonlinear Systems Using Radial Basis Function Neural Network

Authors: C. Pislaru, A. Shebani

Abstract:

This paper uses the radial basis function neural network (RBFNN) for system identification of nonlinear systems. Five nonlinear systems are used to examine the activity of RBFNN in system modeling of nonlinear systems; the five nonlinear systems are dual tank system, single tank system, DC motor system, and two academic models. The feed forward method is considered in this work for modelling the non-linear dynamic models, where the K-Means clustering algorithm used in this paper to select the centers of radial basis function network, because it is reliable, offers fast convergence and can handle large data sets. The least mean square method is used to adjust the weights to the output layer, and Euclidean distance method used to measure the width of the Gaussian function.

Keywords: system identification, nonlinear systems, neural networks, radial basis function, K-means clustering algorithm

Procedia PDF Downloads 474
7891 Evaluating Traffic Congestion Using the Bayesian Dirichlet Process Mixture of Generalized Linear Models

Authors: Ren Moses, Emmanuel Kidando, Eren Ozguven, Yassir Abdelrazig

Abstract:

This study applied traffic speed and occupancy to develop clustering models that identify different traffic conditions. Particularly, these models are based on the Dirichlet Process Mixture of Generalized Linear regression (DML) and change-point regression (CR). The model frameworks were implemented using 2015 historical traffic data aggregated at a 15-minute interval from an Interstate 295 freeway in Jacksonville, Florida. Using the deviance information criterion (DIC) to identify the appropriate number of mixture components, three traffic states were identified as free-flow, transitional, and congested condition. Results of the DML revealed that traffic occupancy is statistically significant in influencing the reduction of traffic speed in each of the identified states. Influence on the free-flow and the congested state was estimated to be higher than the transitional flow condition in both evening and morning peak periods. Estimation of the critical speed threshold using CR revealed that 47 mph and 48 mph are speed thresholds for congested and transitional traffic condition during the morning peak hours and evening peak hours, respectively. Free-flow speed thresholds for morning and evening peak hours were estimated at 64 mph and 66 mph, respectively. The proposed approaches will facilitate accurate detection and prediction of traffic congestion for developing effective countermeasures.

Keywords: traffic congestion, multistate speed distribution, traffic occupancy, Dirichlet process mixtures of generalized linear model, Bayesian change-point detection

Procedia PDF Downloads 294
7890 The Role of Social Capital and Dynamic Capabilities in a Circular Economy: Evidence from German Small and Medium-Sized Enterprises

Authors: Antonia Hoffmann, Andrea Stübner

Abstract:

Resource scarcity and rising material prices are forcing companies to rethink their business models. The conventional linear system of economic growth and rising social needs further exacerbates the problem of resource scarcity. Therefore, it is necessary to separate economic growth from resource consumption. This can be achieved through the circular economy (CE), which focuses on sustainable product life cycles. However, companies face challenges in implementing CE into their businesses. Small and medium-sized enterprises are particularly affected by these problems, as they have a limited resource base. Collaboration and social interaction between different actors can help to overcome these obstacles. Based on a self-generated sample of 1,023 German small and medium-sized enterprises, we use a questionnaire to investigate the influence of social capital and its three dimensions - structural, relational, and cognitive capital - on the implementation of CE and the mediating effect of dynamic capabilities in explaining these relationships. Using regression analyses and structural equation modeling, we find that social capital is positively associated with CE implementation and dynamic capabilities partially mediate this relationship. Interestingly, our findings suggest that not all social capital dimensions are equally important for CE implementation. We theoretically and empirically explore the network forms of social capital and extend the CE literature by suggesting that dynamic capabilities help organizations leverage social capital to drive the implementation of CE practices. The findings of this study allow us to suggest several implications for managers and institutions. From a practical perspective, our study contributes to building circular production and service capabilities in small and medium-sized enterprises. Various CE activities can transform products and services to contribute to a better and more responsible world.

Keywords: circular economy, dynamic capabilities, SMEs, social capital

Procedia PDF Downloads 83
7889 Analyzing the Impact of Spatio-Temporal Climate Variations on the Rice Crop Calendar in Pakistan

Authors: Muhammad Imran, Iqra Basit, Mobushir Riaz Khan, Sajid Rasheed Ahmad

Abstract:

The present study investigates the space-time impact of climate change on the rice crop calendar in tropical Gujranwala, Pakistan. The climate change impact was quantified through the climatic variables, whereas the existing calendar of the rice crop was compared with the phonological stages of the crop, depicted through the time series of the Normalized Difference Vegetation Index (NDVI) derived from Landsat data for the decade 2005-2015. Local maxima were applied on the time series of NDVI to compute the rice phonological stages. Panel models with fixed and cross-section fixed effects were used to establish the relation between the climatic parameters and the time-series of NDVI across villages and across rice growing periods. Results show that the climatic parameters have significant impact on the rice crop calendar. Moreover, the fixed effect model is a significant improvement over cross-sectional fixed effect models (R-squared equal to 0.673 vs. 0.0338). We conclude that high inter-annual variability of climatic variables cause high variability of NDVI, and thus, a shift in the rice crop calendar. Moreover, inter-annual (temporal) variability of the rice crop calendar is high compared to the inter-village (spatial) variability. We suggest the local rice farmers to adapt this change in the rice crop calendar.

Keywords: Landsat NDVI, panel models, temperature, rainfall

Procedia PDF Downloads 205
7888 A Numerical Hybrid Finite Element Model for Lattice Structures Using 3D/Beam Elements

Authors: Ahmadali Tahmasebimoradi, Chetra Mang, Xavier Lorang

Abstract:

Thanks to the additive manufacturing process, lattice structures are replacing the traditional structures in aeronautical and automobile industries. In order to evaluate the mechanical response of the lattice structures, one has to resort to numerical techniques. Ansys is a globally well-known and trusted commercial software that allows us to model the lattice structures and analyze their mechanical responses using either solid or beam elements. In this software, a script may be used to systematically generate the lattice structures for any size. On the one hand, solid elements allow us to correctly model the contact between the substrates (the supports of the lattice structure) and the lattice structure, the local plasticity, and the junctions of the microbeams. However, their computational cost increases rapidly with the size of the lattice structure. On the other hand, although beam elements reduce the computational cost drastically, it doesn’t correctly model the contact between the lattice structures and the substrates nor the junctions of the microbeams. Also, the notion of local plasticity is not valid anymore. Moreover, the deformed shape of the lattice structure doesn’t correspond to the deformed shape of the lattice structure using 3D solid elements. In this work, motivated by the pros and cons of the 3D and beam models, a numerically hybrid model is presented for the lattice structures to reduce the computational cost of the simulations while avoiding the aforementioned drawbacks of the beam elements. This approach consists of the utilization of solid elements for the junctions and beam elements for the microbeams connecting the corresponding junctions to each other. When the global response of the structure is linear, the results from the hybrid models are in good agreement with the ones from the 3D models for body-centered cubic with z-struts (BCCZ) and body-centered cubic without z-struts (BCC) lattice structures. However, the hybrid models have difficulty to converge when the effect of large deformation and local plasticity are considerable in the BCCZ structures. Furthermore, the effect of the junction’s size of the hybrid models on the results is investigated. For BCCZ lattice structures, the results are not affected by the junction’s size. This is also valid for BCC lattice structures as long as the ratio of the junction’s size to the diameter of the microbeams is greater than 2. The hybrid model can take into account the geometric defects. As a demonstration, the point clouds of two lattice structures are parametrized in a platform called LATANA (LATtice ANAlysis) developed by IRT-SystemX. In this process, for each microbeam of the lattice structures, an ellipse is fitted to capture the effect of shape variation and roughness. Each ellipse is represented by three parameters; semi-major axis, semi-minor axis, and angle of rotation. Having the parameters of the ellipses, the lattice structures are constructed in Spaceclaim (ANSYS) using the geometrical hybrid approach. The results show a negligible discrepancy between the hybrid and 3D models, while the computational cost of the hybrid model is lower than the computational cost of the 3D model.

Keywords: additive manufacturing, Ansys, geometric defects, hybrid finite element model, lattice structure

Procedia PDF Downloads 114
7887 Multi-Model Super Ensemble Based Advanced Approaches for Monsoon Rainfall Prediction

Authors: Swati Bhomia, C. M. Kishtawal, Neeru Jaiswal

Abstract:

Traditionally, monsoon forecasts have encountered many difficulties that stem from numerous issues such as lack of adequate upper air observations, mesoscale nature of convection, proper resolution, radiative interactions, planetary boundary layer physics, mesoscale air-sea fluxes, representation of orography, etc. Uncertainties in any of these areas lead to large systematic errors. Global circulation models (GCMs), which are developed independently at different institutes, each of which carries somewhat different representation of the above processes, can be combined to reduce the collective local biases in space, time, and for different variables from different models. This is the basic concept behind the multi-model superensemble and comprises of a training and a forecast phase. The training phase learns from the recent past performances of models and is used to determine statistical weights from a least square minimization via a simple multiple regression. These weights are then used in the forecast phase. The superensemble forecasts carry the highest skill compared to simple ensemble mean, bias corrected ensemble mean and the best model out of the participating member models. This approach is a powerful post-processing method for the estimation of weather forecast parameters reducing the direct model output errors. Although it can be applied successfully to the continuous parameters like temperature, humidity, wind speed, mean sea level pressure etc., in this paper, this approach is applied to rainfall, a parameter quite difficult to handle with standard post-processing methods, due to its high temporal and spatial variability. The present study aims at the development of advanced superensemble schemes comprising of 1-5 day daily precipitation forecasts from five state-of-the-art global circulation models (GCMs), i.e., European Centre for Medium Range Weather Forecasts (Europe), National Center for Environmental Prediction (USA), China Meteorological Administration (China), Canadian Meteorological Centre (Canada) and U.K. Meteorological Office (U.K.) obtained from THORPEX Interactive Grand Global Ensemble (TIGGE), which is one of the most complete data set available. The novel approaches include the dynamical model selection approach in which the selection of the superior models from the participating member models at each grid and for each forecast step in the training period is carried out. Multi-model superensemble based on the training using similar conditions is also discussed in the present study, which is based on the assumption that training with the similar type of conditions may provide the better forecasts in spite of the sequential training which is being used in the conventional multi-model ensemble (MME) approaches. Further, a variety of methods that incorporate a 'neighborhood' around each grid point which is available in literature to allow for spatial error or uncertainty, have also been experimented with the above mentioned approaches. The comparison of these schemes with respect to the observations verifies that the newly developed approaches provide more unified and skillful prediction of the summer monsoon (viz. June to September) rainfall compared to the conventional multi-model approach and the member models.

Keywords: multi-model superensemble, dynamical model selection, similarity criteria, neighborhood technique, rainfall prediction

Procedia PDF Downloads 139
7886 The Impact of Window Opening Occupant Behavior Models on Building Energy Performance

Authors: Habtamu Tkubet Ebuy

Abstract:

Purpose Conventional dynamic energy simulation tools go beyond the static dimension of simplified methods by providing better and more accurate prediction of building performance. However, their ability to forecast actual performance is undermined by a low representation of human interactions. The purpose of this study is to examine the potential benefits of incorporating information on occupant diversity into occupant behavior models used to simulate building performance. The co-simulation of the stochastic behavior of the occupants substantially increases the accuracy of the simulation. Design/methodology/approach In this article, probabilistic models of the "opening and closing" behavior of the window of inhabitants have been developed in a separate multi-agent platform, SimOcc, and implemented in the building simulation, TRNSYS, in such a way that the behavior of the window with the interconnectivity can be reflected in the simulation analysis of the building. Findings The results of the study prove that the application of complex behaviors is important to research in predicting actual building performance. The results aid in the identification of the gap between reality and existing simulation methods. We hope this study and its results will serve as a guide for researchers interested in investigating occupant behavior in the future. Research limitations/implications Further case studies involving multi-user behavior for complex commercial buildings need to more understand the impact of the occupant behavior on building performance. Originality/value This study is considered as a good opportunity to achieve the national strategy by showing a suitable tool to help stakeholders in the design phase of new or retrofitted buildings to improve the performance of office buildings.

Keywords: occupant behavior, co-simulation, energy consumption, thermal comfort

Procedia PDF Downloads 107
7885 A Risk-Based Modeling Approach for Successful Adoption of CAATTs in Audits: An Exploratory Study Applied to Israeli Accountancy Firms

Authors: Alon Cohen, Jeffrey Kantor, Shalom Levy

Abstract:

Technology adoption models are extensively used in the literature to explore drivers and inhibitors affecting the adoption of Computer Assisted Audit Techniques and Tools (CAATTs). Further studies from recent years suggested additional factors that may affect technology adoption by CPA firms. However, the adoption of CAATTs by financial auditors differs from the adoption of technologies in other industries. This is a result of the unique characteristics of the auditing process, which are expressed in the audit risk elements and the risk-based auditing approach, as encoded in the auditing standards. Since these audit risk factors are not part of the existing models that are used to explain technology adoption, these models do not fully correspond to the specific needs and requirements of the auditing domain. The overarching objective of this qualitative research is to fill the gap in the literature, which exists as a result of using generic technology adoption models. Followed by a pretest and based on semi-structured in-depth interviews with 16 Israeli CPA firms of different sizes, this study aims to reveal determinants related to audit risk factors that influence the adoption of CAATTs in audits and proposes a new modeling approach for the successful adoption of CAATTs. The findings emphasize several important aspects: (1) while large CPA firms developed their own inner guidelines to assess the audit risk components, other CPA firms do not follow a formal and validated methodology to evaluate these risks; (2) large firms incorporate a variety of CAATTs, including self-developed advanced tools. On the other hand, small and mid-sized CPA firms incorporate standard CAATTs and still need to catch up to better understand what CAATTs can offer and how they can contribute to the quality of the audit; (3) the top management of mid-sized and small CPA firms should be more proactive and updated about CAATTs capabilities and contributions to audits; and (4) All CPA firms consider professionalism as a major challenge that must be constantly managed to ensure an optimal CAATTs operation. The study extends the existing knowledge of CAATTs adoption by looking at it from a risk-based auditing approach. It suggests a new model for CAATTs adoption by incorporating influencing audit risk factors that auditors should examine when considering CAATTs adoption. Since the model can be used in various audited scenarios and supports strategic, risk-based decisions, it maximizes the great potential of CAATTs on the quality of the audits. The results and insights can be useful to CPA firms, internal auditors, CAATTs developers and regulators. Moreover, it may motivate audit standard-setters to issue updated guidelines regarding CAATTs adoption in audits.

Keywords: audit risk, CAATTs, financial auditing, information technology, technology adoption models

Procedia PDF Downloads 70
7884 Numerical Modelling of Immiscible Fluids Flow in Oil Reservoir Rocks during Enhanced Oil Recovery Processes

Authors: Zahreddine Hafsi, Manoranjan Mishra , Sami Elaoud

Abstract:

Ensuring the maximum recovery rate of oil from reservoir rocks is a challenging task that requires preliminary numerical analysis of different techniques used to enhance the recovery process. After conventional oil recovery processes and in order to retrieve oil left behind after the primary recovery phase, water flooding in one of several techniques used for enhanced oil recovery (EOR). In this research work, EOR via water flooding is numerically modeled, and hydrodynamic instabilities resulted from immiscible oil-water flow in reservoir rocks are investigated. An oil reservoir is a porous medium consisted of many fractures of tiny dimensions. For modeling purposes, the oil reservoir is considered as a collection of capillary tubes which provides useful insights into how fluids behave in the reservoir pore spaces. Equations governing oil-water flow in oil reservoir rocks are developed and numerically solved following a finite element scheme. Numerical results are obtained using Comsol Multiphysics software. The two phase Darcy module of COMSOL Multiphysics allows modelling the imbibition process by the injection of water (as wetting phase) into an oil reservoir. Van Genuchten, Brooks Corey and Levrett models were considered as retention models and obtained flow configurations are compared, and the governing parameters are discussed. For the considered retention models it was found that onset of instabilities viz. fingering phenomenon is highly dependent on the capillary pressure as well as the boundary conditions, i.e., the inlet pressure and the injection velocity.

Keywords: capillary pressure, EOR process, immiscible flow, numerical modelling

Procedia PDF Downloads 133
7883 The Types of Collaboration Models Driven by Public Art Establishment–Case Study of Taichung City

Authors: Cheng-Lung Yu, Ying-His Liao

Abstract:

Some evidence show that public art accelerates local economic growth. Even local governments award the collaboration of public-private partnership to sustain the creation of public art for urban economic development. Through the public-private partnership of public art establishment it is obvious that public construction projects have been led by the governmental policy yet the private developers have played crucial roles to drive the innovative business models such as tourism investment, real estate value up and community participation. This study shows that the types of collaboration have been driven by Taichung city governmental policy from the regulation of public art establishment in the past three years. Through some cases empirical analyzes the authors discover the trends concerning the public art development to support local economic growth in Taiwan.

Keywords: public art, public art establishment regulation, construction management, urban governance

Procedia PDF Downloads 35
7882 WEMax: Virtual Manned Assembly Line Generation

Authors: Won Kyung Ham, Kang Hoon Cho, Sang C. Park

Abstract:

Presented in this paper is a framework of a software ‘WEMax’. The WEMax is invented for analysis and simulation for manned assembly lines to sustain and improve performance of manufacturing systems. In a manufacturing system, performance, such as productivity, is a key of competitiveness for output products. However, the manned assembly lines are difficult to forecast performance, because human labors are not expectable factors by computer simulation models or mathematical models. Existing approaches to performance forecasting of the manned assembly lines are limited to matters of the human itself, such as ergonomic and workload design, and non-human-factor-relevant simulation. Consequently, an approach for the forecasting and improvement of manned assembly line performance is needed to research. As a solution of the current problem, this study proposes a framework that is for generation and simulation of virtual manned assembly lines, and the framework has been implemented as a software.

Keywords: performance forecasting, simulation, virtual manned assembly line, WEMax

Procedia PDF Downloads 329
7881 Parametric Models of Facade Designs of High-Rise Residential Buildings

Authors: Yuchen Sharon Sung, Yingjui Tseng

Abstract:

High-rise residential buildings have become the most mainstream housing pattern in the world’s metropolises under the current trend of urbanization. The facades of high-rise buildings are essential elements of the urban landscape. The skins of these facades are important media between the interior and exterior of high- rise buildings. It not only connects between users and environments, but also plays an important functional and aesthetic role. This research involves a study of skins of high-rise residential buildings using the methodology of shape grammar to find out the rules which determine the combinations of the facade patterns and analyze the patterns’ parameters using software Grasshopper. We chose a number of facades of high-rise residential buildings as source to discover the underlying rules and concepts of the generation of facade skins. This research also provides the rules that influence the composition of facade skins. The items of the facade skins, such as windows, balconies, walls, sun visors and metal grilles are treated as elements in the system of facade skins. The compositions of these elements will be categorized and described by logical rules; and the types of high-rise building facade skins will be modelled by Grasshopper. Then a variety of analyzed patterns can also be applied on other facade skins through this parametric mechanism. Using these patterns established in the models, researchers can analyze each single item to do more detail tests and architects can apply each of these items to construct their facades for other buildings through various combinations and permutations. The goal of these models is to develop a mechanism to generate prototypes in order to facilitate generation of various facade skins.

Keywords: facade skin, grasshopper, high-rise residential building, shape grammar

Procedia PDF Downloads 512
7880 Classification of IoT Traffic Security Attacks Using Deep Learning

Authors: Anum Ali, Kashaf ad Dooja, Asif Saleem

Abstract:

The future smart cities trend will be towards Internet of Things (IoT); IoT creates dynamic connections in a ubiquitous manner. Smart cities offer ease and flexibility for daily life matters. By using small devices that are connected to cloud servers based on IoT, network traffic between these devices is growing exponentially, whose security is a concerned issue, since ratio of cyber attack may make the network traffic vulnerable. This paper discusses the latest machine learning approaches in related work further to tackle the increasing rate of cyber attacks, machine learning algorithm is applied to IoT-based network traffic data. The proposed algorithm train itself on data and identify different sections of devices interaction by using supervised learning which is considered as a classifier related to a specific IoT device class. The simulation results clearly identify the attacks and produce fewer false detections.

Keywords: IoT, traffic security, deep learning, classification

Procedia PDF Downloads 154
7879 Next-Gen Solutions: How Generative AI Will Reshape Businesses

Authors: Aishwarya Rai

Abstract:

This study explores the transformative influence of generative AI on startups, businesses, and industries. We will explore how large businesses can benefit in the area of customer operations, where AI-powered chatbots can improve self-service and agent effectiveness, greatly increasing efficiency. In marketing and sales, generative AI could transform businesses by automating content development, data utilization, and personalization, resulting in a substantial increase in marketing and sales productivity. In software engineering-focused startups, generative AI can streamline activities, significantly impacting coding processes and work experiences. It can be extremely useful in product R&D for market analysis, virtual design, simulations, and test preparation, altering old workflows and increasing efficiency. Zooming into the retail and CPG industry, industry findings suggest a 1-2% increase in annual revenues, equating to $400 billion to $660 billion. By automating customer service, marketing, sales, and supply chain management, generative AI can streamline operations, optimizing personalized offerings and presenting itself as a disruptive force. While celebrating economic potential, we acknowledge challenges like external inference and adversarial attacks. Human involvement remains crucial for quality control and security in the era of generative AI-driven transformative innovation. This talk provides a comprehensive exploration of generative AI's pivotal role in reshaping businesses, recognizing its strategic impact on customer interactions, productivity, and operational efficiency.

Keywords: generative AI, digital transformation, LLM, artificial intelligence, startups, businesses

Procedia PDF Downloads 78
7878 SIM (Subscriber Identity Module) Banking

Authors: Okanta Andrew, Richmond Kweku Frempong

Abstract:

As mobile networks are upgraded with technologies like WAP, GPRS and UMTS to deliver next-generation multimedia services, so are the banks and other financial institutions also getting ready to unleash the financial products on the mobile platform to meet growing demand for mobile based application services. Hence, the onset of Unstructured Supplementary Services (USSD) Banking which would make banking services available at anywhere, anytime through a string of interactive SMS sessions between a mobile device and an application server of a service provider. The aim of this studies was to find out whether the public will accept the sim banking service when it is implemented. Our target group includes: Working class. E. g. Businessmen/women, office workers, fishermen, market women, teachers etc. Nonworking class. E. g. Students (Tertiary, Senior High School), housewives. etc. The survey was in the form of a questionnaire and a verbal interview (video) which was to investigate their idea about the current banking system and the yet to be introduced sim banking concept. Meanwhile, some challenges accompanied the progression of data gathering because some populace showed reluctance in freeing their information. One other suggestion was that government should put measures against foremost challenges obstructing sim banking in Ghana counter to computers hackers. Government and individual have a key role to undertake to give suitable support to facelift the sim banking industry in the country. It was also suggested that Government put strong regulations on the use of sim banking products and services to streamline all the activities and also create awareness of the need for sim banking and emphasize its relevance in the aspect of national GDP.

Keywords: banking, mobile banking, SIM banking, mobile banking in Ghana

Procedia PDF Downloads 485
7877 Finite Element Modeling of Heat and Moisture Transfer in Porous Material

Authors: V. D. Thi, M. Li, M. Khelifa, M. El Ganaoui, Y. Rogaume

Abstract:

This paper presents a two-dimensional model to study the heat and moisture transfer through porous building materials. Dynamic and static coupled models of heat and moisture transfer in porous material under low temperature are presented and the coupled models together with variable initial and boundary conditions have been considered in an analytical way and using the finite element method. The resulting coupled model is converted to two nonlinear partial differential equations, which is then numerically solved by an implicit iterative scheme. The numerical results of temperature and moisture potential changes are compared with the experimental measurements available in the literature. Predicted results demonstrate validation of the theoretical model and effectiveness of the developed numerical algorithms. It is expected to provide useful information for the porous building material design based on heat and moisture transfer model.

Keywords: finite element method, heat transfer, moisture transfer, porous materials, wood

Procedia PDF Downloads 401
7876 Factors Impacting Geostatistical Modeling Accuracy and Modeling Strategy of Fluvial Facies Models

Authors: Benbiao Song, Yan Gao, Zhuo Liu

Abstract:

Geostatistical modeling is the key technic for reservoir characterization, the quality of geological models will influence the prediction of reservoir performance greatly, but few studies have been done to quantify the factors impacting geostatistical reservoir modeling accuracy. In this study, 16 fluvial prototype models have been established to represent different geological complexity, 6 cases range from 16 to 361 wells were defined to reproduce all those 16 prototype models by different methodologies including SIS, object-based and MPFS algorithms accompany with different constraint parameters. Modeling accuracy ratio was defined to quantify the influence of each factor, and ten realizations were averaged to represent each accuracy ratio under the same modeling condition and parameters association. Totally 5760 simulations were done to quantify the relative contribution of each factor to the simulation accuracy, and the results can be used as strategy guide for facies modeling in the similar condition. It is founded that data density, geological trend and geological complexity have great impact on modeling accuracy. Modeling accuracy may up to 90% when channel sand width reaches up to 1.5 times of well space under whatever condition by SIS and MPFS methods. When well density is low, the contribution of geological trend may increase the modeling accuracy from 40% to 70%, while the use of proper variogram may have very limited contribution for SIS method. It can be implied that when well data are dense enough to cover simple geobodies, few efforts were needed to construct an acceptable model, when geobodies are complex with insufficient data group, it is better to construct a set of robust geological trend than rely on a reliable variogram function. For object-based method, the modeling accuracy does not increase obviously as SIS method by the increase of data density, but kept rational appearance when data density is low. MPFS methods have the similar trend with SIS method, but the use of proper geological trend accompany with rational variogram may have better modeling accuracy than MPFS method. It implies that the geological modeling strategy for a real reservoir case needs to be optimized by evaluation of dataset, geological complexity, geological constraint information and the modeling objective.

Keywords: fluvial facies, geostatistics, geological trend, modeling strategy, modeling accuracy, variogram

Procedia PDF Downloads 266