Search results for: reduced order macro models
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 22900

Search results for: reduced order macro models

17110 The Role of Inventory Classification in Supply Chain Responsiveness in a Build-to-Order and Build-To-Forecast Manufacturing Environment: A Comparative Analysis

Authors: Qamar Iqbal

Abstract:

Companies strive to improve their forecasting methods to predict the fluctuations in customer demand. These fluctuation and variation in demand affect the manufacturing operations and can limit a company’s ability to fulfill customer demand on time. Companies keep the inventory buffer and maintain the stocking levels to reduce the impact of demand variation. A mid-size company deals with thousands of stock keeping units (skus). It is neither easy and nor efficient to control and manage each sku. Inventory classification provides a tool to the management to increase their ability to support customer demand. The paper presents a framework that shows how inventory classification can play a role to increase supply chain responsiveness. A case study will be presented to further elaborate the method both for build-to-order and build-to-forecast manufacturing environments. Results will be compared that will show which manufacturing setting has advantage over another under different circumstances. The outcome of this study is very useful to the management because this will give them an insight on how inventory classification can be used to increase their ability to respond to changing customer needs.

Keywords: inventory classification, supply chain responsiveness, forecast, manufacturing environment

Procedia PDF Downloads 597
17109 Application of FT-NIR Spectroscopy and Electronic Nose in On-line Monitoring of Dough Proofing

Authors: Madhuresh Dwivedi, Navneet Singh Deora, Aastha Deswal, H. N. Mishra

Abstract:

FT-NIR spectroscopy and electronic nose was used to study the kinetics of dough proofing. Spectroscopy was conducted with an optic probe in the diffuse reflectance mode. The dough leavening was carried out at different temperatures (25 and 35°C) and constant RH (80%). Spectra were collected in the range of wave numbers from 12,000 to 4,000 cm-1 directly on the samples, every 5 min during proofing, up to 2 hours. NIR spectra were corrected for scatter effect and second order derivatization was done to transform the spectra. Principal component analysis (PCA) was applied for the leavening process and process kinetics was calculated. PCA was performed on data set and loadings were calculated. For leavening, four absorption zones (8,950-8,850, 7,200-6,800, 5,250-5,150 and 4,700-4,250 cm-1) were involved in describing the process. Simultaneously electronic nose was also used for understanding the development of odour compounds during fermentation. The electronic nose was able to differential the sample on the basis of aroma generation at different time during fermentation. In order to rapidly differentiate samples based on odor, a Principal component analysis is performed and successfully demonstrated in this study. The result suggests that electronic nose and FT-NIR spectroscopy can be utilized for the online quality control of the fermentation process during leavening of bread dough.

Keywords: FT-NIR, dough, e-nose, proofing, principal component analysis

Procedia PDF Downloads 396
17108 Determination of Thermophysical Properties of Water Based Magnetic Nanofluids

Authors: Eyüphan Manay, Bayram Sahin, Emre Mandev, Ibrahim Ates, Tuba Yetim

Abstract:

In this study, it was aimed to determine the thermophysical properties of two different magnetic nanofluids (NiFe2O4-water and CoFe2O4-water). Magnetic nanoparticles were dispersed into the pure water at different volume fractions from 0 vol.% to 4 vol.%. The measurements were performed in the temperature range of 15 oC-55 oC. In order to get better idea on the temperature dependent thermophysical properties of magnetic nanofluids (MNFs), viscosity and thermal conductivity measurements were made. SEM images of both NiFe2O4 and CoFe2O4 nanoparticles were used in order to confirm the average dimensions. The measurements showed that the thermal conductivity of MNFs increased with an increase in the volume fraction as well as viscosity. Increase in the temperature of both MNFs resulted in an increase in the thermal conductivity and a decrease in the viscosity. Based on the measured data, the correlations for both the viscosity and the thermal conductivity were presented with respect to solid volume ratio and temperature. Effective thermal conductivity of the prepared MNFs was also calculated. The results indicated that water based NiFe2O4 nanofluid had higher thermal conductivity than that of the CoFe2O4. Once the viscosity values of both MNFs were compared, almost no difference was observed.

Keywords: magnetic nanofluids, thermal conductivity, viscosity, nife2o4-water, cofe2o4-water

Procedia PDF Downloads 266
17107 Different Tillage Possibilities for Second Crop in Green Bean Farming

Authors: Yilmaz Bayhan, Emin Güzel, Ömer Barış Özlüoymak, Ahmet İnce, Abdullah Sessiz

Abstract:

In this study, determining of reduced tillage techniques in green bean farming as a second crop after harvesting wheat was targeted. To this aim, four different soil tillage methods namely, heavy-duty disc harrow (HD), rotary tiller (ROT), heavy-duty disc harrow plus rotary tiller (HD+ROT) and no-tillage (NT) (seeding by direct drill) were examined. Experiments were arranged in a randomized block design with three replications. The highest green beans yields were obtained in HD+ROT and NT as 5,862.1 and 5,829.3 Mg/ha, respectively. The lowest green bean yield was found in HD as 3,076.7 Mg/ha. The highest fuel consumption was measured 30.60 L ha-1 for HD+ROT whereas the lowest value was found 7.50 L ha-1 for NT. No tillage method gave the best results for fuel consumption and effective power requirement. It is concluded that no-tillage method can be used in second crop green bean in the Thrace Region due to economic and erosion conditions.

Keywords: green bean, soil tillage, yield, vegetative

Procedia PDF Downloads 378
17106 Avifauna of Bara Gali Summer Campus, University of Peshawar, Khyber Pakhtunkhwa

Authors: Saif Ullah, Zaigham Hasan, Muhammad Ali, Qaisar Jamal, Kiran Salahuddin, Muhammad Awais

Abstract:

Survey of avian fauna of Bara Gali Summer Campus, University of Peshawar situated in Abbottabad was conducted from April to October, 2013. A total of 21 species belonging to 5 orders and 15 families were recorded. Out of these, 6 were resident, 12 summer visitor and 3 rare. Order Passeriformes was represented by 16 species which are Certhia himalayana, Megalaima virens, Phylloscopus trochiloides, Garrulax lineatus, Passer rutilans, Corvus macrorhynchos, Hypsipetes leucocephalus, Acridotheres tristis, Delichon dasypus cashmeriensis, Hirundo rustica, Muscicapa thalassina, Saxicola ferrea, Myiophoneus caeruleus, Parus melonolophus, Parus rufonuchalis, Parus monticolus, belonging to 11 families. Two species Dendrocopos himalayansis and Picus squamatus belong to only one family Picidae of order Piciformes. Among rest of the three orders each is represented by only a single species; Accipitriformes by Accipiter virgatus, Coraciformes by Upupa epops while order Psittaciformes has been represented by Psittacula himalayana. The distribution and abundance varied with season and maximum number of species were found during the monsoon season when most of the birds migrate for breeding. Some habits and behaviors like nesting, feeding, breeding and vocalizations were also studied which are very unique from other birds found at lower elevations. Among bird species adapted to diverse habitat in the field, Himalayan Jungle Crow, Common Mynas, Bulbuls, Barn Swallows, barbets were prominent. Interesting feature of the avian fauna is its familiarity with flora, was also observed during the present studies that some birds are very quick and active in their movement on a tree surface i.e Certhia himalayana.

Keywords: avifauna diversity, distribution, Bara Gali, Abbottabad

Procedia PDF Downloads 382
17105 Causes and Effects of Delays in Construction Projects in Akure, Ondo State, South-West Nigeria

Authors: K.T Alade, A.F Lawal, A.A Omonori

Abstract:

Construction is an everlasting activity across the globe. Likewise, the problem of delay in the construction industry is a global phenomenon. Although there are several reasons that may be responsible for delay during construction, this may vary from place to place and can be reduced to the minimum when identified. This study considered the major causes and effects of delay in the execution of construction projects in Akure, Ondo State, Nigeria. Using literatures, a total number of 30 causes of construction delays were identified. The convenient sampling technique was used to select sixty respondents for a survey. The respondents comprise twenty-two (22) clients, eighteen consultants (18) and twenty (20) contractors. The analyses of the primary data revealed that the three most important causes of delay in construction projects in Akure, Ondo State Nigeria are poor site management and supervision, inadequate contractors experience and client’s financial difficulties. Based on the findings of this study, recommendations were given on how the causes and effects of delay in construction can be mitigated.

Keywords: Akure, causes, construction projects, delay, effects

Procedia PDF Downloads 515
17104 Second Order Statistics of Dynamic Response of Structures Using Gamma Distributed Damping Parameters

Authors: Badreddine Chemali, Boualem Tiliouine

Abstract:

This article presents the main results of a numerical investigation on the uncertainty of dynamic response of structures with statistically correlated random damping Gamma distributed. A computational method based on a Linear Statistical Model (LSM) is implemented to predict second order statistics for the response of a typical industrial building structure. The significance of random damping with correlated parameters and its implications on the sensitivity of structural peak response in the neighborhood of a resonant frequency are discussed in light of considerable ranges of damping uncertainties and correlation coefficients. The results are compared to those generated using Monte Carlo simulation techniques. The numerical results obtained show the importance of damping uncertainty and statistical correlation of damping coefficients when obtaining accurate probabilistic estimates of dynamic response of structures. Furthermore, the effectiveness of the LSM model to efficiently predict uncertainty propagation for structural dynamic problems with correlated damping parameters is demonstrated.

Keywords: correlated random damping, linear statistical model, Monte Carlo simulation, uncertainty of dynamic response

Procedia PDF Downloads 284
17103 Modelling the Physicochemical Properties of Papaya Based-Cookies Using Response Surface Methodology

Authors: Mayowa Saheed Sanusi A, Musiliu Olushola Sunmonua, Abdulquadri Alakab Owolabi Raheema, Adeyemi Ikimot Adejokea

Abstract:

The development of healthy cookies for health-conscious consumers cannot be overemphasized in the present global health crisis. This study was aimed to evaluate and model the influence of ripeness levels of papaya puree (unripe, ripe and overripe), oven temperature (130°C, 150°C and 170°C) and oven rack speed (stationary, 10 and 20 rpm) on physicochemical properties of papaya-based cookies using Response Surface Methodology (RSM). The physicochemical properties (baking time, cookies mass, cookies thickness, spread ratio, proximate composition, Calcium, Vitamin C and Total Phenolic Content) were determined using standard procedures. The data obtained were statistically analysed at p≤0.05 using ANOVA. The polynomial regression model of response surface methodology was used to model the physicochemical properties. The adequacy of the models was determined using the coefficient of determination (R²) and the response optimizer of RSM was used to determine the optimum physicochemical properties for the papaya-based cookies. Cookies produced from overripe papaya puree were observed to have the shortest baking time; ripe papaya puree favors cookies spread ratio, while the unripe papaya puree gives cookies with the highest mass and thickness. The highest crude protein content, fiber content, calcium content, Vitamin C and Total Phenolic Content (TPC) were observed in papaya based-cookies produced from overripe puree. The models for baking time, cookies mass, cookies thickness, spread ratio, moisture content, crude protein and TPC were significant, with R2 ranging from 0.73 – 0.95. The optimum condition for producing papaya based-cookies with desirable physicochemical properties was obtained at 149°C oven temperature, 17 rpm oven rack speed and with the use of overripe papaya puree. The Information on the use of puree from unripe, ripe and overripe papaya can help to increase the use of underutilized unripe or overripe papaya and also serve as a strategic means of obtaining a fat substitute to produce new products with lower production cost and health benefit.

Keywords: papaya based-cookies, modeling, response surface methodology, physicochemical properties

Procedia PDF Downloads 173
17102 The Volume–Volatility Relationship Conditional to Market Efficiency

Authors: Massimiliano Frezza, Sergio Bianchi, Augusto Pianese

Abstract:

The relation between stock price volatility and trading volume represents a controversial issue which has received a remarkable attention over the past decades. In fact, an extensive literature shows a positive relation between price volatility and trading volume in the financial markets, but the causal relationship which originates such association is an open question, from both a theoretical and empirical point of view. In this regard, various models, which can be considered as complementary rather than competitive, have been introduced to explain this relationship. They include the long debated Mixture of Distributions Hypothesis (MDH); the Sequential Arrival of Information Hypothesis (SAIH); the Dispersion of Beliefs Hypothesis (DBH); the Noise Trader Hypothesis (NTH). In this work, we analyze whether stock market efficiency can explain the diversity of results achieved during the years. For this purpose, we propose an alternative measure of market efficiency, based on the pointwise regularity of a stochastic process, which is the Hurst–H¨older dynamic exponent. In particular, we model the stock market by means of the multifractional Brownian motion (mBm) that displays the property of a time-changing regularity. Mostly, such models have in common the fact that they locally behave as a fractional Brownian motion, in the sense that their local regularity at time t0 (measured by the local Hurst–H¨older exponent in a neighborhood of t0 equals the exponent of a fractional Brownian motion of parameter H(t0)). Assuming that the stock price follows an mBm, we introduce and theoretically justify the Hurst–H¨older dynamical exponent as a measure of market efficiency. This allows to measure, at any time t, markets’ departures from the martingale property, i.e. from efficiency as stated by the Efficient Market Hypothesis. This approach is applied to financial markets; using data for the SP500 index from 1978 to 2017, on the one hand we find that when efficiency is not accounted for, a positive contemporaneous relationship emerges and is stable over time. Conversely, it disappears as soon as efficiency is taken into account. In particular, this association is more pronounced during time frames of high volatility and tends to disappear when market becomes fully efficient.

Keywords: volume–volatility relationship, efficient market hypothesis, martingale model, Hurst–Hölder exponent

Procedia PDF Downloads 83
17101 On the Performance of Improvised Generalized M-Estimator in the Presence of High Leverage Collinearity Enhancing Observations

Authors: Habshah Midi, Mohammed A. Mohammed, Sohel Rana

Abstract:

Multicollinearity occurs when two or more independent variables in a multiple linear regression model are highly correlated. The ridge regression is the commonly used method to rectify this problem. However, the ridge regression cannot handle the problem of multicollinearity which is caused by high leverage collinearity enhancing observation (HLCEO). Since high leverage points (HLPs) are responsible for inducing multicollinearity, the effect of HLPs needs to be reduced by using Generalized M estimator. The existing GM6 estimator is based on the Minimum Volume Ellipsoid (MVE) which tends to swamp some low leverage points. Hence an improvised GM (MGM) estimator is presented to improve the precision of the GM6 estimator. Numerical example and simulation study are presented to show how HLPs can cause multicollinearity. The numerical results show that our MGM estimator is the most efficient method compared to some existing methods.

Keywords: identification, high leverage points, multicollinearity, GM-estimator, DRGP, DFFITS

Procedia PDF Downloads 265
17100 Effect of Contaminants on the Behavior of Shallow Foundations

Authors: Ghazal Horiat, Alireza Hajiani Bushehrian

Abstract:

leakage of contamination from fuel or oil reservoirs can alter the geotechnical properties of the soil under their foundation and finally affect their performance in their service life. This article investigates the behavior of shallow foundations on the soil contaminated with diesel and kerosene using the Plaxis Tunnel3D V1.2 software. The information required for the numerical modeling in the paper was obtained from a similar experimental study. The present study seeks to compare the behavior of square foundations on sandy soil without contamination and the soil contaminated with different percentages of diesel and crude oil. The study was conducted on a small square foundation. The depth of the contamination was assumed constant, and the soil was evaluated with four different percentages of both contaminants. The results of analyses were plotted and assessed in the form of load-displacement curves for the foundation. The results indicate reduced bearing capacity of the foundation with the rise in the contamination percentage.

Keywords: bearing capacity, contaminated soils, shallow foundations, 3D numerical analysis

Procedia PDF Downloads 148
17099 Simulation of Heat Exchanger Behavior during LOCA Accident in THTL Test Loop

Authors: R. Mahmoodi, A. R. Zolfaghari

Abstract:

In nuclear power plants, loss of coolant from the primary system is the type of reduced removed capacity that is given most attention; such an accident is referred as Loss of Coolant Accident (LOCA). In the current study, investigation of shell and tube THTL heat exchanger behavior during LOCA is implemented by ANSYS CFX simulation software in both steady state and transient mode of turbulent fluid flow according to experimental conditions. Numerical results obtained from ANSYS CFX simulation show good agreement with experimental data of THTL heat exchanger. The results illustrate that in large break LOCA as short term accident, heat exchanger could not fast response to temperature variables but in the long term, the temperature of shell side of heat exchanger will be increase.

Keywords: shell-and-tube heat exchanger, shell-side, CFD, flow and heat transfer, LOCA

Procedia PDF Downloads 444
17098 Measuring Human Perception and Negative Elements of Public Space Quality Using Deep Learning: A Case Study of Area within the Inner Road of Tianjin City

Authors: Jiaxin Shi, Kaifeng Hao, Qingfan An, Zeng Peng

Abstract:

Due to a lack of data sources and data processing techniques, it has always been difficult to quantify public space quality, which includes urban construction quality and how it is perceived by people, especially in large urban areas. This study proposes a quantitative research method based on the consideration of emotional health and physical health of the built environment. It highlights the low quality of public areas in Tianjin, China, where there are many negative elements. Deep learning technology is then used to measure how effectively people perceive urban areas. First, this work suggests a deep learning model that might simulate how people can perceive the quality of urban construction. Second, we perform semantic segmentation on street images to identify visual elements influencing scene perception. Finally, this study correlated the scene perception score with the proportion of visual elements to determine the surrounding environmental elements that influence scene perception. Using a small-scale labeled Tianjin street view data set based on transfer learning, this study trains five negative spatial discriminant models in order to explore the negative space distribution and quality improvement of urban streets. Then it uses all Tianjin street-level imagery to make predictions and calculate the proportion of negative space. Visualizing the spatial distribution of negative space along the Tianjin Inner Ring Road reveals that the negative elements are mainly found close to the five key districts. The map of Tianjin was combined with the experimental data to perform the visual analysis. Based on the emotional assessment, the distribution of negative materials, and the direction of street guidelines, we suggest guidance content and design strategy points of the negative phenomena in Tianjin street space in the two dimensions of perception and substance. This work demonstrates the utilization of deep learning techniques to understand how people appreciate high-quality urban construction, and it complements both theory and practice in urban planning. It illustrates the connection between human perception and the actual physical public space environment, allowing researchers to make urban interventions.

Keywords: human perception, public space quality, deep learning, negative elements, street images

Procedia PDF Downloads 121
17097 Spawning Induction and Early Larval Development of the Penshell Atrina maura (Sowerby, 1835) under Controlled Conditions in Ecuador

Authors: Jose Melena, Rosa Santander, Tanya Gonzalez, Richard Duque, Juan Illanes

Abstract:

Ecuador is one of the countries with the greatest aquatic biodiversity worldwide. In particular, there are at least a dozen native marine species with great aquaculture potential locally. This research concerns one of those species. It has proposed to implement experimental protocols in order to induce spawning and to generate the early larval development of the penshell Atrina maura under controlled conditions. Bioassays were carried out with one adult batch (n= 26) with an average valvar length of 307,6 ± 9,4 mm, which were collected in the Puerto El Morro Mangrove (2° 42' 33'' S, 80° 14' 28'' W), Guayas Province. During a short acclimation stage, five adults of penshell A. maura were sacrificed in order to determine their sexual maturity degree and to estimate their sex ratio. Dissection showed that three were ripe females (60%) and two were ripe males (40%). Later, three groups (n= 7 by each) were tested with two treatments in order to induce the broodstock spawning: thermal stress, osmotic shock, and one control. Spawning induction was achieved by the immersion in water to 0 g L⁻¹ per 1 h and immersion in sea water to 34 g L⁻¹ per 1 h. After the delivery of gametes, it was achieved 1,35 × 10⁶ viable zygotes. As results, fertilized eggs had 60 µm diameter; while first and second cell divisions were observed to 1 h post-fertilization, with individual average length of 65 ± 4 µm and polar body. Latter cell divisions, including gastrula stage, appeared at 9 h post-fertilization, with individual average length of 71 ± 4 µm; and trochophore stage at 16 h post-fertilization with individual average length of 75 ± 5 µm. In addition, veliger stage was registered at 20 h post-fertilization with individual average length of 81 ± 5 µm. Umboned larvae appeared at day 8 post-fertilization, with individual average length of 145 ± 6 µm. These pioneering results in Ecuador can strengthen the local conservation process of the overexploited A. maura and to encourage its production for commercial purposes.

Keywords: Atrina maura, Ecuador, larval development, spawning induction

Procedia PDF Downloads 165
17096 Development of an Automatic Calibration Framework for Hydrologic Modelling Using Approximate Bayesian Computation

Authors: A. Chowdhury, P. Egodawatta, J. M. McGree, A. Goonetilleke

Abstract:

Hydrologic models are increasingly used as tools to predict stormwater quantity and quality from urban catchments. However, due to a range of practical issues, most models produce gross errors in simulating complex hydraulic and hydrologic systems. Difficulty in finding a robust approach for model calibration is one of the main issues. Though automatic calibration techniques are available, they are rarely used in common commercial hydraulic and hydrologic modelling software e.g. MIKE URBAN. This is partly due to the need for a large number of parameters and large datasets in the calibration process. To overcome this practical issue, a framework for automatic calibration of a hydrologic model was developed in R platform and presented in this paper. The model was developed based on the time-area conceptualization. Four calibration parameters, including initial loss, reduction factor, time of concentration and time-lag were considered as the primary set of parameters. Using these parameters, automatic calibration was performed using Approximate Bayesian Computation (ABC). ABC is a simulation-based technique for performing Bayesian inference when the likelihood is intractable or computationally expensive to compute. To test the performance and usefulness, the technique was used to simulate three small catchments in Gold Coast. For comparison, simulation outcomes from the same three catchments using commercial modelling software, MIKE URBAN were used. The graphical comparison shows strong agreement of MIKE URBAN result within the upper and lower 95% credible intervals of posterior predictions as obtained via ABC. Statistical validation for posterior predictions of runoff result using coefficient of determination (CD), root mean square error (RMSE) and maximum error (ME) was found reasonable for three study catchments. The main benefit of using ABC over MIKE URBAN is that ABC provides a posterior distribution for runoff flow prediction, and therefore associated uncertainty in predictions can be obtained. In contrast, MIKE URBAN just provides a point estimate. Based on the results of the analysis, it appears as though ABC the developed framework performs well for automatic calibration.

Keywords: automatic calibration framework, approximate bayesian computation, hydrologic and hydraulic modelling, MIKE URBAN software, R platform

Procedia PDF Downloads 312
17095 Characterizing the Rectification Process for Designing Scoliosis Braces: Towards Digital Brace Design

Authors: Inigo Sanz-Pena, Shanika Arachchi, Dilani Dhammika, Sanjaya Mallikarachchi, Jeewantha S. Bandula, Alison H. McGregor, Nicolas Newell

Abstract:

The use of orthotic braces for adolescent idiopathic scoliosis (AIS) patients is the most common non-surgical treatment to prevent deformity progression. The traditional method to create an orthotic brace involves casting the patient’s torso to obtain a representative geometry, which is then rectified by an orthotist to the desired geometry of the brace. Recent improvements in 3D scanning technologies, rectification software, CNC, and additive manufacturing processes have given the possibility to compliment, or in some cases, replace manual methods with digital approaches. However, the rectification process remains dependent on the orthotist’s skills. Therefore, the rectification process needs to be carefully characterized to ensure that braces designed through a digital workflow are as efficient as those created using a manual process. The aim of this study is to compare 3D scans of patients with AIS against 3D scans of both pre- and post-rectified casts that have been manually shaped by an orthotist. Six AIS patients were recruited from the Ragama Rehabilitation Clinic, Colombo, Sri Lanka. All patients were between 10 and 15 years old, were skeletally immature (Risser grade 0-3), and had Cobb angles between 20-45°. Seven spherical markers were placed at key anatomical locations on each patient’s torso and on the pre- and post-rectified molds so that distances could be reliably measured. 3D scans were obtained of 1) the patient’s torso and pelvis, 2) the patient’s pre-rectification plaster mold, and 3) the patient’s post-rectification plaster mold using a Structure Sensor Mark II 3D scanner (Occipital Inc., USA). 3D stick body models were created for each scan to represent the distances between anatomical landmarks. The 3D stick models were used to analyze the changes in position and orientation of the anatomical landmarks between scans using Blender open-source software. 3D Surface deviation maps represented volume differences between the scans using CloudCompare open-source software. The 3D stick body models showed changes in the position and orientation of thorax anatomical landmarks between the patient and the post-rectification scans for all patients. Anatomical landmark position and volume differences were seen between 3D scans of the patient’s torsos and the pre-rectified molds. Between the pre- and post-rectified molds, material removal was consistently seen on the anterior side of the thorax and the lateral areas below the ribcage. Volume differences were seen in areas where the orthotist planned to place pressure pads (usually at the trochanter on the side to which the lumbar curve was tilted (trochanter pad), at the lumbar apical vertebra (lumbar pad), on the rib connected to the apical vertebrae at the mid-axillary line (thoracic pad), and on the ribs corresponding to the upper thoracic vertebra (axillary extension pad)). The rectification process requires the skill and experience of an orthotist; however, this study demonstrates that the brace shape, location, and volume of material removed from the pre-rectification mold can be characterized and quantified. Results from this study can be fed into software that can accelerate the brace design process and make steps towards the automated digital rectification process.

Keywords: additive manufacturing, orthotics, scoliosis brace design, sculpting software, spinal deformity

Procedia PDF Downloads 151
17094 Analysis on the Copyright Protection Dilemma of Webcast in 'Internet Plus' Era

Authors: Yi Yang

Abstract:

In the era of 'Internet plus', the rapid development of webcast has posed new challenges to the intellectual property law. Meanwhile, traditional copyright protection has also exposed the existing theoretical imbalance in webcast. Through the analysis of the outstanding problems in the copyright protection of the network live broadcast, this paper points out that the main causes of the problems are the unclear nature of the copyright of the network live broadcast, the copyright protection system of the game network live broadcast has not yet been constructed, and the copyright infringement of the pan entertainment live broadcast is mostly, and so on. Based on the current practice, this paper puts forward the specific thinking of the protection path of online live broadcast copyright. First of all, to provide a reasonable judicial solution for a large number of online live copyright cases, we need to integrate the right scope and regulatory behavior of broadcasting right and information network communication right. Secondly, in order to protect the rights of network anchors, the webcast should be regarded as works. Thirdly, in order to protect the copyright of webcast and prevent the infringement of copyright by webcast, the webcast platform will be used as an intermediary to provide solutions for solving the judicial dilemma. In the era of 'Internet plus', it is a theoretical attempt to explore the protection and method of copyright protection on webcast, which has positive guiding significance for judicial practice.

Keywords: 'Internet Plus' era, webcast, copyright, protection dilemma

Procedia PDF Downloads 116
17093 A Two-Step Framework for Unsupervised Speaker Segmentation Using BIC and Artificial Neural Network

Authors: Ahmad Alwosheel, Ahmed Alqaraawi

Abstract:

This work proposes a new speaker segmentation approach for two speakers. It is an online approach that does not require a prior information about speaker models. It has two phases, a conventional approach such as unsupervised BIC-based is utilized in the first phase to detect speaker changes and train a Neural Network, while in the second phase, the output trained parameters from the Neural Network are used to predict next incoming audio stream. Using this approach, a comparable accuracy to similar BIC-based approaches is achieved with a significant improvement in terms of computation time.

Keywords: artificial neural network, diarization, speaker indexing, speaker segmentation

Procedia PDF Downloads 510
17092 Digitization and Economic Growth in Africa: The Role of Financial Sector Development

Authors: Abdul Ganiyu Iddrisu, Bei Chen

Abstract:

Digitization is the process of transforming analog material into digital form, especially for storage and use in a computer. Significant development of information and communication technology (ICT) over the past years has encouraged many researchers to investigate its contribution to promoting economic growth and reducing poverty. Yet the compelling empirical evidence on the effects of digitization on economic growth remains weak, particularly in Africa. This is because extant studies that explicitly evaluate digitization and economic growth nexus are mostly reports and desk reviews. This points out an empirical knowledge gap in the literature. Hypothetically, digitization influences financial sector development which in turn influences economic growth. Digitization has changed the financial sector and its operating environment. Obstacles to access to financing, for instance, physical distance, minimum balance requirements, and low-income flows, among others can be circumvented. Savings have increased, micro-savers have opened bank accounts, and banks are now able to price short-term loans. This has the potential to develop the financial sector. However, empirical evidence on the digitization-financial development nexus is dearth. On the other hand, a number of studies maintained that financial sector development greatly influences growth of economies. We, therefore, argue that financial sector development is one of the transmission mechanisms through which digitization affects economic growth. Employing macro-country-level data from African countries and using fixed effects, random effects and Hausman-Taylor estimation approaches, this paper contributes to the literature by analysing economic growth in Africa, focusing on the role of digitization and financial sector development. First, we assess how digitization influences financial sector development in Africa. From an economic policy perspective, it is important to identify digitization determinants of financial sector development so that action can be taken to reduce the economic shocks associated with financial sector distortions. This nexus is rarely examined empirically in the literature. Secondly, we examine the effect of domestic credit to the private sector and stock market capitalization as a percentage of GDP as used to proxy for financial sector development on economic growth. Digitization is represented by the volume of digital/ICT equipment imported and GDP growth is used to proxy economic growth. Finally, we examine the effect of digitization on economic growth in the light of financial sector development. The following key results were found; first, digitalization propels financial sector development in Africa. Second, financial sector development enhances economic growth. Finally, contrary to our expectation, the results also indicate that digitalization conditioned on financial sector development tends to reduce economic growth in Africa. However, results of the net effects suggest that digitalization, overall, improve economic growth in Africa. We, therefore, conclude that, digitalization in Africa does not only develop the financial sector but unconditionally contributes the growth of the continent’s economies.

Keywords: digitalization, financial sector development, Africa, economic growth

Procedia PDF Downloads 148
17091 Pressure Gradient Prediction of Oil-Water Two Phase Flow through Horizontal Pipe

Authors: Ahmed I. Raheem

Abstract:

In this thesis, stratified and stratified wavy flow regimes have been investigated numerically for the oil (1.57 mPa s viscosity and 780 kg/m3 density) and water twophase flow in small and large horizontal steel pipes with a diameter between 0.0254 to 0.508 m by ANSYS Fluent software. Volume of fluid (VOF) with two phases flows using two equations family models (Realizable k-

Keywords: CFD, two-phase flow, pressure gradient, volume of fluid, large diameter, horizontal pipe, oil-water stratified and stratified wavy flow

Procedia PDF Downloads 435
17090 Creative Mathematically Modelling Videos Developed by Engineering Students

Authors: Esther Cabezas-Rivas

Abstract:

Ordinary differential equations (ODE) are a fundamental part of the curriculum for most engineering degrees, and students typically have difficulties in the subsequent abstract mathematical calculations. To enhance their motivation and profit that they are digital natives, we propose a teamwork project that includes the creation of a video. It should explain how to model mathematically a real-world problem transforming it into an ODE, which should then be solved using the tools learned in the lectures. This idea was indeed implemented with first-year students of a BSc in Engineering and Management during the period of online learning caused by the outbreak of COVID-19 in Spain. Each group of 4 students was assigned a different topic: model a hot water heater, search for the shortest path, design the quickest route for delivery, cooling a computer chip, the shape of the hanging cables of the Golden Gate, detecting land mines, rocket trajectories, etc. These topics should be worked out through two complementary channels: a written report describing the problem and a 10-15 min video on the subject. The report includes the following items: description of the problem to be modeled, detailed obtention of the ODE that models the problem, its complete solution, and interpretation in the context of the original problem. We report the outcomes of this teaching in context and active learning experience, including the feedback received by the students. They highlighted the encouragement of creativity and originality, which are skills that they do not typically relate to mathematics. Additionally, the video format (unlike a common presentation) has the advantage of allowing them to critically review and self-assess the recording, repeating some parts until the result is satisfactory. As a side effect, they felt more confident about their oral abilities. In short, students agreed that they had fun preparing the video. They recognized that it was tricky to combine deep mathematical contents with entertainment since, without the latter, it is impossible to engage people to view the video till the end. Despite this difficulty, after the activity, they claimed to understand better the material, and they enjoyed showing the videos to family and friends during and after the project.

Keywords: active learning, contextual teaching, models in differential equations, student-produced videos

Procedia PDF Downloads 151
17089 Development of Web-Based Remote Desktop to Provide Adaptive User Interfaces in Cloud Platform

Authors: Shuen-Tai Wang, Hsi-Ya Chang

Abstract:

Cloud virtualization technologies are becoming more and more prevalent, cloud users usually encounter the problem of how to access to the virtualized remote desktops easily over the web without requiring the installation of special clients. To resolve this issue, we took advantage of the HTML5 technology and developed web-based remote desktop. It permits users to access the terminal which running in our cloud platform from anywhere. We implemented a sketch of web interface following the cloud computing concept that seeks to enable collaboration and communication among users for high performance computing. Given the development of remote desktop virtualization, it allows to shift the user’s desktop from the traditional PC environment to the cloud platform, which is stored on a remote virtual machine rather than locally. This proposed effort has the potential to positively provide an efficient, resilience and elastic environment for online cloud service. This is also made possible by the low administrative costs as well as relatively inexpensive end-user terminals and reduced energy expenses.

Keywords: virtualization, remote desktop, HTML5, cloud computing

Procedia PDF Downloads 344
17088 Closing the Gap: Efficient Voxelization with Equidistant Scanlines and Gap Detection

Authors: S. Delgado, C. Cerrada, R. S. Gómez

Abstract:

This research introduces an approach to voxelizing the surfaces of triangular meshes with efficiency and accuracy. Our method leverages parallel equidistant scan-lines and introduces a Gap Detection technique to address the limitations of existing approaches. We present a comprehensive study showcasing the method's effectiveness, scalability, and versatility in different scenarios. Voxelization is a fundamental process in computer graphics and simulations, playing a pivotal role in applications ranging from scientific visualization to virtual reality. Our algorithm focuses on enhancing the voxelization process, especially for complex models and high resolutions. One of the major challenges in voxelization in the Graphics Processing Unit (GPU) is the high cost of discovering the same voxels multiple times. These repeated voxels incur in costly memory operations with no useful information. Our scan-line-based method ensures that each voxel is detected exactly once when processing the triangle, enhancing performance without compromising the quality of the voxelization. The heart of our approach lies in the use of parallel, equidistant scan-lines to traverse the interiors of triangles. This minimizes redundant memory operations and avoids revisiting the same voxels, resulting in a significant performance boost. Moreover, our method's computational efficiency is complemented by its simplicity and portability. Written as a single compute shader in Graphics Library Shader Language (GLSL), it is highly adaptable to various rendering pipelines and hardware configurations. To validate our method, we conducted extensive experiments on a diverse set of models from the Stanford repository. Our results demonstrate not only the algorithm's efficiency, but also its ability to produce 26 tunnel free accurate voxelizations. The Gap Detection technique successfully identifies and addresses gaps, ensuring consistent and visually pleasing voxelized surfaces. Furthermore, we introduce the Slope Consistency Value metric, quantifying the alignment of each triangle with its primary axis. This metric provides insights into the impact of triangle orientation on scan-line based voxelization methods. It also aids in understanding how the Gap Detection technique effectively improves results by targeting specific areas where simple scan-line-based methods might fail. Our research contributes to the field of voxelization by offering a robust and efficient approach that overcomes the limitations of existing methods. The Gap Detection technique fills a critical gap in the voxelization process. By addressing these gaps, our algorithm enhances the visual quality and accuracy of voxelized models, making it valuable for a wide range of applications. In conclusion, "Closing the Gap: Efficient Voxelization with Equidistant Scan-lines and Gap Detection" presents an effective solution to the challenges of voxelization. Our research combines computational efficiency, accuracy, and innovative techniques to elevate the quality of voxelized surfaces. With its adaptable nature and valuable innovations, this technique could have a positive influence on computer graphics and visualization.

Keywords: voxelization, GPU acceleration, computer graphics, compute shaders

Procedia PDF Downloads 75
17087 Virtual Computing Lab for Phonics Development among Deaf Students

Authors: Ankita R. Bansal, Naren S. Burade

Abstract:

Idea is to create a cloud based virtual lab for Deaf Students, “A language acquisition program using Visual Phonics and Cued Speech” using VMware Virtual Lab. This lab will demonstrate students the sounds of letters associated with the Language, building letter blocks, making words, etc Virtual labs are used for demos, training, for the Lingual development of children in their vernacular language. The main potential benefits are reduced labour and hardware costs, faster response times to users. Virtual Computing Labs allows any of the software as a service solutions, virtualization solutions, and terminal services solutions available today to offer as a service on demand, where a single instance of the software runs on the cloud and services multiple end users. VMWare, XEN, MS Virtual Server, Virtuoso, and Citrix are typical examples.

Keywords: visual phonics, language acquisition, vernacular language, cued speech, virtual lab

Procedia PDF Downloads 600
17086 Spatial Patterns and Temporal Evolution of Octopus Abundance in the Mauritanian Zone

Authors: Dedah Ahmed Babou, Nicolas Bez

Abstract:

The Min-Max autocorrelation factor (MAF) approach makes it possible to express in a space formed by spatially independent factors, spatiotemporal observations. These factors are ordered in decreasing order of spatial autocorrelation. The starting observations are thus expressed in the space formed by these factors according to temporal coordinates. Each vector of temporal coefficients expresses the temporal evolution of the weight of the corresponding factor. Applying this approach has enabled us to achieve the following results: (i) Define a spatially orthogonal space in which the projections of the raw data are determined; (ii) Define a limit threshold for the factors with the strongest structures in order to analyze the weight, and the temporal evolution of these different structures (iii) Study the correlation between the temporal evolution of the persistent spatial structures and that of the observed average abundance (iv) Propose prototypes of campaigns reflecting a high vs. low abundance (v) Propose a classification of campaigns that highlights seasonal and/or temporal similarities. These results were obtained by analyzing the octopus yield during the scientific campaigns of the oceanographic vessel Al Awam during the period 1989-2017 in the Mauritanian exclusive economic zone.

Keywords: spatiotemporal , autocorrelation, kriging, variogram, Octopus vulgaris

Procedia PDF Downloads 152
17085 Mechanical Characterization of Banana by Inverse Analysis Method Combined with Indentation Test

Authors: Juan F. P. Ramírez, Jésica A. L. Isaza, Benjamín A. Rojano

Abstract:

This study proposes a novel use of a method to determine the mechanical properties of fruits by the use of the indentation tests. The method combines experimental results with a numerical finite elements model. The results presented correspond to a simplified numerical modeling of banana. The banana was assumed as one-layer material with an isotropic linear elastic mechanical behavior, the Young’s modulus found is 0.3Mpa. The method will be extended to multilayer models in further studies.

Keywords: finite element method, fruits, inverse analysis, mechanical properties

Procedia PDF Downloads 361
17084 Performance Parameters of an Abbreviated Breast MRI Protocol

Authors: Andy Ho

Abstract:

Breast cancer is a common cancer in Australia. Early diagnosis is crucial for improving patient outcomes, as later-stage detection correlates with poorer prognoses. While multiparametric MRI offers superior sensitivity in detecting invasive and high-grade breast cancers compared to conventional mammography, its extended scan duration and high costs limit widespread application. As a result, full protocol MRI screening is typically reserved for patients at elevated risk. Recent advancements in imaging technology have facilitated the development of Abbreviated MRI protocols, which dramatically reduce scan times (<10 minutes compared to >30 minutes for full protocol). The potential for Abbreviated MRI to offer a more time- and cost-efficient alternative has implications for improving patient accessibility, reducing appointment durations, and enhancing compliance—especially relevant for individuals requiring regular annual screening over several decades. The purpose of this study is to assess the diagnostic efficacy of Abbreviated MRI for breast cancer screening among high-risk patients at the Royal Prince Alfred Hospital (RPA). This study aims to determine the sensitivity, specificity, and inter-reader variability of Abbreviated MRI protocols when interpreted by subspecialty-trained Breast Radiologists. A systematic review of the RPA’s electronic Picture Archive and Communication System identified high-risk patients, defined by Australian ‘Medicare Benefits Schedule’ criteria, who underwent Breast MRI from 2021 to 2022. Eligible participants included asymptomatic patients under 50 years old referred by the High-Risk Clinic due to a high-risk genetic profile or relevant familial history. The MRIs were anonymized, randomized, and interpreted by four Breast Radiologists, each independently completing standardized proforma evaluations. Radiological findings were compared against histopathology as the gold standard or follow-up imaging if biopsies were unavailable. Statistical metrics, including sensitivity, specificity, and inter-reader variability, were assessed. The Fleiss-Kappa analysis demonstrated a fair inter-reader agreement (kappa = 0.25; 95% CI: 0.19–0.32; p < 0.0001). The sensitivity for detecting malignancies was 0.75, with a specificity of 0.84. These findings underline the potential of Abbreviated MRI as a reliable screening tool for malignancies with significant specificity, though reduced sensitivity highlights the importance of robust radiologist training and consistent evaluation standards. Abbreviated MRI protocols exhibit promise as a viable screening option for high-risk patients, combining reduced scan times and acceptable diagnostic accuracy. Further work to refine interpretation practices and optimize training is essential to maximize the protocol’s utility in routine clinical screening and facilitate broader accessibility.

Keywords: abbreviated, breast, cancer, MRI

Procedia PDF Downloads 21
17083 The Markers -mm and dämmo in Amharic: Developmental Approach

Authors: Hayat Omar

Abstract:

Languages provide speakers with a wide range of linguistic units to organize and deliver information. There are several ways to verbally express the mental representations of events. According to the linguistic tools they have acquired, speakers select the one that brings out the most communicative effect to convey their message. Our study focuses on two markers, -mm and dämmo, in Amharic (Ethiopian Semitic language). Our aim is to examine, from a developmental perspective, how they are used by speakers. We seek to distinguish the communicative and pragmatic functions indicated by means of these markers. To do so, we created a corpus of sixty narrative productions of children from 5-6, 7-8 to 10-12 years old and adult Amharic speakers. The experimental material we used to collect our data is a series of pictures without text 'Frog, Where are you?'. Although -mm and dämmo are each used in specific contexts, they are sometimes analyzed as being interchangeable. The suffix -mm is complex and multifunctional. It marks the end of the negative verbal structure, it is found in the relative structure of the imperfect, it creates new words such as adverbials or pronouns, it also serves to coordinate words, sentences and to mark the link between macro-propositions within a larger textual unit. -mm was analyzed as marker of insistence, topic shift marker, element of concatenation, contrastive focus marker, 'bisyndetic' coordinator. On the other hand, dämmo has limited function and did not attract the attention of many authors. The only approach we could find analyzes it in terms of 'monosyndetic' coordinator. The paralleling of these two elements made it possible to understand their distinctive functions and refine their description. When it comes to marking a referent, the choice of -mm or dämmo is not neutral, depending on whether the tagged argument is newly introduced, maintained, promoted or reintroduced. The presence of these morphemes explains the inter-phrastic link. The information is seized by anaphora or presupposition: -mm goes upstream while dämmo arrows downstream, the latter requires new information. The speaker uses -mm or dämmo according to what he assumes to be known to his interlocutors. The results show that -mm and dämmo, although all the speakers use them both, do not always have the same scope according to the speaker and vary according to the age. dämmo is mainly used to mark a contrastive topic to signal the concomitance of events. It is more commonly used in young children’s narratives (F(3,56) = 3,82, p < .01). Some values of -mm (additive) are acquired very early while others are rather late and increase with age (F(3,56) = 3,2, p < .03). The difficulty is due not only because of its synthetic structure but primarily because it is multi-purpose and requires a memory work. It highlights the constituent on which it operates to clarify how the message should be interpreted.

Keywords: acquisition, cohesion, connection, contrastive topic, contrastive focus, discourse marker, pragmatics

Procedia PDF Downloads 139
17082 European Refugee Camps and the Right to an Adequate Standard of Living: Advancing Accountability under International Human Rights Law

Authors: Genevieve Zingg

Abstract:

Since the onset of the 2015 ‘refugee crisis’ in the European Union (EU), migrant deaths have overwhelmingly occurred in the Mediterranean Sea. However, far less attention has been paid to the startling number of injuries, deaths, and allegations of systematic human rights violations occurring within European refugee camps. Most troubling is the assertion that injuries and deaths in EU refugee camps have occurred as a result of negligent management and poor access to healthcare, food, water and sanitation, and other elements that comprise an adequate standard of living under international human rights law. Using available evidence and documentation, this paper will conduct a thorough examination of the causes of death and injury in EU refugee camps, with a specific focus on Greece, in order to identify instances of negligence or conditions that amount to potential breaches of human rights law. Based on its analysis, this paper will subsequently explore potential legal avenues to achieving justice and accountability under international human rights law in order to effectively address and remedy inadequate standards of living causing wrongful death or injury in European refugee camps.

Keywords: European Union, Greece, human rights, international human rights law, migration, refugees

Procedia PDF Downloads 197
17081 A Study for Area-level Mosquito Abundance Prediction by Using Supervised Machine Learning Point-level Predictor

Authors: Theoktisti Makridou, Konstantinos Tsaprailis, George Arvanitakis, Charalampos Kontoes

Abstract:

In the literature, the data-driven approaches for mosquito abundance prediction relaying on supervised machine learning models that get trained with historical in-situ measurements. The counterpart of this approach is once the model gets trained on pointlevel (specific x,y coordinates) measurements, the predictions of the model refer again to point-level. These point-level predictions reduce the applicability of those solutions once a lot of early warning and mitigation actions applications need predictions for an area level, such as a municipality, village, etc... In this study, we apply a data-driven predictive model, which relies on public-open satellite Earth Observation and geospatial data and gets trained with historical point-level in-Situ measurements of mosquito abundance. Then we propose a methodology to extract information from a point-level predictive model to a broader area-level prediction. Our methodology relies on the randomly spatial sampling of the area of interest (similar to the Poisson hardcore process), obtaining the EO and geomorphological information for each sample, doing the point-wise prediction for each sample, and aggregating the predictions to represent the average mosquito abundance of the area. We quantify the performance of the transformation from the pointlevel to the area-level predictions, and we analyze it in order to understand which parameters have a positive or negative impact on it. The goal of this study is to propose a methodology that predicts the mosquito abundance of a given area by relying on point-level prediction and to provide qualitative insights regarding the expected performance of the area-level prediction. We applied our methodology to historical data (of Culex pipiens) of two areas of interest (Veneto region of Italy and Central Macedonia of Greece). In both cases, the results were consistent. The mean mosquito abundance of a given area can be estimated with similar accuracy to the point-level predictor, sometimes even better. The density of the samples that we use to represent one area has a positive effect on the performance in contrast to the actual number of sampling points which is not informative at all regarding the performance without the size of the area. Additionally, we saw that the distance between the sampling points and the real in-situ measurements that were used for training did not strongly affect the performance.

Keywords: mosquito abundance, supervised machine learning, culex pipiens, spatial sampling, west nile virus, earth observation data

Procedia PDF Downloads 151