Search results for: coefficient of consolidation
1861 Effect of Composition on Work Hardening Coefficient of Bismuth-Lead Binary Alloy
Authors: K. A. Mistry, I. B. Patel, A. H. Prajapati
Abstract:
In the present work, the alloy of Bismuth-lead is prepared on the basis of percentage of molecular weight 9:1, 5:5 and 1:9 ratios and grown by Zone- Refining Technique under a vacuum atmosphere. The EDAX of these samples are done and the results are reported. Micro hardness test has been used as an alternative test for measuring material’s tensile properties. The effect of temperature and load on the hardness of the grown alloy has been studied. Further the comparative studies of work hardening coefficients are reported. In the present work, the alloy of Bismuth-lead is prepared on the basis of percentage of molecular weight 9:1, 5:5 and 1:9 ratios and grown by Zone- Refining Technique under a vacuum atmosphere. The EDAX of these samples are done and the results are reported. Micro hardness test has been used as an alternative test for measuring material’s tensile properties. The effect of temperature and load on the hardness of the grown alloy has been studied. Further the comparative studies of work hardening coefficients are reported.Keywords: EDAX, hardening coefficient, micro hardness, Bi-Pb alloy
Procedia PDF Downloads 3071860 Comparison of Developed Statokinesigram and Marker Data Signals by Model Approach
Authors: Boris Barbolyas, Kristina Buckova, Tomas Volensky, Cyril Belavy, Ladislav Dedik
Abstract:
Background: Based on statokinezigram, the human balance control is often studied. Approach to human postural reaction analysis is based on a combination of stabilometry output signal with retroreflective marker data signal processing, analysis, and understanding, in this study. The study shows another original application of Method of Developed Statokinesigram Trajectory (MDST), too. Methods: In this study, the participants maintained quiet bipedal standing for 10 s on stabilometry platform. Consequently, bilateral vibration stimuli to Achilles tendons in 20 s interval was applied. Vibration stimuli caused that human postural system took the new pseudo-steady state. Vibration frequencies were 20, 60 and 80 Hz. Participant's body segments - head, shoulders, hips, knees, ankles and little fingers were marked by 12 retroreflective markers. Markers positions were scanned by six cameras system BTS SMART DX. Registration of their postural reaction lasted 60 s. Sampling frequency was 100 Hz. For measured data processing were used Method of Developed Statokinesigram Trajectory. Regression analysis of developed statokinesigram trajectory (DST) data and retroreflective marker developed trajectory (DMT) data were used to find out which marker trajectories most correlate with stabilometry platform output signals. Scaling coefficients (λ) between DST and DMT by linear regression analysis were evaluated, too. Results: Scaling coefficients for marker trajectories were identified for all body segments. Head markers trajectories reached maximal value and ankle markers trajectories had a minimal value of scaling coefficient. Hips, knees and ankles markers were approximately symmetrical in the meaning of scaling coefficient. Notable differences of scaling coefficient were detected in head and shoulders markers trajectories which were not symmetrical. The model of postural system behavior was identified by MDST. Conclusion: Value of scaling factor identifies which body segment is predisposed to postural instability. Hypothetically, if statokinesigram represents overall human postural system response to vibration stimuli, then markers data represented particular postural responses. It can be assumed that cumulative sum of particular marker postural responses is equal to statokinesigram.Keywords: center of pressure (CoP), method of developed statokinesigram trajectory (MDST), model of postural system behavior, retroreflective marker data
Procedia PDF Downloads 3511859 Accelerating Quantum Chemistry Calculations: Machine Learning for Efficient Evaluation of Electron-Repulsion Integrals
Authors: Nishant Rodrigues, Nicole Spanedda, Chilukuri K. Mohan, Arindam Chakraborty
Abstract:
A crucial objective in quantum chemistry is the computation of the energy levels of chemical systems. This task requires electron-repulsion integrals as inputs, and the steep computational cost of evaluating these integrals poses a major numerical challenge in efficient implementation of quantum chemical software. This work presents a moment-based machine-learning approach for the efficient evaluation of electron-repulsion integrals. These integrals were approximated using linear combinations of a small number of moments. Machine learning algorithms were applied to estimate the coefficients in the linear combination. A random forest approach was used to identify promising features using a recursive feature elimination approach, which performed best for learning the sign of each coefficient but not the magnitude. A neural network with two hidden layers were then used to learn the coefficient magnitudes along with an iterative feature masking approach to perform input vector compression, identifying a small subset of orbitals whose coefficients are sufficient for the quantum state energy computation. Finally, a small ensemble of neural networks (with a median rule for decision fusion) was shown to improve results when compared to a single network.Keywords: quantum energy calculations, atomic orbitals, electron-repulsion integrals, ensemble machine learning, random forests, neural networks, feature extraction
Procedia PDF Downloads 1161858 Modeling of the Flow through an Earth Dam and Geotechnical Slope Analyzes
Authors: Ahmed Ferhati, Arezki Adjrad, Ratiba Mitiche-Kettab, Hakim Djafer Khodja
Abstract:
The porous media are omnipresent around us that they are natural as sand, clay, rocks, or manufactured like concretes, cement, and ceramics. The variety of porous environment indicates a wide material range which can be very different from each other. Their common point is to be made up of a solid matrix and a porous space. In our case of study, we made the modeling of the flows in porous environments through the massives as in the case of an earth dam. The computer code used (PLAXIS) offer the possibility of modeling of various structures, in particular, the works in lands because that it deals with the pore water pressure due to the underground flow and the calculation of the plastic deformations. To confirm results obtained by PLAXIS, GeoStudio SEEP/W code was used. This work treats modeling of flows and mechanical and hydraulic behavior of earth dam. A general framework which can fit the calculation of this kind of structures and the coupling of the soil consolidation and free surface flows was defined. In this study; we have confronted a real case modeling of an earth dam. It was shown, in particular, that it is possible to entirely lead the calculation of real dam and to get encouraging results from the hydraulic and mechanical point of view.Keywords: analyzes, dam, flow, modeling, PLAXIS, seep/w, slope
Procedia PDF Downloads 3101857 Failure Mechanisms of Isolated vs. in Aggregate Historical Buildings: A Case Study for Timisoara, Romania
Authors: I. Apostol, M. Mosoarca
Abstract:
Romania is a seismic country, with two major seismic zones, Vrancea and Banat. One of the most important cities from Banat seismic area is Timisoara, where a lot of valuable historical buildings were built before any design codes, but still they kept their stability during past earthquakes. This article presents the influence of the adjacent buildings during an earthquake and the way that the specific failure mechanism is changed when the building is part of an aggregate. The investigation was made using nonlinear analysis based on Tremuri software, first analyzing the buildings as isolated and second, considering the entire aggregate of buildings. There were noticed significant differences through the two situations regarding the specific failure mechanism activated for each building, showing the fact that in some situations, the presence of the adjacent buildings has positive or negative contribution for the seismic behavior of the analyzed one. The difference between the failure mechanism of the same buildings considered isolated and in aggregate aims to provide explications for the good structural state of the existing historical areas of Timisoara, as part of a larger multidisciplinary study, which will help local authorities to prioritize the consolidation works for the historical buildings in order to assure that the history of the city will be kept alive for the next generations.Keywords: failure mechanism, analysis, aggregate, masonry, earthquake
Procedia PDF Downloads 1451856 Spatial Interpolation of Aerosol Optical Depth Pollution: Comparison of Methods for the Development of Aerosol Distribution
Authors: Sahabeh Safarpour, Khiruddin Abdullah, Hwee San Lim, Mohsen Dadras
Abstract:
Air pollution is a growing problem arising from domestic heating, high density of vehicle traffic, electricity production, and expanding commercial and industrial activities, all increasing in parallel with urban population. Monitoring and forecasting of air quality parameters are important due to health impact. One widely available metric of aerosol abundance is the aerosol optical depth (AOD). The AOD is the integrated light extinction coefficient over a vertical atmospheric column of unit cross section, which represents the extent to which the aerosols in that vertical profile prevent the transmission of light by absorption or scattering. Seasonal aerosol optical depth (AOD) values at 550 nm derived from the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor onboard NASA’s Terra satellites, for the 10 years period of 2000-2010 were used to test 7 different spatial interpolation methods in the present study. The accuracy of estimations was assessed through visual analysis as well as independent validation based on basic statistics, such as root mean square error (RMSE) and correlation coefficient. Based on the RMSE and R values of predictions made using measured values from 2000 to 2010, Radial Basis Functions (RBFs) yielded the best results for spring, summer, and winter and ordinary kriging yielded the best results for fall.Keywords: aerosol optical depth, MODIS, spatial interpolation techniques, Radial Basis Functions
Procedia PDF Downloads 4081855 Artificial Neural Network Approach for Modeling Very Short-Term Wind Speed Prediction
Authors: Joselito Medina-Marin, Maria G. Serna-Diaz, Juan C. Seck-Tuoh-Mora, Norberto Hernandez-Romero, Irving Barragán-Vite
Abstract:
Wind speed forecasting is an important issue for planning wind power generation facilities. The accuracy in the wind speed prediction allows a good performance of wind turbines for electricity generation. A model based on artificial neural networks is presented in this work. A dataset with atmospheric information about air temperature, atmospheric pressure, wind direction, and wind speed in Pachuca, Hidalgo, México, was used to train the artificial neural network. The data was downloaded from the web page of the National Meteorological Service of the Mexican government. The records were gathered for three months, with time intervals of ten minutes. This dataset was used to develop an iterative algorithm to create 1,110 ANNs, with different configurations, starting from one to three hidden layers and every hidden layer with a number of neurons from 1 to 10. Each ANN was trained with the Levenberg-Marquardt backpropagation algorithm, which is used to learn the relationship between input and output values. The model with the best performance contains three hidden layers and 9, 6, and 5 neurons, respectively; and the coefficient of determination obtained was r²=0.9414, and the Root Mean Squared Error is 1.0559. In summary, the ANN approach is suitable to predict the wind speed in Pachuca City because the r² value denotes a good fitting of gathered records, and the obtained ANN model can be used in the planning of wind power generation grids.Keywords: wind power generation, artificial neural networks, wind speed, coefficient of determination
Procedia PDF Downloads 1251854 Government Final Consumption Expenditure and Household Consumption Expenditure NPISHS in Nigeria
Authors: Usman A. Usman
Abstract:
Undeniably, unlike the Classical side, the Keynesian perspective of the aggregate demand side indeed has a significant position in the policy, growth, and welfare of Nigeria due to government involvement and ineffective demand of the population living with poor per capita income. This study seeks to investigate the effect of Government Final Consumption Expenditure, Financial Deepening on Households, and NPISHs Final consumption expenditure using data on Nigeria from 1981 to 2019. This study employed the ADF stationarity test, Johansen Cointegration test, and Vector Error Correction Model. The results of the study revealed that the coefficient of Government final consumption expenditure has a positive effect on household consumption expenditure in the long run. There is a long-run and short-run relationship between gross fixed capital formation and household consumption expenditure. The coefficients cpsgdp (financial deepening and gross fixed capital formation posit a negative impact on household final consumption expenditure. The coefficients money supply lm2gdp, which is another proxy for financial deepening, and the coefficient FDI have a positive effect on household final consumption expenditure in the long run. Therefore, this study recommends that Gross fixed capital formation stimulates household consumption expenditure; a legal framework to support investment is a panacea to increasing hoodmold income and consumption and reducing poverty in Nigeria. Therefore, this should be a key central component of policy.Keywords: government final consumption expenditure, household consumption expenditure, vector error correction model, cointegration
Procedia PDF Downloads 531853 Application of Basic Principles of Educational Administration for the Enhancement of Senior Secondary School Principals in Kano State Nigeria
Authors: Ibrahim Auwal
Abstract:
This study focuses on senior secondary education towards the development of younger generation in general terms, and specifically for the enhancement of senior secondary school principals. Investigation was made to correlate between principals’ application of basic principles of educational administration and principals’ productivity in senior secondary schools in Kano State. The instrument used to collect relevant data was self designed Observation Inventory for School Principals (OISP). The observation inventory items were scrutinized by experts from the School of Education Federal College of Education Kano to ascertain the contents validity, and the reliability coefficient was 0.83. Using purposive sampling technique, 30 schools were chosen from 85 senior secondary schools in Kano state and 30 principals were deliberately sampled due to their small number. Pearson Product Moment Correlation (r) Coefficient was used to test the hypothesis generated for the study. The results of the analysis showed that principals’ application of basic principles of educational administration was significantly correlated with principals’ productivity and it promote the performance of the students. Based on the findings, it was recommended that, government should in as much as possible encourage school principals to obtain degrees in relevant and specialized areas in education specifically educational administration and planning so as to get all the necessary knowledge and skills of leader ship procedures that will definitely promote teachers morale, improve students’ academic performance and enhances principals’ productivity in senior secondary schools in Kano State.Keywords: principles of educational administration, principals of senior secondary schools, Kano, educational sciences
Procedia PDF Downloads 4941852 Managing Fake News for Sustainable Democracy in Enugu State, Nigeria
Authors: Gloria Ebere Amadi, Emeka Promise Ugwunwotti
Abstract:
The study was carried out to determine the strategies for managing fake news for sustainable democracy in Enugu State, Nigeria. Two research questions and two null hypotheses guided the study. A survey research design was used for the study. The population for the study consisted of 100 respondents (from Enugu state House of Assembly). Of the entire population, 24 elected law makers and 76 staff were used; hence there was no sampling since the population was manageable. A 28-item structured questionnaire developed by the researcher was used for data collection. The instrument entitled Managing Fake News Questionnaire (MFNQ) was validated by three experts, two from the Department of Computer Science and one from the Department of Maths and Statistics, all from Enugu State University of Science and Technology. Cronbach Alpha was used to determine the reliability coefficient of the two sections of the instrument, and they are 0.67 and 0.82, while the reliability coefficient of the whole instrument gave a value of 0.81. Mean with standard deviation was used to answer research questions, while the null hypotheses at 0.5 level of significance at 98 degrees of freedom were tested with a t-test. The findings of the study revealed that the respondents agreed that government and citizens-related strategies improve the management of fake news for sustainable democracy in Enugu State. Again, there was no significant difference between the mean response of the lawmakers and staff on government and citizens-related strategies for managing fake news for sustainable democracy in Enugu State. Based on the findings, it was recommended, among others, that there should be regular workshops on the management of fake news for citizens.Keywords: fake news, sustainability, democracy, management
Procedia PDF Downloads 701851 Assessment of Groundwater Quality in Kaltungo Local Government Area of Gombe State
Authors: Rasaq Bello, Grace Akintola Sunday, Yemi Sikiru Onifade
Abstract:
Groundwater is required for the continuity of life and sustainability of the ecosystem. Hence, this research was purposed to assess groundwater quality for domestic use in Kaltungo Local Government Area, Gombe State. The work was also aimed at determining the thickness and resistivity of the topsoil, areas suitable for borehole construction, quality and potentials of groundwater in the study area. The study area extends from latitude N10015’38” - E11008’01” and longitude N10019’29” - E11013’05”. The data was acquired using the Vertical Electrical Sounding (VES) method and processed using IP12win software. Twenty (20) Vertical Electrical Soundings were carried out with a maximum current electrode separation (AB) of 150m. The VES curves generated from the data reveal that all the VES points have five to six subsurface layers. The first layer has a resistivity value of 7.5 to 364.1 Ωm and a thickness ranging from 0.8 to 7.4m, and the second layer has a resistivity value of 1.8 to 600.3 Ωm thickness ranging from 2.6 to 31.4m, the third layer has resistivity value of 23.3 to 564.4 Ωm thickness ranging from 10.3 to 77.8m, the fourth layer has resistivity value of 19.7 to 640.2 Ωm thickness ranging from 8.2m to 120.0m, the fifth layer has resistivity value of 27 to 234 Ωm thickness ranging from 8.2 to 53.7m and the six-layer is the layer that extended beyond the probing depth. The VES curves generated from the data revealed KQHA curve type for VES 1, HKQQ curve for VES 4, HKQ curve for VES 5, KHA curve for VES 11, QQHK curve for VES 12, HAA curve for VES 6 and VES 19, HAKH curve for VES 7, VES 8, VES 10 and VES 18, HKH curve for VES 2, VES 3, VES 9, VES 13, VES 14, VES 15, VES 16, VES 17 and VES 20. Values of the Coefficient of Anisotropy, Reflection Coefficient, and Resistivity Contrast obtained from the Dar-Zarrouk parameters indicated good water prospects for all the VES points in this study, with VES points 4, 9 and 18 having the highest prospects for groundwater exploration.Keywords: formation parameters, groundwater, resistivity, resistivity contrast, vertical electrical sounding
Procedia PDF Downloads 551850 Investigating The Effect Of Convection On The Rating Of Buried Cables Using The Finite Element Method
Authors: Sandy J. M. Balla, Jerry J. Walker, Isaac K. Kyere
Abstract:
The heat transfer coefficient at the soil–air interface is important in calculating underground cable ampacity when convection occurs. Calculating the heat transfer coefficient accurately is complex because of the temperature variations at the earth's surface. This paper presents the effect of convection heat flow across the ground surface on the rating of three single-core, 132kV, XLPE cables buried underground. The Finite element method (FEM) is a numerical analysis technique used to determine the cable rating of buried cables under installation conditions that are difficult to support when using the analytical method. This study demonstrates the use of FEM to investigate the effect of convection on the rating ofburied cables in flat formation using QuickField finite element simulation software. As a result, developing a model to simulate this type of situation necessitates important considerations such as the following boundary conditions: burial depth, soil thermal resistivity, and soil temperature, which play an important role in the simulation's accuracy and reliability. The results show that when the ground surface is taken as a convection interface, the conductor temperature rises and may exceed the maximum permissible temperature when rated current flows. This is because the ground surface acts as a convection interface between the soil and the air (fluid). This result correlates and is compared with the rating obtained using the IEC60287 analytical method, which is based on the condition that the ground surface is an isotherm.Keywords: finite element method, convection, buried cables, steady-state rating
Procedia PDF Downloads 1311849 Reducing Pressure Drop in Microscale Channel Using Constructal Theory
Authors: K. X. Cheng, A. L. Goh, K. T. Ooi
Abstract:
The effectiveness of microchannels in enhancing heat transfer has been demonstrated in the semiconductor industry. In order to tap the microscale heat transfer effects into macro geometries, overcoming the cost and technological constraints, microscale passages were created in macro geometries machined using conventional fabrication methods. A cylindrical insert was placed within a pipe, and geometrical profiles were created on the outer surface of the insert to enhance heat transfer under steady-state single-phase liquid flow conditions. However, while heat transfer coefficient values of above 10 kW/m2·K were achieved, the heat transfer enhancement was accompanied by undesirable pressure drop increment. Therefore, this study aims to address the high pressure drop issue using Constructal theory, a universal design law for both animate and inanimate systems. Two designs based on Constructal theory were developed to study the effectiveness of Constructal features in reducing the pressure drop increment as compared to parallel channels, which are commonly found in microchannel fabrication. The hydrodynamic and heat transfer performance for the Tree insert and Constructal fin (Cfin) insert were studied using experimental methods, and the underlying mechanisms were substantiated by numerical results. In technical terms, the objective is to achieve at least comparable increment in both heat transfer coefficient and pressure drop, if not higher increment in the former parameter. Results show that the Tree insert improved the heat transfer performance by more than 16 percent at low flow rates, as compared to the Tree-parallel insert. However, the heat transfer enhancement reduced to less than 5 percent at high Reynolds numbers. On the other hand, the pressure drop increment stayed almost constant at 20 percent. This suggests that the Tree insert has better heat transfer performance in the low Reynolds number region. More importantly, the Cfin insert displayed improved heat transfer performance along with favourable hydrodynamic performance, as compared to Cfinparallel insert, at all flow rates in this study. At 2 L/min, the enhancement of heat transfer was more than 30 percent, with 20 percent pressure drop increment, as compared to Cfin-parallel insert. Furthermore, comparable increment in both heat transfer coefficient and pressure drop was observed at 8 L/min. In other words, the Cfin insert successfully achieved the objective of this study. Analysis of the results suggests that bifurcation of flows is effective in reducing the increment in pressure drop relative to heat transfer enhancement. Optimising the geometries of the Constructal fins is therefore the potential future study in achieving a bigger stride in energy efficiency at much lower costs.Keywords: constructal theory, enhanced heat transfer, microchannel, pressure drop
Procedia PDF Downloads 3381848 Government Final Consumption Expenditure Financial Deepening and Household Consumption Expenditure NPISHs in Nigeria
Authors: Usman A. Usman
Abstract:
Undeniably, unlike the Classical side, the Keynesian perspective of the aggregate demand side indeed has a significant position in the policy, growth, and welfare of Nigeria due to government involvement and ineffective demand of the population living with poor per capita income. This study seeks to investigate the effect of Government Final Consumption Expenditure, Financial Deepening on Households, and NPISHs Final consumption expenditure using data on Nigeria from 1981 to 2019. This study employed the ADF stationarity test, Johansen Cointegration test, and Vector Error Correction Model. The results of the study revealed that the coefficient of Government final consumption expenditure has a positive effect on household consumption expenditure in the long run. There is a long-run and short-run relationship between gross fixed capital formation and household consumption expenditure. The coefficients cpsgdp financial deepening and gross fixed capital formation posit a negative impact on household final consumption expenditure. The coefficients money supply lm2gdp, which is another proxy for financial deepening, and the coefficient FDI have a positive effect on household final consumption expenditure in the long run. Therefore, this study recommends that Gross fixed capital formation stimulates household consumption expenditure; a legal framework to support investment is a panacea to increasing hoodmold income and consumption and reducing poverty in Nigeria. Therefore, this should be a key central component of policy.Keywords: household, government expenditures, vector error correction model, johansen test
Procedia PDF Downloads 611847 Performance of Autoclaved Aerated Concrete Containing Recycled Ceramic and Gypsum Waste as Partial Replacement for Sand
Authors: Efil Yusrianto, Noraini Marsi, Noraniah Kassim, Izzati Abdul Manaf, Hafizuddin Hakim Shariff
Abstract:
Today, municipal solid waste (MSW), noise pollution, and attack fire are three ongoing issues for inhabitants of urban including in Malaysia. To solve these issues, eco-friendly autoclaved aerated concrete (AAC) containing recycled ceramic and gypsum waste (CGW) as a partial replacement for sand with different ratios (0%, 5%, 10%, 15%, 20%, and 25% wt) has been prepared. The performance of samples, such as the physical, mechanical, sound absorption coefficient, and direct fire resistance, has been investigated. All samples showed normal color behavior, i.e., grey and free crack. The compressive strength was increased in the range of 6.10% to 29.88%. The maximum value of compressive strength was 2.13MPa for 15% wt of CGW. The positive effect of CGW on the compressive strength of AAC has also been confirmed by crystalline phase and microstructure analysis. The acoustic performances, such as sound absorption coefficients of samples at low frequencies (500Hz), are higher than the reference sample (RS). AAC-CGW samples are categorized as AAC material classes B and C. The fire resistance results showed the physical surface of the samples had a free crack and was not burned during the direct fire at 950ºC for 300s. The results showed that CGW succeeded in enhancing the performance of fresh AAC, such as compressive strength, crystalline phase, sound absorption coefficient, and fire resistance of samples.Keywords: physical, mechanical, acoustic, direct fire resistance performance, autoclaved aerated concrete, recycled ceramic-gypsum waste
Procedia PDF Downloads 1401846 Design of a Portable Shielding System for a Newly Installed NaI(Tl) Detector
Authors: Mayesha Tahsin, A.S. Mollah
Abstract:
Recently, a 1.5x1.5 inch NaI(Tl) detector based gamma-ray spectroscopy system has been installed in the laboratory of the Nuclear Science and Engineering Department of the Military Institute of Science and Technology for radioactivity detection purposes. The newly installed NaI(Tl) detector has a circular lead shield of 22 mm width. An important consideration of any gamma-ray spectroscopy is the minimization of natural background radiation not originating from the radioactive sample that is being measured. Natural background gamma-ray radiation comes from naturally occurring or man-made radionuclides in the environment or from cosmic sources. Moreover, the main problem with this system is that it is not suitable for measurements of radioactivity with a large sample container like Petridish or Marinelli beaker geometry. When any laboratory installs a new detector or/and new shield, it “must” first carry out quality and performance tests for the detector and shield. This paper describes a new portable shielding system with lead that can reduce the background radiation. Intensity of gamma radiation after passing the shielding will be calculated using shielding equation I=Ioe-µx where Io is initial intensity of the gamma source, I is intensity after passing through the shield, µ is linear attenuation coefficient of the shielding material, and x is the thickness of the shielding material. The height and width of the shielding will be selected in order to accommodate the large sample container. The detector will be surrounded by a 4π-geometry low activity lead shield. An additional 1.5 mm thick shield of tin and 1 mm thick shield of copper covering the inner part of the lead shielding will be added in order to remove the presence of characteristic X-rays from the lead shield.Keywords: shield, NaI (Tl) detector, gamma radiation, intensity, linear attenuation coefficient
Procedia PDF Downloads 1591845 Powder Flow with Normalized Powder Particles Size Distribution and Temperature Analyses in Laser Melting Deposition: Analytical Modelling and Experimental Validation
Authors: Muhammad Arif Mahmood, Andrei C. Popescu, Mihai Oane, Diana Chioibascu, Carmen Ristoscu, Ion N. Mihailescu
Abstract:
Powder flow and temperature distributions are recognized as influencing factors during laser melting deposition (LMD) process, that not only affect the consolidation rate but also characteristics of the deposited layers. Herewith, two simplified analytical models will be presented to simulate the powder flow with the inclusion of powder particles size distribution in Gaussian form, under three powder jet nozzles, and temperature analyses during LMD process. The output of the 1st model will serve as the input in the 2nd model. The models will be validated with experimental data, i.e., weight measurement method for powder particles distribution and infrared imaging for temperature analyses. This study will increase the cost-efficiency of the LMD process by adjustment of the operating parameters for reaching optimal powder debit and energy. This research has received funds under the Marie Sklodowska-Curie grant agreement No. 764935, from the European Union’s Horizon 2020 research and innovation program.Keywords: laser additive manufacturing, powder particles size distribution in Gaussian form, powder stream distribution, temperature analyses
Procedia PDF Downloads 1361844 Multi-Scale Spatial Difference Analysis Based on Nighttime Lighting Data
Authors: Qinke Sun, Liang Zhou
Abstract:
The ‘Dragon-Elephant Debate’ between China and India is an important manifestation of global multipolarity in the 21st century. The two rising powers have carried out economic reforms one after another in the interval of more than ten years, becoming the fastest growing developing country and emerging economy in the world. At the same time, the development differences between China and India have gradually attracted wide attention of scholars. Based on the continuous annual night light data (DMSP-OLS) from 1992 to 2012, this paper systematically compares and analyses the regional development differences between China and India by Gini coefficient, coefficient of variation, comprehensive night light index (CNLI) and hot spot analysis. The results show that: (1) China's overall expansion from 1992 to 2012 is 1.84 times that of India, in which China's change is 2.6 times and India's change is 2 times. The percentage of lights in unlighted areas in China dropped from 92% to 82%, while that in India from 71% to 50%. (2) China's new growth-oriented cities appear in Hohhot, Inner Mongolia, Ordos, and Urumqi in the west, and the declining cities are concentrated in Liaoning Province and Jilin Province in the northeast; India's new growth-oriented cities are concentrated in Chhattisgarh in the north, while the declining areas are distributed in Uttar Pradesh. (3) China's differences on different scales are lower than India's, and regional inequality of development is gradually narrowing. Gini coefficients at the regional and provincial levels have decreased from 0.29, 0.44 to 0.24 and 0.38, respectively, while regional inequality in India has slowly improved and regional differences are gradually widening, with Gini coefficients rising from 0.28 to 0.32. The provincial Gini coefficient decreased slightly from 0.64 to 0.63. (4) The spatial pattern of China's regional development is mainly east-west difference, which shows the difference between coastal and inland areas; while the spatial pattern of India's regional development is mainly north-south difference, but because the southern states are sea-dependent, it also reflects the coastal inland difference to a certain extent. (5) Beijing and Shanghai present a multi-core outward expansion model, with an average annual CNLI higher than 0.01, while New Delhi and Mumbai present the main core enhancement expansion model, with an average annual CNLI lower than 0.01, of which the average annual CNLI in Shanghai is about five times that in Mumbai.Keywords: spatial pattern, spatial difference, DMSP-OLS, China, India
Procedia PDF Downloads 1571843 Concentrated Whey Protein Drink with Orange Flavor: Protein Modification and Formulation
Authors: Shahram Naghizadeh Raeisi, Ali Alghooneh
Abstract:
The application of whey protein in drink industry to enhance the nutritional value of the products is important. Furthermore, the gelification of protein during thermal treatment and shelf life makes some limitations in its application. So, the main goal of this research is manufacturing of high concentrate whey protein orange drink with appropriate shelf life. In this way, whey protein was 5 to 30% hydrolyzed ( in 5 percent intervals at six stages), then thermal stability of samples with 10% concentration of protein was tested in acidic condition (T= 90 °C, pH=4.2, 5 minutes ) and neutral condition (T=120° C, pH:6.7, 20 minutes.) Furthermore, to study the shelf life of heat treated samples in 4 months at 4 and 24 °C, the time sweep rheological test were done. At neutral conditions, 5 to 20% hydrolyzed sample showed gelling during thermal treatment, whereas at acidic condition, was happened only in 5 to 10 percent hydrolyzed samples. This phenomenon could be related to the difference in hydrodynamic radius and zeta potential of samples with different level of hydrolyzation at acidic and neutral conditions. To study the gelification of heat resistant protein solutions during shelf life, for 4 months with 7 days intervals, the time sweep analysis were performed. Cross over was observed for all heat resistant neutral samples at both storage temperature, while in heat resistant acidic samples with degree of hydrolysis, 25 and 30 percentage at 4 and 20 °C, it was not seen. It could be concluded that the former sample was stable during heat treatment and 4 months storage, which made them a good choice for manufacturing high protein drinks. The Scheffe polynomial model and numerical optimization were employed for modeling and high protein orange drink formula optimization. Scheffe model significantly predicted the overal acceptance index (Pvalue<0.05) of sensorial analysis. The coefficient of determination (R2) of 0.94, the adjusted coefficient of determination (R2Adj) of 0.90, insignificance of the lack-of-fit test and F value of 64.21 showed the accuracy of the model. Moreover, the coefficient of variable (C.V) was 6.8% which suggested the replicability of the experimental data. The desirability function had been achieved to be 0.89, which indicates the high accuracy of optimization. The optimum formulation was found as following: Modified whey protein solution (65.30%), natural orange juice (33.50%), stevia sweetener (0.05%), orange peel oil (0.15%) and citric acid (1 %), respectively. Its worth mentioning that this study made an appropriate model for application of whey protein in drink industry without bitter flavor and gelification during heat treatment and shelf life.Keywords: croos over, orange beverage, protein modification, optimization
Procedia PDF Downloads 621842 Development of IDF Curves for Precipitation in Western Watershed of Guwahati, Assam
Authors: Rajarshi Sharma, Rashidul Alam, Visavino Seleyi, Yuvila Sangtam
Abstract:
The Intensity-Duration-Frequency (IDF) relationship of rainfall amounts is one of the most commonly used tools in water resources engineering for planning, design and operation of water resources project, or for various engineering projects against design floods. The establishment of such relationships was reported as early as in 1932 (Bernard). Since then many sets of relationships have been constructed for several parts of the globe. The objective of this research is to derive IDF relationship of rainfall for western watershed of Guwahati, Assam. These relationships are useful in the design of urban drainage works, e.g. storm sewers, culverts and other hydraulic structures. In the study, rainfall depth for 10 years viz. 2001 to 2010 has been collected from the Regional Meteorological Centre Borjhar, Guwahati. Firstly, the data has been used to construct the mass curve for duration of more than 7 hours rainfall to calculate the maximum intensity and to form the intensity duration curves. Gumbel’s frequency analysis technique has been used to calculate the probable maximum rainfall intensities for a period of 2 yr, 5 yr, 10 yr, 50 yr, 100 yr from the maximum intensity. Finally, regression analysis has been used to develop the intensity-duration-frequency (IDF) curve. Thus, from the analysis the values for the constants ‘a’,‘b’ &‘c’ have been found out. The values of ‘a’ for which the sum of the squared deviation is minimum has been found out to be 40 and when the corresponding value of ‘c’ and ‘b’ for the minimum squared deviation of ‘a’ are 0.744 and 1981.527 respectively. The results obtained showed that in all the cases the correlation coefficient is very high indicating the goodness of fit of the formulae to estimate IDF curves in the region of interest.Keywords: intensity-duration-frequency relationship, mass curve, regression analysis, correlation coefficient
Procedia PDF Downloads 2451841 Banking and Accounting Analysis Researches Effect on Environment
Authors: Michael Saad Thabet Azrek
Abstract:
The advanced facts era is becoming a vital element within the improvement of financial offerings enterprise, in particular, the banking enterprise. It has introduced new approaches to delivering banking to the patron, including Internet Banking. Banks started to observe digital banking (e-banking) as a means to update a number of their conventional branch features using the net as a brand-new distribution channel. A few purchasers have, as a minimum, a couple of accounts across banks and get the right of entry to these accounts using e-banking offerings. To study the contemporary net really worth role, clients ought to log in to each of their debts and get the info and paintings on consolidation. This not simplest takes enough time, but it's also a repetitive hobby at a specific frequency. To cope with this point, an account aggregation idea is introduced as an answer. E-banking account aggregation, as one of the e-banking types, appeared to construct a more potent dating with clients. Account Aggregation provider usually refers to a provider that permits clients to manage their financial institution debts maintained in distinct establishments through a not unusual net banking operating a platform, with an excessive situation to protection and privateness. This paper gives an outline of an e-banking account aggregation method as a new provider in the e-banking field.Keywords: compatibility, complexity, mobile banking, observation, risk banking technology, internet banks, modernization of banks, banks, account aggregation, security, enterprise development
Procedia PDF Downloads 381840 Support Vector Machine Based Retinal Therapeutic for Glaucoma Using Machine Learning Algorithm
Authors: P. S. Jagadeesh Kumar, Mingmin Pan, Yang Yung, Tracy Lin Huan
Abstract:
Glaucoma is a group of visual maladies represented by the scheduled optic nerve neuropathy; means to the increasing dwindling in vision ground, resulting in loss of sight. In this paper, a novel support vector machine based retinal therapeutic for glaucoma using machine learning algorithm is conservative. The algorithm has fitting pragmatism; subsequently sustained on correlation clustering mode, it visualizes perfect computations in the multi-dimensional space. Support vector clustering turns out to be comparable to the scale-space advance that investigates the cluster organization by means of a kernel density estimation of the likelihood distribution, where cluster midpoints are idiosyncratic by the neighborhood maxima of the concreteness. The predicted planning has 91% attainment rate on data set deterrent on a consolidation of 500 realistic images of resolute and glaucoma retina; therefore, the computational benefit of depending on the cluster overlapping system pedestal on machine learning algorithm has complete performance in glaucoma therapeutic.Keywords: machine learning algorithm, correlation clustering mode, cluster overlapping system, glaucoma, kernel density estimation, retinal therapeutic
Procedia PDF Downloads 2541839 Particle Swarm Optimization Algorithm vs. Genetic Algorithm for Image Watermarking Based Discrete Wavelet Transform
Authors: Omaima N. Ahmad AL-Allaf
Abstract:
Over communication networks, images can be easily copied and distributed in an illegal way. The copyright protection for authors and owners is necessary. Therefore, the digital watermarking techniques play an important role as a valid solution for authority problems. Digital image watermarking techniques are used to hide watermarks into images to achieve copyright protection and prevent its illegal copy. Watermarks need to be robust to attacks and maintain data quality. Therefore, we discussed in this paper two approaches for image watermarking, first is based on Particle Swarm Optimization (PSO) and the second approach is based on Genetic Algorithm (GA). Discrete wavelet transformation (DWT) is used with the two approaches separately for embedding process to cover image transformation. Each of PSO and GA is based on co-relation coefficient to detect the high energy coefficient watermark bit in the original image and then hide the watermark in original image. Many experiments were conducted for the two approaches with different values of PSO and GA parameters. From experiments, PSO approach got better results with PSNR equal 53, MSE equal 0.0039. Whereas GA approach got PSNR equal 50.5 and MSE equal 0.0048 when using population size equal to 100, number of iterations equal to 150 and 3×3 block. According to the results, we can note that small block size can affect the quality of image watermarking based PSO/GA because small block size can increase the search area of the watermarking image. Better PSO results were obtained when using swarm size equal to 100.Keywords: image watermarking, genetic algorithm, particle swarm optimization, discrete wavelet transform
Procedia PDF Downloads 2281838 An Overall Evaluation of Food Nanotechnology
Authors: Raana Babadi Fathipour
Abstract:
Nourishment nanotechnology is an range of rising intrigued and opens up a entirety universe of modern conceivable outcomes for the nourishment industry. The essential categories of nanotechnology applications and functionalities right now within the improvement of nourishment bundling incorporate: the enhancement of plastic materials obstructions, the consolidation of dynamic components that can convey utilitarian properties past those of customary dynamic bundling, and the detecting and signaling of significant data. Nano nourishment bundling materials may amplify nourishment life, move forward nourishment security, alarm buyers that nourishment is sullied or ruined, repair tears in bundling, and indeed release preservatives to expand the life of the nourishment within the bundle. Nanotechnology applications within the nourishment industry can be utilized to identify microbes in bundling, or produce stronger flavors and color quality, and security by expanding the obstruction properties. Nanotechnology holds extraordinary guarantee to supply benefits not fair inside nourishment items but too around nourishment items. In reality, nanotechnology presents modern chances for advancement within the nourishment industry at monstrous speed, but instability and wellbeing concerns are moreover developing. EU/WE/global enactment for the direction of nanotechnology in nourishment are scanty. Besides, current enactment shows up unacceptable to nanotechnology specificity.Keywords: nano technology, nano foods, food packaging, nano participle
Procedia PDF Downloads 651837 Opto-Electronic Properties and Structural Phase Transition of Filled-Tetrahedral NaZnAs
Authors: R. Khenata, T. Djied, R. Ahmed, H. Baltache, S. Bin-Omran, A. Bouhemadou
Abstract:
We predict structural, phase transition as well as opto-electronic properties of the filled-tetrahedral (Nowotny-Juza) NaZnAs compound in this study. Calculations are carried out by employing the full potential (FP) linearized augmented plane wave (LAPW) plus local orbitals (lo) scheme developed within the structure of density functional theory (DFT). Exchange-correlation energy/potential (EXC/VXC) functional is treated using Perdew-Burke and Ernzerhof (PBE) parameterization for generalized gradient approximation (GGA). In addition to Trans-Blaha (TB) modified Becke-Johnson (mBJ) potential is incorporated to get better precision for optoelectronic properties. Geometry optimization is carried out to obtain the reliable results of the total energy as well as other structural parameters for each phase of NaZnAs compound. Order of the structural transitions as a function of pressure is found as: Cu2Sb type → β → α phase in our study. Our calculated electronic energy band structures for all structural phases at the level of PBE-GGA as well as mBJ potential point out; NaZnAs compound is a direct (Γ–Γ) band gap semiconductor material. However, as compared to PBE-GGA, mBJ potential approximation reproduces higher values of fundamental band gap. Regarding the optical properties, calculations of real and imaginary parts of the dielectric function, refractive index, reflectivity coefficient, absorption coefficient and energy loss-function spectra are performed over a photon energy ranging from 0.0 to 30.0 eV by polarizing incident radiation in parallel to both [100] and [001] crystalline directions.Keywords: NaZnAs, FP-LAPW+lo, structural properties, phase transition, electronic band-structure, optical properties
Procedia PDF Downloads 4371836 Sound Absorbing and Thermal Insulating Properties of Natural Fibers (Coir/Jute) Hybrid Composite Materials for Automotive Textiles
Authors: Robel Legese Meko
Abstract:
Natural fibers have been used as end-of-life textiles and made into textile products which have become a well-proven and effective way of processing. Nowadays, resources to make primary synthetic fibers are becoming less and less as the world population is rising. Hence it is necessary to develop processes to fabricate textiles that are easily converted to composite materials. Acoustic comfort is closely related to the concept of sound absorption and includes protection against noise. This research paper presents an experimental study on sound absorption coefficients, for natural fiber composite materials: a natural fiber (Coir/Jute) with different blend proportions of raw materials mixed with rigid polyurethane foam as a binder. The natural fiber composite materials were characterized both acoustically (sound absorption coefficient SAC) and also in terms of heat transfer (thermal conductivity). The acoustic absorption coefficient was determined using the impedance tube method according to the ASTM Standard (ASTM E 1050). The influence of the structure of these materials on the sound-absorbing properties was analyzed. The experimental results signify that the porous natural coir/jute composites possess excellent performance in the absorption of high-frequency sound waves, especially above 2000 Hz, and didn’t induce a significant change in the thermal conductivity of the composites. Thus, the sound absorption performances of natural fiber composites based on coir/jute fiber materials promote environmentally friendly solutions.Keywords: coir/jute fiber, sound absorption coefficients, compression molding, impedance tube, thermal insulating properties, SEM analysis
Procedia PDF Downloads 1121835 Relationship between Strengths/Weaknesses of Studying and Subjective Well-Being among Japanese Working Adults
Authors: Katsumi Mera
Abstract:
The purpose of this study was to clarify the relationship between strengths/weaknesses on studying and subjective well-being among Japanese working adults. We conducted an online questionnaire survey of Japanese working adults (aged 20-69) and obtained 1,063 valid responses. Subjective well-being was evaluated using the Cantril Self-Anchoring Striving Scale, the method used in the World Happiness Survey, in which the highest possible life is set at 10, and the lowest possible life is set at 0. The mean subjective well-being of all 1,063 respondents was 5.87, which is a very similar value that was obtained in the Japanese World Happiness Survey in 2023. First, we investigated the respondents' strengths/weaknesses in studying and their subjective well-being. The results showed that the subjective well-being of the group who answered that they were good at studying (6.62) was significantly higher than that of the group who answered that they were poor at studying (5.15). Next, we investigated the subjective well-being of students who answered that they were good at Japanese, Math, English, Social studies, and Science, the major subjects in Japan, and found that the subjective well-being of the group that answered that they were good at these subjects was higher than that of the group that answered that they were poor at these subjects. When the correlation between the score of "strengths/weaknesses on studying" and the score of "subjective well-being" was examined, those who answered that they were good at studying had higher subjective well-being (correlation coefficient r = 0.26, p < 0.001). The correlation between the subjective well-being score and the score of "strengths/weaknesses on studying each subject” was examined. Among the subjects, the subjective well-being score was the highest for those who answered that they were good at English (correlation coefficient r = 0.17, p < 0.001). These results indicate a relationship between strengths/weaknesses on studying and subjective well-being among workers in Japan. These findings may provide an important basis for considering what kind of education should be focused on in order to nurture working adults with high levels of well-being.Keywords: subject well-being, strengths/weaknesses, studying, Japanese working adults
Procedia PDF Downloads 801834 Development of a Direct Immunoassay for Human Ferritin Using Diffraction-Based Sensing Method
Authors: Joel Ballesteros, Harriet Jane Caleja, Florian Del Mundo, Cherrie Pascual
Abstract:
Diffraction-based sensing was utilized in the quantification of human ferritin in blood serum to provide an alternative to label-based immunoassays currently used in clinical diagnostics and researches. The diffraction intensity was measured by the diffractive optics technology or dotLab™ system. Two methods were evaluated in this study: direct immunoassay and direct sandwich immunoassay. In the direct immunoassay, human ferritin was captured by human ferritin antibodies immobilized on an avidin-coated sensor while the direct sandwich immunoassay had an additional step for the binding of a detector human ferritin antibody on the analyte complex. Both methods were repeatable with coefficient of variation values below 15%. The direct sandwich immunoassay had a linear response from 10 to 500 ng/mL which is wider than the 100-500 ng/mL of the direct immunoassay. The direct sandwich immunoassay also has a higher calibration sensitivity with value 0.002 Diffractive Intensity (ng mL-1)-1) compared to the 0.004 Diffractive Intensity (ng mL-1)-1 of the direct immunoassay. The limit of detection and limit of quantification values of the direct immunoassay were found to be 29 ng/mL and 98 ng/mL, respectively, while the direct sandwich immunoassay has a limit of detection (LOD) of 2.5 ng/mL and a limit of quantification (LOQ) of 8.2 ng/mL. In terms of accuracy, the direct immunoassay had a percent recovery of 88.8-93.0% in PBS while the direct sandwich immunoassay had 94.1 to 97.2%. Based on the results, the direct sandwich immunoassay is a better diffraction-based immunoassay in terms of accuracy, LOD, LOQ, linear range, and sensitivity. The direct sandwich immunoassay was utilized in the determination of human ferritin in blood serum and the results are validated by Chemiluminescent Magnetic Immunoassay (CMIA). The calculated Pearson correlation coefficient was 0.995 and the p-values of the paired-sample t-test were less than 0.5 which show that the results of the direct sandwich immunoassay was comparable to that of CMIA and could be utilized as an alternative analytical method.Keywords: biosensor, diffraction, ferritin, immunoassay
Procedia PDF Downloads 3541833 An Approach to Correlate the Statistical-Based Lorenz Method, as a Way of Measuring Heterogeneity, with Kozeny-Carman Equation
Authors: H. Khanfari, M. Johari Fard
Abstract:
Dealing with carbonate reservoirs can be mind-boggling for the reservoir engineers due to various digenetic processes that cause a variety of properties through the reservoir. A good estimation of the reservoir heterogeneity which is defined as the quality of variation in rock properties with location in a reservoir or formation, can better help modeling the reservoir and thus can offer better understanding of the behavior of that reservoir. Most of reservoirs are heterogeneous formations whose mineralogy, organic content, natural fractures, and other properties vary from place to place. Over years, reservoir engineers have tried to establish methods to describe the heterogeneity, because heterogeneity is important in modeling the reservoir flow and in well testing. Geological methods are used to describe the variations in the rock properties because of the similarities of environments in which different beds have deposited in. To illustrate the heterogeneity of a reservoir vertically, two methods are generally used in petroleum work: Dykstra-Parsons permeability variations (V) and Lorenz coefficient (L) that are reviewed briefly in this paper. The concept of Lorenz is based on statistics and has been used in petroleum from that point of view. In this paper, we correlated the statistical-based Lorenz method to a petroleum concept, i.e. Kozeny-Carman equation and derived the straight line plot of Lorenz graph for a homogeneous system. Finally, we applied the two methods on a heterogeneous field in South Iran and discussed each, separately, with numbers and figures. As expected, these methods show great departure from homogeneity. Therefore, for future investment, the reservoir needs to be treated carefully.Keywords: carbonate reservoirs, heterogeneity, homogeneous system, Dykstra-Parsons permeability variations (V), Lorenz coefficient (L)
Procedia PDF Downloads 2221832 Homogenization of a Non-Linear Problem with a Thermal Barrier
Authors: Hassan Samadi, Mustapha El Jarroudi
Abstract:
In this work, we consider the homogenization of a non-linear problem in periodic medium with two periodic connected media exchanging a heat flux throughout their common interface. The interfacial exchange coefficient λ is assumed to tend to zero or to infinity following a rate λ=λ(ε) when the size ε of the basic cell tends to zero. Three homogenized problems are determined according to some critical value depending of λ and ε. Our method is based on Γ-Convergence techniques.Keywords: variational methods, epiconvergence, homogenization, convergence technique
Procedia PDF Downloads 525