Search results for: absolute entropy
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 881

Search results for: absolute entropy

611 Analysis of Creative City Indicators in Isfahan City, Iran

Authors: Reza Mokhtari Malek Abadi, Mohsen Saghaei, Fatemeh Iman

Abstract:

This paper investigates the indices of a creative city in Isfahan. Its main aim is to evaluate quantitative status of the creative city indices in Isfahan city, analyze the dispersion and distribution of these indices in Isfahan city. Concerning these, this study tries to analyze the creative city indices in fifteen area of Isfahan through secondary data, questionnaire, TOPSIS model, Shannon entropy and SPSS. Based on this, the fifteen areas of Isfahan city have been ranked with 12 factors of creative city indices. The results of studies show that fifteen areas of Isfahan city are not equally benefiting from creative indices and there is much difference between the areas of Isfahan city.

Keywords: grading, creative city, creative city evaluation indicators, regional planning model

Procedia PDF Downloads 434
610 Scale-Up Study of Gas-Liquid Two Phase Flow in Downcomer

Authors: Jayanth Abishek Subramanian, Ramin Dabirian, Ilias Gavrielatos, Ram Mohan, Ovadia Shoham

Abstract:

Downcomers are important conduits for multiphase flow transfer from offshore platforms to the seabed. Uncertainty in the predictions of the pressure drop of multiphase flow between platforms is often dominated by the uncertainty associated with the prediction of holdup and pressure drop in the downcomer. The objectives of this study are to conduct experimental and theoretical scale-up study of the downcomer. A 4-in. diameter vertical test section was designed and constructed to study two-phase flow in downcomer. The facility is equipped with baffles for flow area restriction, enabling interchangeable annular slot openings between 30% and 61.7%. Also, state-of-the-art instrumentation, the capacitance Wire-Mesh Sensor (WMS) was utilized to acquire the experimental data. A total of 76 experimental data points were acquired, including falling film under 30% and 61.7% annular slot opening for air-water and air-Conosol C200 oil cases as well as gas carry-under for 30% and 61.7% opening utilizing air-Conosol C200 oil. For all experiments, the parameters such as falling film thickness and velocity, entrained liquid holdup in the core, gas void fraction profiles at the cross-sectional area of the liquid column, the void fraction and the gas carry under were measured. The experimental results indicated that the film thickness and film velocity increase as the flow area reduces. Also, the increase in film velocity increases the gas entrainment process. Furthermore, the results confirmed that the increase of gas entrainment for the same liquid flow rate leads to an increase in the gas carry-under. A power comparison method was developed to enable evaluation of the Lopez (2011) model, which was created for full bore downcomer, with the novel scale-up experiment data acquired from the downcomer with the restricted area for flow. Comparison between the experimental data and the model predictions shows a maximum absolute average discrepancy of 22.9% and 21.8% for the falling film thickness and velocity, respectively; and a maximum absolute average discrepancy of 22.2% for fraction of gas carried with the liquid (oil).

Keywords: two phase flow, falling film, downcomer, wire-mesh sensor

Procedia PDF Downloads 137
609 Different Data-Driven Bivariate Statistical Approaches to Landslide Susceptibility Mapping (Uzundere, Erzurum, Turkey)

Authors: Azimollah Aleshzadeh, Enver Vural Yavuz

Abstract:

The main goal of this study is to produce landslide susceptibility maps using different data-driven bivariate statistical approaches; namely, entropy weight method (EWM), evidence belief function (EBF), and information content model (ICM), at Uzundere county, Erzurum province, in the north-eastern part of Turkey. Past landslide occurrences were identified and mapped from an interpretation of high-resolution satellite images, and earlier reports as well as by carrying out field surveys. In total, 42 landslide incidence polygons were mapped using ArcGIS 10.4.1 software and randomly split into a construction dataset 70 % (30 landslide incidences) for building the EWM, EBF, and ICM models and the remaining 30 % (12 landslides incidences) were used for verification purposes. Twelve layers of landslide-predisposing parameters were prepared, including total surface radiation, maximum relief, soil groups, standard curvature, distance to stream/river sites, distance to the road network, surface roughness, land use pattern, engineering geological rock group, topographical elevation, the orientation of slope, and terrain slope gradient. The relationships between the landslide-predisposing parameters and the landslide inventory map were determined using different statistical models (EWM, EBF, and ICM). The model results were validated with landslide incidences, which were not used during the model construction. In addition, receiver operating characteristic curves were applied, and the area under the curve (AUC) was determined for the different susceptibility maps using the success (construction data) and prediction (verification data) rate curves. The results revealed that the AUC for success rates are 0.7055, 0.7221, and 0.7368, while the prediction rates are 0.6811, 0.6997, and 0.7105 for EWM, EBF, and ICM models, respectively. Consequently, landslide susceptibility maps were classified into five susceptibility classes, including very low, low, moderate, high, and very high. Additionally, the portion of construction and verification landslides incidences in high and very high landslide susceptibility classes in each map was determined. The results showed that the EWM, EBF, and ICM models produced satisfactory accuracy. The obtained landslide susceptibility maps may be useful for future natural hazard mitigation studies and planning purposes for environmental protection.

Keywords: entropy weight method, evidence belief function, information content model, landslide susceptibility mapping

Procedia PDF Downloads 104
608 The Effects of Subjective and Objective Indicators of Inequality on Life Satisfaction in a Comparative Perspective Using a Multi-Level Analysis

Authors: Atefeh Bagherianziarat, Dana Hamplova

Abstract:

The inverse social gradient in life satisfaction (LS) is a well-established research finding. To estimate the influence of inequality on LS, most of the studies have explored the effect of the objective aspects of inequality or individuals’ socioeconomic status (SES). However, relatively fewer studies have confirmed recently the significant effect of the subjective aspect of inequality or subjective socioeconomic status (SSS) on life satisfaction over and above SES. In other words, it is confirmed by some studies that individuals’ perception of their unequal status in society or SSS can moderate the impact of their absolute unequal status on their life satisfaction. Nevertheless, this newly confirmed moderating link has not been affirmed to work likewise in societies with different levels of social inequality and also for people who believe in the value of equality, at different levels. In this study, we compared the moderative influence of subjective inequality on the link between objective inequality and life satisfaction. In particular, we focus on differences across welfare state regimes based on Esping-Andersen's theory. Also, we explored the moderative role of believing in the value of equality on the link between objective and subjective inequality on LS in the given societies. Since our studied variables were measured at both individual and country levels, we applied a multilevel analysis to the European Social Survey data (round 9). The results showed that people in deferent regimes reported statistically meaningful different levels of life satisfaction that is explained to different extends by their household income and their perception of their income inequality. The findings of the study supported the previous findings of the moderator influence of perceived inequality on the link between objective inequality and LS. However, this link is different in various welfare state regimes. The results of the multilevel modeling showed that country-level subjective equality is a positive predictor for individuals’ life satisfaction, while the GINI coefficient that was considered as the indicator of absolute inequality has a smaller effect on life satisfaction. Also, country-level subjective equality moderates the confirmed link between individuals’ income and their life satisfaction. It can be concluded that both individual and country-level subjective inequality slightly moderate the effect of individuals’ income on their life satisfaction.

Keywords: individual values, life satisfaction, multilevel analysis, objective inequality, subjective inequality, welfare regimes status

Procedia PDF Downloads 69
607 Social Structure, Involuntary Relations and Urban Poverty

Authors: Mahmood Niroobakhsh

Abstract:

This article deals with special structuralism approaches to explain a certain kind of social problem. Widespread presence of poverty is a reminder of deep-rooted unresolved problems of social relations. The expected role from an individual for the social system recognizes poverty derived from an interrelated social structure. By the time, enabled to act on his role in the course of social interaction, reintegration of the poor in society may take place. Poverty and housing type are reflections of the underlying social structure, primarily structure’s elements, systemic interrelations, and the overall strength or weakness of that structure. Poverty varies based on social structure in that the stronger structures are less likely to produce poverty.

Keywords: absolute poverty, relative poverty, social structure, urban poverty

Procedia PDF Downloads 641
606 The ‘Quartered Head Technique’: A Simple, Reliable Way of Maintaining Leg Length and Offset during Total Hip Arthroplasty

Authors: M. Haruna, O. O. Onafowokan, G. Holt, K. Anderson, R. G. Middleton

Abstract:

Background: Requirements for satisfactory outcomes following total hip arthroplasty (THA) include restoration of femoral offset, version, and leg length. Various techniques have been described for restoring these biomechanical parameters, with leg length restoration being the most predominantly described. We describe a “quartered head technique” (QHT) which uses a stepwise series of femoral head osteotomies to identify and preserve the centre of rotation of the femoral head during THA in order to ensure reconstruction of leg length, offset and stem version, such that hip biomechanics are restored as near to normal as possible. This study aims to identify whether using the QHT during hip arthroplasty effectively restores leg length and femoral offset to within acceptable parameters. Methods: A retrospective review of 206 hips was carried out, leaving 124 hips in the final analysis. Power analysis indicated a minimum of 37 patients required. All operations were performed using an anterolateral approach by a single surgeon. All femoral implants were cemented, collarless, polished double taper CPT® stems (Zimmer, Swindon, UK). Both cemented, and uncemented acetabular components were used (Zimmer, Swindon, UK). Leg length, version, and offset were assessed intra-operatively and reproduced using the QHT. Post-operative leg length and femoral offset were determined and compared with the contralateral native hip, and the difference was then calculated. For the determination of leg length discrepancy (LLD), we used the method described by Williamson & Reckling, which has been shown to be reproducible with a measurement error of ±1mm. As a reference, the inferior margin of the acetabular teardrop and the most prominent point of the lesser trochanter were used. A discrepancy of less than 6mm LLD was chosen as acceptable. All peri-operative radiographs were assessed by two independent observers. Results: The mean absolute post-operative difference in leg length from the contralateral leg was +3.58mm. 84% of patients (104/124) had LLD within ±6mm of the contralateral limb. The mean absolute post-operative difference in offset from contralateral leg was +3.88mm (range -15 to +9mm, median 3mm). 90% of patients (112/124) were within ±6mm offset of the contralateral limb. There was no statistical difference noted between observer measurements. Conclusion: The QHT provides a simple, inexpensive yet effective method of maintaining femoral leg length and offset during total hip arthroplasty. Combining this technique with pre-operative templating or other techniques described may enable surgeons to reduce even further the discrepancies between pre-operative state and post-operative outcome.

Keywords: leg length discrepancy, technical tip, total hip arthroplasty, operative technique

Procedia PDF Downloads 50
605 Effect of Co Substitution on Structural, Magnetocaloric, Magnetic, and Electrical Properties of Sm0.6Sr0.4CoxMn1-xO3 Synthesized by Sol-gel Method

Authors: A. A. Azab

Abstract:

In this work, Sm0.6Sr0.4CoxMn1-xO3 (x=0, 0.1, 0.2 and 0.3) was synthesized by sol-gel method for magnetocaloric effect (MCE) applications. XRD analysis confirmed formation of the required orthorhombic phase of perovskite, and there is crystallographic phase transition as a result of substitution. Maxwell-Wagner interfacial polarisation and Koops phenomenological theory were used to investigate and analyze the temperature and frequency dependency of the dielectric permittivity. The phase transition from the ferromagnetic to the paramagnetic state was demonstrated to be second order. Based on the isothermal magnetization curves obtained at various temperatures, the magnetic entropy change was calculated. A magnetocaloric effect (MCE) over a wide temperature range was studied by determining DSM and the relative cooling power (RCP).

Keywords: magnetocaloric effect, pperovskite, magnetic phase transition, dielectric permittivity

Procedia PDF Downloads 40
604 Biogas Potential of Deinking Sludge from Wastepaper Recycling Industry: Influence of Dewatering Degree and High Calcium Carbonate Content

Authors: Moses Kolade Ogun, Ina Korner

Abstract:

To improve on the sustainable resource management in the wastepaper recycling industry, studies into the valorization of wastes generated by the industry are necessary. The industry produces different residues, among which is the deinking sludge (DS). The DS is generated from the deinking process and constitutes a major fraction of the residues generated by the European pulp and paper industry. The traditional treatment of DS by incineration is capital intensive due to energy requirement for dewatering and the need for complementary fuel source due to DS low calorific value. This could be replaced by a biotechnological approach. This study, therefore, investigated the biogas potential of different DS streams (different dewatering degrees) and the influence of the high calcium carbonate content of DS on its biogas potential. Dewatered DS (solid fraction) sample from filter press and the filtrate (liquid fraction) were collected from a partner wastepaper recycling company in Germany. The solid fraction and the liquid fraction were mixed in proportion to realize DS with different water content (55–91% fresh mass). Spiked samples of DS using deionized water, cellulose and calcium carbonate were prepared to simulate DS with varying calcium carbonate content (0– 40% dry matter). Seeding sludge was collected from an existing biogas plant treating sewage sludge in Germany. Biogas potential was studied using a 1-liter batch test system under the mesophilic condition and ran for 21 days. Specific biogas potential in the range 133- 230 NL/kg-organic dry matter was observed for DS samples investigated. It was found out that an increase in the liquid fraction leads to an increase in the specific biogas potential and a reduction in the absolute biogas potential (NL-biogas/ fresh mass). By comparing the absolute biogas potential curve and the specific biogas potential curve, an optimal dewatering degree corresponding to a water content of about 70% fresh mass was identified. This degree of dewatering is a compromise when factors such as biogas yield, reactor size, energy required for dewatering and operation cost are considered. No inhibitory influence was observed in the biogas potential of DS due to the reported high calcium carbonate content of DS. This study confirms that DS is a potential bioresource for biogas production. Further optimization such as nitrogen supplementation due to DS high C/N ratio can increase biogas yield.

Keywords: biogas, calcium carbonate, deinking sludge, dewatering, water content

Procedia PDF Downloads 132
603 The Relationship between Ruins and Vegetation: Different Approaches during the Centuries and within the Various Disciplinary Fields, Investigation of Writings and Projects

Authors: Rossana Mancini

Abstract:

The charm of a ruin colonised by wild plants and flowers is part of Western culture. The relationship between ruins and vegetation involves a wide range of different fields of research. During the first phase of the research the most important writings and projects about this argument were investigated, to understand how the perception of the co-existence of ruins and vegetation has changed over time and to investigate the various different approaches that these different fields have adopted when tackling this issue. The paper presents some practical examples of projects carried out from the early 1900s on. The major result is that specifically regards conservation, the best attitude is the management of change, an inevitable process when it comes to the co-existence of ruins and nature and, particularly, ruins and vegetation. Limiting ourselves to adopting measures designed to stop, or rather slow down, the increasing level of entropy (and therefore disorder) may not be enough.

Keywords: ruins, vegetation, conservation, archaeology, architecture

Procedia PDF Downloads 300
602 Phase Transitions of Cerium and Neodymium

Authors: M. Khundadze, V. Varazashvili, N. Lejava, R. Jorbenadze

Abstract:

Phase transitions of cerium and neodymium are investigated by using high-temperature scanning calorimeter (HT-1500 Seteram). For cerium two types of transformation are detected: at 350-372 K - hexagonal close packing (hcp) - face-centered cubic lattice (fcc) transition, and at 880-960K the face-centered cubic lattice (fcc) transformation into body-centered cubic lattice (bcc). For neodymium changing of hexagonal close packing (hcp) into the body-centered cubic lattice (bcc) is detected at 1093-1113K. The thermal characteristics of transitions – enthalpy, entropy, temperature domains – are reported.

Keywords: cerium, calorimetry, enthalpy of phase transitions, neodymium

Procedia PDF Downloads 294
601 Shoulder Range of Motion Measurements using Computer Vision Compared to Hand-Held Goniometric Measurements

Authors: Lakshmi Sujeesh, Aaron Ramzeen, Ricky Ziming Guo, Abhishek Agrawal

Abstract:

Introduction: Range of motion (ROM) is often measured by physiotherapists using hand-held goniometer as part of mobility assessment for diagnosis. Due to the nature of hand-held goniometer measurement procedure, readings often tend to have some variations depending on the physical therapist taking the measurements (Riddle et al.). This study aims to validate computer vision software readings against goniometric measurements for quick and consistent ROM measurements to be taken by clinicians. The use of this computer vision software hopes to improve the future of musculoskeletal space with more efficient diagnosis from recording of patient’s ROM with minimal human error across different physical therapists. Methods: Using the hand-held long arm goniometer measurements as the “gold-standard”, healthy study participants (n = 20) were made to perform 4 exercises: Front elevation, Abduction, Internal Rotation, and External Rotation, using both arms. Assessment of active ROM using computer vision software at different angles set by goniometer for each exercise was done. Interclass Correlation Coefficient (ICC) using 2-way random effects model, Box-Whisker plots, and Root Mean Square error (RMSE) were used to find the degree of correlation and absolute error measured between set and recorded angles across the repeated trials by the same rater. Results: ICC (2,1) values for all 4 exercises are above 0.9, indicating excellent reliability. Lowest overall RMSE was for external rotation (5.67°) and highest for front elevation (8.00°). Box-whisker plots showed have showed that there is a potential zero error in the measurements done by the computer vision software for abduction, where absolute error for measurements taken at 0 degree are shifted away from the ideal 0 line, with its lowest recorded error being 8°. Conclusion: Our results indicate that the use of computer vision software is valid and reliable to use in clinical settings by physiotherapists for measuring shoulder ROM. Overall, computer vision helps improve accessibility to quality care provided for individual patients, with the ability to assess ROM for their condition at home throughout a full cycle of musculoskeletal care (American Academy of Orthopaedic Surgeons) without the need for a trained therapist.

Keywords: physiotherapy, frozen shoulder, joint range of motion, computer vision

Procedia PDF Downloads 68
600 Image Steganography Using Predictive Coding for Secure Transmission

Authors: Baljit Singh Khehra, Jagreeti Kaur

Abstract:

In this paper, steganographic strategy is used to hide the text file inside an image. To increase the storage limit, predictive coding is utilized to implant information. In the proposed plan, one can exchange secure information by means of predictive coding methodology. The predictive coding produces high stego-image. The pixels are utilized to insert mystery information in it. The proposed information concealing plan is powerful as contrasted with the existing methodologies. By applying this strategy, a provision helps clients to productively conceal the information. Entropy, standard deviation, mean square error and peak signal noise ratio are the parameters used to evaluate the proposed methodology. The results of proposed approach are quite promising.

Keywords: cryptography, steganography, reversible image, predictive coding

Procedia PDF Downloads 390
599 Mechanical Properties of Recycled Plasticized PVB/PVC Blends

Authors: Michael Tupý, Dagmar Měřínská, Alice Tesaříková-Svobodová, Christian Carrot, Caroline Pillon, Vít Petránek

Abstract:

The mechanical properties of blends consisting of plasticized poly(vinyl butyral) (PVB) and plasticized poly(vinyl chloride) (PVC) are studied, in order to evaluate the possibility of using recycled PVB waste derived from windshields. PVC was plasticized with 38% of diisononyl phthalate (DINP), while PVB was plasticized with 28% of triethylene glycol, bis(2-ethylhexanoate) (3GO). The optimal process conditions for the PVB/PVC blend in 1:1 ratio were determined. Entropy was used in order to theoretically predict the blends miscibility. The PVB content of each blend composition used was ranging from zero to 100%. Tensile strength and strain were tested. In addition, a comparison between recycled and original PVB, used as constituents of the blend, was performed.

Keywords: poly(vinyl butyral), poly(vinyl chloride), windshield, polymer waste, mechanical properties

Procedia PDF Downloads 414
598 Modelling and Control of Electrohydraulic System Using Fuzzy Logic Algorithm

Authors: Hajara Abdulkarim Aliyu, Abdulbasid Ismail Isa

Abstract:

This research paper studies electrohydraulic system for its role in position and motion control system and develops as mathematical model describing the behaviour of the system. The research further proposes Fuzzy logic and conventional PID controllers in order to achieve both accurate positioning of the payload and overall improvement of the system performance. The simulation result shows Fuzzy logic controller has a superior tracking performance and high disturbance rejection efficiency for its shorter settling time, less overshoot, smaller values of integral of absolute and deviation errors over the conventional PID controller at all the testing conditions.

Keywords: electrohydraulic, fuzzy logic, modelling, NZ-PID

Procedia PDF Downloads 428
597 Brain Age Prediction Based on Brain Magnetic Resonance Imaging by 3D Convolutional Neural Network

Authors: Leila Keshavarz Afshar, Hedieh Sajedi

Abstract:

Estimation of biological brain age from MR images is a topic that has been much addressed in recent years due to the importance it attaches to early diagnosis of diseases such as Alzheimer's. In this paper, we use a 3D Convolutional Neural Network (CNN) to provide a method for estimating the biological age of the brain. The 3D-CNN model is trained by MRI data that has been normalized. In addition, to reduce computation while saving overall performance, some effectual slices are selected for age estimation. By this method, the biological age of individuals using selected normalized data was estimated with Mean Absolute Error (MAE) of 4.82 years.

Keywords: brain age estimation, biological age, 3D-CNN, deep learning, T1-weighted image, SPM, preprocessing, MRI, canny, gray matter

Procedia PDF Downloads 117
596 Performance Evaluation of the CSAN Pronto Point-of-Care Whole Blood Analyzer for Regular Hematological Monitoring During Clozapine Treatment

Authors: Farzana Esmailkassam, Usakorn Kunanuvat, Zahraa Mohammed Ali

Abstract:

Objective: The key barrier in Clozapine treatment of treatment-resistant schizophrenia (TRS) includes frequent bloods draws to monitor neutropenia, the main drug side effect. WBC and ANC monitoring must occur throughout treatment. Accurate WBC and ANC counts are necessary for clinical decisions to halt, modify or continue clozapine treatment. The CSAN Pronto point-of-care (POC) analyzer generates white blood cells (WBC) and absolute neutrophils (ANC) through image analysis of capillary blood. POC monitoring offers significant advantages over central laboratory testing. This study evaluated the performance of the CSAN Pronto against the Beckman DxH900 Hematology laboratory analyzer. Methods: Forty venous samples (EDTA whole blood) with varying concentrations of WBC and ANC as established on the DxH900 analyzer were tested in duplicates on three CSAN Pronto analyzers. Additionally, both venous and capillary samples were concomitantly collected from 20 volunteers and assessed on the CSAN Pronto and the DxH900 analyzer. The analytical performance including precision using liquid quality controls (QCs) as well as patient samples near the medical decision points, and linearity using a mix of high and low patient samples to create five concentrations was also evaluated. Results: In the precision study for QCs and whole blood, WBC and ANC showed CV inside the limits established according to manufacturer and laboratory acceptability standards. WBC and ANC were found to be linear across the measurement range with a correlation of 0.99. WBC and ANC from all analyzers correlated well in venous samples on the DxH900 across the tested sample ranges with a correlation of > 0.95. Mean bias in ANC obtained on the CSAN pronto versus the DxH900 was 0.07× 109 cells/L (95% L.O.A -0.25 to 0.49) for concentrations <4.0 × 109 cells/L, which includes decision-making cut-offs for continuing clozapine treatment. Mean bias in WBC obtained on the CSAN pronto versus the DxH900 was 0.34× 109 cells/L (95% L.O.A -0.13 to 0.72) for concentrations <5.0 × 109 cells/L. The mean bias was higher (-11% for ANC, 5% for WBC) at higher concentrations. The correlations between capillary and venous samples showed more variability with mean bias of 0.20 × 109 cells/L for the ANC. Conclusions: The CSAN pronto showed acceptable performance in WBC and ANC measurements from venous and capillary samples and was approved for clinical use. This testing will facilitate treatment decisions and improve clozapine uptake and compliance.

Keywords: absolute neutrophil counts, clozapine, point of care, white blood cells

Procedia PDF Downloads 56
595 Early Modern Controversies of Mobility within the Spanish Empire: Francisco De Vitoria and the Peaceful Right to Travel

Authors: Beatriz Salamanca

Abstract:

In his public lecture ‘On the American Indians’ given at the University of Salamanca in 1538-39, Francisco de Vitoria presented an unsettling defense of freedom of movement, arguing that the Spanish had the right to travel and dwell in the New World, since it was considered part of the law of nations [ius gentium] that men enjoyed free mutual intercourse anywhere they went. The principle of freedom of movement brought hopeful expectations, promising to bring mankind together and strengthen the ties of fraternity. However, it led to polemical situations when those whose mobility was in question represented a harmful threat or was for some reason undesired. In this context, Vitoria’s argument has been seen on multiple occasions as a justification of the expansion of the Spanish empire. In order to examine the meaning of Vitoria’s defense of free mobility, a more detailed look at Vitoria’s text is required, together with the study of some of his earliest works, among them, his commentaries on Thomas Aquinas’s Summa Theologiae, where he presented relevant insights on the idea of the law of nations. In addition, it is necessary to place Vitoria’s work in the context of the intellectual tradition he belonged to and the responses he obtained from some of his contemporaries who were concerned with similar issues. The claim of this research is that the Spanish right to travel advocated by Vitoria was not intended to be interpreted in absolute terms, for it had to serve the purpose of bringing peace and unity among men, and could not contradict natural law. In addition, Vitoria explicitly observed that the right to travel was only valid if the Spaniards caused no harm, a condition that has been underestimated by his critics. Therefore, Vitoria’s legacy is of enormous value as it initiated a long lasting discussion regarding the question of the grounds under which human mobility could be restricted. Again, under Vitoria’s argument it was clear that this freedom was not absolute, but the controversial nature of his defense of Spanish mobility demonstrates how difficult it was and still is to address the issue of the circulation of peoples across frontiers, and shows the significance of this discussion in today’s globalized world, where the rights and wrongs of notions like immigration, international trade or foreign intervention still lack sufficient consensus. This inquiry about Vitoria’s defense of the principle of freedom of movement is being placed here against the background of the history of political thought, political theory, international law, and international relations, following the methodological framework of contextual history of the ‘Cambridge School’.

Keywords: Francisco de Vitoria, freedom of movement, law of nations, ius gentium, Spanish empire

Procedia PDF Downloads 337
594 Size-Reduction Strategies for Iris Codes

Authors: Jutta Hämmerle-Uhl, Georg Penn, Gerhard Pötzelsberger, Andreas Uhl

Abstract:

Iris codes contain bits with different entropy. This work investigates different strategies to reduce the size of iris code templates with the aim of reducing storage requirements and computational demand in the matching process. Besides simple sub-sampling schemes, also a binary multi-resolution representation as used in the JBIG hierarchical coding mode is assessed. We find that iris code template size can be reduced significantly while maintaining recognition accuracy. Besides, we propose a two stage identification approach, using small-sized iris code templates in a pre-selection satge, and full resolution templates for final identification, which shows promising recognition behaviour.

Keywords: iris recognition, compact iris code, fast matching, best bits, pre-selection identification, two-stage identification

Procedia PDF Downloads 414
593 The Investigation of Effectiveness of Different Concentrations of the Mycotoxin Detoxification Agent Added to Broiler Feed, in the Presence of T-2 Toxin, on Performance, Organ Mass and the Residues T-2 Toxin and His Metabolites in the Broiler Tissues

Authors: Jelena Nedeljković Trailović, Marko Vasiljević, Jog Raj, Hunor Farkaš, Branko Petrujkić, Stamen Radulović, Gorana Popvić

Abstract:

The experiment was performed on a total of 99 one-day-old broilers of Cob 500 provenance, which were divided into IX equal groups. Broilers of the E-I group were fed 0.25 mg T-2 toxin/kg feed, E-II and E-III groups 0.25 mg T-2 toxin/kg feed with the addition of 1 kg/t and 3 kg/t of the mycotoxin detoxification agent MDA, respectively. The E-IV group received 1 mg of T-2 toxin/kg of feed, and the broilers of E-V and E-VI groups received 1 mg of T-2 toxin/kg of feed with the addition of 1 kg/t and 3 kg/t of the MDA detoxification preparation, respectively. The E-VII group received commercial feed without toxins and additives, the E-VIII and E-IX groups received feed with 1kg/t and 3kg/t of the MDA detoxification preparation. The trial lasted 42 days. Observing the results obtained on the 42nd day of the experiment, we can conclude that the change in the absolute mass of the spleen occurred in the broilers of the E-IV group (1.66±0.14)g, which was statistically significantly lower compared to the broilers of the E-V and E-VI groups (2.58±0.15 and 2.68±0.23)g. Heart mass was significantly statistically lower in broilers of group E-IV (9.1±0.38)g compared to broilers of group E-V and E-VI (12.23±0.5 and 11.43±0.51)g. It can be concluded that the broilers that received 1 kg/t and 3 kg/t of the detoxification preparation had an absolute mass of organs within physiological limits. Broilers of the E-IV group achieved the lowest BM during the experiment (on the 42nd day of the experiment 1879±52.73)g, they were significantly statistically lower than the BW of broilers of all experimental groups. This trend is observed from the beginning to the end of the experiment. The protective effect of the detoxification preparation can be seen in broilers of the E-V group, that had a significantly statistically higher BM on the 42nd day of the experiment (2225±58.81)g compared to broilers of group E-IV. Broilers of E-VIII group (2452±46.71) g, which received commercial feed with the addition of 1 kg/t MDA preparation, had the highest BMI at the end of the experiment. At the end of the trial on the 42nd day, blood samples were collected from broilers of the experimental groups that received T-2 toxin and MR detoxification preparations in different concentrations. Also, liver and breast musculature samples were collected for testing for the presence and content of T-2 toxin, HT-2 toxin, T-2 tetraol and T-2 triol. Due to very rapid elimination from the blood, no remains of T-2 toxin and its metabolites were detected in the blood of broilers of groups E-I to E-VI. In the breast muscles, T-2 toxin residues below LoQ < 0.2 (μg/kg) were detected in all groups that received T-2 toxin in food, the highest value was recorded in the E-IV group (0.122 μg/kg and the lowest in E -VI group 0.096 μg/kg). No T-2 toxin residues were detected in the liver. Remains of HT-2 were detected in the breast muscles and livers of broilers from E-IV, E-V and E-VI groups, LoQ < 1 (μg/kg); for the breast muscles: 0.054, 0.044 and 0.041 μg/kg, and for the liver: 0.473, 0.231 and 0.185 μg/kg. Summing up all the results, a partial protective effect of the detoxification preparation, added to food in the amount of 1kg/t, can be seen.

Keywords: T-2 toxin, bloiler, MDA, mycotoxuns

Procedia PDF Downloads 35
592 Constant-Roll Warm Inflation within Rastall Gravity

Authors: Rabia Saleem

Abstract:

This research has a recently proposed strategy to find the exact inflationary solution of the Friedman equations in the context of the Rastall theory of gravity (RTG), known as constant-roll warm inflation, including dissipation effects. We establish the model to evaluate the effective potential of inflation and entropy. We develop the inflationary observable like scalar-tensor power spectra, scalar-tensor spectral indices, tensor-to-scalar ratio, and running of spectral-index. The theory parameter $\lambda$ is constrained to observe the compatibility of our model with Planck 2013, Planck TT, TE, EE+lowP (2015), and Planck 2018 bounds. The results are feasible and interesting up to the 2$\sigma$ confidence level.

Keywords: modified gravity, warm inflation, constant-roll limit, dissipation

Procedia PDF Downloads 66
591 Thermal Effects of Phase Transitions of Cerium and Neodymium

Authors: M. Khundadze, V. Varazashvili, N. Lejava, R. Jorbenadze

Abstract:

Phase transitions of cerium and neodymium are investigated by using high temperature scanning calorimeter (HT-1500 Seteram). For cerium two types of transformation are detected: at 350-372 K - hexagonal close packing (hcp) - face-centered cubic lattice (fcc) transition, and in 880-960K the face-centered cubic lattice (fcc) transformation into body-centered cubic lattice (bcc). For neodymium changing of hexagonal close packing (hcp) into body-centered cubic lattice (bcc) is detected at 1093-1113K. The thermal characteristics of transitions – enthalpy, entropy, temperature domains – are reported.

Keywords: cerium, calorimetry, neodymium, enthalpy of phase transitions, neodymium

Procedia PDF Downloads 335
590 Improvement a Lower Bound of Energy for Some Family of Graphs, Related to Determinant of Adjacency Matrix

Authors: Saieed Akbari, Yousef Bagheri, Amir Hossein Ghodrati, Sima Saadat Akhtar

Abstract:

Let G be a simple graph with the vertex set V (G) and with the adjacency matrix A (G). The energy E (G) of G is defined to be the sum of the absolute values of all eigenvalues of A (G). Also let n and m be number of edges and vertices of the graph respectively. A regular graph is a graph where each vertex has the same number of neighbours. Given a graph G, its line graph L(G) is a graph such that each vertex of L(G) represents an edge of G; and two vertices of L(G) are adjacent if and only if their corresponding edges share a common endpoint in G. In this paper we show that for every regular graphs and also for every line graphs such that (G) 3 we have, E(G) 2nm + n 1. Also at the other part of the paper we prove that 2 (G) E(G) for an arbitrary graph G.

Keywords: eigenvalues, energy, line graphs, matching number

Procedia PDF Downloads 198
589 Removal of Tartrazine Dye Form Aqueous Solutions by Adsorption on the Surface of Polyaniline/Iron Oxide Composite

Authors: Salem Ali Jebreil

Abstract:

In this work, a polyaniline/Iron oxide (PANI/Fe2O3) composite was chemically prepared by oxidative polymerization of aniline in acid medium, in presence of ammonium persulphate as an oxidant and amount of Fe2O3. The composite was characterized by a scanning electron microscopy (SEM). The prepared composite has been used as adsorbent to remove Tartrazine dye form aqueous solutions. The effects of initial dye concentration and temperature on the adsorption capacity of PANI/Fe2O3 for Tartrazine dye have been studied in this paper. The Langmuir and Freundlich adsorption models have been used for the mathematical description of adsorption equilibrium data. The best fit is obtained using the Freundlich isotherm with an R2 value of 0.998. The change of Gibbs energy, enthalpy, and entropy of adsorption has been also evaluated for the adsorption of Tartrazine onto PANI/ Fe2O3. It has been proved according the results that the adsorption process is endothermic in nature.

Keywords: adsorption, composite, dye, polyaniline, tartrazine

Procedia PDF Downloads 249
588 Facial Recognition on the Basis of Facial Fragments

Authors: Tetyana Baydyk, Ernst Kussul, Sandra Bonilla Meza

Abstract:

There are many articles that attempt to establish the role of different facial fragments in face recognition. Various approaches are used to estimate this role. Frequently, authors calculate the entropy corresponding to the fragment. This approach can only give approximate estimation. In this paper, we propose to use a more direct measure of the importance of different fragments for face recognition. We propose to select a recognition method and a face database and experimentally investigate the recognition rate using different fragments of faces. We present two such experiments in the paper. We selected the PCNC neural classifier as a method for face recognition and parts of the LFW (Labeled Faces in the Wild) face database as training and testing sets. The recognition rate of the best experiment is comparable with the recognition rate obtained using the whole face.

Keywords: face recognition, labeled faces in the wild (LFW) database, random local descriptor (RLD), random features

Procedia PDF Downloads 329
587 Magnetic, Magnetocaloric, and Electrical Properties of Pr0.7Ca0.3Mn0.9M0.1O3

Authors: A. Selmi, A. Bettaibi, H. Rahmouni, R. M’nassri, N. Chniba Boudjada, A. Chiekhrouhou, K. Khirouni

Abstract:

Investigation of magnetic and magnetocaloric properties of Pr₀.₇Ca₀.₃Mn₀.₉M₀.₁O₃ perovskite manganites (M=Cr and Ni) has been carried out. Our compounds were prepared by the conventional solid-state reaction method at high temperatures. Rietveld refinement of X-ray diffraction pattern using FULLPROF method shows that all compounds adopt the orthorhombic structure with Pnma space group. The partial substitution of Mn-site drives the system from charge order state to ferromagnetic one with a Curie temperature T𝒸=150K, 118k and 116K for M=Cr and Ni, respectively. Magnetization measurements versus temperature in a magnetic applied field of 0.05T show that all our samples exhibit a paramagnetic–ferromagnetic transition with decreasing temperature. From M(H) isotherms, we have deduced the magnetic entropy change, which present maximum values of 2.37 J/kg.K and 2.94 J/kg.K, in a magnetic field change of 5T for M=Cr and Ni, respectively.

Keywords: manganites, magnetocaloric, magnetic, refrigeration

Procedia PDF Downloads 49
586 Accessibility of the Labor Market in Indonesian Cities

Authors: Hananto Prakoso, Jean-Pierre Orfeuil

Abstract:

The relationship between city size, urban transport efficiency (speed), employment proximity (distance) and accessibility of labour market is rarely examined especially in developing countries. This paper reveals the relationship using 2 points of views (active population and company). Then the analysis is divided according to 3 transport modes (car, public transport and motorcycle) and takes into account the vehicle ownership rate. We employ data across 111 districts in 4 big cities of Indonesia. In our result, speed indicator contributed positively to accessibility of labour market while distance elasticity is negative. In absolute value, elasticity of speed indicator is higher than that of distance.

Keywords: labour market, travel time, travel cost threshold, transportation

Procedia PDF Downloads 346
585 Adaptive Data Approximations Codec (ADAC) for AI/ML-based Cyber-Physical Systems

Authors: Yong-Kyu Jung

Abstract:

The fast growth in information technology has led to de-mands to access/process data. CPSs heavily depend on the time of hardware/software operations and communication over the network (i.e., real-time/parallel operations in CPSs (e.g., autonomous vehicles). Since data processing is an im-portant means to overcome the issue confronting data management, reducing the gap between the technological-growth and the data-complexity and channel-bandwidth. An adaptive perpetual data approximation method is intro-duced to manage the actual entropy of the digital spectrum. An ADAC implemented as an accelerator and/or apps for servers/smart-connected devices adaptively rescales digital contents (avg.62.8%), data processing/access time/energy, encryption/decryption overheads in AI/ML applications (facial ID/recognition).

Keywords: adaptive codec, AI, ML, HPC, cyber-physical, cybersecurity

Procedia PDF Downloads 50
584 Suitability of Black Box Approaches for the Reliability Assessment of Component-Based Software

Authors: Anjushi Verma, Tirthankar Gayen

Abstract:

Although, reliability is an important attribute of quality, especially for mission critical systems, yet, there does not exist any versatile model even today for the reliability assessment of component-based software. The existing Black Box models are found to make various assumptions which may not always be realistic and may be quite contrary to the actual behaviour of software. They focus on observing the manner in which the system behaves without considering the structure of the system, the components composing the system, their interconnections, dependencies, usage frequencies, etc.As a result, the entropy (uncertainty) in assessment using these models is much high.Though, there are some models based on operation profile yet sometimes it becomes extremely difficult to obtain the exact operation profile concerned with a given operation. This paper discusses the drawbacks, deficiencies and limitations of Black Box approaches from the perspective of various authors and finally proposes a conceptual model for the reliability assessment of software.

Keywords: black box, faults, failure, software reliability

Procedia PDF Downloads 420
583 Tracing Sources of Sediment in an Arid River, Southern Iran

Authors: Hesam Gholami

Abstract:

Elevated suspended sediment loads in riverine systems resulting from accelerated erosion due to human activities are a serious threat to the sustainable management of watersheds and ecosystem services therein worldwide. Therefore, mitigation of deleterious sediment effects as a distributed or non-point pollution source in the catchments requires reliable provenance information. Sediment tracing or sediment fingerprinting, as a combined process consisting of sampling, laboratory measurements, different statistical tests, and the application of mixing or unmixing models, is a useful technique for discriminating the sources of sediments. From 1996 to the present, different aspects of this technique, such as grouping the sources (spatial and individual sources), discriminating the potential sources by different statistical techniques, and modification of mixing and unmixing models, have been introduced and modified by many researchers worldwide, and have been applied to identify the provenance of fine materials in agricultural, rural, mountainous, and coastal catchments, and in large catchments with numerous lakes and reservoirs. In the last two decades, efforts exploring the uncertainties associated with sediment fingerprinting results have attracted increasing attention. The frameworks used to quantify the uncertainty associated with fingerprinting estimates can be divided into three groups comprising Monte Carlo simulation, Bayesian approaches and generalized likelihood uncertainty estimation (GLUE). Given the above background, the primary goal of this study was to apply geochemical fingerprinting within the GLUE framework in the estimation of sub-basin spatial sediment source contributions in the arid Mehran River catchment in southern Iran, which drains into the Persian Gulf. The accuracy of GLUE predictions generated using four different sets of statistical tests for discriminating three sub-basin spatial sources was evaluated using 10 virtual sediments (VS) samples with known source contributions using the root mean square error (RMSE) and mean absolute error (MAE). Based on the results, the contributions modeled by GLUE for the western, central and eastern sub-basins are 1-42% (overall mean 20%), 0.5-30% (overall mean 12%) and 55-84% (overall mean 68%), respectively. According to the mean absolute fit (MAF; ≥ 95% for all target sediment samples) and goodness-of-fit (GOF; ≥ 99% for all samples), our suggested modeling approach is an accurate technique to quantify the source of sediments in the catchments. Overall, the estimated source proportions can help watershed engineers plan the targeting of conservation programs for soil and water resources.

Keywords: sediment source tracing, generalized likelihood uncertainty estimation, virtual sediment mixtures, Iran

Procedia PDF Downloads 43
582 Evaluating the Effects of Weather and Climate Change to Risks in Crop Production

Authors: Marcus Bellett-Travers

Abstract:

Different modelling approaches have been used to determine or predict yield of crops in different geographies. Central to the methodologies are the presumption that it is the absolute yield of the crop in a given location that is of the highest priority to those requiring information on crop productivity. Most individuals, companies and organisations within the agri-food sector need to be able to balance the supply of crops with the demand for them. Different modelling approaches have been used to determine and predict crop yield. The growing need to ensure certainty of supply and stability of prices requires an approach that describes the risk in producing a crop. A review of current methodologies to evaluate the risk to food production from changes in the weather and climate is presented.

Keywords: crop production, risk, climate, modelling

Procedia PDF Downloads 357