Search results for: approximate graph matching
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1292

Search results for: approximate graph matching

182 The Effect of Ambient Temperature on the Performance of the Simple and Modified Cycle Gas Turbine Plants

Authors: Ogbe E. E., Ossia. C. V., Saturday. E. G., Ezekwe M. C.

Abstract:

The disparity in power output between a simple and a modified gas turbine plant is noticeable when the gas turbine functions under local environmental conditions that deviate from the standard ISO specifications. Extensive research and literature have demonstrated a well-known direct correlation between ambient temperature and the power output of a gas turbine plant. In this study, the Omotosho gas turbine plant was modified into three different configurations. The reason for the modification is to improve its performance and reduce the fuel consumption and emission rate. Aspen Hysys software was used to simulate both the simple (Omotosho) and the three modified gas turbine plants. The input parameters considered include ambient temperature, air mass flow rate, fuel mass flow rate, water mass flow rate, turbine inlet temperature, compressor efficiency, and turbine efficiency, while the output parameters considered are thermal efficiency, specific fuel consumption, heat rate, emission rate, compressor power, turbine power and power output. The three modified gas turbine power plants incorporate an inlet air cooling system and a heat recovery steam generator. The variations between the modifications are due to additional components or enhancements alongside the inlet air cooling system and heat recovery steam generator incorporated; the first modification has an additional turbine, the second modification has an additional combustion chamber, and the third modification has an additional turbine and combustion chamber. This paper clearly shows ambient temperature effects on both the simple and three modified gas turbine plants. for every 10-degree kelvin increase in ambient temperature, there is an approximate reduction of 3977 kW, 4795 kW, 4681 kW, and 4793 kW of the power output for the simple gas turbine, first, second, and third modifications, respectively. Also, for every 10-degree kelvin increase in temperature, there is a thermal efficiency decrease of 1.22%, 1.45%, 1.43%, and 1.44% for the simple gas turbine, first, second, and third modifications respectively. Low ambient temperature will help save fuel; looking at the high price of fuel presently in Nigeria for every 10 degrees kelvin increase in temperature, there is a specific fuel consumption increase of 0.0074 kg/kWh, 0.0051 kg/kWh, 0.0061 kg/kWh, and 0.0057 kg/kWh for the simple gas turbine, first, second, and third modifications respectively. These findings will aid in accurately evaluating local power generating plants, particularly in hotter regions, for installing gas turbine inlet air cooling (GTIAC) systems.

Keywords: Aspen HYSYS software, Brayton Cycle, modified gas turbine, power plant, simple gas turbine, thermal efficiency.

Procedia PDF Downloads 31
181 Endodontic Pretreatments, Clinical Opportunities and Challenges

Authors: Ilma Robo, Manola Kelmendi, Saimir Heta, Megi Tafa, Vera Ostreni

Abstract:

Preservation of a natural tooth, even if endodontically treated, is more indicated than its replacement with an artificial tooth placed in prosthetic ways or with implant treatment. It is known how technology and endodontic treatment procedures have evolved significantly. It is also known that significant developments have been made in both dental prostheses and implant treatments, and again, in both specialties, it is emphasized that both the tooth placed with dental prostheses and the tooth placed with implant treatment cannot replace the natural tooth. The issue is whether long-term periapical tissue healing is achieved after a successful endodontic treatment, and for this, clinical data should be collected. In the cases when the apical closure or "apical filling" with the endodontic filling was carried out correctly clinically, but for various reasons, the healing of the periapical tissues did not occur, but also for those cases when the endodontic treatment did not reach the "apical filling" of the root canal. Teeth Endodontic retreatments have their clinical difficulty, but knowing the reason why endodontic treatment success has not been achieved clinically, the clinical endodontic approach is easier. In this process, it is important for the dentist to recognize the clinical and radiographic signs of persistent apical periodontitis or renewed apical periodontitis. After this initial procedure, dentists must know and evaluate the possibility of clinical endodontic retreatment by reporting, not precisely, but with very approximate values, the percentage of clinical success of endodontic retreatment. Depending on the reason for the performance, endodontic re-treatment may also need more specialized equipment or tools, for which even the professional who undertakes the re-treatment must be equipped with the relevant knowledge of their use and clinical application. Evaluating the clinical success of endodontic re-treatment is actually a more difficult process and requires more clinical responsibility since it must be considered that the initial treatment was performed by the same specialist as the specialist who undertakes the same endodontic re-treatment. Tooth So, the clinical endodontic re-treatment of a tooth should not be seen as a fund of clinical practice only of a good successful endodontist, but as part of routine endodontic treatments, nor should it be seen as a typical case where the tools and the most advanced technological devices in the endodontic field. So, the clinical picture of endodontic re-treatments offers the possibility of finding endodontic malpractice, the possibility of more accurate assessment of dental morphological anomalies, and above all, the cognitive and professional possibilities of the diagnosis of persistent apical periodontitis. This study offers the possibility of evaluating these three directions by presenting in numbers and in percentage the frequency of the reasons why the endodontic success of the root canal treatment is not always achieved.

Keywords: apical periodontitis, clinical susccess, endodontics, E.faecalis

Procedia PDF Downloads 9
180 The Effectiveness of Probiotics in the Treatment of Minimal Hepatic Encephalopathy Among Patients with Cirrhosis: An Expanded Meta-Analysis

Authors: Erwin Geroleo, Higinio Mappala

Abstract:

Introduction Overt Hepatic Encephalopathy (OHE) is the most dreaded outcome of liver cirrhosis. Aside from the triggering factors which are already known to precipitate OHE, there is growing evidence that an altered gut microbiota profile (dysbiosis) can also trigger OHE. MHE is the mildest form of hepatic encephalopathy(HE), affecting about one-third of patients with cirrhosis, and close 80% of patients with cirrhosis and manifests as abnormalities in central nervous system function. Since these symptoms are subclinical most patients are not being treated to prevent OHE. The gut microbiota have been evaluated by several studies as a therapeutic option for MHE, especially in decreasing the levels of ammonia, thus preventing progression to OHE Objectives This study aims to evaluate the efficacy of probiotics in terms of reduction of ammonia levels in patient with minimal hepatic encephalopathies and to determine if Probiotics has role in the prevention of progression to overt hepatic encephalopathy in adult patients with minimal hepatic encephalopathy (MHE) Methods and Analysis The literature search strategy was restricted to human studies in adults subjects from 2004 to 2022. The Jadad Score Calculation was utilized in the assessment of the final studies included in this study. Eight (8) studies were included. Cochrane’s Revman Web, the Fixed Effects model and the Ztest were all used in the overall analysis of the outcomes. A p value of less than 0.0005 was statistically significant. Results. These results show that Probiotics significantly lowers the level of Ammonia in Cirrhotic patients with OHE. It also shows that the use of Probiotics significantly prevents the progression of MHE to OHE. The overall risk of bias graph indicates low risk of publication bias among the studies included in the meta-analysis. Main findings This research found that plasma ammonia concentration was lower among participants treated with probiotics (p<0.00001).) Ammonia level of the probiotics group is lower by 13.96 μmol/ on the average. Overall risk of developing overt hepatic encephalopathy in the probiotics group is shown to be decreased by 15% as compared to the placebo group Conclusion The analysis showed that compared with placebo, probiotics can decrease serum ammonia, may improve MHE and may prevent OHE.

Keywords: minimal hepatic encephalopathy, probiotics, liver cirrhosis, overt hepatic encephalopathy

Procedia PDF Downloads 46
179 Interference of Mild Drought Stress on Estimation of Nitrogen Status in Winter Wheat by Some Vegetation Indices

Authors: H. Tavakoli, S. S. Mohtasebi, R. Alimardani, R. Gebbers

Abstract:

Nitrogen (N) is one of the most important agricultural inputs affecting crop growth, yield and quality in rain-fed cereal production. N demand of crops varies spatially across fields due to spatial differences in soil conditions. In addition, the response of a crop to the fertilizer applications is heavily reliant on plant available water. Matching N supply to water availability is thus essential to achieve an optimal crop response. The objective of this study was to determine effect of drought stress on estimation of nitrogen status of winter wheat by some vegetation indices. During the 2012 growing season, a field experiment was conducted at the Bundessortenamt (German Plant Variety Office) Marquardt experimental station which is located in the village of Marquardt about 5 km northwest of Potsdam, Germany (52°27' N, 12°57' E). The experiment was designed as a randomized split block design with two replications. Treatments consisted of four N fertilization rates (0, 60, 120 and 240 kg N ha-1, in total) and two water regimes (irrigated (Irr) and non-irrigated (NIrr)) in total of 16 plots with dimension of 4.5 × 9.0 m. The indices were calculated using readings of a spectroradiometer made of tec5 components. The main parts were two “Zeiss MMS1 nir enh” diode-array sensors with a nominal rage of 300 to 1150 nm with less than 10 nm resolutions and an effective range of 400 to 1000 nm. The following vegetation indices were calculated: NDVI, GNDVI, SR, MSR, NDRE, RDVI, REIP, SAVI, OSAVI, MSAVI, and PRI. All the experiments were conducted during the growing season in different plant growth stages including: stem elongation (BBCH=32-41), booting stage (BBCH=43), inflorescence emergence, heading (BBCH=56-58), flowering (BBCH=65-69), and development of fruit (BBCH=71). According to the results obtained, among the indices, NDRE and REIP were less affected by drought stress and can provide reliable wheat nitrogen status information, regardless of water status of the plant. They also showed strong relations with nitrogen status of winter wheat.

Keywords: nitrogen status, drought stress, vegetation indices, precision agriculture

Procedia PDF Downloads 319
178 Analyses of Soil Volatile Contaminants Extraction by Hot Air Injection

Authors: Abraham Dayan

Abstract:

Remediation of soil containing volatile contaminants is often conducted by vapor extraction (SVE) technique. The operation is based on injection of air at ambient temperatures with or without thermal soil warming. Thermal enhancements of soil vapor extraction (TESVE) processes are usually conducted by soil heating, sometimes assisted by added steam injections. The current study addresses a technique which has not received adequate attention and is based on using exclusively hot air as an alternative to the common TESVE practices. To demonstrate the merit of the hot air TESVE technique, a sandy soil containing contaminated water is studied. Numerical and analytical tools were used to evaluate the rate of decontamination processes for various geometries and operating conditions. The governing equations are based on the Darcy law and are applied to an expanding compressible flow within a sandy soil. The equations were solved to determine the minimal time required for complete soil remediation. An approximate closed form solution was developed based on the assumption of local thermodynamic equilibrium and on a linearized representation of temperature dependence of the vapor to air density ratio. The solution is general in nature and offers insight into the governing processes of the soil remediation operation, where self-similar temperature profiles under certain conditions may exist, and the noticeable role of the contaminants evaporation and recondensation processes in affecting the remediation time. Based on analyses of the hot air TESVE technique, it is shown that it is sufficient to heat the air during a certain period of the decontamination process without compromising its full advantage, and thereby, entailing a minimization of the air-heating-energy requirements. This in effect is achieved by regeneration, leaving the energy stored in the soil during the early period of the remediation process to heat the subsequently injected ambient air, which infiltrates through it for the decontamination of the remaining untreated soil zone. The characteristic time required to complete SVE operations are calculated as a function of, both, the injected air temperature and humidity. For a specific set of conditions, it is demonstrated that elevating the injected air temperature by 20oC, the hot air injection technique reduces the soil remediation time by 50%, while requiring 30% of additional energy consumption. Those evaluations clearly unveil the advantage of the hot air SVE process, which for insignificant cost of added air heating energy, the substantial cost expenditures for manpower and equipment utilization are reduced.

Keywords: Porous Media, Soil Decontamination, Hot Air, Vapor Extraction

Procedia PDF Downloads 10
177 Content-Aware Image Augmentation for Medical Imaging Applications

Authors: Filip Rusak, Yulia Arzhaeva, Dadong Wang

Abstract:

Machine learning based Computer-Aided Diagnosis (CAD) is gaining much popularity in medical imaging and diagnostic radiology. However, it requires a large amount of high quality and labeled training image datasets. The training images may come from different sources and be acquired from different radiography machines produced by different manufacturers, digital or digitized copies of film radiographs, with various sizes as well as different pixel intensity distributions. In this paper, a content-aware image augmentation method is presented to deal with these variations. The results of the proposed method have been validated graphically by plotting the removed and added seams of pixels on original images. Two different chest X-ray (CXR) datasets are used in the experiments. The CXRs in the datasets defer in size, some are digital CXRs while the others are digitized from analog CXR films. With the proposed content-aware augmentation method, the Seam Carving algorithm is employed to resize CXRs and the corresponding labels in the form of image masks, followed by histogram matching used to normalize the pixel intensities of digital radiography, based on the pixel intensity values of digitized radiographs. We implemented the algorithms, resized the well-known Montgomery dataset, to the size of the most frequently used Japanese Society of Radiological Technology (JSRT) dataset and normalized our digital CXRs for testing. This work resulted in the unified off-the-shelf CXR dataset composed of radiographs included in both, Montgomery and JSRT datasets. The experimental results show that even though the amount of augmentation is large, our algorithm can preserve the important information in lung fields, local structures, and global visual effect adequately. The proposed method can be used to augment training and testing image data sets so that the trained machine learning model can be used to process CXRs from various sources, and it can be potentially used broadly in any medical imaging applications.

Keywords: computer-aided diagnosis, image augmentation, lung segmentation, medical imaging, seam carving

Procedia PDF Downloads 222
176 Spatial Architecture Impact in Mediation Open Circuit Voltage Control of Quantum Solar Cell Recovery Systems

Authors: Moustafa Osman Mohammed

Abstract:

The photocurrent generations are influencing ultra-high efficiency solar cells based on self-assembled quantum dot (QD) nanostructures. Nanocrystal quantum dots (QD) provide a great enhancement toward solar cell efficiencies through the use of quantum confinement to tune absorbance across the solar spectrum enabled multi-exciton generation. Based on theoretical predictions, QDs have potential to improve systems efficiency in approximate regular electrons excitation intensity greater than 50%. In solar cell devices, an intermediate band formed by the electron levels in quantum dot systems. The spatial architecture is exploring how can solar cell integrate and produce not only high open circuit voltage (> 1.7 eV) but also large short-circuit currents due to the efficient absorption of sub-bandgap photons. In the proposed QD system, the structure allows barrier material to absorb wavelengths below 700 nm while multi-photon processes in the used quantum dots to absorb wavelengths up to 2 µm. The assembly of the electronic model is flexible to demonstrate the atoms and molecules structure and material properties to tune control energy bandgap of the barrier quantum dot to their respective optimum values. In terms of energy virtual conversion, the efficiency and cost of the electronic structure are unified outperform a pair of multi-junction solar cell that obtained in the rigorous test to quantify the errors. The milestone toward achieving the claimed high-efficiency solar cell device is controlling the edge causes of energy bandgap between the barrier material and quantum dot systems according to the media design limits. Despite this remarkable potential for high photocurrent generation, the achievable open-circuit voltage (Voc) is fundamentally limited due to non-radiative recombination processes in QD solar cells. The orientation of voltage recovery system is compared theoretically with experimental Voc variation in mediation upper–limit obtained one diode modeling form at the cells with different bandgap (Eg) as classified in the proposed spatial architecture. The opportunity for improvement Voc is valued approximately greater than 1V by using smaller QDs through QD solar cell recovery systems as confined to other micro and nano operations states.

Keywords: nanotechnology, photovoltaic solar cell, quantum systems, renewable energy, environmental modeling

Procedia PDF Downloads 156
175 Transformation of Periodic Fuzzy Membership Function to Discrete Polygon on Circular Polar Coordinates

Authors: Takashi Mitsuishi

Abstract:

Fuzzy logic has gained acceptance in the recent years in the fields of social sciences and humanities such as psychology and linguistics because it can manage the fuzziness of words and human subjectivity in a logical manner. However, the major field of application of the fuzzy logic is control engineering as it is a part of the set theory and mathematical logic. Mamdani method, which is the most popular technique for approximate reasoning in the field of fuzzy control, is one of the ways to numerically represent the control afforded by human language and sensitivity and has been applied in various practical control plants. Fuzzy logic has been gradually developing as an artificial intelligence in different applications such as neural networks, expert systems, and operations research. The objects of inference vary for different application fields. Some of these include time, angle, color, symptom and medical condition whose fuzzy membership function is a periodic function. In the defuzzification stage, the domain of the membership function should be unique to obtain uniqueness its defuzzified value. However, if the domain of the periodic membership function is determined as unique, an unintuitive defuzzified value may be obtained as the inference result using the center of gravity method. Therefore, the authors propose a method of circular-polar-coordinates transformation and defuzzification of the periodic membership functions in this study. The transformation to circular polar coordinates simplifies the domain of the periodic membership function. Defuzzified value in circular polar coordinates is an argument. Furthermore, it is required that the argument is calculated from a closed plane figure which is a periodic membership function on the circular polar coordinates. If the closed plane figure is continuous with the continuity of the membership function, a significant amount of computation is required. Therefore, to simplify the practice example and significantly reduce the computational complexity, we have discretized the continuous interval and the membership function in this study. In this study, the following three methods are proposed to decide the argument from the discrete polygon which the continuous plane figure is transformed into. The first method provides an argument of a straight line passing through the origin and through the coordinate of the arithmetic mean of each coordinate of the polygon (physical center of gravity). The second one provides an argument of a straight line passing through the origin and the coordinate of the geometric center of gravity of the polygon. The third one provides an argument of a straight line passing through the origin bisecting the perimeter of the polygon (or the closed continuous plane figure).

Keywords: defuzzification, fuzzy membership function, periodic function, polar coordinates transformation

Procedia PDF Downloads 364
174 Comparative Study on Fire Safety Evaluation Methods for External Cladding Systems: ISO 13785-2 and BS 8414

Authors: Kyungsuk Cho, H. Y. Kim, S. U. Chae, J. H. Choi

Abstract:

Technological development has led to the construction of super-tall buildings and insulators are increasingly used as exterior finishing materials to save energy. However, insulators are usually combustible and vulnerable to fire. Fires like that at Wooshin Golden Suite Building in Busan, Korea in 2010 and that at CCTV Building in Beijing, China are the major examples of fire spread accelerated by combustible insulators. The exterior finishing materials of a high-rise building are not made of insulators only, but they are integrated with the building’s external cladding system. There is a limit in evaluating the fire safety of a cladding system with a single small-unit material such as a cone calorimeter. Therefore, countries provide codes to evaluate the fire safety of exterior finishing materials using full-scale tests. This study compares and to examine the applicability of the methods to Korea. Standard analysis showed differences in the type and size of fire sources and duration and exterior finishing materials also differed in size. In order to confirm the differences, fire tests were conducted on identical external cladding systems to compare fire safety. Although the exterior finishing materials were identical, varying degrees of fire spread were observed, which could be considered as differences in the type and size of the fire sources and duration. Therefore, it is deduced that extended studies should be conducted before the evaluation methods and standards are employed in Korea. The two standards for evaluating fire safety provided different results. Peak heat release rate was 5.5MW in ISO method and 3.0±0.5MW in BS method. Peak heat release rate in ISO method continued for 15 minutes. Fire ignition, growth, full development and decay evolved for 30 minutes in BS method where wood cribs were used as fire sources. Therefore, follow-up studies should be conducted to determine which of the two standards provides fire sources that approximate the size of flames coming out from the openings or those spreading to the outside when a fire occurs at a high-rise building.

Keywords: external cladding systems, fire safety evaluation, ISO 13785-2, BS 8414

Procedia PDF Downloads 242
173 Impacts on Marine Ecosystems Using a Multilayer Network Approach

Authors: Nelson F. F. Ebecken, Gilberto C. Pereira, Lucio P. de Andrade

Abstract:

Bays, estuaries and coastal ecosystems are some of the most used and threatened natural systems globally. Its deterioration is due to intense and increasing human activities. This paper aims to monitor the socio-ecological in Brazil, model and simulate it through a multilayer network representing a DPSIR structure (Drivers, Pressures, States-Impacts-Responses) considering the concept of Management based on Ecosystems to support decision-making under the National/State/Municipal Coastal Management policy. This approach considers several interferences and can represent a significant advance in several scientific aspects. The main objective of this paper is the coupling of three different types of complex networks, the first being an ecological network, the second a social network, and the third a network of economic activities, in order to model the marine ecosystem. Multilayer networks comprise two or more "layers", which may represent different types of interactions, different communities, different points in time, and so on. The dependency between layers results from processes that affect the various layers. For example, the dispersion of individuals between two patches affects the network structure of both samples. A multilayer network consists of (i) a set of physical nodes representing entities (e.g., species, people, companies); (ii) a set of layers, which may include multiple layering aspects (e.g., time dependency and multiple types of relationships); (iii) a set of state nodes, each of which corresponds to the manifestation of a given physical node in a layer-specific; and (iv) a set of edges (weighted or not) to connect the state nodes among themselves. The edge set includes the intralayer edges familiar and interlayer ones, which connect state nodes between layers. The applied methodology in an existent case uses the Flow cytometry process and the modeling of ecological relationships (trophic and non-trophic) following fuzzy theory concepts and graph visualization. The identification of subnetworks in the fuzzy graphs is carried out using a specific computational method. This methodology allows considering the influence of different factors and helps their contributions to the decision-making process.

Keywords: marine ecosystems, complex systems, multilayer network, ecosystems management

Procedia PDF Downloads 113
172 Performance Evaluation of Routing Protocols in Vehicular Adhoc Networks

Authors: Salman Naseer, Usman Zafar, Iqra Zafar

Abstract:

This study explores the implication of Vehicular Adhoc Network (VANET) - in the rural and urban scenarios that is one domain of Mobile Adhoc Network (MANET). VANET provides wireless communication between vehicle to vehicle and also roadside units. The Federal Commission Committee of United States of American has been allocated 75 MHz of the spectrum band in the 5.9 GHz frequency range for dedicated short-range communications (DSRC) that are specifically designed to enhance any road safety applications and entertainment/information applications. There are several vehicular related projects viz; California path, car 2 car communication consortium, the ETSI, and IEEE 1609 working group that have already been conducted to improve the overall road safety or traffic management. After the critical literature review, the selection of routing protocols is determined, and its performance was well thought-out in the urban and rural scenarios. Numerous routing protocols for VANET are applied to carry out current research. Its evaluation was conceded with the help of selected protocols through simulation via performance metric i.e. throughput and packet drop. Excel and Google graph API tools are used for plotting the graphs after the simulation results in order to compare the selected routing protocols which result with each other. In addition, the sum of the output from each scenario was computed to undoubtedly present the divergence in results. The findings of the current study present that DSR gives enhanced performance for low packet drop and high throughput as compared to AODV and DSDV in an urban congested area and in rural environments. On the other hand, in low-density area, VANET AODV gives better results as compared to DSR. The worth of the current study may be judged as the information exchanged between vehicles is useful for comfort, safety, and entertainment. Furthermore, the communication system performance depends on the way routing is done in the network and moreover, the routing of the data based on protocols implement in the network. The above-presented results lead to policy implication and develop our understanding of the broader spectrum of VANET.

Keywords: AODV, DSDV, DSR, Adhoc network

Procedia PDF Downloads 286
171 The On-Board Critical Message Transmission Design for Navigation Satellite Delay/Disruption Tolerant Network

Authors: Ji-yang Yu, Dan Huang, Guo-ping Feng, Xin Li, Lu-yuan Wang

Abstract:

The navigation satellite network, especially the Beidou MEO Constellation, can relay data effectively with wide coverage and is applied in navigation, detection, and position widely. But the constellation has not been completed, and the amount of satellites on-board is not enough to cover the earth, which makes the data-relay disrupted or delayed in the transition process. The data-relay function needs to tolerant the delay or disruption in some extension, which make the Beidou MEO Constellation a delay/disruption-tolerant network (DTN). The traditional DTN designs mainly employ the relay table as the basic of data path schedule computing. But in practical application, especially in critical condition, such as the war-time or the infliction heavy losses on the constellation, parts of the nodes may become invalid, then the traditional DTN design could be useless. Furthermore, when transmitting the critical message in the navigation system, the maximum priority strategy is used, but the nodes still inquiry the relay table to design the path, which makes the delay more than minutes. Under this circumstances, it needs a function which could compute the optimum data path on-board in real-time according to the constellation states. The on-board critical message transmission design for navigation satellite delay/disruption-tolerant network (DTN) is proposed, according to the characteristics of navigation satellite network. With the real-time computation of parameters in the network link, the least-delay transition path is deduced to retransmit the critical message in urgent conditions. First, the DTN model for constellation is established based on the time-varying matrix (TVM) instead of the time-varying graph (TVG); then, the least transition delay data path is deduced with the parameters of the current node; at last, the critical message transits to the next best node. For the on-board real-time computing, the time delay and misjudges of constellation states in ground stations are eliminated, and the residual information channel for each node can be used flexibly. Compare with the minute’s delay of traditional DTN; the proposed transmits the critical message in seconds, which improves the re-transition efficiency. The hardware is implemented in FPGA based on the proposed model, and the tests prove the validity.

Keywords: critical message, DTN, navigation satellite, on-board, real-time

Procedia PDF Downloads 343
170 Predicting Child Attachment Style Based on Positive and Safe Parenting Components and Mediating Maternal Attachment Style in Children With ADHD

Authors: Alireza Monzavi Chaleshtari, Maryam Aliakbari

Abstract:

Objective: The aim of this study was to investigate the prediction of child attachment style based on a positive and safe combination parenting method mediated by maternal attachment styles in children with attention deficit hyperactivity disorder. Method: The design of the present study was descriptive of correlation and structural equations and applied in terms of purpose. The population of this study includes all children with attention deficit hyperactivity disorder living in Chaharmahal and Bakhtiari province and their mothers. The sample size of the above study includes 165children with attention deficit hyperactivity disorder in Chaharmahal and Bakhtiari province with their mothers, who were selected by purposive sampling method based on the inclusion criteria. The obtained data were analyzed in two sections of descriptive and inferential statistics. In the descriptive statistics section, statistical indices of mean, standard deviation, frequency distribution table and graph were used. In the inferential section, according to the nature of the hypotheses and objectives of the research, the data were analyzed using Pearson correlation coefficient tests, Bootstrap test and structural equation model. findings:The results of structural equation modeling showed that the research models fit and showed a positive and safe combination parenting style mediated by the mother attachment style has an indirect effect on the child attachment style. Also, a positive and safe combined parenting style has a direct relationship with child attachment style, and She has a mother attachment style. Conclusion:The results and findings of the present study show that there is a significant relationship between positive and safe combination parenting methods and attachment styles of children with attention deficit hyperactivity disorder with maternal attachment style mediation. Therefore, it can be expected that parents using a positive and safe combination232 parenting method can effectively lead to secure attachment in children with attention deficit hyperactivity disorder.

Keywords: child attachment style, positive and safe parenting, maternal attachment style, ADHD

Procedia PDF Downloads 66
169 Visual Preferences of Elementary School Children with Autism Spectrum Disorder: An Experimental Study

Authors: Larissa Pliska, Isabel Neitzel, Michael Buschermöhle, Olga Kunina-Habenicht, Ute Ritterfeld

Abstract:

Visual preferences, which can be assessed using eye tracking technologies, are considered one of the defining hallmarks of Autism Spectrum Disorder (ASD). Specifically, children with ASD show a decreased preference for social images rather than geometric images compared to typically developed (TD) children. Such differences are already prevalent at a very early age and indicate the severity of the disorder: toddlers with ASD who preferred geometric images when confronted with social and geometric images showed higher ASD symptom severity than toddlers with ASD who showed higher social attention. Furthermore, the complexity of social pictures (one child playing vs. two children playing together) as well as the mode of stimulus presentation (video or image), are not decisive for the marker. The average age of diagnosis for ASD in Germany is 6.5 years, and visual preference data on this age group is missing. In the present study, we therefore investigated whether visual preferences persist into school age. We examined the visual preferences of 16 boys aged 6 to 11 with ASD and unimpaired cognition as well as TD children (1:1 matching based on children's age and the parent's level of education) within an experimental setting. Different stimulus presentation formats (images vs. videos) and different levels of stimulus complexity were included. Children with and without ASD received pairs of social and non-social images and video stimuli on a screen while eye movements (i.e., eye position and gaze direction) were recorded. For this specific use case, KIZMO GmbH developed a customized, native iOS app (KIZMO Face-Analyzer) for use on iPads. Neither the format of stimulus presentation nor the complexity of the social images had a significant effect on the visual preference of children with and without ASD in this study. Despite the tendency for a difference between the groups for the video stimuli, there were no significant differences. Overall, no statistical differences in visual preference occurred between boys with and without ASD, suggesting that gaze preference in these groups is similar at primary school age. One limitation is that the children with ASD were already receiving Autism-specific intervention. The potential of a visual preference task as an indicator of ASD can be emphasized. The article discusses the clinical relevance of this marker in elementary school children.

Keywords: autism spectrum disorder, eye tracking, hallmark, visual preference

Procedia PDF Downloads 60
168 Drying Shrinkage of Concrete: Scale Effect and Influence of Reinforcement

Authors: Qier Wu, Issam Takla, Thomas Rougelot, Nicolas Burlion

Abstract:

In the framework of French underground disposal of intermediate level radioactive wastes, concrete is widely used as a construction material for containers and tunnels. Drying shrinkage is one of the most disadvantageous phenomena of concrete structures. Cracks generated by differential shrinkage could impair the mechanical behavior, increase the permeability of concrete and act as a preferential path for aggressive species, hence leading to an overall decrease in durability and serviceability. It is of great interest to understand the drying shrinkage phenomenon in order to predict and even to control the strains of concrete. The question is whether the results obtained from laboratory samples are in accordance with the measurements on a real structure. Another question concerns the influence of reinforcement on drying shrinkage of concrete. As part of a global project with Andra (French National Radioactive Waste Management Agency), the present study aims to experimentally investigate the scale effect as well as the influence of reinforcement on the development of drying shrinkage of two high performance concretes (based on CEM I and CEM V cements, according to European standards). Various sizes of samples are chosen, from ordinary laboratory specimens up to real-scale specimens: prismatic specimens with different volume-to-surface (V/S) ratios, thin slices (thickness of 2 mm), cylinders with different sizes (37 and 160 mm in diameter), hollow cylinders, cylindrical columns (height of 1000 mm) and square columns (320×320×1000 mm). The square columns have been manufactured with different reinforcement rates and can be considered as mini-structures, to approximate the behavior of a real voussoir from the waste disposal facility. All the samples are kept, in a first stage, at 20°C and 50% of relative humidity (initial conditions in the tunnel) in a specific climatic chamber developed by the Laboratory of Mechanics of Lille. The mass evolution and the drying shrinkage are monitored regularly. The obtained results show that the specimen size has a great impact on water loss and drying shrinkage of concrete. The specimens with a smaller V/S ratio and a smaller size have a bigger drying shrinkage. The correlation between mass variation and drying shrinkage follows the same tendency for all specimens in spite of the size difference. However, the influence of reinforcement rate on drying shrinkage is not clear based on the present results. The second stage of conservation (50°C and 30% of relative humidity) could give additional results on these influences.

Keywords: concrete, drying shrinkage, mass evolution, reinforcement, scale effect

Procedia PDF Downloads 183
167 Automatic Target Recognition in SAR Images Based on Sparse Representation Technique

Authors: Ahmet Karagoz, Irfan Karagoz

Abstract:

Synthetic Aperture Radar (SAR) is a radar mechanism that can be integrated into manned and unmanned aerial vehicles to create high-resolution images in all weather conditions, regardless of day and night. In this study, SAR images of military vehicles with different azimuth and descent angles are pre-processed at the first stage. The main purpose here is to reduce the high speckle noise found in SAR images. For this, the Wiener adaptive filter, the mean filter, and the median filters are used to reduce the amount of speckle noise in the images without causing loss of data. During the image segmentation phase, pixel values are ordered so that the target vehicle region is separated from other regions containing unnecessary information. The target image is parsed with the brightest 20% pixel value of 255 and the other pixel values of 0. In addition, by using appropriate parameters of statistical region merging algorithm, segmentation comparison is performed. In the step of feature extraction, the feature vectors belonging to the vehicles are obtained by using Gabor filters with different orientation, frequency and angle values. A number of Gabor filters are created by changing the orientation, frequency and angle parameters of the Gabor filters to extract important features of the images that form the distinctive parts. Finally, images are classified by sparse representation method. In the study, l₁ norm analysis of sparse representation is used. A joint database of the feature vectors generated by the target images of military vehicle types is obtained side by side and this database is transformed into the matrix form. In order to classify the vehicles in a similar way, the test images of each vehicle is converted to the vector form and l₁ norm analysis of the sparse representation method is applied through the existing database matrix form. As a result, correct recognition has been performed by matching the target images of military vehicles with the test images by means of the sparse representation method. 97% classification success of SAR images of different military vehicle types is obtained.

Keywords: automatic target recognition, sparse representation, image classification, SAR images

Procedia PDF Downloads 366
166 An Intelligent Text Independent Speaker Identification Using VQ-GMM Model Based Multiple Classifier System

Authors: Ben Soltane Cheima, Ittansa Yonas Kelbesa

Abstract:

Speaker Identification (SI) is the task of establishing identity of an individual based on his/her voice characteristics. The SI task is typically achieved by two-stage signal processing: training and testing. The training process calculates speaker specific feature parameters from the speech and generates speaker models accordingly. In the testing phase, speech samples from unknown speakers are compared with the models and classified. Even though performance of speaker identification systems has improved due to recent advances in speech processing techniques, there is still need of improvement. In this paper, a Closed-Set Tex-Independent Speaker Identification System (CISI) based on a Multiple Classifier System (MCS) is proposed, using Mel Frequency Cepstrum Coefficient (MFCC) as feature extraction and suitable combination of vector quantization (VQ) and Gaussian Mixture Model (GMM) together with Expectation Maximization algorithm (EM) for speaker modeling. The use of Voice Activity Detector (VAD) with a hybrid approach based on Short Time Energy (STE) and Statistical Modeling of Background Noise in the pre-processing step of the feature extraction yields a better and more robust automatic speaker identification system. Also investigation of Linde-Buzo-Gray (LBG) clustering algorithm for initialization of GMM, for estimating the underlying parameters, in the EM step improved the convergence rate and systems performance. It also uses relative index as confidence measures in case of contradiction in identification process by GMM and VQ as well. Simulation results carried out on voxforge.org speech database using MATLAB highlight the efficacy of the proposed method compared to earlier work.

Keywords: feature extraction, speaker modeling, feature matching, Mel frequency cepstrum coefficient (MFCC), Gaussian mixture model (GMM), vector quantization (VQ), Linde-Buzo-Gray (LBG), expectation maximization (EM), pre-processing, voice activity detection (VAD), short time energy (STE), background noise statistical modeling, closed-set tex-independent speaker identification system (CISI)

Procedia PDF Downloads 309
165 Flexural Properties of Carbon/Polypropylene Composites: Influence of Matrix Forming Polypropylene in Fiber, Powder, and Film States

Authors: Vijay Goud, Ramasamy Alagirusamy, Apurba Das, Dinesh Kalyanasundaram

Abstract:

Thermoplastic composites render new opportunities as effective processing technology while crafting newer complications into processing. One of the notable challenges is in achieving thorough wettability that is significantly deterred by the high viscosity of the long molecular chains of the thermoplastics. As a result of high viscosity, it is very difficult to impregnate the resin into a tightly interlaced textile structure to fill the voids present in the structure. One potential solution to the above problem, is to pre-deposit resin on the fiber, prior to consolidation. The current study compares DREF spinning, powder coating and film stacking methods of predeposition of resin onto fibers. An investigation into the flexural properties of unidirectional composites (UDC) produced from blending of carbon fiber and polypropylene (PP) matrix in varying forms of fiber, powder and film are reported. Dr. Ernst Fehrer (DREF) yarns or friction spun hybrid yarns were manufactured from PP fibers and carbon tows. The DREF yarns were consolidated to yield unidirectional composites (UDCs) referred to as UDC-D. PP in the form of powder was coated on carbon tows by electrostatic spray coating. The powder-coated towpregs were consolidated to form UDC-P. For the sake of comparison, a third UDC referred as UDC-F was manufactured by the consolidation of PP films stacked between carbon tows. The experiments were designed to yield a matching fiber volume fraction of about 50 % in all the three UDCs. A comparison of mechanical properties of the three composites was studied to understand the efficiency of matrix wetting and impregnation. Approximately 19% and 68% higher flexural strength were obtained for UDC-P than UDC-D and UDC-F respectively. Similarly, 25% and 81% higher modulus were observed in UDC-P than UDC-D and UDC-F respectively. Results from micro-computed tomography, scanning electron microscopy, and short beam tests indicate better impregnation of PP matrix in UDC-P obtained through electrostatic spray coating process and thereby higher flexural strength and modulus.

Keywords: DREF spinning, film stacking, flexural strength, powder coating, thermoplastic composite

Procedia PDF Downloads 222
164 Electromagnetic Energy Harvesting by Using a Rectenna with a Metamaterial Lens

Authors: Ursula D. C. Resende, Fabiano S. Bicalho, Sandro T. M. Gonçalves

Abstract:

The growing demand for cheap and clean energy sources have been motivated by the study and development of distinct technologies and devices able to provide different amounts of energy. In order to supply energy for small loads, the energy from the electromagnetic spectrum can be harvested. This possibility is particularly interesting because this kind of energy is constantly available in the environment and the number of radiofrequency sources is permanently increasing, due to advances in telecommunications services. A rectenna, which is a combination of an antenna and a rectifier circuit, is an equipment that can efficiently perform the electromagnetic energy harvesting. However, since the amount of electromagnetic energy available in the environment is very small, limited values of power can be harvested by the rectenna. Therefore, several technical strategies have been investigated in order to increase this amount of power. In this work, a metamaterial electromagnetic lens is used to improve the electromagnetic energy harvesting. The rectenna investigated was designed and optimized to charge a Li-Ion battery using the electromagnetic energy from an internet Wi-Fi commercial router model TL-WR841HP operating in 2.45 GHz with maximal output power equal to 18 dBm. The rectenna consists of a high directive antenna, a double voltage rectifier circuit and a metamaterial lens. The printed antenna, constituted of two rectangular radiator elements, was projected and optimized by using the Computer Simulation Software (CST) in order to obtain high directivities and values of S11 parameter below -10 dB in 2.45 GHz. The antenna was printed over a double-sided copper fiberglass substrate, FR4, with characterized relative electric permittivity εr = 4.3 and tangent of losses δ = 0.01. The rectifier circuit, which incorporates a circuit for impedance matching and uses the Schottky diode HSMS-2852, was projected and optimized by using Advanced Design Software (ADS) and built over the same FR4 substrate. The metamaterial cell is composed of two Square Split Ring Resonator (S-SRR) and a thin wire in order to operate with negative values of εr and relative magnetic permeability in 2.45 GHz. In order to evaluate the performance of the purposed rectenna two experimental charging tests were performed, one without and other with the metamaterial lens. The result obtained demonstrate that the electromagnetic lens was able to significantly increase the levels of electric current delivered to the battery, approximately 44%.

Keywords: electromagnetic energy harvesting, electromagnetic lens, metamaterial, rectenna

Procedia PDF Downloads 143
163 Imbalance on the Croatian Housing Market in the Aftermath of an Economic Crisis

Authors: Tamara Slišković, Tomislav Sekur

Abstract:

This manuscript examines factors that affect demand and supply of the housing market in Croatia. The period from the beginning of this century, until 2008, was characterized by a strong expansion of construction, housing and real estate market in general. Demand for residential units was expanding, and this was supported by favorable lending conditions of banks. Indicators on the supply side, such as the number of newly built houses and the construction volume index were also increasing. Rapid growth of demand, along with the somewhat slower supply growth, led to the situation in which new apartments were sold before the completion of residential buildings. This resulted in a rise of housing price which was indication of a clear link between the housing prices with the supply and demand in the housing market. However, after 2008 general economic conditions in Croatia worsened and demand for housing has fallen dramatically, while supply descended at much slower pace. Given that there is a gap between supply and demand, it can be concluded that the housing market in Croatia is in imbalance. Such trend is accompanied by a relatively small decrease in housing price. The final result of such movements is the large number of unsold housing units at relatively high price levels. For this reason, it can be argued that housing prices are sticky and that, consequently, the price level in the aftermath of a crisis does not correspond to the discrepancy between supply and demand on the Croatian housing market. The degree of rigidity of the housing price can be determined by inclusion of the housing price as the explanatory variable in the housing demand function. Other independent variables are demographic variable (e.g. the number of households), the interest rate on housing loans, households' disposable income and rent. The equilibrium price is reached when the demand for housing equals its supply, and the speed of adjustment of actual prices to equilibrium prices reveals the extent to which the prices are rigid. The latter requires inclusion of the housing prices with time lag as an independent variable in estimating demand function. We also observe the supply side of the housing market, in order to explain to what extent housing prices explain the movement of new construction activity, and other variables that describe the supply. In this context, we test whether new construction on the Croatian market is dependent on current prices or prices with a time lag. Number of dwellings is used to approximate new construction (flow variable), while the housing prices (current or lagged), quantity of dwellings in the previous period (stock variable) and a series of costs related to new construction are independent variables. We conclude that the key reason for the imbalance in the Croatian housing market should be sought in the relative relationship of price elasticities of supply and demand.

Keywords: Croatian housing market, economic crisis, housing prices, supply imbalance, demand imbalance

Procedia PDF Downloads 272
162 A pilot Study of Umbilical Cord Mini-Clamp

Authors: Seng Sing Tan

Abstract:

Clamping of the umbilical cord after birth is widely practiced as a part of labor management. Further improvements were proposed to produce a smaller, lighter and more comfortable clamp while still maintaining current standards of clamping. A detachable holder was also developed to facilitate the clamping process. This pilot study on the efficacy of the mini-clamp was conducted to evaluate a tightness of the seal and a firm grip of the clamp on the umbilical cord. The study was carried out at National University Hospital, using 5 sets of placental cord. 18 samples of approximate 10 cm each were harvested. The test results showed that the mini-clamp was able to stop the flow through the cord after clamping without rupturing the cord. All slip tests passed with a load of 0.2 kg. In the pressure testing, 30kPa of saline was exerted into the umbilical veins. Although there was no physical sign of fluid leaking through the end secured by the mini-clamp, the results showed the pressure was not able to sustain the pressure set during the tests. 12 out of the 18 test samples have more than 7% of pressure drop in 30 seconds. During the pressure leak test, it was observed on several samples that when pressurized, small droplets of saline were growing on the outer surface of the cord lining membrane. It was thus hypothesized that the pressure drop was likely caused by the perfusion of the injected saline through the Wharton’s jelly and the cord lining membrane. The average pressure in the umbilical vein is roughly 2.67kPa (20 mmHg), less than 10% of 30kPa (~225mmHg), set for the pressure testing. As such, the pressure set could be over-specified, leading to undesirable outcomes. The development of the mini-clamp was an attempt to increase the comfort of newly born babies while maintaining the usability and efficacy of hospital grade umbilical cord clamp. The pressure leak in this study would be unfair to fully attribute it to the design and efficacy of the mini-clamp. Considering the unexpected leakage of saline through the umbilical membrane due to over-specified pressure exerted on the umbilical veins, improvements can definitely be made to the existing experimental setup to obtain a more accurate and conclusive outcome. If proven conclusive and effective, the mini-clamp with a detachable holder could be a smaller and potentially cheaper alternative to existing umbilical cord clamps. In addition, future clinical trials could be conducted to determine the user-friendliness of the mini-clamp and evaluate its practicality in the clinical setting by labor ward clinicians. A further potential improvement could be proposed on the sustainability factor of the mini-clamp. A biodegradable clamp would revolutionise the industry in this increasingly environmentally sustainability world.

Keywords: leak test, mini-clamp, slip test, umbilical cord

Procedia PDF Downloads 132
161 A Numerical Investigation of Segmental Lining Joints Interactions in Tunnels

Authors: M. H. Ahmadi, A. Mortazavi, H. Zarei

Abstract:

Several authors have described the main mechanism of formation of cracks in the segment lining during the construction of tunnels with tunnel boring machines. A comprehensive analysis of segmental lining joints may help to guarantee a safe construction during Tunneling and serviceable stages. The most frequent types of segment damage are caused by a condition of uneven segment matching due to contact deficiencies. This paper investigated the interaction mechanism of precast concrete lining joints in tunnels. The Discrete Element Method (DEM) was used to analyze a typical segmental lining model consisting of six segment rings. In the analyses, typical segmental lining design parameters of the Ghomrood water conveyance tunnel, Iran were employed in the study. In the conducted analysis, the worst-case scenario of loading faced during the boring of Ghomrood tunnel was considered. This was associated with the existence of a crushed zone dipping at 75 degree at the location of the key segment. In the analysis, moreover, the effect of changes in horizontal stress ratio on the loads on the segment was assessed. The boundary condition associated with K (ratio of the horizontal to the vertical stress) values of 0.5, 1, 1.5 and 2 were applied to the model and separate analysis was conducted for each case. Important parameters such as stress, moments, and displacements were measured at joint locations and the surrounding rock. Accordingly, the segment joint interactions were assessed and analyzed. Moreover, rock mass properties of the Ghomrood in Ghom were adopted. In this study, the load acting on segments joints are included a crushed zone stratum force that intersect tunnel with 75 slopes in the location of the key segment, gravity force of segments and earth pressures. A numerical investigation was used for different coefficients of stress concentration of 0.5, 1, 1.5, 2 and different geological conditions of saturated crushed zone under the critical scenario. The numerical results also demonstrate that maximum bending moments in longitudinal joints occurred for crushed zone with the weaken strengths (Sandstone). Besides that, increasing the load in segment-stratum interfaces affected radial stress in longitudinal joints and finally the opening of joints occurred.

Keywords: joint, interface, segment, contact

Procedia PDF Downloads 258
160 AS-Geo: Arbitrary-Sized Image Geolocalization with Learnable Geometric Enhancement Resizer

Authors: Huayuan Lu, Chunfang Yang, Ma Zhu, Baojun Qi, Yaqiong Qiao, Jiangqian Xu

Abstract:

Image geolocalization has great application prospects in fields such as autonomous driving and virtual/augmented reality. In practical application scenarios, the size of the image to be located is not fixed; it is impractical to train different networks for all possible sizes. When its size does not match the size of the input of the descriptor extraction model, existing image geolocalization methods usually directly scale or crop the image in some common ways. This will result in the loss of some information important to the geolocalization task, thus affecting the performance of the image geolocalization method. For example, excessive down-sampling can lead to blurred building contour, and inappropriate cropping can lead to the loss of key semantic elements, resulting in incorrect geolocation results. To address this problem, this paper designs a learnable image resizer and proposes an arbitrary-sized image geolocation method. (1) The designed learnable image resizer employs the self-attention mechanism to enhance the geometric features of the resized image. Firstly, it applies bilinear interpolation to the input image and its feature maps to obtain the initial resized image and the resized feature maps. Then, SKNet (selective kernel net) is used to approximate the best receptive field, thus keeping the geometric shapes as the original image. And SENet (squeeze and extraction net) is used to automatically select the feature maps with strong contour information, enhancing the geometric features. Finally, the enhanced geometric features are fused with the initial resized image, to obtain the final resized images. (2) The proposed image geolocalization method embeds the above image resizer as a fronting layer of the descriptor extraction network. It not only enables the network to be compatible with arbitrary-sized input images but also enhances the geometric features that are crucial to the image geolocalization task. Moreover, the triplet attention mechanism is added after the first convolutional layer of the backbone network to optimize the utilization of geometric elements extracted by the first convolutional layer. Finally, the local features extracted by the backbone network are aggregated to form image descriptors for image geolocalization. The proposed method was evaluated on several mainstream datasets, such as Pittsburgh30K, Tokyo24/7, and Places365. The results show that the proposed method has excellent size compatibility and compares favorably to recently mainstream geolocalization methods.

Keywords: image geolocalization, self-attention mechanism, image resizer, geometric feature

Procedia PDF Downloads 214
159 Pre-Implementation of Total Body Irradiation Using Volumetric Modulated Arc Therapy: Full Body Anthropomorphic Phantom Development

Authors: Susana Gonçalves, Joana Lencart, Anabela Gregório Dias

Abstract:

Introduction: In combination with chemotherapy, Total Body Irradiation (TBI) is most used as part of the conditioning regimen prior to allogeneic hematopoietic stem cell transplantation. Conventional TBI techniques have a long application time but non-conformality of beam-application with the inability to individually spare organs at risk. Our institution’s intention is to start using Volumetric Modulated Arc Therapy (VMAT) techniques to increase homogeneity of delivered radiation. As a first approach, a dosimetric plan was performed on a computed tomography (CT) scan of a Rando Alderson antropomorfic phantom (head and torso), using a set of six arcs distributed along the phantom. However, a full body anthropomorphic phantom is essential to carry out technique validation and implementation. Our aim is to define the physical and chemical characteristics and the ideal manufacturing procedure of upper and lower limbs to our anthropomorphic phantom, for later validate TBI using VMAT. Materials and Methods: To study the better fit between our phantom and limbs, a CT scan of Rando Alderson anthropomorphic phantom was acquired. CT was performed on GE Healthcare equipment (model Optima CT580 W), with slice thickness of 2.5 mm. This CT was also used to access the electronic density of soft tissue and bone through Hounsfield units (HU) analysis. Results: CT images were analyzed and measures were made for the ideal upper and lower limbs. Upper limbs should be build under the following measures: 43cm length and 7cm diameter (next to the shoulder section). Lower limbs should be build under the following measures: 79cm length and 16.5cm diameter (next to the thigh section). As expected, soft tissue and bone have very different electronic density. This is important to choose and analyze different materials to better represent soft tissue and bone characteristics. The approximate HU values of the soft tissue and for bone shall be 35HU and 250HU, respectively. Conclusion: At the moment, several compounds are being developed based on different types of resins and additives in order to be able to control and mimic the various constituent densities of the tissues. Concurrently, several manufacturing techniques are being explored to make it possible to produce the upper and lower limbs in a simple and non-expensive way, in order to finally carry out a systematic and appropriate study of the total body irradiation. This preliminary study was a good starting point to demonstrate the feasibility of TBI with VMAT.

Keywords: TBI, VMAT, anthropomorphic phantom, tissue equivalent materials

Procedia PDF Downloads 80
158 Energy Efficient Refrigerator

Authors: Jagannath Koravadi, Archith Gupta

Abstract:

In a world with constantly growing energy prices, and growing concerns about the global climate changes caused by increased energy consumption, it is becoming more and more essential to save energy wherever possible. Refrigeration systems are one of the major and bulk energy consuming systems now-a-days in industrial sectors, residential sectors and household environment. Refrigeration systems with considerable cooling requirements consume a large amount of electricity and thereby contribute greatly to the running costs. Therefore, a great deal of attention is being paid towards improvement of the performance of the refrigeration systems in this regard throughout the world. The Coefficient of Performance (COP) of a refrigeration system is used for determining the system's overall efficiency. The operating cost to the consumer and the overall environmental impact of a refrigeration system in turn depends on the COP or efficiency of the system. The COP of a refrigeration system should therefore be as high as possible. Slight modifications in the technical elements of the modern refrigeration systems have the potential to reduce the energy consumption, and improvements in simple operational practices with minimal expenses can have beneficial impact on COP of the system. Thus, the challenge is to determine the changes that can be made in a refrigeration system in order to improve its performance, reduce operating costs and power requirement, improve environmental outcomes, and achieve a higher COP. The opportunity here, and a better solution to this challenge, will be to incorporate modifications in conventional refrigeration systems for saving energy. Energy efficiency, in addition to improvement of COP, can deliver a range of savings such as reduced operation and maintenance costs, improved system reliability, improved safety, increased productivity, better matching of refrigeration load and equipment capacity, reduced resource consumption and greenhouse gas emissions, better working environment, and reduced energy costs. The present work aims at fabricating a working model of a refrigerator that will provide for effective heat recovery from superheated refrigerant with the help of an efficient de-superheater. The temperature of the refrigerant and water in the de-super heater at different intervals of time are measured to determine the quantity of waste heat recovered. It is found that the COP of the system improves by about 6% with the de-superheater and the power input to the compressor decreases by 4 % and also the refrigeration capacity increases by 4%.

Keywords: coefficiency of performance, de-superheater, refrigerant, refrigeration capacity, heat recovery

Procedia PDF Downloads 320
157 Structuring Paraphrases: The Impact Sentence Complexity Has on Key Leader Engagements

Authors: Meaghan Bowman

Abstract:

Soldiers are taught about the importance of effective communication with repetition of the phrase, “Communication is key.” They receive training in preparing for, and carrying out, interactions between foreign and domestic leaders to gain crucial information about a mission. These interactions are known as Key Leader Engagements (KLEs). For the training of KLEs, doctrine mandates the skills needed to conduct these “engagements” such as how to: behave appropriately, identify key leaders, and employ effective strategies. Army officers in training learn how to confront leaders, what information to gain, and how to ask questions respectfully. Unfortunately, soldiers rarely learn how to formulate questions optimally. Since less complex questions are easier to understand, we hypothesize that semantic complexity affects content understanding, and that age and education levels may have an effect on one’s ability to form paraphrases and judge their quality. In this study, we looked at paraphrases of queries as well as judgments of both the paraphrases’ naturalness and their semantic similarity to the query. Queries were divided into three complexity categories based on the number of relations (the first number) and the number of knowledge graph edges (the second number). Two crowd-sourced tasks were completed by Amazon volunteer participants, also known as turkers, to answer the research questions: (i) Are more complex queries harder to paraphrase and judge and (ii) Do age and education level affect the ability to understand complex queries. We ran statistical tests as follows: MANOVA for query understanding and two-way ANOVA to understand the relationship between query complexity and education and age. A probe of the number of given-level queries selected for paraphrasing by crowd-sourced workers in seven age ranges yielded promising results. We found significant evidence that age plays a role and marginally significant evidence that education level plays a role. These preliminary tests, with output p-values of 0.0002 and 0.068, respectively, suggest the importance of content understanding in a communication skill set. This basic ability to communicate, which may differ by age and education, permits reproduction and quality assessment and is crucial in training soldiers for effective participation in KLEs.

Keywords: engagement, key leader, paraphrasing, query complexity, understanding

Procedia PDF Downloads 161
156 An Assessment of the Impacts of Agro-Ecological Practices towards the Improvement of Crop Health and Yield Capacity: A Case of Mopani District, Limpopo, South Africa

Authors: Tshilidzi C. Manyanya, Nthaduleni S. Nethengwe, Edmore Kori

Abstract:

The UNFCCC, FAO, GCF, IPCC and other global structures advocate for agro-ecology do address food security and sovereignty. However, most of the expected outcomes concerning agro-ecological were not empirically tested for universal application. Agro-ecology is theorised to increase crop health over ago-ecological farms and decrease over conventional farms. Increased crop health means increased carbon sequestration and thus less CO2 in the atmosphere. This is in line with the view that global warming is anthropogenically enhanced through GHG emissions. Agro-ecology mainly affects crop health, soil carbon content and yield on the cultivated land. Economic sustainability is directly related to yield capacity, which is theorized to increase by 3-10% in a space of 3 - 10 years as a result of agro-ecological implementation. This study aimed to empirically assess the practicality and validity of these assumptions. The study utilized mainly GIS and RS techniques to assess the effectiveness of agro-ecology in crop health improvement from satellite images. The assessment involved a longitudinal study (2013 – 2015) assessing the changes that occur after a farm retrofits from conventional agriculture to agro-ecology. The assumptions guided the objectives of the study. For each objective, an agro-ecological farm was compared with a conventional farm in the same climatic conditional occupying the same general location. Crop health was assessed using satellite images analysed through ArcGIS and Erdas. This entailed the production of NDVI and Re-classified outputs of the farm area. The NDVI ranges of the entire period of study were thus compared in a stacked histogram for each farm to assess for trends. Yield capacity was calculated based on the production records acquired from the farmers and plotted in a stacked bar graph as percentages of a total for each farm. The results of the study showed decreasing crop health trends over 80% of the conventional farms and an increase over 80% of the organic farms. Yield capacity showed similar patterns to those of crop health. The study thus showed that agro-ecology is an effective strategy for crop-health improvement and yield increase.

Keywords: agro-ecosystem, conventional farm, dialectical, sustainability

Procedia PDF Downloads 216
155 Recursion, Merge and Event Sequence: A Bio-Mathematical Perspective

Authors: Noury Bakrim

Abstract:

Formalization is indeed a foundational Mathematical Linguistics as demonstrated by the pioneering works. While dialoguing with this frame, we nonetheless propone, in our approach of language as a real object, a mathematical linguistics/biosemiotics defined as a dialectical synthesis between induction and computational deduction. Therefore, relying on the parametric interaction of cycles, rules, and features giving way to a sub-hypothetic biological point of view, we first hypothesize a factorial equation as an explanatory principle within Category Mathematics of the Ergobrain: our computation proposal of Universal Grammar rules per cycle or a scalar determination (multiplying right/left columns of the determinant matrix and right/left columns of the logarithmic matrix) of the transformable matrix for rule addition/deletion and cycles within representational mapping/cycle heredity basing on the factorial example, being the logarithmic exponent or power of rule deletion/addition. It enables us to propone an extension of minimalist merge/label notions to a Language Merge (as a computing principle) within cycle recursion relying on combinatorial mapping of rules hierarchies on external Entax of the Event Sequence. Therefore, to define combinatorial maps as language merge of features and combinatorial hierarchical restrictions (governing, commanding, and other rules), we secondly hypothesize from our results feature/hierarchy exponentiation on graph representation deriving from Gromov's Symbolic Dynamics where combinatorial vertices from Fe are set to combinatorial vertices of Hie and edges from Fe to Hie such as for all combinatorial group, there are restriction maps representing different derivational levels that are subgraphs: the intersection on I defines pullbacks and deletion rules (under restriction maps) then under disjunction edges H such that for the combinatorial map P belonging to Hie exponentiation by intersection there are pullbacks and projections that are equal to restriction maps RM₁ and RM₂. The model will draw on experimental biomathematics as well as structural frames with focus on Amazigh and English (cases from phonology/micro-semantics, Syntax) shift from Structure to event (especially Amazigh formant principle resolving its morphological heterogeneity).

Keywords: rule/cycle addition/deletion, bio-mathematical methodology, general merge calculation, feature exponentiation, combinatorial maps, event sequence

Procedia PDF Downloads 127
154 An Evolutionary Approach for QAOA for Max-Cut

Authors: Francesca Schiavello

Abstract:

This work aims to create a hybrid algorithm, combining Quantum Approximate Optimization Algorithm (QAOA) with an Evolutionary Algorithm (EA) in the place of traditional gradient based optimization processes. QAOA’s were first introduced in 2014, where, at the time, their algorithm performed better than the traditional best known classical algorithm for Max-cut graphs. Whilst classical algorithms have improved since then and have returned to being faster and more efficient, this was a huge milestone for quantum computing, and their work is often used as a benchmarking tool and a foundational tool to explore variants of QAOA’s. This, alongside with other famous algorithms like Grover’s or Shor’s, highlights to the world the potential that quantum computing holds. It also presents the reality of a real quantum advantage where, if the hardware continues to improve, this could constitute a revolutionary era. Given that the hardware is not there yet, many scientists are working on the software side of things in the hopes of future progress. Some of the major limitations holding back quantum computing are the quality of qubits and the noisy interference they generate in creating solutions, the barren plateaus that effectively hinder the optimization search in the latent space, and the availability of number of qubits limiting the scale of the problem that can be solved. These three issues are intertwined and are part of the motivation for using EAs in this work. Firstly, EAs are not based on gradient or linear optimization methods for the search in the latent space, and because of their freedom from gradients, they should suffer less from barren plateaus. Secondly, given that this algorithm performs a search in the solution space through a population of solutions, it can also be parallelized to speed up the search and optimization problem. The evaluation of the cost function, like in many other algorithms, is notoriously slow, and the ability to parallelize it can drastically improve the competitiveness of QAOA’s with respect to purely classical algorithms. Thirdly, because of the nature and structure of EA’s, solutions can be carried forward in time, making them more robust to noise and uncertainty. Preliminary results show that the EA algorithm attached to QAOA can perform on par with the traditional QAOA with a Cobyla optimizer, which is a linear based method, and in some instances, it can even create a better Max-Cut. Whilst the final objective of the work is to create an algorithm that can consistently beat the original QAOA, or its variants, due to either speedups or quality of the solution, this initial result is promising and show the potential of EAs in this field. Further tests need to be performed on an array of different graphs with the parallelization aspect of the work commencing in October 2023 and tests on real hardware scheduled for early 2024.

Keywords: evolutionary algorithm, max cut, parallel simulation, quantum optimization

Procedia PDF Downloads 60
153 The Mechanism Study on the Difference between High and Low Voltage Performance of Li3V2(PO4)3

Authors: Enhui Wang, Qingzhu Ou, Yan Tang, Xiaodong Guo

Abstract:

As one of most popular polyanionic compounds in lithium-ion cathode materials, Li3V2(PO4)3 has always suffered from the low rate capability especially during 3~4.8V, which is considered to be related with the ion diffusion resistance and structural transformation during the Li+ de/intercalation. Here, as the change of cut-off voltages, cycling numbers and current densities, the process of SEI interfacial film’s formation-growing- destruction-repair on the surface of the cathode, the structural transformation during the charge and discharge, the de/intercalation kinetics reflected by the electrochemical impedance and the diffusion coefficient, have been investigated in detail. Current density, cycle numbers and cut-off voltage impacting on interfacial film and structure was studied specifically. Firstly, the matching between electrolyte and material was investigated, it turned out that the batteries with high voltage electrolyte showed the best electrochemical performance and high voltage electrolyte would be the best electrolyte. Secondly, AC impedance technology was used to study the changes of interface impedance and lithium ion diffusion coefficient, the results showed that current density, cycle numbers and cut-off voltage influenced the interfacial film together and the one who changed the interfacial properties most was the key factor. Scanning electron microscopy (SEM) analysis confirmed that the attenuation of discharge specific capacity was associated with the destruction and repair process of the SEI film. Thirdly, the X-ray diffraction was used to study the changes of structure, which was also impacted by current density, cycle numbers and cut-off voltage. The results indicated that the cell volume of Li3V2 (PO4 )3 increased as the current density increased; cycle numbers merely influenced the structure of material; the cell volume decreased first and moved back gradually after two Li-ion had been deintercalated as the charging cut-off voltage increased, and increased as the intercalation number of Li-ion increased during the discharging process. Then, the results which studied the changes of interface impedance and lithium ion diffusion coefficient turned out that the interface impedance and lithium ion diffusion coefficient increased when the cut-off voltage passed the voltage platforms and decreased when the cut-off voltage was between voltage platforms. Finally, three-electrode system was first adopted to test the activation energy of the system, the results indicated that the activation energy of the three-electrode system (22.385 KJ /mol) was much smaller than that of two-electrode system (40.064 KJ /mol).

Keywords: cut-off voltage, de/intercalation kinetics, solid electrolyte interphase film, structural transformation

Procedia PDF Downloads 296