Search results for: bathtub curve
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 1034

Search results for: bathtub curve

644 Endocardial Ultrasound Segmentation using Level Set method

Authors: Daoudi Abdelaziz, Mahmoudi Saïd, Chikh Mohamed Amine

Abstract:

This paper presents a fully automatic segmentation method of the left ventricle at End Systolic (ES) and End Diastolic (ED) in the ultrasound images by means of an implicit deformable model (level set) based on Geodesic Active Contour model. A pre-processing Gaussian smoothing stage is applied to the image, which is essential for a good segmentation. Before the segmentation phase, we locate automatically the area of the left ventricle by using a detection approach based on the Hough Transform method. Consequently, the result obtained is used to automate the initialization of the level set model. This initial curve (zero level set) deforms to search the Endocardial border in the image. On the other hand, quantitative evaluation was performed on a data set composed of 15 subjects with a comparison to ground truth (manual segmentation).

Keywords: level set method, transform Hough, Gaussian smoothing, left ventricle, ultrasound images.

Procedia PDF Downloads 437
643 A Holistic Study of the Beta Lyrae Systems V0487 Lac, V0566 Hya and V0666 Lac

Authors: Moqbil S. Alenazi, Magdy. M. Elkhateeb

Abstract:

A comprehensive photometric study and evolutionary state for the newly discovered Beta Lyr systems V0487 Lac, V0566 Hya, and V0666 Lac were carried out by means of their first photometric observations. New times of minima were estimated from the observed light curves, and first (O-C) curves were established for all systems. A windows interface version of the Wilson and Devinney code (W-D) based on model atmospheres and a pass band prescription have been used for the radiative treatment. The accepted models reveal some absolute parameters for the studied systems, which are used in adopting the spectral type of the system's components and their evolutionary status. Distances to each system were calculated, and physical properties were estimated. Locations of the systems on the theoreticalmass–luminosity and mass–radius relations revealed a good fit for all systems components except for the secondary component of the system V0487 Lac.

Keywords: eclipsing binaries, light curve modelling, evolutionary state

Procedia PDF Downloads 47
642 Definition of Service Angle of Android’S Robot Hand by Method of Small Movements of Gripper’S Axis Synthesis by Speed Vector

Authors: Valeriy Nebritov

Abstract:

The paper presents a generalized method for determining the service solid angle based on the assigned gripper axis orientation with a stationary grip center. Motion synthesis in this work is carried out in the vector of velocities. As an example, a solid angle of the android robot arm is determined, this angle being formed by the longitudinal axis of a gripper. The nature of the method is based on the study of sets of configuration positions, defining the end point positions of the unit radius sphere sweep, which specifies the service solid angle. From this the spherical curve specifying the shape of the desired solid angle was determined. The results of the research can be used in the development of control systems of autonomous android robots.

Keywords: android robot, control systems, motion synthesis, service angle

Procedia PDF Downloads 171
641 A Crowdsourced Homeless Data Collection System and Its Econometric Analysis

Authors: Praniil Nagaraj

Abstract:

This paper proposes a method to collect homeless data using crowdsourcing and presents an approach to analyze the data, demonstrating its potential to strengthen existing and future policies aimed at promoting socio-economic equilibrium. The 2022 Annual Homeless Assessment Report (AHAR) to Congress highlighted alarming statistics, emphasizing the need for effective decision-making and budget allocation within local planning bodies known as Continuums of Care (CoC). This paper's contributions can be categorized into three main areas. Firstly, a unique method for collecting homeless data is introduced, utilizing a user-friendly smartphone app (currently available for Android). The app enables the general public to quickly record information about homeless individuals, including the number of people and details about their living conditions. The collected data, including date, time, and location, is anonymized and securely transmitted to the cloud. It is anticipated that an increasing number of users motivated to contribute to society will adopt the app, thus expanding the data collection efforts. Duplicate data is addressed through simple classification methods, and historical data is utilized to fill in missing information. The second contribution of this paper is the description of data analysis techniques applied to the collected data. By combining this new data with existing information, statistical regression analysis is employed to gain insights into various aspects, such as distinguishing between unsheltered and sheltered homeless populations, as well as examining their correlation with factors like unemployment rates, housing affordability, and labor demand. Initial data is collected in San Francisco, while pre-existing information is drawn from three cities: San Francisco, New York City, and Washington D.C., facilitating the conduction of simulations. The third contribution focuses on demonstrating the practical implications of the data processing results. The challenges faced by key stakeholders, including charitable organizations and local city governments, are taken into consideration. Two case studies are presented as examples. The first case study explores improving the efficiency of food and necessities distribution, as well as medical assistance, driven by charitable organizations. The second case study examines the correlation between micro-geographic budget expenditure by local city governments and homeless information to justify budget allocation and expenditures. The ultimate objective of this endeavor is to enable the continuous enhancement of the quality of life for the underprivileged. It is hoped that through increased crowdsourcing of data from the public, the Generosity Curve and the Need Curve will intersect, leading to a better world for all.

Keywords: crowdsourcing, homelessness, socio-economic policies, statistical analysis

Procedia PDF Downloads 36
640 The Nursing Profession in Algeria between Humane Treatment and Work Environment Problems - A Field Study

Authors: Bacha Zakaria

Abstract:

This study aimed to investigate the reality of humane treatment and work environment problems for nurses in public hospitals and their repercussions on the patients arriving there. In this curve, our field study was based on a sample of nurses in Algiers hospitals estimated at 100 nurses. The questionnaire prepared by the two researchers was applied face to face with the nurses, and after obtaining and analyzing the data, we concluded the most important results: The presence of many problems in the work environment, such as work pressures, lack of appreciation, verbal and physical violence, risk of infection, poor salary and incentives, working during fatigue, administrative problems etc. And accordingly, The embodiment of humane dealing with patients requires providing a humane work environment for nurses and dealing with them humanely so that they embody positive behaviors while dealing with patients.

Keywords: nursing, future, family-focused care, health equity

Procedia PDF Downloads 63
639 Model Order Reduction for Frequency Response and Effect of Order of Method for Matching Condition

Authors: Aref Ghafouri, Mohammad javad Mollakazemi, Farhad Asadi

Abstract:

In this paper, model order reduction method is used for approximation in linear and nonlinearity aspects in some experimental data. This method can be used for obtaining offline reduced model for approximation of experimental data and can produce and follow the data and order of system and also it can match to experimental data in some frequency ratios. In this study, the method is compared in different experimental data and influence of choosing of order of the model reduction for obtaining the best and sufficient matching condition for following the data is investigated in format of imaginary and reality part of the frequency response curve and finally the effect and important parameter of number of order reduction in nonlinear experimental data is explained further.

Keywords: frequency response, order of model reduction, frequency matching condition, nonlinear experimental data

Procedia PDF Downloads 372
638 Potential Biosorption of Rhodococcus erythropolis, an Isolated Strain from Sossego Copper Mine, Brazil

Authors: Marcela dos P. G. Baltazar, Louise H. Gracioso, Luciana J. Gimenes, Bruno Karolski, Ingrid Avanzi, Elen A. Perpetuo

Abstract:

In this work, bacterial strains were isolated from environmental samples from a copper mine and three of them presented potential for bioremediation of copper. All the strains were identified by mass spectrometry (MALDI-TOF-Biotyper) and grown in three diferent media supplemented with 100 ppm of copper chloride in flasks of 500mL and it was incubated at 28 °C and 180 rpm. Periodically, samples were taken and monitored for cellular growth and copper biosorption by spectrophotometer UV-Vis (600 nm) and Inductively Coupled Plasma Optical Emission Spectrometry (ICP-OES), respectively. At the end of exponential phase of cellular growth, the biomass was utilized to construct a correlation curve between absorbance and dry mass of the cells. Among the three isolates with potential for biorremediation, 1 strain exhibit capacity the most for bioremediation of effluents contaminated by copper being identified as Rhodococcus erythropolis.

Keywords: bioprocess, bioremediation, biosorption, copper

Procedia PDF Downloads 358
637 Re-Evaluation of Field X Located in Northern Lake Albert Basin to Refine the Structural Interpretation

Authors: Calorine Twebaze, Jesca Balinga

Abstract:

Field X is located on the Eastern shores of L. Albert, Uganda, on the rift flank where the gross sedimentary fill is typically less than 2,000m. The field was discovered in 2006 and encountered about 20.4m of net pay across three (3) stratigraphic intervals within the discovery well. The field covers an area of 3 km2, with the structural configuration comprising a 3-way dip-closed hanging wall anticline that seals against the basement to the southeast along the bounding fault. Field X had been mapped on reprocessed 3D seismic data, which was originally acquired in 2007 and reprocessed in 2013. The seismic data quality is good across the field, and reprocessing work reduced the uncertainty in the location of the bounding fault and enhanced the lateral continuity of reservoir reflectors. The current study was a re-evaluation of Field X to refine fault interpretation and understand the structural uncertainties associated with the field. The seismic data, and three (3) wells datasets were used during the study. The evaluation followed standard workflows using Petrel software and structural attribute analysis. The process spanned from seismic- -well tie, structural interpretation, and structural uncertainty analysis. Analysis of three (3) well ties generated for the 3 wells provided a geophysical interpretation that was consistent with geological picks. The generated time-depth curves showed a general increase in velocity with burial depth. However, separation in curve trends observed below 1100m was mainly attributed to minimal lateral variation in velocity between the wells. In addition to Attribute analysis, three velocity modeling approaches were evaluated, including the Time-Depth Curve, Vo+ kZ, and Average Velocity Method. The generated models were calibrated at well locations using well tops to obtain the best velocity model for Field X. The Time-depth method resulted in more reliable depth surfaces with good structural coherence between the TWT and depth maps with minimal error at well locations of 2 to 5m. Both the NNE-SSW rift border fault and minor faults in the existing interpretation were reevaluated. However, the new interpretation delineated an E-W trending fault in the northern part of the field that had not been interpreted before. The fault was interpreted at all stratigraphic levels and thus propagates from the basement to the surface and is an active fault today. It was also noted that the entire field is less faulted with more faults in the deeper part of the field. The major structural uncertainties defined included 1) The time horizons due to reduced data quality, especially in the deeper parts of the structure, an error equal to one-third of the reflection time thickness was assumed, 2) Check shot analysis showed varying velocities within the wells thus varying depth values for each well, and 3) Very few average velocity points due to limited wells produced a pessimistic average Velocity model.

Keywords: 3D seismic data interpretation, structural uncertainties, attribute analysis, velocity modelling approaches

Procedia PDF Downloads 26
636 Decomposition of the Discount Function Into Impatience and Uncertainty Aversion. How Neurofinance Can Help to Understand Behavioral Anomalies

Authors: Roberta Martino, Viviana Ventre

Abstract:

Intertemporal choices are choices under conditions of uncertainty in which the consequences are distributed over time. The Discounted Utility Model is the essential reference for describing the individual in the context of intertemporal choice. The model is based on the idea that the individual selects the alternative with the highest utility, which is calculated by multiplying the cardinal utility of the outcome, as if the reception were instantaneous, by the discount function that determines a decrease in the utility value according to how the actual reception of the outcome is far away from the moment the choice is made. Initially, the discount function was assumed to have an exponential trend, whose decrease over time is constant, in line with a profile of a rational investor described by classical economics. Instead, empirical evidence called for the formulation of alternative, hyperbolic models that better represented the actual actions of the investor. Attitudes that do not comply with the principles of classical rationality are termed anomalous, i.e., difficult to rationalize and describe through normative models. The development of behavioral finance, which describes investor behavior through cognitive psychology, has shown that deviations from rationality are due to the limited rationality condition of human beings. What this means is that when a choice is made in a very difficult and information-rich environment, the brain does a compromise job between the cognitive effort required and the selection of an alternative. Moreover, the evaluation and selection phase of the alternative, the collection and processing of information, are dynamics conditioned by systematic distortions of the decision-making process that are the behavioral biases involving the individual's emotional and cognitive system. In this paper we present an original decomposition of the discount function to investigate the psychological principles of hyperbolic discounting. It is possible to decompose the curve into two components: the first component is responsible for the smaller decrease in the outcome as time increases and is related to the individual's impatience; the second component relates to the change in the direction of the tangent vector to the curve and indicates how much the individual perceives the indeterminacy of the future indicating his or her aversion to uncertainty. This decomposition allows interesting conclusions to be drawn with respect to the concept of impatience and the emotional drives involved in decision-making. The contribution that neuroscience can make to decision theory and inter-temporal choice theory is vast as it would allow the description of the decision-making process as the relationship between the individual's emotional and cognitive factors. Neurofinance is a discipline that uses a multidisciplinary approach to investigate how the brain influences decision-making. Indeed, considering that the decision-making process is linked to the activity of the prefrontal cortex and amygdala, neurofinance can help determine the extent to which abnormal attitudes respect the principles of rationality.

Keywords: impatience, intertemporal choice, neurofinance, rationality, uncertainty

Procedia PDF Downloads 101
635 Groundwater Recharge Suitability Mapping Using Analytical Hierarchy Process Based-Approach

Authors: Aziza Barrek, Mohamed Haythem Msaddek, Ismail Chenini

Abstract:

Excessive groundwater pumping due to the increasing water demand, especially in the agricultural sector, causes groundwater scarcity. Groundwater recharge is the most important process that contributes to the water's durability. This paper is based on the Analytic Hierarchy Process multicriteria analysis to establish a groundwater recharge susceptibility map. To delineate aquifer suitability for groundwater recharge, eight parameters were used: soil type, land cover, drainage density, lithology, NDVI, slope, transmissivity, and rainfall. The impact of each factor was weighted. This method was applied to the El Fahs plain shallow aquifer. Results suggest that 37% of the aquifer area has very good and good recharge suitability. The results have been validated by the Receiver Operating Characteristics curve. The accuracy of the prediction obtained was 89.3%.

Keywords: AHP, El Fahs aquifer, empirical formula, groundwater recharge zone, remote sensing, semi-arid region

Procedia PDF Downloads 83
634 Computer Aide Discrimination of Benign and Malignant Thyroid Nodules by Ultrasound Imaging

Authors: Akbar Gharbali, Ali Abbasian Ardekani, Afshin Mohammadi

Abstract:

Introduction: Thyroid nodules have an incidence of 33-68% in the general population. More than 5-15% of these nodules are malignant. Early detection and treatment of thyroid nodules increase the cure rate and provide optimal treatment. Between the medical imaging methods, Ultrasound is the chosen imaging technique for assessment of thyroid nodules. The confirming of the diagnosis usually demands repeated fine-needle aspiration biopsy (FNAB). So, current management has morbidity and non-zero mortality. Objective: To explore diagnostic potential of automatic texture analysis (TA) methods in differentiation benign and malignant thyroid nodules by ultrasound imaging in order to help for reliable diagnosis and monitoring of the thyroid nodules in their early stages with no need biopsy. Material and Methods: The thyroid US image database consists of 70 patients (26 benign and 44 malignant) which were reported by Radiologist and proven by the biopsy. Two slices per patient were loaded in Mazda Software version 4.6 for automatic texture analysis. Regions of interests (ROIs) were defined within the abnormal part of the thyroid nodules ultrasound images. Gray levels within an ROI normalized according to three normalization schemes: N1: default or original gray levels, N2: +/- 3 Sigma or dynamic intensity limited to µ+/- 3σ, and N3: present intensity limited to 1% - 99%. Up to 270 multiscale texture features parameters per ROIs per each normalization schemes were computed from well-known statistical methods employed in Mazda software. From the statistical point of view, all calculated texture features parameters are not useful for texture analysis. So, the features based on maximum Fisher coefficient and the minimum probability of classification error and average correlation coefficients (POE+ACC) eliminated to 10 best and most effective features per normalization schemes. We analyze this feature under two standardization states (standard (S) and non-standard (NS)) with Principle Component Analysis (PCA), Linear Discriminant Analysis (LDA) and Non-Linear Discriminant Analysis (NDA). The 1NN classifier was performed to distinguish between benign and malignant tumors. The confusion matrix and Receiver operating characteristic (ROC) curve analysis were used for the formulation of more reliable criteria of the performance of employed texture analysis methods. Results: The results demonstrated the influence of the normalization schemes and reduction methods on the effectiveness of the obtained features as a descriptor on discrimination power and classification results. The selected subset features under 1%-99% normalization, POE+ACC reduction and NDA texture analysis yielded a high discrimination performance with the area under the ROC curve (Az) of 0.9722, in distinguishing Benign from Malignant Thyroid Nodules which correspond to sensitivity of 94.45%, specificity of 100%, and accuracy of 97.14%. Conclusions: Our results indicate computer-aided diagnosis is a reliable method, and can provide useful information to help radiologists in the detection and classification of benign and malignant thyroid nodules.

Keywords: ultrasound imaging, thyroid nodules, computer aided diagnosis, texture analysis, PCA, LDA, NDA

Procedia PDF Downloads 258
633 Modelling for Temperature Non-Isothermal Continuous Stirred Tank Reactor Using Fuzzy Logic

Authors: Nasser Mohamed Ramli, Mohamad Syafiq Mohamad

Abstract:

Many types of controllers were applied on the continuous stirred tank reactor (CSTR) unit to control the temperature. In this research paper, Proportional-Integral-Derivative (PID) controller are compared with Fuzzy Logic controller for temperature control of CSTR. The control system for temperature non-isothermal of a CSTR will produce a stable response curve to its set point temperature. A mathematical model of a CSTR using the most general operating condition was developed through a set of differential equations into S-function using MATLAB. The reactor model and S-function are developed using m.file. After developing the S-function of CSTR model, User-Defined functions are used to link to SIMULINK file. Results that are obtained from simulation and temperature control were better when using Fuzzy logic control compared to PID control.

Keywords: CSTR, temperature, PID, fuzzy logic

Procedia PDF Downloads 423
632 Quick Covering Machine for Grain Drying Pavement

Authors: Fatima S. Rodriguez, Victorino T. Taylan, Manolito C. Bulaong, Helen F. Gavino, Vitaliana U. Malamug

Abstract:

In sundrying, the quality of the grains are greatly reduced when paddy grains were caught by the rain unsacked and unstored resulting to reduced profit. The objectives of this study were to design and fabricate a quick covering machine for grain drying pavement to test and evaluate the operating characteristics of the machine according to its deployment speed, recovery speed, deployment time, recovery time, power consumption, aesthetics of laminated sack, conducting partial budget, and cost curve analysis. The machine was able to cover the grains in a 12.8 m x 22.5 m grain drying pavement at an average time of 17.13 s. It consumed 0 .53 W-hr for the deployment and recovery of the cover. The machine entailed an investment cost of $1,344.40 and an annual cost charge of $647.32. Moreover, the savings per year using the quick covering machine was $101.83.

Keywords: quick, covering machine, grain, drying pavement

Procedia PDF Downloads 336
631 Biogas Separation, Alcohol Amine Solutions

Authors: Jingxiao Liang, David Rooneyman

Abstract:

Biogas, which is a valuable renewable energy source, can be produced by anaerobic fermentation of agricultural waste, manure, municipal waste, plant material, sewage, green waste, or food waste. It is composed of methane (CH4) and carbon dioxide (CO2) but also contains significant quantities of undesirable compounds such as hydrogen sulfide (H2S), ammonia (NH3), and siloxanes. Since typical raw biogas contains 25–45% CO2, The requirements for biogas quality depend on its further application. Before biogas is being used more efficiently, CO2 should be removed. One of the existing options for biogas separation technologies is based on chemical absorbents, in particular, mono-, di- and tri-alcohol amine solutions. Such amine solutions have been applied as highly efficient CO2 capturing agents. The benchmark in this experiment is N-methyldiethanolamine (MDEA) with piperazine (PZ) as an activator, from CO2 absorption Isotherm curve, optimization conditions are collected, such as activator percentage, temperature etc. This experiment makes new alcohol amines, which could have the same CO2 absorbing ability as activated MDEA, using glycidol as one of reactant, the result is quite satisfying.

Keywords: biogas, CO2, MDEA, separation

Procedia PDF Downloads 595
630 Dy³+/Eu³+ Co-Activated Gadolinium Aluminate Borate Phosphor: Enhanced Luminescence and Color Output Tuning

Authors: Osama Madkhali

Abstract:

GdAl₃(BO₃)₄ phosphors, incorporating Dy³+ and Dy³+/Eu³+ activators, were successfully synthesized via the gel combustion method. Powder X-ray diffraction (XRD) was utilized to ascertain phase purity and assess the impact of dopant concentration on the crystallographic structure. Photoluminescence (PL) measurements revealed that luminescence properties' intensity and lifetime varied with Dy³+ and Eu³+ ion concentrations. The relationship between luminescence intensity and doping concentration was explored in the context of the energy transfer process between Eu³+ and Dy³+ ions. An increase in Eu³+ co-doping concentrations resulted in a decrease in luminescence lifetime. Energy transfer efficiency was significantly enhanced from 26% to 84% with Eu³+ co-doping, as evidenced by decay curve analysis. These findings position GdAl₃(BO₃)4: Dy³+, Eu³+ phosphors as promising candidates for LED applications in solid-state lighting and displays.

Keywords: GdAl₃(BO₃)₄ phosphors, Dy³+/Eu³+ co-doping, photoluminescence (PL) measurements, luminescence properties, LED applications, solid-state lighting

Procedia PDF Downloads 29
629 Using Multi-Level Analysis to Identify Future Trends in Small Device Digital Communication Examinations

Authors: Mark A. Spooner

Abstract:

The growth of technological advances in the digital communications industry has dictated the way forensic examination laboratories receive, analyze, and report on digital evidence. This study looks at the trends in a medium sized digital forensics lab that examines small communications devices (i.e., cellular telephones, tablets, thumb drives, etc.) over the past five years. As law enforcement and homeland security organizations budgets shrink, many agencies are being asked to perform more examinations with less resources available. Using multi-level statistical analysis using five years of examination data, this research shows the increasing technological demand trend. The research then extrapolates the current data into the model created and finds a continued exponential growth curve of said demands is well within the parameters defined earlier on in the research.

Keywords: digital forensics, forensic examination, small device, trends

Procedia PDF Downloads 175
628 Graphene Metamaterials Supported Tunable Terahertz Fano Resonance

Authors: Xiaoyong He

Abstract:

The manipulation of THz waves is still a challenging task due to lack of natural materials interacted with it strongly. Designed by tailoring the characters of unit cells (meta-molecules), the advance of metamaterials (MMs) may solve this problem. However, because of Ohmic and radiation losses, the performance of MMs devices is subjected to the dissipation and low quality factor (Q-factor). This dilemma may be circumvented by Fano resonance, which arises from the destructive interference between a bright continuum mode and dark discrete mode (or a narrow resonance). Different from symmetric Lorentz spectral curve, Fano resonance indicates a distinct asymmetric line-shape, ultrahigh quality factor, steep variations in spectrum curves. Fano resonance is usually realized through symmetry breaking. However, if concentric double rings (DR) are placed closely to each other, the near-field coupling between them gives rise to two hybridized modes (bright and narrowband dark modes) because of the local asymmetry, resulting into the characteristic Fano line shape. Furthermore, from the practical viewpoint, it is highly desirable requirement that to achieve the modulation of Fano spectral curves conveniently, which is an important and interesting research topics. For current Fano systems, the tunable spectral curves can be realized by adjusting the geometrical structural parameters or magnetic fields biased the ferrite-based structure. But due to limited dispersion properties of active materials, it is still a tough work to tailor Fano resonance conveniently with the fixed structural parameters. With the favorable properties of extreme confinement and high tunability, graphene is a strong candidate to achieve this goal. The DR-structure possesses the excitation of so-called “trapped modes,” with the merits of simple structure and high quality of resonances in thin structures. By depositing graphene circular DR on the SiO2/Si/ polymer substrate, the tunable Fano resonance has been theoretically investigated in the terahertz regime, including the effects of graphene Fermi level, structural parameters and operation frequency. The results manifest that the obvious Fano peak can be efficiently modulated because of the strong coupling between incident waves and graphene ribbons. As Fermi level increases, the peak amplitude of Fano curve increases, and the resonant peak position shifts to high frequency. The amplitude modulation depth of Fano curves is about 30% if Fermi level changes in the scope of 0.1-1.0 eV. The optimum gap distance between DR is about 8-12 μm, where the value of figure of merit shows a peak. As the graphene ribbon width increases, the Fano spectral curves become broad, and the resonant peak denotes blue shift. The results are very helpful to develop novel graphene plasmonic devices, e.g. sensors and modulators.

Keywords: graphene, metamaterials, terahertz, tunable

Procedia PDF Downloads 323
627 Non-Linear Static Pushover Analysis of 15 Storied Reinforced Concrete Building Structure with Shear Wall

Authors: Hamid Nikzad, Shinta Yoshitomi

Abstract:

In this paper, nonlinear static pushover analysis is performed on 15 storied RC building structure with a shear wall to evaluate the seismic performance of the building. Section sizes of the members are obtained based on structural optimization method utilizing MATLAB frame optimizer, then the structure is simulated and designed in ETABS program conforming ACI 318-14 design code. The pushover curve has been generated by pushing the top node of the structure to the limited target displacement. Members failure due to the formation of plastic hinges, considering shear wall-frame structure was observed and the result of this study is presented based on current regulation of FEMA356, ASCE7-10, and ACI 318-14 design criteria

Keywords: structural optimization, linear static analysis, ETABS, MATLAB, RC moment frame, RC shear wall structures

Procedia PDF Downloads 132
626 Dynamic Analysis of Offshore 2-HUS/U Parallel Platform

Authors: Xie Kefeng, Zhang He

Abstract:

For the stability and control demand of offshore small floating platform, a 2-HUS/U parallel mechanism was presented as offshore platform. Inverse kinematics was obtained by institutional constraint equation, and the dynamic model of offshore 2-HUS/U parallel platform was derived based on rigid body’s Lagrangian method. The equivalent moment of inertia, damping and driving force/torque variation of offshore 2-HUS/U parallel platform were analyzed. A numerical example shows that, for parallel platform of given motion, system’s equivalent inertia changes 1.25 times maximally. During the movement of platform, they change dramatically with the system configuration and have coupling characteristics. The maximum equivalent drive torque is 800 N. At the same time, the curve of platform’s driving force/torque is smooth and has good sine features. The control system needs to be adjusted according to kinetic equation during stability and control and it provides a basis for the optimization of control system.

Keywords: 2-HUS/U platform, dynamics, Lagrange, parallel platform

Procedia PDF Downloads 319
625 A Crowdsourced Homeless Data Collection System And Its Econometric Analysis: Strengthening Inclusive Public Administration Policies

Authors: Praniil Nagaraj

Abstract:

This paper proposes a method to collect homeless data using crowdsourcing and presents an approach to analyze the data, demonstrating its potential to strengthen existing and future policies aimed at promoting socio-economic equilibrium. The 2022 Annual Homeless Assessment Report (AHAR) to Congress highlighted alarming statistics, emphasizing the need for effective decision-making and budget allocation within local planning bodies known as Continuums of Care (CoC). This paper's contributions can be categorized into three main areas. Firstly, a unique method for collecting homeless data is introduced, utilizing a user-friendly smartphone app (currently available for Android). The app enables the general public to quickly record information about homeless individuals, including the number of people and details about their living conditions. The collected data, including date, time, and location, is anonymized and securely transmitted to the cloud. It is anticipated that an increasing number of users motivated to contribute to society will adopt the app, thus expanding the data collection efforts. Duplicate data is addressed through simple classification methods, and historical data is utilized to fill in missing information. The second contribution of this paper is the description of data analysis techniques applied to the collected data. By combining this new data with existing information, statistical regression analysis is employed to gain insights into various aspects, such as distinguishing between unsheltered and sheltered homeless populations, as well as examining their correlation with factors like unemployment rates, housing affordability, and labor demand. Initial data is collected in San Francisco, while pre-existing information is drawn from three cities: San Francisco, New York City, and Washington D.C., facilitating the conduction of simulations. The third contribution focuses on demonstrating the practical implications of the data processing results. The challenges faced by key stakeholders, including charitable organizations and local city governments, are taken into consideration. Two case studies are presented as examples. The first case study explores improving the efficiency of food and necessities distribution, as well as medical assistance, driven by charitable organizations. The second case study examines the correlation between micro-geographic budget expenditure by local city governments and homeless information to justify budget allocation and expenditures. The ultimate objective of this endeavor is to enable the continuous enhancement of the quality of life for the underprivileged. It is hoped that through increased crowdsourcing of data from the public, the Generosity Curve and the Need Curve will intersect, leading to a better world for all.

Keywords: crowdsourcing, homelessness, socio-economic policies, statistical regression

Procedia PDF Downloads 56
624 Determination of Effect Factor for Effective Parameter on Saccharification of Lignocellulosic Material by Concentrated Acid

Authors: Sina Aghili, Ali Arasteh Nodeh

Abstract:

Tamarisk usage as a new group of lignocelluloses material to produce fermentable sugars in bio-ethanol process was studied. The overall aim of this work was to establish the optimum condition for acid hydrolysis of this new material and a mathematical model predicting glucose release as a function of operation variable. Sulfuric acid concentration in the range of 20 to 60%(w/w), process temperature between 60 to 95oC, hydrolysis time from 120 to 240 min and solid content 5,10,15%(w/w) were used as hydrolysis conditions. HPLC was used to analysis of the product. This analysis indicated that glucose was the main fermentable sugar and was increased with time, temperature and solid content and acid concentration was a parabola influence in glucose production.The process was modeled by a quadratic equation. Curve study and model were found that 42% acid concentration, 15 % solid content and 90oC were in optimum condition.

Keywords: fermentable sugar, saccharification, wood, hydrolysis

Procedia PDF Downloads 313
623 Parametric Analysis of Water Lily Shaped Split Ring Resonator Loaded Fractal Monopole Antenna for Multiband Applications

Authors: C. Elavarasi, T. Shanmuganantham

Abstract:

A coplanar waveguide (CPW) feed is presented, and comprising a split ring resonator (SRR) loaded fractal with water lily shape is used for multi band applications. The impedance matching of the antenna is determined by the number of Koch curve fractal unit cells. The antenna is designed on a FR4 substrate with a permittivity of εr = 4.4 and size of 14 x 16 x 1.6 mm3 to generate multi resonant mode at 3.8 GHz covering S band, 8.68 GHz at X band, 13.96 GHz at Ku band, and 19.74 GHz at K band with reflection coefficient better than -10 dB. Simulation results show that the antenna exhibits the desired voltage standing wave ratio (VSWR) level and radiation patterns across the wide frequency range. The fundamental parameters of the antenna such as return loss, VSWR, good radiation pattern with reasonable gain across the operating bands are obtained.

Keywords: fractal, metamaterial, split ring resonator, waterlily shape

Procedia PDF Downloads 247
622 Design and Biomechanical Analysis of a Transtibial Prosthesis for Cyclists of the Colombian Team Paralympic

Authors: Jhonnatan Eduardo Zamudio Palacios, Oscar Leonardo Mosquera Dussan, Daniel Guzman Perez, Daniel Alfonso Botero Rosas, Oscar Fabian Rubiano Espinosa, Jose Antonio Garcia Torres, Ivan Dario Chavarro, Ivan Ramiro Rodriguez Camacho, Jaime Orlando Rodriguez

Abstract:

The training of cilsitas with some type of disability finds in the technological development an indispensable ally, generating every day advances to contribute to the quality of life allowing to maximize the capacities of the athletes. The performance of a cyclist depends on physiological and biomechanical factors, such as aerodynamic profile, bicycle measurements, connecting rod length, pedaling systems, type of competition, among others. This study particularly focuses on the description of the dynamic model of a transtibial prosthesis for Paralympic cyclists. To make the model, two points are chosen: in the radius centers of rotation of the plate and pinion of the track bicycle. The parametric scheme of the track bike represents a model of 6 degrees of freedom due to the displacement in X - Y of each of the reference points of the angles of the curve profile β, cant of the velodrome α and the angle of rotation of the connecting rod φ. The force exerted on the crank of the bicycle varies according to the angles of the curve profile β, the velodrome cant of α and the angle of rotation of the crank φ. The behavior is analyzed through the Matlab R2015a software. The average strength that a cyclist exerts on the cranks of a bicycle is 1,607.1 N, the Paralympic cyclist must perform a force on each crank about 803.6 N. Once the maximum force associated with the movement has been determined, it is continued to the dynamic modeling of the transtibial prosthesis that represents a model of 6 degrees of freedom with displacement in X - Y in relation to the angles of rotation of the hip π, knee γ and ankle λ. Subsequently, an analysis of the kinematic behavior of the prosthesis was carried out by means of SolidWorks 2017 and Matlab R2015a, which was used to model and analyze the variation of the hip angles π, knee γ and ankle of the λ prosthesis. The reaction forces generated in the prosthesis were performed on the ankle of the prosthesis, performing the summation of forces on the X and Y axes. The same analysis was then applied to the tibia of the prosthesis and the socket. The reaction force of the parts of the prosthesis varies according to the hip angles π, knee γ and ankle of the prosthesis λ. Therefore, it can be deduced that the maximum forces experienced by the ankle of the prosthesis is 933.6 N on the X axis and 2.160.5 N on the Y axis. Finally, it is calculated that the maximum forces experienced by the tibia and the socket of the transtibial prosthesis in high performance competitions is 3.266 N on the X axis and 1.357 N on the Y axis. In conclusion, it can be said that the performance of the cyclist depends on several physiological factors, linked to biomechanics of training. The influence of biomechanical factors such as aerodynamics, bicycle measurements, connecting rod length, or non-circular pedaling systems on the cyclist performance.

Keywords: biomechanics, dynamic model, paralympic cyclist, transtibial prosthesis

Procedia PDF Downloads 306
621 Classification of ECG Signal Based on Mixture of Linear and Non-Linear Features

Authors: Mohammad Karimi Moridani, Mohammad Abdi Zadeh, Zahra Shahiazar Mazraeh

Abstract:

In recent years, the use of intelligent systems in biomedical engineering has increased dramatically, especially in the diagnosis of various diseases. Also, due to the relatively simple recording of the electrocardiogram signal (ECG), this signal is a good tool to show the function of the heart and diseases associated with it. The aim of this paper is to design an intelligent system for automatically detecting a normal electrocardiogram signal from abnormal one. Using this diagnostic system, it is possible to identify a person's heart condition in a very short time and with high accuracy. The data used in this article are from the Physionet database, available in 2016 for use by researchers to provide the best method for detecting normal signals from abnormalities. Data is of both genders and the data recording time varies between several seconds to several minutes. All data is also labeled normal or abnormal. Due to the low positional accuracy and ECG signal time limit and the similarity of the signal in some diseases with the normal signal, the heart rate variability (HRV) signal was used. Measuring and analyzing the heart rate variability with time to evaluate the activity of the heart and differentiating different types of heart failure from one another is of interest to the experts. In the preprocessing stage, after noise cancelation by the adaptive Kalman filter and extracting the R wave by the Pan and Tampkinz algorithm, R-R intervals were extracted and the HRV signal was generated. In the process of processing this paper, a new idea was presented that, in addition to using the statistical characteristics of the signal to create a return map and extraction of nonlinear characteristics of the HRV signal due to the nonlinear nature of the signal. Finally, the artificial neural networks widely used in the field of ECG signal processing as well as distinctive features were used to classify the normal signals from abnormal ones. To evaluate the efficiency of proposed classifiers in this paper, the area under curve ROC was used. The results of the simulation in the MATLAB environment showed that the AUC of the MLP and SVM neural network was 0.893 and 0.947, respectively. As well as, the results of the proposed algorithm in this paper indicated that the more use of nonlinear characteristics in normal signal classification of the patient showed better performance. Today, research is aimed at quantitatively analyzing the linear and non-linear or descriptive and random nature of the heart rate variability signal, because it has been shown that the amount of these properties can be used to indicate the health status of the individual's heart. The study of nonlinear behavior and dynamics of the heart's neural control system in the short and long-term provides new information on how the cardiovascular system functions, and has led to the development of research in this field. Given that the ECG signal contains important information and is one of the common tools used by physicians to diagnose heart disease, but due to the limited accuracy of time and the fact that some information about this signal is hidden from the viewpoint of physicians, the design of the intelligent system proposed in this paper can help physicians with greater speed and accuracy in the diagnosis of normal and patient individuals and can be used as a complementary system in the treatment centers.

Keywords: neart rate variability, signal processing, linear and non-linear features, classification methods, ROC Curve

Procedia PDF Downloads 225
620 Electrical Performance Analysis of Single Junction Amorphous Silicon Solar (a-Si:H) Modules Using IV Tracer (PVPM)

Authors: Gilbert Omorodion Osayemwenre, Edson Meyer, R. T. Taziwa

Abstract:

The electrical analysis of single junction amorphous silicon solar modules is carried out using outdoor monitoring technique. Like crystalline silicon PV modules, the electrical characterisation and performance of single junction amorphous silicon modules are best described by its current-voltage (IV) characteristic. However, IV curve has a direct dependence on the type of PV technology and material properties used. The analysis reveals discrepancies in the modules performance parameter even though they are of similar technology. The aim of this work is to compare the electrical performance output of each module, using electrical parameters with the aid of PVPM 100040C IV tracer. These results demonstrated the relevance of standardising the performance parameter for effective degradation analysis of a-Si:H.

Keywords: PVPM 100040C IV tracer, SolarWatt part, single junction amorphous silicon module (a-Si:H), Staebler-Wronski (S-W) degradation effect

Procedia PDF Downloads 289
619 Numerical Investigation of Flow and Heat Transfer Characteristics of a Natural Refrigerant within a Vortex Tube

Authors: Mirza Popovac

Abstract:

This paper investigates the application of the vortex tubes towards increasing the efficiency of high temperature heat pumps based on natural refrigerants, by recovering a part of the expansion work within the refrigerant cycle. To this purpose the 3D Navier-Stokes solver is used to perform a set of numerical simulations, investigating the vortex tube performance. Firstly, the fluid flow and heat transfer characteristics are analyzed for standard configurations of vortex tubes, and the obtained results are validated against the experimental and numerical data available in literature. Subsequently, different geometry specifications are analyzed, as well as the interplay between relevant heat pump operating conditions and the properties of natural refrigerants. Finally, the characteristic curve of performance will be derived for investigated vortex tubes specifications when used within high temperature heat pumps.

Keywords: heat pump, vortex tube, CFD, natural refrigerant

Procedia PDF Downloads 115
618 Electrochemical Studies of Si, Si-Ge- and Ge-Air Batteries

Authors: R. C. Sharma, Rishabh Bansal, Prajwal Menon, Manoj K. Sharma

Abstract:

Silicon-air battery is highly promising for electric vehicles due to its high theoretical energy density (8470 Whkg⁻¹) and its discharge products are non-toxic. For the first time, pure silicon and germanium powders are used as anode material. Nickel wire meshes embedded with charcoal and manganese dioxide powder as cathode and concentrated potassium hydroxide is used as electrolyte. Voltage-time curves have been presented in this study for pure silicon and germanium powder and 5% and 10% germanium with silicon powder. Silicon powder cell assembly gives a stable voltage of 0.88 V for ~20 minutes while Si-Ge provides cell voltage of 0.80-0.76 V for ~10-12 minutes, and pure germanium cell provides cell voltage 0.80-0.76 V for ~30 minutes. The cell voltage is higher for concentrated (10%) sodium hydroxide solution (1.08 V) and it is stable for ~40 minutes. A sharp decrease in cell voltage beyond 40 min may be due to rapid corrosion.

Keywords: Silicon-air battery, Germanium-air battery, voltage-time curve, open circuit voltage, Anodic corrosion

Procedia PDF Downloads 208
617 Mathematical Model for Progressive Phase Distribution of Ku-band Reflectarray Antennas

Authors: M. Y. Ismail, M. Inam, A. F. M. Zain, N. Misran

Abstract:

Progressive phase distribution is an important consideration in reflect array antenna design which is required to form a planar wave in front of the reflect array aperture. This paper presents a detailed mathematical model in order to determine the required reflection phase values from individual element of a reflect array designed in Ku-band frequency range. The proposed technique of obtaining reflection phase can be applied for any geometrical design of elements and is independent of number of array elements. Moreover the model also deals with the solution of reflect array antenna design with both centre and off-set feed configurations. The theoretical modeling has also been implemented for reflect arrays constructed on 0.508 mm thickness of different dielectric substrates. The results show an increase in the slope of the phase curve from 4.61°/mm to 22.35°/mm by varying the material properties.

Keywords: mathematical modeling, progressive phase distribution, reflect array antenna, reflection phase

Procedia PDF Downloads 351
616 A Theoretical Overview of Thermoluminescence

Authors: Sadhana Agrawal, Tarkeshwari Verma, Shmbhavi Katyayan

Abstract:

The magnificently accentuating phenomenon of luminescence has gathered a lot of attentions from last few decades. Probably defined as the one involving emission of light from certain kinds of substances on absorbing various energies in the form of external stimulus, the phenomenon claims a versatile pertinence. First observed and reported in an extract of Ligrium Nephriticum by Monards, the phenomenon involves turning of crystal clear water into colorful fluid when comes in contact with the special wood. In words of Sir G.G. Stokes, the phenomenon actually involves three different techniques – absorption, excitation and emission. With variance in external stimulus, the corresponding luminescence phenomenon is obtained. Here, this paper gives a concise discussion of thermoluminescence which is one of the types of luminescence obtained when the external stimulus is given in form of heat energy. A deep insight of thermoluminescence put forward a qualitative analysis of various parameters such as glow curves peaks, trap depth, frequency factors and order of kinetics.

Keywords: frequency factor, glow curve peaks, thermoluminescence, trap depth

Procedia PDF Downloads 368
615 Efficient Internal Generator Based on Random Selection of an Elliptic Curve

Authors: Mustapha Benssalah, Mustapha Djeddou, Karim Drouiche

Abstract:

The random number generation (RNG) presents a significant importance for the security and the privacy of numerous applications, such as RFID technology and smart cards. Since, the quality of the generated bit sequences is paramount that a weak internal generator for example, can directly cause the entire application to be insecure, and thus it makes no sense to employ strong algorithms for the application. In this paper, we propose a new pseudo random number generator (PRNG), suitable for cryptosystems ECC-based, constructed by randomly selecting points from several elliptic curves randomly selected. The main contribution of this work is the increasing of the generator internal states by extending the set of its output realizations to several curves auto-selected. The quality and the statistical characteristics of the proposed PRNG are validated using the Chi-square goodness of fit test and the empirical Special Publication 800-22 statistical test suite issued by NIST.

Keywords: PRNG, security, cryptosystem, ECC

Procedia PDF Downloads 417