Search results for: optimal tracking
2966 Spatial Ecology of an Endangered Amphibian Litoria Raniformis within Modified Tasmanian Landscapes
Authors: Timothy Garvey, Don Driscoll
Abstract:
Within Tasmania, the growling grass frog (Litoria raniformis) has experienced a rapid contraction in distribution. This decline is primarily attributed to habitat loss through landscape modification and improved land drainage. Reductions in seasonal water-sources have placed increasing importance on permanent water bodies for reproduction and foraging. Tasmanian agricultural and commercial forestry landscapes often feature small artificial ponds, utilized for watering livestock and fighting wildfires. Improved knowledge of how L. raniformis may be exploiting anthropogenic ponds is required for improved conservation management. We implemented telemetric tracking in order to evaluate the spatial ecology of L. raniformis (n = 20) within agricultural and managed forestry sites, with tracking conducted periodically over the breeding season (November/December, January/February, March/April). We investigated (1) potential differences in habitat utilization between agricultural and plantation sites, and (2) the post-breeding dispersal of individual frogs. Frogs were found to remain in close proximity to ponds throughout November/December, with individuals occupying vegetative depauperate water bodies beginning to disperse by January/February. Dispersing individuals traversed exposed plantation understory and agricultural pasture land in order to enter patches of native scrubland. By March/April all individuals captured at minimally vegetated ponds had retreated to adjacent scrub corridors. Animals found in ponds featuring dense riparian vegetation were not recorded to disperse. No difference in behavior was recorded between sexes. Rising temperatures coincided with increased movement by individuals towards native scrub refugia. The patterns of movement reported in this investigation emphasize the significant contribution of manmade water-bodies towards the conservation of L. raniformis within modified landscapes. The use of natural scrubland as cyclical retreats between breeding seasons also highlights the importance of the continued preservation of remnant vegetation corridors. Loss of artificial dams or buffering scrubland in heavily altered landscapes could see the breakdown of the greater L. raniformis meta-population further threatening their regional persistence.Keywords: habitat loss, modified landscapes, spatial ecology, telemetry
Procedia PDF Downloads 1172965 Effect of Modified Atmosphere Packaging and Storage Temperatures on Quality of Shelled Raw Walnuts
Authors: M. Javanmard
Abstract:
This study was aimed at analyzing the effects of packaging (MAP) and preservation conditions on the packaged fresh walnut kernel quality. The central composite plan was used for evaluating the effect of oxygen (0–10%), carbon dioxide (0-10%), and temperature (4-26 °C) on qualitative characteristics of walnut kernels. Also, the response level technique was used to find the optimal conditions for interactive effects of factors, as well as estimating the best conditions of process using least amount of testing. Measured qualitative parameters were: peroxide index, color, decreased weight, mould and yeast counting test, and sensory evaluation. The results showed that the defined model for peroxide index, color, weight loss, and sensory evaluation is significant (p < 0.001), so that increase of temperature causes the peroxide value, color variation, and weight loss to increase and it reduces the overall acceptability of walnut kernels. An increase in oxygen percentage caused the color variation level and peroxide value to increase and resulted in lower overall acceptability of the walnuts. An increase in CO2 percentage caused the peroxide value to decrease, but did not significantly affect other indices (p ≥ 0.05). Mould and yeast were not found in any samples. Optimal packaging conditions to achieve maximum quality of walnuts include: 1.46% oxygen, 10% carbon dioxide, and temperature of 4 °C.Keywords: shelled walnut, MAP, quality, storage temperature
Procedia PDF Downloads 3882964 Material and Parameter Analysis of the PolyJet Process for Mold Making Using Design of Experiments
Authors: A. Kampker, K. Kreisköther, C. Reinders
Abstract:
Since additive manufacturing technologies constantly advance, the use of this technology in mold making seems reasonable. Many manufacturers of additive manufacturing machines, however, do not offer any suggestions on how to parameterize the machine to achieve optimal results for mold making. The purpose of this research is to determine the interdependencies of different materials and parameters within the PolyJet process by using design of experiments (DoE), to additively manufacture molds, e.g. for thermoforming and injection molding applications. Therefore, the general requirements of thermoforming molds, such as heat resistance, surface quality and hardness, have been identified. Then, different materials and parameters of the PolyJet process, such as the orientation of the printed part, the layer thickness, the printing mode (matte or glossy), the distance between printed parts and the scaling of parts, have been examined. The multifactorial analysis covers the following properties of the printed samples: Tensile strength, tensile modulus, bending strength, elongation at break, surface quality, heat deflection temperature and surface hardness. The key objective of this research is that by joining the results from the DoE with the requirements of the mold making, optimal and tailored molds can be additively manufactured with the PolyJet process. These additively manufactured molds can then be used in prototyping processes, in process testing and in small to medium batch production.Keywords: additive manufacturing, design of experiments, mold making, PolyJet, 3D-Printing
Procedia PDF Downloads 2552963 Algorithmic Skills Transferred from Secondary CSI Studies into Tertiary Education
Authors: Piroska Biró, Mária Csernoch, János Máth, Kálmán Abari
Abstract:
Testing the first year students of Informatics at the University of Debrecen revealed that students start their tertiary studies in programming with a low level of programming knowledge and algorithmic skills. The possible reasons which lead the students to this very unfortunate result were examined. The results of the test were compared to the students’ results in the school leaving exams and to their self-assessment values. It was found that there is only a slight connection between the students’ results in the test and in the school leaving exams, especially at intermediate level. Beyond this, the school leaving exams do not seem to enable students to evaluate their own abilities.Keywords: deep and surface approaches, metacognitive abilities, programming and algorithmic skills, school leaving exams, tracking code
Procedia PDF Downloads 3842962 Water Immersion Recovery for Swimmers in Hot Environments
Authors: Thanura Randula Abeywardena
Abstract:
This study recognized the effectiveness of cold-water immersion recovery post exhaustive short-term exercise. The purpose of this study was to understand if 16- 20°C of cold-water immersion would be beneficial in a tropical environment to achieve optimal recovery in sprint swim performance in comparison to 10-15°C of water immersion. Two 100m-sprint swim performance times were measured along with blood lactate (BLa), heart rate (HR) and rate of perceived exertion (RPE) in a 25m swimming pool with full body head out horizontal water immersions of 10-15°C, 16-20°C and 29-32°C (pool temperature) for 10 minutes followed by 5 minutes of seated passive rest outside; in between the two swim performances. Twelve well-trained adult swimmers (5 male and 5 female) within the top twenty in the Sri Lankan national swimming championships in 100m Butterfly and Freestyle in the years 2020 & 2021 volunteered for this study. One-way ANOVA analysis (p<0.05) suggested performance time, Bla and HR had no significant differences between the 3 conditions after the second sprint; however, RPE was significantly different with p=0.034 between 10-15°C and 16-20°C immersion conditions. The study suggested that the recovery post the two cold-water immersion conditions were similar in terms of performance and physiological factors; however, the 16-20°C temperature had a better “feel good” factor post sprint 2. Further study is recommended as there was participant bias with the swimmers not reaching optimal levels in sprint 1. Therefore, they might have possibly fully recovered before sprint 2, invalidating the physiological effect of recovery.Keywords: hydrotherapy, blood lactate, fatigue, recovery, sprint-performance, sprint-swimming
Procedia PDF Downloads 1022961 Julia-Based Computational Tool for Composite System Reliability Assessment
Authors: Josif Figueroa, Kush Bubbar, Greg Young-Morris
Abstract:
The reliability evaluation of composite generation and bulk transmission systems is crucial for ensuring a reliable supply of electrical energy to significant system load points. However, evaluating adequacy indices using probabilistic methods like sequential Monte Carlo Simulation can be computationally expensive. Despite this, it is necessary when time-varying and interdependent resources, such as renewables and energy storage systems, are involved. Recent advances in solving power network optimization problems and parallel computing have improved runtime performance while maintaining solution accuracy. This work introduces CompositeSystems, an open-source Composite System Reliability Evaluation tool developed in Julia™, to address the current deficiencies of commercial and non-commercial tools. This work introduces its design, validation, and effectiveness, which includes analyzing two different formulations of the Optimal Power Flow problem. The simulations demonstrate excellent agreement with existing published studies while improving replicability and reproducibility. Overall, the proposed tool can provide valuable insights into the performance of transmission systems, making it an important addition to the existing toolbox for power system planning.Keywords: open-source software, composite system reliability, optimization methods, Monte Carlo methods, optimal power flow
Procedia PDF Downloads 732960 Research on the Optimization of the Facility Layout of Efficient Cafeterias for Troops
Authors: Qing Zhang, Jiachen Nie, Yujia Wen, Guanyuan Kou, Peng Yu, Kun Xia, Qin Yang, Li Ding
Abstract:
BACKGROUND: A facility layout problem (FLP) is an NP-complete (non-deterministic polynomial) problem, which is hard to obtain an exact optimal solution. FLP has been widely studied in various limited spaces and workflows. For example, cafeterias with many types of equipment for troops cause chaotic processes when dining. OBJECTIVE: This article tried to optimize the layout of troops’ cafeteria and to improve the overall efficiency of the dining process. METHODS: First, the original cafeteria layout design scheme was analyzed from an ergonomic perspective and two new design schemes were generated. Next, three facility layout models were designed, and further simulation was applied to compare the total time and density of troops between each scheme. Last, an experiment of the dining process with video observation and analysis verified the simulation results. RESULTS: In a simulation, the dining time under the second new layout is shortened by 2.25% and 1.89% (p<0.0001, p=0.0001) compared with the other two layouts, while troops-flow density and interference both greatly reduced in the two new layouts. In the experiment, process completing time and the number of interference reduced as well, which verified corresponding simulation results. CONCLUSIONS: Our two new layout schemes are tested to be optimal by a series of simulation and space experiments. In future research, similar approaches could be applied when taking layout-design algorithm calculation into consideration.Keywords: layout optimization, dining efficiency, troops’ cafeteria, anylogic simulation, field experiment
Procedia PDF Downloads 1432959 Performance Evaluation of Adsorption Refrigerating Systems
Authors: Nadia Allouache, Omar Rahli
Abstract:
Many promising technologies have been developed to harness the sun's energy. These technologies help in economizing energy and environmental protection. The solar refrigerating systems are one of these important technologies. In addition to environmental benefits and energy saving, adsorption refrigerating systems have many advantages such as lack of moving parts, simplicity of construction and low operating costs. The work aimed to establish the main factors that affect the performances of an adsorption refrigerating system using different geometries of adsorbers and different adsorbent-adsorbate pairs. The numerical modeling of the heat and mass transfer in the system, using various working pairs, such as: activated carbon-ammonia, calcium chlorid-ammonia, activated carbon fiber- methanol and activated carbon AC35-methanol, show that the adsorber design can influence the system performances; The thermal performances of system are better in the annular configuration case. An optimal value of generating temperature is observed in annular adsorber case for which the thermal performance of the cooling system is maximal. While in the plate adsorber, above a certain value of generating temperature, the performance of the system remains almost constant. The environmental conditions such as solar radiation and pressure have a great influence in the system efficiency, and the choice of the working pair depends on the environmental conditions and the geometry of the adsorber.Keywords: adsorber geometry, numerical modeling, optimal environmental conditions, working pairs.
Procedia PDF Downloads 792958 Integrated Gesture and Voice-Activated Mouse Control System
Authors: Dev Pratap Singh, Harshika Hasija, Ashwini S.
Abstract:
The project aims to provide a touchless, intuitive interface for human-computer interaction, enabling users to control their computers using hand gestures and voice commands. The system leverages advanced computer vision techniques using the Media Pipe framework and OpenCV to detect and interpret real-time hand gestures, transforming them into mouse actions such as clicking, dragging, and scrolling. Additionally, the integration of a voice assistant powered by the speech recognition library allows for seamless execution of tasks like web searches, location navigation, and gesture control in the system through voice commands.Keywords: gesture recognition, hand tracking, machine learning, convolutional neural networks, natural language processing, voice assistant
Procedia PDF Downloads 102957 Optimal Consume of NaOH in Starches Gelatinization for Froth Flotation
Authors: André C. Silva, Débora N. Sousa, Elenice M. S. Silva, Thales P. Fontes, Raphael S. Tomaz
Abstract:
Starches are widely used as depressant in froth flotation operations in Brazil due to their efficiency, increasing the selectivity in the inverse flotation of quartz depressing iron ore. Starches market have been growing and improving in recent years, leading to better products attending the requirements of the mineral industry. The major source of starch used for iron ore is corn starch, which needs to be gelatinized with sodium hydroxide (NaOH) prior to use. This stage has a direct impact on industrials costs, once the lowest consumption of NaOH in gelatinization provides better control of the pH in the froth flotation and reduces the amount of electrolytes present in the pulp. In order to evaluate the gelatinization degree of different starches and flour were subjected to the addiction of NaOH and temperature variation experiments. Samples of starch (corn, cassava, HIPIX 100, HIPIX 101 and HIPIX 102 commercialized by Ingredion) and flour (cassava and potato) were tested. The starch samples were characterized through Scanning Electronic Microscopy and the amylose content were determined through spectrometry, swelling and solubility tests. The gelatinization was carried out through titration with NaOH, keeping the solution temperature constant at 40 oC. At the end of the tests, the optimal amount of NaOH consumed to gelatinize the starch or flour from different botanical sources was established and a correlation between the content of amylopectin in the starch and the starch/NaOH ratio needed for its gelatinization.Keywords: froth flotation, gelatinization, sodium hydroxide, starches and flours
Procedia PDF Downloads 3632956 An Exploration of Health Promotion Approach to Increase Optimal Complementary Feeding among Pastoral Mothers Having Children between 6 and 23 Months in Dikhil, Djibouti
Authors: Haruka Ando
Abstract:
Undernutrition of children is a critical issue, especially for people in the remote areas of the Republic of Djibouti, since household food insecurity, inadequate child caring and feeding, unhealthy environment and lack of clean water, as well as insufficient maternal and child healthcare, are underlying causes which affect. Nomadic pastoralists living in the Dikhil region (Dikhil) are socio-economically and geographically more vulnerable due to displacement, which in turn worsens the situation of child stunting. A high prevalence of inappropriate complementary feeding among pastoral mothers might be a significant barrier to child growth. This study aims to identify health promotion intervention strategies that would support an increase in optimal complementary feeding among pastoral mothers of children aged 6-23 months in Dikhil. There are four objectives; to explore and to understand the existing practice of complementary feeding among pastoral mothers in Dikhil; to identify the barriers in appropriate complementary feeding among the mothers; to critically explore and analyse the strategies for an increase in complementary feeding among the mothers; to make pragmatic recommendations to address the barriers in Djibouti. This is an in-depth study utilizing a conceptual framework, the behaviour change wheel, to analyse the determinants of complementary feeding and categorize health promotion interventions for increasing optimal complementary feeding among pastoral mothers living in Dikhil. The analytical tool was utilized to appraise the strategies to mitigate the selected barriers against optimal complementary feeding. The data sources were secondary literature from both published and unpublished sources. The literature was systematically collected. The findings of the determinants including the barriers of optimal complementary feeding were identified: heavy household workload, caring for multiple children under five, lack of education, cultural norms and traditional eating habits, lack of husbands' support, poverty and food insecurity, lack of clean water, low media coverage, insufficient health services on complementary feeding, fear, poor personal hygiene, and mothers' low decision-making ability and lack of motivation for food choice. To mitigate selected barriers of optimal complementary feeding, four intervention strategies based on interpersonal communication at the community-level were chosen: scaling up mothers' support groups, nutrition education, grandmother-inclusive approach, and training for complementary feeding counseling. The strategies were appraised through the criteria of effectiveness and feasibility. Scaling up mothers' support groups could be the best approach. Mid-term and long-term recommendations are suggested based on the situation analysis and appraisal of intervention strategies. Mid-term recommendations include complementary feeding promotion interventions are integrated into the healthcare service providing system in Dikhil, and donor agencies advocate and lobby the Ministry of Health Djibouti (MoHD) to increase budgetary allocation on complementary feeding promotion to implement interventions at a community level. Moreover, the recommendations include a community health management team in Dikhil training healthcare workers and mother support groups by using complementary feeding communication guidelines and monitors behaviour change of pastoral mothers and health outcome of their children. Long-term recommendations are the MoHD develops complementary feeding guidelines to cover sector-wide collaboration for multi-sectoral related barriers.Keywords: Afar, child food, child nutrition, complementary feeding, complementary food, developing countries, Djibouti, East Africa, hard-to-reach areas, Horn of Africa, nomad, pastoral, rural area, Somali, Sub-Saharan Africa
Procedia PDF Downloads 1252955 Application of Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Multipoint Optimal Minimum Entropy Deconvolution in Railway Bearings Fault Diagnosis
Authors: Yao Cheng, Weihua Zhang
Abstract:
Although the measured vibration signal contains rich information on machine health conditions, the white noise interferences and the discrete harmonic coming from blade, shaft and mash make the fault diagnosis of rolling element bearings difficult. In order to overcome the interferences of useless signals, a new fault diagnosis method combining Complete Ensemble Empirical Mode Decomposition with adaptive noise (CEEMDAN) and Multipoint Optimal Minimum Entropy Deconvolution (MOMED) is proposed for the fault diagnosis of high-speed train bearings. Firstly, the CEEMDAN technique is applied to adaptively decompose the raw vibration signal into a series of finite intrinsic mode functions (IMFs) and a residue. Compared with Ensemble Empirical Mode Decomposition (EEMD), the CEEMDAN can provide an exact reconstruction of the original signal and a better spectral separation of the modes, which improves the accuracy of fault diagnosis. An effective sensitivity index based on the Pearson's correlation coefficients between IMFs and raw signal is adopted to select sensitive IMFs that contain bearing fault information. The composite signal of the sensitive IMFs is applied to further analysis of fault identification. Next, for propose of identifying the fault information precisely, the MOMED is utilized to enhance the periodic impulses in composite signal. As a non-iterative method, the MOMED has better deconvolution performance than the classical deconvolution methods such Minimum Entropy Deconvolution (MED) and Maximum Correlated Kurtosis Deconvolution (MCKD). Third, the envelope spectrum analysis is applied to detect the existence of bearing fault. The simulated bearing fault signals with white noise and discrete harmonic interferences are used to validate the effectiveness of the proposed method. Finally, the superiorities of the proposed method are further demonstrated by high-speed train bearing fault datasets measured from test rig. The analysis results indicate that the proposed method has strong practicability.Keywords: bearing, complete ensemble empirical mode decomposition with adaptive noise, fault diagnosis, multipoint optimal minimum entropy deconvolution
Procedia PDF Downloads 3742954 Optimization of Hepatitis B Surface Antigen Purifications to Improving the Production of Hepatitis B Vaccines on Pichia pastoris
Authors: Rizky Kusuma Cahyani
Abstract:
Hepatitis B is a liver inflammatory disease caused by hepatitis B virus (HBV). This infection can be prevented by vaccination which contains HBV surface protein (sHBsAg). However, vaccine supply is limited. Several attempts have been conducted to produce local sHBsAg. However, the purity degree and protein yield are still inadequate. Therefore optimization of HBsAg purification steps is required to obtain high yield with better purification fold. In this study, optimization of purification was done in 2 steps, precipitation using variation of NaCl concentration (0,3 M; 0,5 M; 0,7 M) and PEG (3%, 5%, 7%); ion exchange chromatography (IEC) using NaCl 300-500 mM elution buffer concentration.To determine HBsAg protein, bicinchoninic acid assay (BCA) and enzyme-linked immunosorbent assay (ELISA) was used in this study. Visualization of HBsAg protein was done by SDS-PAGE analysis. Based on quantitative analysis, optimal condition at precipitation step was given 0,3 M NaCl and PEG 3%, while in ion exchange chromatography step, the optimum condition when protein eluted with NaCl 500 mM. Sodium dodecyl sulfate-polyacrylamide gel electrophoresis (SDS-PAGE) analysis indicates that the presence of protein HBsAg with a molecular weight of 25 kDa (monomer) and 50 kDa (dimer). The optimum condition for purification of sHBsAg produced in Pichia pastoris gave a yield of 47% and purification fold 17x so that it would increase the production of hepatitis B vaccine to be more optimal.Keywords: hepatitis B virus, HBsAg, hepatitis B surface antigen, Pichia pastoris, purification
Procedia PDF Downloads 1512953 Emerging Methods as a Tool for Obtaining Subconscious Feedback in E-Commerce and Marketplace
Authors: J. Berčík, A. Mravcová, A. Rusková, P. Jurčišin, R. Virágh
Abstract:
The online world is changing every day. With this comes the emergence and development of new business models. One of them is the sale of several types of products in one place. This type of sales in the form of online marketplaces has undergone a positive development in recent years and represents a kind of alternative to brick-and-mortar shopping centres. The main philosophy is to buy several products under one roof. Examples of popular e-commerce marketplaces are Amazon, eBay, and Allegro. Their share of total e-commerce turnover is expected to even double in the coming years. The paper highlights possibilities for testing web applications and online marketplace using emerging methods like stationary eye cameras (eye tracking) and facial analysis (FaceReading).Keywords: emerging methods, consumer neuroscience, e-commerce, marketplace, user experience, user interface
Procedia PDF Downloads 712952 Load-Enabled Deployment and Sensing Range Optimization for Lifetime Enhancement of WSNs
Authors: Krishan P. Sharma, T. P. Sharma
Abstract:
Wireless sensor nodes are resource constrained battery powered devices usually deployed in hostile and ill-disposed areas to cooperatively monitor physical or environmental conditions. Due to their limited power supply, the major challenge for researchers is to utilize their battery power for enhancing the lifetime of whole network. Communication and sensing are two major sources of energy consumption in sensor networks. In this paper, we propose a deployment strategy for enhancing the average lifetime of a sensor network by effectively utilizing communication and sensing energy to provide full coverage. The proposed scheme is based on the fact that due to heavy relaying load, sensor nodes near to the sink drain energy at much faster rate than other nodes in the network and consequently die much earlier. To cover this imbalance, proposed scheme finds optimal communication and sensing ranges according to effective load at each node and uses a non-uniform deployment strategy where there is a comparatively high density of nodes near to the sink. Probable relaying load factor at particular node is calculated and accordingly optimal communication distance and sensing range for each sensor node is adjusted. Thus, sensor nodes are placed at locations that optimize energy during network operation. Formal mathematical analysis for calculating optimized locations is reported in present work.Keywords: load factor, network lifetime, non-uniform deployment, sensing range
Procedia PDF Downloads 3832951 A Review of Ultralightweight Mutual Authentication Protocols
Authors: Umar Mujahid, Greatzel Unabia, Hongsik Choi, Binh Tran
Abstract:
Radio Frequency Identification (RFID) is one of the most commonly used technologies in IoTs and Wireless Sensor Networks which makes the devices identification and tracking extremely easy to manage. Since RFID uses wireless channel for communication, which is open for all types of adversaries, researchers have proposed many Ultralightweight Mutual Authentication Protocols (UMAPs) to ensure security and privacy in a cost-effective manner. These UMAPs involve simple bitwise logical operators such as XOR, AND, OR & Rot, etc., to design the protocol messages. However, most of these UMAPs were later reported to be vulnerable against many malicious attacks. In this paper, we have presented a detailed overview of some eminent UMAPs and also discussed the many security attacks on them. Finally, some recommendations and suggestions have been discussed, which can improve the design of the UMAPs.Keywords: RFID, Ultralightweight, UMAP, SASI
Procedia PDF Downloads 1532950 A Comparison of Proxemics and Postural Head Movements during Pop Music versus Matched Music Videos
Authors: Harry J. Witchel, James Ackah, Carlos P. Santos, Nachiappan Chockalingam, Carina E. I. Westling
Abstract:
Introduction: Proxemics is the study of how people perceive and use space. It is commonly proposed that when people like or engage with a person/object, they will move slightly closer to it, often quite subtly and subconsciously. Music videos are known to add entertainment value to a pop song. Our hypothesis was that by adding appropriately matched video to a pop song, it would lead to a net approach of the head to the monitor screen compared to simply listening to an audio-only version of the song. Methods: We presented to 27 participants (ages 21.00 ± 2.89, 15 female) seated in front of 47.5 x 27 cm monitor two musical stimuli in a counterbalanced order; all stimuli were based on music videos by the band OK Go: Here It Goes Again (HIGA, boredom ratings (0-100) = 15.00 ± 4.76, mean ± SEM, standard-error-of-the-mean) and Do What You Want (DWYW, boredom ratings = 23.93 ± 5.98), which did not differ in boredom elicited (P = 0.21, rank-sum test). Each participant experienced each song only once, and one song (counterbalanced) as audio-only versus the other song as a music video. The movement was measured by video-tracking using Kinovea 0.8, based on recording from a lateral aspect; before beginning, each participant had a reflective motion tracking marker placed on the outer canthus of the left eye. Analysis of the Kinovea X-Y coordinate output in comma-separated-variables format was performed in Matlab, as were non-parametric statistical tests. Results: We found that the audio-only stimuli (combined for both HIGA and DWYW, mean ± SEM, 35.71 ± 5.36) were significantly more boring than the music video versions (19.46 ± 3.83, P = 0.0066 Wilcoxon Signed Rank Test (WSRT), Cohen's d = 0.658, N = 28). We also found that participants' heads moved around twice as much during the audio-only versions (speed = 0.590 ± 0.095 mm/sec) compared to the video versions (0.301 ± 0.063 mm/sec, P = 0.00077, WSRT). However, the participants' mean head-to-screen distances were not detectably smaller (i.e. head closer to the screen) during the music videos (74.4 ± 1.8 cm) compared to the audio-only stimuli (73.9 ± 1.8 cm, P = 0.37, WSRT). If anything, during the audio-only condition, they were slightly closer. Interestingly, the ranges of the head-to-screen distances were smaller during the music video (8.6 ± 1.4 cm) compared to the audio-only (12.9 ± 1.7 cm, P = 0.0057, WSRT), the standard deviations were also smaller (P = 0.0027, WSRT), and their heads were held 7 mm higher (video 116.1 ± 0.8 vs. audio-only 116.8 ± 0.8 cm above floor, P = 0.049, WSRT). Discussion: As predicted, sitting and listening to experimenter-selected pop music was more boring than when the music was accompanied by a matched, professionally-made video. However, we did not find that the proxemics of the situation led to approaching the screen. Instead, adding video led to efforts to control the head to a more central and upright viewing position and to suppress head fidgeting.Keywords: boredom, engagement, music videos, posture, proxemics
Procedia PDF Downloads 1672949 Characterising the Performance Benefits of a 1/7-Scale Morphing Rotor Blade
Authors: Mars Burke, Alvin Gatto
Abstract:
Rotary-wing aircraft serve as indispensable components in the advancement of aviation, valued for their ability to operate in diverse and challenging environments without the need for conventional runways. This versatility makes them ideal for applications like environmental conservation, precision agriculture, emergency medical support, and rapid-response operations in rugged terrains. However, although highly maneuverable, rotary-wing platforms generally have lower aerodynamic efficiency than fixed-wing aircraft. This study takes the view of improving aerodynamic performance by examining a 1/7th scale rotor blade model with a NACA0012 airfoil using CROTOR software. The analysis focuses on optimal spanwise locations for separating morphing and fixed blade sections at 85%, 90%, and 95% of the blade radius (r/R) with up to +20 degrees of twist incorporated to the design.. Key performance metrics assessed include lift coefficient (CL), drag coefficient (CD), lift-to-drag ratio (CL / CD), Mach number, power, thrust coefficient, and Figure of Merit (FOM). Results indicate that the 0.90 r/R position is optimal for dividing the morphing and fixed sections, achieving a significant improvement of over 7% in both lift-to-drag ratio and FOM. These findings underscoring the substantial impact on overall performance of the rotor system and rotational aerodynamics that geometric modifications through the inclusion of a morphing capability can ultimately realise.Keywords: rotary morphing, rotational aerodynamics, rotorcraft morphing, rotor blade, twist morphing
Procedia PDF Downloads 122948 Optimal Pricing Based on Real Estate Demand Data
Authors: Vanessa Kummer, Maik Meusel
Abstract:
Real estate demand estimates are typically derived from transaction data. However, in regions with excess demand, transactions are driven by supply and therefore do not indicate what people are actually looking for. To estimate the demand for housing in Switzerland, search subscriptions from all important Swiss real estate platforms are used. These data do, however, suffer from missing information—for example, many users do not specify how many rooms they would like or what price they would be willing to pay. In economic analyses, it is often the case that only complete data is used. Usually, however, the proportion of complete data is rather small which leads to most information being neglected. Also, the data might have a strong distortion if it is complete. In addition, the reason that data is missing might itself also contain information, which is however ignored with that approach. An interesting issue is, therefore, if for economic analyses such as the one at hand, there is an added value by using the whole data set with the imputed missing values compared to using the usually small percentage of complete data (baseline). Also, it is interesting to see how different algorithms affect that result. The imputation of the missing data is done using unsupervised learning. Out of the numerous unsupervised learning approaches, the most common ones, such as clustering, principal component analysis, or neural networks techniques are applied. By training the model iteratively on the imputed data and, thereby, including the information of all data into the model, the distortion of the first training set—the complete data—vanishes. In a next step, the performances of the algorithms are measured. This is done by randomly creating missing values in subsets of the data, estimating those values with the relevant algorithms and several parameter combinations, and comparing the estimates to the actual data. After having found the optimal parameter set for each algorithm, the missing values are being imputed. Using the resulting data sets, the next step is to estimate the willingness to pay for real estate. This is done by fitting price distributions for real estate properties with certain characteristics, such as the region or the number of rooms. Based on these distributions, survival functions are computed to obtain the functional relationship between characteristics and selling probabilities. Comparing the survival functions shows that estimates which are based on imputed data sets do not differ significantly from each other; however, the demand estimate that is derived from the baseline data does. This indicates that the baseline data set does not include all available information and is therefore not representative for the entire sample. Also, demand estimates derived from the whole data set are much more accurate than the baseline estimation. Thus, in order to obtain optimal results, it is important to make use of all available data, even though it involves additional procedures such as data imputation.Keywords: demand estimate, missing-data imputation, real estate, unsupervised learning
Procedia PDF Downloads 2852947 A Systematic Review Investigating the Use of EEG Measures in Neuromarketing
Authors: A. M. Byrne, E. Bonfiglio, C. Rigby, N. Edelstyn
Abstract:
Introduction: Neuromarketing employs numerous methodologies when investigating products and advertisement effectiveness. Electroencephalography (EEG), a non-invasive measure of electrical activity from the brain, is commonly used in neuromarketing. EEG data can be considered using time-frequency (TF) analysis, where changes in the frequency of brainwaves are calculated to infer participant’s mental states, or event-related potential (ERP) analysis, where changes in amplitude are observed in direct response to a stimulus. This presentation discusses the findings of a systematic review of EEG measures in neuromarketing. A systematic review summarises evidence on a research question, using explicit measures to identify, select, and critically appraise relevant research papers. Thissystematic review identifies which EEG measures are the most robust predictor of customer preference and purchase intention. Methods: Search terms identified174 papers that used EEG in combination with marketing-related stimuli. Publications were excluded if they were written in a language other than English or were not published as journal articles (e.g., book chapters). The review investigated which TF effect (e.g., theta-band power) and ERP component (e.g., N400) most consistently reflected preference and purchase intention. Machine-learning prediction was also investigated, along with the use of EEG combined with physiological measures such as eye-tracking. Results: Frontal alpha asymmetry was the most reliable TF signal, where an increase in activity over the left side of the frontal lobe indexed a positive response to marketing stimuli, while an increase in activity over the right side indexed a negative response. The late positive potential, a positive amplitude increase around 600 ms after stimulus presentation, was the most reliable ERP component, reflecting the conscious emotional evaluation of marketing stimuli. However, each measure showed mixed results when related to preference and purchase behaviour. Predictive accuracy was greatly improved through machine-learning algorithms such as deep neural networks, especially when combined with eye-tracking or facial expression analyses. Discussion: This systematic review provides a novel catalogue of the most effective use of each EEG measure commonly used in neuromarketing. Exciting findings to emerge are the identification of the frontal alpha asymmetry and late positive potential as markers of preferential responses to marketing stimuli. Predictive accuracy using machine-learning algorithms achieved predictive accuracies as high as 97%, and future research should therefore focus on machine-learning prediction when using EEG measures in neuromarketing.Keywords: EEG, ERP, neuromarketing, machine-learning, systematic review, time-frequency
Procedia PDF Downloads 1112946 Optimal 3D Deployment and Path Planning of Multiple Uavs for Maximum Coverage and Autonomy
Authors: Indu Chandran, Shubham Sharma, Rohan Mehta, Vipin Kizheppatt
Abstract:
Unmanned aerial vehicles are increasingly being explored as the most promising solution to disaster monitoring, assessment, and recovery. Current relief operations heavily rely on intelligent robot swarms to capture the damage caused, provide timely rescue, and create road maps for the victims. To perform these time-critical missions, efficient path planning that ensures quick coverage of the area is vital. This study aims to develop a technically balanced approach to provide maximum coverage of the affected area in a minimum time using the optimal number of UAVs. A coverage trajectory is designed through area decomposition and task assignment. To perform efficient and autonomous coverage mission, solution to a TSP-based optimization problem using meta-heuristic approaches is designed to allocate waypoints to the UAVs of different flight capacities. The study exploits multi-agent simulations like PX4-SITL and QGroundcontrol through the ROS framework and visualizes the dynamics of UAV deployment to different search paths in a 3D Gazebo environment. Through detailed theoretical analysis and simulation tests, we illustrate the optimality and efficiency of the proposed methodologies.Keywords: area coverage, coverage path planning, heuristic algorithm, mission monitoring, optimization, task assignment, unmanned aerial vehicles
Procedia PDF Downloads 2152945 Design and Implementation of DC-DC Converter with Inc-Cond Algorithm
Authors: Mustafa Engin Başoğlu, Bekir Çakır
Abstract:
The most important component affecting the efficiency of photovoltaic power systems are solar panels. Efficiency of these systems are significantly affected because of being low efficiency of solar panel. Therefore, solar panels should be operated under maximum power point conditions through a power converter. In this study, design boost converter with maximum power point tracking (MPPT) operation has been designed and performed with Incremental Conductance (Inc-Cond) algorithm by using direct duty control. Furthermore, it is shown that performance of boost converter with MPPT operation fails under low load resistance connection.Keywords: boost converter, incremental conductance (Inc-Cond), MPPT, solar panel
Procedia PDF Downloads 10462944 A Study on the Effect of Design Factors of Slim Keyboard’s Tactile Feedback
Authors: Kai-Chieh Lin, Chih-Fu Wu, Hsiang Ling Hsu, Yung-Hsiang Tu, Chia-Chen Wu
Abstract:
With the rapid development of computer technology, the design of computers and keyboards moves towards a trend of slimness. The change of mobile input devices directly influences users’ behavior. Although multi-touch applications allow entering texts through a virtual keyboard, the performance, feedback, and comfortableness of the technology is inferior to traditional keyboard, and while manufacturers launch mobile touch keyboards and projection keyboards, the performance has not been satisfying. Therefore, this study discussed the design factors of slim pressure-sensitive keyboards. The factors were evaluated with an objective (accuracy and speed) and a subjective evaluation (operability, recognition, feedback, and difficulty) depending on the shape (circle, rectangle, and L-shaped), thickness (flat, 3mm, and 6mm), and force (35±10g, 60±10g, and 85±10g) of the keyboard. Moreover, MANOVA and Taguchi methods (regarding signal-to-noise ratios) were conducted to find the optimal level of each design factor. The research participants, by their typing speed (30 words/ minute), were divided in two groups. Considering the multitude of variables and levels, the experiments were implemented using the fractional factorial design. A representative model of the research samples were established for input task testing. The findings of this study showed that participants with low typing speed primarily relied on vision to recognize the keys, and those with high typing speed relied on tactile feedback that was affected by the thickness and force of the keys. In the objective and subjective evaluation, a combination of keyboard design factors that might result in higher performance and satisfaction was identified (L-shaped, 3mm, and 60±10g) as the optimal combination. The learning curve was analyzed to make a comparison with a traditional standard keyboard to investigate the influence of user experience on keyboard operation. The research results indicated the optimal combination provided input performance to inferior to a standard keyboard. The results could serve as a reference for the development of related products in industry and for applying comprehensively to touch devices and input interfaces which are interacted with people.Keywords: input performance, mobile device, slim keyboard, tactile feedback
Procedia PDF Downloads 2992943 Evaluation of the Power Generation Effect Obtained by Inserting a Piezoelectric Sheet in the Backlash Clearance of a Circular Arc Helical Gear
Authors: Barenten Suciu, Yuya Nakamoto
Abstract:
Power generation effect, obtained by inserting a piezo- electric sheet in the backlash clearance of a circular arc helical gear, is evaluated. Such type of screw gear is preferred since, in comparison with the involute tooth profile, the circular arc profile leads to reduced stress-concentration effects, and improved life of the piezoelectric film. Firstly, geometry of the circular arc helical gear, and properties of the piezoelectric sheet are presented. Then, description of the test-rig, consisted of a right-hand thread gear meshing with a left-hand thread gear, and the voltage measurement procedure are given. After creating the tridimensional (3D) model of the meshing gears in SolidWorks, they are 3D-printed in acrylonitrile butadiene styrene (ABS) resin. Variation of the generated voltage versus time, during a meshing cycle of the circular arc helical gear, is measured for various values of the center distance. Then, the change of the maximal, minimal, and peak-to-peak voltage versus the center distance is illustrated. Optimal center distance of the gear, to achieve voltage maximization, is found and its significance is discussed. Such results prove that the contact pressure of the meshing gears can be measured, and also, the electrical power can be generated by employing the proposed technique.Keywords: circular arc helical gear, contact problem, optimal center distance, piezoelectric sheet, power generation
Procedia PDF Downloads 1672942 Argos System: Improvements and Future of the Constellation
Authors: Sophie Baudel, Aline Duplaa, Jean Muller, Stephan Lauriol, Yann Bernard
Abstract:
Argos is the main satellite telemetry system used by the wildlife research community, since its creation in 1978, for animal tracking and scientific data collection all around the world, to analyze and understand animal migrations and behavior. The marine mammals' biology is one of the major disciplines which had benefited from Argos telemetry, and conversely, marine mammals biologists’ community has contributed a lot to the growth and development of Argos use cases. The Argos constellation with 6 satellites in orbit in 2017 (Argos 2 payload on NOAA 15, NOAA 18, Argos 3 payload on NOAA 19, SARAL, METOP A and METOP B) is being extended in the following years with Argos 3 payload on METOP C (launch in October 2018), and Argos 4 payloads on Oceansat 3 (launch in 2019), CDARS in December 2021 (to be confirmed), METOP SG B1 in December 2022, and METOP-SG-B2 in 2029. Argos 4 will allow more frequency bands (600 kHz for Argos4NG, instead of 110 kHz for Argos 3), new modulation dedicated to animal (sea turtle) tracking allowing very low transmission power transmitters (50 to 100mW), with very low data rates (124 bps), enhancement of high data rates (1200-4800 bps), and downlink performance, at the whole contribution to enhance the system capacity (50,000 active beacons per month instead of 20,000 today). In parallel of this ‘institutional Argos’ constellation, in the context of a miniaturization trend in the spatial industry in order to reduce the costs and multiply the satellites to serve more and more societal needs, the French Space Agency CNES, which designs the Argos payloads, is innovating and launching the Argos ANGELS project (Argos NEO Generic Economic Light Satellites). ANGELS will lead to a nanosatellite prototype with an Argos NEO instrument (30 cm x 30 cm x 20cm) that will be launched in 2019. In the meantime, the design of the renewal of the Argos constellation, called Argos For Next Generations (Argos4NG), is on track and will be operational in 2022. Based on Argos 4 and benefitting of the feedback from ANGELS project, this constellation will allow revisiting time of fewer than 20 minutes in average between two satellite passes, and will also bring more frequency bands to improve the overall capacity of the system. The presentation will then be an overview of the Argos system, present and future and new capacities coming with it. On top of that, use cases of two Argos hardware modules will be presented: the goniometer pathfinder allowing recovering Argos beacons at sea or on the ground in a 100 km radius horizon-free circle around the beacon location and the new Argos 4 chipset called ‘Artic’, already available and tested by several manufacturers.Keywords: Argos satellite telemetry, marine protected areas, oceanography, maritime services
Procedia PDF Downloads 1812941 Increasing Performance of Autopilot Guided Small Unmanned Helicopter
Authors: Tugrul Oktay, Mehmet Konar, Mustafa Soylak, Firat Sal, Murat Onay, Orhan Kizilkaya
Abstract:
In this paper, autonomous performance of a small manufactured unmanned helicopter is tried to be increased. For this purpose, a small unmanned helicopter is manufactured in Erciyes University, Faculty of Aeronautics and Astronautics. It is called as ZANKA-Heli-I. For performance maximization, autopilot parameters are determined via minimizing a cost function consisting of flight performance parameters such as settling time, rise time, overshoot during trajectory tracking. For this purpose, a stochastic optimization method named as simultaneous perturbation stochastic approximation is benefited. Using this approach, considerable autonomous performance increase (around %23) is obtained.Keywords: small helicopters, hierarchical control, stochastic optimization, autonomous performance maximization, autopilots
Procedia PDF Downloads 5822940 Hydrometallurgical Recovery of Cobalt, Nickel, Lithium, and Manganese from Spent Lithium-Ion Batteries
Authors: E. K. Hardwick, L. B. Siwela, J. G. Falconer, M. E. Mathibela, W. Rolfe
Abstract:
Lithium-ion battery (LiB) demand has increased with the advancement in technologies. The applications include electric vehicles, cell phones, laptops, and many more devices. Typical components of the cathodes include lithium, cobalt, nickel, and manganese. Recycling the spent LiBs is necessary to reduce the ecological footprint of their production and use and to have a secondary source of valuable metals. A hydrometallurgical method was investigated for the recovery of cobalt and nickel from LiB cathodes. The cathodes were leached using a chloride solution. Ion exchange was then used to recover the chloro-complexes of the metals. The aim of the research was to determine the efficiency of a chloride leach, as well as ion exchange operating capacities that can be achieved for LiB recycling, and to establish the optimal operating conditions (ideal pH, temperature, leachate and eluant, flowrate, and reagent concentrations) for the recovery of the cathode metals. It was found that the leaching of the cathodes could be hindered by the formation of refractory metal oxides of cathode components. A reducing agent was necessary to improve the leaching rate and efficiency. Leaching was achieved using various chloride-containing solutions. The chloro-complexes were absorbed by the ion exchange resin and eluted to produce concentrated cobalt, nickel, lithium, and manganese streams. Chromatographic separation of these elements was achieved. Further work is currently underway to determine the optimal operating conditions for the recovery by ion exchange.Keywords: cobalt, ion exchange, leachate formation, lithium-ion batteries, manganese, nickel
Procedia PDF Downloads 982939 Internet of Things Based Process Model for Smart Parking System
Authors: Amjaad Alsalamah, Liyakathunsia Syed
Abstract:
Transportation is an essential need for many people to go to their work, school, and home. In particular, the main common method inside many cities is to drive the car. Driving a car can be an easy job to reach the destination and load all stuff in a reasonable time. However, deciding to find a parking lot for a car can take a long time using the traditional system that can issue a paper ticket for each customer. The old system cannot guarantee a parking lot for all customers. Also, payment methods are not always available, and many customers struggled to find their car among a numerous number of cars. As a result, this research focuses on providing an online smart parking system in order to save time and budget. This system provides a flexible management system for both parking owner and customers by receiving all request via the online system and it gets an accurate result for all available parking and its location.Keywords: smart parking system, IoT, tracking system, process model, cost, time
Procedia PDF Downloads 3362938 Systems Approach on Thermal Analysis of an Automatic Transmission
Authors: Sinsze Koo, Benjin Luo, Matthew Henry
Abstract:
In order to increase the performance of an automatic transmission, the automatic transmission fluid is required to be warm up to an optimal operating temperature. In a conventional vehicle, cold starts result in friction loss occurring in the gear box and engine. The stop and go nature of city driving dramatically affect the warm-up of engine oil and automatic transmission fluid and delay the time frame needed to reach an optimal operating temperature. This temperature phenomenon impacts both engine and transmission performance but also increases fuel consumption and CO2 emission. The aim of this study is to develop know-how of the thermal behavior in order to identify thermal impacts and functional principles in automatic transmissions. Thermal behavior was studied using models and simulations, developed using GT-Suit, on a one-dimensional thermal and flow transport. A power train of a conventional vehicle was modeled in order to emphasis the thermal phenomena occurring in the various components and how they impact the automatic transmission performance. The simulation demonstrates the thermal model of a transmission fluid cooling system and its component parts in warm-up after a cold start. The result of these analyses will support the future designs of transmission systems and components in an attempt to obtain better fuel efficiency and transmission performance. Therefore, these thermal analyses could possibly identify ways that improve existing thermal management techniques with prioritization on fuel efficiency.Keywords: thermal management, automatic transmission, hybrid, and systematic approach
Procedia PDF Downloads 3772937 Genomic Prediction Reliability Using Haplotypes Defined by Different Methods
Authors: Sohyoung Won, Heebal Kim, Dajeong Lim
Abstract:
Genomic prediction is an effective way to measure the abilities of livestock for breeding based on genomic estimated breeding values, statistically predicted values from genotype data using best linear unbiased prediction (BLUP). Using haplotypes, clusters of linked single nucleotide polymorphisms (SNPs), as markers instead of individual SNPs can improve the reliability of genomic prediction since the probability of a quantitative trait loci to be in strong linkage disequilibrium (LD) with markers is higher. To efficiently use haplotypes in genomic prediction, finding optimal ways to define haplotypes is needed. In this study, 770K SNP chip data was collected from Hanwoo (Korean cattle) population consisted of 2506 cattle. Haplotypes were first defined in three different ways using 770K SNP chip data: haplotypes were defined based on 1) length of haplotypes (bp), 2) the number of SNPs, and 3) k-medoids clustering by LD. To compare the methods in parallel, haplotypes defined by all methods were set to have comparable sizes; in each method, haplotypes defined to have an average number of 5, 10, 20 or 50 SNPs were tested respectively. A modified GBLUP method using haplotype alleles as predictor variables was implemented for testing the prediction reliability of each haplotype set. Also, conventional genomic BLUP (GBLUP) method, which uses individual SNPs were tested to evaluate the performance of the haplotype sets on genomic prediction. Carcass weight was used as the phenotype for testing. As a result, using haplotypes defined by all three methods showed increased reliability compared to conventional GBLUP. There were not many differences in the reliability between different haplotype defining methods. The reliability of genomic prediction was highest when the average number of SNPs per haplotype was 20 in all three methods, implying that haplotypes including around 20 SNPs can be optimal to use as markers for genomic prediction. When the number of alleles generated by each haplotype defining methods was compared, clustering by LD generated the least number of alleles. Using haplotype alleles for genomic prediction showed better performance, suggesting improved accuracy in genomic selection. The number of predictor variables was decreased when the LD-based method was used while all three haplotype defining methods showed similar performances. This suggests that defining haplotypes based on LD can reduce computational costs and allows efficient prediction. Finding optimal ways to define haplotypes and using the haplotype alleles as markers can provide improved performance and efficiency in genomic prediction.Keywords: best linear unbiased predictor, genomic prediction, haplotype, linkage disequilibrium
Procedia PDF Downloads 141