Search results for: computer simulated person
838 An Approximate Formula for Calculating the Fundamental Mode Period of Vibration of Practical Building
Authors: Abdul Hakim Chikho
Abstract:
Most international codes allow the use of an equivalent lateral load method for designing practical buildings to withstand earthquake actions. This method requires calculating an approximation to the fundamental mode period of vibrations of these buildings. Several empirical equations have been suggested to calculate approximations to the fundamental periods of different types of structures. Most of these equations are knowing to provide an only crude approximation to the required fundamental periods and repeating the calculation utilizing a more accurate formula is usually required. In this paper, a new formula to calculate a satisfactory approximation of the fundamental period of a practical building is proposed. This formula takes into account the mass and the stiffness of the building therefore, it is more logical than the conventional empirical equations. In order to verify the accuracy of the proposed formula, several examples have been solved. In these examples, calculating the fundamental mode periods of several farmed buildings utilizing the proposed formula and the conventional empirical equations has been accomplished. Comparing the obtained results with those obtained from a dynamic computer has shown that the proposed formula provides a more accurate estimation of the fundamental periods of practical buildings. Since the proposed method is still simple to use and requires only a minimum computing effort, it is believed to be ideally suited for design purposes.Keywords: earthquake, fundamental mode period, design, building
Procedia PDF Downloads 284837 High-Resolution ECG Automated Analysis and Diagnosis
Authors: Ayad Dalloo, Sulaf Dalloo
Abstract:
Electrocardiogram (ECG) recording is prone to complications, on analysis by physicians, due to noise and artifacts, thus creating ambiguity leading to possible error of diagnosis. Such drawbacks may be overcome with the advent of high resolution Methods, such as Discrete Wavelet Analysis and Digital Signal Processing (DSP) techniques. This ECG signal analysis is implemented in three stages: ECG preprocessing, features extraction and classification with the aim of realizing high resolution ECG diagnosis and improved detection of abnormal conditions in the heart. The preprocessing stage involves removing spurious artifacts (noise), due to such factors as muscle contraction, motion, respiration, etc. ECG features are extracted by applying DSP and suggested sloping method techniques. These measured features represent peak amplitude values and intervals of P, Q, R, S, R’, and T waves on ECG, and other features such as ST elevation, QRS width, heart rate, electrical axis, QR and QT intervals. The classification is preformed using these extracted features and the criteria for cardiovascular diseases. The ECG diagnostic system is successfully applied to 12-lead ECG recordings for 12 cases. The system is provided with information to enable it diagnoses 15 different diseases. Physician’s and computer’s diagnoses are compared with 90% agreement, with respect to physician diagnosis, and the time taken for diagnosis is 2 seconds. All of these operations are programmed in Matlab environment.Keywords: ECG diagnostic system, QRS detection, ECG baseline removal, cardiovascular diseases
Procedia PDF Downloads 297836 Georgia Case: Tourism Expenses of International Visitors on the Basis of Growing Attractiveness
Authors: Nino Abesadze, Marine Mindorashvili, Nino Paresashvili
Abstract:
At present actual tourism indicators cannot be calculated in Georgia, making it impossible to perform their quantitative analysis. Therefore, the study conducted by us is highly important from a theoretical as well as practical standpoint. The main purpose of the article is to make complex statistical analysis of tourist expenses of foreign visitors and to calculate statistical attractiveness indices of the tourism potential of Georgia. During the research, the method involving random and proportional selection has been applied. Computer software SPSS was used to compute statistical data for corresponding analysis. Corresponding methodology of tourism statistics was implemented according to international standards. Important information was collected and grouped from major Georgian airports, and a representative population of foreign visitors and a rule of selection of respondents were determined. The results show a trend of growth in tourist numbers and the share of tourists from post-soviet countries are constantly increasing. The level of satisfaction with tourist facilities and quality of service has improved, but still we have a problem of disparity between the service quality and the prices. The design of tourist expenses of foreign visitors is diverse; competitiveness of tourist products of Georgian tourist companies is higher. Attractiveness of popular cities of Georgia has increased by 43%.Keywords: tourist, expenses, indexes, statistics, analysis
Procedia PDF Downloads 333835 Dynamic Distribution Calibration for Improved Few-Shot Image Classification
Authors: Majid Habib Khan, Jinwei Zhao, Xinhong Hei, Liu Jiedong, Rana Shahzad Noor, Muhammad Imran
Abstract:
Deep learning is increasingly employed in image classification, yet the scarcity and high cost of labeled data for training remain a challenge. Limited samples often lead to overfitting due to biased sample distribution. This paper introduces a dynamic distribution calibration method for few-shot learning. Initially, base and new class samples undergo normalization to mitigate disparate feature magnitudes. A pre-trained model then extracts feature vectors from both classes. The method dynamically selects distribution characteristics from base classes (both adjacent and remote) in the embedding space, using a threshold value approach for new class samples. Given the propensity of similar classes to share feature distributions like mean and variance, this research assumes a Gaussian distribution for feature vectors. Subsequently, distributional features of new class samples are calibrated using a corrected hyperparameter, derived from the distribution features of both adjacent and distant base classes. This calibration augments the new class sample set. The technique demonstrates significant improvements, with up to 4% accuracy gains in few-shot classification challenges, as evidenced by tests on miniImagenet and CUB datasets.Keywords: deep learning, computer vision, image classification, few-shot learning, threshold
Procedia PDF Downloads 66834 Effect of Spelling on Communicative Competence: A Case Study of Registry Staff of the University of Ibadan, Nigeria
Authors: Lukman Omobola Adisa
Abstract:
Spelling is rule bound in a written discourse. It, however, calls into question, when such convention is grossly contravened in a formal setting revered as citadel of learning, despite availability of computer spell-checker, human knowledge, and lexicon. The foregoing reveals the extent of decadence pervading education sector in Nigeria. It is on this premise that this study reviews the effect of spelling on communicative competence of the University of Ibadan Registry Staff. The theoretical framework basically evaluates diverse scholars’ views on communicative competence and how spelling influences the intended meaning of a word/ sentence as a result of undue infringement on grammatical (spelling) rule. Newsletter, bulletin, memo, and letter are four print materials purposively selected while the methodology adopted is content analysis. Similarly, five categories, though not limited to, through which spelling blunders are committed are considered: effect of spelling (omission, addition, and substitution); sound ( homophone); transposition (heading/body: content) and ambiguity (capitalisation, space, and acronym). Subsequently, the analyses, findings, and recommendations are equally looked into. Summarily, the study x-rays effective role(s) plays by spelling in enhancing communicative competence through appropriate usage of linguistic registers.Keywords: communicative competence, content analysis, effect of spelling, linguistics registers
Procedia PDF Downloads 218833 An In-Depth Definition of the 24 Levels of Consciousness and Its Relationship to Buddhism and Artificial Intelligence
Authors: James V. Luisi
Abstract:
Understanding consciousness requires a synthesis of ideas from multiple disciplines, including obvious ones like psychology, biology, evolution, neurology, and neuroscience, as well as less obvious ones like protozoology, botany, entomology, carcinology, herpetology, mammalogy, and computer sciences. Furthermore, to incorporate the necessary backdrop, it is best presented in a theme of Eastern philosophy, specifically leveraging the teachings of Buddhism for its relevance to early thought on consciousness. These ideas are presented as a multi-level framework that illustrates the various aspects of consciousness within a tapestry of foundational and dependent building blocks as to how living organisms evolved to understand elements of their reality sufficiently to survive, and in the case of Homo sapiens, eventually move beyond meeting the basic needs of survival, but to also achieve survival of the species beyond the eventual fate of our planet. This is not a complete system of thought, but just a framework of consciousness gathering some of the key elements regarding the evolution of consciousness and the advent of free will, and presenting them in a unique way that encourages readers to continue the dialog and thought process as an experience to enjoy long after reading the last page. Readers are encouraged to think for themselves about the issues raised herein and to question every facet presented, as much further exploration is needed. Needless to say, this subject will remain a rapidly evolving one for quite some time to come, and it is probably in the interests of everyone to at least consider attaining both an ability and willingness to participate in the dialog.Keywords: consciousness, sentience, intelligence, artificial intelligence, Buddhism
Procedia PDF Downloads 108832 Speeding Up Lenia: A Comparative Study Between Existing Implementations and CUDA C++ with OpenGL Interop
Authors: L. Diogo, A. Legrand, J. Nguyen-Cao, J. Rogeau, S. Bornhofen
Abstract:
Lenia is a system of cellular automata with continuous states, space and time, which surprises not only with the emergence of interesting life-like structures but also with its beauty. This paper reports ongoing research on a GPU implementation of Lenia using CUDA C++ and OpenGL Interoperability. We demonstrate how CUDA as a low-level GPU programming paradigm allows optimizing performance and memory usage of the Lenia algorithm. A comparative analysis through experimental runs with existing implementations shows that the CUDA implementation outperforms the others by one order of magnitude or more. Cellular automata hold significant interest due to their ability to model complex phenomena in systems with simple rules and structures. They allow exploring emergent behavior such as self-organization and adaptation, and find applications in various fields, including computer science, physics, biology, and sociology. Unlike classic cellular automata which rely on discrete cells and values, Lenia generalizes the concept of cellular automata to continuous space, time and states, thus providing additional fluidity and richness in emerging phenomena. In the current literature, there are many implementations of Lenia utilizing various programming languages and visualization libraries. However, each implementation also presents certain drawbacks, which serve as motivation for further research and development. In particular, speed is a critical factor when studying Lenia, for several reasons. Rapid simulation allows researchers to observe the emergence of patterns and behaviors in more configurations, on bigger grids and over longer periods without annoying waiting times. Thereby, they enable the exploration and discovery of new species within the Lenia ecosystem more efficiently. Moreover, faster simulations are beneficial when we include additional time-consuming algorithms such as computer vision or machine learning to evolve and optimize specific Lenia configurations. We developed a Lenia implementation for GPU using the C++ and CUDA programming languages, and CUDA/OpenGL Interoperability for immediate rendering. The goal of our experiment is to benchmark this implementation compared to the existing ones in terms of speed, memory usage, configurability and scalability. In our comparison we focus on the most important Lenia implementations, selected for their prominence, accessibility and widespread use in the scientific community. The implementations include MATLAB, JavaScript, ShaderToy GLSL, Jupyter, Rust and R. The list is not exhaustive but provides a broad view of the principal current approaches and their respective strengths and weaknesses. Our comparison primarily considers computational performance and memory efficiency, as these factors are critical for large-scale simulations, but we also investigate the ease of use and configurability. The experimental runs conducted so far demonstrate that the CUDA C++ implementation outperforms the other implementations by one order of magnitude or more. The benefits of using the GPU become apparent especially with larger grids and convolution kernels. However, our research is still ongoing. We are currently exploring the impact of several software design choices and optimization techniques, such as convolution with Fast Fourier Transforms (FFT), various GPU memory management scenarios, and the trade-off between speed and accuracy using single versus double precision floating point arithmetic. The results will give valuable insights into the practice of parallel programming of the Lenia algorithm, and all conclusions will be thoroughly presented in the conference paper. The final version of our CUDA C++ implementation will be published on github and made freely accessible to the Alife community for further development.Keywords: artificial life, cellular automaton, GPU optimization, Lenia, comparative analysis.
Procedia PDF Downloads 41831 Personas Help Understand Users’ Needs, Goals and Desires in an Online Institutional Repository
Authors: Maha ALjohani, James Blustein
Abstract:
Communicating users' needs, goals and problems help designers and developers overcome challenges faced by end users. Personas are used to represent end users’ needs. In our research, creating personas allowed the following questions to be answered: Who are the potential user groups? What do they want to achieve by using the service? What are the problems that users face? What should the service provide to them? To develop realistic personas, we conducted a focus group discussion with undergraduate and graduate students and also interviewed a university librarian. The personas were created to help evaluating the Institutional Repository that is based on the DSpace system. The profiles helped to communicate users' needs, abilities, tasks, and problems, and the task scenarios used in the heuristic evaluation were based on these personas. Four personas resulted of a focus group discussion with undergraduate and graduate students and from interviewing a university librarian. We then used these personas to create focused task-scenarios for a heuristic evaluation on the system interface to ensure that it met users' needs, goals, problems and desires. In this paper, we present the process that we used to create the personas that led to devise the task scenarios used in the heuristic evaluation as a follow up study of the DSpace university repository.Keywords: heuristic evaluation, institutional repositories, user experience, human computer interaction, user profiles, personas, task scenarios, heuristics
Procedia PDF Downloads 499830 Methods for Restricting Unwanted Access on the Networks Using Firewall
Authors: Bhagwant Singh, Sikander Singh Cheema
Abstract:
This paper examines firewall mechanisms routinely implemented for network security in depth. A firewall can't protect you against all the hazards of unauthorized networks. Consequently, many kinds of infrastructure are employed to establish a secure network. Firewall strategies have already been the subject of significant analysis. This study's primary purpose is to avoid unnecessary connections by combining the capability of the firewall with the use of additional firewall mechanisms, which include packet filtering and NAT, VPNs, and backdoor solutions. There are insufficient studies on firewall potential and combined approaches, but there aren't many. The research team's goal is to build a safe network by integrating firewall strength and firewall methods. The study's findings indicate that the recommended concept can form a reliable network. This study examines the characteristics of network security and the primary danger, synthesizes existing domestic and foreign firewall technologies, and discusses the theories, benefits, and disadvantages of different firewalls. Through synthesis and comparison of various techniques, as well as an in-depth examination of the primary factors that affect firewall effectiveness, this study investigated firewall technology's current application in computer network security, then introduced a new technique named "tight coupling firewall." Eventually, the article discusses the current state of firewall technology as well as the direction in which it is developing.Keywords: firewall strategies, firewall potential, packet filtering, NAT, VPN, proxy services, firewall techniques
Procedia PDF Downloads 101829 Recent Developments in Artificial Intelligence and Information Communications Technology
Authors: Dolapo Adeyemo
Abstract:
Technology can be designed specifically for geriatrics and persons with disabilities or ICT accessibility solutions. Both solutions stand to benefit from advances in Artificial intelligence, which are computer systems that perform tasks that require human intelligence. Tasks such as decision making, visual perception, speech recognition, and even language translation are useful in both situation and will provide significant benefits to people with temporarily or permanent disabilities. This research’s goal is to review innovations focused on the use of artificial intelligence that bridges the accessibility gap in technology from a user-centered perspective. A mixed method approach that utilized a comprehensive review of academic literature on the subject combined with semi structure interviews of users, developers, and technology product owners. The internet of things and artificial intelligence technology is creating new opportunities in the assistive technology space and proving accessibility to existing technology. Device now more adaptable to the needs of the user by learning the behavior of users as they interact with the internet. Accessibility to devices have witnessed significant enhancements that continue to benefit people with disabilities. Examples of other advances identified are prosthetic limbs like robotic arms supported by artificial intelligence, route planning software for the visually impaired, and decision support tools for people with disabilities and even clinicians that provide care.Keywords: ICT, IOT, accessibility solutions, universal design
Procedia PDF Downloads 87828 Virtual and Visual Reconstructions in Museum Expositions
Authors: Ekaterina Razuvalova, Konstantin Rudenko
Abstract:
In this article the most successful examples of international visual and virtual reconstructions of historical and culture objects, which are based on informative and communicative technologies, are represented. 3D reconstructions can demonstrate outward appearance, visualize different hypothesis, connected to represented object. Virtual reality can give us any daytime and season, any century and environment. We can see how different people from different countries and different era lived; we can get different information about any object; we can see historical complexes in real city environment, which are damaged or vanished. These innovations confirm the fact, that 3D reconstruction is important in museum development. Considering the most interesting examples of visual and virtual reconstructions, we can notice, that visual reconstruction is a 3D image of different objects, historical complexes, buildings and phenomena. They are constant and we can see them only as momentary objects. And virtual reconstruction is some environment with its own time, rules and phenomena. These reconstructions are continuous; seasons, daytime and natural conditions can change there. They can demonstrate abilities of virtual world existence. In conclusion: new technologies give us opportunities to expand the boundaries of museum space, improve abilities of museum expositions, create emotional atmosphere of game immersion, which can interest visitor. Usage of network sources allows increasing the number of visitors and virtual reconstruction opportunities show creative side of museum business.Keywords: computer technologies, historical reconstruction, museums, museum expositions, virtual reconstruction
Procedia PDF Downloads 329827 Environmental Fate and Toxicity of Aged Titanium Dioxide Nano-Composites Used in Sunscreen
Authors: Danielle Slomberg, Jerome Labille, Riccardo Catalano, Jean-Claude Hubaud, Alexandra Lopes, Alice Tagliati, Teresa Fernandes
Abstract:
In the assessment and management of cosmetics and personal care products, sunscreens are of emerging concern regarding both human and environmental health. Organic UV blockers in many sunscreens have been evidenced to undergo rapid photodegradation, induce dermal allergic reactions due to skin penetration, and to cause adverse effects on marine systems. While mineral UV-blockers may offer a safer alternative, their fate and impact and resulting regulation are still under consideration, largely related to the potential influence of nanotechnology-based products on both consumers and the environment. Nanometric titanium dioxide (TiO₂) UV-blockers have many advantages in terms of sun protection and asthetics (i.e., transparency). These UV-blockers typically consist of rutile nanoparticles coated with a primary mineral layer (silica or alumina) aimed at blocking the nanomaterial photoactivity and can include a secondary organic coating (e.g., stearic acid, methicone) aimed at favouring dispersion of the nanomaterial in the sunscreen formulation. The nanomaterials contained in the sunscreen can leave the skin either through a bathing of everyday usage, with subsequent release into rivers, lakes, seashores, and/or sewage treatment plants. The nanomaterial behaviour, fate and impact in these different systems is largely determined by its surface properties, (e.g. the nanomaterial coating type) and lifetime. The present work aims to develop the eco-design of sunscreens through the minimisation of risks associated with nanomaterials incorporated into the formulation. All stages of the sunscreen’s life cycle must be considered in this aspect, from its manufacture to its end-of-life, through its use by the consumer to its impact on the exposed environment. Reducing the potential release and/or toxicity of the nanomaterial from the sunscreen is a decisive criterion for its eco-design. TiO₂ UV-blockers of varied size and surface coating (e.g., stearic acid and silica) have been selected for this study. Hydrophobic TiO₂ UV-blockers (i.e., stearic acid-coated) were incorporated into a typical water-in-oil (w/o) formulation while hydrophilic, silica-coated TiO₂ UV-blockers were dispersed into an oil-in-water (o/w) formulation. The resulting sunscreens were characterised in terms of nanomaterial localisation, sun protection factor, and photo-passivation. The risk to the direct aquatic environment was assessed by evaluating the release of nanomaterials from the sunscreen through a simulated laboratory aging procedure. The size distribution, surface charge, and degradation state of the nano-composite by-products, as well as their nanomaterial concentration and colloidal behaviour were determined in a variety of aqueous environments (e.g., seawater and freshwater). Release of the hydrophobic nanocomposites into the aqueous environment was driven by oil droplet formation while hydrophilic nano-composites were readily dispersed. Ecotoxicity of the sunscreen by-products (from both w/o and o/w formulations) and their risk to marine organisms were assessed using coral symbiotes and tropical corals, evaluating both lethal and sublethal toxicities. The data dissemination and provided risk knowledge from the present work will help guide regulation related to nanomaterials in sunscreen, provide better information for consumers, and allow for easier decision-making for manufacturers.Keywords: alteration, environmental fate, sunscreens, titanium dioxide nanoparticles
Procedia PDF Downloads 262826 Finite Volume Method Simulations of GaN Growth Process in MOVPE Reactor
Authors: J. Skibinski, P. Caban, T. Wejrzanowski, K. J. Kurzydlowski
Abstract:
In the present study, numerical simulations of heat and mass transfer during gallium nitride growth process in Metal Organic Vapor Phase Epitaxy reactor AIX-200/4RF-S is addressed. Existing knowledge about phenomena occurring in the MOVPE process allows to produce high quality nitride based semiconductors. However, process parameters of MOVPE reactors can vary in certain ranges. Main goal of this study is optimization of the process and improvement of the quality of obtained crystal. In order to investigate this subject a series of computer simulations have been performed. Numerical simulations of heat and mass transfer in GaN epitaxial growth process have been performed to determine growth rate for various mass flow rates and pressures of reagents. According to the fact that it’s impossible to determine experimentally the exact distribution of heat and mass transfer inside the reactor during the process, modeling is the only solution to understand the process precisely. Main heat transfer mechanisms during MOVPE process are convection and radiation. Correlation of modeling results with the experiment allows to determine optimal process parameters for obtaining crystals of highest quality.Keywords: Finite Volume Method, semiconductors, epitaxial growth, metalorganic vapor phase epitaxy, gallium nitride
Procedia PDF Downloads 399825 Effects of Delphinidin on Lipid Metabolism in HepG2 Cells and Diet-Induced Obese Mice
Authors: Marcela Parra-Vargas, Ana Sandoval-Rodriguez, Roberto Rodriguez-Echevarria, Jose Dominguez-Rosales, Juan Armendariz-Borunda
Abstract:
Non-alcoholic fatty liver disease (NAFLD) is characterized by an excess of hepatic lipids, and it is to author’s best knowledge, the most prevalent chronic liver disorder. Anthocyanin-rich food consumption is linked to health benefits in metabolic disorders associated with obesity and NAFLD, although the precise functional role of anthocyanidin delphinidin (Dp) has yet to be established. The aim of this study was to investigate the effect of the Dp in NAFLD metabolic alterations by evaluating prevention or amelioration of hepatic lipid accumulation, as well as molecular mechanisms in two experimental obesity-related models of NALFD. In vitro: HepG2 cells were incubated with sodium palmitate (PA, 1 mM) to induce lipotoxic damage, and concomitantly treated with Dp (180 uM) for 24 h. Subsequently, total lipid accumulation was measured by colorimetric staining with Oil Red O, and total intrahepatic triglycerides were determined by an enzymatic assay. To assess molecular mechanisms, cells were pre-treated with PA for 24 h and then exposed to Dp for 1 h. In vivo: four-week-old male C57BL/6Nhsd mice were allocated in two main groups. Mice were fed with standard diet (control) or high-fat and high-carbohydrate diet (45% fat, HFD) for 16 wk to induce NAFLD. Then HFD was divided into subgroups: one treated orally with Dp (15 mg/kg bw, HFD-Dp) every day for 4 wk, while HFD group treated with vehicle (DMSO). Weight and fasting glucose were recorded weekly, while dietary ingestion was measured daily. Insulin tolerance test was performed at the end of treatment. Liver histology was evaluated with H&E and Masson’s trichrome stain. RT-PCR was used to evaluate gene expression and Western Blot to determine levels of protein in both experimental models. Parametric data were analyzed with one-way ANOVA and Tukey’s post-hoc test. Kruskal-Wallis and Mann-Whitney U test for non-parametric data, and P < 0.5 were considered significant. Dp prevented hepatic lipid accumulation by PA in HepG2 hepatocytes. Furthermore, Dp down-regulated gene expression of SREBP1c, FAS, and CPT1a without modifying AMPK phosphorylation levels. In vivo, Dp oral administration did not ameliorate lipid metabolic alterations raised by HFD. Adiposity, dietary ingestion, fasting glucose, and insulin sensitivity after Dp treatment remained similar to HFD group. Histological analysis showed hepatic damage in HFD groups and no differences between HFD and HFD-Dp groups were found. Hepatic gene expression of ACC and FAS were not altered by HFD. SREBP1c was similar in both HFD and HFD-Dp groups. No significant changes were observed in SREBP1c, ACC, and FAS adipose tissue gene expression by HFD or Dp treatment. Additionally, immunoblotting analysis revealed no changes in pathway SIRT1-LKB-AMPK and PPAR alpha by both HFD groups compared to control. In conclusion, the antioxidant Dp may provoke beneficial effects in the prevention of hepatic lipid accumulation. Nevertheless, the oral dose administrated in mice that simulated the total intake of anthocyanins consumed daily by humans has no effect as a treatment on hepatic lipid metabolic alterations and histological abnormalities associated with exposure to chronic HFD. A healthy lifestyle with regular intake of antioxidants such as anthocyanins may prevent metabolic alterations in NAFLD.Keywords: anthocyanins, antioxidants, delphinidin, non-alcoholic fatty liver disease, obesity
Procedia PDF Downloads 202824 The Characteristics of Quantity Operation for 2nd and 3rd Grade Mathematics Slow Learners
Authors: Pi-Hsia Hung
Abstract:
The development of mathematical competency has individual benefits as well as benefits to the wider society. Children who begin school behind their peers in their understanding of number, counting, and simple arithmetic are at high risk of staying behind throughout their schooling. The development of effective strategies for improving the educational trajectory of these individuals will be contingent on identifying areas of early quantitative knowledge that influence later mathematics achievement. A computer-based quantity assessment was developed in this study to investigate the characteristics of 2nd and 3rd grade slow learners in quantity. The concept of quantification involves understanding measurements, counts, magnitudes, units, indicators, relative size, and numerical trends and patterns. Fifty-five tasks of quantitative reasoning—such as number sense, mental calculation, estimation and assessment of reasonableness of results—are included as quantity problem solving. Thus, quantity is defined in this study as applying knowledge of number and number operations in a wide variety of authentic settings. Around 1000 students were tested and categorized into 4 different performance levels. Students’ quantity ability correlated higher with their school math grade than other subjects. Around 20% students are below basic level. The intervention design implications of the preliminary item map constructed are discussed.Keywords: mathematics assessment, mathematical cognition, quantity, number sense, validity
Procedia PDF Downloads 247823 A Generic Middleware to Instantly Sync Intensive Writes of Heterogeneous Massive Data via Internet
Authors: Haitao Yang, Zhenjiang Ruan, Fei Xu, Lanting Xia
Abstract:
Industry data centers often need to sync data changes reliably and instantly from a large-scale of heterogeneous autonomous relational databases accessed via the not-so-reliable Internet, for which a practical universal sync middle of low maintenance and operation costs is most wanted, but developing such a product and adapting it for various scenarios are a very sophisticated and continuous practice. The authors have been devising, applying, and optimizing a generic sync middleware system, named GSMS since 2006, holding the principles or advantages that the middleware must be SyncML-compliant and transparent to data application layer logic, need not refer to implementation details of databases synced, does not rely on host computer operating systems deployed, and its construction is light weighted and hence, of low cost. A series of ultimate experiments with GSMS sync performance were conducted for a persuasive example of a source relational database that underwent a broad range of write loads, say, from one thousand to one million intensive writes within a few minutes. The tests proved that GSMS has achieved an instant sync level of well below a fraction of millisecond per record sync, and GSMS’ smooth performances under ultimate write loads also showed it is feasible and competent.Keywords: heterogeneous massive data, instantly sync intensive writes, Internet generic middleware design, optimization
Procedia PDF Downloads 121822 3D Liver Segmentation from CT Images Using a Level Set Method Based on a Shape and Intensity Distribution Prior
Authors: Nuseiba M. Altarawneh, Suhuai Luo, Brian Regan, Guijin Tang
Abstract:
Liver segmentation from medical images poses more challenges than analogous segmentations of other organs. This contribution introduces a liver segmentation method from a series of computer tomography images. Overall, we present a novel method for segmenting liver by coupling density matching with shape priors. Density matching signifies a tracking method which operates via maximizing the Bhattacharyya similarity measure between the photometric distribution from an estimated image region and a model photometric distribution. Density matching controls the direction of the evolution process and slows down the evolving contour in regions with weak edges. The shape prior improves the robustness of density matching and discourages the evolving contour from exceeding liver’s boundaries at regions with weak boundaries. The model is implemented using a modified distance regularized level set (DRLS) model. The experimental results show that the method achieves a satisfactory result. By comparing with the original DRLS model, it is evident that the proposed model herein is more effective in addressing the over segmentation problem. Finally, we gauge our performance of our model against matrices comprising of accuracy, sensitivity and specificity.Keywords: Bhattacharyya distance, distance regularized level set (DRLS) model, liver segmentation, level set method
Procedia PDF Downloads 313821 Simulation, Design, and 3D Print of Novel Highly Integrated TEG Device with Improved Thermal Energy Harvest Efficiency
Abstract:
Despite the remarkable advancement of solar cell technology, the challenge of optimizing total solar energy harvest efficiency persists, primarily due to significant heat loss. This excess heat not only diminishes solar panel output efficiency but also curtails its operational lifespan. A promising approach to address this issue is the conversion of surplus heat into electricity. In recent years, there is growing interest in the use of thermoelectric generators (TEG) as a potential solution. The integration of efficient TEG devices holds the promise of augmenting overall energy harvest efficiency while prolonging the longevity of solar panels. While certain research groups have proposed the integration of solar cells and TEG devices, a substantial gap between conceptualization and practical implementation remains, largely attributed to low thermal energy conversion efficiency of TEG devices. To bridge this gap and meet the requisites of practical application, a feasible strategy involves the incorporation of a substantial number of p-n junctions within a confined unit volume. However, the manufacturing of high-density TEG p-n junctions presents a formidable challenge. The prevalent solution often leads to large device sizes to accommodate enough p-n junctions, consequently complicating integration with solar cells. Recently, the adoption of 3D printing technology has emerged as a promising solution to address this challenge by fabricating high-density p-n arrays. Despite this, further developmental efforts are necessary. Presently, the primary focus is on the 3D printing of vertically layered TEG devices, wherein p-n junction density remains constrained by spatial limitations and the constraints of 3D printing techniques. This study proposes a novel device configuration featuring horizontally arrayed p-n junctions of Bi2Te3. The structural design of the device is subjected to simulation through the Finite Element Method (FEM) within COMSOL Multiphysics software. Various device configurations are simulated to identify optimal device structure. Based on the simulation results, a new TEG device is fabricated utilizing 3D Selective laser melting (SLM) printing technology. Fusion 360 facilitates the translation of the COMSOL device structure into a 3D print file. The horizontal design offers a unique advantage, enabling the fabrication of densely packed, three-dimensional p-n junction arrays. The fabrication process entails printing a singular row of horizontal p-n junctions using the 3D SLM printing technique in a single layer. Subsequently, successive rows of p-n junction arrays are printed within the same layer, interconnected by thermally conductive copper. This sequence is replicated across multiple layers, separated by thermal insulating glass. This integration created in a highly compact three-dimensional TEG device with high density p-n junctions. The fabricated TEG device is then attached to the bottom of the solar cell using thermal glue. The whole device is characterized, with output data closely matching with COMSOL simulation results. Future research endeavors will encompass the refinement of thermoelectric materials. This includes the advancement of high-resolution 3D printing techniques tailored to diverse thermoelectric materials, along with the optimization of material microstructures such as porosity and doping. The objective is to achieve an optimal and highly integrated PV-TEG device that can substantially increase the solar energy harvest efficiency.Keywords: thermoelectric, finite element method, 3d print, energy conversion
Procedia PDF Downloads 62820 Synthetic Method of Contextual Knowledge Extraction
Authors: Olga Kononova, Sergey Lyapin
Abstract:
Global information society requirements are transparency and reliability of data, as well as ability to manage information resources independently; particularly to search, to analyze, to evaluate information, thereby obtaining new expertise. Moreover, it is satisfying the society information needs that increases the efficiency of the enterprise management and public administration. The study of structurally organized thematic and semantic contexts of different types, automatically extracted from unstructured data, is one of the important tasks for the application of information technologies in education, science, culture, governance and business. The objectives of this study are the contextual knowledge typologization, selection or creation of effective tools for extracting and analyzing contextual knowledge. Explication of various kinds and forms of the contextual knowledge involves the development and use full-text search information systems. For the implementation purposes, the authors use an e-library 'Humanitariana' services such as the contextual search, different types of queries (paragraph-oriented query, frequency-ranked query), automatic extraction of knowledge from the scientific texts. The multifunctional e-library «Humanitariana» is realized in the Internet-architecture in WWS-configuration (Web-browser / Web-server / SQL-server). Advantage of use 'Humanitariana' is in the possibility of combining the resources of several organizations. Scholars and research groups may work in a local network mode and in distributed IT environments with ability to appeal to resources of any participating organizations servers. Paper discusses some specific cases of the contextual knowledge explication with the use of the e-library services and focuses on possibilities of new types of the contextual knowledge. Experimental research base are science texts about 'e-government' and 'computer games'. An analysis of the subject-themed texts trends allowed to propose the content analysis methodology, that combines a full-text search with automatic construction of 'terminogramma' and expert analysis of the selected contexts. 'Terminogramma' is made out as a table that contains a column with a frequency-ranked list of words (nouns), as well as columns with an indication of the absolute frequency (number) and the relative frequency of occurrence of the word (in %% ppm). The analysis of 'e-government' materials showed, that the state takes a dominant position in the processes of the electronic interaction between the authorities and society in modern Russia. The media credited the main role in these processes to the government, which provided public services through specialized portals. Factor analysis revealed two factors statistically describing the used terms: human interaction (the user) and the state (government, processes organizer); interaction management (public officer, processes performer) and technology (infrastructure). Isolation of these factors will lead to changes in the model of electronic interaction between government and society. In this study, the dominant social problems and the prevalence of different categories of subjects of computer gaming in science papers from 2005 to 2015 were identified. Therefore, there is an evident identification of several types of contextual knowledge: micro context; macro context; dynamic context; thematic collection of queries (interactive contextual knowledge expanding a composition of e-library information resources); multimodal context (functional integration of iconographic and full-text resources through hybrid quasi-semantic algorithm of search). Further studies can be pursued both in terms of expanding the resource base on which they are held, and in terms of the development of appropriate tools.Keywords: contextual knowledge, contextual search, e-library services, frequency-ranked query, paragraph-oriented query, technologies of the contextual knowledge extraction
Procedia PDF Downloads 359819 Descriptive Study of Adverse Drug Reactions in a Paediatric Hospital in Mongolia from 2015 to 2019
Authors: Khaliun Nyambayar, Nomindari Azzaya, Batkhuyag Purevjav
Abstract:
Pharmacovigilance was officially introduced in Mongolia in 2003, in accordance with the Health Minister Order 183 for the registry of adverse drug reactions (ADR), approved in 2006 and was reviewed in 2010. This study was designed to evaluate the incidence and common types of adverse drug reactions among hospitalized children, the frequency of adverse drug reaction reported by health care providers, and the follow-up processes resulting from adverse drug reactions. A retrospective study of paediatric patients who experienced an adverse drug reaction from 2015 to 2019, extracted from the “yellow” card at the State Research Center for Maternal and Child Health, (city). A total of 417 adverse drug reactions were reported with an overall incidence was 80 (21.5%). Adverse reactions resulting from the use of antibiotics (particularly gentamycin, cephalosporins, and vancomycin) were usually mild. ADR’s were reported by physicians and nurses (93.8%), pharmacists (6.25%). Although documentation of physician notification occurred for 93% of adverse drug reactions, only 29% of cases were documented in the patient's medical chart, 13% included follow-up education for individuals involved, and 10% were updated in the allergy profile of the hospital computer system. Measures to improve the detection and reporting of adverse drug reactions by all health care professionals should be improved, to enhance our understanding of the nature and impact of these reactions in children.Keywords: adverse drug reaction, pediatric, yellow card, Mongolia
Procedia PDF Downloads 113818 The Effect of Gamma rays on Physicochemical Properties of Carboxymethyl Starch
Authors: N. Rajeswara Rao, T. Venkatappa Rao, K. Sowri Babu, N. Srinivas Rao, P. S. V. Shanmukhi
Abstract:
Carboxymethyl Starch (CMS) is a biopolymer derived from starch by the substitution method. CMS is proclaimed to have improved physicochemical properties than native starch. The present work deals with the effect of gamma radiation on the physicochemical properties of CMS. The samples were exposed to gamma irradiation of doses 30, 60 and 90 kGy. The resultant properties were studied with electron spin resonance (ESR), fourier transform infrared spectrometer (FTIR), differential scanning calorimeter (DSC), X-ray diffractometer (XRD) and scanning electron microscopy. Irradiation of CMS by gamma rays initiates cleavage of glucosidic bonds producing different types of radicals. Some of these radicals convert to peroxy radicals by abstracting oxygen. The ESR spectrum of CMS is anisotropic and is thought to be due to the superposition of various component spectra. In order to analyze the ESR spectrum, computer simulations were also employed. ESR spectra are also recorded under different conditions like post-irradiation times, variable temperatures and saturation behavior in order to evaluate the stability of free radicals produced on irradiation. Thermal studies from DSC depict that for CMS the gelatization process was absconded at higher doses. Relative crystallinity was reduced significantly after irradiation from XRD Studies. FTIR studies also confirm the same aspect. From ESR studies, it was concluded that irradiated CMS could be a potential reference material in ESR dosimetry.Keywords: gamma rays, free radicals, ESR simulations, gelatization
Procedia PDF Downloads 103817 3D Printing Perceptual Models of Preference Using a Fuzzy Extreme Learning Machine Approach
Authors: Xinyi Le
Abstract:
In this paper, 3D printing orientations were determined through our perceptual model. Some FDM (Fused Deposition Modeling) 3D printers, which are widely used in universities and industries, often require support structures during the additive manufacturing. After removing the residual material, some surface artifacts remain at the contact points. These artifacts will damage the function and visual effect of the model. To prevent the impact of these artifacts, we present a fuzzy extreme learning machine approach to find printing directions that avoid placing supports in perceptually significant regions. The proposed approach is able to solve the evaluation problem by combing both the subjective knowledge and objective information. Our method combines the advantages of fuzzy theory, auto-encoders, and extreme learning machine. Fuzzy set theory is applied for dealing with subjective preference information, and auto-encoder step is used to extract good features without supervised labels before extreme learning machine. An extreme learning machine method is then developed successfully for training and learning perceptual models. The performance of this perceptual model will be demonstrated on both natural and man-made objects. It is a good human-computer interaction practice which draws from supporting knowledge on both the machine side and the human side.Keywords: 3d printing, perceptual model, fuzzy evaluation, data-driven approach
Procedia PDF Downloads 438816 The Malfatti’s Problem in Reuleaux Triangle
Authors: Ching-Shoei Chiang
Abstract:
The Malfatti’s Problem is to ask for fitting 3 circles into a right triangle such that they are tangent to each other, and each circle is also tangent to a pair of the triangle’s side. This problem has been extended to any triangle (called general Malfatti’s Problem). Furthermore, the problem has been extended to have 1+2+…+n circles, we call it extended general Malfatti’s problem, these circles whose tangency graph, using the center of circles as vertices and the edge connect two circles center if these two circles tangent to each other, has the structure as Pascal’s triangle, and the exterior circles of these circles tangent to three sides of the triangle. In the extended general Malfatti’s problem, there are closed-form solutions for n=1, 2, and the problem becomes complex when n is greater than 2. In solving extended general Malfatti’s problem (n>2), we initially give values to the radii of all circles. From the tangency graph and current radii, we can compute angle value between two vectors. These vectors are from the center of the circle to the tangency points with surrounding elements, and these surrounding elements can be the boundary of the triangle or other circles. For each circle C, there are vectors from its center c to its tangency point with its neighbors (count clockwise) pi, i=0, 1,2,..,n. We add all angles between cpi to cp(i+1) mod (n+1), i=0,1,..,n, call it sumangle(C) for circle C. Using sumangle(C), we can reduce/enlarge the radii for all circles in next iteration, until sumangle(C) is equal to 2πfor all circles. With a similar idea, this paper proposed an algorithm to find the radii of circles whose tangency has the structure of Pascal’s triangle, and the exterior circles of these circles are tangent to the unit Realeaux Triangle.Keywords: Malfatti’s problem, geometric constraint solver, computer-aided geometric design, circle packing, data visualization
Procedia PDF Downloads 132815 Multi-Layer Multi-Feature Background Subtraction Using Codebook Model Framework
Authors: Yun-Tao Zhang, Jong-Yeop Bae, Whoi-Yul Kim
Abstract:
Background modeling and subtraction in video analysis has been widely proved to be an effective method for moving objects detection in many computer vision applications. Over the past years, a large number of approaches have been developed to tackle different types of challenges in this field. However, the dynamic background and illumination variations are two of the most frequently occurring issues in the practical situation. This paper presents a new two-layer model based on codebook algorithm incorporated with local binary pattern (LBP) texture measure, targeted for handling dynamic background and illumination variation problems. More specifically, the first layer is designed by block-based codebook combining with LBP histogram and mean values of RGB color channels. Because of the invariance of the LBP features with respect to monotonic gray-scale changes, this layer can produce block-wise detection results with considerable tolerance of illumination variations. The pixel-based codebook is employed to reinforce the precision from the outputs of the first layer which is to eliminate false positives further. As a result, the proposed approach can greatly promote the accuracy under the circumstances of dynamic background and illumination changes. Experimental results on several popular background subtraction datasets demonstrate a very competitive performance compared to previous models.Keywords: background subtraction, codebook model, local binary pattern, dynamic background, illumination change
Procedia PDF Downloads 217814 A Case Study in Using the Can-Sized Satellite Platforms for Interdisciplinary Problem-Based Learning in Aeronautical and Electronic Engineering
Authors: Michael Johnson, Vincenzo Oliveri
Abstract:
This work considers an interdisciplinary Problem-Based Learning (PBL) project developed by lecturers from the Aeronautical and Electronic and Computer Engineering departments at the University of Limerick. This “CANSAT” project utilises the CanSat can-sized satellite platform in order to allow students from aeronautical and electronic engineering to engage in a mixed format (online/face-to-face), interdisciplinary PBL assignment using a real-world platform and application. The project introduces students to the design, development, and construction of the CanSat system over the course of a single semester, enabling student(s) to apply their aeronautical and technical skills/capabilities to the realisation of a working CanSat system. In this case study, the CanSat kits are used to pivot the real-world, discipline-relevant PBL goal of designing, building, and testing the CanSat system with payload(s) from a traditional module-based setting to an online PBL setting. Feedback, impressions, benefits, and challenges identified through the semester are presented. Students found the project to be interesting and rewarding, with the interdisciplinary nature of the project appealing to them. Challenges and difficulties encountered are also addressed, with solutions developed between the students and facilitators to overcoming these discussed.Keywords: problem-based learning, interdisciplinary, engineering, CanSATs
Procedia PDF Downloads 129813 Local Buckling of Web-Core and Foam-Core Sandwich Panels
Authors: Ali N. Suri, Ahmad A. Al-Makhlufi
Abstract:
Sandwich construction is widely accepted as a method of construction especially in the aircraft industry. It is a type of stressed skin construction formed by bonding two thin faces to a thick core, the faces resist all of the applied edge loads and provide all or nearly all of the required rigidities, the core spaces the faces to increase cross section moment of inertia about common neutral axis and transmit shear between them provides a perfect bond between core and faces is made. Material for face sheets can be of metal or reinforced plastics laminates, core material can be metallic cores of thin sheets forming corrugation or honeycomb, or non-metallic core of Balsa wood, plastic foams, or honeycomb made of reinforced plastics. For in plane axial loading web core and web-foam core Sandwich panels can fail by local buckling of plates forming the cross section with buckling wave length of the order of length of spacing between webs. In this study local buckling of web core and web-foam core Sandwich panels is carried out for given materials of facing and core, and given panel overall dimension for different combinations of cross section geometries. The Finite Strip Method is used for the analysis, and Fortran based computer program is developed and used.Keywords: local buckling, finite strip, sandwich panels, web and foam core
Procedia PDF Downloads 351812 Autonomous Kuka Youbot Navigation Based on Machine Learning and Path Planning
Authors: Carlos Gordon, Patricio Encalada, Henry Lema, Diego Leon, Dennis Chicaiza
Abstract:
The following work presents a proposal of autonomous navigation of mobile robots implemented in an omnidirectional robot Kuka Youbot. We have been able to perform the integration of robotic operative system (ROS) and machine learning algorithms. ROS mainly provides two distributions; ROS hydro and ROS Kinect. ROS hydro allows managing the nodes of odometry, kinematics, and path planning with statistical and probabilistic, global and local algorithms based on Adaptive Monte Carlo Localization (AMCL) and Dijkstra. Meanwhile, ROS Kinect is responsible for the detection block of dynamic objects which can be in the points of the planned trajectory obstructing the path of Kuka Youbot. The detection is managed by artificial vision module under a trained neural network based on the single shot multibox detector system (SSD), where the main dynamic objects for detection are human beings and domestic animals among other objects. When the objects are detected, the system modifies the trajectory or wait for the decision of the dynamic obstacle. Finally, the obstacles are skipped from the planned trajectory, and the Kuka Youbot can reach its goal thanks to the machine learning algorithms.Keywords: autonomous navigation, machine learning, path planning, robotic operative system, open source computer vision library
Procedia PDF Downloads 177811 Topics of Blockchain Technology to Teach at Community College
Authors: Penn P. Wu, Jeannie Jo
Abstract:
Blockchain technology has rapidly gained popularity in industry. This paper attempts to assist academia to answer four questions. First, should community colleges begin offering education to nurture blockchain-literate students for the job market? Second, what are the appropriate topical areas to cover? Third, should it be an individual course? And forth, should it be a technical or management course? This paper starts with identifying the knowledge domains of blockchain technology and the topical areas each domain has, and continues with placing them in appropriate academic territories (Computer Sciences vs. Business) and subjects (programming, management, marketing, and laws), and then develops an evaluation model to determine the appropriate topical area for community colleges to teach. The evaluation is based on seven factors: maturity of technology, impacts on management, real-world applications, subject classification, knowledge prerequisites, textbook readiness, and recommended pedagogies. The evaluation results point to an interesting direction that offering an introductory course is an ideal option to guide students through the learning journey of what blockchain is and how it applies to business. Such an introductory course does not need to engage students in the discussions of mathematics and sciences that make blockchain technologies possible. While it is inevitable to brief technical topics to help students build a solid knowledge foundation of blockchain technologies, community colleges should avoid offering students a course centered on the discussion of developing blockchain applications.Keywords: blockchain, pedagogies, blockchain technologies, blockchain course, blockchain pedagogies
Procedia PDF Downloads 133810 Rapid Evidence Remote Acquisition in High-Availability Server and Storage System for Digital Forensic to Unravel Academic Crime
Authors: Bagus Hanindhito, Fariz Azmi Pratama, Ulfah Nadiya
Abstract:
Nowadays, digital system including, but not limited to, computer and internet have penetrated the education system widely. Critical information such as students’ academic records is stored in a server off- or on-campus. Although several countermeasures have been taken to protect the vital resources from outsider attack, the defense from insiders threat is not getting serious attention. At the end of 2017, a security incident that involved academic information system in one of the most respected universities in Indonesia affected not only the reputation of the institution and its academia but also academic integrity in Indonesia. In this paper, we will explain our efforts in investigating this security incident where we have implemented a novel rapid evidence remote acquisition method in high-availability server and storage system thus our data collection efforts do not disrupt the academic information system and can be conducted remotely minutes after incident report has been received. The acquired evidence is analyzed during digital forensic by constructing the model of the system in an isolated environment which allows multiple investigators to work together. In the end, the suspect is identified as a student (insider), and the investigation result is used by prosecutors to charge the suspect as an academic crime.Keywords: academic information system, academic crime, digital forensic, high-availability server and storage, rapid evidence remote acquisition, security incident
Procedia PDF Downloads 152809 Finite Element Method (FEM) Simulation, design and 3D Print of Novel Highly Integrated PV-TEG Device with Improved Solar Energy Harvest Efficiency
Abstract:
Despite the remarkable advancement of solar cell technology, the challenge of optimizing total solar energy harvest efficiency persists, primarily due to significant heat loss. This excess heat not only diminishes solar panel output efficiency but also curtails its operational lifespan. A promising approach to address this issue is the conversion of surplus heat into electricity. In recent years, there is growing interest in the use of thermoelectric generators (TEG) as a potential solution. The integration of efficient TEG devices holds the promise of augmenting overall energy harvest efficiency while prolonging the longevity of solar panels. While certain research groups have proposed the integration of solar cells and TEG devices, a substantial gap between conceptualization and practical implementation remains, largely attributed to low thermal energy conversion efficiency of TEG devices. To bridge this gap and meet the requisites of practical application, a feasible strategy involves the incorporation of a substantial number of p-n junctions within a confined unit volume. However, the manufacturing of high-density TEG p-n junctions presents a formidable challenge. The prevalent solution often leads to large device sizes to accommodate enough p-n junctions, consequently complicating integration with solar cells. Recently, the adoption of 3D printing technology has emerged as a promising solution to address this challenge by fabricating high-density p-n arrays. Despite this, further developmental efforts are necessary. Presently, the primary focus is on the 3D printing of vertically layered TEG devices, wherein p-n junction density remains constrained by spatial limitations and the constraints of 3D printing techniques. This study proposes a novel device configuration featuring horizontally arrayed p-n junctions of Bi2Te3. The structural design of the device is subjected to simulation through the Finite Element Method (FEM) within COMSOL Multiphysics software. Various device configurations are simulated to identify optimal device structure. Based on the simulation results, a new TEG device is fabricated utilizing 3D Selective laser melting (SLM) printing technology. Fusion 360 facilitates the translation of the COMSOL device structure into a 3D print file. The horizontal design offers a unique advantage, enabling the fabrication of densely packed, three-dimensional p-n junction arrays. The fabrication process entails printing a singular row of horizontal p-n junctions using the 3D SLM printing technique in a single layer. Subsequently, successive rows of p-n junction arrays are printed within the same layer, interconnected by thermally conductive copper. This sequence is replicated across multiple layers, separated by thermal insulating glass. This integration created in a highly compact three-dimensional TEG device with high density p-n junctions. The fabricated TEG device is then attached to the bottom of the solar cell using thermal glue. The whole device is characterized, with output data closely matching with COMSOL simulation results. Future research endeavors will encompass the refinement of thermoelectric materials. This includes the advancement of high-resolution 3D printing techniques tailored to diverse thermoelectric materials, along with the optimization of material microstructures such as porosity and doping. The objective is to achieve an optimal and highly integrated PV-TEG device that can substantially increase the solar energy harvest efficiency.Keywords: thermoelectric, finite element method, 3d print, energy conversion
Procedia PDF Downloads 67