Search results for: numerical solution
883 Evaluation of NoSQL in the Energy Marketplace with GraphQL Optimization
Authors: Michael Howard
Abstract:
The growing popularity of electric vehicles in the United States requires an ever-expanding infrastructure of commercial DC fast charging stations. The U.S. Department of Energy estimates 33,355 publicly available DC fast charging stations as of September 2023. In 2017, 115,370 gasoline stations were operating in the United States, much more ubiquitous than DC fast chargers. Range anxiety is an important impediment to the adoption of electric vehicles and is even more relevant in underserved regions in the country. The peer-to-peer energy marketplace helps fill the demand by allowing private home and small business owners to rent their 240 Volt, level-2 charging facilities. The existing, publicly accessible outlets are wrapped with a Cloud-connected microcontroller managing security and charging sessions. These microcontrollers act as Edge devices communicating with a Cloud message broker, while both buyer and seller users interact with the framework via a web-based user interface. The database storage used by the marketplace framework is a key component in both the cost of development and the performance that contributes to the user experience. A traditional storage solution is the SQL database. The architecture and query language have been in existence since the 1970s and are well understood and documented. The Structured Query Language supported by the query engine provides fine granularity with user query conditions. However, difficulty in scaling across multiple nodes and cost of its server-based compute have resulted in a trend in the last 20 years towards other NoSQL, serverless approaches. In this study, we evaluate the NoSQL vs. SQL solutions through a comparison of Google Cloud Firestore and Cloud SQL MySQL offerings. The comparison pits Google's serverless, document-model, non-relational, NoSQL against the server-base, table-model, relational, SQL service. The evaluation is based on query latency, flexibility/scalability, and cost criteria. Through benchmarking and analysis of the architecture, we determine whether Firestore can support the energy marketplace storage needs and if the introduction of a GraphQL middleware layer can overcome its deficiencies.Keywords: non-relational, relational, MySQL, mitigate, Firestore, SQL, NoSQL, serverless, database, GraphQL
Procedia PDF Downloads 60882 Enhancing Students’ Academic Engagement in Mathematics through a “Concept+Language Mapping” Approach
Authors: Jodie Lee, Lorena Chan, Esther Tong
Abstract:
Hong Kong students face a unique learning environment. Starting from the 2010/2011 school year, The Education Bureau (EDB) of the Government of the Hong Kong Special Administrative Region implemented the fine-tuned Medium of Instruction (MOI) arrangements for secondary schools. Since then, secondary schools in Hong Kong have been given the flexibility to decide the most appropriate MOI arrangements for their schools and under the new academic structure for senior secondary education, particularly on the compulsory part of the mathematics curriculum. In 2019, Hong Kong Diploma of Secondary Education Examination (HKDSE), over 40% of school day candidates attempted the Mathematics Compulsory Part examination in the Chinese version while the rest took the English version. Moreover, only 14.38% of candidates sat for one of the extended Mathematics modules. This results in a serious of intricate issues to students’ learning in post-secondary education programmes. It is worth to note that when students further pursue to an higher education in Hong Kong or even oversea, they may facing substantial difficulties in transiting learning from learning mathematics in their mother tongue in Chinese-medium instruction (CMI) secondary schools to an English-medium learning environment. Some students understood the mathematics concepts were found to fail to fulfill the course requirements at college or university due to their learning experience in secondary study at CMI. They are particularly weak in comprehending the mathematics questions when they are doing their assessment or attempting the test/examination. A government funded project was conducted with the aims of providing integrated learning context and language support to students with a lower level of numeracy and/or with CMI learning experience. By introducing this “integrated concept + language mapping approach”, students can cope with the learning challenges in the compulsory English-medium mathematics and statistics subjects in their tertiary education. Ultimately, in the hope that students can enhance their mathematical ability, analytical skills, and numerical sense for their lifelong learning. The “Concept + Language Mapping “(CLM) approach was adopted and tried out in the bridging courses for students with a lower level of numeracy and/or with CMI learning experiences. At the beginning of each class, a pre-test was conducted, and class time was then devoted to introducing the concepts by CLM approach. For each concept, the key thematic items and their different semantic relations are presented using graphics and animations via the CLM approach. At the end of each class, a post-test was conducted. Quantitative data analysis was performed to study the effect on students’ learning via the CLM approach. Stakeholders' feedbacks were collected to estimate the effectiveness of the CLM approach in facilitating both content and language learning. The results based on both students’ and lecturers’ feedback indicated positive outcomes on adopting the CLM approach to enhance the mathematical ability and analytical skills of CMI students.Keywords: mathematics, Concept+Language Mapping, level of numeracy, medium of instruction
Procedia PDF Downloads 79881 Informal Economy: Case Study of Street Vendors in Bangkok
Authors: Kangrij Roeksiripat
Abstract:
Street vending is one of the informal economy activities which considered significance to Thai people in the economic and the day-to-day social life. It had been believed that the street vendor is a group of the poor and uneducated people. With the increasing numbers of the street vendor occupying space on public sidewalks especially in central business districts, it becomes unclear whether street vending continues as a solution to unemployment for access labors. This research attempts to study and analyze types of street vendors in Bangkok under the informal economy framework. The debate on the heterogeneous informal economy has categorized into four schools; the dualism, the structuralism, the legalism and the voluntarism. The examination also embodies with market concept with Porter’s Five Forces of Competitive Position Model analysis and the interviews with the street vendors in three case study areas: Inner zone (Pathumwan district - the sidewalk on the opposite side of Siam Paragon mall), Middle zone (Ramkhamhaeng district - the sidewalk on the opposite side of Ramkhamhaeng University) and Outer zone (Minburi district- the sidewalk of Sriburanukit Road). The result indicates that most of street vendors in Siam square are voluntarily choose to make a living in vending on a sidewalk and tend to take it as a long-term occupation even though they can be in formal wage employment. Moreover, average income and positive attitude towards self-employed are the important factors that drive them to operate street vending businesses. Meanwhile, street vending is often a family enterprise in Ramkhamhaeng area and most vendors do not wish to transform their businesses into the formal sectors. Whereas the survey conducted in Sriburankit Road reveals that almost all of street vendors migrated from other provinces and were previously paid as the unskilled workers in formal sectors. They moved to informal trades because of the uncertainty of employment in the mainstream sectors and the inconsistent income with knowledge support of friends and relatives from the same hometown. In particular, the result reveals a common pattern that street vending is the very first occupation of some group of vendors and they will continue to engage in this activity. Thus, it is important for the government to design optimal policy which not only integrating informal workers into the formal economy but also monitoring the enforcement of regulations on the modern informal economy.Keywords: informal economy, sidewalks, street vendors, occupation
Procedia PDF Downloads 284880 Quality of Education in Dilla Zone
Authors: Gezahegn Bekele Welldgiyorgise
Abstract:
It is obvious that the economics, politics and social conditions of a country are determined by the quality and standard of its education. Indeed, education plays a vital role in changing the consciousness and awareness of society and transforming it on a large scale. Moreover, education contributes a lot to the advancement of science and technology, information and communication, and above all, it speeds up its progress in no time if it focuses mainly on the qualitative approach to education. Education brings about universal change and transformation and lightens mankind in all dimensions. It creates an educated, enlightened and brightened generation in society. The generation will be sharped, sharpened and well-oriented if it gets modern, sophisticated and standardized education in its field of study. The main goal of education is to produce well-qualified, well-trained and disciplined young offers in a given community. If the youth is well trained and well-mannered, he will certainly be enlightened, problem solvers and solution seekers, researchers, and innovators. In this respect, we have to provide the youth with modern education, a teaching-learning process led by active learning and a participatory approach with a new curriculum preparation for the age of children supported by modern facilities (ICT).In addition to that, the curriculum should have to give attention to mathematics and science lessons that include international experience in a comfortable school and classrooms. Therefore, the generation that will be created through such kinds of the guided education system will make the students active participants, self-confident, researchers and problem solvers, besides that result in changed life standards and a developed country. Similarly, our country, Ethiopia, has aimed to get such change in youth (generation) through modern education, designing a new educational policy and curriculum which was implemented for many years, although the goal of education has not reached the required level. To get the main idea of the article, I should have answered the question of why our country's educational goal had not reached the desired level because it is necessary to lay the foundation for research in finding out problems seen through students learning performance, the first task is selecting primary-school as a sample. Therefore, we selected “Dilla primary school (5-8)” which is a workplace for a teacher and gives me a chance to recognize students’ learning performance to recognize their learning grades (internal and external) and measure performance (achievement) of students easily’.Keywords: curriculum, performance, innovation, learning
Procedia PDF Downloads 75879 Improving Rural Access to Specialist Emergency Mental Health Care: Using a Time and Motion Study in the Evaluation of a Telepsychiatry Program
Authors: Emily Saurman, David Lyle
Abstract:
In Australia, a well serviced rural town might have a psychiatrist visit once-a-month with more frequent visits from a psychiatric nurse, but many have no resident access to mental health specialists. Access to specialist care, would not only reduce patient distress and benefit outcomes, but facilitate the effective use of limited resources. The Mental Health Emergency Care-Rural Access Program (MHEC-RAP) was developed to improve access to specialist emergency mental health care in rural and remote communities using telehealth technologies. However, there has been no current benchmark to gauge program efficiency or capacity; to determine whether the program activity is justifiably sufficient. The evaluation of MHEC-RAP used multiple methods and applied a modified theory of access to assess the program and its aim of improved access to emergency mental health care. This was the first evaluation of a telepsychiatry service to include a time and motion study design examining program time expenditure, efficiency, and capacity. The time and motion study analysis was combined with an observational study of the program structure and function to assess the balance between program responsiveness and efficiency. Previous program studies have demonstrated that MHEC-RAP has improved access and is used and effective. The findings from the time and motion study suggest that MHEC-RAP has the capacity to manage increased activity within the current model structure without loss to responsiveness or efficiency in the provision of care. Enhancing program responsiveness and efficiency will also support a claim of the program’s value for money. MHEC-RAP is a practical telehealth solution for improving access to specialist emergency mental health care. The findings from this evaluation have already attracted the attention of other regions in Australia interested in implementing emergency telepsychiatry programs and are now informing the progressive establishment of mental health resource centres in rural New South Wales. Like MHEC-RAP, these centres will provide rapid, safe, and contextually relevant assessments and advice to support local health professionals to manage mental health emergencies in the smaller rural emergency departments. Sharing the application of this methodology and research activity may help to improve access to and future evaluations of telehealth and telepsychiatry services for others around the globe.Keywords: access, emergency, mental health, rural, time and motion
Procedia PDF Downloads 232878 Strengthening Reinforced Concrete Beams Using Carbon Fibre Reinforced Polymer Strips
Authors: Mina Iskander, Mina Melad, Mourad Yasser, Waleed Abdel Rahim, Amr Mosa, Mohamed El Lahamy, Ezzeldin Sayed-Ahmed, Mohamed Abou-Zeid
Abstract:
Strengthening of reinforced concrete beams in flexure using externally bonded composite laminate of high tensile strength is easy and of the minimum cost compared to traditional methods such as increasing the concrete section depth or reinforcement that requires formwork and curing which affect the structure usability. One of the main limitations of this technique is debonding of the externally bonded laminate, either by end delamination or by mid-span flexural crack-induced debonding. ACI 440.2-08 suggests that using side-bonded FRP laminate in the flexural strengthening of RC beams may serve to limit the extent and width of flexural cracks. Consequently, this technique may decrease the effect of flexural cracks on initiating the mid-span debonding; i.e. delays the flexural crack-induced debonding. Furthermore, bonding the FRP strips to the side of the beam may offer an attractive, practical solution when the soffit of this beam is not accessible. This paper presents an experimental programme designed to investigate the effect of using externally bonded CFRP laminate on the sides of reinforced concrete beams and compares the results to those of bonding the CFRP laminate to the soffit of the beams. In addition, the paper discusses the effect of using end anchorage by U-wrapping the CFRP strips at their end zones with CFRP sheets for beams strengthened with soffit-bonded and side-bonded CFRP strips. Thus, ten rectangular reinforced concrete beams were tested to failure in order to study the effect of changing the location of the externally bonded laminate on the flexural capacity and ductility of the strengthened beams. Pultruded CFRP strips were bonded to the soffit of the beams or their sides to check the possibility of limiting the flexural cracking in mid-span region, which is the main reason for mid-span debonding. Pre-peg CFRP sheets were used near the support as U-wrap for the beam to act as an end-anchorage for the externally bonded strips in order to delay/prevent the end delamination. Strength gains of 38% and 43% were recorded for the soffit-bonded and the side-bonded composite strips with end U-wrapped sheets, respectively. Furthermore, beams with end sheets applied as an end anchorage showed higher ductility than those without these sheets.Keywords: flexural strengthening, externally bonded CFRP, side-bonded CFRP, CFRP laminates
Procedia PDF Downloads 353877 Modern Agriculture and Employment Generation in Nigeria: A Recursive Model Approach
Authors: Ese Urhie, Olabisi Popoola, Obindah Gershon
Abstract:
Several policies and programs initiated to address the challenge of unemployment in Nigeria seem to be inadequate. The desired structural transformation which is expected to absorb the excess labour in the economy is yet to be achieved. The agricultural sector accounts for almost half of the labour force with very low productivity. This could partly explain why the much anticipated structural transformation has not been achieved. A major reason for the low productivity is the fact that the production process is predominantly based on the use of traditional tools. In view of the underdeveloped nature of the agricultural sector, Nigeria still has huge potentials for productivity enhancement through modern technology. Aside from productivity enhancement, modern agriculture also stimulates both backward and forward linkages that promote investment and thus generate employment. Contrary to the apprehension usually expressed by many stake-holders about the adoption of modern technology by labour-abundant less-developed countries, this study showed that though there will be job loss initially, the reverse will be the case in the long-run. The outcome of this study will enhance the understanding of all stakeholders in the sector and also encourage them to adopt modern techniques of farming. It will also aid policy formulation at both sectoral and national levels. The recursive model and analysis adopted in the study is useful because it exhibits a unilateral cause-and-effect relationship which most simultaneous equation models do not. It enables the structural equations to be ordered in such a way that the first equation includes only predetermined variables on the right-hand side, while the solution for the final endogenous variable is completely determined by all equations of the system. The study examines the transmission channels and effect of modern agriculture on agricultural productivity and employment growth in Nigeria, via its forward and backward linkages. Using time series data spanning 1980 to 2014, the result of the analyses shows that: (i) a significant and positive relationship between agricultural productivity growth and modern agriculture; (ii) a significant and negative relationship between export price index and agricultural productivity growth; (iii) a significant and positive relationship between export and investment; and (iv) a significant and positive relationship between investment and employment growth. The unbalanced growth theory will be a good strategy to adopt by developing countries such as Nigeria.Keywords: employment, modern agriculture, productivity, recursive model
Procedia PDF Downloads 264876 Verification of the Supercavitation Phenomena: Investigation of the Cavity Parameters and Drag Coefficients for Different Types of Cavitator
Authors: Sezer Kefeli, Sertaç Arslan
Abstract:
Supercavitation is a pressure dependent process which gives opportunity to eliminate the wetted surface effects on the underwater vehicle due to the differences of viscosity and velocity effects between liquid (freestream) and gas phase. Cavitation process occurs depending on rapid pressure drop or temperature rising in liquid phase. In this paper, pressure based cavitation is investigated due to the fact that is encountered in the underwater world, generally. Basically, this vapor-filled pressure based cavities are unstable and harmful for any underwater vehicle because these cavities (bubbles or voids) lead to intense shock waves while collapsing. On the other hand, supercavitation is a desired and stabilized phenomena than general pressure based cavitation. Supercavitation phenomena offers the idea of minimizing form drag, and thus supercavitating vehicles are revived. When proper circumstances are set up, which are either increasing the operating speed of the underwater vehicle or decreasing the pressure difference between free stream and artificial pressure, the continuity of the supercavitation is obtainable. There are 2 types of supercavitation to obtain stable and continuous supercavitation, and these are called as natural and artificial supercavitation. In order to generate natural supercavitation, various mechanical structures are discovered, which are called as cavitators. In literature, a lot of cavitator types are studied either experimentally or numerically on a CFD platforms with intent to observe natural supercavitation since the 1900s. In this paper, firstly, experimental results are obtained, and trend lines are generated based on supercavitation parameters in terms of cavitation number (), form drag coefficientC_D, dimensionless cavity diameter (d_m/d_c), and length (L_c/d_c). After that, natural cavitation verification studies are carried out for disk and cone shape cavitators. In addition, supercavitation parameters are numerically analyzed at different operating conditions, and CFD results are fitted into trend lines of experimental results. The aims of this paper are to generate one generally accepted drag coefficient equation for disk and cone cavitators at different cavitator half angle and investigation of the supercavitation parameters with respect to cavitation number. Moreover, 165 CFD analysis are performed at different cavitation numbers on FLUENT version 21R2. Five different cavitator types are modeled on SCDM with respect tocavitator’s half angles. After that, CFD database is generated depending on numerical results, and new trend lines are generated based on supercavitation parameters. These trend lines are compared with experimental results. Finally, the generally accepted drag coefficient equation and equations of supercavitation parameters are generated.Keywords: cavity envelope, CFD, high speed underwater vehicles, supercavitation, supercavitating flows, supercavitation parameters, drag reduction, viscous force elimination, natural cavitation verification
Procedia PDF Downloads 130875 Design of Identification Based Adaptive Control for Fermentation Process in Bioreactor
Authors: J. Ritonja
Abstract:
The biochemical technology has been developing extremely fast since the middle of the last century. The main reason for such development represents a requirement for large production of high-quality biologically manufactured products such as pharmaceuticals, foods, and beverages. The impact of the biochemical industry on the world economy is enormous. The great importance of this industry also results in intensive development in scientific disciplines relevant to the development of biochemical technology. In addition to developments in the fields of biology and chemistry, which enable to understand complex biochemical processes, development in the field of control theory and applications is also very important. In the paper, the control for the biochemical reactor for the milk fermentation was studied. During the fermentation process, the biophysical quantities must be precisely controlled to obtain the high-quality product. To control these quantities, the bioreactor’s stirring drive and/or heating system can be used. Available commercial biochemical reactors are equipped with open loop or conventional linear closed loop control system. Due to the outstanding parameters variations and the partial nonlinearity of the biochemical process, the results obtained with these control systems are not satisfactory. To improve the fermentation process, the self-tuning adaptive control system was proposed. The use of the self-tuning adaptive control is suggested because the parameters’ variations of the studied biochemical process are very slow in most cases. To determine the linearized mathematical model of the fermentation process, the recursive least square identification method was used. Based on the obtained mathematical model the linear quadratic regulator was tuned. The parameters’ identification and the controller’s synthesis are executed on-line and adapt the controller’s parameters to the fermentation process’ dynamics during the operation. The use of the proposed combination represents the original solution for the control of the milk fermentation process. The purpose of the paper is to contribute to the progress of the control systems for the biochemical reactors. The proposed adaptive control system was tested thoroughly. From the obtained results it is obvious that the proposed adaptive control system assures much better following of the reference signal as a conventional linear control system with fixed control parameters.Keywords: adaptive control, biochemical reactor, linear quadratic regulator, recursive least square identification
Procedia PDF Downloads 124874 Simulation of Optimum Sculling Angle for Adaptive Rowing
Authors: Pornthep Rachnavy
Abstract:
The purpose of this paper is twofold. First, we believe that there are a significant relationship between sculling angle and sculling style among adaptive rowing. Second, we introduce a methodology used for adaptive rowing, namely simulation, to identify effectiveness of adaptive rowing. For our study we simulate the arms only single scull of adaptive rowing. The method for rowing fastest under the 1000 meter was investigated by study sculling angle using the simulation modeling. A simulation model of a rowing system was developed using the Matlab software package base on equations of motion consist of many variation for moving the boat such as oars length, blade velocity and sculling style. The boat speed, power and energy consumption on the system were compute. This simulation modeling can predict the force acting on the boat. The optimum sculling angle was performing by computer simulation for compute the solution. Input to the model are sculling style of each rower and sculling angle. Outputs of the model are boat velocity at 1000 meter. The present study suggests that the optimum sculling angle exist depends on sculling styles. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the first style is -57.00 and 22.0 degree. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the second style is -57.00 and 22.0 degree. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the third style is -51.57 and 28.65 degree. The optimum angle for blade entry and release with respect to the perpendicular through the pin of the fourth style is -45.84 and 34.38 degree. A theoretical simulation for rowing has been developed and presented. The results suggest that it may be advantageous for the rowers to select the sculling angles proper to sculling styles. The optimum sculling angles of the rower depends on the sculling styles made by each rower. The investigated of this paper can be concludes in three directions: 1;. There is the optimum sculling angle in arms only single scull of adaptive rowing. 2. The optimum sculling angles depend on the sculling styles. 3. Computer simulation of rowing can identify opportunities for improving rowing performance by utilizing the kinematic description of rowing. The freedom to explore alternatives in speed, thrust and timing with the computer simulation will provide the coach with a tool for systematic assessments of rowing technique In addition, the ability to use the computer to examine the very complex movements during rowing will help both the rower and the coach to conceptualize the components of movements that may have been previously unclear or even undefined.Keywords: simulation, sculling, adaptive, rowing
Procedia PDF Downloads 464873 A Geo DataBase to Investigate the Maximum Distance Error in Quality of Life Studies
Authors: Paolino Di Felice
Abstract:
The background and significance of this study come from papers already appeared in the literature which measured the impact of public services (e.g., hospitals, schools, ...) on the citizens’ needs satisfaction (one of the dimensions of QOL studies) by calculating the distance between the place where they live and the location on the territory of the services. Those studies assume that the citizens' dwelling coincides with the centroid of the polygon that expresses the boundary of the administrative district, within the city, they belong to. Such an assumption “introduces a maximum measurement error equal to the greatest distance between the centroid and the border of the administrative district.”. The case study, this abstract reports about, investigates the implications descending from the adoption of such an approach but at geographical scales greater than the urban one, namely at the three levels of nesting of the Italian administrative units: the (20) regions, the (110) provinces, and the 8,094 municipalities. To carry out this study, it needs to be decided: a) how to store the huge amount of (spatial and descriptive) input data and b) how to process them. The latter aspect involves: b.1) the design of algorithms to investigate the geometry of the boundary of the Italian administrative units; b.2) their coding in a programming language; b.3) their execution and, eventually, b.4) archiving the results in a permanent support. The IT solution we implemented is centered around a (PostgreSQL/PostGIS) Geo DataBase structured in terms of three tables that fit well to the hierarchy of nesting of the Italian administrative units: municipality(id, name, provinceId, istatCode, regionId, geometry) province(id, name, regionId, geometry) region(id, name, geometry). The adoption of the DBMS technology allows us to implement the steps "a)" and "b)" easily. In particular, step "b)" is simplified dramatically by calling spatial operators and spatial built-in User Defined Functions within SQL queries against the Geo DB. The major findings coming from our experiments can be summarized as follows. The approximation that, on the average, descends from assimilating the residence of the citizens with the centroid of the administrative unit of reference is of few kilometers (4.9) at the municipalities level, while it becomes conspicuous at the other two levels (28.9 and 36.1, respectively). Therefore, studies such as those mentioned above can be extended up to the municipal level without affecting the correctness of the interpretation of the results, but not further. The IT framework implemented to carry out the experiments can be replicated for studies referring to the territory of other countries all over the world.Keywords: quality of life, distance measurement error, Italian administrative units, spatial database
Procedia PDF Downloads 371872 Going beyond Elementary Algebraic Identities: The Expectation of a Gifted Child, an Indian Scenario
Authors: S. R. Santhanam
Abstract:
A gifted child is one who gives evidence of creativity, good memory, rapid learning. In mathematics, a teacher often comes across some gifted children and they exhibit the following characteristics: unusual alertness, enjoying solving problems, getting bored on repetitions, self-taught, going beyond what teacher taught, ask probing questions, connecting unconnected concepts, vivid imagination, readiness for research work, perseverance of a topic. There are two main areas of research carried out on them: 1)identifying gifted children, 2) interacting and channelizing them. A lack of appropriate recognition will lead the gifted child demotivated. One of the main findings is if proper attention and nourishment are not given then it leads a gifted child to become depressed, underachieving, fail to reach their full potential and sometimes develop negative attitude towards school and study. After identifying them, a mathematics teacher has to develop them into a fall fledged achiever. The responsibility of the teacher is enormous. The teacher has to be resourceful and patient. But interacting with them one finds a lot of surprises and awesomeness. The elementary algebraic identities like (a+b)(a-b)=a²-b², expansion of like (a+b)²(a-b)² and others are taught to students, of age group 13-15 in India. An average child will be satisfied with a single proof and immediate application of these identities. But a gifted child expects more from the teacher and at one stage after a little training will surpass the teacher also. In this short paper, the author shares his experience regarding teaching algebraic identities to gifted children. The following problem was given to a set of 10 gifted children of the specified age group: If a natural number ‘n’ to expressed as the sum of the two squares, will 2n also be expressed as the sum of two squares? An investigation has been done on what multiples of n satisfying the criterion. The attempts of the gifted children were consolidated and conclusion was drawn. A second problem was given to them as: can two natural numbers be found such that the difference of their square is 3? After a successful solution, more situations were analysed. As a third question, the finding of the sign of an algebraic expression in three variables was analysed. As an example: if a,b,c are real and unequal what will be sign of a²+4b²+9c²-4ab-12bc-6ca? Apart from an expression as a perfect square what other methods can be employed to prove an algebraic expression as positive negative or non negative has been analysed. Expressions like 4x²+2y²+13y²-2xy-4yz-6zx were given, and the children were asked to find the sign of the expression for all real values of x,y and z. In all investigations, only basic algebraic identities were used. As a next probe, a divisibility problem was initiated. When a,b,c are natural numbers such that a+b+c is at least 6, and if a+b+c is divisible by 6 then will 6 divide a³+b³+c³. The gifted children solved it in two different ways.Keywords: algebraic identities, gifted children, Indian scenario, research
Procedia PDF Downloads 179871 Control of Biofilm Formation and Inorganic Particle Accumulation on Reverse Osmosis Membrane by Hypochlorite Washing
Authors: Masaki Ohno, Cervinia Manalo, Tetsuji Okuda, Satoshi Nakai, Wataru Nishijima
Abstract:
Reverse osmosis (RO) membranes have been widely used for desalination to purify water for drinking and other purposes. Although at present most RO membranes have no resistance to chlorine, chlorine-resistant membranes are being developed. Therefore, direct chlorine treatment or chlorine washing will be an option in preventing biofouling on chlorine-resistant membranes. Furthermore, if particle accumulation control is possible by using chlorine washing, expensive pretreatment for particle removal can be removed or simplified. The objective of this study was to determine the effective hypochlorite washing condition required for controlling biofilm formation and inorganic particle accumulation on RO membrane in a continuous flow channel with RO membrane and spacer. In this study, direct chlorine washing was done by soaking fouled RO membranes in hypochlorite solution and fluorescence intensity was used to quantify biofilm on the membrane surface. After 48 h of soaking the membranes in high fouling potential waters, the fluorescence intensity decreased to 0 from 470 using the following washing conditions: 10 mg/L chlorine concentration, 2 times/d washing interval, and 30 min washing time. The chlorine concentration required to control biofilm formation decreased as the chlorine concentration (0.5–10 mg/L), the washing interval (1–4 times/d), or the washing time (1–30 min) increased. For the sample solutions used in the study, 10 mg/L chlorine concentration with 2 times/d interval, and 5 min washing time was required for biofilm control. The optimum chlorine washing conditions obtained from soaking experiments proved to be applicable also in controlling biofilm formation in continuous flow experiments. Moreover, chlorine washing employed in controlling biofilm with suspended particles resulted in lower amounts of organic (0.03 mg/cm2) and inorganic (0.14 mg/cm2) deposits on the membrane than that for sample water without chlorine washing (0.14 mg/cm2 and 0.33 mg/cm2, respectively). The amount of biofilm formed was 79% controlled by continuous washing with 10 mg/L of free chlorine concentration, and the inorganic accumulation amount decreased by 58% to levels similar to that of pure water with kaolin (0.17 mg/cm2) as feed water. These results confirmed the acceleration of particle accumulation due to biofilm formation, and that the inhibition of biofilm growth can almost completely reduce further particle accumulation. In addition, effective hypochlorite washing condition which can control both biofilm formation and particle accumulation could be achieved.Keywords: reverse osmosis, washing condition optimization, hypochlorous acid, biofouling control
Procedia PDF Downloads 348870 Impact of Chess Intervention on Cognitive Functioning of Children
Authors: Ebenezer Joseph
Abstract:
Chess is a useful tool to enhance general and specific cognitive functioning in children. The present study aims to assess the impact of chess on cognitive in children and to measure the differential impact of socio-demographic factors like age and gender of the child on the effectiveness of the chess intervention.This research study used an experimental design to study the impact of the Training in Chess on the intelligence of children. The Pre-test Post-test Control Group Design was utilized. The research design involved two groups of children: an experimental group and a control group. The experimental group consisted of children who participated in the one-year Chess Training Intervention, while the control group participated in extra-curricular activities in school. The main independent variable was training in chess. Other independent variables were gender and age of the child. The dependent variable was the cognitive functioning of the child (as measured by IQ, working memory index, processing speed index, perceptual reasoning index, verbal comprehension index, numerical reasoning, verbal reasoning, non-verbal reasoning, social intelligence, language, conceptual thinking, memory, visual motor and creativity). The sample consisted of 200 children studying in Government and Private schools. Random sampling was utilized. The sample included both boys and girls falling in the age range 6 to 16 years. The experimental group consisted of 100 children (50 from Government schools and 50 from Private schools) with an equal representation of boys and girls. The control group similarly consisted of 100 children. The dependent variables were assessed using Binet-Kamat Test of Intelligence, Wechsler Intelligence Scale for Children - IV (India) and Wallach Kogan Creativity Test. The training methodology comprised Winning Moves Chess Learning Program - Episodes 1–22, lectures with the demonstration board, on-the-board playing and training, chess exercise through workbooks (Chess school 1A, Chess school 2, and tactics) and working with chess software. Further students games were mapped using chess software and the brain patterns of the child were understood. They were taught the ideas behind chess openings and exposure to classical games were also given. The children participated in mock as well as regular tournaments. Preliminary analysis carried out using independent t tests with 50 children indicates that chess training has led to significant increases in the intelligent quotient. Children in the experimental group have shown significant increases in composite scores like working memory and perceptual reasoning. Chess training has significantly enhanced the total creativity scores, line drawing and pattern meaning subscale scores. Systematically learning chess as part of school activities appears to have a broad spectrum of positive outcomes.Keywords: chess, intelligence, creativity, children
Procedia PDF Downloads 256869 Good Governance Complementary to Corruption Abatement: A Cross-Country Analysis
Authors: Kamal Ray, Tapati Bhattacharya
Abstract:
Private use of public office for private gain could be a tentative definition of corruption and most distasteful event of corruption is that it is not there, nor that it is pervasive, but it is socially acknowledged in the global economy, especially in the developing nations. We attempted to assess the interrelationship between the Corruption perception index (CPI) and the principal components of governance indicators as per World Bank like Control of Corruption (CC), rule of law (RL), regulatory quality (RQ) and government effectiveness (GE). Our empirical investigation concentrates upon the degree of reflection of governance indicators upon the CPI in order to single out the most powerful corruption-generating indicator in the selected countries. We have collected time series data on above governance indicators such as CC, RL, RQ and GE of the selected eleven countries from the year of 1996 to 2012 from World Bank data set. The countries are USA, UK, France, Germany, Greece, China, India, Japan, Thailand, Brazil, and South Africa. Corruption Perception Index (CPI) of the countries mentioned above for the period of 1996 to 2012is also collected. Graphical method of simple line diagram against the time series data on CPI is applied for quick view for the relative positions of different trend lines of different nations. The correlation coefficient is enough to assess primarily the degree and direction of association between the variables as we get the numerical data on governance indicators of the selected countries. The tool of Granger Causality Test (1969) is taken into account for investigating causal relationships between the variables, cause and effect to speak of. We do not need to verify stationary test as length of time series is short. Linear regression is taken as a tool for quantification of a change in explained variables due to change in explanatory variable in respect of governance vis a vis corruption. A bilateral positive causal link between CPI and CC is noticed in UK, index-value of CC increases by 1.59 units as CPI increases by one unit and CPI rises by 0.39 units as CC rises by one unit, and hence it has a multiplier effect so far as reduction in corruption is concerned in UK. GE causes strongly to the reduction of corruption in UK. In France, RQ is observed to be a most powerful indicator in reducing corruption whereas it is second most powerful indicator after GE in reducing of corruption in Japan. Governance-indicator like GE plays an important role to push down the corruption in Japan. In China and India, GE is proactive as well as influencing indicator to curb corruption. The inverse relationship between RL and CPI in Thailand indicates that ongoing machineries related to RL is not complementary to the reduction of corruption. The state machineries of CC in S. Africa are highly relevant to reduce the volume of corruption. In Greece, the variations of CPI positively influence the variations of CC and the indicator like GE is effective in controlling corruption as reflected by CPI. All the governance-indicators selected so far have failed to arrest their state level corruptions in USA, Germany and Brazil.Keywords: corruption perception index, governance indicators, granger causality test, regression
Procedia PDF Downloads 303868 Specification Requirements for a Combined Dehumidifier/Cooling Panel: A Global Scale Analysis
Authors: Damien Gondre, Hatem Ben Maad, Abdelkrim Trabelsi, Frédéric Kuznik, Joseph Virgone
Abstract:
The use of a radiant cooling solution would enable to lower cooling needs which is of great interest when the demand is initially high (hot climate). But, radiant systems are not naturally compatibles with humid climates since a low-temperature surface leads to condensation risks as soon as the surface temperature is close to or lower than the dew point temperature. A radiant cooling system combined to a dehumidification system would enable to remove humidity for the space, thereby lowering the dew point temperature. The humidity removal needs to be especially effective near the cooled surface. This requirement could be fulfilled by a system using a single desiccant fluid for the removal of both excessive heat and moisture. This task aims at providing an estimation of the specification requirements of such system in terms of cooling power and dehumidification rate required to fulfill comfort issues and to prevent any condensation risk on the cool panel surface. The present paper develops a preliminary study on the specification requirements, performances and behavior of a combined dehumidifier/cooling ceiling panel for different operating conditions. This study has been carried using the TRNSYS software which allows nodal calculations of thermal systems. It consists of the dynamic modeling of heat and vapor balances of a 5m x 3m x 2.7m office space. In a first design estimation, this room is equipped with an ideal heating, cooling, humidification and dehumidification system so that the room temperature is always maintained in between 21◦C and 25◦C with a relative humidity in between 40% and 60%. The room is also equipped with a ventilation system that includes a heat recovery heat exchanger and another heat exchanger connected to a heat sink. Main results show that the system should be designed to meet a cooling power of 42W.m−2 and a desiccant rate of 45 gH2O.h−1. In a second time, a parametric study of comfort issues and system performances has been achieved on a more realistic system (that includes a chilled ceiling) under different operating conditions. It enables an estimation of an acceptable range of operating conditions. This preliminary study is intended to provide useful information for the system design.Keywords: dehumidification, nodal calculation, radiant cooling panel, system sizing
Procedia PDF Downloads 174867 Biomimetic Systems to Reveal the Action Mode of Epigallocatechin-3-Gallate in Lipid Membrane
Authors: F. Pires, V. Geraldo, O. N. Oliveira Jr., M. Raposo
Abstract:
Catechins are powerful antioxidants which have attractive properties useful for tumor therapy. Considering their antioxidant activity, these molecules can act as a scavenger of the reactive oxygen species (ROS), alleviating the damage of cell membrane induced by oxidative stress. The complexity and dynamic nature of the cell membrane compromise the analysis of the biophysical interactions between drug and cell membrane and restricts the transport or uptake of the drug by intracellular targets. To avoid the cell membrane complexity, we used biomimetic systems as liposomes and Langmuir monolayers to study the interaction between catechin and membranes at the molecular level. Liposomes were formed after the dispersion of anionic 1,2-dipalmitoyl-sn-glycero-3-[phospho-rac-(1-glycerol)(sodium salt) (DPPG) phospholipids in an aqueous solution, which mimic the arrangement of lipids in natural cell membranes and allows the entrapment of catechins. Langmuir monolayers were formed after dropping amphiphilic molecules, DPPG phospholipids, dissolved in an organic solvent onto the water surface. In this work, we mixed epigallocatechin-3-gallate (EGCG) with DPPG liposomes and exposed them to ultra-violet radiation in order to evaluate the antioxidant potential of these molecules against oxidative stress induced by radiation. The presence of EGCG in the mixture decreased the rate of lipid peroxidation, proving that EGCG protects membranes through the quenching of the reactive oxygen species. Considering the high amount of hydroxyl groups (OH groups) on structure of EGCG, a possible mechanism to these molecules interact with membrane is through hydrogen bonding. We also investigated the effect of EGCG at various concentrations on DPPG Langmuir monolayers. The surface pressure isotherms and infrared reflection-absorption spectroscopy (PM-IRRAS) results corroborate with absorbance results preformed on liposome-model, showing that EGCG interacts with polar heads of the monolayers. This study elucidates the physiological action of EGCG which can be incorporated in lipid membrane. These results are also relevant for the improvement of the current protocols used to incorporate catechins in drug delivery systems.Keywords: catechins, lipid membrane, anticancer agent, molecular interactions
Procedia PDF Downloads 231866 Insights into Particle Dispersion, Agglomeration and Deposition in Turbulent Channel Flow
Authors: Mohammad Afkhami, Ali Hassanpour, Michael Fairweather
Abstract:
The work described in this paper was undertaken to gain insight into fundamental aspects of turbulent gas-particle flows with relevance to processes employed in a wide range of applications, such as oil and gas flow assurance in pipes, powder dispersion from dry powder inhalers, and particle resuspension in nuclear waste ponds, to name but a few. In particular, the influence of particle interaction and fluid phase behavior in turbulent flow on particle dispersion in a horizontal channel is investigated. The mathematical modeling technique used is based on the large eddy simulation (LES) methodology embodied in the commercial CFD code FLUENT, with flow solutions provided by this approach coupled to a second commercial code, EDEM, based on the discrete element method (DEM) which is used for the prediction of particle motion and interaction. The results generated by LES for the fluid phase have been validated against direct numerical simulations (DNS) for three different channel flows with shear Reynolds numbers, Reτ = 150, 300 and 590. Overall, the LES shows good agreement, with mean velocities and normal and shear stresses matching those of the DNS in both magnitude and position. The research work has focused on the prediction of those conditions favoring particle aggregation and deposition within turbulent flows. Simulations have been carried out to investigate the effects of particle size, density and concentration on particle agglomeration. Furthermore, particles with different surface properties have been simulated in three channel flows with different levels of flow turbulence, achieved by increasing the Reynolds number of the flow. The simulations mimic the conditions of two-phase, fluid-solid flows frequently encountered in domestic, commercial and industrial applications, for example, air conditioning and refrigeration units, heat exchangers, oil and gas suction and pressure lines. The particle size, density, surface energy and volume fractions selected are 45.6, 102 and 150 µm, 250, 1000 and 2159 kg m-3, 50, 500, and 5000 mJ m-2 and 7.84 × 10-6, 2.8 × 10-5, and 1 × 10-4, respectively; such particle properties are associated with particles found in soil, as well as metals and oxides prevalent in turbulent bounded fluid-solid flows due to erosion and corrosion of inner pipe walls. It has been found that the turbulence structure of the flow dominates the motion of the particles, creating particle-particle interactions, with most of these interactions taking place at locations close to the channel walls and in regions of high turbulence where their agglomeration is aided both by the high levels of turbulence and the high concentration of particles. A positive relationship between particle surface energy, concentration, size and density, and agglomeration was observed. Moreover, the results derived for the three Reynolds numbers considered show that the rate of agglomeration is strongly influenced for high surface energy particles by, and increases with, the intensity of the flow turbulence. In contrast, for lower surface energy particles, the rate of agglomeration diminishes with an increase in flow turbulence intensity.Keywords: agglomeration, channel flow, DEM, LES, turbulence
Procedia PDF Downloads 316865 Design of the Ice Rink of the Future
Authors: Carine Muster, Prina Howald Erika
Abstract:
Today's ice rinks are important energy consumers for the production and maintenance of ice. At the same time, users demand that the other rooms should be tempered or heated. The building complex must equally provide cooled and heated zones, which does not translate as carbon-zero ice rinks. The study provides an analysis of how the civil engineering sector can significantly impact minimizing greenhouse gas emissions and optimizing synergies across an entire ice rink complex. The analysis focused on three distinct aspects: the layout, including the volumetric layout of the premises present in an ice rink; the materials chosen that can potentially use the most ecological structural approach; and the construction methods based on innovative solutions to reduce carbon footprint. The first aspect shows that the organization of the interior volumes and defining the shape of the rink play a significant role. Its layout makes the use and operation of the premises as efficient as possible, thanks to the differentiation between heated and cooled volumes while optimising heat loss between the different rooms. The sprayed concrete method, which is still little known, proves that it is possible to achieve the strength of traditional concrete for the structural aspect of the load-bearing and non-load-bearing walls of the ice rink by using materials excavated from the construction site and providing a more ecological and sustainable solution. The installation of an empty sanitary space underneath the ice floor, making it independent of the rest of the structure, provides a natural insulating layer, preventing the transfer of cold to the rest of the structure and reducing energy losses. The addition of active pipes as part of the foundation of the ice floor, coupled with a suitable system, gives warmth in the winter and storage in the summer; this is all possible thanks to the natural heat in the ground. In conclusion, this study provides construction recommendations for future ice rinks with a significantly reduced energy demand, using some simple preliminary design concepts. By optimizing the layout, materials, and construction methods of ice rinks, the civil engineering sector can play a key role in reducing greenhouse gas emissions and promoting sustainability.Keywords: climate change, energy optimization, green building, sustainability
Procedia PDF Downloads 66864 Empowering Leaders: Strategies for Effective Management in a Changing World
Authors: Shahid Ali
Abstract:
Leadership and management are essential components of running successful organizations. Both concepts are closely related but serve different purposes in the overall management of a company. Leadership focuses on inspiring and motivating employees towards a common goal, while management involves coordinating and directing resources to achieve organizational objectives efficiently. Objectives of Leadership and Management: Inspiring and motivating employees: A key objective of leadership is to inspire and motivate employees to work towards achieving the organization’s goals. Effective leaders create a vision that employees can align with and provide the necessary motivation to drive performance. Setting goals and objectives: Both leadership and management play a crucial role in setting goals and objectives for the organization. Leaders create a vision for the future, while managers develop plans to achieve specific objectives within the given timeframe. Implementing strategies: Leaders come up with innovative strategies to drive the organization forward, while managers are responsible for implementing these strategies effectively. Together, leadership and management ensure that the organization’s plans are executed efficiently. Contributions of Leadership and Management: Employee Engagement: Effective leadership and management can increase employee engagement and satisfaction. When employees feel motivated and inspired by their leaders, they are more likely to be engaged in their work and contribute to the organization’s success. Organizational Success: Good leadership and management are essential for navigating the challenges and changes that organizations face. By setting clear goals, inspiring employees, and making strategic decisions, leaders and managers can drive organizational success. Talent Development: Leaders and managers are responsible for identifying and developing talent within the organization. By providing feedback, training, and coaching, they can help employees reach their full potential and contribute effectively to the organization. Research Type: The research on leadership and management is typically quantitative and qualitative in nature. Quantitative research involves the collection and analysis of numerical data to understand the impact of leadership and management practices on organizational outcomes. This type of research often uses surveys, questionnaires, and statistical analysis to measure variables such as employee satisfaction, performance, and organizational success. Qualitative research, on the other hand, involves exploring the subjective experiences and perspectives of individuals related to leadership and management. This type of research may include interviews, observations, and case studies to gain a deeper understanding of how leadership and management practices influence organizational behavior and outcomes. In conclusion, leadership and management play a critical role in the success of organizations. Through effective leadership and management practices, organizations can inspire and motivate employees, set goals, and implement strategies to achieve their objectives. Research on leadership and management helps to understand the impact of these practices on organizational outcomes and provides valuable insights for improving leadership and management practices in the future.Keywords: empowering, leadership, management, adaptability
Procedia PDF Downloads 48863 Probabilistic Study of Impact Threat to Civil Aircraft and Realistic Impact Energy
Authors: Ye Zhang, Chuanjun Liu
Abstract:
In-service aircraft is exposed to different types of threaten, e.g. bird strike, ground vehicle impact, and run-way debris, or even lightning strike, etc. To satisfy the aircraft damage tolerance design requirements, the designer has to understand the threatening level for different types of the aircraft structures, either metallic or composite. Exposing to low-velocity impacts may produce very serious internal damages such as delaminations and matrix cracks without leaving visible mark onto the impacted surfaces for composite structures. This internal damage can cause significant reduction in the load carrying capacity of structures. The semi-probabilistic method provides a practical and proper approximation to establish the impact-threat based energy cut-off level for the damage tolerance evaluation of the aircraft components. Thus, the probabilistic distribution of impact threat and the realistic impact energy level cut-offs are the essential establishments required for the certification of aircraft composite structures. A new survey of impact threat to civil aircraft in-service has recently been carried out based on field records concerning around 500 civil aircrafts (mainly single aisles) and more than 4.8 million flight hours. In total 1,006 damages caused by low-velocity impact events had been screened out from more than 8,000 records including impact dents, scratches, corrosions, delaminations, cracks etc. The impact threat dependency on the location of the aircraft structures and structural configuration was analyzed. Although the survey was mainly focusing on the metallic structures, the resulting low-energy impact data are believed likely representative to general civil aircraft, since the service environments and the maintenance operations are independent of the materials of the structures. The probability of impact damage occurrence (Po) and impact energy exceedance (Pe) are the two key parameters for describing the statistic distribution of impact threat. With the impact damage events from the survey, Po can be estimated as 2.1x10-4 per flight hour. Concerning the calculation of Pe, a numerical model was developed using the commercial FEA software ABAQUS to backward estimate the impact energy based on the visible damage characteristics. The relationship between the visible dent depth and impact energy was established and validated by drop-weight impact experiments. Based on survey results, Pe was calculated and assumed having a log-linear relationship versus the impact energy. As the product of two aforementioned probabilities, Po and Pe, it is reasonable and conservative to assume Pa=PoxPe=10-5, which indicates that the low-velocity impact events are similarly likely as the Limit Load events. Combing Pa with two probabilities Po and Pe obtained based on the field survey, the cutoff level of realistic impact energy was estimated and valued as 34 J. In summary, a new survey was recently done on field records of civil aircraft to investigate the probabilistic distribution of impact threat. Based on the data, two probabilities, Po and Pe, were obtained. Considering a conservative assumption of Pa, the cutoff energy level for the realistic impact energy has been determined, which provides potential applicability in damage tolerance certification of future civil aircraft.Keywords: composite structure, damage tolerance, impact threat, probabilistic
Procedia PDF Downloads 307862 Impact of Instructional Designing in Digital Game-Based Learning for Enhancing Students' Motivation
Authors: Shafaq Rubab
Abstract:
The primary reason for dropping out of school is associated with students’ lack of motivation in class, especially in mathematics. Digital game-based learning is an approach that is being actively explored; there are very few learning games based on proven instructional design models or frameworks due to which the effectiveness of the learning games suffers. The purpose of this research was twofold: first, developing an appropriate instructional design model and second, evaluating the impact of the instructional design model on students’ motivation. This research contributes significantly to the existing literature in terms of student motivation and the impact of instructional design model in digital game-based learning. The sample size for this study consists of two hundred out-of-school students between the age of 6 and 12 years. The research methodology used for this research was a quasi-experimental approach and data was analyzed by using the instructional material motivational survey questionnaire which is adapted from the Keller Arcs model. Control and experimental groups consisting of two hundred students were analyzed by utilizing instructional material motivational survey (IMMS), and comparison of result from both groups showed the difference in the level of motivation of the students. The result of the research showed that the motivational level of student in the experimental group who were taught by the game was higher than the student in control group (taught by conventional methodology). The mean score of the experimental group against all subscales (attention, relevance, confidence, and satisfaction) of IMMS survey was higher; however, no statistical significance was found between the motivational scores of control and experimental group. The positive impact of game-based learning on students’ level of motivation, as measured in this study, strengthens the case for the use of pedagogically sound instructional design models in the design of interactive learning applications. In addition, the present study suggests learning from interactive, immersive applications as an alternative solution for children, especially in Third World countries, who, for various reasons, do not attend school. The mean score of experimental group against all subscales of IMMS survey was higher; however, no statistical significance was found between motivational scores of control and experimental group.Keywords: digital game-based learning, students’ motivation, and instructional designing, instructional material motivational survey
Procedia PDF Downloads 418861 Sustainable Geographic Information System-Based Map for Suitable Landfill Sites in Aley and Chouf, Lebanon
Authors: Allaw Kamel, Bazzi Hasan
Abstract:
Municipal solid waste (MSW) generation is among the most significant sources which threaten the global environmental health. Solid Waste Management has been an important environmental problem in developing countries because of the difficulties in finding sustainable solutions for solid wastes. Therefore, more efforts are needed to be implemented to overcome this problem. Lebanon has suffered a severe solid waste management problem in 2015, and a new landfill site was proposed to solve the existing problem. The study aims to identify and locate the most suitable area to construct a landfill taking into consideration the sustainable development to overcome the present situation and protect the future demands. Throughout the article, a landfill site selection methodology was discussed using Geographic Information System (GIS) and Multi Criteria Decision Analysis (MCDA). Several environmental, economic and social factors were taken as criterion for selection of a landfill. Soil, geology, and LUC (Land Use and Land Cover) indices with the Sustainable Development Index were main inputs to create the final map of Environmentally Sensitive Area (ESA) for landfill site. Different factors were determined to define each index. Input data of each factor was managed, visualized and analyzed using GIS. GIS was used as an important tool to identify suitable areas for landfill. Spatial Analysis (SA), Analysis and Management GIS tools were implemented to produce input maps capable of identifying suitable areas related to each index. Weight has been assigned to each factor in the same index, and the main weights were assigned to each index used. The combination of the different indices map generates the final output map of ESA. The output map was reclassified into three suitability classes of low, moderate, and high suitability. Results showed different locations suitable for the construction of a landfill. Results also reflected the importance of GIS and MCDA in helping decision makers finding a solution of solid wastes by a sanitary landfill.Keywords: sustainable development, landfill, municipal solid waste (MSW), geographic information system (GIS), multi criteria decision analysis (MCDA), environmentally sensitive area (ESA)
Procedia PDF Downloads 148860 A Stepwise Approach for Piezoresistive Microcantilever Biosensor Optimization
Authors: Amal E. Ahmed, Levent Trabzon
Abstract:
Due to the low concentration of the analytes in biological samples, the use of Biological Microelectromechanical System (Bio-MEMS) biosensors for biomolecules detection results in a minuscule output signal that is not good enough for practical applications. In response to this, a need has arisen for an optimized biosensor capable of giving high output signal in response the detection of few analytes in the sample; the ultimate goal is being able to convert the attachment of a single biomolecule into a measurable quantity. For this purpose, MEMS microcantilevers based biosensors emerged as a promising sensing solution because it is simple, cheap, very sensitive and more importantly does not need analytes optical labeling (Label-free). Among the different microcantilever transducing techniques, piezoresistive based microcantilever biosensors became more prominent because it works well in liquid environments and has an integrated readout system. However, the design of piezoresistive microcantilevers is not a straightforward problem due to coupling between the design parameters, constraints, process conditions, and performance. It was found that the parameters that can be optimized to enhance the sensitivity of Piezoresistive microcantilever-based sensors are: cantilever dimensions, cantilever material, cantilever shape, piezoresistor material, piezoresistor doping level, piezoresistor dimensions, piezoresistor position, Stress Concentration Region's (SCR) shape and position. After a systematic analyzation of the effect of each design and process parameters on the sensitivity, a step-wise optimization approach was developed in which almost all these parameters were variated one at each step while fixing the others to get the maximum possible sensitivity at the end. At each step, the goal was to optimize the parameter in a way that it maximizes and concentrates the stress in the piezoresistor region for the same applied force thus get the higher sensitivity. Using this approach, an optimized sensor that has 73.5x times higher electrical sensitivity (ΔR⁄R) than the starting sensor was obtained. In addition to that, this piezoresistive microcantilever biosensor it is more sensitive than the other similar sensors previously reported in the open literature. The mechanical sensitivity of the final senior is -1.5×10-8 Ω/Ω ⁄pN; which means that for each 1pN (10-10 g) biomolecules attach to this biosensor; the piezoresistor resistivity will decrease by 1.5×10-8 Ω. Throughout this work COMSOL Multiphysics 5.0, a commercial Finite Element Analysis (FEA) tool, has been used to simulate the sensor performance.Keywords: biosensor, microcantilever, piezoresistive, stress concentration region (SCR)
Procedia PDF Downloads 569859 Algae Biofertilizers Promote Sustainable Food Production and Nutrient Efficiency: An Integrated Empirical-Modeling Study
Authors: Zeenat Rupawalla, Nicole Robinson, Susanne Schmidt, Sijie Li, Selina Carruthers, Elodie Buisset, John Roles, Ben Hankamer, Juliane Wolf
Abstract:
Agriculture has radically changed the global biogeochemical cycle of nitrogen (N). Fossil fuel-enabled synthetic N-fertiliser is a foundation of modern agriculture but applied to soil crops only use about half of it. To address N-pollution from cropping and the large carbon and energy footprint of N-fertiliser synthesis, new technologies delivering enhanced energy efficiency, decarbonisation, and a circular nutrient economy are needed. We characterised algae fertiliser (AF) as an alternative to synthetic N-fertiliser (SF) using empirical and modelling approaches. We cultivated microalgae in nutrient solution and modelled up-scaled production in nutrient-rich wastewater. Over four weeks, AF released 63.5% of N as ammonium and nitrate, and 25% of phosphorous (P) as phosphate to the growth substrate, while SF released 100% N and 20% P. To maximise crop N-use and minimise N-leaching, we explored AF and SF dose-response-curves with spinach in glasshouse conditions. AF-grown spinach produced 36% less biomass than SF-grown plants due to AF’s slower and linear N-release, while SF resulted in 5-times higher N-leaching loss than AF. Optimised blends of AF and SF boosted crop yield and minimised N-loss due to greater synchrony of N-release and crop uptake. Additional benefits of AF included greener leaves, lower leaf nitrate concentration, and higher microbial diversity and water holding capacity in the growth substrate. Life-cycle-analysis showed that replacing the most effective SF dosage with AF lowered the carbon footprint of fertiliser production from 2.02 g CO₂ (C-producing) to -4.62 g CO₂ (C-sequestering), with a further 12% reduction when AF is produced on wastewater. Embodied energy was lowest for AF-SF blends and could be reduced by 32% when cultivating algae on wastewater. We conclude that (i) microalgae offer a sustainable alternative to synthetic N-fertiliser in spinach production and potentially other crop systems, and (ii) microalgae biofertilisers support the circular nutrient economy and several sustainable development goals.Keywords: bioeconomy, decarbonisation, energy footprint, microalgae
Procedia PDF Downloads 136858 Handy EKG: Low-Cost ECG For Primary Care Screening In Developing Countries
Authors: Jhiamluka Zservando Solano Velasquez, Raul Palma, Alejandro Calderon, Servio Paguada, Erick Marin, Kellyn Funes, Hana Sandoval, Oscar Hernandez
Abstract:
Background: Screening cardiac conditions in primary care in developing countries can be challenging, and Honduras is not the exception. One of the main limitations is the underfunding of the Healthcare System in general, causing conventional ECG acquisition to become a secondary priority. Objective: Development of a low-cost ECG to improve screening of arrhythmias in primary care and communication with a specialist in secondary and tertiary care. Methods: Design a portable, pocket-size low-cost 3 lead ECG (Handy EKG). The device is autonomous and has Wi-Fi/Bluetooth connectivity options. A mobile app was designed which can access online servers with machine learning, a subset of artificial intelligence to learn from the data and aid clinicians in their interpretation of readings. Additionally, the device would use the online servers to transfer patient’s data and readings to a specialist in secondary and tertiary care. 50 randomized patients volunteer to participate to test the device. The patients had no previous cardiac-related conditions, and readings were taken. One reading was performed with the conventional ECG and 3 readings with the Handy EKG using different lead positions. This project was possible thanks to the funding provided by the National Autonomous University of Honduras. Results: Preliminary results show that the Handy EKG performs readings of the cardiac activity similar to those of a conventional electrocardiograph in lead I, II, and III depending on the position of the leads at a lower cost. The wave and segment duration, amplitude, and morphology of the readings were similar to the conventional ECG, and interpretation was possible to conclude whether there was an arrhythmia or not. Two cases of prolonged PR segment were found in both ECG device readings. Conclusion: Using a Frugal innovation approach can allow lower income countries to develop innovative medical devices such as the Handy EKG to fulfill unmet needs at lower prices without compromising effectiveness, safety, and quality. The Handy EKG provides a solution for primary care screening at a much lower cost and allows for convenient storage of the readings in online servers where clinical data of patients can then be accessed remotely by Cardiology specialists.Keywords: low-cost hardware, portable electrocardiograph, prototype, remote healthcare
Procedia PDF Downloads 179857 Linguistic Cyberbullying, a Legislative Approach
Authors: Simona Maria Ignat
Abstract:
Bullying online has been an increasing studied topic during the last years. Different approaches, psychological, linguistic, or computational, have been applied. To our best knowledge, a definition and a set of characteristics of phenomenon agreed internationally as a common framework are still waiting for answers. Thus, the objectives of this paper are the identification of bullying utterances on Twitter and their algorithms. This research paper is focused on the identification of words or groups of words, categorized as “utterances”, with bullying effect, from Twitter platform, extracted on a set of legislative criteria. This set is the result of analysis followed by synthesis of law documents on bullying(online) from United States of America, European Union, and Ireland. The outcome is a linguistic corpus with approximatively 10,000 entries. The methods applied to the first objective have been the following. The discourse analysis has been applied in identification of keywords with bullying effect in texts from Google search engine, Images link. Transcription and anonymization have been applied on texts grouped in CL1 (Corpus linguistics 1). The keywords search method and the legislative criteria have been used for identifying bullying utterances from Twitter. The texts with at least 30 representations on Twitter have been grouped. They form the second corpus linguistics, Bullying utterances from Twitter (CL2). The entries have been identified by using the legislative criteria on the the BoW method principle. The BoW is a method of extracting words or group of words with same meaning in any context. The methods applied for reaching the second objective is the conversion of parts of speech to alphabetical and numerical symbols and writing the bullying utterances as algorithms. The converted form of parts of speech has been chosen on the criterion of relevance within bullying message. The inductive reasoning approach has been applied in sampling and identifying the algorithms. The results are groups with interchangeable elements. The outcomes convey two aspects of bullying: the form and the content or meaning. The form conveys the intentional intimidation against somebody, expressed at the level of texts by grammatical and lexical marks. This outcome has applicability in the forensic linguistics for establishing the intentionality of an action. Another outcome of form is a complex of graphemic variations essential in detecting harmful texts online. This research enriches the lexicon already known on the topic. The second aspect, the content, revealed the topics like threat, harassment, assault, or suicide. They are subcategories of a broader harmful content which is a constant concern for task forces and legislators at national and international levels. These topic – outcomes of the dataset are a valuable source of detection. The analysis of content revealed algorithms and lexicons which could be applied to other harmful contents. A third outcome of content are the conveyances of Stylistics, which is a rich source of discourse analysis of social media platforms. In conclusion, this corpus linguistics is structured on legislative criteria and could be used in various fields.Keywords: corpus linguistics, cyberbullying, legislation, natural language processing, twitter
Procedia PDF Downloads 85856 Investigating Acute and Chronic Pain after Bariatric Surgery
Authors: Patti Kastanias, Wei Wang, Karyn Mackenzie, Sandra Robinson, Susan Wnuk
Abstract:
Obesity is a worldwide epidemic and is recognized as a chronic disease. Pain in the obese individual is a multidimensional issue. An increase in BMI is positively correlated with pain incidence and severity, especially in central obesity where individuals are twice as likely to have chronic pain. Both obesity and chronic pain are also associated with mood disorders. Pain is worse among obese individuals with depression and anxiety. Bariatric surgery provides patients with an effective solution for long-term weight loss and associated health problems. However, not much is known about acute and chronic pain after bariatric surgery and its contributing factors, including mood disorders. Nurse practitioners (NPs) at one large multidisciplinary bariatric surgery centre led two studies to examine acute and chronic pain and pain management over time after bariatric surgery. The purpose of the initial study was to examine the incidence and severity of acute and chronic pain after bariatric surgery. The aim of the secondary study was to further examine chronic pain, specifically looking at psychological factors that influence severity or incidence of both neuropathic and somatic pain as well as changes in opioid use. The initial study was a prospective, longitudinal study where patients having bariatric surgery at one surgical center were followed up to 6 months postop. Data was collected at 7 time points using validated instruments for pain severity, pain interference, and patient satisfaction. In the second study, subjects were followed longitudinally starting preoperatively and then at 6 months and 1 year postoperatively to capture changes in chronic pain and influencing variables over time. Valid and reliable instruments were utilized for all major study outcomes. In the first study, there was a trend towards decreased acute post-operative pain over time. The incidence and severity of chronic pain was found to be significantly reduced at 6 months post bariatric surgery. Interestingly, interference of chronic pain in daily life such as normal work, mood, and walking ability was significantly improved at 6 months postop however; this was not the case with sleep. Preliminary results of the secondary study indicate that pain severity, pain interference, anxiety and depression are significantly improved at 6 months postoperatively. In addition, preoperative anxiety, depression and emotional regulation were predictive of pain interference, but not pain severity. The results of our regression analyses provide evidence for the impact of pre-existing psychological factors on pain, particularly anxiety in obese populations.Keywords: bariatric surgery, mood disorders, obesity, pain
Procedia PDF Downloads 303855 AI Predictive Modeling of Excited State Dynamics in OPV Materials
Authors: Pranav Gunhal., Krish Jhurani
Abstract:
This study tackles the significant computational challenge of predicting excited state dynamics in organic photovoltaic (OPV) materials—a pivotal factor in the performance of solar energy solutions. Time-dependent density functional theory (TDDFT), though effective, is computationally prohibitive for larger and more complex molecules. As a solution, the research explores the application of transformer neural networks, a type of artificial intelligence (AI) model known for its superior performance in natural language processing, to predict excited state dynamics in OPV materials. The methodology involves a two-fold process. First, the transformer model is trained on an extensive dataset comprising over 10,000 TDDFT calculations of excited state dynamics from a diverse set of OPV materials. Each training example includes a molecular structure and the corresponding TDDFT-calculated excited state lifetimes and key electronic transitions. Second, the trained model is tested on a separate set of molecules, and its predictions are rigorously compared to independent TDDFT calculations. The results indicate a remarkable degree of predictive accuracy. Specifically, for a test set of 1,000 OPV materials, the transformer model predicted excited state lifetimes with a mean absolute error of 0.15 picoseconds, a negligible deviation from TDDFT-calculated values. The model also correctly identified key electronic transitions contributing to the excited state dynamics in 92% of the test cases, signifying a substantial concordance with the results obtained via conventional quantum chemistry calculations. The practical integration of the transformer model with existing quantum chemistry software was also realized, demonstrating its potential as a powerful tool in the arsenal of materials scientists and chemists. The implementation of this AI model is estimated to reduce the computational cost of predicting excited state dynamics by two orders of magnitude compared to conventional TDDFT calculations. The successful utilization of transformer neural networks to accurately predict excited state dynamics provides an efficient computational pathway for the accelerated discovery and design of new OPV materials, potentially catalyzing advancements in the realm of sustainable energy solutions.Keywords: transformer neural networks, organic photovoltaic materials, excited state dynamics, time-dependent density functional theory, predictive modeling
Procedia PDF Downloads 116854 Utilization of Sludge in the Manufacturing of Fired Clay Bricks
Authors: Anjali G. Pillai, S. Chadrakaran
Abstract:
The extensive amount of sludge generated throughout the world, as a part of water treatment works, have caused various social and economic issues, such as a demand on landfill spaces, increase in environmental pollution and raising the waste management cost. With growing social awareness about toxic incinerator emissions and the increasing concern over the disposal of sludge on the agricultural land, the recovery of sewage sludge as a building and construction raw material can be considered as an innovative approach to tackle the sludge disposal problem. The proposed work aims at studying the recycling ability of the sludge, generated from the water treatment process, by incorporating it into the fired clay brick units. The work involves initial study of the geotechnical characteristics of the brick-clay and the sludge. Chemical compatibility of both the materials will be analyzed by X-ray fluorescence technique. The variation in the strength aspects with varying proportions of sludge i.e. 10%, 20%, 30% and 40% in the sludge-clay mix will also be determined by the proctor density test. Based on the optimum moisture content, the sludge-clay bricks will be manufactured in a brick manufacturing plant and the modified brick units will be tested to determine the variation in compressive strength, bulk density, firing shrinkage, shrinkage loss and initial water absorption rate with respect to the conventional clay bricks. The results will be compared with the specifications given in Indian Standards to arrive at the potential use of the new bricks. The durability aspect will be studied by conducting the leachate analysis test using atomic adsorption spectrometry. The lightweight characteristics of the sludge modified bricks will be ascertained with the scanning electron microscope technique which will be indicative of the variation in pore structure with the increase in sludge content within the bricks. The work will determine the suitable proportion of the sludge – clay mix in the brick which can then be effectively implemented. The feasibility aspect of the work will be determined for commercial production of the units. The work involves providing a strategy for conversion of waste to resource. Moreover, it provides an alternative solution to the problem of growing scarcity of brick-clay for the manufacturing of fired clay bricks.Keywords: eco-bricks, green construction material, sludge amended bricks, sludge disposal, waste management
Procedia PDF Downloads 304