Search results for: visual design element
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 15809

Search results for: visual design element

7679 Regional Flood Frequency Analysis in Narmada Basin: A Case Study

Authors: Ankit Shah, R. K. Shrivastava

Abstract:

Flood and drought are two main features of hydrology which affect the human life. Floods are natural disasters which cause millions of rupees’ worth of damage each year in India and the whole world. Flood causes destruction in form of life and property. An accurate estimate of the flood damage potential is a key element to an effective, nationwide flood damage abatement program. Also, the increase in demand of water due to increase in population, industrial and agricultural growth, has let us know that though being a renewable resource it cannot be taken for granted. We have to optimize the use of water according to circumstances and conditions and need to harness it which can be done by construction of hydraulic structures. For their safe and proper functioning of hydraulic structures, we need to predict the flood magnitude and its impact. Hydraulic structures play a key role in harnessing and optimization of flood water which in turn results in safe and maximum use of water available. Mainly hydraulic structures are constructed on ungauged sites. There are two methods by which we can estimate flood viz. generation of Unit Hydrographs and Flood Frequency Analysis. In this study, Regional Flood Frequency Analysis has been employed. There are many methods for estimating the ‘Regional Flood Frequency Analysis’ viz. Index Flood Method. National Environmental and Research Council (NERC Methods), Multiple Regression Method, etc. However, none of the methods can be considered universal for every situation and location. The Narmada basin is located in Central India. It is drained by most of the tributaries, most of which are ungauged. Therefore it is very difficult to estimate flood on these tributaries and in the main river. As mentioned above Artificial Neural Network (ANN)s and Multiple Regression Method is used for determination of Regional flood Frequency. The annual peak flood data of 20 sites gauging sites of Narmada Basin is used in the present study to determine the Regional Flood relationships. Homogeneity of the considered sites is determined by using the Index Flood Method. Flood relationships obtained by both the methods are compared with each other, and it is found that ANN is more reliable than Multiple Regression Method for the present study area.

Keywords: artificial neural network, index flood method, multi layer perceptrons, multiple regression, Narmada basin, regional flood frequency

Procedia PDF Downloads 396
7678 PSS and SVC Controller Design by BFA to Enhance the Power System Stability

Authors: Saeid Jalilzadeh

Abstract:

Designing of PSS and SVC controller based on Bacterial Foraging Algorithm (BFA) to improve the stability of power system is proposed in this paper. Same controllers for PSS and SVC has been considered and Single machine infinite bus (SMIB) system with SVC located at the terminal of generator is used to evaluate the proposed controllers. BFA is used to optimize the coefficients of the controllers. Finally simulation for a special disturbance as an input power of generator with the proposed controllers in order to investigate the dynamic behavior of generator is done. The simulation results demonstrate that the system composed with optimized controllers has an outstanding operation in fast damping of oscillations of power system.

Keywords: PSS, SVC, SMIB, optimize controller

Procedia PDF Downloads 436
7677 Using Optical Character Recognition to Manage the Unstructured Disaster Data into Smart Disaster Management System

Authors: Dong Seop Lee, Byung Sik Kim

Abstract:

In the 4th Industrial Revolution, various intelligent technologies have been developed in many fields. These artificial intelligence technologies are applied in various services, including disaster management. Disaster information management does not just support disaster work, but it is also the foundation of smart disaster management. Furthermore, it gets historical disaster information using artificial intelligence technology. Disaster information is one of important elements of entire disaster cycle. Disaster information management refers to the act of managing and processing electronic data about disaster cycle from its’ occurrence to progress, response, and plan. However, information about status control, response, recovery from natural and social disaster events, etc. is mainly managed in the structured and unstructured form of reports. Those exist as handouts or hard-copies of reports. Such unstructured form of data is often lost or destroyed due to inefficient management. It is necessary to manage unstructured data for disaster information. In this paper, the Optical Character Recognition approach is used to convert handout, hard-copies, images or reports, which is printed or generated by scanners, etc. into electronic documents. Following that, the converted disaster data is organized into the disaster code system as disaster information. Those data are stored in the disaster database system. Gathering and creating disaster information based on Optical Character Recognition for unstructured data is important element as realm of the smart disaster management. In this paper, Korean characters were improved to over 90% character recognition rate by using upgraded OCR. In the case of character recognition, the recognition rate depends on the fonts, size, and special symbols of character. We improved it through the machine learning algorithm. These converted structured data is managed in a standardized disaster information form connected with the disaster code system. The disaster code system is covered that the structured information is stored and retrieve on entire disaster cycle such as historical disaster progress, damages, response, and recovery. The expected effect of this research will be able to apply it to smart disaster management and decision making by combining artificial intelligence technologies and historical big data.

Keywords: disaster information management, unstructured data, optical character recognition, machine learning

Procedia PDF Downloads 111
7676 Value Co-Creation in Used-Car Auctions: A Service Scientific Perspective

Authors: Safdar Muhammad Usman, Youji Kohda, Katsuhiro Umemoto

Abstract:

Electronic market place plays an important intermediary role for connecting dealers and retail customers. The main aim of this paper is to design a value co-creation model in used-car auctions. More specifically, the study has been designed in order to describe the process of value co-creation in used-car auctions, to explore the co-created values in used-car auctions, and finally conclude the paper indicating the future research directions. Our analysis shows that economic values as well as non-economic values are co-created in used-car auctions. In addition, this paper contributes to the academic society broadening the view of value co-creation in service science.

Keywords: value co-creation, used-car auctions, non-financial values, service science

Procedia PDF Downloads 340
7675 Microbiological Analysis on Anatomical Specimens of Cats for Use in Veterinary Surgery

Authors: Raphael C. Zero, Marita V. Cardozo, Thiago A. S. S. Rocha, Mariana T. Kihara, Fernando A. Ávila, Fabrício S. Oliveira

Abstract:

There are several fixative and preservative solutions for use on cadavers, many of them using formaldehyde as the fixative or anatomical part preservative. In some countries, such as Brazil, this toxic agent has been increasingly restricted. The objective of this study was to microbiologically identify and quantify the key agents in tanks containing 96GL ethanol or sodium chloride solutions, used respectively as fixatives and preservatives of cat cadavers. Eight adult cat corpses, three females and five males, with an average weight of 4.3 kg, were used. After injection via the external common carotid artery (120 ml/kg, 95% 96GL ethyl alcohol and 5% pure glycerin), the cadavers were fixed in a plastic tank with 96GL ethanol for 60 days. After fixing, they were stored in a 30% sodium chloride aqueous solution for 120 days in a similar tank. Samples were collected at the start of the experiment - before the animals were placed in the ethanol tanks, and monthly thereafter. The bacterial count was performed by Pour Plate Method in BHI agar (Brain Heart Infusion) and the plates were incubated aerobically and anaerobically for 24h at 37ºC. MacConkey agar, SPS agar (Sulfite Polymyxin Sulfadizine) and MYP Agar Base were used to isolate the microorganisms. There was no microbial growth in the samples prior to alcohol fixation. After 30 days of fixation in the alcohol solution, total aerobic and anaerobic (<1.0 x 10 CFU/ml) were found and Pseudomonas sp., Staphylococcus sp., Clostridium sp. were the identified agents. After 60 days in the alcohol fixation solution, total aerobes (<1.0 x 10 CFU/ml) and total anaerobes (<2.2 x 10 CFU/mL) were found, and the identified agents were the same. After 30 days of storage in the aqueous solution of 30% sodium chloride, total aerobic (<5.2 x 10 CFU/ml) and total anaerobes (<3.7 x 10 CFU/mL) were found and the agents identified were Staphylococcus sp., Clostridium sp., and fungi. After 60 days of sodium chloride storage, total aerobic (<3.0 x 10 CFU / ml) and total anaerobes (<7.0 x 10 CFU/mL) were found and the identified agents remained the same: Staphylococcus sp., Clostridium sp., and fungi. The microbiological count was low and visual inspection did not reveal signs of contamination in the tanks. There was no strong odor or purification, which proved the technique to be microbiologically effective in fixing and preserving the cat cadavers for the four-month period in which they are provided to undergraduate students of University of Veterinary Medicine for surgery practice. All experimental procedures were approved by the Municipal Legal Department (protocol 02.2014.000027-1). The project was funded by FAPESP (protocol 2015-08259-9).

Keywords: anatomy, fixation, microbiology, small animal, surgery

Procedia PDF Downloads 264
7674 On the Utility of Bidirectional Transformers in Gene Expression-Based Classification

Authors: Babak Forouraghi

Abstract:

A genetic circuit is a collection of interacting genes and proteins that enable individual cells to implement and perform vital biological functions such as cell division, growth, death, and signaling. In cell engineering, synthetic gene circuits are engineered networks of genes specifically designed to implement functionalities that are not evolved by nature. These engineered networks enable scientists to tackle complex problems such as engineering cells to produce therapeutics within the patient's body, altering T cells to target cancer-related antigens for treatment, improving antibody production using engineered cells, tissue engineering, and production of genetically modified plants and livestock. Construction of computational models to realize genetic circuits is an especially challenging task since it requires the discovery of the flow of genetic information in complex biological systems. Building synthetic biological models is also a time-consuming process with relatively low prediction accuracy for highly complex genetic circuits. The primary goal of this study was to investigate the utility of a pre-trained bidirectional encoder transformer that can accurately predict gene expressions in genetic circuit designs. The main reason behind using transformers is their innate ability (attention mechanism) to take account of the semantic context present in long DNA chains that are heavily dependent on the spatial representation of their constituent genes. Previous approaches to gene circuit design, such as CNN and RNN architectures, are unable to capture semantic dependencies in long contexts, as required in most real-world applications of synthetic biology. For instance, RNN models (LSTM, GRU), although able to learn long-term dependencies, greatly suffer from vanishing gradient and low-efficiency problem when they sequentially process past states and compresses contextual information into a bottleneck with long input sequences. In other words, these architectures are not equipped with the necessary attention mechanisms to follow a long chain of genes with thousands of tokens. To address the above-mentioned limitations, a transformer model was built in this work as a variation to the existing DNA Bidirectional Encoder Representations from Transformers (DNABERT) model. It is shown that the proposed transformer is capable of capturing contextual information from long input sequences with an attention mechanism. In previous works on genetic circuit design, the traditional approaches to classification and regression, such as Random Forrest, Support Vector Machine, and Artificial Neural Networks, were able to achieve reasonably high R2 accuracy levels of 0.95 to 0.97. However, the transformer model utilized in this work, with its attention-based mechanism, was able to achieve a perfect accuracy level of 100%. Further, it is demonstrated that the efficiency of the transformer-based gene expression classifier is not dependent on the presence of large amounts of training examples, which may be difficult to compile in many real-world gene circuit designs.

Keywords: machine learning, classification and regression, gene circuit design, bidirectional transformers

Procedia PDF Downloads 46
7673 Inverse Mapping of Weld Bead Geometry in Shielded Metal Arc-Welding: Genetic Algorithm Approach

Authors: D. S. Nagesh, G. L. Datta

Abstract:

In the field of welding, various studies had been made by some of the previous investigators to predict as well as optimize weld bead geometric descriptors. Modeling of weld bead shape is important for predicting the quality of welds. In most of the cases, design of experiments technique to postulate multiple linear regression equations have been used. Nowadays, Genetic Algorithm (GA) an intelligent information treatment system with the characteristics of treating complex relationships as seen in welding processes used as a tool for inverse mapping/optimization of the process is attempted.

Keywords: smaw, genetic algorithm, bead geometry, optimization/inverse mapping

Procedia PDF Downloads 438
7672 Genetic Algorithm Approach for Inverse Mapping of Weld Bead Geometry in Shielded Metal Arc-Welding

Authors: D. S. Nagesh, G. L. Datta

Abstract:

In the field of welding, various studies had been made by some of the previous investigators to predict as well as optimize weld bead geometric descriptors. Modeling of weld bead shape is important for predicting the quality of welds. In most of the cases design of experiments technique to postulate multiple linear regression equations have been used. Nowadays Genetic Algorithm (GA) an intelligent information treatment system with the characteristics of treating complex relationships as seen in welding processes used as a tool for inverse mapping/optimization of the process is attempted.

Keywords: SMAW, genetic algorithm, bead geometry, optimization/inverse mapping

Procedia PDF Downloads 410
7671 The Evolution of Architecture through Digital: A Survey on Fashion Catwalk Becoming Digital

Authors: Valeria Minucciani, Maria Maddalena Margaria

Abstract:

While mathematical tools that make digital architecture possible are very sophisticated and advanced, theoretical development of digital architecture (intended as a discipline that integrates or replaces the real architecture) is not. The fashion show, that involves interiors architecture, exhibit design and scenography, has been exploiting for ten years the opportunities offered by digital technologies. To gain greater visibility and to reach a wider audience, high-level experimentations have been performed. The aim of this paper is in investigating, through the analysis of some cases of virtual fashion shows, the 'architectural' impact of the virtual conception of interior space.

Keywords: digital interiors, exhibit, fashion catwalk, architectural theory

Procedia PDF Downloads 401
7670 An Experimental Study of Diffuser-Enhanced Propeller Hydrokinetic Turbines

Authors: Matheus Nunes, Rafael Mendes, Taygoara Felamingo Oliveira, Antonio Brasil Junior

Abstract:

Wind tunnel experiments of horizontal axis propeller hydrokinetic turbines model were carried out, in order to determine the performance behavior for different configurations and operational range. The present experiments introduce the use of two different geometries of rear diffusers to enhance the performance of the free flow machine. The present paper reports an increase of the power coefficient about 50%-80%. It represents an important feature that has to be taken into account in the design of this kind of machine.

Keywords: diffuser-enhanced turbines, hydrokinetic turbine, wind tunnel experiments, micro hydro

Procedia PDF Downloads 251
7669 Research on the Conservation Strategy of Territorial Landscape Based on Characteristics: The Case of Fujian, China

Authors: Tingting Huang, Sha Li, Geoffrey Griffiths, Martin Lukac, Jianning Zhu

Abstract:

Territorial landscapes have experienced a gradual loss of their typical characteristics during long-term human activities. In order to protect the integrity of regional landscapes, it is necessary to characterize, evaluate and protect them in a graded manner. The study takes Fujian, China, as an example and classifies the landscape characters of the site at the regional scale, middle scale, and detailed scale. A multi-scale approach combining parametric and holistic approaches is used to classify and partition the landscape character types (LCTs) and landscape character areas (LCAs) at different scales, and a multi-element landscape assessment approach is adopted to explore the conservation strategies of the landscape character. Firstly, multiple fields and multiple elements of geography, nature and humanities were selected as the basis of assessment according to the scales. Secondly, the study takes a parametric approach to the classification and partitioning of landscape character, Principal Component Analysis, and two-stage cluster analysis (K-means and GMM) in MATLAB software to obtain LCTs, combines with Canny Operator Edge Detection Algorithm to obtain landscape character contours and corrects LCTs and LCAs by field survey and manual identification methods. Finally, the study adopts the Landscape Sensitivity Assessment method to perform landscape character conservation analysis and formulates five strategies for different LCAs: conservation, enhancement, restoration, creation, and combination. This multi-scale identification approach can efficiently integrate multiple types of landscape character elements, reduce the difficulty of broad-scale operations in the process of landscape character conservation, and provide a basis for landscape character conservation strategies. Based on the natural background and the restoration of regional characteristics, the results of landscape character assessment are scientific and objective and can provide a strong reference in regional and national scale territorial spatial planning.

Keywords: parameterization, multi-scale, landscape character identify, landscape character assessment

Procedia PDF Downloads 76
7668 Influence of Temperature on Properties of MOSFETs

Authors: Azizi Cherifa, O. Benzaoui

Abstract:

The thermal aspects in the design of power circuits often deserve as much attention as pure electric components aspects as the operating temperature has a direct influence on their static and dynamic characteristics. MOSFET is fundamental in the circuits, it is the most widely used device in the current production of semiconductor components using their honorable performance. The aim of this contribution is devoted to the effect of the temperature on the properties of MOSFETs. The study enables us to calculate the drain current as function of bias in both linear and saturated modes. The effect of temperature is evaluated using a numerical simulation, using the laws of mobility and saturation velocity of carriers as a function of temperature.

Keywords: temperature, MOSFET, mobility, transistor

Procedia PDF Downloads 332
7667 The Significance of Urban Space in Death Trilogy of Alejandro González Iñárritu

Authors: Marta Kaprzyk

Abstract:

The cinema of Alejandro González Iñárritu hasn’t been subjected to a lot of detailed analysis yet, what makes it an exceptionally interesting research material. The purpose of this presentation is to discuss the significance of urban space in three films of this Mexican director, that forms Death Trilogy: ‘Amores Perros’ (2000), ‘21 Grams’ (2003) and ‘Babel’ (2006). The fact that in the aforementioned movies the urban space itself becomes an additional protagonist with its own identity, psychology and the ability to transform and affect other characters, in itself warrants for independent research and analysis. Independently, such mode of presenting urban space has another function; it enables the director to complement the rest of characters. The basis for methodology of this description of cinematographic space is to treat its visual layer as a point of departure for a detailed analysis. At the same time, the analysis itself will be supported by recognised academic theories concerning special issues, which are transformed here into essential tools necessary to describe the world (mise-en-scène) created by González Iñárritu. In ‘Amores perros’ the Mexico City serves as a scenery – a place full of contradictions- in the movie depicted as a modern conglomerate and an urban jungle, as well as a labyrinth of poverty and violence. In this work stylistic tropes can be found in an intertextual dialogue of the director with photographies of Nan Goldin and Mary Ellen Mark. The story recounted in ‘21 Grams’, the most tragic piece in the trilogy, is characterised by almost hyperrealistic sadism. It takes place in Memphis, which on the screen turns into an impersonal formation full of heterotopias described by Michel Foucault and non-places, as defined by Marc Augé in his essay. By contrast, the main urban space in ‘Babel’ is Tokio, which seems to perfectly correspond with the image of places discussed by Juhani Pallasmaa in his works concerning the reception of the architecture by ‘pathological senses’ in the modern (or, even more adequately, postmodern) world. It’s portrayed as a city full of buildings that look so surreal, that they seem to be completely unsuitable for the humans to move between them. Ultimately, the aim of this paper is to demonstrate the coherence of the manner in which González Iñárritu designs urban spaces in his Death Trilogy. In particular, the author attempts to examine the imperative role of the cities that form three specific microcosms in which the protagonists of the Mexican director live their overwhelming tragedies.

Keywords: cinematographic space, Death Trilogy, film Studies, González Iñárritu Alejandro, urban space

Procedia PDF Downloads 312
7666 Design, Construction and Performance Evaluation of a HPGe Detector Shield

Authors: M. Sharifi, M. Mirzaii, F. Bolourinovin, H. Yousefnia, M. Akbari, K. Yousefi-Mojir

Abstract:

A multilayer passive shield composed of low-activity lead (Pb), copper (Cu), tin (Sn) and iron (Fe) was designed and manufactured for a coaxial HPGe detector placed at a surface laboratory for reducing background radiation and radiation dose to the personnel. The performance of the shield was evaluated and efficiency curves of the detector were plotted by using of the various standard sources in different distances. Monte Carlo simulations and a set of TLD chips were used for dose estimation in two distances of 20 and 40 cm. The results show that the shield reduced background spectrum and the personnel dose more than 95%.

Keywords: HPGe shield, background count, personnel dose, efficiency curve

Procedia PDF Downloads 438
7665 Two Kinds of Self-Oscillating Circuits Mechanically Demonstrated

Authors: Shiang-Hwua Yu, Po-Hsun Wu

Abstract:

This study introduces two types of self-oscillating circuits that are frequently found in power electronics applications. Special effort is made to relate the circuits to the analogous mechanical systems of some important scientific inventions: Galileo’s pendulum clock and Coulomb’s friction model. A little touch of related history and philosophy of science will hopefully encourage curiosity, advance the understanding of self-oscillating systems and satisfy the aspiration of some students for scientific literacy. Finally, the two self-oscillating circuits are applied to design a simple class-D audio amplifier.

Keywords: self-oscillation, sigma-delta modulator, pendulum clock, Coulomb friction, class-D amplifier

Procedia PDF Downloads 337
7664 Call-Back Laterality and Bilaterality: Possible Screening Mammography Quality Metrics

Authors: Samson Munn, Virginia H. Kim, Huija Chen, Sean Maldonado, Michelle Kim, Paul Koscheski, Babak N. Kalantari, Gregory Eckel, Albert Lee

Abstract:

In terms of screening mammography quality, neither the portion of reports that advise call-back imaging that should be bilateral versus unilateral nor how much the unilateral call-backs may appropriately diverge from 50–50 (left versus right) is known. Many factors may affect detection laterality: display arrangement, reflections preferentially striking one display location, hanging protocols, seating positions with respect to others and displays, visual field cuts, health, etc. The call-back bilateral fraction may reflect radiologist experience (not in our data) or confidence level. Thus, laterality and bilaterality of call-backs advised in screening mammography reports could be worthy quality metrics. Here, laterality data did not reveal a concern until drilling down to individuals. Bilateral screening mammogram report recommendations by five breast imaging, attending radiologists at Harbor-UCLA Medical Center (Torrance, California) 9/1/15--8/31/16 and 9/1/16--8/31/17 were retrospectively reviewed. Recommended call-backs for bilateral versus unilateral, and for left versus right, findings were counted. Chi-square (χ²) statistic was applied. Year 1: of 2,665 bilateral screening mammograms, reports of 556 (20.9%) recommended call-back, of which 99 (17.8% of the 556) were for bilateral findings. Of the 457 unilateral recommendations, 222 (48.6%) regarded the left breast. Year 2: of 2,106 bilateral screening mammograms, reports of 439 (20.8%) recommended call-back, of which 65 (14.8% of the 439) were for bilateral findings. Of the 374 unilateral recommendations, 182 (48.7%) regarded the left breast. Individual ranges of call-backs that were bilateral were 13.2–23.3%, 10.2–22.5%, and 13.6–17.9%, by year(s) 1, 2, and 1+2, respectively; these ranges were unrelated to experience level; the two-year mean was 15.8% (SD=1.9%). The lowest χ² p value of the group's sidedness disparities years 1, 2, and 1+2 was > 0.4. Regarding four individual radiologists, the lowest p value was 0.42. However, the fifth radiologist disfavored the left, with p values of 0.21, 0.19, and 0.07, respectively; that radiologist had the greatest number of years of experience. There was a concerning, 93% likelihood that bias against left breast findings evidenced by one of our radiologists was not random. Notably, very soon after the period under review, he retired, presented with leukemia, and died. We call for research to be done, particularly by large departments with many radiologists, of two possible, new, quality metrics in screening mammography: laterality and bilaterality. (Images, patient outcomes, report validity, and radiologist psychological confidence levels were not assessed. No intervention nor subsequent data collection was conducted. This uncomplicated collection of data and simple appraisal were not designed, nor had there been any intention to develop or contribute, to generalizable knowledge (per U.S. DHHS 45 CFR, part 46)).

Keywords: mammography, screening mammography, quality, quality metrics, laterality

Procedia PDF Downloads 145
7663 Experimental and Analytical Design of Rigid Pavement Using Geopolymer Concrete

Authors: J. Joel Bright, P. Peer Mohamed, M. Aswin SAangameshwaran

Abstract:

The increasing usage of concrete produces 80% of carbon dioxide in the atmosphere. Hence, this results in various environmental effects like global warming. The amount of the carbon dioxide released during the manufacture of OPC due to the calcination of limestone and combustion of fossil fuel is in the order of one ton for every ton of OPC produced. Hence, to minimize this Geo Polymer Concrete was introduced. Geo polymer concrete is produced with 0% cement, and hence, it is eco-friendly and it also uses waste product from various industries like thermal power plant, steel manufacturing plant, and paper waste materials. This research is mainly about using Geo polymer concrete for pavement which gives very high strength than conventional concrete and at the same time gives way for sustainable development.

Keywords: activator solution, GGBS, fly ash, metakaolin

Procedia PDF Downloads 448
7662 Standard Essential Patents for Artificial Intelligence Hardware and the Implications For Intellectual Property Rights

Authors: Wendy de Gomez

Abstract:

Standardization is a critical element in the ability of a society to reduce uncertainty, subjectivity, misrepresentation, and interpretation while simultaneously contributing to innovation. Technological standardization is critical to codify specific operationalization through legal instruments that provide rules of development, expectation, and use. In the current emerging technology landscape Artificial Intelligence (AI) hardware as a general use technology has seen incredible growth as evidenced from AI technology patents between 2012 and 2018 in the United States Patent Trademark Office (USPTO) AI dataset. However, as outlined in the 2023 United States Government National Standards Strategy for Critical and Emerging Technology the codification through standardization of emerging technologies such as AI has not kept pace with its actual technological proliferation. This gap has the potential to cause significant divergent possibilities for the downstream outcomes of AI in both the short and long term. This original empirical research provides an overview of the standardization efforts around AI in different geographies and provides a background to standardization law. It quantifies the longitudinal trend of Artificial Intelligence hardware patents through the USPTO AI dataset. It seeks evidence of existing Standard Essential Patents from these AI hardware patents through a text analysis of the Statement of patent history and the Field of the invention of these patents in Patent Vector and examines their determination as a Standard Essential Patent and their inclusion in existing AI technology standards across the four main AI standards bodies- European Telecommunications Standards Institute (ETSI); International Telecommunication Union (ITU)/ Telecommunication Standardization Sector (-T); Institute of Electrical and Electronics Engineers (IEEE); and the International Organization for Standardization (ISO). Once the analysis is complete the paper will discuss both the theoretical and operational implications of F/Rand Licensing Agreements for the owners of these Standard Essential Patents in the United States Court and Administrative system. It will conclude with an evaluation of how Standard Setting Organizations (SSOs) can work with SEP owners more effectively through various forms of Intellectual Property mechanisms such as patent pools.

Keywords: patents, artifical intelligence, standards, F/Rand agreements

Procedia PDF Downloads 61
7661 Design of Experiment for Optimizing Immunoassay Microarray Printing

Authors: Alex J. Summers, Jasmine P. Devadhasan, Douglas Montgomery, Brittany Fischer, Jian Gu, Frederic Zenhausern

Abstract:

Immunoassays have been utilized for several applications, including the detection of pathogens. Our laboratory is in the development of a tier 1 biothreat panel utilizing Vertical Flow Assay (VFA) technology for simultaneous detection of pathogens and toxins. One method of manufacturing VFA membranes is with non-contact piezoelectric dispensing, which provides advantages, such as low-volume and rapid dispensing without compromising the structural integrity of antibody or substrate. Challenges of this processinclude premature discontinuation of dispensing and misaligned spotting. Preliminary data revealed the Yp 11C7 mAb (11C7)reagent to exhibit a large angle of failure during printing which may have contributed to variable printing outputs. A Design of Experiment (DOE) was executed using this reagent to investigate the effects of hydrostatic pressure and reagent concentration on microarray printing outputs. A Nano-plotter 2.1 (GeSIM, Germany) was used for printing antibody reagents ontonitrocellulose membrane sheets in a clean room environment. A spotting plan was executed using Spot-Front-End software to dispense volumes of 11C7 reagent (20-50 droplets; 1.5-5 mg/mL) in a 6-test spot array at 50 target membrane locations. Hydrostatic pressure was controlled by raising the Pressure Compensation Vessel (PCV) above or lowering it below our current working level. It was hypothesized that raising or lowering the PCV 6 inches would be sufficient to cause either liquid accumulation at the tip or discontinue droplet formation. After aspirating 11C7 reagent, we tested this hypothesis under stroboscope.75% of the effective raised PCV height and of our hypothesized lowered PCV height were used. Humidity (55%) was maintained using an Airwin BO-CT1 humidifier. The number and quality of membranes was assessed after staining printed membranes with dye. The droplet angle of failure was recorded before and after printing to determine a “stroboscope score” for each run. The DOE set was analyzed using JMP software. Hydrostatic pressure and reagent concentration had a significant effect on the number of membranes output. As hydrostatic pressure was increased by raising the PCV 3.75 inches or decreased by lowering the PCV -4.5 inches, membrane output decreased. However, with the hydrostatic pressure closest to equilibrium, our current working level, membrane output, reached the 50-membrane target. As the reagent concentration increased from 1.5 to 5 mg/mL, the membrane output also increased. Reagent concentration likely effected the number of membrane output due to the associated dispensing volume needed to saturate the membranes. However, only hydrostatic pressure had a significant effect on stroboscope score, which could be due to discontinuation of dispensing, and thus the stroboscope check could not find a droplet to record. Our JMP predictive model had a high degree of agreement with our observed results. The JMP model predicted that dispensing the highest concentration of 11C7 at our current PCV working level would yield the highest number of quality membranes, which correlated with our results. Acknowledgements: This work was supported by the Chemical Biological Technologies Directorate (Contract # HDTRA1-16-C-0026) and the Advanced Technology International (Contract # MCDC-18-04-09-002) from the Department of Defense Chemical and Biological Defense program through the Defense Threat Reduction Agency (DTRA).

Keywords: immunoassay, microarray, design of experiment, piezoelectric dispensing

Procedia PDF Downloads 162
7660 Textile-Based Sensing System for Sleep Apnea Detection

Authors: Mary S. Ruppert-Stroescu, Minh Pham, Bruce Benjamin

Abstract:

Sleep apnea is a condition where a person stops breathing and can lead to cardiovascular disease, hypertension, and stroke. In the United States, approximately forty percent of overnight sleep apnea detection tests are cancelled. The purpose of this study was to develop a textile-based sensing system that acquires biometric signals relevant to cardiovascular health, to transmit them wirelessly to a computer, and to quantitatively assess the signals for sleep apnea detection. Patient interviews, literature review and market analysis defined a need for a device that ubiquitously integrated into the patient’s lifestyle. A multi-disciplinary research team of biomedical scientists, apparel designers, and computer engineers collaborated to design a textile-based sensing system that gathers EKG, Sp02, and respiration, then wirelessly transmits the signals to a computer in real time. The electronic components were assembled from existing hardware, the Health Kit which came pre-set with EKG and Sp02 sensors. The respiration belt was purchased separately and its electronics were built and integrated into the Health Kit mother board. Analog ECG signals were amplified and transmitted to the Arduino™ board where the signal was converted from analog into digital. By using textile electrodes, ECG lead-II was collected, and it reflected the electrical activity of the heart. Signals were collected when the subject was in sitting position and at sampling rate of 250 Hz. Because sleep apnea most often occurs in people with obese body types, prototypes were developed for a man’s size medium, XL, and XXL. To test user acceptance and comfort, wear tests were performed on 12 subjects. Results of the wear tests indicate that the knit fabric and t-shirt-like design were acceptable from both lifestyle and comfort perspectives. The airflow signal and respiration signal sensors return good signals regardless of movement intensity. Future study includes reconfiguring the hardware to a smaller size, developing the same type of garment for the female body, and further enhancing the signal quality.

Keywords: sleep apnea, sensors, electronic textiles, wearables

Procedia PDF Downloads 251
7659 Mapping Context, Roles, and Relations for Adjudicating Robot Ethics

Authors: Adam J. Bowen

Abstract:

Abstract— Should robots have rights or legal protections. Often debates concerning whether robots and AI should be afforded rights focus on conditions of personhood and the possibility of future advanced forms of AI satisfying particular intrinsic cognitive and moral attributes of rights-holding persons. Such discussions raise compelling questions about machine consciousness, autonomy, and value alignment with human interests. Although these are important theoretical concerns, especially from a future design perspective, they provide limited guidance for addressing the moral and legal standing of current and near-term AI that operate well below the cognitive and moral agency of human persons. Robots and AI are already being pressed into service in a wide range of roles, especially in healthcare and biomedical contexts. The design and large-scale implementation of robots in the context of core societal institutions like healthcare systems continues to rapidly develop. For example, we bring them into our homes, hospitals, and other care facilities to assist in care for the sick, disabled, elderly, children, or otherwise vulnerable persons. We enlist surgical robotic systems in precision tasks, albeit still human-in-the-loop technology controlled by surgeons. We also entrust them with social roles involving companionship and even assisting in intimate caregiving tasks (e.g., bathing, feeding, turning, medicine administration, monitoring, transporting). There have been advances to enable severely disabled persons to use robots to feed themselves or pilot robot avatars to work in service industries. As the applications for near-term AI increase and the roles of robots in restructuring our biomedical practices expand, we face pressing questions about the normative implications of human-robot interactions and collaborations in our collective worldmaking, as well as the moral and legal status of robots. This paper argues that robots operating in public and private spaces be afforded some protections as either moral patients or legal agents to establish prohibitions on robot abuse, misuse, and mistreatment. We already implement robots and embed them in our practices and institutions, which generates a host of human-to-machine and machine-to-machine relationships. As we interact with machines, whether in service contexts, medical assistance, or home health companions, these robots are first encountered in relationship to us and our respective roles in the encounter (e.g., surgeon, physical or occupational therapist, recipient of care, patient’s family, healthcare professional, stakeholder). This proposal aims to outline a framework for establishing limiting factors and determining the extent of moral or legal protections for robots. In doing so, it advocates for a relational approach that emphasizes the priority of mapping the complex contextually sensitive roles played and the relations in which humans and robots stand to guide policy determinations by relevant institutions and authorities. The relational approach must also be technically informed by the intended uses of the biomedical technologies in question, Design History Files, extensive risk assessments and hazard analyses, as well as use case social impact assessments.

Keywords: biomedical robots, robot ethics, robot laws, human-robot interaction

Procedia PDF Downloads 94
7658 The Investigate Relationship between Moral Hazard and Corporate Governance with Earning Forecast Quality in the Tehran Stock Exchange

Authors: Fatemeh Rouhi, Hadi Nassiri

Abstract:

Earning forecast is a key element in economic decisions but there are some situations, such as conflicts of interest in financial reporting, complexity and lack of direct access to information has led to the phenomenon of information asymmetry among individuals within the organization and external investors and creditors that appear. The adverse selection and moral hazard in the investor's decision and allows direct assessment of the difficulties associated with data by users makes. In this regard, the role of trustees in corporate governance disclosure is crystallized that includes controls and procedures to ensure the lack of movement in the interests of the company's management and move in the direction of maximizing shareholder and company value. Therefore, the earning forecast of companies in the capital market and the need to identify factors influencing this study was an attempt to make relationship between moral hazard and corporate governance with earning forecast quality companies operating in the capital market and its impact on Earnings Forecasts quality by the company to be established. Getting inspiring from the theoretical basis of research, two main hypotheses and sub-hypotheses are presented in this study, which have been examined on the basis of available models, and with the use of Panel-Data method, and at the end, the conclusion has been made at the assurance level of 95% according to the meaningfulness of the model and each independent variable. In examining the models, firstly, Chow Test was used to specify either Panel Data method should be used or Pooled method. Following that Housman Test was applied to make use of Random Effects or Fixed Effects. Findings of the study show because most of the variables are positively associated with moral hazard with earnings forecasts quality, with increasing moral hazard, earning forecast quality companies listed on the Tehran Stock Exchange is increasing. Among the variables related to corporate governance, board independence variables have a significant relationship with earnings forecast accuracy and earnings forecast bias but the relationship between board size and earnings forecast quality is not statistically significant.

Keywords: corporate governance, earning forecast quality, moral hazard, financial sciences

Procedia PDF Downloads 299
7657 Colonialism and Modernism in Architecture, the Case of a Blank Page Opportunity in Casablanka

Authors: Nezha Alaoui

Abstract:

The early 1950s French colonial context in Morocco provided an opportunity for architects to question the modernist established order by building dwellings for the local population. The dwellings were originally designed to encourage Muslims to adopt an urban lifestyle based on local customs. However, the inhabitants transformed their dwelling into a hybrid habitation. This paper aims to prove the relevance of the design process in accordance with the local colonial context by analyzing the dwellers' appropriation process and the modification of their habitat.

Keywords: colonial heritage, appropriation process, islamic spatial habit, housing experiment, modernist mass housing

Procedia PDF Downloads 115
7656 Two Degree of Freedom Spherical Mechanism Design for Exact Sun Tracking

Authors: Osman Acar

Abstract:

Sun tracking systems are the systems following the sun ray by a right angle or by predetermined certain angle. In this study, we used theoretical trajectory of sun for latitude of central Anatolia in Turkey. A two degree of freedom spherical mechanism was designed to have a large workspace able to follow the sun's theoretical motion by the right angle during the whole year. An inverse kinematic analysis was generated to find the positions of mechanism links for the predicted trajectory. Force and torque analysis were shown for the first day of the year.

Keywords: sun tracking, theoretical sun trajectory, spherical mechanism, inverse kinematic analysis

Procedia PDF Downloads 400
7655 Enhancing the Network Security with Gray Code

Authors: Thomas Adi Purnomo Sidhi

Abstract:

Nowadays, network is an essential need in almost every part of human daily activities. People now can seamlessly connect to others through the Internet. With advanced technology, our personal data now can be more easily accessed. One of many components we are concerned for delivering the best network is a security issue. This paper is proposing a method that provides more options for security. This research aims to improve network security by focusing on the physical layer which is the first layer of the OSI model. The layer consists of the basic networking hardware transmission technologies of a network. With the use of observation method, the research produces a schematic design for enhancing the network security through the gray code converter.

Keywords: network, network security, grey code, physical layer

Procedia PDF Downloads 484
7654 Chilled Books: Managing Defamatory Content in Non-fiction Trade Publishing

Authors: Katherine Day

Abstract:

Non-fiction genres (autobiographies and biographies, true stories and criticism, investigative journalism and narrative journalism) have enjoyed increasing sales in the English-language publishing territories over the last decade, but writing the tell-all or exposé is not without consequences: defamation laws cast a “chilling effect” by regarding reputation above publications with a public interest element. This is evident in the many publications that have been amended or pulped after publication. These communications, alterations and negotiations indicate that the threat of legal action forms part of the editorial decision-making around such publications, the presence of which could be attributed to strict defamation laws. In the UK and Australia, particularly, defamation law has proved notoriously biased in favour of plaintiffs. The legal obstacles have prompted law reform by way of section 4 of the UK Defamation Act, which allows for editorial assessment into whether the statement/s made are in the public interest; as of July 1st 2021, the NSW Government in Australia also implemented reforms to help steer the law towards more flexibility in the digital age – the most interesting of these developments for commercial publishing being the new ‘public interest’ defence (s 29A), which is modelled on the UK’s section 4 and which most states in Australia have now integrated into their respective state laws (Queensland, new South Wales, Victoria and South Australia, with the remaining states committing at a later date). This paper will outline and discuss the preliminary findings of a 1-year project that aims to explore how potentially litigious content is managed in unpublished non-fiction manuscripts in two countries identified as having strict defamation laws: Australia and the UK. Significantly, it expects to indicate the burden of current defamation laws on publishing practice and publishing outputs in these countries by interrogating in-house editorial processes and the likelihood of editorial management in a ‘post negotiation space’, where the activities and communication between authors and editors are reconstructed, if necessary, to correct the author/publisher power balance and affirm the business relationship. In doing so, the project asks: has the threat, explicit or implicit, of defamation action produced a significant chilling effect in trade non-fiction publishing in the UK and Australia?

Keywords: defamation, publishing, socio-legal, authorship, editing, literature

Procedia PDF Downloads 47
7653 Predictability of Thermal Response in Housing: A Case Study in Australia, Adelaide

Authors: Mina Rouhollahi, J. Boland

Abstract:

Changes in cities’ heat balance due to rapid urbanization and the urban heat island (UHI) have increased energy demands for space cooling and have resulted in uncomfortable living conditions for urban residents. Climate resilience and comfortable living spaces can be addressed through well-designed urban development. The sustainable housing can be more effective in controlling high levels of urban heat. In Australia, to mitigate the effects of UHIs and summer heat waves, one solution to sustainable housing has been the trend to compact housing design and the construction of energy efficient dwellings. This paper analyses whether current housing configurations and orientations are effective in avoiding increased demands for air conditioning and having an energy efficient residential neighborhood. A significant amount of energy is consumed to ensure thermal comfort in houses. This paper reports on the modelling of heat transfer within the homes using the measurements of radiation, convection and conduction between exterior/interior wall surfaces and outdoor/indoor environment respectively. The simulation was tested on selected 7.5-star energy efficient houses constructed of typical material elements and insulation in Adelaide, Australia. The chosen design dwellings were analyzed in extremely hot weather through one year. The data were obtained via a thermal circuit to accurately model the fundamental heat transfer mechanisms on both boundaries of the house and through the multi-layered wall configurations. The formulation of the Lumped capacitance model was considered in discrete time steps by adopting a non-linear model method. The simulation results focused on the effects of orientation of the solar radiation on the dynamic thermal characteristics of the houses orientations. A high star rating did not necessarily coincide with a decrease in peak demands for cooling. A more effective approach to avoid increasing the demands for air conditioning and energy may be to integrate solar–climatic data to evaluate the performance of energy efficient houses.

Keywords: energy-efficient residential building, heat transfer, neighborhood orientation, solar–climatic data

Procedia PDF Downloads 119
7652 Approximate Spring Balancing for Swimming Pool Lift Mechanism to Reduce Actuator Torque

Authors: Apurva Patil, Sujatha Srinivasan

Abstract:

Reducing actuator loads is important for applications in which human effort is required for actuation. The potential benefit of applying spring balancing to rehabilitation devices which work against gravity on a nonhorizontal plane is well recognized, but practical applications have been elusive. Although existing methods provide exact spring balance, they require additional masses or auxiliary links, or all the springs used originate from the ground, which makes the resulting device bulky and space-inefficient. This paper uses a method of static balancing of mechanisms with conservative loads such as gravity and spring loads using non-zero-free-length springs and no auxiliary links. Application of this method to a manually operated swimming pool lift mechanism which lowers and raises the physically challenged users into or out of the swimming pool is presented here. Various possible configurations using extension and compression springs as well as gas spring in the mechanism are compared. This work involves approximate spring balancing of the mechanism using minimization of potential energy variance. It uses the approach of flattening the potential energy distribution over the workspace and fuses it with numerical optimization. The results show the considerable reduction in actuator torque requirement with practical spring design and arrangement. Although the method provides only an approximate balancing, it is versatile, flexible in choosing appropriate control variables that are relevant to the design problem and easy to implement. The true potential of this technique lies in the fact that it uses a very simple optimization to find the spring constant, free length of the spring and the optimal attachment points subject to the optimization constraints. Also, it uses physically realizable non-zero-free-length springs directly, thereby reducing the complexity involved in simulating zero-free-length springs from non-zero-free-length springs. This method allows springs to be attached inside the mechanism, which makes the implementation of spring balancing practical. Because auxiliary linkages can be avoided, the resultant swimming pool lift mechanism is compact. The cost benefits and reduced complexity can be significant advantages in the development of this user-actuated swimming pool lift for developing countries.

Keywords: gas spring, rehabilitation device, spring balancing, swimming pool lift

Procedia PDF Downloads 223
7651 Application of Multilinear Regression Analysis for Prediction of Synthetic Shear Wave Velocity Logs in Upper Assam Basin

Authors: Triveni Gogoi, Rima Chatterjee

Abstract:

Shear wave velocity (Vs) estimation is an important approach in the seismic exploration and characterization of a hydrocarbon reservoir. There are varying methods for prediction of S-wave velocity, if recorded S-wave log is not available. But all the available methods for Vs prediction are empirical mathematical models. Shear wave velocity can be estimated using P-wave velocity by applying Castagna’s equation, which is the most common approach. The constants used in Castagna’s equation vary for different lithologies and geological set-ups. In this study, multiple regression analysis has been used for estimation of S-wave velocity. The EMERGE module from Hampson-Russel software has been used here for generation of S-wave log. Both single attribute and multi attributes analysis have been carried out for generation of synthetic S-wave log in Upper Assam basin. Upper Assam basin situated in North Eastern India is one of the most important petroleum provinces of India. The present study was carried out using four wells of the study area. Out of these wells, S-wave velocity was available for three wells. The main objective of the present study is a prediction of shear wave velocities for wells where S-wave velocity information is not available. The three wells having S-wave velocity were first used to test the reliability of the method and the generated S-wave log was compared with actual S-wave log. Single attribute analysis has been carried out for these three wells within the depth range 1700-2100m, which corresponds to Barail group of Oligocene age. The Barail Group is the main target zone in this study, which is the primary producing reservoir of the basin. A system generated list of attributes with varying degrees of correlation appeared and the attribute with the highest correlation was concerned for the single attribute analysis. Crossplot between the attributes shows the variation of points from line of best fit. The final result of the analysis was compared with the available S-wave log, which shows a good visual fit with a correlation of 72%. Next multi-attribute analysis has been carried out for the same data using all the wells within the same analysis window. A high correlation of 85% has been observed between the output log from the analysis and the recorded S-wave. The almost perfect fit between the synthetic S-wave and the recorded S-wave log validates the reliability of the method. For further authentication, the generated S-wave data from the wells have been tied to the seismic and correlated them. Synthetic share wave log has been generated for the well M2 where S-wave is not available and it shows a good correlation with the seismic. Neutron porosity, density, AI and P-wave velocity are proved to be the most significant variables in this statistical method for S-wave generation. Multilinear regression method thus can be considered as a reliable technique for generation of shear wave velocity log in this study.

Keywords: Castagna's equation, multi linear regression, multi attribute analysis, shear wave logs

Procedia PDF Downloads 205
7650 The Development of an Anaesthetic Crisis Manual for Acute Critical Events: A Pilot Study

Authors: Jacklyn Yek, Clara Tong, Shin Yuet Chong, Yee Yian Ong

Abstract:

Background: While emergency manuals and cognitive aids (CA) have been used in high-hazard industries for decades, this has been a nascent field in healthcare. CAs can potentially offset the large cognitive load involved in crisis resource management and possibly facilitate the efficient performance of key steps in treatment. A crisis manual was developed based on local guidelines and the latest evidence-based information and introduced to a tertiary hospital setting in Singapore. Hence, the objective of this study is to evaluate the effectiveness of the crisis manual in guiding response and management of critical events. Methods: 7 surgical teams were recruited to participate in a series of simulated emergencies in high-fidelity operating room simulator over the period of April to June 2018. All teams consisted of a surgical consultant and medical officer/registrar, anesthesia consultant and medical officer/registrar; as well as a circulating, scrub and anesthetic nurse. Each team performed a simulated operation in which 1 or more of the crisis events occurred. The teams were randomly assigned to a scenario of the crisis manual and all teams were deemed to be equal in experience and knowledge. Before the simulation, teams were instructed on proper checklist use but the use of the checklist was optional. Results: 7 simulation sessions were performed, consisting of the following scenarios: Airway fire, Massive Transfusion Protocol, Malignant Hyperthermia, Eclampsia, and Difficult Airway. Out of the 7 surgical teams, 2 teams made use of the crisis manual – of which both teams had encountered a ‘Malignant Hyperthermia’ scenario. These team members reflected that the crisis manual assisted allowed them to work in a team, especially being able to involve the surgical doctors who were unfamiliar with the condition and management. A run chart plotted showed a possible upward trend, suggesting that with increasing awareness and training, staff would become more likely to initiate the use of the crisis manual. Conclusion: Despite the high volume load in this tertiary hospital, certain crises remain rare and clinicians are often caught unprepared. A crisis manual is an effective tool and easy-to-use repository that can improve patient outcome and encourage teamwork. With training, familiarity would allow clinicians to be increasingly comfortable with reaching out for the crisis manual. More simulation training would need to be conducted to determine its effectiveness.

Keywords: crisis resource management, high fidelity simulation training, medical errors, visual aids

Procedia PDF Downloads 108