Search results for: the front-end
Commenced in January 2007
Frequency: Monthly
Edition: International
Paper Count: 4

Search results for: the front-end

4 Analysis of a Power Factor Correction Converter for Light Emitting Diode Driver Application

Authors: Edwina G. Rodrigues, S. J. Bindhu, A. V. Rajesh

Abstract:

This paper proposes a switched capacitor based driver circuit for high power light emitting diodes with a front end rectifier. LEDs are low-voltage light sources, requiring a constant DC voltage or current to operate optimally. LEDs, therefore, require a device that can convert incoming AC power to the proper DC voltage, and regulate the current flowing through the LED during operation. Proposed topology has a front end converter. It is an AC-DC rectifier that works on bridgeless boost topology which shapes the input current waveform. The front end converter is followed by a DC-DC converter which provides a constant DC voltage across the LEDs. A 12V AC input is given to the input of frontend converter which rectifies and boost the voltage to 24v DC and gives it to the DC-DC converter. The DC-DC converter converts the 24V DC and regulates this constant DC voltage across the LEDs.

Keywords: bridgeless rectifier, power factor correction(PFC), SC converter, total harmonic distortion (THD)

Procedia PDF Downloads 840
3 Digital Forensic Exploration Framework for Email and Instant Messaging Applications

Authors: T. Manesh, Abdalla A. Alameen, M. Mohemmed Sha, A. Mohamed Mustaq Ahmed

Abstract:

Email and instant messaging applications are foremost and extensively used electronic communication methods in this era of information explosion. These applications are generally used for exchange of information using several frontend applications from various service providers by its users. Almost all such communications are now secured using SSL or TLS security over HTTP communication. At the same time, it is also noted that cyber criminals and terrorists have started exchanging information using these methods. Since communication is encrypted end-to-end, tracing significant forensic details and actual content of messages are found to be unattended and severe challenges by available forensic tools. These challenges seriously affect in procuring substantial evidences against such criminals from their working environments. This paper presents a vibrant forensic exploration and architectural framework which not only decrypts any communication or network session but also reconstructs actual message contents of email as well as instant messaging applications. The framework can be effectively used in proxy servers and individual computers and it aims to perform forensic reconstruction followed by analysis of webmail and ICQ messaging applications. This forensic framework exhibits a versatile nature as it is equipped with high speed packet capturing hardware, a well-designed packet manipulating algorithm. It regenerates message contents over regular as well as SSL encrypted SMTP, POP3 and IMAP protocols and catalyzes forensic presentation procedure for prosecution of cyber criminals by producing solid evidences of their actual communication as per court of law of specific countries.

Keywords: forensics, network sessions, packet reconstruction, packet reordering

Procedia PDF Downloads 304
2 Impulsivity Leads to Compromise Effect

Authors: Sana Maidullah, Ankita Sharma

Abstract:

The present study takes naturalistic decision-making approach to examine the role of personality in information processing in consumer decision making. In the technological era, most of the information comes in form of HTML or similar language via the internet; processing of this situation could be ambiguous, laborious and painful. The present study explores the role of impulsivity in creating an extreme effect on consumer decision making. Specifically, the study explores the role of impulsivity in extreme effect, i.e., extremeness avoidance (compromise effect) and extremeness seeking; the role of demographic variables, i.e. age and gender, in the relation between impulsivity and extreme effect. The study was conducted with the help of a questionnaire and two experiments. The experiment was designed in the form of two shopping websites with two product types: Hotel choice and Mobile choice. Both experimental interfaces were created with the Xampp software, the frontend of interfaces was HTML CSS JAVASCRIPT and backend was PHP MySQL. The mobile experiment was designed to measure the extreme effect and hotel experiment was designed to measure extreme effect with alignability of attributes. To observe the possibilities of the combined effect of individual difference and context effects, the manipulation of price, a number of alignable attributes and number of the non-alignable attributes is done. The study was conducted on 100 undergraduate and post-graduate engineering students within the age range of 18-35. The familiarity and level of use of internet and shopping website were assessed and controlled in the analysis. The analysis was done by using a t-test, ANOVA and regression analysis. The results indicated that the impulsivity leads to compromise effect and at the same time it also increases the relationship between alignability of attribute among choices and the compromise effect. The demographic variables were found to play a significant role in the relationship. The subcomponents of impulsivity were significantly influencing compromise effect, but the cognitive impulsivity was significant for women, and motor impulsivity was significant for males only. The impulsivity was significantly positively predicted by age, though there were no significant gender differences in impulsivity. The results clearly indicate the importance of individual factors in decision making. The present study, with precise and direct results, provides a significant suggestion for market analyst and business providers.

Keywords: impulsivity, extreme effect, personality, alignability, consumer decision making

Procedia PDF Downloads 164
1 Modelling of Reactive Methodologies in Auto-Scaling Time-Sensitive Services With a MAPE-K Architecture

Authors: Óscar Muñoz Garrigós, José Manuel Bernabeu Aubán

Abstract:

Time-sensitive services are the base of the cloud services industry. Keeping low service saturation is essential for controlling response time. All auto-scalable services make use of reactive auto-scaling. However, reactive auto-scaling has few in-depth studies. This presentation shows a model for reactive auto-scaling methodologies with a MAPE-k architecture. Queuing theory can compute different properties of static services but lacks some parameters related to the transition between models. Our model uses queuing theory parameters to relate the transition between models. It associates MAPE-k related times, the sampling frequency, the cooldown period, the number of requests that an instance can handle per unit of time, the number of incoming requests at a time instant, and a function that describes the acceleration in the service's ability to handle more requests. This model is later used as a solution to horizontally auto-scale time-sensitive services composed of microservices, reevaluating the model’s parameters periodically to allocate resources. The solution requires limiting the acceleration of the growth in the number of incoming requests to keep a constrained response time. Business benefits determine such limits. The solution can add a dynamic number of instances and remains valid under different system sizes. The study includes performance recommendations to improve results according to the incoming load shape and business benefits. The exposed methodology is tested in a simulation. The simulator contains a load generator and a service composed of two microservices, where the frontend microservice depends on a backend microservice with a 1:1 request relation ratio. A common request takes 2.3 seconds to be computed by the service and is discarded if it takes more than 7 seconds. Both microservices contain a load balancer that assigns requests to the less loaded instance and preemptively discards requests if they are not finished in time to prevent resource saturation. When load decreases, instances with lower load are kept in the backlog where no more requests are assigned. If the load grows and an instance in the backlog is required, it returns to the running state, but if it finishes the computation of all requests and is no longer required, it is permanently deallocated. A few load patterns are required to represent the worst-case scenario for reactive systems: the following scenarios test response times, resource consumption and business costs. The first scenario is a burst-load scenario. All methodologies will discard requests if the rapidness of the burst is high enough. This scenario focuses on the number of discarded requests and the variance of the response time. The second scenario contains sudden load drops followed by bursts to observe how the methodology behaves when releasing resources that are lately required. The third scenario contains diverse growth accelerations in the number of incoming requests to observe how approaches that add a different number of instances can handle the load with less business cost. The exposed methodology is compared against a multiple threshold CPU methodology allocating/deallocating 10 or 20 instances, outperforming the competitor in all studied metrics.

Keywords: reactive auto-scaling, auto-scaling, microservices, cloud computing

Procedia PDF Downloads 66