_id
stringlengths 40
40
| text
stringlengths 0
10k
|
---|---|
084459cc2b4499ae8ae4a0edea46687900a0c4d4
| |
e11cc2a3cd150927ec61e596296ceb9b0c20bda4
|
Frequency diverse array (FDA) is an emerging technology, the hybrid of FDA and multiple-input-multiple-output (FDA-MIMO) under monostatic scenarios has received much attention in recent years. However, little work have been done for bistatic FDA-MIMO radar. In this paper, we investigate strategies on estimating direction-of-departure (DOD), direction-of-arrival and range for bistatic FDA-MIMO radar. Our strategies have two aspects. First, nonlinear frequency increments including both subarray and random modes are employed to overcome the problem that the DOD and range parameters of FDA transmitting steering vectors are coupled. Second, in order to reduce the computational complexity associated with the 3-D spectral peak searching algorithms, estimation of signal parameters via rotational invariance technique and parallel factor algorithms with their corresponding phase ambiguity resolving methods, are proposed for subarray and random modes, respectively. Both of the two algorithms perform well while the range parameter of targets satisfy a range constraint criterion. This criterion can also be used for designing frequency increments of bistatic FDA-MIMO radar. Additionally, the Cramér-Rao bound of bistatic FDA-MIMO radar and the algorithm performance analysis consist of identifiability and complexity are derived. All the proposed methods are verified by both theoretical analysis and numerical simulations. And satisfactory results are achieved.
|
04250e037dce3a438d8f49a4400566457190f4e2
|
However, for a task with very high-dimensional data such as images, the traditional LDA algorithm encounters several di$culties. Consider face recognition for example. A low-de"nition face image of size 64 64 implies a feature space of 64 64"4096 dimensions, and therefore scatter matrices of size 4096 4096"16M. First, it is computationally challenging to handle big matrices (such as computing eigenvalues). Second, those matrices are almost always singular, as the number of training images needs to be at least 16M for them to be non-degenerate. Due to these di$culties, it is commonly believed that a direct LDA solution for such high-dimensional data is infeasible. Thus, ironically, before LDA can be used to reduce dimensionality, another procedure has to be "rst applied for dimensionality reduction. In face recognition, many techniques have been proposed (for a good review, see Ref. [1]). Among them, the most notable is a two-stage PCA#LDA approach [2,3]:
|
3b6c2c76420dee7413691c2fcd157a13bd1ce69f
|
Modern cryptocurrencies exploit decentralised ledgers — the so-called blockchains — to record a public and unalterable history of transactions. These ledgers represent a rich, and increasingly growing, source of information, in part of difficult interpretation and undisclosed meaning. Many analytics, mostly based on ad-hoc engineered solutions, are being developed to discover relevant knowledge from these data. We introduce a framework for the development of custom analytics on Bitcoin — the most preeminent cryptocurrency — which also allows to integrate data within the blockchain with data retrieved form external sources. We illustrate the flexibility and effectiveness of our analytics framework by means of paradigmatic use cases.
|
76415cffc9c0c9585307696da09a0d082f8e8df4
| |
496691c466c736ac02c89e36491c7da2c4d58650
|
Phenomenal growth in the telecommunication industry in recent years has brought significant advances in filter technology as new communication systems emerged, demanding more stringent filter characteristics. In particular, the growth of the wireless communication industry has spurred tremendous activity in the area of microwave filter miniaturization and has been responsible for many advances made in this field. The filters that are currently being used in wireless base stations can be divided into two main categories: coaxial cavity resonator filters and dielectric resonator (DR) filters. While coaxial cavity filters have limited quality factor (Q) values, they offer the lowest cost design and are still being widely employed, particularly in wide bandwidth applications. With increased demands for high performance wireless systems, dielectric resonator filters are emerging as the baseline design for wireless base stations. Over the next five years, dielectric resonator filters are expected to have a significant share of the overall wireless base station filter market. High-temperature superconductor (HTS) filters are also expected to have a share of this market, particularly for systems, which have very stringent requirements for out-of-band interference. In this article, we begin by reviewing the main filter requirements, highlighting the technologies that are being currently employed. Emerging filter technologies that have the potential to replace the existing technologies are then described.
|
538c4e4327c706a9346472661b3256c39b07df08
|
Causal knowledge plays a crucial role in human thought, but the nature of causal representation and inference remains a puzzle. Can human causal inference be captured by relations of probabilistic dependency, or does it draw on richer forms of representation? This article explores this question by reviewing research in reasoning, decision making, various forms of judgment, and attribution. We endorse causal Bayesian networks as the best normative framework and as a productive guide to theory building. However, it is incomplete as an account of causal thinking. On the basis of a range of experimental work, we identify three hallmarks of causal reasoning-the role of mechanism, narrative, and mental simulation-all of which go beyond mere probabilistic knowledge. We propose that the hallmarks are closely related. Mental simulations are representations over time of mechanisms. When multiple actors are involved, these simulations are aggregated into narratives.
|
2b735a5cd94b0b5868e071255bd187a901cb975a
| |
4017f984d1b4b8748a06da2739183782bbe9b46d
| |
7be0f21125b4accf24ed8884e32f723606a48b9e
| |
c24ae0a7b8c0b4557690c247accee55747e39acc
|
In this paper, we evaluate handover (HO) performance when we apply various time-to-trigger (TTT) methods in self-organizing network (SON)-based long term evolution (LTE) systems. Although TTT can mitigate the wasteful ping-pong HO effect, it also can cause undesirable radio link failure (RLF) due to delayed HO. The optimal HO timings that produce the lowest ping-pong rate within allowable RLF rate vary depending on user equipment (UE) speeds and neighboring cell configurations. To achieve efficient HO timings, we propose and investigate two methods: “adaptive” and “grouping.” In the “adaptive method,” we select the adaptive TTT value for each UE speed based on RLF rate of 2%. The “grouping method” classifies UE speeds into three ranges and assigns the proper TTT value to each range. To apply the LTE specification more effectively, we suggest the criteria of grouping, and propose the proper TTT value for each range. We consider HO in two neighboring cell configurations: from a macro cell to either a macro cell or a pico cell. Simulation results show that the HO performance of the “adaptive method” is greatly improved compared to that of applying fixed TTT values. The results also show that the performance of the “grouping method”, using proposed criteria and proper TTT values, is comparable to that of the “adaptive method.”
|
2c56368255ca6a86ef0a26466b9d0a2425f55e9d
|
Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at http://www.jstor.org/about/terms.html. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you may use content in the JSTOR archive only for your personal, non-commercial use.
|
2d7622afe72922de5c430014965cfddf33692885
|
Manufacturing flexibility is widely recognized as a critical component to achieving a competitive advantage in the marketplace. A comprehensive look at the empirical research pertaining to manufacturing flexibility highlights the very fragmented nature of this body of work. We present a comprehensive contingency-based framework for examining the content related issues involving the relationships and variables included in past studies. We also examine several important Ž . research designrmethodology issues e.g., sampling, data collection and measurement and propose solutions to some identified problems. q 2000 Elsevier Science B.V. All rights reserved.
|
df7c11bf2eaec672904440901efaeba2ef601fef
|
This paper develops a general framework for analyzing corporate risk management policies. We begin by observing that if external sources of finance are more costly to corporations than internally generated funds, there will typically be a benefit to hedging: hedging adds value to the extent that it helps ensure that a corporation has sufficient internal funds available to take advantage of attractive investment opportunities. We then argue that this simple observation has wide ranging implications for the design of risk management strategies. We delineate how these strategies should depend on such factors as shocks to investment and financing opportunities. We also discuss exchange rate hedging strategies for multinationals, as well as strategies involving "nonlinear" instruments like options. CORPORATIONS TAKE RISK MANAGEMENT very seriously—recent surveys find that risk management is ranked by financial executives as one of their most important objectives.^ Given its real-world prominence, one might guess that the topic of risk management would command a great deal of attention from researchers in finance, and that practitioners would therefore have a welldeveloped body of wisdom from which to draw in formulating hedging strategies. Such a guess would, however, be at best only partially correct. Finance theory does do a good job of instructing firms on the implementation of hedges. For example, if a refining company decides that it wants to use options to reduce its exposure to oil prices by a certain amount, a BlackScholes type model can help the company calculate the number of contracts needed. Indeed, there is an extensive literature that covers numerous practical aspects of what might be termed "hedging mechanics," from the computation of hedge ratios to the institutional peculiarities of individual contracts. Unfortunately, finance theory has had much less clear cut guidance to offer on the logically prior questions of hedging strategy: What sorts of risks * Froot is from Harvard and NBER, Scharfstein is from MIT and NBER, and Stein is from MIT and NBER. We thank Don Lessard, Tim Luehrman, Andre Perold, Raghuram Rajan, Julio Rotemberg, and Stew Myers for helpful discussions. We are also grateful to the IFSRC and the Center for Energy Policy Research at MIT, the Department of Research at Harvard Business School, the National Science Foundation, and Batterymarch Financial Management for generous financial support. ' See Rawls and Smithson (1990).
|
3ab1c861c4be472a9b672214f472e37ad82931bb
|
EMPIRICAL models of business investment rely generally on the assumption of a "representative firm" that responds to prices set in centralized securities markets. Indeed, if all firms have equal access to capital markets, firms' responses to changes in the cost of capital or tax-based investment incentives differ only because of differences in investment demand. A firm's financial structure is irrelevant to investment because external funds provide a perfect substitute for internal capital. In general, with perfect capital markets, a firm's investment decisions are independent of its financial condition. An alternative research agenda, however, has been based on the view that internal and external capital are not perfect substitutes. According to this view, investment may depend on financial factors, such as the availability of internal finance, access to new debt or equity finance, or the functioning of particular credit markets. For example, a firm's internal cash flow may affect investment spending because of a "financ-
|
329cf7a6528f339560ff9fb617f140b1a54a067e
|
The Internet and World Wide Web have brought us into a world of endless possibiliti es: interactive Web sites to experience, music to li sten to, conversations to participate in, and every conceivable consumer item to order. But this world also is one of endless choice: how can we select from a huge universe of items of widely varying quality? Computational recommender systems have emerged to address this issue. They enable people to share their opinions and benefit from each other’s experience. We present a framework for understanding recommender systems and survey a number of distinct approaches in terms of this framework. We also suggest two main research challenges: (1) helping people form communities of interest while respecting personal privacy, and (2) developing algorithms that combine multiple types of information to compute recommendations.
|
462881e12a6708ddc520fa4d58c99ba73b18a6ab
|
Continuous electrowetting (CEW) is demonstrated to be an effective actuation mechanism for reconfigurable radio frequency (RF) devices that use non-toxic liquid-metal tuning elements. Previous research has shown CEW is an efficient means of electrically inducing motion in a liquid-metal slug, but precise control of the slug's position within fluidic channels has not been demonstrated. Here, the precise positioning of liquid-metal slugs is achieved using CEW actuation in conjunction with channels designed to minimize the liquid-metal surface energy at discrete locations. This approach leverages the high surface tension of liquid metal to control its resting position with submillimeter accuracy. The CEW actuation and fluidic channel design were optimized to create reconfigurable RF devices. In addition, solutions for the reliable actuation of a gallium-based, non-toxic liquid-metal alloy (Galinstan) are presented that mitigate the tendency of the alloy to form a surface oxide layer capable of wetting to the channel walls, inhibiting motion. A reconfigurable slot antenna utilizing these techniques to achieve a 15.2% tunable frequency bandwidth is demonstrated.
|
334d6c71b6bce8dfbd376c4203004bd4464c2099
|
References [1] Wang, P., Shen, C., van den Hengel, A., A fast semidefinite approach to solving binary quadratic problems. CVPR 2013. [2] Candes, E.J., Li, X., Soltanolkotabi, M., Phase retrieval via wirtinger flow: Theory and algorithms. IEEE Transaction on Information Theory, 2015. [3] Zheng, Q., Lafferty, J., A convergent gradient descent algorithm for rank minimization and semidefinite programming from random linear measurements. NIPS 2015. [4] Netrapalli, P., Jain, P., Sanghavi, S., Phase retrieval using alternating minimization. NIPS 2013. Many important problems requires solving for low rank SDPs with PSD constraint matrices.
|
5ac65efcf8db05e8f1f70e09d51f275caa9d0aae
|
In this work we describe a Convolutional Neural Network (CNN) to accurately predict image quality without a reference image. Taking image patches as input, the CNN works in the spatial domain without using hand-crafted features that are employed by most previous methods. The network consists of one convolutional layer with max and min pooling, two fully connected layers and an output node. Within the network structure, feature learning and regression are integrated into one optimization process, which leads to a more effective model for estimating image quality. This approach achieves state of the art performance on the LIVE dataset and shows excellent generalization ability in cross dataset experiments. Further experiments on images with local distortions demonstrate the local quality estimation ability of our CNN, which is rarely reported in previous literature.
|
62d37831e00bc0511480f7cb78874ffb30331382
|
Fingerprint matching is affected by the nonlinear distortion introduced in fingerprint impressions during the image acquisition process. This nonlinear deformation causes fingerprint features such as minutiae points and ridge curves to be distorted in a complex manner. In this paper we develop an average deformation model for a fingerprint impression (baseline impression) by observing its relative distortion with respect to several other impressions of the same finger. The deformation is computed using a Thin Plate Spline (TPS) model that relies on ridge curve correspondences between image pairs. The estimated average deformation is used to distort the minutiae template of the baseline impression prior to matching. An index of deformation has been proposed to select the average deformation model with the least variability corresponding to a finger. Preliminary results indicate that the average deformation model can improve the matching performance of a fingerprint matcher.
|
cf20e34a1402a115523910d2a4243929f6704db1
| |
b9843426b745ea575276fce3b527a2dc4d2eebb3
|
Nowadays smartphones are ubiquitous in various aspects of our lives. The processing power, communication bandwidth, and the memory capacity of these devices have surged considerably in recent years. Besides, the variety of sensor types, such as accelerometer, gyroscope, humidity sensor, and bio-sensors, which are embedded in these devices, opens a new horizon in self-monitoring of physical daily activities. One of the primary steps for any research in the area of detecting daily life activities is to test a detection method on benchmark datasets. Most of the early datasets limited their work to collecting only a single type of sensor data such as accelerometer data. While some others do not consider age, weight, and gender of the subjects who have participated in collecting their activity data. Finally, part of the previous works collected data without considering the smartphone's position. In this paper, we introduce a new dataset, called Position-Aware Multi-Sensor (PAMS). The dataset contains both accelerometer and gyroscope data. The gyroscope data boosts the accuracy of activity recognition methods as well as enabling them to detect a wider range of activities. We also take the user information into account. Based on the biometric attributes of the participants, a separate learned model is generated to analyze their activities. We concentrate on several major activities, including sitting, standing, walking, running, ascending/descending stairs, and cycling. To evaluate the dataset, we use various classifiers, and the outputs are compared to the WISDM. The results show that using aforementioned classifiers, the average precision for all activities is above 88.5%. Besides, we measure the CPU, memory, and bandwidth usage of the application collecting data on the smartphone.
|
22d35b27bab295efe9d5a28cdf15ed7c4fbcf25c
|
The emerging discipline of algorithm engineering has primarily focussed on transforming pencil-and-paper sequential algorithms into robust, efficient, well tested, and easily used implementations. As parallel computing becomes ubiquitous, we need to extend algorithm engineering techniques to parallel computation. Such an extension adds significant complications. After a short review of algorithm engineering achievements for sequential computing, we review the various complications caused by parallel computing, present some examples of successful efforts, and give a personal view of possible future research.
|
7fdf8e2393ee280c9a3aeb94aae375411358e45a
|
Article history: Received 1 August 2007 Received in revised form 22 May 2009 Accepted 26 May 2009
|
28e702e1a352854cf0748b9a6a9ad6679b1d4e83
|
The skyline of a d-dimensional dataset contains the points that are not dominated by any other point on all dimensions. Skyline computation has recently received considerable attention in the database community, especially for progressive methods that can quickly return the initial results without reading the entire database. All the existing algorithms, however, have some serious shortcomings which limit their applicability in practice. In this article we develop branch-and-bound skyline (BBS), an algorithm based on nearest-neighbor search, which is I/O optimal, that is, it performs a single access only to those nodes that may contain skyline points. BBS is simple to implement and supports all types of progressive processing (e.g., user preferences, arbitrary dimensionality, etc). Furthermore, we propose several interesting variations of skyline computation, and show how BBS can be applied for their efficient processing.
|
a2e738c4107a8d123a6be42d34c02b9f9939b50d
|
Fog Computing (FC) has extended the services of cloud computing to the edge of the network. It inherits some of the characteristics from cloud computing but FC also have some distinguished features such as geo-distribution, location awareness and low latency. Along with the inherited characteristics, it also inherits the issues and problems of cloud computing like energy efficiency, resource management and security issues. This paper presents the critical analysis of the fog architecture with respect to security. The state of the art work done since 2012 is critical analyzed on the bases of security techniques and security threats. We grouped the existing security techniques on the bases of security goals achieved by each. It will provide a clear and comprehensive distinction between the security areas explored and those, which still need researchers' attention.
|
e2e9eb8d0ac182b9db04a4fa833ee078e04a10c3
|
Learning from expert demonstrations has received a lot of attention in artificial intelligence and machine learning. The goal is to infer the underlying reward function that an agent is optimizing given a set of observations of the agent’s behavior over time in a variety of circumstances, the system state trajectories, and a plant model specifying the evolution of the system state for different agent’s actions. The system is often modeled as a Markov decision process (Puterman 2014), that is, the next state depends only on the current state and agent’s action, and the the agent’s choice of action depends only on the current state. While the former is a Markovian assumption on the evolution of system state, the later assumes that the target reward function is itself Markovian. In this work, we explore learning a class of non-Markovian reward functions, known in the formal methods literature as specifications. These specifications offer better composition, transferability, and interpretability. We then show that inferring the specification can be done efficiently without unrolling the transition system. We demonstrate on a 2-d grid world example.
|
6bd3544905ec46cb321bc06c7937110da44f90ed
| |
2151a214aca6e72ee2980ae8cbf7be47fed0cb7a
|
Simulators have played a critical role in robotics research as tools for quick and efficient testing of new concepts, strategies, and algorithms. To date, most simulators have been restricted to 2D worlds, and few have matured to the point where they are both highly capable and easily adaptable. Gazebo is designed to fill this niche by creating a 3D dynamic multi-robot environment capable of recreating the complex worlds that would be encountered by the next generation of mobile robots. Its open source status, fine grained control, and high fidelity place Gazebo in a unique position to become more than just a stepping stone between the drawing board and real hardware: data visualization, simulation of remote environments, and even reverse engineering of blackbox systems are all possible applications. Gazebo is developed in cooperation with the Player and Stage projects (Gerkey, B. P., et al., July 2003), (Gerkey, B. P., et al., May 2001), (Vaughan, R. T., et al., Oct. 2003), and is available from http://playerstage.sourceforge.net/gazebo/ gazebo.html.
|
2fa2af72590819d7a4be995baa9809060f9c815a
|
Many important problems involve clustering large datasets. Although naive implementations of clustering are computationally expensive, there are established efficient techniques for clustering when the dataset has either (1) a limited number of clusters, (2) a low feature dimensionality, or (3) a small number of data points. However, there has been much less work on methods of efficiently clustering datasets that are large in all three ways at once—for example, having millions of data points that exist in many thousands of dimensions representing many thousands of clusters. We present a new technique for clustering these large, highdimensional datasets. The key idea involves using a cheap, approximate distance measure to efficiently divide the data into overlapping subsets we call canopies. Then clustering is performed by measuring exact distances only between points that occur in a common canopy. Using canopies, large clustering problems that were formerly impossible become practical. Under reasonable assumptions about the cheap distance metric, this reduction in computational cost comes without any loss in clustering accuracy. Canopies can be applied to many domains and used with a variety of clustering approaches, including Greedy Agglomerative Clustering, K-means and Expectation-Maximization. We present experimental results on grouping bibliographic citations from the reference sections of research papers. Here the canopy approach reduces computation time over a traditional clustering approach by more than an order of magnitude and decreases error in comparison to a previously used algorithm by 25%.
|
9ebecb9581185af467e8c770254a269244513d4b
|
Graph mining and management has become a popular area of research in recent years because of its numerous applications in a wide variety of practical fields, including computational biology, software bug localization and computer networking. Different applications result in graphs of different sizes and complexities. Correspondingly, the applications have different requirements for the underlying mining algorithms. In this chapter, we will provide a survey of different kinds of graph mining and management algorithms. We will also discuss a number of applications, which are dependent upon graph representations. We will discuss how the different graph mining algorithms can be adapted for different applications. Finally, we will discuss important avenues of future research
|
3d7348c63309ddb68b4e69782bc6bf516bb1ced7
|
This paper presents a high step-down tranformerless single-stage single-switch ac/dc converter suitable for universal line applications (90-270 Vrms) . The topology integrates a buck-type power-factor correction (PFC) cell with a buck-boost dc/dc cell and part of the input power is coupled to the output directly after the first power processing. With this direct power transfer feature and sharing capacitor voltages, the converter is able to achieve efficient power conversion, high power factor, low voltage stress on intermediate bus (less than 130 V) and low output voltage without a high step-down transformer. The absence of transformer reduces the component counts and cost of the converter. Unlike most of the boost-type PFC cell, the main switch of the proposed converter only handles the peak inductor current of dc/dc cell rather than the superposition of both inductor currents. Detailed analysis and design procedures of the proposed circuit are given and verified by experimental results.
|
0889019b395890f57bfae3ce7d8391649ae68de4
|
The large scale of Q&A archives accumulated in community based question answering (CQA) servivces are important information and knowledge resource on the web. Question and answer matching task has been attached much importance to for its ability to reuse knowledge stored in these systems: it can be useful in enhancing user experience with recurrent questions. In this paper, a Word Embedding based Correlation (WEC) model is proposed by integrating advantages of both the translation model and word embedding. Given a random pair of words, WEC can score their co-occurrence probability in Q&A pairs, while it can also leverage the continuity and smoothness of continuous space word representation to deal with new pairs of words that are rare in the training parallel text. An experimental study on Yahoo! Answers dataset and Baidu Zhidao dataset shows this new method’s promising
|
31ace8c9d0e4550a233b904a0e2aabefcc90b0e3
|
Face representation is a crucial step of face recognition systems. An optimal face representation should be discriminative, robust, compact, and very easyto-implement. While numerous hand-crafted and learning-based representations have been proposed, considerable room for improvement is still present. In this paper, we present a very easy-to-implement deep learning framework for face representation. Our method bases on a new structure of deep network (called Pyramid CNN). The proposed Pyramid CNN adopts a greedy-filter-and-down-sample operation, which enables the training procedure to be very fast and computationefficient. In addition, the structure of Pyramid CNN can naturally incorporate feature sharing across multi-scale face representations, increasing the discriminative ability of resulting representation. Our basic network is capable of achieving high recognition accuracy (85.8% on LFW benchmark) with only 8 dimension representation. When extended to feature-sharing Pyramid CNN, our system achieves the state-of-the-art performance (97.3%) on LFW benchmark. We also introduce a new benchmark of realistic face images on social network and validate our proposed representation has a good ability of generalization.
|
4281046803e75e1ad7144bc1adec7a3757de7e8d
|
Recent years have witnessed significant progress in example-based texture synthesis algorithms. Given an example texture, these methods produce a larger texture that is tailored to the user’s needs. In this state-of-the-art report, we aim to achieve three goals: (1) provide a tutorial that is easy to follow for readers who are not already familiar with the subject, (2) make a comprehensive survey and comparisons of different methods, and (3) sketch a vision for future work that can help motivate and guide readers that are interested in texture synthesis research. We cover fundamental algorithms as well as extensions and applications of texture synthesis.
|
21d470547b836d6e561a1cc86f24bbb6d1ee83b1
|
Fundamental movement competency is essential for participation in physical activity and for mitigating the risk of injury, which are both key elements of health throughout life. The squat movement pattern is arguably one of the most primal and critical fundamental movements necessary to improve sport performance, to reduce injury risk and to support lifelong physical activity. Based on current evidence, this first (1 of 2) report deconstructs the technical performance of the back squat as a foundation training exercise and presents a novel dynamic screening tool that incorporates identification techniques for functional deficits that limit squat performance and injury resilience. The follow-up report will outline targeted corrective methodology for each of the functional deficits presented in the assessment tool.
|
3037897d2fd1cc72dfc5a5b79cf9c0e8cdae083e
|
Semantic analysis of multimodal video aims to index segments of interest at a conceptual level. In reaching this goal, it requires an analysis of several information streams. At some point in the analysis these streams need to be fused. In this paper, we consider two classes of fusion schemes, namely early fusion and late fusion. The former fuses modalities in feature space, the latter fuses modalities in semantic space. We show by experiment on 184 hours of broadcast video data and for 20 semantic concepts, that late fusion tends to give slightly better performance for most concepts. However, for those concepts where early fusion performs better the difference is more significant.
|
fc211e6b7a112982bd96a9aa04144a0a06e86a97
|
This paper proposes an interleaved zero voltage transition (ZVT) PWM buck converter. The proposed converter consists of two identical buck converter modules and an auxiliary circuit. The auxiliary circuit provides zero voltage switching condition for the main switches. Also, zero current switching condition for the auxiliary switch and diodes is achieved. The interleaved buck converter with the proposed auxiliary switch is analyzed. The simulation results of a 300W ZVT interleaved buck converter operating at 100 KHz are presented to justify the validity of theoretical analysis.
|
0407d72c2e773aec18a4be6e2bcbdf1f91f032bb
|
This paper asks how complexity theory informs the role of leadership in organizations. Complexity theory is a science of complexly interacting systems; it explores the nature of interaction and adaptation in such systems and how they influence such things as emergence, innovation, and fitness. We argue that complexity theory focuses leadership efforts on behaviors that enable organizational effectiveness, as opposed to determining or guiding effectiveness. Complexity science broadens conceptualizations of leadership from perspectives that are heavily invested in psychology and social psychology (e.g., human relations models) to include processes for managing dynamic systems and interconnectivity. We develop a definition of organizational complexity and apply it to leadership science, discuss strategies for enabling complexity and effectiveness, and delve into the relationship between complexity theory and other currently important leadership theories. The paper concludes with a discussion of possible implications for research strategies in the social
|
2d59338108d3333890089305f15a60b6e5f00c54
|
Your use of the JSTOR archive indicates your acceptance of JSTOR's Terms and Conditions of Use, available at http://www.jstor.org/page/info/about/policies/terms.jsp. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained prior permission, you may not download an entire issue of a journal or multiple copies of articles, and you may use content in the JSTOR archive only for your personal, non-commercial use.
|
3813fade6b111f08636ad220ef32bd95b57d3e03
|
Theoretical performance limits of twophase switched-capacitor (SC) dc-dc converters are discussed in this paper. For a given number of capacitors k, the complete set of attainable dc conversion ratios is found. The maximum step-up or stepdown ratio is given by the k t h Fibonacci number, while the bound on the number of switches required in any SC circuit is 3k 2. Practical implications, illustrated by several SC converter examples, include savings in the number of components required for a given application, and the ability to construct SC converters that can maintain the output voltage regulation and high conversion efficiency over a wide range of input voltage variations. Limits found for the output resistance and efflciency can be used for selection and comparison of SC converters.
|
7cde4cf792f2be12deb8d5410170a003375397d5
|
The paper describes switched-capacitor dc-dc converters (charge pumps) suitable for on-chip, low-power applications. The proposed configurations are based on connecting two identical but opposite-phase SC converters in parallel, thus eliminating the need for separate bootstrap gate drivers. We focus on emerging very low-power VLSI applications such as batterypowered or self-powered signal processors where high power conversion efficiency is important and where power levels are in the milliwatt range. Conduction and switching losses are considered to allow design optimization in terms of switching frequency and component sizes. Open-loop and closed-loop operation of an experimental, fully integrated, 10MHz voltage doubler is described. The doubler has 2V or 3V input and generates 3.3V or 5V output at up to 5mW load. The converter circuit fabricated in a standard 1.2μ CMOS technology takes 0.7mm of the chip area.
|
bbad8eb4fd80411f99b08a5b2aa11eaed6b6f51a
|
A compact circularly polarized (CP) co-designed filtering antenna is reported. The device is based on a patch radiator seamlessly integrated with a bandpass filter composed of coupled stripline open-loop resonators, which are designed together as a system. In the proposed design, the patch functions simultaneously as the radiator and the last stage resonator of the filter, resulting in a low-profile integrated radiating and filtering module with a small overall form factor of 0.53λ0 × 0.53λ0 × 0.07λ0. It is shown that the filtering circuit not only ensures frequency selectivity but also provides impedance matching functionality, which serves to broaden both the impedance and axial ratio bandwidths. The designed filtering antenna was fabricated and measured, experimentally achieving an S11 <; -13.5 dB, an axial ratio of less than 3 dB and a gain higher than 5.2 dBi over a bandwidth from 3.77 to 4.26 GHz, i.e., around 12.2%, which makes it an excellent candidate for integration into a variety of wireless systems. A linearly polarized version of the integrated filtering antenna was also demonstrated. In addition, further full-wave simulations and experiments were carried out to verify that the designed CP filtering antenna maintains its properties even when mounted on different positions of the human body with various body gestures. The stable impedance and radiation properties also make it a suitable candidate as a wearable antenna for off-body wireless communications.
|
31d15e69389efeda5e21adfa888dffa8018523c0
| |
21e1a50ead66ac791db4ae9afd917f2b3adf28cc
|
Millimeter-wave MIMO systems are one of the candidate schemes for 5G wireless standardization efforts. In this context, the main contributions of this article are three-fold. First, we describe parallel sets of measurements at identical transmit-receive location pairs with 2.9, 29 and 61 GHz carrier frequencies in indoor office, shopping mall, and outdoor settings. These measurements provide insights on propagation, blockage and material penetration losses, and the key elements necessary in system design to make mm-Wave systems viable in practice. Second, one of these elements is hybrid beamforming necessary for better link margins by reaping the array gain with large antenna dimensions. From the class of fully-flexible hybrid beamformers, we describe a robust class of directional beamformers toward meeting the high data-rate requirements of mm-Wave systems. Third, leveraging these design insights, we then describe an experimental prototype system at 28 GHz that realizes high data rates on both the downlink and uplink and robustly maintains these rates in outdoor and indoor mobility scenarios. In addition to maintaining large signal constellation sizes in spite of radio frequency challenges, this prototype leverages the directional nature of the mm-Wave channel to perform seamless beam switching and handover across mm-Wave base stations, thereby overcoming the path losses in non-line-of-sight links and blockages encountered at mm-Wave frequencies.
|
ebbe56d235e0812d99609d4cff98473bfb5a7e33
|
The interworking of the LTE system and WLAN technologies has drawn much attention lately, due to the growing demands for various multimedia services and large data traffic in hotspot areas. Existing research studies have mostly investigated the coupling architectures for these two wireless communication standards at the network layer. However, in the current architectures, many important coordination functions and joint optimizations cannot be accomplished efficiently. To tackle this problem, a new CBS solution is proposed, which integrates different RATs at layer 2 in the true sense of convergence. We design a unified protocol stack that includes all the original functions of both LTE and WLAN systems. Then we propose a convergence architecture, the RMC sublayer, for joint management of these two RATs. The proposed CBS solution can support seamless offloading through soft handover, guaranteed QoS, forwarding management by a single IP address, and customized bandwidth aggregation service. Finally, our simulation and initial experiment results demonstrate the feasibility and efficiency of the CBS solution in future mobile converged networks.
|
3e5dbb5fd3460e020c321bc48b7d0fa6c3324dbc
|
Equivalent circuits for piezoelectric devices and ultrasonic transmission media can be used to cosimulate electronics and ultrasound parts in simulators originally intended for electronics. To achieve efficient system- level optimization, it is important to simulate correct, absolute amplitude of the ultrasound signal in the system, as this determines the requirements on the electronics regarding dynamic range, circuit noise, and power consumption. This paper presents methods to achieve correct, absolute amplitude of an ultrasound signal in a simulation of a pulse-echo system using equivalent circuits. This is achieved by taking into consideration loss due to diffraction and the effect of the cable that connects the electronics and the piezoelectric transducer. The conductive loss in the transmission line that models the propagation media of the ultrasound pulse is used to model the loss due to diffraction. Results show that the simulated amplitude of the echo follows measured values well in both near and far fields, with an offset of about 10%. The use of a coaxial cable introduces inductance and capacitance that affect the amplitude of a received echo. Amplitude variations of 60% were observed when the cable length was varied between 0.07 m and 2.3 m, with simulations predicting similar variations. The high precision in the achieved results show that electronic design and system optimization can rely on system simulations alone. This will simplify the development of integrated electronics aimed at ultrasound systems.
|
60afb1828de4efa1588401f87caa55ac3a0cd820
|
A 9-year-old girl developed hidradenitis suppurativa 3 months after the first signs of adrenarche. Such a close temporal relationship is consistent with the hypothesis that the disease is androgen dependent. Less than 2% of patients have onset of the disease before the age of 11 years. The exceptionally early age of onset in our patient may be partly explained by the fact that she had an early puberty.
|
7f60b70dede16fe4d7b412674929b4805c9b5c95
|
Hidradenitis suppurativa (HS) is a chronic suppurative scarring disease of apocrine sweat gland-bearing skin in the axillary, anogenital, and, rarely, the breast and scalp regions. Females are more commonly affected than males and it is usually seen at puberty or later. We report two girls with prepubertal hidradenitis suppurativa whose initial presentation predated any signs of puberty. This early onset is very rare and its etiology remains unknown. Severe disease can be seen in prepubertal children and surgical intervention is effective in these cases.
|
8b85f287c144aad5ff038ec0b140e0c4e210990b
|
IMPORTANCE
Hidradenitis suppurativa (HS) is a chronic debilitating cutaneous disease for which there is no universally effective treatment. Patients typically present at puberty with tender subcutaneous nodules that can progress to dermal abscess formation. Antiandrogens have been used in the treatment of HS, and studies have primarily focused on adult patients.
OBSERVATIONS
We present a case series of 3 pediatric patients with HS who were successfully treated with oral finasteride, resulting in decreased frequency and severity of disease flares with no significant adverse effects.
CONCLUSIONS AND RELEVANCE
Finasteride is a therapeutic option that provides benefit for pediatric patients with HS. Further prospective data and randomized controlled studies will provide helpful information in the management of this disease.
|
e72e10ad6228bd3dcee792f6f571c5ffed37266f
|
Hidradenitis suppurativa (HS) is a burdensome disease and has the potential to affect the life course of patients. It is a rare disease in children, and the recorded literature is correspondingly scarce. This article reviews the therapeutic options for HS in children and adolescents, and highlights particular differences or challenges with treating patients in this age group compared with adults. The work-up of paediatric patients with HS should include considerations of possible endocrine co-morbidities and obesity. Medical therapy of lesions may include topical clindamycin. Systemic therapy may include analgesics, clindamycin and rifampicin, finasteride, corticosteroids or tumour necrosis factor alpha (TNFα) blockers. Superinfections should be appropriately treated. Scarring lesions generally require surgery.
|
f4f6ae0b10f2cc26e5cf75b4c39c70703f036e9b
|
BACKGROUND
Although skin diseases are often immediately visible to both patients and society, the morbidity they cause is only poorly defined. It has been suggested that quality-of-life measures may be a relevant surrogate measure of skin disease. Hidradenitis suppurativa (HS) leads to painful eruptions and malodorous discharge and is assumed to cause a significant degree of morbidity. The resulting impairment of life quality has not previously been quantitatively assessed, although such an assessment may form a pertinent measure of disease severity in HS.
OBJECTIVES
To measure the impairment of life quality in patients with HS.
METHODS
In total, 160 patients suffering from HS were approached. The following data were gathered: quality-of-life data (Dermatology Life Quality Index, DLQI questionnaire), basic demographic data, age at onset of the condition and the average number of painful lesions per month.
RESULTS
One hundred and fourteen patients participated in the study. The mean +/- SD age of the patients was 40.9 +/- 11.7 years, the mean +/- SD age at onset 21.8 +/- 9.9 years and the mean +/- SD duration of the disease 18.8 +/- 11.4 years. Patients had a mean +/- SD DLQI score of 8.9 +/- 8.3 points. The highest mean score out of the 10 DLQI questions was recorded for question 1, which measures the level of pain, soreness, stinging or itching (mean 1.55 points, median 2 points). Patients experienced a mean of 5.1 lesions per month.
CONCLUSIONS
HS causes a high degree of morbidity, with the highest scores obtained for the level of pain caused by the disease. The mean DLQI score for HS was higher than for previously studied skin diseases, and correlated with disease intensity as expressed by lesions per month. This suggests that the DLQI may be a relevant outcome measure in future therapeutic trials in HS.
|
2bf11b00938e73468e3ab02dfe678985212f1aea
|
Mobile location-based services are thriving, providing an unprecedented opportunity to collect fine grained spatio-temporal data about the places users visit. This multi-dimensional source of data offers new possibilities to tackle established research problems on human mobility, but it also opens avenues for the development of novel mobile applications and services. In this work we study the problem of predicting the next venue a mobile user will visit, by exploring the predictive power offered by different facets of user behavior. We first analyze about 35 million check-ins made by about 1 million Foursquare users in over 5 million venues across the globe, spanning a period of five months. We then propose a set of features that aim to capture the factors that may drive users' movements. Our features exploit information on transitions between types of places, mobility flows between venues, and spatio-temporal characteristics of user check-in patterns. We further extend our study combining all individual features in two supervised learning models, based on linear regression and M5 model trees, resulting in a higher overall prediction accuracy. We find that the supervised methodology based on the combination of multiple features offers the highest levels of prediction accuracy: M5 model trees are able to rank in the top fifty venues one in two user check-ins, amongst thousands of candidate items in the prediction list.
|
011dcf6b9fa8d64e508ecead47c1a9a9521a3e59
|
The growing volume and variety of data presents both opportunities and challenges for visual analytics. Addressing these challenges is needed for big data to provide valuable insights and novel solutions for business, security, social media, and healthcare. In the case of temporal event sequence analytics it is the number of events in the data and variety of temporal sequence patterns that challenges users of visual analytic tools. This paper describes 15 strategies for sharpening analytic focus that analysts can use to reduce the data volume and pattern variety. Four groups of strategies are proposed: (1) extraction strategies, (2) temporal folding, (3) pattern simplification strategies, and (4) iterative strategies. For each strategy, we provide examples of the use and impact of this strategy on volume and/or variety. Examples are selected from 20 case studies gathered from either our own work, the literature, or based on email interviews with individuals who conducted the analyses and developers who observed analysts using the tools. Finally, we discuss how these strategies might be combined and report on the feedback from 10 senior event sequence analysts.
|
fe4f36731311fa013cd083a7e3c961f392325afc
| |
bff411fd40bf3e40c9b4f61554dc3370710e2d43
|
With a stringent input-to-output conversion ratio (CR) of 20 (24V input and 1.2V output) for a DC-DC converter, two stage cascaded architectures are easy to implement, but suffer from poor efficiency and doubled number of power components. This paper presents a single-stage solution with a proposed Adaptive ON-OFF Time (AO2T) control. In steady state, the control works as an adaptive ON-time valley current mode control to accommodate the large CR. During load transient periods, both ON- and OFF-time are adaptively adjustable to instantaneous load change, in order to accomplish fast load transient response within one switching cycle. To facilitate the high speed current mode control, a sensorless current detection circuit is also proposed. Operating at 5MHz, the converter achieves a maximum efficiency of 89.8% at 700mA and an efficiency of 85% at 2A full load. During a load current slew rate of 1.8A/200ns, the undershoot/overshoot voltages at VO are 23mV and 37mV respectively.
|
e928564981b35eccc1035df3badf74de7611d9cc
|
This article presents an algorithm for inducing multiclass decision trees with multivariate tests at internal decision nodes. Each test is constructed by training a linear machine and eliminating variables in a controlled manner. Empirical results demonstrate that the algorithm builds small accurate trees across a variety of tasks.
|
ada53a115e1551f3fbad3dc5930c1187473a78a4
|
We introduce a new descriptor for images which allows the construction of efficient and compact classifiers with good accuracy on object category recognition. The descriptor is the output of a large number of weakly trained object category classifiers on the image. The trained categories are selected from an ontology of visual concepts, but the intention is not to encode an explicit decomposition of the scene. Rather, we accept that existing object category classifiers often encode not the category per se but ancillary image characteristics; and that these ancillary characteristics can combine to represent visual classes unrelated to the constituent categories’ semantic meanings. The advantage of this descriptor is that it allows object-category queries to be made against image databases using efficient classifiers (efficient at test time) such as linear support vector machines, and allows these queries to be for novel categories. Even when the representation is reduced to 200 bytes per image, classification accuracy on object category recognition is comparable with the state of the art (36% versus 42%), but at orders of magnitude lower computational cost.
|
e99bd9fd681d64d209a462692fe80d8e28e0db1c
|
Most research to date has approached employee empowerment as an individual-level phenomenon. In this study we proposed a work-unit-level construct, empowerment climate, and tested a multiple-level model integrating macro and micro approaches to empowerment. Empowerment climate was shown to be empirically distinct from psychological empowerment and positively related to manager ratings of work-unit performance. A cross-level mediation analysis using hierarchical linear modeling showed that psychological empowerment mediated the relationships between empowerment climate and individual performance and job satisfaction.
|
0786d19321c380f98ade66e4c9c8c9380ac89beb
|
Many hierarchical multi-label classification systems predict a real valued score for every (instance, class) couple, with a higher score reflecting more confidence that the instance belongs to that class. These classifiers leave the conversion of these scores to an actual label set to the user, who applies a cut-off value to the scores. The predictive performance of these classifiers is usually evaluated using threshold independent measures like precision-recall curves. However, several applications require actual label sets, and thus an automatic labelling strategy. In this article, we present and evaluate different alternatives to perform the actual labelling in hierarchical multi-label classification. We investigate the selection of both single and multiple thresholds. Despite the existence of multiple threshold selection strategies in non-hierarchical multi-label classification, they can not be applied directly to the hierarchical context. The proposed strategies are implemented within two main approaches: optimisation of a certain performance measure of interest (such as F-measure or hierarchical loss), and simulating training set properties (such as class distribution or label cardinality) in the predictions. We assess the performance of the proposed labelling schemes on 10 datasets from different application domains. Our results show that selecting multiple thresholds may result in ∗Corresponding author. Tel : +32(0)9 331 36 93 Fax : +32(0)9 221 76 73 Email addresses: [email protected] (Isaac Triguero), [email protected] (Celine Vens) Preprint submitted to Elsevier January 29, 2016 an efficient and effective solution for hierarchical multi-label problems.
|
5a459e764726c4ffceba51c76baa7f511ee5b1b8
|
Future computing environments will free the user from the constraints of the desktop. Applications for a mobile environment should take advantage of contextual information, such as position, to o er greater services to the user. In this paper, we present the Cyberguide project, in which we are building prototypes of a mobile context-aware tour guide. Knowledge of the user's current location, as well as a history of past locations, are used to provide more of the kind of services that we come to expect from a real tour guide. We describe the architecture and features of a variety of Cyberguide prototypes developed for indoor and outdoor use on a number of di erent hand-held platforms. We also discuss the general research issues that have emerged in our context-aware applications development in a mobile environment.
|
31b6b7a1e00ada40a674f0fa13fa695245058f97
|
Corresponding Author: Michelle Lim Sern Mi Institute of Microengineering and Nanoelectronics (IMEN), UKM, 43600 Bangi, Selangor, Malaysia Email: [email protected] Abstract: This paper reviews CMOS based charge pump topologies used within autonomous embedded micro-systems. These charge pump structures have evolved from its simplistic diode-tied, single-branches with major threshold drops to exponential type, dual-branches with sophisticated gate and substrate control for lower voltage operation. Published charge pumps are grouped based on architecture, operation principles and pump optimization techniques with their pros and cons compared and results contrasted. The various charge pump topologies and schemes used are considered based on pumping efficiency, power efficiency, charge transferability, circuit complexity, pumping capacitors, form factor and minimum supply voltages with an optimum load. This article concludes with an overview of suitable techniques and recommendations that will aid a designer in selecting the most suitable charge pump topology especially for low ambient micro energy harvesting applications.
|
958e19db35adf74d6f74e36f3ade4494bd9829f6
|
This paper presents the findings of a steerable higher order mode (TE<inline-formula> <tex-math notation="LaTeX">$^{\mathrm {y}}_{1\delta 3}$ </tex-math></inline-formula>) dielectric resonator antenna with parasitic elements. The beam steering was successfully achieved by switching the termination capacitor on the parasitic element. In this light, all of the dielectric resonator antennas (DRAs) have the same dielectric permittivity similar to that of ten and excited by a <inline-formula> <tex-math notation="LaTeX">$50\Omega $ </tex-math></inline-formula> microstrip with a narrow aperture. The effect of the mutual coupling on the radiation pattern and the reflection coefficient, as well as the array factor, was investigated clearly using MATLAB version 2014b and ANSYS HFSS version 16. As the result, the antenna beam of the proposed DRA array managed to steer from −32° to +32° at 15 GHz. Furthermore, the measured antenna array showed the maximum gain of 9.25 dBi and the reflection coefficients which are less than −10 dB with the bandwidth more than 1.3 GHz, which is viewed as desirable for device-to-device communication in 5G Internet of Things applications.
|
c60c6632548f09f066ccb693dd2e1738ca012d6c
|
This communication presents a 32 × 32 high-gain patch array antenna fed by the substrate integrated waveguide (SIW) structure at W-band. The array antenna consists of two layers to achieve a compact topology, which allows for mass production using a standard PCB fabrication process. The wideband feeding network is placed in the bottom layer while the radiating patches are on the top layer. This configuration also resolves the trade-off between gain and bandwidth of conventional SIW array antennas. Measured gain of the 32 × 32 antenna array is within the range 28.81-29.97 dBi in the working bandwidth of 91-97 GHz. Measured impedance bandwidth covers the same frequency band for . The cross-polarization of the antenna array is less than 40 dB at the beam direction. Good agreement between the simulated and measured results validates our design.
|
61f30c93c68064d8706ba7b6d3b57701bd1b9ffc
| |
5f8fa49eea09a43a6e6f6e7fdc4387751aee8973
|
The most common way of constructing a hash function (e.g., SHA-1) is to iterate a compression function on the input message. The compression function is usually designed from scratch or made out of a block-cipher. In this paper, we introduce a new security notion for hash-functions, stronger than collision-resistance. Under this notion, the arbitrary length hash function H must behave as a random oracle when the fixed-length building block is viewed as an ideal primitive. This enables to eliminate all possible generic attacks against iterative hash-functions. In this paper, we show that the current design principle behind hash functions such as SHA-1 and MD5 — the (strengthened) Merkle-Damg̊ard transformation — does not satisfy this security notion. We provide several constructions that provably satisfy this notion; those new constructions introduce minimal changes to the plain Merkle-Damg̊ard construction and are easily implementable in practice. This paper is a modified version of a paper to appear at Crypto 2005.
|
eea181af6fc81ac0897c79a8bdb1c2dcbe410863
|
As the development of mobile devices and applications, mobile privacy has become a very important issue. Current researches on mobile privacy mainly focus on potential leakages on a particular device. However, leakage of sensitive data on a mobile device not only violates the privacy of the phone (or data) owner, but also violates the privacy of many other people whose information are contained in the data directly or indirectly (they are called data involvers). To address such problems, we introduce a collaborative privacy management framework, which aims to provide fine-grained data privacy protection for both data owners and data involvers in a distributed manner. Based on individual privacy policies specified by each user, a collaborative privacy policy is generated and enforced on different devices automatically. As a proof-of-concept prototype, we implement the proposed framework on Android and demonstrate its applicability with two case studies.
|
8e04afb34228a7fbb3f6ef3af8cfe85e0e74c34b
|
To save resources for Internet of Things (IoT) devices, a proposed approach segments operands and corresponding basic arithmetic operations that can be carried out by approximate function units for almost all applications. The approach also increases the security of IoT devices by hiding information for IP watermarking, digital fingerprinting, and lightweight encryption.
|
27d2ee0a25f97137aaea666a1d39350cd7f1c4ba
|
doesn't need to be secure? Almost every softwarecontrolled system faces threats from potential adversaries, from Internet-aware client applications running on PCs, to complex telecommunications and power systems accessible over the Internet, to commodity software with copy protection mechanisms. Software engineers must be cognizant of these threats and engineer systems with credible defenses, while still delivering value to customers. In this paper, we present our perspectives on the research issues that arise in the interactions between software engineering and security.
|
7133ed98bfefa2e7a177f88e8c100562fca82b3a
|
The OpenStreetMap project is a knowledge collective that provides user-generated street maps. OSM follows the peer production model that created Wikipedia; its aim is to create a set of map data that's free to use, editable, and licensed under new copyright schemes. A considerable number of contributors edit the world map collaboratively using the OSM technical infrastructure, and a core group, estimated at approximately 40 volunteers, dedicate their time to creating and improving OSM's infrastructure, including maintaining the server, writing the core software that handles the transactions with the server, and creating cartographical outputs. There's also a growing community of software developers who develop software tools to make OSM data available for further use across different application domains, software platforms, and hardware devices. The OSM project's hub is the main OSM Web site.
|
b563b7e6e14661d488bb082bffe7c2837e56c022
|
OBJECTIVE
Our purpose was to describe the surgical procedure, its results, and its complications and to determine whether patients are satisfied with surgical reduction of labia minora in cases of hypertrophy.
STUDY DESIGN
The records of 163 patients who underwent reduction of the labia minora during a 9-year period were reviewed. The ages of the patients ranged from 12 to 67 years (median, 26). Motives for requesting surgery were aesthetic concerns in 87% of the cases, discomfort in clothing in 64%, discomfort with exercise in 26%, and entry dyspareunia in 43%. Anatomic results were assessed 1 month postoperatively. Patient satisfaction was assessed by means of a mailed questionnaire.
RESULTS
No surgery-related significant complications were noticed. Anatomic results were satisfactory for 151 patients (93%). Ninety-eight completed questionnaires were returned. Eighty-one patients (83%) found that the results after surgery were satisfactory. Eighty-seven (89%) were satisfied with the aesthetic result, and 91 (93%) approved the functional outcome. Four patients (4%) would not undergo the same procedure again.
CONCLUSION
Labia minora reduction is a simple surgical procedure associated with a high degree of patient satisfaction.
|
6e07e1390c6a6cd6ca52ac909e1db807d7ba19be
|
Flexible electronics are increasingly being used in a number of applications which benefit from their low profile, light weight, and favorable dielectric properties. However, despite these advantages, the range of practical, high-volume, applications for flexible electronics will remain limited in the future unless a number of challenges related to lithographic patterning on flexible substrates are successfully addressed. The most critical of these pertain to system parameters that affect the cost and performance of flexible circuits, including the resolution, panel size, process throughput, substrate distortion, material handling, and yield. We present a new class of roll-to-roll lithography systems, developed in recent years, that were designed to address the challenges in each of these critical areas. These systems provides high-resolution projection imaging over very large exposure areas, on flexible substrate materials. Additionally, they achieve high-precision alignment by means of image scaling to compensate for substrate distortion due to processing; and they also performs high-throughput photoablation, patterning millions of pixels simultaneously, by means of projection imaging. This technology is attractive for current and emerging applications, such as flexible circuit boards and flexible chip carriers, as well as for potential future applications such as such as flexible displays and macroelectronic systems.
|
2744ea7c1f495d97e0cfbecf3e6a315a34d71b6a
|
Technologies now exist for implementing dense surface-normal optical interconnections for silicon CMOS VLSI using hybrid integration techniques. The critical factors in determining the performance of the resulting photonic chip are the yield on the transceiver device arrays, the sensitivity and power dissipation of the receiver and transmitter circuits, and the total optical power budget available. The use of GaAs–AlGaAs multiple-quantum-well p-i-n diodes for on-chip detection and modulation is one effective means of implementing the optoelectronic transceivers. We discuss a potential roadmap for the scaling of this hybrid optoelectronic VLSI technology as CMOS linewidths shrink and the characteristics of the hybrid optoelectronic tranceiver technology improve. An important general conclusion is that, unlike electrical interconnects, such dense optical interconnections directly to an electronic circuit will likely be able to scale in capacity to match the improved performance of future CMOS technology.
|
2a748cc66531dd7f4d122e66cc0cb461d1205fc0
|
In this paper we present an approach to enhance existing maps with fine grained segmentation categories such as parking spots and sidewalk, as well as the number and location of road lanes. Towards this goal, we propose an efficient approach that is able to estimate these fine grained categories by doing joint inference over both, monocular aerial imagery, as well as ground images taken from a stereo camera pair mounted on top of a car. Important to this is reasoning about the alignment between the two types of imagery, as even when the measurements are taken with sophisticated GPS+IMU systems, this alignment is not sufficiently accurate. We demonstrate the effectiveness of our approach on a new dataset which enhances KITTI [8] with aerial images taken with a camera mounted on an airplane and flying around the city of Karlsruhe, Germany.
|
b833ee2d196180e11eb4d93f793acd66ff1e2dbe
|
Accurate metrical localization is one of the central challenges in mobile robotics. Many existing methods aim at localizing after building a map with the robot. In this paper, we present a novel approach that instead uses geo-tagged panoramas from the Google Street View as a source of global positioning. We model the problem of localization as a non-linear least squares estimation in two phases. The first estimates the 3D position of tracked feature points from short monocular camera sequences. The second computes the rigid body transformation between the Street View panoramas and the estimated points. The only input of this approach is a stream of monocular camera images and odometry estimates. We quantified the accuracy of the method by running the approach on a robotic platform in a parking lot by using visual fiducials as ground truth. Additionally, we applied the approach in the context of personal localization in a real urban scenario by using data from a Google Tango tablet.
|
0f8c30445f3d994ac220dd101de6999cb6eaf911
|
In the past, there has been a tremendous progress in the area of autonomous robot navigation and a large variety of robots have been developed who demonstrated robust navigation capabilities indoors, in non-urban outdoor environments, or on roads and relatively few approaches focus on navigation in urban environments such as city centers. Urban areas, however, introduce numerous challenges for autonomous robots as they are rather unstructured and dynamic. In this paper, we present a navigation system for mobile robots designed to operate in crowded city environments and pedestrian zones. We describe the different components of this system including a SLAM module for dealing with huge maps of city centers, a planning component for inferring feasible paths taking also into account the traversability and type of terrain, a module for accurate localization in dynamic environments, and means for calibrating and monitoring the platform. Our navigation system has been implemented and tested in several large-scale field tests, in which a real robot autonomously navigated over several kilometers in a complex urban environment. This also included a public demonstration, during which the robot autonomously traveled along a more than three kilometer long route through the city center of Freiburg, Germany.
|
1233f38bddaebafe9f4ae676bb2f8671f6c4821a
|
We describe linear-time algorithms for solving a class of problems that involve transforming a cost function on a grid using spatial information. These problems can be viewed as a generalization of classical distance transforms of binary images, where the binary image is replaced by an arbitrary function on a grid. Alternatively they can be viewed in terms of the minimum convolution of two functions, which is an important operation in grayscale morphology. A consequence of our techniques is a simple and fast method for computing the Euclidean distance transform of a binary image. Our algorithms are also applicable to Viterbi decoding, belief propagation, and optimal control. ACM Classification: F.2.1, I.4 AMS Classification: 68T45, 68W40
|
6e37979d2a910e8a2337927731619fd789a5213b
|
In this survey we discuss various approximation-theoretic problems that arise in the multilayer feedforward perceptron (MLP) model in neural networks. The MLP model is one of the more popular and practical of the many neural network models. Mathematically it is also one of the simpler models. Nonetheless the mathematics of this model is not well understood, and many of these problems are approximation-theoretic in character. Most of the research we will discuss is of very recent vintage. We will report on what has been done and on various unanswered questions. We will not be presenting practical (algorithmic) methods. We will, however, be exploring the capabilities and limitations of this model. In the rst two sections we present a brief introduction and overview of neural networks and the multilayer feedforward perceptron model. In Section 3 we discuss in great detail the question of density. When does this model have the theoretical ability to approximate any reasonable function arbritrarily well? In Section 4 we present conditions for simultaneously approximating a function and its derivatives. Section 5 considers the interpolation capability of this model. In Section 6 we study upper and lower bounds on the order of approximation of this model. The material presented in Sections 3{6 treats the single hidden layer MLP model. In Section 7 we discuss some of the diierences that arise when considering more than one hidden layer. The lengthy list of references includes many papers not cited in the text, but relevant to the subject matter of this survey.
|
60cd946e854e2adf256358d2e5e17b0459ba80c6
| |
dbe1cb8e323da5bf045b51534e7ed06e69ca53df
|
While increasingly complex approaches to question answering (QA) have been proposed, the true gain of these systems, particularly with respect to their expensive training requirements, can be in- flated when they are not compared to adequate baselines. Here we propose an unsupervised, simple, and fast alignment and informa- tion retrieval baseline that incorporates two novel contributions: a one-to-many alignment between query and document terms and negative alignment as a proxy for discriminative information. Our approach not only outperforms all conventional baselines as well as many supervised recurrent neural networks, but also approaches the state of the art for supervised systems on three QA datasets. With only three hyperparameters, we achieve 47% P@1 on an 8th grade Science QA dataset, 32.9% P@1 on a Yahoo! answers QA dataset and 64% MAP on WikiQA.
|
36fb553aa996885017afe3489a8377eceddc08ee
|
We present a novel approach for determining robot movements that efficiently accomplish the robot's tasks while not hindering the movements of people within the environment. Our approach models the goal-directed trajectories of pedestrians using maximum entropy inverse optimal control. The advantage of this modeling approach is the generality of its learned cost function to changes in the environment and to entirely different environments. We employ the predictions of this model of pedestrian trajectories in a novel incremental planner and quantitatively show the improvement in hindrance-sensitive robot trajectory planning provided by our approach.
|
35582a30685083c62dca992553eec44123be9d07
|
We study the construction of prediction algorithms in a situation in which a learner faces a sequence of trials with a prediction to be made in each and the goal of the learner is to make few mistakes We are interested in the case that the learner has reason to believe that one of some pool of known algorithms will perform well but the learner does not know which one A simple and e ective method based on weighted voting is introduced for constructing a compound algorithm in such a circumstance We call this method the Weighted Majority Algorithm We show that this algorithm is robust in the presence of errors in the data We discuss various versions of the Weighted Majority Algorithm and prove mistake bounds for them that are closely related to the mistake bounds of the best algorithms of the pool For example given a sequence of trials if there is an algorithm in the pool A that makes at most m mistakes then the Weighted Majority Algorithm will make at most c log jAj m mistakes on that sequence where c is xed constant
|
f017d6284b6526790dca6bff0bb0495231534e2a
|
A new type of metallic electromagnetic structure has been developed that is characterized by having high surface impedance. Although it is made of continuous metal, and conducts dc currents, it does not conduct ac currents within a forbidden frequency band. Unlike normal conductors, this new surface does not support propagating surface waves, and its image currents are not phase reversed. The geometry is analogous to a corrugated metal surface in which the corrugations have been folded up into lumped-circuit elements, and distributed in a two-dimensional lattice. The surface can be described using solid-state band theory concepts, even though the periodicity is much less than the free-space wavelength. This unique material is applicable to a variety of electromagnetic problems, including new kinds of low-profile antennas.
|
9e4291de6cdce8e6f247effa308d72e2ec3f6122
| |
2cb46d5cab5590ef9950bd303bdfae41e7a98b1a
|
Cloud computing promises a number of advantages for the deployment of data-intensive applications. One important promise is reduced cost with a pay-as-you-go business model. Another promise is (virtually) unlimited throughput by adding servers if the workload increases. This paper lists alternative architectures to effect cloud computing for database applications and reports on the results of a comprehensive evaluation of existing commercial cloud services that have adopted these architectures. The focus of this work is on transaction processing (i.e., read and update workloads), rather than analytics or OLAP workloads, which have recently gained a great deal of attention. The results are surprising in several ways. Most importantly, it seems that all major vendors have adopted a different architecture for their cloud services. As a result, the cost and performance of the services vary significantly depending on the workload.
|
98354bb3d015d684d9248589191367fd7069cbc6
| |
212fc5ddeb4416aa7e1435f4c69391d0ad4fb18d
|
We consider the problem of learning distributed representations for documents in data streams. The documents are represented as low-dimensional vectors and are jointly learned with distributed vector representations of word tokens using a hierarchical framework with two embedded neural language models. In particular, we exploit the context of documents in streams and use one of the language models to model the document sequences, and the other to model word sequences within them. The models learn continuous vector representations for both word tokens and documents such that semantically similar documents and words are close in a common vector space. We discuss extensions to our model, which can be applied to personalized recommendation and social relationship mining by adding further user layers to the hierarchy, thus learning user-specific vectors to represent individual preferences. We validated the learned representations on a public movie rating data set from MovieLens, as well as on a large-scale Yahoo News data comprising three months of user activity logs collected on Yahoo servers. The results indicate that the proposed model can learn useful representations of both documents and word tokens, outperforming the current state-of-the-art by a large margin.
|
b8e7dfa21aac846cb52848e54a68dd822ced20dd
|
Mainstream parallel algorithms for mining frequent itemsets (patterns) were designed by implementing FP-Growth or Apriori algorithms on MapReduce (MR) framework. Existing MR FP-Growth algorithms can not distribute data equally among nodes, and MR Apriori algorithms utilize multiple map/reduce procedures and generate too many key-value pairs with value of 1; these disadvantages hinder their performance. This paper proposes an algorithm FIMMR: it firstly mines local frequent itemsets for each data chunk as candidates, applies prune strategies to the candidates, and then identifies global frequent itemsets from candidates. Experimental results show that the time efficiency of FIMMR outperforms PFP and SPC significantly; and under small minimum support threshold, FIMMR can achieve one order of magnitude improvement than the other two algorithms; meanwhile, the speedup of FIMMR is also satisfactory.
|
c759c1f4376e322943d8a463064367fcee5f4eb6
|
We propose a scalable Gaussian process model for regression by applying a deep neural network as the feature-mapping function. We first pre-train the deep neural network with a stacked denoising auto-encoder in an unsupervised way. Then, we perform a Bayesian linear regression on the top layer of the pre-trained deep network. The resulting model, Deep-Neural-Network-based Gaussian Process (DNN-GP), can learn much more meaningful representation of the data by the finite-dimensional but deep-layered feature-mapping function. Unlike standard Gaussian processes, our model scales well with the size of the training set due to the avoidance of kernel matrix inversion. Moreover, we present a mixture of DNN-GPs to further improve the regression performance. For the experiments on three representative large datasets, our proposed models significantly outperform the state-of-the-art algorithms of Gaussian process regression.
|
ddaa3e20c1cd31c8f511bdaf5d84a80aa6660737
|
Robotic lower limb exoskeletons that can alter joint mechanical power output are novel tools for studying the relationship between the mechanics and energetics of human locomotion. We built pneumatically powered ankle exoskeletons controlled by the user's own soleus electromyography (i.e. proportional myoelectric control) to determine whether mechanical assistance at the ankle joint could reduce the metabolic cost of level, steady-speed human walking. We hypothesized that subjects would reduce their net metabolic power in proportion to the average positive mechanical power delivered by the bilateral ankle exoskeletons. Nine healthy individuals completed three 30 min sessions walking at 1.25 m s(-1) while wearing the exoskeletons. Over the three sessions, subjects' net metabolic energy expenditure during powered walking progressed from +7% to -10% of that during unpowered walking. With practice, subjects significantly reduced soleus muscle activity (by approximately 28% root mean square EMG, P<0.0001) and negative exoskeleton mechanical power (-0.09 W kg(-1) at the beginning of session 1 and -0.03 W kg(-1) at the end of session 3; P=0.005). Ankle joint kinematics returned to similar patterns to those observed during unpowered walking. At the end of the third session, the powered exoskeletons delivered approximately 63% of the average ankle joint positive mechanical power and approximately 22% of the total positive mechanical power generated by all of the joints summed (ankle, knee and hip) during unpowered walking. Decreases in total joint positive mechanical power due to powered ankle assistance ( approximately 22%) were not proportional to reductions in net metabolic power ( approximately 10%). The ;apparent efficiency' of the ankle joint muscle-tendon system during human walking ( approximately 0.61) was much greater than reported values of the ;muscular efficiency' of positive mechanical work for human muscle ( approximately 0.10-0.34). High ankle joint ;apparent efficiency' suggests that recoiling Achilles' tendon contributes a significant amount of ankle joint positive power during the push-off phase of walking in humans.
|
3f63b12887fd8d29bb6de64fce506b2a388ae3ed
|
E-commerce customer churn rate is high and the customer churn dataset is seriously imbalanced. In order to improve the prediction accuracy of churn customers as well as strengthen to identify non-churn customers, this paper presents e-commerce customer churn prediction model based on improved SMOTE and AdaBoost. First, processing the churn data with improved SMOTE, which combines oversampling and undersampling methods to address the imbalance problem and subsequently integrates AdaBoost algorithm to predict. Finally, the empirical study on B2C E-commerce platform proves that this model has better efficiency and accuracy compared with the mature customer churn prediction algorithms.
|
22e584677475a4a807b852a1ced75d5cdf24e23c
|
Arrau is a new corpus annotated for anaphoric relations, with information about agreement and explicit representation of multiple antecedents for ambiguous anaphoric expressions and discourse antecedents for expressions which refer to abstract entities such as events, actions and plans. The corpus contains texts from different genres: task-oriented dialogues from the Trains-91 and Trains-93 corpus, narratives from the English Pear Stories corpus, newspaper articles from the Wall Street Journal portion of the Penn Treebank, and mixed text from the Gnome corpus.
|
162d50e6e2c000baf10148f761cc0929aad48ca2
|
On-line analytical processing (OLAP) is a recent and important application of database systems. Typically, OLAP data is presented as a multidimensional \data cube." OLAP queries are complex and can take many hours or even days to run, if executed directly on the raw data. The most common method of reducing execution time is to precompute some of the queries into summary tables (subcubes of the data cube) and then to build indexes on these summary tables. In most commercial OLAP systems today, the summary tables that are to be precomputed are picked rst, followed by the selection of the appropriate indexes on them. A trial-and-error approach is used to divide the space available between the summary tables and the indexes. This two-step process can perform very poorly. Since both summary tables and indexes consume the same resource |space | their selection should be done together for the most e cient use of space. In this paper, we give algorithms that automate the selection of summary tables and indexes. In particular, we present a family of algorithms of increasing time complexities, and prove strong performance bounds for them. The algorithms with higher complexities have better performance bounds. However, the increase in the performance bound is diminishing, and we show that an algorithm of moderate complexity can perform fairly close to the optimal. This work was supported by NSF grant IRI{92{23405, ARO grant DAAH04{95{1{0192, and Air Force Contract F33615{93{ 1{1339. Present address of V. Harinarayan and A. Rajaraman: Junglee Corp., Palo Alto, CA.
|
6b509a872a23abf233cf212303eae31eed6e02c7
|
A complete radar transceiver for 77 GHz with two integrated antenna elements is presented. Based on a previously published design [1], two of the transmit and receive channels of the transceiver are supplemented with integrated antenna elements. The antennas exhibit a well defined antenna pattern with an efficiency of better than 50%.
|
afeee08b2f21f2408efd1af442cbdca545b86e3d
|
We analyze the value-loading problem. This is the problem of robustly encoding moral values into an AI agent interacting with a complex environment. Like many before, we argue that this is both a major concern and an extremely challenging problem. Solving it will likely require years, if not decades, of multidisciplinary work by teams of top scientists and experts. Given how uncertain the timeline of human-level AI research is, we thus argue that a pragmatic partial solution should be designed as
|
28b39a2e9ab23f1ea7ebd0ef9a239dab37b51cca
|
Matrix factorization has been recently utilized for the task of multi-modal hashing for cross-modality visual search, where basis functions are learned to map data from different modalities to the same Hamming embedding. In this paper, we propose a novel cross-modality hashing algorithm termed Supervised Matrix Factorization Hashing (SMFH) which tackles the multi-modal hashing problem with a collective non-matrix factorization across the different modalities. In particular, SMFH employs a well-designed binary code learning algorithm to preserve the similarities among multi-modal original features through a graph regularization. At the same time, semantic labels, when available, are incorporated into the learning procedure. We conjecture that all these would facilitate to preserve the most relevant information during the binary quantization process, and hence improve the retrieval accuracy. We demonstrate the superior performance of SMFH on three cross-modality visual search benchmarks, i.e., the PASCAL-Sentence, Wiki, and NUS-WIDE, with quantitative comparison to various stateof-the-art methods [Kumar and Udupa, 2011; Rastegari et al., 2013; Zhang and Li, 2014; Ding et al., 2014].
|
7d07425a77f3042264be780bc8319a6fd8134502
|
The recent growth of IoT use-cases in a wide array of industrial, utility and environmental applications has necessitated the need for connectivity solutions with diverse requirements. Connectivity through BLE, Zigbee and 6LoPAN are examples of short-range IoT deployments. But to provide connectivity to a high density of devices over larger coverage areas, Low-Power Wide-Area Network (LPWAN) technologies in both licensed as well as unlicensed bands have been considered. In this paper, we consider modelling the traffic from IoT devices connected through LPWAN technologies. Due to diverse applications of IoT, it is not trivial to have a single traffic model to represent all of them, but the traffic can be broadly classified as either periodic, event-triggered, or a combination of both. We evaluate the performance of LoRaWAN, one such LPWAN technology, in the presence of a hybrid of both traffic types, where the event propagates spatially over time. In a practical deployment of sensor based IoT devices, the devices are usually densely deployed to ensure sufficient & reliable measurement. Thereby, when an event occurs, they exhibit spatial & temporal correlation in their traffic rate due to the natural phenomena of the metric they measure. We use the CMMPP model to represent such characteristic traffic from independent IoT devices triggered by an event. The characteristics of LoRa, the physical layer of LoRaWAN, is abstracted based on required signal strength and interference thresholds for different modulation parameters. Through system simulations, we demonstrate that there is a significant performance hit in LoRaWAN based networks, during the occurrence of events. In particular, using the packet delivery rate (PDR) as the metric, we found that while the system was able to handle regular updates from the devices with a PDR > 80%, event-driven traffic nearly impaired the network causing the PDR to drop below 10%.
|
88fc1622d9964c43b28810c7db5eedf6d26c2447
|
We propose a novel semi supervised, Multi Level Sequential Generative Adversarial Network (MLS-GAN ) architecture for group activity recognition. In contrast to previous works which utilise manually annotated individual human action predictions, we allow the models to learn it’s own internal representations to discover pertinent subactivities that aid the final group activity recognition task. The generator is fed with person-level and scene-level features that are mapped temporally through LSTM networks. Action-based feature fusion is performed through novel gated fusion units that are able to consider long-term dependancies, exploring the relationships among all individual actions, to learn an intermediate representation or ‘action code’ for the current group activity. The network achieves it’s semi-supervised behaviour by allowing it to perform group action classification together with the adversarial real/fake validation. We perform extensive evaluations on different architectural variants to demonstrate the importance of the proposed architecture. Furthermore, we show that utilising both person-level and scene-level features facilitates the group activity prediction better than using only person-level features. Our proposed architecture outperforms current state-of-the-art results for sports and pedestrian based classification tasks on Volleyball and Collective Activity datasets, showing it’s flexible nature for effective learning of group activities. 1
|
d278da6edddd56001c991a48279422b9a623d8ce
| |
9c13d3584332e3670b73b119d6661bccb10e240e
|
One of the main orientations in power electronics in the last decade has been the development of switching-mode converters without inductors and transformers. Light weight, small size and high power density are the result of using only switches and capacitors in the power stage of these converters. Thus, they serve as ideal power supplies for mobile electronic systems (e.g. cellular phones, personal digital assistants, and so forth). Switched-capacitor (SC) converters, with their large voltage conversion ratio, promise to be a response to such challenges of the 21st century as high-efficiency converters with low EMI emissions and the ability to realize steep step-down of the voltage (to 3V or even a smaller supply voltage for integrated circuits) or steep step-up of the voltage for automotive industry or internet services in the telecom industry. This paper is a tutorial of the main results in SC-converter research and design. Switched-Capacitor Converters— A Typical Power Electronics Contribution of the CAS Society From all of the research subjects in the Power Systems and Power Electronics Circuits Technical Committee—the stability of power systems, analysis of power electronics circuits (which are time-variable circuits with internally-controlled switches), chaos, modeling and simulation of converters, hardswitching and soft-switching converters, and so forth—the SCconverter is probably the most pertinent contribution of the CAS Society to power electronics. In fact, most of the researchers in this field belong to our Society’s technical committee and most of the breakthrough contributions on this subject have appeared in CAS publications. The reason for this is clear. Pursuing small size filters in the 1950’s, the circuit theory community found a solution in the elimination of bulky inductors from the structure of a passive filter. Active and then switched-capacitor filters showed the possibility of implementing the filtering function without using magnetic devices. The power electronics designer faced a similar challenge: the request for miniaturization of the power supply. This could be achieved only by eliminating the inductors and transformers. In a converter, the inductor serves the two purposes of processing energy and filtering the output voltage. The use of a switched-capacitor circuit for processing energy seemed doomed to failure because it is known that charging a capacitor from zero is achieved at a 50% efficiency. Ten years of research were needed to overcome the efficiency dilemma and to develop high-efficiency switched-capacitor energy-processing circuits. First SC Converters and Basic Principles The primary goal of any switching-mode power converter is to provide a constant (DC or AC) output voltage at its load, despite variations in the input voltage or load. A control element, therefore, has to be introduced in the process of transmitting energy so that the converter (power stage) changes its topology cyclically, and the durations of the switching topologies are adjusted for regulation purposes. The first SC converters were developed by a group of researchers from Kumamoto, Japan, who processed a DC unregulated voltage toward a DC regulated voltage [1, 2, 4]. These DC-DC converters were soon followed by AC-DC converters [3], DC-AC inverters [5] and AC-AC transformers [6]. By avoiding magnetic elements, these circuits, realized in a hybrid technology, featured a high power density (23W/inch). Figure 1(a). Basic SC step-down DC-DC converter. *A. Ioinovici is with the Department of Electrical and Electronics Engineering, Holon Academic Institute of Technology, Israel. R V S
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.