Based on current literature and cutting-edge advances in the machine learning field, there are four algorithms whose usage in new application domains must be explored: neural networks, rule induction algorithms, tree-based algorithms, and density-based algorithms. A number of machine learning related algorithms have been derived from these four algorithms. Consequently, they represent excellent underlying methods for extracting hidden knowledge from unstructured data, as essential data mining tasks. Implementation of Machine Learning Algorithms Using Control-Flow and Dataflow Paradigms presents widely used data-mining algorithms and explains their advantages and disadvantages, their mathematical treatment, applications, energy efficient implementations, and more. It presents research of energy efficient accelerators for machine learning algorithms. Covering topics such as control-flow implementation, approximate computing, and decision tree algorithms, this book is an essential resource for computer scientists, engineers, students and educators of higher education, researchers, and academicians.
Advances in Computers, Volume 126 presents innovations in computer hardware, software, theory, design and applications, with this updated volume including new chapters on VLSI for Super-Computing: Creativity in R+D from Applications and Algorithms to Masks and Chips, Bulk Bitwise Execution Model in Memory: Mechanisms, Implementation, and Evaluation, Embracing the Laws of Physics: Three Reversible Models of Computation, WSNs in Environmental Monitoring: Data Acquisition and Dissemination Aspects, Energy efficient implementation of tensor operations using dataflow paradigm for machine learning, and A Run-Time Job Scheduling Algorithm for Cluster Architectures with DataFlow Accelerators. Contains novel subject matter that is relevant to computer science Includes the expertise of contributing authorsPresents an easy to comprehend writing style
Innovation is the key to maintain competitive advantage. Innovation in products, processes, and business models help companies to provide economic value to their customers. Identifying the innovative ideas, implementing those ideas, and absorbing them in the market requires investing many resources that could incur large costs. Technology encourages companies to foster innovation to remain competitive in the marketplace. Emerging Technologies for Innovation Management in the Software Industry serves as a resource for technology absorption in companies supporting innovation. It highlights the role of technology to assist software companies—especially small start-ups—to innovate their products, processes, and business models. This book provides the necessary guidelines of which tools to use and under what situations. Covering topics such as risk management, prioritization approaches, and digitally-enabled innovation processes, this premier reference source is an ideal resource for entrepreneurs, software developers, software managers, business leaders, engineers, students and faculty of higher education, researchers, and academicians.
The most powerful computers work by harnessing the combined computational power of millions of processors, and exploiting the full potential of such large-scale systems is something which becomes more difficult with each succeeding generation of parallel computers. Alternative architectures and computer paradigms are increasingly being investigated in an attempt to address these difficulties. Added to this, the pervasive presence of heterogeneous and parallel devices in consumer products such as mobile phones, tablets, personal computers and servers also demands efficient programming environments and applications aimed at small-scale parallel systems as opposed to large-scale supercomputers. This book presents a selection of papers presented at the conference: Parallel Computing (ParCo2017), held in Bologna, Italy, on 12 to 15 September 2017. The conference included contributions about alternative approaches to achieving High Performance Computing (HPC) to potentially surpass exa- and zetascale performances, as well as papers on the application of quantum computers and FPGA processors. These developments are aimed at making available systems better capable of solving intensive computational scientific/engineering problems such as climate models, security applications and classic NP-problems, some of which cannot currently be managed by even the most powerful supercomputers available. New areas of application, such as robotics, AI and learning systems, data science, the Internet of Things (IoT), and in-car systems and autonomous vehicles were also covered. As always, ParCo2017 attracted a large number of notable contributions covering present and future developments in parallel computing, and the book will be of interest to all those working in the field.
A key focus in recent years has been on sustainable development and promoting environmentally conscious practices. In today’s rapidly evolving technological world, it is important to consider how technology can be applied to solve problems across disciplines and fields in these areas. Further study is needed in order to understand how technology can be applied to sustainability and the best practices, considerations, and challenges that follow. Futuristic Trends for Sustainable Development and Sustainable Ecosystems discusses recent advances and innovative research in the area of information and communication technology for sustainable development and covers practices in several artificial intelligence fields such as knowledge representation and reasoning, natural language processing, machine learning, and the semantic web. Covering topics such as blockchain, deep learning, and renewable energy, this reference work is ideal for computer scientists, industry professionals, researchers, academicians, scholars, instructors, and students.
Advances in Computers, Volume 116, presents innovations in computer hardware, software, theory, design, and applications, with this updated volume including new chapters on Teaching Graduate Students How to Review Research Articles and How to Respond to Reviewer Comments, ALGATOR - An Automatic Algorithm Evaluation System, Graph Grammar Induction, Asymmetric Windows in Digital Signal Processing, Intelligent Agents in Games: Review With an Open-Source Tool, Using Clickstream Data to Enhance Reverse Engineering of Web Applications, and more. Contains novel subject matter that is relevant to computer science Includes the expertise of contributing authors Presents an easy to comprehend writing style
This book constitutes the refereed post-conference proceedings of 13 workshops held at the 33rd International ISC High Performance 2018 Conference, in Frankfurt, Germany, in June 2018: HPC I/O in the Data Center, HPC-IODC 2018; Workshop on Performance and Scalability of Storage Systems, WOPSSS 2018; 13th Workshop on Virtualization in High-Performance Cloud Computing, VHPC 2018; Third International Workshop on In Situ Visualization, WOIV 2018; 4th International Workshop on Communication Architectures for HPC, Big Data, Deep Learning and Clouds at Extreme Scale, ExaComm 2018; International Workshop on OpenPOWER for HPC, IWOPH 2018; IXPUG Workshop: Many-Core Computing on Intel Processors; Workshop on Sustainable Ultrascale Computing Systems; Approximate and Transprecision Computing on Emerging Technologies, ATCET 2018; First Workshop on the Convergence of Large-Scale Simulation and Artificial Intelligence; Third Workshop for Open Source Supercomputing, OpenSuCo 2018; First Workshop on Interactive High-Performance Computing; Workshop on Performance Portable Programming Models for Accelerators, P^3MA 2018. The 53 full papers included in this volume were carefully reviewed and selected from 80 submissions. They cover all aspects of research, development, and application of large-scale, high performance experimental and commercial systems. Topics include HPC computer architecture and hardware; programming models, system software, and applications; solutions for heterogeneity, reliability, power efficiency of systems; virtualization and containerized environments; big data and cloud computing; and artificial intelligence.
This illuminating text/reference reviews the fundamentals of programming for effective DataFlow computing. The DataFlow paradigm enables considerable increases in speed and reductions in power consumption for supercomputing processes, yet the programming model requires a distinctly different approach. The algorithms and examples showcased in this book will help the reader to develop their understanding of the advantages and unique features of this methodology. This work serves as a companion title to DataFlow Supercomputing Essentials: Research, Development and Education, which analyzes the latest research in this area, and the training resources available. Topics and features: presents an implementation of Neural Networks using the DataFlow paradigm, as an alternative to the traditional ControlFlow approach; discusses a solution to the three-dimensional Poisson equation, using the Fourier method and DataFlow technology; examines how the performance of the Binary Search algorithm can be improved through implementation on a DataFlow architecture; reviews the different way of thinking required to best configure the DataFlow engines for the processing of data in space flowing through the devices; highlights how the DataFlow approach can efficiently support applications in big data analytics, deep learning, and the Internet of Things. This indispensable volume will benefit all researchers interested in supercomputing in general, and DataFlow computing in particular. Advanced undergraduate and graduate students involved in courses on Data Mining, Microprocessor Systems, and VLSI Systems, will also find the book to be an invaluable resource.
This useful text/reference describes the implementation of a varied selection of algorithms in the DataFlow paradigm, highlighting the exciting potential of DataFlow computing for applications in such areas as image understanding, biomedicine, physics simulation, and business. The mapping of additional algorithms onto the DataFlow architecture is also covered in the following Springer titles from the same team: DataFlow Supercomputing Essentials: Research, Development and Education, DataFlow Supercomputing Essentials: Algorithms, Applications and Implementations, and Guide to DataFlow Supercomputing. Topics and Features: introduces a novel method of graph partitioning for large graphs involving the construction of a skeleton graph; describes a cloud-supported web-based integrated development environment that can develop and run programs without DataFlow hardware owned by the user; showcases a new approach for the calculation of the extrema of functions in one dimension, by implementing the Golden Section Search algorithm; reviews algorithms for a DataFlow architecture that uses matrices and vectors as the underlying data structure; presents an algorithm for spherical code design, based on the variable repulsion force method; discusses the implementation of a face recognition application, using the DataFlow paradigm; proposes a method for region of interest-based image segmentation of mammogram images on high-performance reconfigurable DataFlow computers; surveys a diverse range of DataFlow applications in physics simulations, and investigates a DataFlow implementation of a Bitcoin mining algorithm. This unique volume will prove a valuable reference for researchers and programmers of DataFlow computing, and supercomputing in general. Graduate and advanced undergraduate students will also find that the book serves as an ideal supplementary text for courses on Data Mining, Microprocessor Systems, and VLSI Systems.
High Performance Computing (HPC) remains a driver that offers huge potentials and benefits for science and society. However, a profound understanding of the computational matters and specialized software is needed to arrive at effective and efficient simulations. Dedicated software tools are important parts of the HPC software landscape, and support application developers. Even though a tool is by definition not a part of an application, but rather a supplemental piece of software, it can make a fundamental difference during the development of an application. Such tools aid application developers in the context of debugging, performance analysis, and code optimization, and therefore make a major contribution to the development of robust and efficient parallel software. This book introduces a selection of the tools presented and discussed at the 9th International Parallel Tools Workshop held in Dresden, Germany, September 2-3, 2015, which offered an established forum for discussing the latest advances in parallel tools.
This book presents the lecture notes of the 1st Summer School on Methods and Tools for the Design of Digital Systems, 2015, held in Bremen, Germany. The topic of the summer school was devoted to modeling and verification of cyber-physical systems. This covers several aspects of the field, including hybrid systems and model checking, as well as applications in robotics and aerospace systems. The main chapters have been written by leading scientists, who present their field of research, each providing references to introductory material as well as latest scientific advances and future research directions. This is complemented by short papers submitted by the participating PhD students.