Welcome to Loot.co.za!
Sign in / Register |Wishlists & Gift Vouchers |Help | Advanced search
|
Your cart is empty |
|||
Books > Computing & IT > Computer programming > Programming languages
This book constitutes the thoroughly refereed post-conference proceedings of the Second International Conference on Data Engineering and Management, ICDEM 2010, held in Tiruchirappalli, India, in July 2010. The 46 revised full papers presented together with 1 keynote paper and 2 tutorial papers were carefully reviewed and selected from numerous submissions. The papers are organized in topical sections on Digital Library; Knowledge and Mulsemedia; Data Management and Knowledge Extraction; Natural Language Processing; Workshop on Data Mining with Graphs and Matrices.
This volume contains the proceedings of the 4th International Conference on Verified Software: Theories, Tools, and Experiments, VSTTE 2012, held in Philadelphia, PA, USA, in January 2012. The 20 revised full papers presented together with 2 invited talks and 2 tutorials were carefully revised and selected from 54 initial submissions for inclusion in the book. The goal of the VSTTE conference is to advance the state of the art through the interaction of theory development, tool evolution, and experimental validation. The papers address topics such as: specification and verification techniques, tool support for specification languages, tool for various design methodologies, tool integration and plug-ins, automation in formal verification, tool comparisons and benchmark repositories, combination of tools and techniques, customizing tools for particular applications, challenge problems, refinement methodologies, requirements modeling, specification languages, specification/verification case-studies, software design methods, and program logic.
This book constitutes the thoroughly refereed post-proceedings of the 20th International Workshop on Implementation and Applications of Functional Languages, IFL 2008, held in Hatfield, UK, in September 2008. The 15 revised full papers presented were carefully reviewed and selected from 31 submissions. Topics of interest cover a wide range from novel language designs, theoretical underpinnings, compilation and optimisation techniques for diverse hardware architectures, to applications, programming techniques and novel tools.
The focus in development methodologies of large and complex software systems has switched in the last two decades from functional issues to structural issues; this holds for both the object-oriented and the more recent component-based software engineering paradigms. Formal methods have been applied successfully to the verification of medium-sized programs in protocol and hardware design for quite a long time. However, their application to the development of large systems requires more emphasis on specification, modeling and validation techniques supporting the concepts of reusability and modifiability, and their implementation in new extensions of existing programming languages like Java. This state-of-the-art survey presents the outcome of the 9th Symposium on Formal Methods for Components and Objects, held in Graz, Austria, in November/December 2010. The volume contains 20 revised contributions submitted after the symposium by speakers from each of the following European IST projects: the FP7-IST project AVANTSSAR on automated validation of trust and security of service-oriented architectures; the FP7-IST project DEPLOY on industrial deployment of advanced system engineering methods for high productivity and dependability; the ESF-COST Action IC0701 on formal verification of object-oriented software; the FP7-IST project HATS on highly adaptable and trustworthy software using formal models; the FP7-SST project INESS on an integrated European railway signalling system; the FP7-IST project MADES on a model-driven approach to improve the current practice in the development of embedded systems; the FP7-IST project MOGENTES on model-based generation of tests for dependable embedded systems; as well as the FP7-IST project MULTIFORM on integrated multi-formalism tool support for the design of networked embedded control systems.
This book constitutes the refereed proceedings of the 23rd IFIP WG 6.1 International Conference on Testing Software and Systems, ICTSS 2011, held in Paris, France, in November 2011. The 13 revised full papers presented together with 2 invited talks were carefully selected from 40 submissions. The papers address the conceptual, theoretic, and practical problems of testing software systems, including communication protocols, services, distributed platforms, middleware, controllers, and security infrastructures.
This book constitutes the refereed proceedings of the 13th International Conference on Formal Engineering Methods, ICFEM 2011, held in Durham, UK, October 2011. The 40 revised full papers together with 3 invited talks presented were carefully reviewed and selected from 103 submissions. The papers address all current issues in formal methods and their applications in software engineering. They are organized in topical sections on formal models; model checking and probability; specification and development; security; formal verification; cyber physical systems; event-B; verification, analysis and testing; refinement; as well as theorem proving and rewriting.
This book constitutes the refereed proceedings of the 49th International Conference on Objects, Models, Components, Patterns, held in Zurich, Switzerland, in June 2011. The 19 revised full papers presented together with the abstracts of 2 invited papers were carefully reviewed and selected from a total of 68 submissions. The papers discuss all aspects of object technology and related fields, in particular model-based development, component-based development, language implementation and patterns, in a holistic way. The conference has a strong practical bias, without losing sight of the importance of correctness and performance.
This book constitutes the refereed proceedings of the 5th International Conference on Tests and Proofs, TAP 2011, held in Zurich, Switzerland in June/July 2011. The 12 revised full papers presented together with 2 invited papers were carefully reviewed and selected from 27 submissions. Among the topics covered are model checking, testing systems, test generation, symbolic testing, SAT solvers, SMT solvers, property-based testing, automated test generation, learning-based testing, UML, OCL, specification-based testing, and network testing.
This book constitutes the throughly refereed post-proceedings of the 6th International Workshop on Systems Analysis and Modeling, SAM 2010, held in collocation with MODELS 2010 in Oslo, Norway in October 2010. The 15 revised full papers presented went through two rounds of reviewing and improvement. The papers are organized in topical sections on modularity, composition, choreography, application of SDL and UML; SDL language profiles; code generation and model transformations; verification and analysis; and user requirements notification.
Advanced Data Science and Analytics with Python enables data scientists to continue developing their skills and apply them in business as well as academic settings. The subjects discussed in this book are complementary and a follow-up to the topics discussed in Data Science and Analytics with Python. The aim is to cover important advanced areas in data science using tools developed in Python such as SciKit-learn, Pandas, Numpy, Beautiful Soup, NLTK, NetworkX and others. The model development is supported by the use of frameworks such as Keras, TensorFlow and Core ML, as well as Swift for the development of iOS and MacOS applications. Features: Targets readers with a background in programming, who are interested in the tools used in data analytics and data science Uses Python throughout Presents tools, alongside solved examples, with steps that the reader can easily reproduce and adapt to their needs Focuses on the practical use of the tools rather than on lengthy explanations Provides the reader with the opportunity to use the book whenever needed rather than following a sequential path The book can be read independently from the previous volume and each of the chapters in this volume is sufficiently independent from the others, providing flexibility for the reader. Each of the topics addressed in the book tackles the data science workflow from a practical perspective, concentrating on the process and results obtained. The implementation and deployment of trained models are central to the book. Time series analysis, natural language processing, topic modelling, social network analysis, neural networks and deep learning are comprehensively covered. The book discusses the need to develop data products and addresses the subject of bringing models to their intended audiences - in this case, literally to the users' fingertips in the form of an iPhone app. About the Author Dr. Jesus Rogel-Salazar is a lead data scientist in the field, working for companies such as Tympa Health Technologies, Barclays, AKQA, IBM Data Science Studio and Dow Jones. He is a visiting researcher at the Department of Physics at Imperial College London, UK and a member of the School of Physics, Astronomy and Mathematics at the University of Hertfordshire, UK.
This book constitutes the refereed proceedings of the 13th IFIP WG 6.1 International Conference on Formal Methods for Open Object-Based Distributed Systems, FMOODS 2011, and the 31st IFIP WG 6.1 Formal Techniques for Networked and Distributed Systems, FORTE 2011, held in Reykjavik, Island, in June 2011, as one of the DisCoTec 2011 events. The 21 revised full papers presented together with one invited talk were carefully reviewed and selected from 65 submissions. The topics covered are formal verification, formal modeling and specification, run-time monitoring, and testing to address challenges in many different application areas, including dynamic and ad hoc networks, mobile and adaptive computation, reactive and timed systems, business processes, and distributed and concurrent systems and algorithms.
This book presents a comprehensive documentation of the scientific outcome of 14 satellite events held at the 13th International Conference on Model-Driven Engineering, Languages and Systems, MODELS 2010, held in Oslo, Norway, in October 2010. Besides the 21 revised best papers selected from 12 topically focused workshops, the post-proceedings also covers the doctoral symposium and the educators symposium; each of the 14 satellite events covered is introduced by a summary of the respective organizers. All relevant current aspects in model-based systems design and analysis are addressed. This book is the companion of the MODELS 2010 main conference proceedings LNCS 6394/6395.
Transition from a back-end developer to a full-stack developer with knowledge of all the dimensions of web application development, namely, front-end, back-end and server-side software. This book provides a comprehensive overview of Streamlit, allowing developers and programmers of all backgrounds to get up to speed in as little time as possible. Streamlit is a pure Python web framework that will bridge the skills gap and shorten development time from weeks to hours. This book walks you through the complete cycle of web application development, from an introductory to advanced level with accompanying source code and resources. You will be exposed to developing basic, intermediate, and sophisticated user interfaces and subsequently you will be acquainted with data visualization, database systems, application security, and cloud deployment in Streamlit. In a market with a surplus demand for full stack developers, this skill set could not possibly come at a better time. In one sentence, Streamlit is a means for the empowerment of developers everywhere and all stand to gain from it. What You'll Learn Mutate big data in real-time Visualize big data interactively Implement web application security and privacy protocols Deploy Streamlit web applications to the cloud using Streamlit, Linux and Windows servers Who is this Book for? Developers with solid programming experience wanting to learn Streamlit; Back-end developers looking to upskill and transition to become a full-stack developers; Those who wish to learn and become more acquainted with data visualization, database systems, security and cloud deployment with Steamlit
Data Science and Analytics with Python is designed for practitioners in data science and data analytics in both academic and business environments. The aim is to present the reader with the main concepts used in data science using tools developed in Python, such as SciKit-learn, Pandas, Numpy, and others. The use of Python is of particular interest, given its recent popularity in the data science community. The book can be used by seasoned programmers and newcomers alike. The book is organized in a way that individual chapters are sufficiently independent from each other so that the reader is comfortable using the contents as a reference. The book discusses what data science and analytics are, from the point of view of the process and results obtained. Important features of Python are also covered, including a Python primer. The basic elements of machine learning, pattern recognition, and artificial intelligence that underpin the algorithms and implementations used in the rest of the book also appear in the first part of the book. Regression analysis using Python, clustering techniques, and classification algorithms are covered in the second part of the book. Hierarchical clustering, decision trees, and ensemble techniques are also explored, along with dimensionality reduction techniques and recommendation systems. The support vector machine algorithm and the Kernel trick are discussed in the last part of the book. About the Author Dr. Jesus Rogel-Salazar is a Lead Data scientist with experience in the field working for companies such as AKQA, IBM Data Science Studio, Dow Jones and others. He is a visiting researcher at the Department of Physics at Imperial College London, UK and a member of the School of Physics, Astronomy and Mathematics at the University of Hertfordshire, UK, He obtained his doctorate in physics at Imperial College London for work on quantum atom optics and ultra-cold matter. He has held a position as senior lecturer in mathematics as well as a consultant in the financial industry since 2006. He is the author of the book Essential Matlab and Octave, also published by CRC Press. His interests include mathematical modelling, data science, and optimization in a wide range of applications including optics, quantum mechanics, data journalism, and finance.
C++ is capable of tackling a whole range of programming tasks. The purpose of this book is to give breadth and depth to C++ programmers’ existing experience of the language by representing a large number of algorithms, most of them implemented as ready-to-run (and standalone) programs. The programs are as readable as possible without sacrificing too great a degree of efficiency, generality, portability and robustness. Both the classes and programs are designed to demonstrate major programming principles. There is coverage of two key language features - templates and exception handling - apart from which the reader is assumed to have working knowledge of C++. Besides traditional subjects, such as quicksort and binary trees, this book also covers some less well-known topics, including multi-precision arithmetic, route planning and external sorting. Demonstration programs for these and many other exciting applications are based on C++ classes which you can also use in programs of your own.
This book constitutes the refereed proceedings of the 14th International Conference on Fundamental Approaches to Software Engineering, FASE 2011, held in Saarbrucken, Germany, March 26 April 3, 2011, as part of ETAPS 2011, the European Joint Conferences on Theory and Practice of Software. The 29 revised full papers presented together with one full length invited talk were carefully reviewed and selected from 99 full paper submissions. The papers are organized in topical sections on verification, specification and modeling, reachability and model checking, model driven engineering, software development for QoS, testing: theory and new trends, testing in practice, code development and analysis, and empirical studies."
This volume constitutes the thoroughly refereed post-conference proceedings of the 10th International Workshop on Agent-Oriented Software Engineering, AOSE 2009, held in Budapest, Hungary, in May 2009 as part of AAMAS 2009, the 8th International Conference on Autonomous Agents and Multiagent Systems. The 10 revised full papers presented were carefully selected from numerous initial submissions during two rounds of reviewing and improvement. The papers have been organized into three sections on multi-agent organizations, concrete development techniques, and - one step higher - going beyond the concrete technique and proposing a development method for designing concrete types of systems. This state-of-the-art survey is rounded off by five additional lectures addressing key areas in development: agent-oriented modelling languages, implementation of MAS, testing of MAS, software processes, and formal methods for the development of MAS. They permit analysis of the current state in the generation of specifications of MAS, the way these specifications can be implemented, how they can be validated, and what steps are necessary to do so.
Transactions on HiPEAC aims at the timely dissemination of research contributions in computer architecture and compilation methods for high-performance embedded computer systems. Recognizing the convergence of embedded and general-purpose computer systems, this journal publishes original research on systems targeted at specific computing tasks as well as systems with broad application bases. The scope of the journal therefore covers all aspects of computer architecture, code generation and compiler optimization methods of interest to researchers and practitioners designing future embedded systems. This third issue contains 14 papers carefully reviewed and selected out of numerous submissions and is divided into four sections. The first section contains the top four papers from the Third International Conference on High-Performance Embedded Architectures and Compilers, HiPEAC 2008, held in Goeteborg, Sweden, in January 2008. The second section consists of four papers from the 8th MEDEA Workshop held in conjunction with PACT 2007 in Brasov, Romania, in September 2007. The third section contains two regular papers and the fourth section provides a snapshot from the First Workshop on Programmability Issues for Multicore Computers, MULTIPROG, held in conjunction with HiPEAC 2008.
This book constitutes the refereed proceedings of the 12th International Conference on Verification, Model Checking, and Abstract Interpretation, VMCAI 2011, held in Austin, TX, USA, in January 2011, co-located with the Symposium on Principles of Programming Languages, POPL 2011. The 24 revised full papers presented together with 4 invited talks were carefully reviewed and selected from 71 initial submissions. The papers showcases state-of-the-art research in areas such as verification, model checking, abstract interpretation and address any programming paradigm, including concurrent, constraint, functional, imperative, logic and object-oriented programming. Further topics covered are static analysis, deductive methods, program certification, debugging techniques, abstract domains, type systems, and optimization.
This volume contains the proceedings of the Eighth International Workshop on Rewriting Logic and its Applications (WRLA 2010) that was held in Paphos, Cyprus, March 20-21, 2010, as a satellite workshop of the European Joint C- ferences on Theory and Practice of Software (ETAPS 2010). Rewriting logic is a natural semantic framework for representing conc- rency,parallelism,communicationandinteraction,aswellasbeing anexpressive (meta)logical framework for representing logics. It can then be used for spe- fying a wide range of systems and programming languages in various appli- tion ?elds. In recent years, several executable speci?cation languages based on rewriting logic (ASF+SDF, CafeOBJ, ELAN, Maude) have been designed and implemented. The aim of the WRLA workshop series is to bring together - searchers with a common interest in rewriting logic and its applications, and to give them the opportunity to present their recent works, discuss future research directions, and exchange ideas. Previous WRLA workshops were held in Asilomar (1996), Pont-a-Mousson ' (1998), Kanazawa (2000), Pisa (2002), Barcelona (2004), Vienna (2006), and Budapest (2008), and their proceedings have been published inElectronic Notes in Theoretical Computer Science. In addition, selected papers from WRLA 1996 have been published in a special issue of Theoretical Computer Science,and selected papers from WRLA 2004 appeared in a special issue of Higher-Order and Symbolic Computation.
The 8th IFIP Workshop on Software Technologies for Embedded and Ubiq- tous Systems (SEUS 2010) in Waidhofen/Ybbs, Austria, October 13-15, 2010, succeeded the seven previous workshops in Newport Beach, USA (2009); Capri, Italy (2008); Santorini, Greece (2007); Gyeongju, Korea (2006); Seattle, USA (2005); Vienna, Austria (2004); and Hokodate, Japan (2003); installing SEUS as a successfully established workshop in the ?eld of embedded and ubiquitous systems. SEUS 2010 continued the tradition of fostering cross-community scienti?c excellence and establishing strong links between research and industry. SEUS 2010 provided a forum where researchers and practitioners with substantial - periences and serious interests in advancing the state of the art and the state of practice in the ?eld of embedded and ubiquitous computing systems gathered with the goal of fostering new ideas, collaborations, and technologies. The c- tributions in this volume present advances in integrating the ?elds of embedded computing and ubiquitous systems. The call for papers attracted 30 submissions from all around the world. Each submission was assigned to at least four members of the Program Committee for review. The Program Committee decided to accept 21 papers, which were arranged in eight sessions. The accepted papers are from Austria, Denmark, France, Germany, Italy, Japan, Korea, Portugal, Taiwan, UK, and USA. Two keynotes complemented the strong technical program.
Wearedelightedtopresenttheproceedingsofthe5thInternationalWorkshopon Modeling Autonomic Communication Environments (MACE 2010). This wo- shopwasheldaspartofthe6thInternationalConferenceonNetworkandService Management (CNSM 2010), formerly known as and building on the success of the MANWEEK conference series. This year we met just a hundred yards away from Niagara Falls in Canada, a very exciting location. MACE started as an experiment and over the past years has created a small yet very active community that convened again this year to discuss and ev- uate new advances, innovative ideas, and solid developments. The main focus of MACE, combining modeling with communications, is certainly a hard topic that requires a lot of discussion, thus the work presented at the workshop is - trinsically debatable and might not be as practiced as in other well-established workshops, but this was the nature of MACE from the beginning. New ideas, sometimes more,sometimes less rougharoundthe edges (and someof them even inside) are submitted and provoke extensive discussions. The ?eld in which we areworkingreliesonthesediscussions,orevenadventures,andwehavethis year again strongly motivated and supported a variety of novel work in the technical program. This year, the submissions, while being closely related to the main themes, brought some new areas into the workshop. We still see architectural design and theapplicationofautonomicprinciplestonetworksandservices,butwealsonow have submissions looking into previously unexplored areas such as Home Area Networks,multimedia streaming,virtualization,federation,anduserexperience. This portrays a maturity in the domain, which has by now gone through several cycles, and improves its outputs by applying the lessons learned.
These proceedings contain the papers presented at the 8th Internationl S- posium on Automated Technology for Veri?cation and Analysis held during September 21-24, 2010 in Singapore. The primary objective of the ATVA c- ferences remains the same: to exchange and promote the latest advances of state-of-the-art research on theoretical and practical aspects of automated an- ysis, veri?cation and synthesis. From 72 papers submitted to ATVA 2010 in response to our call for papers, the Program Committee accepted 21 regular papers and 9 tool papers. Each paper received at least three reviews. The Program Committee worked hard to ensure that every submission received a rigorous and fair evaluation, with the ?nalprogramselectedaftera10-dayonlinediscussionsviatheEasychairsystem. OurprogramalsoincludedthreekeynotetalksandinvitedtutorialsbyThomas A.Henzinger(ISTAustria),JoxanJa?ar(NationalUniversityofSingapore)and IgorWalukiewicz(CNRS, France).Theconferenceorganizersweretrulygrateful to have such distinguished researchers as keynote speakers for the symposium. A new feature for the ATVA symposium this year were the two co-located workshops, In?nity 2010 (co-chaired by Yu-Fang Chen and Ahmed Rezine) and PMCW 2010 (co-chaired by Jun Sun and Hai Wang). We are delighted with the expanded scope, interactions and depth that the two workshops helped bring to the symposium. Many people worked hard and o?ered their valuable time so generously to make ATVA 2010 successful. First and foremost, we would like to thank all authors who worked hard to complete and submit papers to the conference. The ProgramCommittee members, reviewersand Steering Committee members alsodeservespecialrecognition.Without them, a competitive andpeer-reviewed international symposium simply cannot take place.
The First International Workshop on FOundational and Practical Aspects of Resource Analysis (FOPARA) was initiated to serve as a forum for presenting originalresearchresultsthatarerelevanttothe analysisofresource(time,space) consumption by computer programs. FOPARA aimed to bring together the - searchers working on foundational issues with the researchers focusing more on practical results. Therefore, both theoretical and practical contributions were encouraged.The contributions coveredthe following topics: resourceanalysisfor embedded systems, logical and machine-independent characterizations of c- plexity classes, logics closely related to complexity classes, type systems for c- trolling complexity, semantic methods to analyze resources,including quasi- and sup-interpretations, practical applications of resource analysis, etc. This ?rst FOPARA brought two di?erent groups of researchers together. In 2006 and 2008 informal application-oriented resource analysis workshops (an EmBoundedOpen WorkshopinBudapest,2006,andaResourceAnalysisWo- shop in Hertfordshire, 2008) were held as a?liated events of the International Symposium on the Implementation and Application of Functional Languages (IFL). Participants in these workshops were the University of St. Andrew (UK), Heriot-Watt University of Edinburgh (UK), Ludwig Maximilians University of Munich (Germany), Complutense University of Madrid (Spain) and the Po- technical University of Madrid (Spain). Another group of researchers was active in the series of informal workshops on implicit computational complexity (see, for instance, WICC 2008 in Paris). That series gathers researchers working in theoretical foundations of resource analysis, mainly from France (universities of ParisDiderotandParisNord,LORIANancy),Italy(universitiesofBolognaand Turin), Norway, Germany and Portugal.
The 16th annual International Conference on the Principles and Practice of Constraint Programming (CP 2010) was held in St. Andrews, Scotland, during September 6-10, 2010. We would like to thank our sponsors for their generous support of this event. This conference is concerned with all aspects of computing with constraints, including:theory,algorithms,applications,environments,languages,modelsand systems. We received a wide variety of submissions, each of which was reviewed by at least three referees. Referees were chosen for each submission by an initial bidding process where Program Committee members chose papers from their area of interest. The range of expertise represented by the large Program C- mittee meant that almost all submissions were reviewed by subject experts on the Program Committee, or by colleagues chosen by members of the Program Committee for their particular expertise. Papers weresolicitedeither as long (15 page), or short (8 page) submissions. Short-paper submissions were refereed to exactly the same high standards as long-paper submissions but naturally were expected to contain a smaller quantity of new material. Thus there is no disti- tion in these proceedings between short and long papers. I used the excellent EasyChair conference management system to support this process of reviewing, and for the collation and organization of these proceedings. Submissions were made either to the applications track or to the research track. Therewere101(23short)researchtracksubmissionsofwhich36(8short) wereaccepted,whichisa36%(35%ofshort)acceptancerate. Applicationstrack submissions received special consideration and the acceptance rate was sign- cantly higher than for the research track. |
You may like...
C++ How to Program: Horizon Edition
Harvey Deitel, Paul Deitel
Paperback
R1,857
Discovery Miles 18 570
VBA and Macros for Microsoft Office…
Bill Jelen, Tracy Syrstad
Paperback
R1,247
Discovery Miles 12 470
|